<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI Unlimited: AI Research Simplified]]></title><description><![CDATA[AI research papers simplified and explained]]></description><link>https://ai-unltd.com/s/ai-research-simplified</link><image><url>https://substackcdn.com/image/fetch/$s_!A0L1!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8255d44-7fec-4989-82c0-961bf24ab714_256x256.png</url><title>AI Unlimited: AI Research Simplified</title><link>https://ai-unltd.com/s/ai-research-simplified</link></image><generator>Substack</generator><lastBuildDate>Mon, 27 Apr 2026 11:38:34 GMT</lastBuildDate><atom:link href="https://ai-unltd.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[AI Unlimited]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[aiunltd@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[aiunltd@substack.com]]></itunes:email><itunes:name><![CDATA[Arjun]]></itunes:name></itunes:owner><itunes:author><![CDATA[Arjun]]></itunes:author><googleplay:owner><![CDATA[aiunltd@substack.com]]></googleplay:owner><googleplay:email><![CDATA[aiunltd@substack.com]]></googleplay:email><googleplay:author><![CDATA[Arjun]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Reduce GPT Costs with Prompt Compression]]></title><description><![CDATA[Prompt compression is a technique that reduces the length of inputs given to large language models (LLMs).]]></description><link>https://ai-unltd.com/p/reduce-gpt-costs-with-prompt-compression</link><guid isPermaLink="false">https://ai-unltd.com/p/reduce-gpt-costs-with-prompt-compression</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 11 Jul 2024 16:58:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!I0i_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I0i_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I0i_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I0i_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:359902,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I0i_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Prompt compression is a technique that reduces the length of inputs given to large language models (LLMs). It aims to maintain output quality while using fewer tokens. This is important because LLM APIs (such as OpenAI, Anthropic, etc.) charge based on the number of tokens processed.</p><p>LLMs break down text into tokens, which are chunks of characters. Each token has a cost associated with it. Longer prompts use more tokens, leading to higher costs. Prompt compression helps you stay within token limits and reduce processing time.</p><h2>Main Compression Techniques</h2><p>There are three primary methods for compressing prompts:</p><ol><li><p>Knowledge Distillation: This involves summarization or rewriting the sentences to reduce the number of tokens and make things more precise and short.<br><strong>Example:</strong><br>Original: "Explain the process of photosynthesis in detail, including the light-dependent and light-independent reactions, and how this process is crucial for life on Earth."<br>Compressed: "Describe photosynthesis: light reactions, dark reactions, importance for life."</p></li><li><p>Encoding: This technique transforms text into a format that we might not be able to comprehend easily, but LLMs can make sense of them.<br><strong>Example:</strong><br>Original: "The quick brown fox jumps over the lazy dog. This sentence contains every letter of the English alphabet."<br>Encoded: &#8220;VGhlIHF1aWNrIGJyb3duIGZveCBqdW1wcyBvdmVyIHRoZ&#8230;&#8221;</p></li><li><p>Filtering: This method removes unnecessary parts of prompts. It keeps only the most relevant information, reducing token count. This can be done at various levels, such as sentences, phrases, or tokens.<br><strong>Example:</strong><br>Original: "Can you please provide me with a comprehensive and detailed explanation of the fundamental principles of quantum mechanics, including its historical development and key concepts such as superposition, entanglement, and wave-particle duality?"<br>Filtered: "Explain quantum mechanics: principles, history, superposition, entanglement, wave-particle duality."</p></li></ol><p>Each of these techniques can significantly reduce the number of tokens in a prompt while preserving its core meaning. The choice of technique depends on your specific use case and the type of information you're working with.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Implementing Prompt Compression</h2><p>To implement prompt compression, you can use various tools and libraries. One popular option is <a href="https://www.microsoft.com/en-us/research/blog/llmlingua-innovating-llm-efficiency-with-prompt-compression/">LLMLingua</a>, developed by Microsoft. It uses advanced techniques to refine prompts into key components.</p><p>Here are examples of how you might implement basic versions of our three main compression techniques using Python.</p><h4><strong>Knowledge Distillation (Simplified example using a pre-trained model)</strong></h4><pre><code>from transformers import pipeline

def compress_prompt_distill(prompt, max_length=50):
    summarizer = pipeline("summarization", model="facebook/bart-large-cnn")
    summary = summarizer(prompt, max_length=max_length, min_length=10, do_sample=False)
    return summary[0]['summary_text']

original_prompt = "Explain the process of photosynthesis in detail, including the light-dependent and light-independent reactions, and how this process is crucial for life on Earth."
distilled_prompt = compress_prompt_distill(original_prompt)
print(f"Distilled prompt: {distilled_prompt}")</code></pre><h4><strong>Encoding (Using sentence embeddings)</strong></h4><pre><code>from sentence_transformers import SentenceTransformer

def compress_prompt_encode(prompt):
    model = SentenceTransformer('all-MiniLM-L6-v2')
    return model.encode(prompt)

original_prompt = "The quick brown fox jumps over the lazy dog."
encoded_prompt = compress_prompt_encode(original_prompt)
print(f"Encoded prompt (first 5 values): {encoded_prompt[:5]}")</code></pre><h4><strong>Filtering (Keyword Extraction)</strong></h4><pre><code>import nltk
from nltk.corpus import stopwords
from nltk.tokenize import word_tokenize

nltk.download('punkt')
nltk.download('stopwords')

def compress_prompt_filter(prompt, num_keywords=10):
    words = word_tokenize(prompt.lower())
    stop_words = set(stopwords.words('english'))
    keywords = [word for word in words if word.isalnum() and word not in stop_words]
    freq_dist = nltk.FreqDist(keywords)
    top_keywords = [word for word, _ in freq_dist.most_common(num_keywords)]
    return " ".join(top_keywords)

original_prompt = "Prompt compression is a technique used to optimize inputs given to large language models (LLMs) by reducing their length while maintaining output quality and relevance."
compressed_prompt = compress_prompt_filter(original_prompt)
print(f"Filtered prompt: {compressed_prompt}")</code></pre><p>When implementing prompt compression, keep these best practices in mind:</p><ul><li><p>Maintain a balance between compression and preserving essential information.</p></li><li><p>Test compressed prompts to ensure they still produce high-quality outputs.</p></li><li><p>Be aware that excessive compression can lead to loss of important nuances.</p></li></ul><p>Common challenges include handling diverse types of input data and ensuring consistent output quality. Overcome these by testing your compression methods on a variety of prompts and fine-tuning your approach.</p><p>Case studies often show significant cost reductions, sometimes up to 50% or more, without major impact on output quality. However, results can vary based on the specific use case and compression method.</p><p>The field of prompt compression is rapidly evolving. New techniques are emerging, such as dynamic compression ratios and context-aware compression. In the future, we may see prompt compression integrated directly into LLM architectures, making it even more efficient and effective.</p><p>By implementing prompt compression, you can significantly reduce your LLM token costs while maintaining the quality of your AI applications. Start with simple techniques and gradually explore more advanced methods as you become comfortable with the concept.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Tutorial: Chat with PDFs on your Google Drive]]></title><description><![CDATA[6 easy steps to chat with your document in Google Drive. It costs less than a ChatGPT subscription and is more secure.]]></description><link>https://ai-unltd.com/p/tutorial-chat-with-pdfs-in-google</link><guid isPermaLink="false">https://ai-unltd.com/p/tutorial-chat-with-pdfs-in-google</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Mon, 05 Feb 2024 19:00:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!crj2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!crj2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!crj2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!crj2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!crj2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:537200,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!crj2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!crj2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Unlock the ability to interact with PDFs in your Google Drive as if you're having a conversation with them. This simple guide is crafted for anyone who wants immediate results without delving into the technicalities. Get ready to chat with your documents!</p><div class="preformatted-block" data-component-name="PreformattedTextBlockToDOM"><label class="hide-text" contenteditable="false">Text within this block will maintain its original spacing when published</label><pre class="text">Find a sample google colab document all the way at the bottom of this article.</pre></div><h2>What You'll Need</h2><p>Before we begin, ensure you have the following:</p><ul><li><p>A PDF uploaded to your Google Drive.</p></li><li><p>A Google Colab account ready for use.</p></li><li><p>An <a href="https://platform.openai.com/api-keys">OpenAI account to access your API key</a>. Using OpenAI's API comes with costs, so please <a href="https://platform.openai.com/usage">monitor your usage</a> regularly.</p></li><li><p>And, you need to subscribe to this newsletter&#8230; Not mandatory, but please? &#128517;</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><h2>Setting Up Your Workspace</h2><h4>Step 1: Open Your Google Colab Notebook</h4><p>Visit <a href="https://colab.google/">Google Colab</a>, login with your google account and create a new colab notebook. That should look something like this. &#128071;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!djbW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!djbW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 424w, https://substackcdn.com/image/fetch/$s_!djbW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 848w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1272w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!djbW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png" width="1456" height="822" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:822,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:203875,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!djbW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 424w, https://substackcdn.com/image/fetch/$s_!djbW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 848w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1272w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Step 2: Connect to Google Drive</h4><p>Within your new notebook, look for the file icon on the left sidebar. Click on it and wait for a few seconds for the Colab environment to load.</p><p>Then click the Google Drive symbol to mount your drive. This should ask you a couple of permissions to enable you to connect with your google drive.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lUkT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lUkT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 424w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 848w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1272w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lUkT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png" width="1012" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1012,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:52459,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lUkT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 424w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 848w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1272w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Once this is done, you should see two things:</p><ol><li><p>A drive folder which is nothing but your google drive that&#8217;s connected to the colab notebook.</p></li><li><p>A piece of code that can be used to connect to your colab notebook. The code should look something like below.</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kj_C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kj_C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kj_C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png" width="1352" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1352,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:56063,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kj_C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now run the code by pressing the play button on the left side of the code. This might again ask you for a few permissions, but once it finishes running, your google drive will be connected with the colab notebook.</p><h4>Step 3: Installing Embedchain package</h4><p>Before we install, lets add a new cell by hovering on the centre of the last cell and clicking on &#8220;+ Code&#8221; button.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Brdc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Brdc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Brdc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png" width="1352" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/eca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1352,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:82334,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Brdc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"></figcaption></figure></div><p>Once we do that, let&#8217;s add the following code to the new cell and run it again by clicking the play button.</p><pre><code>!pip install embedchain</code></pre><p>This will generate a lot of output messages, wait for it to finish, scroll all the way to the bottom and look for the "Successfully installed" message to confirm everything went smoothly.</p><h4>Step 4: Getting Ready to Chat</h4><p>Import Packages and Load Your OpenAI API Key</p><p>Here&#8217;s where you bring in the tools you need. Add the following code to a new cell:</p><pre><code>import os
from embedchain import App

os.environ["OPENAI_API_KEY"] = "sk-yourapikey"
bot = App()</code></pre><p>Don't forget to replace <code>"sk-yourapikey"</code> with the actual key you get from your <a href="https://platform.openai.com/api-keys">OpenAI account</a>.</p><h4>Step 5: Adding Your PDF to the Conversation</h4><p>Add Your PDF to the Bot</p><p>With this step, you tell the bot about the PDF you want to chat with. Add the following code to a new cell and run it:</p><pre><code>bot.add('/content/drive/MyDrive/PDFs/grammy_awards.pdf', data_type='pdf_file')</code></pre><p>I have my PDF (grammy_awards.pdf) in a folder called &#8220;PDFs&#8221; in my google drive. Make sure to update the above code to match the path of your pdf in your google drive.</p><h4>Step 6: Let&#8217;s start the chat!</h4><p>Start Chatting with Your PDF!</p><p>It's showtime! Ask your PDF anything by adding the below code to a new cell and running it:</p><pre><code>bot.query('Who hosted the grammy awards of the year 2014?')</code></pre><p>Replace the above question with your own and you can start chatting with the PDF in your google drive!</p><p>Here&#8217;s a sample output:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UMTU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UMTU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 424w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 848w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1272w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UMTU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png" width="1456" height="304" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:304,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:55141,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UMTU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 424w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 848w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1272w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p><strong><a href="https://colab.research.google.com/drive/1h6W_tf47ebtXH8yfFXojVgVwp1qxjAY7?usp=sharing">Here is the link to a sample code on Google colab.</a></strong></p><h2>Conclusion</h2><p>You've just stepped into the future of document interaction. With these few steps, you can effortlessly extract information from your PDFs. Remember to monitor your OpenAI API usage to manage any associated costs.</p><p>Enjoy your new-found efficiency in data retrieval and happy chatting with your documents!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Rephrase and Respond (RaR): A New Way to Prompt ChatGPT for Accurate Responses]]></title><description><![CDATA[RaR methodology surpasses Chain-of-Thought in multiple benchmarks.]]></description><link>https://ai-unltd.com/p/rephrase-and-response-rar-a-new-way</link><guid isPermaLink="false">https://ai-unltd.com/p/rephrase-and-response-rar-a-new-way</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sun, 17 Dec 2023 19:58:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!X2Qi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!X2Qi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!X2Qi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1164141,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!X2Qi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Contents</h3><ol><li><p>Introduction</p></li><li><p>Understanding the Need for Better Questioning in LLMs</p></li><li><p>The RaR Method Explained <em>(With prompt examples)</em></p></li><li><p>Benefits of RaR in Enhancing LLM Responses</p></li><li><p>RaR vs. Chain-of-Thought (CoT) Method</p></li><li><p>Conclusion</p></li></ol><h2>Introduction</h2><p>In the evolving landscape of artificial intelligence, particularly with Large Language Models (LLMs) like ChatGPT, <strong>the method of interaction is crucial in eliciting accurate and meaningful responses</strong>. The Rephrase and Respond (RaR) method enhances the way we prompt these models by offering a new perspective on optimizing our dialogue with AI. By rephrasing questions and responses, <em>RaR aims to address common communication gaps, ensuring that LLMs like ChatGPT understand and reply with greater precision</em>. <strong>The best part is, that this method has been tested and benchmarked on GPT-4, one of the top LLMs out there.</strong></p><h2>Understanding the Need for Better Questioning in LLMs</h2><blockquote><p>The effectiveness of ChatGPT heavily relies on the quality of the prompts it receives. </p></blockquote><p><em>When posed with the query, &#8220;Was Mother Teresa born on an even month?&#8221; GPT-4 might mistakenly assert that August is an odd month.</em></p><p>Often, users face challenges in framing questions that elicit accurate and comprehensive responses. This gap arises due to the inherent limitations of LLMs in understanding nuanced or complex queries. <strong>Misinterpretations, lack of context, and ambiguous phrasing lead to suboptimal responses</strong>.</p><p>Rephrase and Respond (RaR) method emerges as a solution, aimed at refining the way questions are posed to LLMs. <strong>By focusing on the art of rephrasing, RaR addresses the core issue of communication breakdown, ensuring that the queries align more closely with the LLM's processing capabilities.</strong></p><h2>The RaR Method Explained <em>(With prompt examples)</em></h2><p>The RaR method elevates the effectiveness of LLMs like ChatGPT. <strong>It encompasses two distinct strategies: the one-step RaR and the two-step RaR.</strong></p><p><strong>One-Step RaR</strong>: <em>This involves the LLM rephrasing the user's query into a single, more precise question before responding</em>. This rephrasing aims to clarify the query's intent and ensure a correct understanding, leading to a more accurate answer.</p><pre><code>"{question}"
Rephrase and expand the question, and respond.</code></pre><p><a href="https://chat.openai.com/share/0ab795a3-d640-4e20-b841-89b45b5b306b">One-Step RaR ChatGPT Link</a></p><p><strong>Two-Step RaR</strong>: In this approach, the <em>LLM first rephrases the query and then, in a separate step, responds to this refined query</em>. This two-step process allows for even greater clarity and specificity in understanding and addressing the user's needs.</p><pre><code># Step 1

"{question}"
Given the above question, rephrase and expand it to help you
do better answering. Maintain all information in the original question. 

# Step 2

(original) "{question}"
(rephrased) "{rephrased_question}"
Use your answer for the rephrased question to answer the original question.</code></pre><p><a href="https://chat.openai.com/share/87fc6da5-d233-4342-83bb-b9927d9f8d96">Two-Step RaR ChatGPT Link</a></p><p><strong>The key advantage of RaR over traditional questioning methods is its focus on question clarity and precision</strong>.</p><p><em>With RaR, the LLM first seeks to disentangle the query's ambiguities, asking for clarifications or rephrasing it to grasp the user's true intent. This ensures that the response is more relevant and informative.</em></p><h2>Benefits of RaR in Enhancing LLM Responses</h2><blockquote><p>The RaR method significantly elevates the performance of ChatGPT.</p></blockquote><p>Key benefits include:</p><ol><li><p><strong>Improved Accuracy</strong>: RaR, in both its one-step and two-step forms, <em>substantially increases response accuracy in ChatGPT</em>. This is evident in the improved performance metrics across a range of tasks.</p></li><li><p><strong>Enhanced Relevance</strong>: Responses become more contextually relevant, as ChatGPT better grasps the specifics of the query through rephrasing.</p></li><li><p><strong>Increased Efficiency</strong>: RaR can streamline interactions, <em>particularly in cases where initial queries may be too vague or broad</em>.</p></li><li><p><strong>Versatility in Application</strong>: <em>RaR's adaptability makes it suitable for diverse applications, from simple Q&amp;A to complex problem-solving scenarios.</em></p></li></ol><p><a href="https://arxiv.org/abs/2311.04205">Research paper link.</a></p><h2>RaR vs. Chain-of-Thought (CoT) Method</h2><p>Comparing the RaR method with the Chain-of-Thought (CoT) approach reveals distinct features and applications:</p><ol><li><p><strong>Approach</strong>:</p><ul><li><p>RaR focuses on refining queries through <strong>rephrasing and iterative clarification</strong>, enhancing understanding before responding.</p></li><li><p>CoT involves the <strong>LLM explicating its reasoning process step-by-step</strong>, akin to a human solving a problem out loud.</p></li></ul></li><li><p><strong>Accuracy and Efficiency</strong>:</p><ul><li><p>RaR aims to <strong>increase accuracy by ensuring the question is well-understood before answering</strong>, which can be more efficient in obtaining precise information.</p></li><li><p>CoT, by <strong>elaborating the thought process, may provide deeper insights into complex problems</strong> but can be more time-consuming.</p></li></ul></li><li><p><strong>User Interaction</strong>:</p><ul><li><p>RaR <strong>encourages active user participation in refining the query</strong>, making it more interactive.</p></li><li><p>CoT is more AI-centric, with the <strong>model displaying its reasoning without direct user intervention</strong> in the thought process.</p></li></ul></li><li><p><strong>Applicability</strong>:</p><ul><li><p>RaR is <strong>versatile, and suitable for a wide range of queries</strong>, especially where precision and clarity are key.</p></li><li><p>CoT <strong>excels in scenarios requiring detailed explanations or step-by-step reasoning</strong>, such as complex problem-solving.</p></li></ul></li></ol><p>Each method has its strengths, and choosing between them depends on the specific needs of the interaction, whether it's <strong>clarity and precision (RaR) or detailed understanding and explanation (CoT)</strong>.</p><p><em>Additionally, CoT and RaR can be combined to get the best of both.</em></p><h2>Conclusion</h2><p>RaR enables LLMs to rephrase queries for better comprehension and more accurate responses. As we move forward, <strong>RaR promises to revolutionize LLM interactions, making them more intuitive, efficient, and aligned with human communication</strong>. We encourage you to experiment with RaR in their interactions with ChatGPT, to experience first-hand the advancements it brings to human-AI communication.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To read more, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Beyond Chain-of-Thought: The Evolution of AI Problem-Solving with Least-to-Most Prompting]]></title><description><![CDATA[Least-to-most prompting outperforms chain-of-thought in complex problem-solving.]]></description><link>https://ai-unltd.com/p/beyond-chain-of-thought-the-evolution</link><guid isPermaLink="false">https://ai-unltd.com/p/beyond-chain-of-thought-the-evolution</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 07 Dec 2023 20:43:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!dgPa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dgPa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dgPa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 424w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 848w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1272w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dgPa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png" width="554" height="554" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:554,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dgPa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 424w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 848w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1272w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: Least-to-most prompting applied to GPT-3 code-davinci-002, surpasses chain-of-thought methods in solving complex problems. This technique uses problem decomposition and subproblem solving, achieving 99% accuracy on the SCAN benchmark. This innovative approach demonstrates a leap towards deep learning systems with capabilities akin to human-like reasoning, bridging a significant gap in AI problem-solving.</p><div><hr></div><p>In the evolving world of AI prompting, the ability to solve complex problems has been a perennial challenge. Traditional chain-of-thought prompting, while effective in various reasoning tasks, often falters when faced with complexities beyond its programmed examples. This limitation sets the stage for a groundbreaking solution: least-to-most prompting.</p><h2>Bridging the Human-AI Gap</h2><p>Deep learning and human intelligence have always been worlds apart. Humans learn and reason from minimal examples, a feat AI has struggled to replicate. Chain-of-thought prompting made strides in this direction, offering improved interpretability and performance, yet it still stumbled in generalizing to more complex tasks. This is where least-to-most prompting shines, aligning more closely with how humans tackle difficult problems.</p><h2>The Rise of Least-to-Most Prompting</h2><p>Designed to tackle the challenge of easy-to-hard generalization, least-to-most prompting is a novel strategy that dissects complex problems into simpler subproblems. These subproblems are then solved sequentially, leveraging the answers of preceding ones, without necessitating extra training. It mirrors a teaching technique from educational psychology, sequentially guiding learners to build upon simpler concepts.</p><h2>Methodology and Execution</h2><p>The process involves two stages: <strong>decomposition</strong> and <strong>subproblem solving</strong>. Initially, the problem is broken down. Then, using a blend of constant examples and previously solved subquestions, each subproblem is tackled in turn. The culmination of this approach is the resolution of the original complex problem, an elegant demonstration of sequential reasoning.</p><h2>Experimental Results and Comparative Analysis</h2><p>In practice, least-to-most prompting has shown remarkable results. When applied to the GPT-3 code-davinci-002 model, it reached a staggering <strong>99% accuracy</strong> on the SCAN benchmark, significantly outperforming chain-of-thought's 16% accuracy. This impressive feat was achieved with just a handful of examples, highlighting the efficiency and effectiveness of this method.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pQkn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pQkn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 424w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 848w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1272w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pQkn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png" width="1286" height="218" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:218,&quot;width&quot;:1286,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:42985,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pQkn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 424w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 848w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1272w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Accuracy of Chain-of-Thought vs Least-to-Most</figcaption></figure></div><p><a href="https://arxiv.org/abs/2205.10625">Link to paper</a></p><h2>Advantages and Limitations</h2><p>This strategy not only excels in longer and more intricate problem sets but also demonstrates integration potential with other prompting methods. However, its brilliance has its bounds. The technique&#8217;s accuracy can be marred by minor errors like concatenation mistakes, and its application is less effective in tasks requiring domain-specific decomposition.</p><h2>Conclusion</h2><p>In conclusion, least-to-most prompting is not just a step forward in AI's problem-solving capabilities; it's a leap towards bridging the gap between machine learning and the nuanced reasoning of the human mind. While it's not the ultimate solution for teaching reasoning to AI, it marks a significant advancement, nudging AI towards more efficient and human-like learning processes.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Get Over Chain-of-Thought, Analogical Prompting is Here! [Prompt Examples Included]]]></title><description><![CDATA[Analogical Prompting, utilising analogical reasoning, offers a refined approach.]]></description><link>https://ai-unltd.com/p/get-over-chain-of-thought-analogical</link><guid isPermaLink="false">https://ai-unltd.com/p/get-over-chain-of-thought-analogical</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Mon, 30 Oct 2023 21:01:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WLAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WLAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WLAn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WLAn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg" width="592" height="444" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:592,&quot;bytes&quot;:455217,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WLAn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: Large Language Models (LLMs) using Chain-of-Thought (CoT) face challenges in 0-shot and few-shot reasoning. Analogical Prompting, utilizing analogical reasoning, offers a refined approach. Tapping into past experiences and generating high-level takeaways, it focuses on core problem-solving concepts. Testing on LLMs, including GPT-3.5-turbo and GPT-4, showed superior results in mathematical reasoning and code generation. As Analogical Prompting integrates knowledge generation with example crafting, it has the potential to set new benchmarks in the LLM domain, making reasoning efficient and intuitive.</p><h2>Business Implications</h2><ul><li><p>Its universal problem-solving approach can enhance product versatility, catering to a wider array of business challenges.</p></li><li><p>Enhanced AI capabilities via Analogical Prompting can optimize ROI by reducing manual efforts and expanding AI application areas.</p></li></ul><p><a href="https://chat.openai.com/share/544a97d2-e0ec-4805-97ae-71247c8cbe4a">Prompt Example 1</a></p><p><a href="https://chat.openai.com/share/9ca26c5a-5621-4c52-8f22-1d7c2c4ae1df">Prompt Example 2</a></p><div><hr></div><p>In the rapidly evolving world of artificial intelligence, Large Language Models (LLMs) have been at the forefront of pushing the boundaries of what machines can achieve. Among the strategies employed to amplify their potential, Chain-of-Thought (CoT) emerged as a frontrunner. However, like every innovation, CoT has its challenges. </p><p>Enter Analogical Prompting, a technique that promises to revolutionize the way we understand and guide LLMs.</p><h2>Chain-of-Thought (CoT) Unpacked</h2><p>At its core, CoT for LLMs showcased remarkable performance across an array of reasoning tasks. Yet, its efficacy came at a cost: the <strong>need for labelled examples to depict the reasoning process</strong>. In practice, this meant feeding LLMs with predefined instances. The 0-shot CoT offered a general instruction akin to "think step by step", serving as a generic reasoning guide. However, <strong>its broad approach sometimes fell short for intricate tasks</strong>. In contrast, few-shot CoT provided a more targeted approach, with multiple examples illustrating the reasoning process, but it had its own baggage &#8211; the onus of obtaining these labelled examples for every task. Such challenges paved the way for a quest: Could there be a technique that married the best of both worlds without their drawbacks?</p><h2>Enter Analogical Prompting: The New Kid on the Block</h2><p>What if there was a way to guide the reasoning process of LLMs seamlessly, drawing inspiration from the cognitive processes of humans? This is where Analogical Prompting enters the frame. Rooted in the principle of analogical reasoning, this method helps LLMs tap into past experiences, much like how we humans recall related problems and their solutions when faced with new challenges. For instance, calculating the area of a square becomes intuitive when one remembers the need to find its side length. By imitating this human-like reasoning, Analogical Prompting paves the way for LLMs to tackle novel problems with unprecedented efficiency.</p><h2>Advantages of Analogical Prompting Over CoT</h2><p>The magic of Analogical Prompting lies in its adaptability and self-sufficiency. By self-generating examples, it eliminates the labour of manually crafting reasoning instances for every task, a hurdle that both 0-shot and few-shot CoT grappled with. What&#8217;s more, these examples aren't just generic placeholders; they're tailored to individual problems, be it geometry or probability, ensuring a more nuanced approach. And the cherry on top? There&#8217;s no more scouring through external data sources to find relevant examples.</p><h2>Digging Deeper: How Analogical Prompting Works</h2><p>The brilliance of Analogical Prompting is not just in its outcomes but also in its methodology. Recognizing the importance of diverse examples, LLMs are directed to generate a range of 3 to 5 distinct instances in one go. But here&#8217;s the game-changer: before diving into examples, LLMs are instructed to generate high-level takeaways to complement the examples. By prioritizing this, LLMs hone in on the core concepts of the problem, ensuring that generated examples resonate with fundamental problem-solving strategies over mere superficial resemblances.</p><h2>Real-World Evaluations: Putting Analogical Prompting to the Test</h2><p>To determine its prowess, Analogical Prompting was put through its paces across a spectrum of tasks. <strong>From elementary math word conundrums and advanced high school math challenges to code generation involving intricate algorithms, and other reasoning tasks spanning logical deduction and formal fallacies</strong>. Two formidable LLMs, <strong>GPT-3.5-turbo and GPT-4</strong>, were the chosen ones for these experiments.</p><h2>Impressive Results: A Comparative Analysis</h2><p>The outcomes? Nothing short of impressive. In mathematical reasoning, Analogical Prompting left both 0-shot and few-shot CoT in the dust. It showcased similar superiority in code generation and other reasoning tasks, making a compelling case for its effectiveness. Not to mention, the fusion of knowledge generation with example crafting added another dimension to its efficacy.</p><p><a href="https://arxiv.org/abs/2310.01714">Paper Link</a></p><h2>Conclusion</h2><p>The world of LLMs stands at a fascinating juncture. While CoT laid the groundwork, Analogical Prompting is setting new benchmarks. By addressing the challenges that plagued its predecessor, it offers a glimpse into the future of machine reasoning &#8212; a future that's efficient, adaptable, and, most importantly, intuitive. As we march ahead, the exciting journey of LLMs and their capabilities is something to watch out for, and Analogical Prompting will undoubtedly be at its heart.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Retrieval-Augmented-Generation (RAG) vs Long-context LLMs: Which to Choose?]]></title><description><![CDATA[Is RAG superior to longer context LLMs?]]></description><link>https://ai-unltd.com/p/retrieval-augmented-generation-rag</link><guid isPermaLink="false">https://ai-unltd.com/p/retrieval-augmented-generation-rag</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Tue, 10 Oct 2023 04:30:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!E33R!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!E33R!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!E33R!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!E33R!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!E33R!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:181428,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!E33R!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!E33R!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: LLMs like GPT-4, <a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2 LONG</a> and Claude 2 boast Long context windows due to technological advancements in GPUs and memory-efficient exact attention. Retrieval Augmentation employs retrievers like Dragon, Contriever, and OpenAI&#8217;s text-embedding-ada-002 to provide LLMs with crucial context. A pivotal finding is that a 4K LLM augmented by retrieval can rival the efficacy of a 16K context LLM. Models like Nemo GPT-43B and LLAMA2-70B were pivotal in these comparisons, indicating that RAG could enhance the performance of an LLM irrespective of its context length.</p><h3>Business Implications</h3><ol><li><p>Retrieval-augmented LLMs can achieve top-tier AI capabilities without escalating operational costs, ensuring a competitive advantage.</p></li><li><p>Harnessing open-sourced retriever tools can drive superior AI performance, elevating customer satisfaction and trust, while keeping the costs low.</p></li><li><p>Retrieval-augmented models can provide high precision in responses, derived from expansive or latest content.</p></li></ol><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/p/retrieval-augmented-generation-rag?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Like these quick takeaways? Share this article with people who might find it interesting&#8230;</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/p/retrieval-augmented-generation-rag?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/p/retrieval-augmented-generation-rag?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><div><hr></div><p>In the rapidly evolving world of Large Language Models(LLMs), the two competing strategies have been contending for the spotlight: extending the context window of Large Language Models (LLMs) and augmenting LLMs with retrieval mechanisms. As the latter has been a familiar solution for a long time, it prompts us to ask two key questions:</p><ol><li><p>Is retrieval augmentation superior to longer context LLMs?</p></li><li><p>Could a synthesis of both strategies unlock unprecedented capabilities?</p></li></ol><h2>Long Context LLMs</h2><p>Long-context LLMs have recently become the focal point of discussions within research, production, and open-source domains. Spearheading this momentum is the advent of faster GPUs complemented by memory-efficient exact attention. These technologies have allowed for building Long Context LLMs with up to 32K (<a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2 LONG</a>, GPT-4) and 100K &#128561; (Claude 2).</p><h2>Conceptual Understanding of Retrieval Augmentation</h2><p>Beyond the context window expansion lies the retrieval method, a well-established, alternative approach. In this method, L<strong>LMs are provided with only the relevant context fetched by a retriever</strong>, offering higher scalability and speed. This strategy essentially transforms the retrieval-augmented decoder-only LLM into a model with sparse attention, where the attention pattern is determined not beforehand but by the retriever's discernment. In simpler words, <strong>the non-retrieved data (context) is treated as irrelevant and has zero-values attention weights</strong>.</p><h2>Experiment Overview</h2><p>The primary objective is to juxtapose the retrieval augmented method (using a retriever) and the everything-in-context method (adding content to the LLM context). Venturing beyond smaller models, the investigation centres on models larger than 40B parameters.</p><p>The two models used are:</p><ul><li><p><strong>Nemo GPT-43B</strong>: A model boasting 43 billion parameters and trained on a staggering 1.1T tokens. Its training data contains a 70% English corpus, enriched by multilingual and code data sources like Common Crawl, Wikipedia, and StackExchange.</p></li><li><p><strong><a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2-70B</a></strong>: A publicly accessible model with 70B parameters, trained on approximately 2T tokens dominated by English data.</p></li></ul><p>The experimental landscape spanned seven datasets, encompassing single-document QA, multi-document QA, and query-based summarization for zero-shot evaluations.</p><h2>Retrieval Mechanism Explored</h2><p>Three retrievers were employed:</p><ul><li><p><strong>Dragon</strong>: Renowned for benchmark-setting performances across supervised and zero-shot information retrieval.</p></li><li><p><strong>Contriever model</strong>: Another proficient retriever.</p></li><li><p><strong>OpenAI embedding</strong>: Specifically text-embedding-ada-002.</p></li></ul><p>The retrieval process: <strong>Questions and documents are encoded into vectors, rankings are established via methods like cosine similarity, and then the top tokens, based on their similarities, are selected and dispatched alongside the question in the LLM prompt.</strong></p><h2>Key Findings and Observations</h2><p>The findings are not either/or but an and:</p><ul><li><p>A retrieval-augmented LLM with a 4K context window astonishingly mirrored the prowess of an LLM with a 16K context window on long-context tasks, albeit with significantly reduced computational demands.</p></li><li><p>Retrieval-augmented <a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2-70B with a 32K context window</a> overshadowed GPT-3.5-Turbo-16K and Davinci003 across various long-context tasks.</p></li><li><p><strong>The experiments cemented the understanding that irrespective of context window size, retrieval enhances LLM performance.</strong></p></li><li><p>Interestingly, when presented with identical evidence chunks, long-context LLMs (16K, 32K) outshine their 4K context counterparts.</p></li><li><p><strong>Public retrievers often performed better than proprietary solutions like OpenAI embeddings.</strong></p></li></ul><p><a href="https://arxiv.org/abs/2310.03025">Paper Link</a></p><h2>When To Use RAG?</h2><ul><li><p>When you have question-answering tasks based on long-form documents (PDFs, Text files).</p></li><li><p>When you have to support an LLM by passing extra information that is not already available in its context to reply to a user.</p></li><li><p>When you want the LLM to always have the latest information.</p></li><li><p>When you want the LLM to not hallucinate and make up random information &#8212; tho using a RAG is not a guarantee of eliminating hallucination completely.</p></li></ul><h2>When Is LLM Context Just Enough?</h2><ul><li><p>When answering questions based on short texts that can fit well inside the context of LLMs.</p></li><li><p>Summarization tasks that involve the LLM being able to see the whole document.</p></li><li><p>When the conversation with LLM is based on its already trained/fine-tuned data.</p></li></ul><h2>Closing Thoughts&#8230;</h2><p>Retrieval augmentation Generation in LLMs significantly amplifies their strengths, enhancing perplexity, accuracy, and learning capacities. The evidence is compelling: a 4K context LLM, when powered by retrieval, can go toe-to-toe with a 16K context counterpart, ensuring computational efficiency during inference.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Enjoyed diving deep into the world of RAG and long-context LLMs? Subscribe to the <strong>AI Newsletter Today</strong> to keep up with the latest advancements in AI!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Extending LLAMA to 32K tokens - Catching Up with ChatGPT]]></title><description><![CDATA[LLAMA 2 LONG defeats GPT-3.5-16K ChatGPT model]]></description><link>https://ai-unltd.com/p/extending-llama-to-32k-tokens-catching</link><guid isPermaLink="false">https://ai-unltd.com/p/extending-llama-to-32k-tokens-catching</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Fri, 06 Oct 2023 04:30:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GjRB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GjRB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GjRB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GjRB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg" width="524" height="393" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:524,&quot;bytes&quot;:409726,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GjRB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: Meta's advancement in LLAMA 2 models manifests in handling long contexts, defeating the GPT-3.5-16K model. The newly introduced models: LLAMA 2 LONG 7B, 13B, 34B, and 70B, are centered around long-context continual pretraining, negating the need for a vast volume of long texts. The smaller 7B/13B variants were trained with 32,768-token sequences while the larger 34B/70B variants with 16,384-token sequences, demonstrated a significant performance improvement with an increased context window.</p><h3>Business Implications</h3><ul><li><p><strong>Enhanced Data Security</strong>: By self-hosting LLAMA 2 LONG models, companies can bolster data security while sustaining a model performance comparable to GPT, mitigating reliance on external providers.</p></li><li><p><strong>Cost-Effectiveness</strong>: The adoption of LLAMA 2 LONG models presents a more economical alternative to OpenAI&#8217;s GPT, with expenses relegated primarily to server hosting, thus reducing operational costs.</p></li><li><p><strong>Efficient Long-Context Handling</strong>: The proficiency of LLAMA 2 LONG in managing long-context tasks paves the way for developers to create chat solutions devoid of semantic retrievers, enhancing LLMs' capacity in question-answering tasks over extensive documents, streamlining the development process.</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Meta's recent stride in Large Language Models (LLMs) lays down a milestone in extending the capabilities of open-source LLMs, particularly LLAMA 2, to handle long contexts efficiently, nudging closer to the prowess of models like GPT-4. There are 4 new models:</p><ul><li><p>LLAMA 2 LONG 7B</p></li><li><p>LLAMA 2 LONG 13B</p></li><li><p>LLAMA 2 LONG 34B</p></li><li><p>LLAMA 2 LONG 70B</p></li></ul><h2>Expanding the Horizons of Language Models</h2><p>Meta has taken a giant leap by introducing a series of long-context LLMs supporting effective context windows of up to 32,768 tokens. Through the lens of continual pretraining, LLAMA 2 has been extended with longer training sequences on a dataset where long texts were upsampled, starting a new frontier for open-source language models.</p><h2>Achieving Superior Performance</h2><p>Meta's research challenges the assumption that a wealth of long texts in the pre-train dataset is crucial for excelling in long-context tasks. It reveals that the method of long-context continual pretraining, not the volume of long texts, is key to achieving superior performance.</p><p>The strategy of long-context continual pretraining builds upon the existing knowledge and architecture of LLAMA 2, rather than starting from scratch with long sequences. This approach is less resource-intensive and time-efficient, showcasing notable improvement in long-context tasks. It utilizes a dataset where long texts are upsampled, providing a solid foundation for the LLAMA models to handle extended contexts efficiently.</p><p>Furthermore, the research uncovers a significant between performance improvement and context length, indicating continual performance improvement as the context length increases, up to 32,768 tokens.</p><h2>A Glimpse into the Training Arena</h2><p>The training regime embarked upon was both simple and cost-effective. A total of 400 billion tokens, formed as long training sequences, were utilized to train the existing LLAMA 2 model. The smaller 7B/13B variants were trained with 32,768-token sequences while the larger 34B/70B variants with 16,384-token sequences. The positional encoding modifications were a cornerstone to ensure the model's adeptness at attending to longer sequences.</p><h2>Instruction Tuning for Long-Context Tasks</h2><p>Instruction tuning emerged as a key ingredient in navigating the challenges of LLM alignment, especially under long-context scenarios. A simple yet effective approach leveraging a pre-built short-prompt dataset exhibited surprising efficacy on long-context benchmarks.</p><h2>Comparative Analysis and Results</h2><p>When pitted against benchmarks, the LLAMA 2 models trained with long sequences showcased significant improvements, especially on long-context tasks. The end result was a chat model, devoid of any human-annotated data, displaying a stronger overall performance than gpt-3.5-turbo-16k and other open-source models across a suite of long-context benchmarks. However, when thrown in the ring with the GPT-4 32K model, the LLAMA 2 Long 70B model fell short, indicating there's still ground to cover.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ys_-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ys_-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 424w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 848w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1272w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ys_-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png" width="795" height="166" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:166,&quot;width&quot;:795,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:44405,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ys_-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 424w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 848w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1272w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Source: <strong>Effective Long-Context Scaling of Foundation Models</strong></figcaption></figure></div><p><a href="https://arxiv.org/abs/2309.16039">Paper Link</a></p><h2>Bridging the Future: LLMs in Complex Use Cases and Beyond</h2><p>The narrative of LLMs is evolving rapidly with each passing day. They now stand on the verge of serving more intricate use cases, from analyzing knowledge-rich documents to powering more genuine chat interactions. This expedition of Meta reflects not just a technical advancement but a step towards a future where human-digital interactions are more intuitive and enriched.</p><p>The journey of extending LLAMA 2 to 32k tokens while keeping an eye on GPT models reflects a competitive spirit driving the field towards uncharted territories. The ingenuity in training methodologies and instruction tuning, as demonstrated, not only propels LLAMA 2 closer to the prowess of GPT but also lights the way for future endeavours in the domain.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Reducing Hallucinations in ChatGPT with Chain-of-Verification (CoVe)]]></title><description><![CDATA[Using Chain-of-Verification, Llama outperformed leading models like ChatGPT, InstructGPT, and PerplexityAI.]]></description><link>https://ai-unltd.com/p/reducing-hallucinations-in-chatgpt</link><guid isPermaLink="false">https://ai-unltd.com/p/reducing-hallucinations-in-chatgpt</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Mon, 25 Sep 2023 19:41:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9NfE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9NfE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9NfE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9NfE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg" width="1456" height="822" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:822,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1530747,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9NfE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR:</strong> Hallucinations in LLMs refer to incorrect yet seemingly plausible outputs. The Chain-of-Verification (CoVe) method seeks to mitigate this by having LLMs draft, verify, and refine responses. <strong>Llama 65B, using CoVe, surpassed models like ChatGPT in long-form tasks</strong>. CoVe's efficacy was notably tested on Wikidata and other tasks, showing an improvement in precision, F1 score, and fact score. Despite its potential, CoVe still has some limitations in completely eradicating hallucinations.</p><h3>Business Implications</h3><ol><li><p>CoVe's improved LLM accuracy can lead to better AI-driven decision-making for businesses.</p></li><li><p>Enhanced trust in AI outputs can foster stronger customer relationships and brand reputation.</p></li><li><p>Llama 65B with CoVe might provide superior AI performance and security, differentiating businesses in the market.</p></li><li><p>Despite CoVe's advancements, businesses should maintain a hybrid approach, combining AI insights with human judgment.</p></li><li><p>CoVe's proficiency in short-form tasks indicates its potential for efficient chatbots and automated customer interactions.</p></li></ol><p>Here is a sample ChatGPT conversation to show the effectiveness of Chain-of-Verification (CoVe): <a href="https://chat.openai.com/share/c716c501-2d41-478a-a436-019aa378b717">Chat Link</a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.ainewsletter.today/&quot;,&quot;text&quot;:&quot;Subscribe to read more such articles&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.ainewsletter.today/"><span>Subscribe to read more such articles</span></a></p><div><hr></div><p>Hallucinations in LLMs refer to the generation of plausible yet factually incorrect information. As LLMs are trained on an enormous text corpus, spanning billions of tokens, their performance generally improves with an increase in model parameters. Yet, even the most advanced models can falter, especially on tasks less represented in their training data. These errors often appear credible yet are factually incorrect.</p><h2>Chain-of-Verification (CoVe): A Solution to Hallucinations</h2><p>Enter the Chain-of-Verification (CoVe) method, a novel approach designed to curb the hallucination issue in LLMs. The CoVe method is a systematic process where the LLM:</p><ul><li><p>Draft an initial response</p></li><li><p>Plan verification questions to fact-check the draft</p></li><li><p>Answer the planned verification questions independently to avoid bias</p></li><li><p>Generates a final, verified response</p></li></ul><h2>Deep Dive: How CoVe Works</h2><ul><li><p><strong>Baseline Response</strong>: A simple output from LLM is obtained as the starting point, which is typically prone to hallucinations.</p></li><li><p><strong>Plan Verifications</strong>: Using the baseline response, LLM generates a series of verification questions that test the factual claims of the baseline responses.</p></li><li><p><strong>Execute Verifications</strong>: LLM answers the planned verification questions through several variations like <em>Joint, 2-Step, Factored, and Factor+Revise</em>, each with its unique approach and level of sophistication.</p></li><li><p><strong>Generate Final Verified Response</strong>: The improved response that takes verification into account is generated, incorporating any discovered inconsistencies.</p></li></ul><h2>CoVe in Action: Experimental Results</h2><p>CoVe's efficacy was tested across various tasks, including Wikidata, Wikipedia Category List, MultiSpanQA, and long-form biographies. The results are promising:</p><ul><li><p>Significant precision improvements in list-based tasks.</p></li><li><p>Enhanced performance in closed book QA, with a 23% F1 score improvement (This represents an improvement to both precision and recall).</p></li><li><p>A 28% increase in fact score for long-form generations.</p></li><li><p><strong>Notably, with CoVe, Llama 65B outperformed leading models like ChatGPT, InstructGPT, and PerplexityAI in long-form generation tasks, marking a significant achievement in the realm of open-source LLMs.</strong></p></li></ul><p><a href="https://arxiv.org/abs/2309.11495">Paper Link</a>.</p><p><a href="https://chat.openai.com/share/c716c501-2d41-478a-a436-019aa378b717">Check this ChatGPT conversation for prompt examples.</a></p><h2>Additional Insights from the Study</h2><p>The experiments also revealed that short-form verification questions were more accurately answered than long-form ones. Additionally, LLM-based verification questions surpassed heuristic-based ones, and open-ended questions proved more effective than yes/no formats.</p><h2>Limitations of the CoVe Method</h2><p>Despite its groundbreaking approach, CoVe isn't without limitations. While it significantly reduces hallucinations, it doesn't eradicate them entirely. There's still a possibility of the model generating misleading information. Moreover, hallucinations might manifest in other forms, such as during incorrect reasoning or when expressing opinions in long-form answers.</p><h2>Conclusion</h2><p>The Chain-of-Verification (CoVe) method represents a significant stride in reducing hallucinations in Large Language Models, enhancing their reliability and accuracy across various tasks. By enabling models to verify their responses, CoVe brings us closer to more dependable and error-free artificial intelligence, although some limitations and challenges still need to be addressed.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[PDF Triage: Elevating ChatGPT's Question Answering Capabilities]]></title><description><![CDATA[Unlocking Advanced Document Insights with Enhanced QA Techniques]]></description><link>https://ai-unltd.com/p/pdf-triage-elevating-chatgpts-question</link><guid isPermaLink="false">https://ai-unltd.com/p/pdf-triage-elevating-chatgpts-question</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Wed, 20 Sep 2023 04:30:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!IxM9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IxM9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IxM9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IxM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg" width="528" height="396" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:528,&quot;bytes&quot;:480239,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IxM9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: PDFTriage enhances LLMs' ability to handle large documents by leveraging the Adobe Extract API. With functions such as fetch_pages and fetch_table, it addresses document structure and table reasoning questions as well. The gpt-35-turbo-0613 model, when using PDFTriage, produces answers with fewer retrieved tokens, increasing efficiency.</p><h3>Business Implications</h3><ol><li><p>By mirroring users' perceptions of documents, PDFTriage can elevate product UX, potentially leading to increased adoption and retention.</p></li><li><p>Its robust performance across varying document lengths offers businesses a versatile tool, reducing the need for multiple solutions.</p></li><li><p>Precise and efficient data extraction can provide CXOs with actionable insights, optimizing business strategies.</p></li></ol><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>The ability to extract precise information from documents is important in the digital age. Large Language Models (LLMs) have been at the forefront of this, but they face challenges when the document's size exceeds its context length. This limitation often leads to inefficiencies in document question answering (QA).</p><h2>The Problem with Current LLM Approaches</h2><p>LLMs, despite their prowess, falter when the document doesn't fit within their context length. The prevailing solution has been to retrieve relevant contexts from the document and present them as plain text.</p><p>However, this approach overlooks the inherent structure of many documents. When users think of a PDF or a webpage, they visualize pages, tables, and sections. Representing these as mere text creates a disconnect between the user's mental model and the system's representation. This incongruity becomes glaringly evident when seemingly simple questions stump the QA system. </p><p>For instance:</p><ul><li><p>"Can you summarize the key takeaways from pages 5-7?"</p></li><li><p>"What year [in Table 3] has the maximum revenue?"</p></li></ul><p>Both questions require an understanding of the document's structure, something that plain text representation lacks.</p><h2>Introducing PDFTriage: Bridging the Gap</h2><p>Enter PDFTriage, a groundbreaking approach that enables models to retrieve context based on either the document's structure or its content. This method proves effective where traditional retrieval-augmented LLMs fall short. By giving models access to a document's structural metadata, PDFTriage can handle a variety of questions that stump plain retrieval-augmented LLMs.</p><h2>How PDFTriage Works</h2><p>The genius of PDFTriage lies in its three-step method:</p><ol><li><p><strong>Generate Document Metadata</strong>: Using the Adobe Extract API, PDFs are transformed into an HTML-like tree. This tree, rich with metadata like section titles, tables, and figures, is then parsed to extract valuable structural information.</p></li><li><p><strong>LLM-based Triage</strong>: The LLM queries the document, selecting precise content based on the question at hand.</p></li><li><p><strong>Answer Using Retrieved Content</strong>: With the relevant context retrieved, the LLM generates a comprehensive answer.</p></li></ol><p>PDFTriage employs five functions to achieve this:</p><ul><li><p><strong>fetch_pages</strong>: Retrieves text from specified pages.</p></li><li><p><strong>fetch_sections</strong>: Extracts text from a given section.</p></li><li><p><strong>fetch_table</strong>: Gathers text from a specified table caption.</p></li><li><p><strong>fetch_figure</strong>: Obtains text surrounding a particular figure caption.</p></li><li><p><strong>retrieve</strong>: Issues a natural language query over the document, fetching pertinent chunks.</p></li></ul><p><strong>These functions will be called by LLMs like GPT to synthesise various pieces of information to craft the final answer.</strong></p><h2>PDFTriage in Action: Testing and Results</h2><p>To validate PDFTriage's capabilities, a dataset comprising roughly 900 human-written questions spanning 90 documents was curated. These questions spanned categories like "document structure questions," "table reasoning questions," and even "trick questions". PDFTriage was tested on the gpt-35-turbo-0613 model.</p><p>The results were illuminating. Human evaluators consistently favoured PDFTriage over traditional retrieval methods. Specifically, PDFTriage was preferred 50.7% of the time, outperforming both Page Retrieval and Chunk Retrieval methods. </p><p>Moreover, PDFTriage showcased its efficiency by requiring fewer tokens to produce superior answers. Impressively, the length of the document had a negligible effect on PDFTriage's performance, underscoring its adaptability to both short and long documents.</p><p><a href="https://arxiv.org/abs/2309.08872">Paper Link</a></p><h2>Benefits of PDFTriage</h2><p>PDFTriage's approach aligns seamlessly with the user's perception of structured documents. By recognizing and utilizing a document's structure, it offers:</p><ul><li><p>Enhanced answer quality and accuracy.</p></li><li><p>Improved readability and informativeness.</p></li><li><p>Efficient answers with fewer retrieved tokens.</p></li><li><p>Consistent performance across varying document lengths.</p></li></ul><h2>Conclusion</h2><p>PDFTriage is set to redefine the realm of document QA. By aligning more closely with users' perceptions of structured documents, it offers a more intuitive and efficient solution to information retrieval. Its potential impact on future QA systems is immense, promising more accurate, efficient, and user-friendly outcomes.</p><p>For those keen on harnessing the power of advanced QA systems, it's time to delve deeper into PDFTriage. Its applications are vast, and its promise is undeniable. Embrace the future of structured document querying today.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[High-Quality Summaries with ChatGPT: Chain of Density (Prompt Included)]]></title><description><![CDATA[This is a groundbreaking method to generate increasingly dense summaries.]]></description><link>https://ai-unltd.com/p/high-quality-summaries-with-chatgpt</link><guid isPermaLink="false">https://ai-unltd.com/p/high-quality-summaries-with-chatgpt</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sun, 17 Sep 2023 19:06:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Vrhh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Vrhh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Vrhh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg" width="528" height="396" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:528,&quot;bytes&quot;:387342,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Vrhh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In today's online world, where there's so much information, turning big chunks of text into short, clear summaries is very important. Here comes ChatGPT (GPT-4). Among the many things it can do with a simple prompt, making quick summaries is one of its top skills, helping users get the main idea of long articles, reports, or data easily and quickly.</p><h2>The Challenge of Summarization</h2><p>Crafting the perfect summary is no easy feat. It involves selecting just the right amount of information&#8212;enough to be detailed and entity-centric, but not so much that it becomes dense and hard to follow. This delicate balance is what researchers from Salesforce AI, MIT, and Columbia University aimed to strike with their innovative "Chain of Density" (CoD) prompt.</p><h2>Introducing the Chain of Density (CoD) Prompting</h2><p>The CoD approach is a groundbreaking method that seeks to generate increasingly dense GPT-4 summaries. The process begins with a very simple summary. From there, GPT-4 iteratively incorporates missing important information, all without increasing the summary's length.</p><h2>How Chain of Density Works</h2><p>The CoD method is iterative. It starts with a summary that focuses on just 1-3 initial key pieces of information (entities). As the process continues, 1-3 missing entities are identified from the source text and seamlessly fused into the summary, all while maintaining the original length. This is achieved through a combination of abstraction, compression, and fusion. The aim is to convey more information within a fixed token budget, ensuring the summary remains legible and accurate. <strong>There are 5 total iterations that happen and all these iterations can happen within a single prompt.</strong></p><h2>Comparing CoD Summaries with Traditional GPT-4 Summaries</h2><p>Summaries produced using the CoD method have distinct advantages. They are more abstractive, exhibit greater fusion, and have less lead bias than those generated by a vanilla GPT-4 prompt. A human preference study, conducted on 100 CNN/DailyMail articles, revealed that humans favoured GPT-4 summaries produced using the CoD method. These summaries were almost as dense as human-written ones and made more sense than those generated by a vanilla prompt.</p><h2>Practical Application: The CoD Prompt in Action</h2><p>The Chain of Density prompt is a marvel of innovation. Here&#8217;s the exact prompt:</p><pre><code>Article: {{ ARTICLE}}

You will generate increasingly concise, entity-dense summaries of the above Article.

Repeat the following 2 steps 5 times.

Step 1. Identify 1-3 informative Entities (";" delimited) from the Article which are missing from the previously generated summary.
Step 2. Write a new, denser summary of identical length which covers every entity and detail from the previous summary plus the Missing Entities.

A Missing Entity is:
- Relevant: to the main story.
- Specific: descriptive yet concise (5 words or fewer).
- Novel: not in the previous summary.
- Faithful: present in the Article.
- Anywhere: located anywhere in the Article.

Guidelines:
- The first summary should be long (4-5 sentences, ~80 words) yet highly non-specific, containing little information beyond the entities marked as missing. Use overly verbose language and fillers (e.g., "this article discusses") to reach ~80 words.
- Make every word count: re-write the previous summary to improve flow and make space for additional entities.
- Make space with fusion, compression, and removal of uninformative phrases like "the article discusses".
- The summaries should become highly dense and concise yet self-contained, e.g., easily understood without the Article.
- Missing entities can appear anywhere in the new summary.
- Never drop entities from the previous summary. If space cannot be made, add fewer new entities. Remember, use the exact same number of words for each summary. 

Answer in JSON. The JSON should be a list ( length 5) of dictionaries whose keys are "Missing_Entities" and "Denser_Summary"</code></pre><p><em>Tip: Include the article in XML tags such as </em><code>&lt;Article&gt;&lt;/Article&gt;</code><em> for good results</em></p><p>When applied to an article titled "<a href="https://edition.cnn.com/cnn-underscored/electronics/iphone-15-hands-on">iPhone 15 hands-on: pre-orders, release date, price</a>", the following output was produced:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!v7l7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!v7l7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 424w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 848w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1272w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!v7l7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png" width="876" height="608" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:608,&quot;width&quot;:876,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:124041,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!v7l7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 424w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 848w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1272w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The 3rd output seems to be the best.</strong></p><p><a href="https://arxiv.org/abs/2309.04269">Paper Link</a></p><h2>Evaluating the Quality of CoD Summaries</h2><p>The quality of CoD summaries was put to the test through both human and GPT-4 evaluations. Human annotators were presented with randomly shuffled CoD summaries from 100 CNN/DailyMail articles. Their feedback was enlightening: for 3 out of 4 annotators, <strong>the first iteration of CoD received the most first-place votes across the 100 examples. However, in aggregate, 61% of top-ranked summaries had undergone at least three iterations</strong>. This suggests that a significant portion of annotators favoured summaries that had undergone further densification. The inferred entity density, calculated as the ratio of entities mentioned per token, after three iterations was approximately 0.15, closely mirroring human-written summaries (0.151) and surpassing those produced by a vanilla GPT-4 prompt.</p><p>GPT-4 itself showcased its versatility by being able to evaluate the quality of summaries. When prompted, it provided ratings, revealing a <strong>preference for the middle iterations with scores of 4.78, 4.77, and 4.76, while the first and last iterations were less favoured</strong>. This dual approach to evaluation, combining human insights with machine precision and specific ratings, offers a comprehensive understanding of summary quality.</p><h2>Limitations and Future Directions</h2><p>While the CoD method has shown promise, it's worth noting that its analysis has been limited to news summarization. Additionally, CoD has been exclusively tested on GPT-4, a closed-source model, and hasn't been evaluated on other large language models (LLMs). There's potential for its application across various domains, and the quest continues to determine the optimal density for summaries. The broader applicability of CoD on different LLMs remains an area ripe for exploration.</p><h2>Conclusion</h2><p>The Chain of Density prompting method offers a fresh perspective on automated text summarization. It's an invitation for readers and researchers alike to experiment with CoD and GPT-4, pushing the boundaries of what's possible in the realm of concise, high-quality summaries.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Enjoyed diving deep into the world of GPT-4 and the Chain of Density Prompting? Subscribe to <strong>AI Newsletter Today</strong> to keep up with the latest advancements in AI!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Automatic Audiobook Creation Using Neural Text-To-Speech]]></title><description><![CDATA[Over 5000 audiobooks were generated using Neural Text-To-Speech and Project Gutenberg e-books.]]></description><link>https://ai-unltd.com/p/automatic-audiobook-creation-using</link><guid isPermaLink="false">https://ai-unltd.com/p/automatic-audiobook-creation-using</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 14 Sep 2023 04:31:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!cfU2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cfU2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cfU2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cfU2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg" width="560" height="420" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:560,&quot;bytes&quot;:492223,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cfU2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR:</strong> Researchers from Microsoft, MIT, Project Gutenberg and Google have collaborated to revolutionize audiobook creation. Using neural text-to-speech technology and the SynapseML framework, they've automated the process, turning e-books into high-quality, customizable audiobooks. This system overcomes previous robotic voice challenges, offers voice customization, and introduces emotive reading, making literature more accessible and engaging.</p><p>Here&#8217;s an AI-generated audiobook:</p><iframe class="spotify-wrap podcast" data-attrs="{&quot;image&quot;:&quot;https://i.scdn.co/image/ab6765630000ba8ab5b43154e3b1dc9ff7dfa08e&quot;,&quot;title&quot;:&quot;Trent's Trust, and Other Stories by Bret Harte&quot;,&quot;subtitle&quot;:&quot;Project Gutenberg&quot;,&quot;description&quot;:&quot;Episode&quot;,&quot;url&quot;:&quot;https://open.spotify.com/episode/2mFIf7SFpoY2JP8LRbjN0K&quot;,&quot;belowTheFold&quot;:false,&quot;noScroll&quot;:false}" src="https://open.spotify.com/embed/episode/2mFIf7SFpoY2JP8LRbjN0K" frameborder="0" gesture="media" allowfullscreen="true" allow="encrypted-media" data-component-name="Spotify2ToDOM"></iframe><div><hr></div><p>In the digital age, audiobooks have emerged as a pivotal medium, enhancing the accessibility and engagement of literature for audiences worldwide. However, the traditional methods of creating these audiobooks are fraught with challenges, from the extensive time and effort required to the inconsistencies in quality. Enter the groundbreaking collaboration between researchers from tech and academic giants: Microsoft, MIT, Project Gutenberg, and Google. Together, they're reshaping the audiobook landscape, introducing automation and innovation where it's needed most.</p><p><a href="https://marhamilresearch4.blob.core.windows.net/gutenberg-public/Website/index.html#Code">Website Link</a></p><p><a href="https://arxiv.org/abs/2309.03926">Paper Link</a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><h2>The Need for Automated Audiobook Creation</h2><p>Historically, producing an audiobook has been a labour-intensive process. Whether it's the meticulous narration by professionals or the passionate efforts of volunteers, the journey from text to audio is long and winding. Platforms like LibriVox, driven by human volunteers, have made commendable strides in making audiobooks accessible. However, the variability in recording quality and environments can lead to inconsistent outputs. On the other end of the spectrum, platforms like Audible offer high-quality audiobooks but at a price, both monetarily and in terms of open access.</p><h2>Unveiling the Automated System</h2><p>The collaborative project introduces a system that harnesses the power of neural text-to-speech technology, promising to revolutionize the way we perceive audiobooks. Imagine generating thousands of human-quality audiobooks from the vast collection of Project Gutenberg, all at the click of a button. Beyond this, the system offers unparalleled customization. Listeners can adjust speaking speeds, choose different styles, and even modify emotional intonations. And for those who've always dreamt of hearing a book in their voice, this system can make that dream a reality with just a snippet of sample audio.</p><h2>Overcoming Traditional Challenges</h2><p>Past attempts at automated audiobook creation often stumbled at two major hurdles: the unmistakably robotic tone of text-to-speech systems and the challenge of discerning which text segments should be vocalized (you don&#8217;t want a huge index table readout in an audiobook!). This new system not only produces high-quality, human-like audio but also intelligently navigates the content of diverse e-books, ensuring a seamless listening experience.</p><h2>The Technical Backbone: SynapseML</h2><p>At the heart of this revolution is SynapseML, a robust scalable machine learning framework. It orchestrates the entire audiobook creation process, ensuring efficiency and quality at every step. From parsing thousands of e-books from Project Gutenberg to generating emotive, lifelike speech, SynapseML is the unsung hero behind the scenes.</p><h2>From E-books to Audiobooks: The Process</h2><p>The journey begins with e-books, specifically those in the HTML format from Project Gutenberg. Given the non-standardized nature of these files, which can contain everything from footnotes to transcriber notes, the system employs clustering on the HTML Document Object Model (DOM) tree. This approach allows for the efficient parsing of vast collections, ensuring only the relevant text makes its way to the listener's ears.</p><h2>Achieving High-Quality Speech</h2><p>The system's prowess doesn't stop at parsing. It recognizes the nuances required in different genres. While a non-fiction work might demand a clear, neutral voice, fiction, with its dialogues and drama, comes alive with emotive reading. Using the zero-shot text-to-speech method, it can even clone a user's voice from minimal recordings, adding a personal touch to the listening experience.</p><h2>Breathing Life into Text: Emotive Reading</h2><p>What truly sets this system apart is its ability to infuse emotion into the narrative. By segmenting text into narration and dialogue, identifying speakers, and predicting emotions, it crafts a dynamic listening experience. Passages with multiple characters and emotional dialogues are no longer monotonous; they're vibrant and engaging, much like a theatrical performance.</p><h2>Conclusion</h2><p>The collaboration between researchers of Microsoft, MIT, Project Gutenberg, and Google marks a significant leap in the world of audiobooks. By automating the creation process, they're not only making literature more accessible but also ensuring consistent quality. As we stand on the cusp of this revolution, one thing is clear: the future of audiobooks is bright, inclusive, and incredibly exciting.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[ChatGPT Can Build a Complete Software in Less Than 7 Minutes: ChatDev]]></title><description><![CDATA[Dive into the world of ChatDev, and experience the future of software development.]]></description><link>https://ai-unltd.com/p/chatgpt-can-build-a-software-development</link><guid isPermaLink="false">https://ai-unltd.com/p/chatgpt-can-build-a-software-development</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 07 Sep 2023 04:30:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bg8e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bg8e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bg8e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bg8e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg" width="570" height="427.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:570,&quot;bytes&quot;:468459,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bg8e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The process of building software is often seen as a lengthy and complex endeavour, filled with layers of decisions, consultations, and nuanced intuition. Enter ChatDev, an innovative paradigm powered by ChatGPT, which promises to flip this narrative, presenting a virtual software development company that operates at lightning speed and minimal cost.</p><h2>The ChatDev Framework</h2><p>ChatDev mirrors an entire software development company, it offers a comprehensive approach to building software applications. The secret sauce? A structured waterfall model that segregates the software creation process into distinct phases. Adding another layer of effectiveness, the chat chain deconstructs these phases into granular subtasks, ensuring precision and coherence at every step.</p><p>A standout feature of ChatDev is its innovative use of personas, all simulated by Large Language Models (LLMs). These personas, ranging from CEOs to programmers, art designers to testers, embody the roles of a real-world software development team. <strong>Each persona has its unique expertise and perspective, bringing depth and diversity to the decision-making process</strong>. They interact, collaborate, and debate just as human counterparts would in a traditional software company. This multi-agent approach, entirely powered by ChatGPT, creates a dynamic and responsive virtual development environment.</p><h2>The Four Phases of ChatDev</h2><h4>1. Designing</h4><p>The journey begins with the 'Designing' phase. Here, innovative ideas are crafted. The CEO, CPO, and CTO personas (all played by ChatGPT agents) bring their expertise to the table, turning abstract concepts into well-defined technical designs. They discuss and decide on the software's modality and the ideal programming language to bring it to life.</p><h4>2. Coding</h4><p>With a design in hand, the 'Coding' phase takes centre stage. The CTO collaborates with the programmer and the art designer, translating designs into functional code and captivating user interfaces. Every line of code, every design element, is meticulously crafted in this collaborative endeavour.</p><h4>3. Testing</h4><p>But what's coding without rigorous testing? In the 'Testing' phase, the software undergoes thorough scrutiny. The programmer, reviewer, and tester personas ensure that the code doesn't just work; it excels. From peer reviews to system testing, every potential glitch is identified and rectified.</p><h4>4. Documenting</h4><p>The final stage in this journey is 'Documenting'. This isn't just about creating a user manual; it's about crafting a comprehensive guide that details environment specifications and everything a user needs to make the most of the software. The CEO, CPO, CTO, and programmer unite to ensure that users have clarity from the get-go.</p><h2>Addressing the Hallucination Challenge</h2><p>Hallucinations in ChatGPT agents could lead to incomplete functions, missing dependencies, or even undiscovered bugs. However, ChatDev's granular task approach and cross-examination mechanism serve as a safeguard against these hallucinations, ensuring clarity and precision.</p><h2>The Role of Communication in ChatDev</h2><p>Communication is the heart of ChatDev. The framework thrives on context-aware, multi-turn discussions, ensuring that every decision, from software design to bug fixes, is a result of collaborative dialogue. This context-rich communication is further enhanced by the thought instruction mechanism.</p><p>The thought instruction mechanism introduces a "role flip" during the code completion, reviewing, and testing stages. An instructor agent provides specific guidelines for code modifications, directing the assistant programmer agent. This ensures precision and alignment with desired outputs, minimizing errors and enhancing the overall quality of the software being developed.</p><h2>Experiments, Results, and Findings</h2><p>Rigorous experiments using the GPT-3.5 Turbo version of ChatGPT validate ChatDev's prowess. Not only does it highlight the framework's efficacy, but it also showcases tangible metrics of its efficiency. ChatDev impressively produces software solutions for 70 unique user requirements, generating an average of 17.04 files per software. <strong>The speed is unmatched, wrapping up the entire process in just 409.84 seconds (that's under 7 minutes &#128561;) and at a staggering low cost of approximately $0.2967 (&#129327;)</strong>. These metrics elucidate the unparalleled efficiency and cost-effectiveness of the ChatDev approach.</p><p><a href="https://arxiv.org/abs/2307.07924v3?s=09">Paper Link</a></p><p><a href="https://github.com/OpenBMB/ChatDev">Code Link</a></p><h2>Conclusion</h2><p>ChatDev is more than just a software development tool; it's a testament to the future of software engineering. By harnessing the power of ChatGPT, ChatDev offers a glimpse into a world where building software is not just fast and affordable but also efficient and precise. With its structured approach, collaborative communication, and unmatched speed, ChatDev promises to reshape the software development landscape, one line of code at a time.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Fact Checking ChatGPT: Using FacTool to Detect Factual Errors]]></title><description><![CDATA[FacTool - A Framework Designed to Detect Factual Errors in LLM Outputs.]]></description><link>https://ai-unltd.com/p/fact-checking-chatgpt-using-factool</link><guid isPermaLink="false">https://ai-unltd.com/p/fact-checking-chatgpt-using-factool</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Wed, 06 Sep 2023 04:31:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!g2ML!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g2ML!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g2ML!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g2ML!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg" width="542" height="406.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:542,&quot;bytes&quot;:477668,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g2ML!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The rise of Large Language Models (LLMs) like ChatGPT has revolutionized the AI domain. These models offer high-quality text outputs, but they're not without challenges. Key among these is the issue of factual errors in generated texts. This article delves into the world of FacTool, a beacon in the murky waters of LLM-generated content.</p><h2>Challenges and Limitations of LLM-Generated Texts</h2><p>LLMs have ushered in a new age of content creation, handling a vast array of tasks from question answering to code generation. However, with their prowess come certain limitations:</p><ul><li><p>A heightened risk of factual inconsistencies across diverse tasks.</p></li><li><p>Outputs that, while detailed, often blur individual factual boundaries.</p></li><li><p>A notable absence of concrete sources/evidence accompanying the generated content.</p></li><li><p>An inherent tendency to produce text that, while sounding credible, might be riddled with inaccuracies.</p></li></ul><p>These limitations are especially critical in high-stakes domains like healthcare, finance, and law, emphasizing the dire need for rigorous fact-checking.</p><h2>The FacTool Solution</h2><p>Addressing the above challenges is FacTool &#8212; a domain-agnostic framework designed to detect factual errors in texts produced by LLMs. Think of FacTool as a guardian, ensuring the veracity of every piece of information generated by models like ChatGPT.</p><h2>Tested Domains for FacTool</h2><p>FacTool's versatility shines through its testing across multiple tasks:</p><ul><li><p><strong>Knowledge-based QA</strong>: Validating the accuracy of answers.</p></li><li><p><strong>Code Generation</strong>: Validating the efficacy of generated codes.</p></li><li><p><strong>Mathematical Reasoning</strong>: Ensuring accurate problem-solving.</p></li><li><p><strong>Scientific Literature Review</strong>: Validating the credibility of AI-generated citations and reviews.</p></li></ul><h2>How FacTool Works</h2><p>Harnessing the power of tool augmentation, FacTool operates through:</p><ol><li><p><strong>Claim extraction</strong>: LLMs, like ChatGPT, extract claims from the generated response. Claims are the sentences that need to be fact-checked. For instance, in Knowledge-based QA, each claim is a part of the answer generated, while in code generation, every snippet becomes a claim to verify.</p></li><li><p><strong>Query generation</strong>: These claims are transformed into queries that can be sent to external tools to retrieve factual responses.</p></li><li><p><strong>Tool querying &amp; Evidence collection</strong>: Relevant evidence is collated by sending the generated query to external tools such as Google Search, Python interpreters or Google Scholar APIs.</p></li><li><p><strong>Agreement verification</strong>: The final step involves presenting the claim and evidence back to the LLM, which then verifies the claim's authenticity.</p></li></ol><h2>FacTool's Stellar Performance</h2><p>Benchmarked against established methods like 3-shot chain-of-thought, FacTool, especially when powered by GPT-4, consistently outperformed competitors. Key findings include:</p><ul><li><p>Superior performance in scientific literature review tasks.</p></li><li><p>Enhanced sensitivity in error detection compared to self-check chain-of-thought methods.</p></li><li><p>In Knowledge-based QA, FacTool debunked false claims, such as "Argentina has not won the World Cup since 1986".</p></li><li><p>In code generation, it surpassed other baselines, showcasing its capability in technical domains.</p></li><li><p>Both GPT-4 and ChatGPT versions of FacTool excelled in mathematical reasoning, identifying errors in calculations.</p></li></ul><p><a href="https://arxiv.org/abs/2307.13528v2">Paper link</a></p><p><a href="https://github.com/GAIR-NLP/factool">Code link</a></p><h2>Conclusion</h2><p>With the world leaning heavily on LLMs for varied tasks, the importance of tools like FacTool cannot be overstated. Ensuring the factuality of generated content paves the way for a future where we can trust AI-generated outputs without second-guessing.</p><p>For those venturing into the world of LLMs, integrating FacTool can be a game-changer. Especially in critical domains, it's more than a tool&#8212;it's a necessity. As generative AI continues to evolve, ensuring the accuracy of generated content will remain paramount, and FacTool is leading the charge in this endeavour.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Using AI to Build AI: Prompt2Model]]></title><description><![CDATA[Building deployable NLP models using GPT-3.5 Turbo and Prompt2Model Framework.]]></description><link>https://ai-unltd.com/p/using-ai-to-build-ai-prompt2model</link><guid isPermaLink="false">https://ai-unltd.com/p/using-ai-to-build-ai-prompt2model</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sat, 02 Sep 2023 04:30:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hhDS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hhDS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hhDS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hhDS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg" width="728" height="546" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:424086,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hhDS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Building an AI model from the ground up can be tough. It includes deciding on the job the model will do, finding the right data, picking a model, training it, checking how good it is, and then using it. However, with the rise of Large Language Models (LLMs) like GPT-3, we can build rapid prototypes without diving deep into the coding intricacies. But is relying solely on LLMs the silver bullet we've been waiting for? Unfortunately, while LLMs promise convenience, they come with their own set of challenges, from escalating costs and slower predictions to potential privacy pitfalls. To counter this, Prompt2Model proposes that we build smaller models for specific tasks that need to be accomplished.</p><h2>The Promise of Prompt2Model</h2><p>Enter Prompt2Model &#8211; an innovative method designed to convert task descriptions into specialized, deployable models. This AI-driven tool is not just a mechanism for swift NLP system creation, but it also stands as a comprehensive research platform. Researchers can explore avenues in model distillation, synthetic evaluation, dataset retrieval, and more, all under the Prompt2Model umbrella.</p><h2>How Prompt2Model Works: An Overview</h2><p>At its core, Prompt2Model operates through a multi-pronged strategy. It harnesses existing datasets and pre-trained models, generates datasets with the assistance of LLMs like GPT-3.5 Turbo, and fine-tunes through supervised methodologies. What&#8217;s truly remarkable is its prowess. Tests have shown that Prompt2Model, with just a few-shot prompt as input, can outdo GPT-3.5 Turbo's results by an impressive average of 20%. This performance spike is achieved even when the resultant Prompt2Model is up to 700 times smaller in scale.</p><p>Note: The model that Prompt2Model builds is fine-tuned for a single task that it needs to accomplish while GPT-3.5 Turbo can accomplish multiple tasks.</p><h2>Delving Deeper: The Prompt2Model Framework</h2><p>Understanding Prompt2Model requires a dive into its intricate framework:</p><ul><li><p><strong>Prompt Parser</strong>: Here, user prompts, which could be simple task instructions or demonstrations, are decoded into tangible, actionable steps. This process ensures that the ML pipeline is optimized for user input.</p></li><li><p><strong>Dataset Retriever</strong>: With the support of DataFinder, this module scours the vast expanse of the digital realm to retrieve datasets that resonate with the task at hand.</p></li><li><p><strong>Dataset Generator</strong>: Recognizing that not all tasks are backed by existing data, this tool creates synthetic data using GPT-3.5 Turbo.</p></li><li><p><strong>Model Retriever</strong>: This module pulls an apt pre-trained model from the treasure trove of Hugging Face, ensuring the model aligns with the user's intent.</p></li><li><p><strong>Model Trainer</strong>: The chosen model undergoes fine-tuning, leveraging both the retrieved and the freshly minted datasets.</p></li><li><p><strong>Model Evaluator</strong>: Post-training, it's time for a performance check. This segment evaluates the model's accuracy and reliability.</p></li><li><p><strong>Demo Creator</strong>: An optional yet valuable feature, this tool crafts a graphical interface, allowing users to interact seamlessly with the trained model.</p></li></ul><p><a href="https://arxiv.org/abs/2308.12261?s=09">Paper Link</a></p><p><a href="https://github.com/neulab/prompt2model">Code Link</a></p><h2>The Real-World Benefits of Prompt2Model</h2><p>Prompt2Model isn't just a theoretical marvel, it promises tangible benefits. By navigating beyond the constraints of zero-shot and few-shot prompting, it delivers robust performance. In certain tasks, it even managed to overshadow GPT-3.5-turbo by an impressive average of 20%, showcasing its potential for real-world applications, especially given its efficiency and compact size.</p><h2>Limitations and Challenges</h2><p>However, Prompt2Model has a few limitations and challenges as well. Its reliance on GPT-3.5 Turbo, a paid and closed-source entity, raises eyebrows, especially when legal concerns come into play. OpenAI's policies might restrict Prompt2Model's commercial exploits. Additionally, there's a language barrier. Currently, tasks beyond the English realm might encounter hiccups, given GPT-3.5 Turbo's predominant English training.</p><h2>Conclusion</h2><p>Prompt2Model heralds a paradigm shift, underscoring the potential of using AI to construct AI models. As we stand on the cusp of this AI revolution, it's exhilarating to envision the future prospects, improvements, and the transformative impact Prompt2Model might usher in for the NLP and broader AI domain.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Making ChatGPT Think Like a Human: Unlocking Its Potential with Graph of Thoughts]]></title><description><![CDATA[Dive deeper into the intricate weave of Graph of Thoughts, edging ever closer to how humans think.]]></description><link>https://ai-unltd.com/p/making-chatgpt-think-like-a-human</link><guid isPermaLink="false">https://ai-unltd.com/p/making-chatgpt-think-like-a-human</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Fri, 01 Sep 2023 04:30:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!lvcw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lvcw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lvcw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lvcw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lvcw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lvcw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lvcw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg" width="510" height="382.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:510,&quot;bytes&quot;:355665,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lvcw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lvcw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lvcw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lvcw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff9ef9c54-62d0-42ee-b3a3-83962cb7ce9c_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The landscape of the Large Language Model (LLM) prompting frameworks has seen a new contender &#8211; the Graph of Thoughts (GoT) framework. Going beyond the capabilities of Chain-of-thought (CoT) and Tree-of-thoughts (ToT), GoT presents a fresh perspective on how we can make LLMs think like humans for better results.</p><h2>Background Information: What is a Graph?</h2><p>In essence, a graph is a structure consisting of vertices and edges. Think of vertices as dots and edges as the lines connecting these dots. In the world of problem-solving, these graphs, especially directed acyclic ones, play a pivotal role in representing complex structures.</p><h2>The Evolution of Prompting Structures</h2><h4>Chain-of-thought (CoT)</h4><p>CoT emerged as an innovative prompting approach by weaving intermediate reasoning steps directly into the prompt alongside the primary task input/output. By visualizing the thought process as a linear chain where each link represents a step of reasoning, CoT significantly amplified the ability of LLMs to tackle problems. Each "link" or step in this chain paved the way for the next, ensuring a more structured approach to problem-solving.</p><h4>Tree of Thoughts (ToT)</h4><p>Taking inspiration from CoT, the Tree of Thoughts (ToT) was designed to provide even more depth to the LLM reasoning process. Instead of a linear chain, ToT models reasoning as a tree, with branches representing different paths of thought. This branching mechanism introduced novel capabilities like backtracking from outcomes that didn't seem promising. However, while ToT added multiple pathways of reasoning, its tree-like structure sometimes proved restrictive, confining the thought process to its branches without allowing for more intricate interconnections.</p><p>With these two as the backdrop, the emergence of the Graph of Thoughts (GoT) represents a leap forward. Instead of chains or trees, GoT envisions the reasoning process as an intricate web, akin to graphs, allowing for a more interconnected and dynamic form of reasoning.</p><h2>Introducing Graph of Thoughts (GoT)</h2><h4>Core Concept of GoT</h4><p>GoT is motivated by numerous phenomena such as human reasoning, brain structure, or algorithm execution. When working on a novel idea, a human would generally form a more complex network of thoughts. For example, one could explore a certain chain of reasoning, backtrack, and start a new one, then realize that a certain idea from the previous chain could be combined with the currently explored one, and merge them both into a new solution, taking advantage of their strengths and eliminate their weaknesses.</p><p>GoT's brilliance lies in its ability to represent LLM-generated information as an arbitrary graph, where thoughts are vertices and their dependencies, the edges. This means GoT can amalgamate multiple thoughts, refine vast networks of thoughts, or even augment individual thoughts.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dOPf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dOPf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 424w, https://substackcdn.com/image/fetch/$s_!dOPf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 848w, https://substackcdn.com/image/fetch/$s_!dOPf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 1272w, https://substackcdn.com/image/fetch/$s_!dOPf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dOPf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png" width="311" height="368" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:368,&quot;width&quot;:311,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:48986,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dOPf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 424w, https://substackcdn.com/image/fetch/$s_!dOPf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 848w, https://substackcdn.com/image/fetch/$s_!dOPf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 1272w, https://substackcdn.com/image/fetch/$s_!dOPf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95f28f55-1b46-47cc-adc6-28e1bdf4abee_311x368.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Components of GoT</h4><ul><li><p><strong>Prompter:</strong> Transforms graph structures into LLM prompts.</p></li><li><p><strong>Parser:</strong> Decodes information from LLMs' output into usable 'thought states'.</p></li><li><p><strong>Scoring &amp; Validation:</strong> Determines the accuracy of LLM outputs, often quantified through scores.</p></li><li><p><strong>Graph of Operations:</strong> A predefined plan detailing the flow of thought operations.</p></li><li><p><strong>Graph Reasoning State:</strong> Maintains a record of the LLM reasoning process.</p></li><li><p><strong>Controller:</strong> Orchestrates the flow, deciding when to loop back or conclude the process.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PJd1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PJd1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 424w, https://substackcdn.com/image/fetch/$s_!PJd1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 848w, https://substackcdn.com/image/fetch/$s_!PJd1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 1272w, https://substackcdn.com/image/fetch/$s_!PJd1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PJd1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png" width="465" height="311" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bc8de185-fc37-45f7-b961-2582fdb72213_465x311.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:311,&quot;width&quot;:465,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:75652,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PJd1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 424w, https://substackcdn.com/image/fetch/$s_!PJd1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 848w, https://substackcdn.com/image/fetch/$s_!PJd1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 1272w, https://substackcdn.com/image/fetch/$s_!PJd1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc8de185-fc37-45f7-b961-2582fdb72213_465x311.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Evaluation and Results</h2><h4>Methodology</h4><p>The researchers utilized 100 input samples for each task, relying on a 4k context model. However, due to constraints, their focus remained on GPT-3.5 over GPT-4.</p><h4>Analysis of Outcomes</h4><p>The results are clear: GoT outshines both ToT and CoT. Compared to ToT, GoT has lower costs and reduces median errors by 62%. Against CoT, GoT improves results by 65%, albeit at a slightly higher cost. This showcases GoT's aptness for intricate problems, especially as they scale in complexity.</p><h4>Human Thought and GoT: Drawing Parallels</h4><p>Our brains don't think linearly. We merge ideas, backtrack, and often change our reasoning direction based on new insights. GoT mirrors this non-linear, dynamic approach, resembling the intricate networks formed in human reasoning.</p><p><a href="https://arxiv.org/abs/2308.09687">Paper link</a></p><p><a href="https://github.com/spcl/graph-of-thoughts">Code link</a></p><h2>Conclusion</h2><p>The Graph of Thoughts framework is an upgrade in the way we leverage Large Language Models. By drawing inspiration from the intricate networks of human reasoning and bridging gaps in previous models, GoT stands poised to inspire a new wave of LLM prompting frameworks.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Google's Solution to Detect AI Generated Images]]></title><description><![CDATA[In today's digital era, the boundary between reality and artificiality is rapidly blurring.]]></description><link>https://ai-unltd.com/p/googles-solution-to-detect-ai-generated</link><guid isPermaLink="false">https://ai-unltd.com/p/googles-solution-to-detect-ai-generated</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 31 Aug 2023 04:30:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xD6z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xD6z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xD6z!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xD6z!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xD6z!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xD6z!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xD6z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg" width="468" height="351" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:468,&quot;bytes&quot;:473566,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xD6z!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!xD6z!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!xD6z!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!xD6z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F518afc99-8cde-4625-ba4e-85e3dcb58adf_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In today's digital era, the boundary between reality and artificiality is rapidly blurring. AI-generated images, fueled by models such as stable diffusion, mid-journey, Google's Bard, and OpenAI's DALL-E 2, are gaining traction. But how do we discern their authenticity, especially when they appear convincingly realistic?</p><h2>The Problem: Realism and Deception in AI-Generated Images</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!W9k8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!W9k8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 424w, https://substackcdn.com/image/fetch/$s_!W9k8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 848w, https://substackcdn.com/image/fetch/$s_!W9k8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!W9k8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!W9k8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg" width="444" height="296" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:640,&quot;width&quot;:960,&quot;resizeWidth&quot;:444,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;An AI-generated image of \&quot;Balenciaga\&quot; Pope Francis&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="An AI-generated image of &quot;Balenciaga&quot; Pope Francis" title="An AI-generated image of &quot;Balenciaga&quot; Pope Francis" srcset="https://substackcdn.com/image/fetch/$s_!W9k8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 424w, https://substackcdn.com/image/fetch/$s_!W9k8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 848w, https://substackcdn.com/image/fetch/$s_!W9k8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!W9k8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805c0873-3306-4ffb-91fb-02a9c8cf4175_960x640.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">"Balenciaga" Pope Francis REDDIT/U/TRIPPY_ART_SPECIAL</figcaption></figure></div><p>The AI landscape is rife with instances where the line between real and fabricated is challenging to discern. A widely circulated image of Pope Francis clad in a Balenciaga puffer jacket and another portraying a meeting between Putin and Xi are prime examples. Such episodes underscore the urgent need for effective tools to identify AI-crafted visuals.</p><h2>DeepMind's Response: SynthID Unveiled</h2><p>Recognizing this challenge, Google's DeepMind offers a promising solution: SynthID. Designed as a watermarking tool, its primary function is to detect AI-generated images, a crucial step towards fostering trust in digital information. Though not a panacea for misinformation, SynthID emerges as a pioneering technical remedy to a growing AI safety concern.</p><h2>The Mechanics of SynthID</h2><p>SynthID operates through two interconnected deep-learning models. One focuses on watermarking, while the other zeroes in on identifying AI-generated content. These models, trained conjointly on a myriad of images, aim to detect watermarked content accurately and ensure the watermark's invisibility aligns seamlessly with the original image.</p><p><em><strong>Watermarking</strong></em>: Employing embedded watermarking technology, SynthID infuses a digital watermark within the pixels of AI-generated images, rendering it undetectable to human eyes. Prioritizing image quality, this watermark remains discernible even amidst image modifications, be it filter additions, colour alterations, or conversions to diverse formats.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3W9h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3W9h!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 424w, https://substackcdn.com/image/fetch/$s_!3W9h!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 848w, https://substackcdn.com/image/fetch/$s_!3W9h!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 1272w, https://substackcdn.com/image/fetch/$s_!3W9h!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3W9h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:948466,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/gif&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3W9h!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 424w, https://substackcdn.com/image/fetch/$s_!3W9h!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 848w, https://substackcdn.com/image/fetch/$s_!3W9h!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 1272w, https://substackcdn.com/image/fetch/$s_!3W9h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d36742f-91e8-46af-a29e-328de765f0d0_960x540.gif 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>Identification</strong></em>: Upon scanning an image, SynthID checks for the watermark's presence, offering users three levels of confidence:</p><ul><li><p>&#9989; Digital watermark detected: Image is AI-generated.</p></li><li><p>&#10060; Digital watermark absent: Image likely isn't AI-generated.</p></li><li><p>&#9888;&#65039; Digital watermark possibly found: Approach with caution, potential AI origin.</p></li></ul><p><a href="https://www.deepmind.com/blog/identifying-ai-generated-images-with-synthid">Read more here</a>.</p><h2>SynthID and Other Identification Techniques</h2><p>SynthID complements the broader gamut of digital content identification methods. A prevalent approach involves metadata, offering insights into the content's creator and creation time. Digital signatures within metadata can indicate image alterations. Yet, <strong>metadata's vulnerability to removal or edits poses a challenge</strong>. <strong>SynthID's pixel-embedded watermark, however, stands resilient even in the absence of metadata</strong>, ensuring compatibility with metadata-based identifiers.</p><h2>Conclusion</h2><p>As AI-generated content proliferates, tools like SynthID become imperative. By seamlessly integrating watermarking and metadata techniques, we can ensure a trustworthy digital environment. It's a collective call to action: tech giants, content creators, and consumers must unite to preserve the authenticity of our digital realm.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[RecMind: Personalised Recommendation using ChatGPT]]></title><description><![CDATA[An innovative approach to use LLMs for recommendation based tasks]]></description><link>https://ai-unltd.com/p/recmind-personalised-recommendation</link><guid isPermaLink="false">https://ai-unltd.com/p/recmind-personalised-recommendation</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Wed, 30 Aug 2023 04:30:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!R8f4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R8f4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R8f4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!R8f4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!R8f4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!R8f4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R8f4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg" width="598" height="448.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:598,&quot;bytes&quot;:358899,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!R8f4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!R8f4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!R8f4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!R8f4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f8b361-3d1f-4378-9394-d31672ef0bf7_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Large Language Models (LLMs) have made a mark by showcasing abilities to execute complex tasks, from solving math problems to sparking creative writing. Yet, they falter when faced with personalized queries, especially recommendation requests. Enter RecMind, an innovative research born from the collaboration between Amazon Alexa AI and Arizona State University. This LLM-powered autonomous recommender agent is tailored to address this glaring gap.</p><h2>The Imperative Role of Recommender Systems</h2><p>Every time you search on Google, shop on Amazon, or scroll through your social media feeds, you're interacting with a Recommender System. These systems, pivotal in various internet platforms, suggest potential items or content based on your past interactions. Modern Recommender Systems, supercharged by Deep Neural Networks (DNNs), are getting better at understanding user behaviours and preferences. However, many still struggle with capturing the depth of textual knowledge about users and items, particularly due to the limitations in model size and data volume.</p><h2>RecMind: A New LLM Innovation In Recommendation</h2><p>RecMind stands out by tapping into external tools, bringing in real-time information and domain-specific knowledge.</p><p>Structured meticulously, RecMind is divided into three core parts:</p><ul><li><p><strong>Planning</strong>: Akin to breaking down a problem, this helps in decomposing complex recommendation tasks into smaller, digestible chunks and prompts.</p></li><li><p><strong>Memory</strong>: Moving beyond the inherent data within LLMs, it comprises Personalized Memory, storing individual user data, and World Knowledge, a reservoir of real-time and domain-specific knowledge.</p></li><li><p><strong>Tools</strong>: These are the powerhouses like the Database tool, Search tool, and Text summarization tool that amplify RecMind's functionality.</p></li></ul><p>One of RecMind's standout features is its unique "Self-Inspiring" algorithm. This allows the model to reflect on previously explored paths and use that historical data for better recommendations.</p><h2>RecMind in Action</h2><p>RecMind is no slouch when it comes to practical application. It excels in:</p><ul><li><p><strong>Rating prediction</strong>: Predicting how a user would rate a particular item.</p></li><li><p><strong>Sequential recommendation</strong>: Suggesting items based on the user's past interactions.</p></li><li><p><strong>Direct recommendation</strong>: Predicting future user interactions based on a dataset.</p></li><li><p><strong>Explanation generation</strong>: Crafting textual explanations for user-item interactions.</p></li><li><p><strong>Review summarization</strong>: Condensing lengthy reviews into succinct titles.</p></li></ul><p>When measured against its contemporaries, RecMind frequently surpasses other LLM-based recommendation methods. <strong>Remarkably, its performance is on par with the P5 model, a fully pre-trained giant</strong>. What sets RecMind apart even further is its efficiency. While P5 demands a hefty investment in terms of training time, effort, and resources, RecMind delivers similar results without such exhaustive prerequisites. <strong>This makes RecMind not only a competitive alternative but perhaps a more pragmatic choice for many applications</strong>.</p><p><a href="https://arxiv.org/abs/2308.14296?s=09">Paper link here</a>.</p><h2>Conclusion</h2><p>In the digital age where personalization is key, RecMind offers a promising horizon for recommendation systems. With its ability to efficiently plan, remember, and utilize a suite of tools, it might just redefine the future of user interactions and suggestions. Given the resource-intensive nature of competitors like P5, RecMind not only matches up but could potentially lead the way.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Jailbreaking (Hacking) ChatGPT]]></title><description><![CDATA[Jailbreak prompts can make ChatGPT generate harmful content]]></description><link>https://ai-unltd.com/p/jailbreaking-hacking-chatgpt</link><guid isPermaLink="false">https://ai-unltd.com/p/jailbreaking-hacking-chatgpt</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sun, 27 Aug 2023 18:03:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JQ6L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JQ6L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JQ6L!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JQ6L!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JQ6L!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JQ6L!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JQ6L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg" width="564" height="564" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:564,&quot;bytes&quot;:521461,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JQ6L!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!JQ6L!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!JQ6L!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!JQ6L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952cbfa5-89f3-4416-a5a4-ce90afba19e3_1024x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Recently, Large Language Models (LLMs) have emerged as powerful tools with transformative potential. However, alongside their capabilities come concerns. Recent incidents have underscored LLMs' risks, from spreading misinformation and conspiracy theories to facilitating spear phishing attacks and hate campaigns. A security firm's report only adds fuel to the fire, shedding light on the exploitation of ChatGPT for cybercriminal activities.</p><h2><strong>The Defensive Measures: LLM Safety Protocols</strong></h2><p>OpenAI introduced reinforcement learning from human feedback (RLHF) to combat these risks. This technique aligns ChatGPT with human values, ensuring the model's responses align more closely with human intent. Furthermore, external safeguards have been developed, acting as a secondary layer of defence. These safeguards detect and block inputs or outputs that fall into predefined harmful or inappropriate categories, substantially reducing potential harm. <strong>But, that&#8217;s not enough!</strong></p><h2><strong>Jailbreak Prompts: The New Adversary</strong></h2><p>However, a new challenge has arisen: jailbreak prompts. These craftily designed adversarial prompts bypass existing safeguards, manipulating LLMs to generate harmful content. Evolving continuously, these prompts are a testament to the ingenuity of adversaries seeking to exploit LLMs for malicious ends.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1EGS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1EGS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 424w, https://substackcdn.com/image/fetch/$s_!1EGS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 848w, https://substackcdn.com/image/fetch/$s_!1EGS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 1272w, https://substackcdn.com/image/fetch/$s_!1EGS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1EGS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png" width="528" height="430.96216216216214" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:604,&quot;width&quot;:740,&quot;resizeWidth&quot;:528,&quot;bytes&quot;:142445,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1EGS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 424w, https://substackcdn.com/image/fetch/$s_!1EGS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 848w, https://substackcdn.com/image/fetch/$s_!1EGS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 1272w, https://substackcdn.com/image/fetch/$s_!1EGS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2cf8ef4-aeff-448e-b4ab-95c4316c89a3_740x604.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Example Jailbreak prompt</figcaption></figure></div><h2><strong>The Study: Unmasking the Efficacy of Jailbreak Prompts</strong></h2><p>A comprehensive study was undertaken to understand the potency of these jailbreak prompts. Spanning from December 2022 to May 2023, the study collected a staggering 6,387 prompts across platforms such as Reddit, Discord, websites, and open-source datasets. Among these, 666 were identified as jailbreak prompts.</p><p>Four major LLMs - ChatGPT (both GPT-3.5 and GPT-4), ChatGLM, Golly, and Vicuna - served as subjects for these experiments. Results were startling: while some LLMs exhibited resistance against certain harmful queries, they struggled against jailbreak prompts. In fact, two prompts stood out, boasting a whopping 99% attack success rate against ChatGPT.</p><p><a href="https://arxiv.org/abs/2308.03825">Research paper link</a> (Disclaimer: This paper contains examples of harmful language. Reader discretion is recommended).</p><h2><strong>Distinguishing Characteristics of Jailbreak Prompts</strong></h2><p>Delving deeper, the study sought to understand what made jailbreak prompts so effective. On Reddit, while a regular prompt averaged a token count of 178.686, a jailbreak prompt clocked in at 502.249 tokens. This significant discrepancy is believed to arise from the necessity for attackers to use more intricate instructions to bypass safeguards.</p><p>Moreover, the toxicity level of jailbreak prompts was found to be higher than regular prompts. Regular prompts had a toxicity score of 0.066, while jailbreak prompts averaged a score of 0.150. These prompts not only contained more instructions but also exhibited higher toxicity and closely resembled regular prompts in the semantic space.</p><h2><strong>The Resistance: How LLMs Stand Against Jailbreak Prompts</strong></h2><p>Certain LLMs, especially those trained with RLHF, did show initial resistance to forbidden topics. Models like Vicuna, which were fine-tuned on data generated by RLHF-trained models, also demonstrated some degree of resistance. This indicates the effectiveness of built-in safeguards like RLHF in certain scenarios. However, the reality remains: these safeguards are not foolproof against the rising tide of jailbreak prompts.</p><h2>Platforms Fueling Jailbreak Prompt Evolution</h2><p>The evolution and discussions surrounding jailbreak prompts have found a home on platforms like Reddit, Discord, and certain Websites. Subreddits such as <code>r/ChatGPT</code>, <code>r/ChatGPTPromptGenius</code>, and <code>r/ChatGPTJailbreak</code> have become hotbeds for refining and sharing these jailbreaks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CfiW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CfiW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 424w, https://substackcdn.com/image/fetch/$s_!CfiW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 848w, https://substackcdn.com/image/fetch/$s_!CfiW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 1272w, https://substackcdn.com/image/fetch/$s_!CfiW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CfiW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png" width="490" height="373.21867321867325" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/93a34e53-3398-432f-b289-35bc823da905_814x620.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:620,&quot;width&quot;:814,&quot;resizeWidth&quot;:490,&quot;bytes&quot;:120973,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CfiW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 424w, https://substackcdn.com/image/fetch/$s_!CfiW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 848w, https://substackcdn.com/image/fetch/$s_!CfiW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 1272w, https://substackcdn.com/image/fetch/$s_!CfiW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93a34e53-3398-432f-b289-35bc823da905_814x620.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Data sources</figcaption></figure></div><p><code>#Posts = Number of posts</code></p><p><code>#P = Number of prompts</code></p><p><code>#J = Number of jailbreak prompts</code></p><h2>Conclusion: The Road Ahead</h2><p>The existence and evolution of jailbreak prompts underscore the need for continuous research and development in LLM safety. While current safeguards have made strides in ensuring model safety, the dynamic nature of adversarial threats means there's no room for complacency. As we stride forward, the onus is on both the research community and LLM vendors to ensure these powerful tools are used safely and responsibly. The research paper referred to in this post is aimed at facilitating the research community and LLM vendors in promoting safer and regulated LLMs.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Getting Accurate Math Answers With ChatGPT (GPT-4)]]></title><description><![CDATA[Using GPT-4's Code Interpreter and Code-base Self-Verification (CSV)]]></description><link>https://ai-unltd.com/p/getting-accurate-math-answers-with</link><guid isPermaLink="false">https://ai-unltd.com/p/getting-accurate-math-answers-with</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sat, 26 Aug 2023 04:30:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!FY2d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FY2d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FY2d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!FY2d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!FY2d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!FY2d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FY2d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg" width="552" height="414" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:552,&quot;bytes&quot;:425966,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FY2d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!FY2d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!FY2d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!FY2d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfc74139-d0f8-4192-9fa7-1f40415249a4_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">R2-D2 solving math</figcaption></figure></div><p>In recent times, the world of Large Language Models (LLMs) has seen exponential growth, notably with models like GPT-4 and PaLM-2. These models have revolutionized how we approach math reasoning problems, and with OpenAI's latest innovations, they're only getting better.</p><h2><strong>GPT-4's Recent New Avatar: The Code Interpreter</strong></h2><p>OpenAI fine-tuned GPT-4 to run Python code, it&#8217;s called the GPT-4 <a href="https://openai.com/blog/chatgpt-plugins#code-interpreter">Code Interpreter</a>, this model has showcased exemplary performance on some of the most challenging math datasets available. This Code Interpreter, dubbed GPT4-Code, has the prowess to:</p><ul><li><p>Offer logical natural language reasoning.</p></li><li><p>Generate and execute Python code step by step.</p></li><li><p>Deliver the executed code's results back to the LLM, enhancing its decision-making.</p></li></ul><h2><strong>The Magic Ingredient: Code-based Self-Verification (CSV)</strong></h2><p>GPT-4's success isn't magic, but it sure feels like it. The key lies in its unmatched capability to generate and execute code, evaluate the outcomes, and make necessary corrections based on unreasonable outputs. This entire mechanism is backed by the GPT-4 Code Interpreter.</p><p>To further enhance this process, Code-based-Self-Verification (CSV) can be used. This method amplifies GPT-4's mathematical reasoning potential. In cases where the self-verification process identifies discrepancies, GPT-4 takes the lead, amending its solutions, and rectifying errors. This is not just about generating and executing code; it's about the model's ability to adjust its strategies based on feedback.</p><p>CSV guides GPT4-Code to:</p><ul><li><p>Generate additional verification code.</p></li><li><p>Refine reasoning steps in case of discrepancies.</p></li></ul><p>The outcomes?</p><ul><li><p>Incorrect solutions are promptly rectified (corrected).</p></li><li><p>Solutions that are verified resemble the reliability of human problem-solving.</p></li></ul><h2><strong>From GPT-4 to GPT4-Code: A Comparative Analysis</strong></h2><p>When GPT4-Code was put to the test, it recorded an impressive <strong>69.7%</strong> accuracy on the intricate MATH dataset. A significant leap from GPT-4's previous score of 42.2%. But with the Code-based Self-Verification approach, <strong>GPT4-Code further pushed its boundaries, achieving an astounding accuracy of 84.32%.</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mrRe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mrRe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 424w, https://substackcdn.com/image/fetch/$s_!mrRe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 848w, https://substackcdn.com/image/fetch/$s_!mrRe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 1272w, https://substackcdn.com/image/fetch/$s_!mrRe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mrRe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png" width="1322" height="588" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:588,&quot;width&quot;:1322,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:144002,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mrRe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 424w, https://substackcdn.com/image/fetch/$s_!mrRe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 848w, https://substackcdn.com/image/fetch/$s_!mrRe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 1272w, https://substackcdn.com/image/fetch/$s_!mrRe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbaeee1a-57e7-415f-b394-acdb7d3bfa51_1322x588.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Prompt Examples: GCD and LCM</strong></h2><ul><li><p><em>Basic prompt:</em> "<strong>Solve the problem and put your answer in \bracket{}. The problem is:</strong> The greatest common divisor of positive integers m and n is 6. The least common multiple of m and n is 16. What is the least possible value of <em>m </em>+ <em>n</em>?"</p></li><li><p><em>Code-based Self-Evaluation prompt:</em> "<strong>Solve the problem using the code interpreter step by step, and please verify your answer using the code interpreter. This problem is:</strong> The greatest common divisor of positive integers m and n is 6. The least common multiple of m and n is 16. What is the least possible value of <em>m </em>+ <em>n</em>?"</p></li></ul><h2><strong>Experience it Yourself: Trying Out GPT4-Code</strong></h2><p>For the enthusiasts, here's how you can dive in:</p><ol><li><p>Acquire a ChatGPT Plus subscription.</p></li><li><p>Navigate to 'settings' -&gt; 'beta features' and activate the Code Interpreter.</p></li><li><p>Choose the GPT-4 tab and opt for the Code Interpreter.</p></li><li><p>Insert the Code-based Self-Evaluation prompt with your problem and watch the magic unfold.</p></li></ol><p><a href="https://arxiv.org/abs/2308.07921?s=09">Link to paper</a>.</p><h2><strong>Conclusion</strong></h2><p>GPT-4 and its Code Interpreter are changing the game in mathematical problem-solving. As technology progresses, we're on the brink of reshaping the landscape of automated math challenges, all thanks to models like GPT-4.</p><p>Want to stay ahead in the world of AI-driven problem-solving? &#128640; Dive deeper into the wonders of GPT-4, mathematical reasoning, and the future of automated solutions. Subscribe to this newsletter now and never miss an update on the revolutionary advancements of ChatGPT and more!</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Code Llama: The Future of Writing Code is Here]]></title><description><![CDATA[Meta launches Code Llama - A comprehensive coding companion, ready to assist, guide, and innovate]]></description><link>https://ai-unltd.com/p/code-llama-the-future-of-writing</link><guid isPermaLink="false">https://ai-unltd.com/p/code-llama-the-future-of-writing</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Fri, 25 Aug 2023 03:30:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Id4I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Id4I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Id4I!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Id4I!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Id4I!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Id4I!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Id4I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg" width="552" height="414" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:552,&quot;bytes&quot;:393355,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Id4I!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Id4I!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Id4I!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Id4I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd4952251-8e6a-4fa7-8434-639f24dc88a3_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In the world of software development, the concept of an AI coder is just taking flight, starting a new era of automated programming and enhanced developer productivity. Many developers have already experienced the magic of GitHub Copilot. It's like having a co-pilot for your coding journey, making the process smoother and more efficient. But what if I told you the AI coder landscape just got a significant upgrade?</p><p><em><strong>Enter Code Llama.</strong></em></p><h2>What is Code Llama?</h2><p>Code Llama is an open-sourced Large Language Model(LLM) that can <strong>assist in coding-related tasks</strong>. In short, it&#8217;s a <strong>fine-tuned Llama 2 model that can code</strong>. While GitHub Copilot is a game-changer, Code Llama takes it a notch higher. It's a state-of-the-art Large Language Model (LLM) designed to generate code and even provide natural language explanations about the code it produces. Think of it as an enhanced co-pilot, not just suggesting code but also explaining its logic.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;66bf3a67-76eb-4e88-8ada-37fed5f9f29a&quot;,&quot;duration&quot;:null}"></div><h2>What Can Code Llama Do?</h2><p>Code Llama isn't just another AI tool; it's a game-changer. Here's why:</p><ul><li><p><strong>Code Generation &amp; Explanation:</strong> It crafts code and provides natural language explanations, making complex tasks understandable.</p></li><li><p><strong>Code Completion:</strong> With its fill-in-the-middle capability, it seamlessly completes code in real-time.</p></li><li><p><strong>Multilingual Support:</strong> From Python to Java to C#, Code Llama has a wide language range.</p></li><li><p><strong>Deep Context:</strong> It understands up to 100,000 tokens of context, ensuring accurate code generations.</p></li><li><p><strong>Specialized Versions:</strong> With variants like Code Llama - Python and Code Llama - Instruct, it caters to specific needs.</p></li></ul><h2>The Power Behind Code Llama</h2><p>Built on the robust foundation of Llama 2, Code Llama comes in three specialized variations:</p><ul><li><p>A foundational model for general coding tasks.</p></li><li><p>A Python-specialized version that can also <strong>write deep learning models in PyTorch</strong>.</p></li><li><p>An instruction-fine-tuned version, designed to understand and generate code from natural language prompts.</p></li></ul><h2>Performance Metrics</h2><p>For those familiar with GitHub Copilot's efficiency, Code Llama sets new benchmarks. In tests like HumanEval and Mostly Basic Python Programming (MBPP), Code Llama showcased superior performance, making it a formidable tool in the AI coder arsenal.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WYD3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WYD3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WYD3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WYD3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WYD3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WYD3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg" width="506" height="617.7142857142857" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1316,&quot;width&quot;:1078,&quot;resizeWidth&quot;:506,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WYD3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WYD3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WYD3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WYD3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F90f29868-f1c3-4867-8495-b56f474acd9a_1078x1316.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Code Llama comparison results from Meta</figcaption></figure></div><p>Further reading:</p><ul><li><p><a href="https://github.com/facebookresearch/codellama">Code link</a></p></li><li><p><a href="https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/">Research paper link</a></p></li><li><p><a href="https://ai.meta.com/blog/code-llama-large-language-model-coding">Blog post link</a></p></li></ul><h2>The Future of AI Coders</h2><p>With tools like GitHub Copilot setting the stage and Code Llama elevating the game, the future of AI coders looks promising. As these tools become more integrated into developers' workflows, I can anticipate faster development cycles, fewer bugs, and more time for developers to focus on innovative solutions.</p><p>In conclusion, the era of AI coders is here to stay. From GitHub Copilot's initial introduction which was widely adopted by developers to the advanced capabilities of Code Llama, AI is reshaping the coding landscape, making it more efficient, intuitive, and accessible.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>