<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[The Reasoning Engine]]></title><description><![CDATA[Musings on LLMs for Software Engineers]]></description><link>https://reasoningengine.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!F4OX!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30fd94e4-bea0-409d-821d-4fcd2a0074c0_400x400.jpeg</url><title>The Reasoning Engine</title><link>https://reasoningengine.ai</link></image><generator>Substack</generator><lastBuildDate>Sat, 18 Apr 2026 00:49:55 GMT</lastBuildDate><atom:link href="https://reasoningengine.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Rogerio Chaves]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[reasoningengine@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[reasoningengine@substack.com]]></itunes:email><itunes:name><![CDATA[Rogerio Chaves]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rogerio Chaves]]></itunes:author><googleplay:owner><![CDATA[reasoningengine@substack.com]]></googleplay:owner><googleplay:email><![CDATA[reasoningengine@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rogerio Chaves]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Data Flywheel]]></title><description><![CDATA[Using your production data to build better LLM products]]></description><link>https://reasoningengine.ai/p/data-flywheel</link><guid isPermaLink="false">https://reasoningengine.ai/p/data-flywheel</guid><dc:creator><![CDATA[Rogerio Chaves]]></dc:creator><pubDate>Fri, 28 Jun 2024 10:21:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4L8T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post was originally shared here: <a href="https://langwatch.ai/blog/data-flywheel">https://langwatch.ai/blog/data-flywheel</a></em></p><p>It does not come as a shock for anybody when I say that data is one of the most valuable assets a company can have, this has been true for some years now. Valuable data can be used to build very powerful products, and with the invention of LLMs, it became even more obvious that companies that already have valuable data internally can leverage AI to create a bigger edge.</p><p>But internal data is not the one I want to talk about today, I want to talk about the data that is produced as a byproduct of your AI being in production: all the interactions your users have with it, both with happy and unhappy outcomes. This is super valuable data, it&#8217;s literally the best insight on how users expect your product to behave, and now it comes in conversational text format, it's not just numbers like traditional product analytics, so it can directly feed back into the AI.</p><p>The thing with LLMs, is that they are super powerful and general, this is the exact reason we want to use them that, they can answer any questions and execute a large range of tasks, however exactly because of that, their surface area to the real world usage is extremely large, the more they do the larger it is, resulting in countless different ways it can fall short. It is hopeless to try and predict all the ways your users will want to use your product, and try covering all the possibilities and edge cases beforehand.</p><h4>How can a Data Flywheel help LLM apps?</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4L8T!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4L8T!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 424w, https://substackcdn.com/image/fetch/$s_!4L8T!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 848w, https://substackcdn.com/image/fetch/$s_!4L8T!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 1272w, https://substackcdn.com/image/fetch/$s_!4L8T!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4L8T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png" width="928" height="844" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:844,&quot;width&quot;:928,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:48115,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4L8T!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 424w, https://substackcdn.com/image/fetch/$s_!4L8T!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 848w, https://substackcdn.com/image/fetch/$s_!4L8T!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 1272w, https://substackcdn.com/image/fetch/$s_!4L8T!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7ff8160-4aa0-4ee9-ae13-1ce0a29e4424_928x844.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We call it a flywheel when an improvement brings business value and that loops back on further improvements. Since the number of possibilities of an LLM product is too vast - think of all possible ways a user can try using your product, or even compose a sentence - nothing is better at selecting what is more valuable and relevant like the real world, this makes it essential to monitor and get insights on where the product is suboptimal, or where it could bring even more value.</p><p>Then by using those insights for improving the product and making it more valuable, it will receive more usage and more users, it will gain more trust and cover more cases to be used in more places. By getting more usage, even more scenarios are revealed, new paths to bring even more value, and new shortcomings to overcome and make the product more robust, which will lead to even more usage an so on. That&#8217;s the flywheel.</p><p>The advantage with AI over traditional software improvements, is that this connection is even more direct, LLMs can be used also to evaluate, classify and synthesize data that will improve the product, so this flywheel is closed even tighter together.</p><p>There is, however, no shortcuts, if your solution has been finding new unexpected edge cases for months or years, it is not possible for someone else to just guess which ones they are and port them over elsewhere without going through the same discovery process. The improvement process you done on your product for your specific domain then creates an edge that sets you apart from others as soon as you start it.</p><p>This is a good news, if you are doing it.</p><h4>LangWatch</h4><p>I started building <a href="https://langwatch.ai/">LangWatch</a> on top of this realization, to help you maximize your data flywheel. Our platform helps both on collection part, with monitoring, evaluations, insights, collecting user feedback and annotating with domain experts from your team, and on the improvement part, running experiments, validating and guaranteeing the improved quality of each iteration.</p><p>If you are interested, help us out with a <a href="https://github.com/langwatch/langwatch">star on github</a> &#11088;</p>]]></content:encoded></item><item><title><![CDATA[Adventures with DSPy]]></title><description><![CDATA[LLM development getting to its maturity stage]]></description><link>https://reasoningengine.ai/p/adventures-with-dspy</link><guid isPermaLink="false">https://reasoningengine.ai/p/adventures-with-dspy</guid><dc:creator><![CDATA[Rogerio Chaves]]></dc:creator><pubDate>Sat, 01 Jun 2024 14:22:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VMeh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the past few weeks I&#8217;ve became obsessed with <a href="https://dspy-docs.vercel.app/">DSPy</a>, when I first saw the video <a href="https://www.youtube.com/watch?v=41EfOY0Ldkc">DSPy Explained</a> by Connor Shorten everything immediately clicked.</p><p>This is the LLM framework I was waiting for, this is the idea I was trying to get to but didn&#8217;t knew exactly how, and oh how coherent all the pieces are! It&#8217;s a great conceptual abstraction, I noticed myself not needing the docs all the time, even though the library is not properly typed right now, because things just *make sense* together.</p><p>If you haven&#8217;t dig into it yet, DSPy main selling point is that you just define your pipeline declaratively without worrying about the prompts to the LLM, and then it can optimize the prompts for you automatically. Of course, why should humans be trying to prompt engineer, when we have the best machines to do that: LLMs and trial-and-error.</p><p>And that&#8217;s kinda obvious right, of course I thought of that before, of course a lot of people thought about that before, having another LLM to improve your prompts, in fact we already have plenty of those like <a href="https://promptperfect.jina.ai">PromptPerfect</a>, however things still felt iffy, kinda unreliable, relying just on feeling, or on a bit of uncanny abstraction. But DSPy looked at PyTorch for inspiration, and this, turns out, was the right abstraction.</p><p>Many devs adventured themselves at the problem, and different angles were attempted, I personally took a <a href="https://github.com/langwatch/langevals/#unit-test-evaluations-with-pytest">TDD approach</a> at the problem, which I still think is the right one for sanity checks, I&#8217;ve also tried <a href="https://github.com/rogeriochaves/langstream">Reactive Streams</a> FP-style approach, some inventented this weird <a href="https://www.langchain.com/">&#8220;chains&#8221; concept</a> which nobody really likes, and are now attempting <a href="https://python.langchain.com/v0.1/docs/langgraph/">Graphs</a> angle.</p><p>But turns out, comming from the Machine Learning angle with PyTorch abstraction like DSPy did was the actual right approach, the one that makes me and all devs I talk to go like &#8220;oooh that is genius! It makes a lot of sense&#8221; and have this sense of relief that finally looks like someone had the right jab at the problem and you&#8217;ll be able to build something reliable.</p><p>Of course DSPy doesn&#8217;t seem the best for every problem right now, it seems to make the most sense for problems with well defined discrete outputs, like traditional supervised learning, for example when you want to use LLMs as a classifier. However, it is a clear a step on the right direction, and you can see a near future where more &#8220;feeling-based&#8221; evaluators will be also very effective and ubiquotous, for example, if you are writing a children&#8217;s stories chatbot and want to reliably evaluate how boring or exciting a story is.</p><p>The main value that DSPy brings to the industry then, is not (only) that it optimizes prompts for you allowing you to switch models whenever you like, but that it actually turns the problem on its head: it forces you to change your mindset, thinking first about what does it mean to be a good output or not with examples (like TDD promotes!) before the prompt, freeing you think in a more abstract level about the structure and so, and let the machine do machine-things of finding the best fit given the problem at hand.</p><p>Being so bearish in DSPy, I really want to help it grow and contribute to the community because I think it has a lot of future and it is the ahead. And I mean conceptually, even if something else would superseed DSPy with a different name.</p><p>Turns out, I already have a startup in the LLM space, <a href="https://langwatch.ai/">LangWatch</a>, so as I played with DSPy for LangWatch experiments, I naturally noticed a piece missing that could help me a lot. Thus, DSPy Visualizer was born.</p><h2>DSPy Visualizer</h2><p>DSPy Visualizer is open-source, and it allows you to log your DSPy training sessions, track the performance, costs, compare runs and debug them in detail.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VMeh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VMeh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 424w, https://substackcdn.com/image/fetch/$s_!VMeh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 848w, https://substackcdn.com/image/fetch/$s_!VMeh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 1272w, https://substackcdn.com/image/fetch/$s_!VMeh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VMeh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png" width="1456" height="900" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:221195,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VMeh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 424w, https://substackcdn.com/image/fetch/$s_!VMeh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 848w, https://substackcdn.com/image/fetch/$s_!VMeh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 1272w, https://substackcdn.com/image/fetch/$s_!VMeh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfcaad3b-33f8-43be-ab1d-f50b79c9dd09_1573x972.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">DSPy Visualizer</figcaption></figure></div><p>It&#8217;s still early days for DSPy or any kinda of automated LLM optimizer, so it really helps a lot being able to understand what is the optimizer trying to do, where is it going, and specially, how much is it costing, as too many examples and too complex pipelines can make your OpenAI credits go down the drain in a heartbeat. Even though <a href="https://reasoningengine.ai/p/llms-price-reduction-timeline">LLMs are getting cheaper</a>, we are pushing the automation limits here.</p><p>Iteration is the word of order with DSPy, although a much more confortable, reliable iteration than &#8220;classical&#8221; prompt engineering of course, and we found that looking at the examples being used, the LLM calls being made, and what is happening on all those optimization attempts that is helping it go up or down, gave us lots of ideas of what we can try next, or what is wrong so we can stop the iteration earlier and fix it.</p><p>I&#8217;ve posted an official blogpost on the DSPy Visualizer on our <a href="https://langwatch.ai/blog/dspy-visualizer">LangWatch blog</a>, if you want to try it out, check out the <a href="https://docs.langwatch.ai/dspy-visualization/quickstart">quickstart guide</a>.</p><p>If you are in for more DSPy content, <a href="https://twitter.com/_rchaves_">follow me on twitter</a> and subscribe to this substack!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://reasoningengine.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The Reasoning Engine! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[In defense of vibe-checking]]></title><description><![CDATA[As AIs approach humans, so does our evaluation on how well they are working.]]></description><link>https://reasoningengine.ai/p/in-defense-of-vibe-checking</link><guid isPermaLink="false">https://reasoningengine.ai/p/in-defense-of-vibe-checking</guid><dc:creator><![CDATA[Rogerio Chaves]]></dc:creator><pubDate>Tue, 02 Apr 2024 10:30:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!DPM5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DPM5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DPM5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DPM5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DPM5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DPM5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DPM5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg" width="577" height="432" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:432,&quot;width&quot;:577,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:89714,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DPM5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 424w, https://substackcdn.com/image/fetch/$s_!DPM5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 848w, https://substackcdn.com/image/fetch/$s_!DPM5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!DPM5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2873a72-6a08-4eea-9244-9376fe410120_577x432.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As AIs approach humans, so does our evaluation on how well they are working. &#8220;Vibe-checking&#8221; is the act of evaluating an LLM and deploying it in production just by eye-balling and *feeling* if it is good enough, but is that a problem?</p><p>Engineering is regarded as applied science to solve practical problems. We think it&#8217;s like applied physics, where knowledge is derived from first principles, in all reality, it looks much more like natural sciences, where constant tinkering leads us to new discoveries. Just look at all the recent advances in AI and the papers coming out of it, they are all about tinkering, trying stuff until it works. Attempts to explain *why* they work comes much later.</p><p>Even though this has always been the case for all areas of engineering (or do you think humans only started building houses *after* we had geometry? Nope, practice comes first, theory after), we have this general feeling in retrospect that engineers first do the calculations and make something work in theory before they do in practice. This makes us feel weird and funny about people deploying all those chatbots without knowing how they will act and what they will do, and having no answers whatsoever if it&#8217;s good or not other a few query examples and then &#8220;vibe-checking&#8221;.</p><p>How do we know that something is working well? If you go back to civil engineering, you can actually make the calculations of the structure and a simulation of the whole building, which matches pretty close to real life. If you go to a car maker manufactured, they have all those crash test dummies, that&#8217;s how they know their theorized safety really works on real life. You plan, you building, you run a battery of tests trying to cover all real life scenarios, and you deploy, with a high success rate.</p><p>On software, it&#8217;s even easier. Computers follow a completely predictable logic, that&#8217;s literally their definition, so <a href="https://www.bbc.com/future/article/20221011-how-space-weather-causes-computer-errors">unless you are in outer space</a>, a computer program will be executed the same and behave the same 100% of the times. This means if we have a financial application, we can guarantee the transactions calculations are working correctly, by having not only real world examples for the unit tests, but really deriving from first principles to proof the algorithm is correct if needed so. How come software has bugs then? Well, because we can take bugs, we are comfortable with a certain level of them, so we increased the complexity of our software by a lot (the whole world is connected), and our testing and safety practices followed just enough behind.</p><p>Then, in the past decade, as ML gained more popularity, we started to adapt to a somewhat level of uncertainty in technology, but still, a measured uncertainty. When a classification model has 90% accuracy, you know what to expect how often it will be wrong, in what domain. When a regression has a mean absolute error of $50K for predicting apartment prices in New York, you know that&#8217;s still a pretty damn accurate model.</p><p>But as AI get more and more general, executing more and more tasks, how do you evaluate how good they are? You cannot cover all real-life scenarios, if you could, then you wouldn&#8217;t need a general-purpose agent.</p><p>Well, turns out, other than technology, there was something else all along with a lot of unpredictable performance: humans. Yes, turns out we have been handling billions of those completely unpredictable creatures for a while now. They have an advantage though: they are very general and can execute about every type of task there is.</p><p>How can you uniformly evaluate humans? Is there even a way? Well, turns out, there is, we do it all the time with children, by scoring their school tests, having even standardized exams on a national level like the SATs, or law school and medical school exams.</p><p>This is pretty much the same as we have been doing with AI now, we give them standardized benchmarks, and as they approach human-level and benchmarks quickly get obsolete, we come up with more and more benchmarks to somehow measure how good they are, even the actual SATs we give humans.</p><p>But both you and I know that&#8217;s not enough, if it were, why would companies still do interviews with candidates before hiring them as employee? If they passed all standardized tests, why do we still need to talk to them, what are companies doing there? Well, they are vibe-checking of course!</p><p>As the work becomes too general and too nuanced, so it becomes the evaluation of what good or bad means, and the effort to capture all of it increases exponentially. As you remove the biggest uncertainties, what&#8217;s left brings steep diminish returns. That&#8217;s why companies don&#8217;t keep evaluating forever, they deploy employees, put them on probation, put some guardrails around, and out they go.</p><p>This is not to say that standardized exams are not useful, of course they are, and that&#8217;s why all LLMs try to compete for <a href="https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard">the leaderboard</a> to prove their value, and why companies prefer to hire employees from high-regarded institutions rather than little known ones. Before you have any insights on how well it is going to work out for your specific scenario, having higher priors is a safer bet, and then you vibe-check on top of that.</p><p>That&#8217;s the same as a lot of companies have been doing deploying their LLMs right now, but not without guilt, everybody is ashamed to admit that they are really just vibe-checking, it makes us uncomfortable, there is lack of confidence, lack of control, it&#8217;s an unpredictable employee with too high risk for brand reputation.</p><p>So I wanted to build up to bring you to this thought: in many ways LLMs are like an employee of yours, you give it the job description, you interview it with some test scenarios and then you deploy it, without knowing what it is going to say or do. It&#8217;s okay, it&#8217;s not a problem, as long as you also add some guardrails around it to keep it in check for the most important rules of your business, and monitor it on how it is doing it&#8217;s work, as a manager does, evaluating its performance and giving feedbacks to improve (iterating), not only based on cold-hard metrics though, but also on further vibe-checking for constant growth (with regression-prevention tests being possible with LLMs).</p><p>On one hand, being machines, of course AI can scale and be way more impactful than a single individual, for good or for worse, so then we justifiably hold them accountable to higher standards than we would a lot of people. On the other hand, being machines, there is a maybe a lot we can automate and do moving forward to increase the standards in a scalable way as we expect. Meanwhile, iterative vibe-checking is still a very valid way for our brains to align on weather general-purpose machines are good or not for our specific scenario.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://reasoningengine.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The Reasoning Engine! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[LLMs price reduction timeline]]></title><description><![CDATA[The fastest price reduction in technology]]></description><link>https://reasoningengine.ai/p/llms-price-reduction-timeline</link><guid isPermaLink="false">https://reasoningengine.ai/p/llms-price-reduction-timeline</guid><dc:creator><![CDATA[Rogerio Chaves]]></dc:creator><pubDate>Wed, 06 Mar 2024 10:30:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4f3ab7c9-1521-4ee5-a11a-7dba44c8af93_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>As ChatGPT launched the world into the AI and LLMs race, enormous amounts of efforts and capital went into improving the technology, not only in quality, but also in speed and price. As I wrote earlier, LLMs saw an <a href="https://reasoningengine.ai/i/140093213/sharp-decline-of-cost">86.5% reduction in cost just in 2023</a> (Mar/2023 to Nov/2023) if we take GPT 3.5-turbo as a baseline (if we would go a step before turbo, plain GPT-3.5 on Nov/2022, this actually jumps to 92.5% reduction for the first year!). For contrast, IT equipment historically reduced &#8220;just&#8221; <a href="https://en.wikipedia.org/wiki/Moore%27s_law#Density_at_minimum_cost_per_transistor">23% per year on its best years</a>, even tech that faced impressive sharp cost declines like storage, solar, gene sequencing or LED bulbs were not that fast. </p><p>And the price cuts continues steadily as new challengers try to take both OpenAI and NVIDIA&#8217;s crowns. Sure, a lot of price cuts might come from VCs and Big Tech deep pockets, but given that open-source LLMs are definitively here to stay, predatory pricing just to get monopoly might not be the best investment, and they need to be sure to recover the costs somehow, so it certainly feels like the price cuts are in a big part due to real optimizations, not just hype dynamics, and papers comming out give a good evidence support for that.<br></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://reasoningengine.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The Reasoning Engine! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3>Timeline with GPT-3.5 as a baseline<br></h3><p><strong>Nov/2022</strong> - <a href="https://en.wikipedia.org/wiki/GPT-3#GPT-3.5">OpenAI launches GPT-3.5 (text-davinci-003)</a> for <a href="https://platform.openai.com/docs/deprecations/instructgpt-models">$20 / 1M tokens</a></p><p><strong>Mar/2023</strong> - <a href="https://openai.com/blog/introducing-chatgpt-and-whisper-apis">OpenAI launches GPT-3.5-turbo, for $2 / 1M tokens</a>, a 10x reduction</p><p><strong>Jun/2023</strong> - <a href="https://openai.com/blog/function-calling-and-other-api-updates">OpenAI reduces GPT-3.5-turbo input cost by 25%</a>, at $1.5 / 1M tokens, keeping output at $2 / 1M tokens</p><p><strong>Nov/2023</strong> - <a href="https://openai.com/blog/new-models-and-developer-products-announced-at-devday">OpenAI announces on it&#8217;s DevDay a reduction in GPT-3.5 turbo input cost again</a> to $1 / 1M tokens (output still at $2 / 1M tokens), while increasing the context window to 16K</p><p><strong>Dec/2023</strong> - <a href="https://mistral.ai/news/mixtral-of-experts/">Mistral announces Mixtral 8x7B</a>, achieving GPT 3-5&#8217;s performance, and costing $0.7 / 1M tokens. Given it&#8217;s open-source nature, providers rush to host it at lower prices, with <a href="https://twitter.com/JosephJacks_/status/1735756308496667101">DeepInfra offering it at $0.27 / 1M tokens</a></p><p><strong>Jan/2024</strong> - <a href="https://openai.com/blog/new-embedding-models-and-api-updates">OpenAI announces GPT-3.5 price reduction again</a>, with $0.5 / 1M tokens input and $1.5 / 1M tokens output (lower than Mistral for input, but not the lowest anymore at this quality)</p><p><strong>Feb/2024</strong> - <a href="https://wow.groq.com/news_press/groq-opens-api-access-to-real-time-inference/">Groq opens it&#8217;s API access with it&#8217;s new LPU engine</a>, offering a <a href="https://wow.groq.com/">lower price guarantee</a> per million of tokens, with Mixtral 8x7B at the same $0.27 / 1M tokens, but at ~480 tokens/s speed</p><p><strong>Mar/2024</strong> - <a href="https://www.anthropic.com/news/claude-3-family">Anthropic announces the new Claude 3 family</a>, with it&#8217;s smaller Haiku model at GPT 3-5&#8217;s capability but for half of the price for input, at $0.25 / 1M tokens for input and $1.25 / 1M tokens for output</p><p></p><p>In summary, this represents ~92.5% reduction from Nov/2022 to Nov/2023, and ~86.5% from Mar/2023 to Mar/2024, I&#8217;ll keep this timeline up to date, let&#8217;s see how it will look like this november, if the cost reduction trend will reduce, or keep accelerating!</p><p>Now, you might be wondering, do we even care about GPT-3.5 when GPT-4 is out there? Well, yes, it is a quite capable model and good baseline to start with and look at price reductions, specially now that other models are catching up in performance.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://twitter.com/mattshumer_/status/1764738098389225759" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EycA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 424w, https://substackcdn.com/image/fetch/$s_!EycA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 848w, https://substackcdn.com/image/fetch/$s_!EycA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 1272w, https://substackcdn.com/image/fetch/$s_!EycA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EycA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png" width="338" height="437.27272727272725" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:740,&quot;width&quot;:572,&quot;resizeWidth&quot;:338,&quot;bytes&quot;:193158,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://twitter.com/mattshumer_/status/1764738098389225759&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EycA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 424w, https://substackcdn.com/image/fetch/$s_!EycA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 848w, https://substackcdn.com/image/fetch/$s_!EycA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 1272w, https://substackcdn.com/image/fetch/$s_!EycA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F999c1c27-7f22-40ac-a422-7d0567db6120_572x740.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>However, let&#8217;s also take a look at a timeline for GPT-4 as a baseline.<br></p><h3>Timeline with GPT-4 as a baseline<br></h3><p>Mar/2023 - OpenAI launched GPT-4 behind a waitlist for developers, <a href="https://en.wikipedia.org/wiki/GPT-4#Usage">costing $30 / 1M tokens for input and $60 / 1M tokens for output</a> </p><p>July/2023 - <a href="https://openai.com/blog/gpt-4-api-general-availability">GPT-4 API generally available for all paying customers</a></p><p>Nov/2023 - <a href="https://openai.com/blog/new-models-and-developer-products-announced-at-devday">OpenAI announces GPT-4-turbo at the DevDay</a>, with 128K context window, and reducing the prices to $10 / 1M tokens for input and $30 / 1M tokens for output</p><p>Mar/2024 - <a href="https://www.anthropic.com/news/claude-3-family">Anthropic seems to be the first one to really beat GPT-4 with Claude 3 Opus</a>, but at a slighlty saltier price of $15 / 1M tokens for input and $75 / 1M tokens for output. However, the runner up Claude 3 Sonnet does beat GPT-4 in some benchmarks and get quite close on others, at a lower $3 / 1M tokens for input and $15 / 1M tokens output prices</p><p>That&#8217;s a ~33% to ~90% input price reduction from Mar/2023 to Mar/2023, depending how you look at it.<br></p><p>Now, you might notice that Google&#8217;s Gemini is missing from those lists, this is because Google is being weird with their prices, telling that the <a href="https://blog.google/technology/ai/google-gemini-next-generation-model-february-2024/">prices for Gemini Ultra are &#8220;comming soon&#8221;</a>, offering Gemini Pro <a href="https://blog.google/technology/ai/gemini-api-developers-cloud/">completely for free</a> (desperate much?) and losing public confidence, like <a href="https://www.reddit.com/r/ChatGPT/comments/18c76c6/google_gemini_claim_to_outperform_gpt4_5shot/">comparing 5-show with CoT</a>, trying to pull marketing stunts with <a href="https://mashable.com/article/google-gemini-demo-video-editing">very heavily edited demo videos</a>, and so on. When they can actually show comparable performances and prices without gimmicks, then we can put them as a good comparison point as well.<br></p><p>That&#8217;s it, GPT-4 seems like a tougher game to play with Anthropic reaching it just now, but regardless of that, the LLM game doesn&#8217;t seem to be slowing down anytime soon, the remaining question to answer is if this remaining GPUs after optimization will transformed in business value and cost recovery, or if it means we can go stronger, bigger, GPT-5 and ahead.<br><br>Cheers!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://reasoningengine.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The Reasoning Engine! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Reasoning Engine]]></title><description><![CDATA[Intelligence got cheaper]]></description><link>https://reasoningengine.ai/p/the-reasoning-engine</link><guid isPermaLink="false">https://reasoningengine.ai/p/the-reasoning-engine</guid><dc:creator><![CDATA[Rogerio Chaves]]></dc:creator><pubDate>Wed, 27 Dec 2023 11:31:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2eebd25c-f930-42cb-a508-e2ca435edd9f_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the world of software development, we&#8217;re accustomed to connect many components together to build complex systems, those components have been stable for a while: you have a database, which is your storage engine for transactional data, you have a queue, to transport data reliably and at scale to different parts of the system, you have load balancers and CDNs, and other <a href="https://docs.aws.amazon.com/whitepapers/latest/aws-overview/introduction.html">200+ AWS services at your disposal</a>. However, never before we had a component for reasoning, an engine where we throw data at it, it reasons for a bit, and it gives an answer out. Now we do.</p><p>Some may argue that what LLMs do is not really reasoning, even so, they definitively pushed the boundaries of what is possible to automate. As we automated away mechanical and repetitive tasks, we moved more and more to become knowledge workers, doing non-repetitive work or at least very flexible types of work, which cannot be described by a formula to be automated, or at least it is not worth to do so. Now, for the first time, we can also automate tasks that require knowledge and intelligence, and that are somewhat non-repetitive, tasks that we always said, required <em>reasoning.</em></p><p>For the first time we can, as part of our software, have a function call to execute a task with the specification of the task in natural human language, a task that requires some level of intelligence to be executed, and have the computer to carry it over for you instead of a fellow co-worker.</p><p>Let&#8217;s take the example of a retail ecommerce, which has a catalog of over 40 thousand items from various suppliers, adding a special &#8220;playful and funny&#8221; branding on top of it. To have this branding throughout the whole store, the ecommerce needs to rewrite the description given by the suppliers of each and every item. To do that, the ecommerce pays a wholesome amount to an agency, which has an army of humans rewriting those products in the business tone of voice. This costs millions per year for the business, but there is no other way, rewriting requires a level of intelligence. Turns out, rewriting with a different tone of voice can now be automated with LLMs, the cost can be brought down from a few hundred dollars per item, to <strong>cents</strong>.<strong> </strong>Even if there is a trade-off in quality, it has to be huge to cover the difference in price, and as LLMs get smarter, the gap is definitively shrinking. You can now execute tasks that requires intelligence for a cheaper price. Intelligence got cheaper.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g2YY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g2YY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 424w, https://substackcdn.com/image/fetch/$s_!g2YY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 848w, https://substackcdn.com/image/fetch/$s_!g2YY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 1272w, https://substackcdn.com/image/fetch/$s_!g2YY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g2YY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp" width="1024" height="672" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:672,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:156460,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g2YY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 424w, https://substackcdn.com/image/fetch/$s_!g2YY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 848w, https://substackcdn.com/image/fetch/$s_!g2YY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 1272w, https://substackcdn.com/image/fetch/$s_!g2YY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa375514-d7eb-4e6b-a3e3-b2f2b8bb32db_1024x672.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Reasoning Engine as imagined by Dalle-3</figcaption></figure></div><p>Humans need not to be displaced for the reasoning engine to add value either, automation is probably the most clich&#233; example, but there is actually much more to be gained in the accretion of new ideas, tiny gaps that requires intelligence, but would not be carried unless really cheap to do so.</p><p>For example, consider an online travel agency selling tickets to attractions and things to do around the world, and it wants to help travelers to have fun even if they got unlucky with the weather, under a heavy rain for vacations. They don&#8217;t have the information of which things would be fun to do in bad weather conditions, they could ask the suppliers, but suppliers may have quite biased opinions of themselves. They could crowdsource it from their community, but it could take years until they have a decent amount of data worldwide to give out recommendations. They could ask their own travel experts and employees, or hire an agency to mark each attraction as &#8220;good for heavy rain&#8221; or not, but if they have a hundred thousand products, the task becomes incredibly boring and cost prohibitive for such a minor feature. There is no return over investment, so this feature would never happen. Guess what, LLMs have the world knowledge compared to a travel expert, they can be good estimators if an attraction is good for rainy days or not, and can carry over the task for a modest sum&#8202;&#8212;&#8202; suddenly, this feature becomes ROI positive and pops into existence.</p><p>Now where does traditional Machine Learning fit into this? This seems like a task that could be easily carried over by a classification model, and moreover, isn&#8217;t ML already this engine capable of more than just the logical processing of traditional software? Well, yes and no. Sure, it can carry over a task that requires some intelligence, but it doesn&#8217;t reason about the task itself, it just mindlessly executes this one thing. This means that for this specific example, unless there is some off-the-shelf attractions-to-rain classification model out there, then there would be also a high cost associated with hiring data scientists to develop, train and run a specific model for your business, with the timelines also measured in months, making it still cost prohibitive.</p><p>Those examples are just scratching the surface, considering the abstraction of having a Reasoning Engine anywhere can take software much further, think of having a system that dynamically integrates with any other system on the fly, by simply generating the API integrations as users need them, or an exception handler that can auto-recover by understanding the cause of issue and patching the data to get it to pass through, maybe UIs generated on the fly for user amusement, and a supervision tree agent that quarantines and review all actions executed by the thousands of mini automated agents across the system, reporting to the human in charge. The possibilities are much larger than my creativity can contain.</p><h3>Sharp Decline of&nbsp;Cost</h3><p>Two biggest blockers for having Reasoning Engines more and more ingrained in our software are quality and cost. Quality is where the biggest challenge and most popular race lies of course, however, even given lower quality models, we can play into computers advantages, which is being easily scalable and never getting tired. Take software development for example, even GPT-4 is often not better than Junior developers, however, there is no reason why can&#8217;t we employ an army of a 1000 almost-junior LLM developers to work 24/7 until they crack a problem, the main limiting factor right now is cost, and this has been in sharp decline.</p><p>This year, OpenAI announced a <a href="https://openai.com/blog/new-models-and-developer-products-announced-at-devday">3X reduction in GPT-4</a> cost even as they massively increase the context window size; there is a <a href="https://fortune.com/2023/11/19/sam-altman-was-fundraising-for-chip-venture-to-rival-nvidia-before-openai-ouster/">race starting</a> to try outperform NVIDIA GPUs with <a href="https://www.etched.ai/">specialized transformers chips</a>; open-source is catching up on performance with tiny models that run on your phone and architectural innovations such as mixture-of-experts reducing inference cost to ~25% for the same number of total parameters, as the <a href="https://mistral.ai/news/mixtral-of-experts/">8x7B MoE model by Mistral</a>. The race to the bottom is so intense that Google decided to provide <a href="https://blog.google/technology/ai/gemini-api-developers-cloud/">Gemini Pro completely for free</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://twitter.com/JosephJacks_/status/1735756308496667101" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!P_4K!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 424w, https://substackcdn.com/image/fetch/$s_!P_4K!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 848w, https://substackcdn.com/image/fetch/$s_!P_4K!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 1272w, https://substackcdn.com/image/fetch/$s_!P_4K!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!P_4K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png" width="492" height="416.2391304347826" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:934,&quot;width&quot;:1104,&quot;resizeWidth&quot;:492,&quot;bytes&quot;:188522,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://twitter.com/JosephJacks_/status/1735756308496667101&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!P_4K!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 424w, https://substackcdn.com/image/fetch/$s_!P_4K!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 848w, https://substackcdn.com/image/fetch/$s_!P_4K!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 1272w, https://substackcdn.com/image/fetch/$s_!P_4K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79c02536-4041-4c3d-8644-fb61e5b7ee3e_1104x934.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Considering that GPT-3.5, which launched the LLM race at the beginning of 2023 came with the cost of $2.00 / 1M tokens (advertised as <a href="https://www.reddit.com/r/OpenAI/comments/126mrss/am_i_doing_my_math_right_is_the_gpt_35_api_really/">$0.002 / 1K tokens</a>) and Mixtral 8x7B at the end of 2023 is <a href="https://mistral.ai/news/mixtral-of-experts/">comparable or even better</a> in performance, now costing $0.27 / 1M tokens at <a href="https://twitter.com/DeepInfra/status/1735468890413776932">DeepInfra</a>, this means a <strong>86.5% reduction in cost</strong> just this year, an impressive feat. For contrast, to give an idea, IT equipment historically reduced &#8220;just&#8221; <a href="https://en.wikipedia.org/wiki/Moore%27s_law#Density_at_minimum_cost_per_transistor">23% per year on its best years</a>.</p><h3>LLM as the&nbsp;Kernel</h3><p>Taking a step forward, some people, most notably <a href="https://twitter.com/karpathy">Andrej Karpathy</a>, came up with the conceptual idea of a LLM OS, having the LLM in very core just for it&#8217;s reasoning capabilities similar to a CPU or our brain, having other components connected to allow it to extend its limitations, such as giving it longer term memory, peripherals like audio and video, communication with the network and other LLMs and so on.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7KmJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7KmJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 424w, https://substackcdn.com/image/fetch/$s_!7KmJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 848w, https://substackcdn.com/image/fetch/$s_!7KmJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 1272w, https://substackcdn.com/image/fetch/$s_!7KmJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7KmJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png" width="1456" height="824" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:824,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7KmJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 424w, https://substackcdn.com/image/fetch/$s_!7KmJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 848w, https://substackcdn.com/image/fetch/$s_!7KmJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 1272w, https://substackcdn.com/image/fetch/$s_!7KmJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77faa323-4bd5-465b-8cc1-89ed8c5c16c9_1600x906.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://youtu.be/zjkBMFhNj_g?si=OBJ_t0lFOc0VpvI9&amp;t=2535">LLM OS by Andrej&nbsp;Karpathy</a></figcaption></figure></div><p>This has as its very core the idea that the LLM is capable of general reasoning to orchestrate all those elements, which is only true at a shallow level as of now, but as context windows and multi-modal capabilities expand, models become better and better at it, already controlling <a href="https://github.com/rogeriochaves/driver">computers</a> and <a href="https://twitter.com/_akhaliq/status/1738050817100325354">mobile phones</a> while doing all the reasoning to carry over a task.</p><p>Kernel or not, being at the core or at the edge of software, either way I cannot even begin to imagine the possibilities of what having an engine capable of reasoning in the creative hands of developers can lead to, and I&#8217;m starting this blog&#8217;s precisely to follow and write about it as it develops.</p><h3>This Blog and Newsletter</h3><p>Hello my name is Rogerio <a href="https://emojipedia.org/waving-hand">&#128075;</a></p><p>I&#8217;ve been a software engineer for 14 years now, I&#8217;ve started this blog to write on musings and developments of the Reasoning Engine through the lenses of a developer as it eats software. It doesn&#8217;t really matter if this will indeed be advanced by LLMs or if some better idea is going to replace it, but the fact is, we won&#8217;t go back from here, the journey ahead is exciting for builders, and I will be following it closely.</p><p>My goal is to understand how devs are using LLMs, how both the foundational technology and the libraries and techniques on top advances, for dev productivity but specially when embedded in the software itself, and what possibilities will be unlocked that we could never guess before.</p><p><a href="https://reasoningengine.substack.com/welcome">Subscribe to the newsletter</a> if you want to enjoy the ride too, and <a href="https://twitter.com/_rchaves_">follow me on twitter</a> if you prefer AI memes instead.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://reasoningengine.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading The Reasoning Engine! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Coming soon]]></title><description><![CDATA[This is The Reasoning Engine.]]></description><link>https://reasoningengine.ai/p/coming-soon</link><guid isPermaLink="false">https://reasoningengine.ai/p/coming-soon</guid><dc:creator><![CDATA[Rogerio Chaves]]></dc:creator><pubDate>Tue, 26 Dec 2023 15:00:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!F4OX!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F30fd94e4-bea0-409d-821d-4fcd2a0074c0_400x400.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is The Reasoning Engine.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://reasoningengine.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://reasoningengine.ai/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item></channel></rss>