<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Supervised]]></title><description><![CDATA[Covering innovation and emerging technology in big data and AI.]]></description><link>https://www.supervised.news</link><image><url>https://substackcdn.com/image/fetch/$s_!kpIY!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F76c85142-baa2-4de5-bdd2-d3d2932bbdb2_1200x1200.png</url><title>Supervised</title><link>https://www.supervised.news</link></image><generator>Substack</generator><lastBuildDate>Tue, 28 Apr 2026 20:14:36 GMT</lastBuildDate><atom:link href="https://www.supervised.news/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Matthew Lynley]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[supervised@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[supervised@substack.com]]></itunes:email><itunes:name><![CDATA[Matthew Lynley]]></itunes:name></itunes:owner><itunes:author><![CDATA[Matthew Lynley]]></itunes:author><googleplay:owner><![CDATA[supervised@substack.com]]></googleplay:owner><googleplay:email><![CDATA[supervised@substack.com]]></googleplay:email><googleplay:author><![CDATA[Matthew Lynley]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[What's next for Supervised]]></title><description><![CDATA[I've made the extremely difficult decision to end my full-time job with the newsletter, which will shift to more of a hobby. Plus, is the thinning out in the modern data stack finally here?]]></description><link>https://www.supervised.news/p/whats-next-for-supervised</link><guid isPermaLink="false">https://www.supervised.news/p/whats-next-for-supervised</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 16 May 2025 16:20:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Y9CP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Y9CP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Y9CP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!Y9CP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!Y9CP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!Y9CP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Y9CP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1173563,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.supervised.news/i/156027303?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Y9CP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!Y9CP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!Y9CP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!Y9CP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41e9bf34-aa19-44e2-94b9-22d6c108b965_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a sleek, friendly pixar-style robot carrying a box of office supplies out of a dim room &#8212; midjourney</figcaption></figure></div><p>If you haven&#8217;t heard already, a few months ago I made the very, very difficult decision to wind down my full-time job writing this newsletter due to a back injury. Now that I&#8217;m feeling better post-surgery and getting eager to participate in all the exciting stuff going on in this space, I&#8217;m going to start writing again &#8212; for free &#8212; until I find a full-time gig.<br><br>If you want my thoughts on the modern data stack, skip ahead. If you want the full story of my newsletter decision and my takeaways from the experience, here it is: </p><p>A lower back injury causing extreme pain left me barely able to get write, much less actively try to report&#8212;which involves a lot of meetings, events, and being able to sit for long periods of time and focus on conversations. I spent six months, unsuccessfully, trying to rehab my injury and treat it through physical therapy, acupuncture and lifestyle changes. That unfortunately didn&#8217;t work, and I&#8217;m now writing this on the other end of back surgery.</p><p>Before launching the newsletter, I set several ambitious growth targets&#8212;ones that would make it viable to run a newsletter as a full-time gig in the Bay Area. As a result of the injury, I wasn&#8217;t able to hit them.</p><p>Momentum is <em>everything</em> when you are running your own thing. This was growing quickly for a year, and then came to a crashing halt once I stopped publishing and turned off billing as I tried to rehab my injury.</p><p>Launching my own project has always been a dream of mine, and I had hoped if it were to come to an end, it would be because of a series of my own mistakes and errors. Unfortunately, I massively underestimated the value your physical health plays into the actual health&#8212;and survival&#8212;of your product.</p><p>For now, I finally have some reprieve from the pain (and some creative energy) to write for you while I recover. But I&#8217;m now going to have to do it as a hobby while I figure out what to do next. (If anyone has ideas, I&#8217;m all ears!)</p><p>Words cannot express my gratitude if you&#8217;ve been a paid supporter of Supervised&#8212;especially those who were there from day one. You all made it possible to chase that dream, even if it ended in devastating fashion. I hope I was able to deliver on the value and promises I made when you came along for the ride.</p><p>To everyone who was even a casual reader&#8212;I know your inbox is crowded! Thank you for taking even a little look at what I was working on over here.</p><p>Finally, I want to say this over, and over, and over again for others that are running their own thing (or thinking of starting one): <strong>take care of your health</strong>. I know I did my best work when I was fully healthy.</p><p>Some final logistics to address:</p><ul><li><p>Billing will remain off for the foreseeable future. This also means only existing paid subscribers will be able to see what is below a paywall in older posts.</p></li><li><p>I&#8217;ll continue to write as opportunities arise, though on a much more limited publishing schedule.</p></li><li><p>If you&#8217;re a paid annual subscriber and you are interested in a pro-rated refund, please reach out to me directly at m@supervised.news. I&#8217;ll do my best to process them in the coming months.</p></li></ul><p>Again, my deepest thanks to everyone: my readers, my sources, and all the companies and startups that took the time to talk to me&#8212;especially right as I was getting off the ground. As always, you can reach out to me at any point for now, well, anything. My line will remain open, and I want to know about your life too! </p><div><hr></div><h2>Early signs of a consolidation of the modern data stack</h2><p><em>(Full disclosure: I spoke with a handful of the companies in the broader modern data stack and MLops stack about potential roles. That has no impact whatsoever on my analysis here, which is based on historical reporting and prior newsletters.)</em></p><p>The looming consolidation of the modern data stack has been a constant theme (and somewhat of a joke) for the better half of a decade. Venture money flooded the ecosystem going back to the early 2020s, sending companies that were rewriting parts of the analytics and data science stack to multi-billion dollar valuations. In the same way that AI is <em>incredibly</em> crowded, the modern data stack has been <em>incredibly</em> crowded for a significantly longer period of time. </p><p>But in the last few months, the largest players in the industry have indeed finally started picking off smaller companies that made sense within a company, and not necessarily a full-fledged public company. We&#8217;re starting to see signs of it in <em>both</em> the modern data stack and the machine learning ops stack, which you could even argue are kind of interchangeable by now.</p><p>While not exactly part of that consolidation, my general observation here is that this kicked off in earnest with <a href="https://www.getdbt.com/blog/dbt-labs-acquires-sdf-labs">Dbt Labs acquiring SDF</a> in January this year, which in some circles was considered a potential challenger to parts of Dbt as a kind of pre-compiled SQL tool. And it also wasn&#8217;t solely interesting for being in the category of &#8220;what if X, but Rust?&#8221; (Tobiko is also in this conversation, <a href="https://www.tobikodata.com/blog/the-future-of-tobiko---from-sinks-to-sql">which in June last year raised $21.8 million</a> from Theory Ventures and Unusual Ventures&#8212;as well as MotherDuck CEO Jordan Tigani, Fivetran CEO George Fraser, and Census CEO Boris Jabes.)</p><p>Since then <a href="https://wandb.ai/site/">Weights &amp; Biases</a>, basically the main option in model experiment and evaluation&#8212;<a href="https://wandb.ai/site/llm-monitoring-sign-up/">and in some ways now observability</a>&#8212;<a href="https://www.coreweave.com/blog/coreweave-and-weights-biases-to-join-forces">was acquired by CoreWeave</a>. The Information reported that <a href="https://www.theinformation.com/articles/coreweave-in-talks-to-buy-ai-startup-weights-biases-for-around-1-7-billion?rc=u5devd">earlier talks between the companies landed around $1.7 billion</a>. (I <a href="https://www.supervised.news/p/the-second-act-of-the-pre-gpt-ai">originally reported that Weights &amp; Biases held talks for a $2 billion funding round</a> around the beginning of March 2023, and <a href="https://www.supervised.news/p/vibes-based-evals-and-weights-and">that Weights &amp; Biases had passed $30 million in recurring revenue in April last year</a>.)</p><p>Databricks also <a href="https://www.databricks.com/company/newsroom/press-releases/databricks-agrees-acquire-neon-help-developers-deliver-ai-systems">acquired Neon, a serverless Postgres,</a> <a href="https://www.wsj.com/articles/databricks-to-buy-startup-neon-for-1-billion-fdded971">for around $1 billion</a>. Neon&#8217;s (correct) bet was that Postgres was never going to go away, and there was an opportunity to potentially extend it to AI applications with embedding and retrieval tools like pgvector. It also bet early that, though there would be a need for some vector format, <a href="https://www.supervised.news/p/lets-check-back-in-on-the-vector">the massive scale of independent vector databases</a> like <a href="https://www.lancedb.com/">LanceDB</a> wasn&#8217;t going to be necessary for most companies. Indeed, Databricks largely <a href="https://www.databricks.com/blog/databricks-neon">positioned the announcement as important for agentic workflows</a>.</p><p>The list goes on&#8212;both in the form of acquisitions, but also in plenty of conversations about acquisitions. </p><ul><li><p>Fivetran, the largest provider for ETL tools, <a href="https://www.fivetran.com/press/fivetran-signs-agreement-to-acquire-census-delivering-the-first-end-to-end-data-movement-platform-for-the-ai-era">acquired Sequoia-backed reverse ETL startup Census</a>. The reverse ETL category contained not one, but <em>three</em> companies that touched it in some form or another: Census, <a href="https://hightouch.com/">Hightouch</a>, and <a href="https://www.rudderstack.com/">Rudderstack</a>.</p></li><li><p>OpenAI <a href="https://www.bloomberg.com/news/articles/2025-05-06/openai-reaches-agreement-to-buy-startup-windsurf-for-3-billion?embedded-checkout=true">acquired Cursor competitor Windsurf for $3 billion</a>, bringing a proper IDE into the fold rather than plugging a key into Cursor (or having extensive back-and-forths with Claude Sonnet). You could call this an app on top of AI, but really, this is potentially so deeply integrated into workflows it&#8217;s simply part of the stack.</p></li><li><p>Observability colossus Datadog <a href="https://www.geteppo.com/blog/eppo-is-now-part-of-datadog">acquired feature flagging and experimentation startup Eppo</a>, which Alex Konrad over at his new publication Upstarts Media <a href="https://www.upstartsmedia.com/p/datadog-buys-startup-220-million">reported was for $220 million</a>. This is after Datadog first whiffed on its <a href="https://www.supervised.news/p/a-rare-opening-against-datadog">AI model evaluation last year while a number of startups were gaining momentum</a>, and maybe signals an openness to buy into areas rather than build.</p></li><li><p>MongoDB <a href="https://www.mongodb.com/blog/post/redefining-database-ai-why-mongodb-acquired-voyage-ai">acquired Voyage AI</a>, a specialist in embedding products, <a href="https://www.bloomberg.com/news/articles/2025-02-24/mongodb-buys-voyage-ai-for-220-million-to-bolster-ai-search">for $220 million</a> in February. This <em>also</em> came after a lot of hand-wringing last year over whether MongoDB would launch its own embedding model after playing the field for so long.</p></li><li><p>The Information reported that <a href="https://www.theinformation.com/articles/snowflake-in-talks-to-acquire-analytics-startup-redpanda?rc=u5devd">Snowflake held talks to acquire Redpanda</a>, a streaming data service. This is <em>also</em> part of a potential shift (and ongoing joke) about all data coming in through streaming data pipelines, rather than batch data processing. (I&#8217;ll let you all argue amongst yourselves the difference between streaming and micro-batch.)</p></li></ul><p>The very crowded space that was, by some (most?) accounts, drastically over-funded has begun an early thinning out process through acquisitions. And this in some ways is in service to two impending changes that are intertwined: data lakes en route to being an increasingly attractive option for inexpensive storage; and the fight over owning the data control plane for a company&#8217;s data. (For the former, look no further than the battle for Tabular, which Databricks snatched from Snowflake by more than <a href="https://www.supervised.news/p/lakes-catalogs-and-snowflakes-full">doubling an early offer around $600 million</a> that I first reported.)</p><p>Late last year (before I had to begin my rehab attempt), a number of categories entered the conversation for ownership of a more complete data control plane. Dbt largely owned the data transformation layer. Atlan, Alation, and Collibra were dueling for ownership of the data catalog covering lineage and governance (<a href="https://www.supervised.news/p/lakes-catalogs-and-snowflakes-full">which Snowflake released its own open source product</a>). Fivetran owned classic ETL and (still) stands to grow into AI workflows&#8212;especially as lakes move toward a preferred option. The customer data pipeline was (and seemingly still is) up for grabs, with Hightouch&#8212;<a href="https://hightouch.com/blog/hightouch-funding-series-c">recently valued at $1.2 billion</a> in a round that included Bain Capital Ventures, Y Combinator, and Amplify partners&#8212;and Rudderstack still going after it.</p><p>(One outcome I&#8217;m keeping my eye on right now is <a href="https://www.supervised.news/p/a-new-whisper-in-the-ai-analytics">what happens with Informatica</a> and whether it is able to modernize in a way that it can truly make a run for ownership of the data control plane. As one source once joked to me, &#8220;I don&#8217;t want to have to deal with five logins.&#8221;)</p><p>The rise of Snowflake and Databricks gave birth to all of these micro-categories that are seemingly consolidating out to a smaller number of broader categories that make sense with expanded ownership of a general data pipeline. And while it doesn&#8217;t seem like there will be a <em>collapse</em>, there are a number of clear leaders emerging&#8212;all with slight overlap with one another, but largely playing nicely.</p><p>(To just also toot my own horn here, the majority of companies here were on my 2023 startups <a href="https://www.supervised.news/p/29-startups-in-ai-im-watching-closely">watch lists for AI</a> and <a href="https://www.supervised.news/p/27-big-data-startups-im-watching">Big Data</a>! I&#8217;ll revisit that in the next issue and provide a few updates.)</p><p>The difference now is that all of these companies have since adjusted their branding-slash-products to fit the mold of modern AI. And this whole process of the modern data stack extending itself into AI has been a <a href="https://www.supervised.news/p/december-in-ai-whats-old-is-new-and">constant theme since the launch of ChatGPT</a>. Enterprises have stricter requirements like robust data lineage and governance, which also feed directly into systems like those using RAG. And good AI is predicated on good data, after all, whether that&#8217;s flowing into. model customization or data retrieval.</p><p>As a final note, one thread I am following <em>very</em> closely is the second coming of query engines like <a href="https://www.starburst.io/">Starburst</a>, another startup that was in my early watch list. Starburst&#8212;which includes Andreessen-Horowitz, Altimeter Capital, Coatue, and Index as investors&#8212;raised an enormous amount of funding <a href="https://www.starburst.io/blog/starburst-announces-250m-series-d/">that valued it at $3.35 billion in the mania</a> around the modern data stack. But the shift in focus towards lakes with query engines on top of it has the potential to <em>enormously</em> benefit Starburst.</p><p>I suspect we are nowhere close to the end of this thinning out, which has been a very <em>long</em> time coming. The bigger question I have is whether this is going to be a thinning out leads to significant outcomes for firms and companies, or whether we&#8217;re talking about dollar-in-dollar-out acquisitions.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.supervised.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Supervised! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>If you did indeed make it all the way down here, I want to say it one more time: thank you, from the bottom of my heart, for being a supporter and reader of my work as I went on this insane adventure. I&#8217;m forever in your debt.</em></p><p></p><p><em>As always, if you have any complaints, ideas, or thoughts you&#8217;d like to share, feel free to reach out to me. You can find me at m@supervised.news or on Signal at (415) 690-7086.</em></p>]]></content:encoded></item><item><title><![CDATA[Please send OpenAI your good vibes as it dives into model evaluation]]></title><description><![CDATA[OpenAI has decided it's time to try to handle one of AI's existential crises. Maybe this time it'll work?]]></description><link>https://www.supervised.news/p/please-send-openai-your-good-vibes</link><guid isPermaLink="false">https://www.supervised.news/p/please-send-openai-your-good-vibes</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Tue, 01 Oct 2024 20:26:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!5-Vs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5-Vs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5-Vs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 424w, https://substackcdn.com/image/fetch/$s_!5-Vs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 848w, https://substackcdn.com/image/fetch/$s_!5-Vs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 1272w, https://substackcdn.com/image/fetch/$s_!5-Vs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5-Vs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png" width="1456" height="1306" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1306,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3551465,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5-Vs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 424w, https://substackcdn.com/image/fetch/$s_!5-Vs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 848w, https://substackcdn.com/image/fetch/$s_!5-Vs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 1272w, https://substackcdn.com/image/fetch/$s_!5-Vs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b0f680b-d4f3-4ce8-bca4-e10ffcb62a80_1868x1676.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><strong>a sleek friendly pixar-styled white robot trying to shut the door to a closet that is filled with hundreds of millions of documents, the documents are starting to fall out of the room as the robot shoves the door closed, sunday funny comics aesthetic &#8212;midjourney</strong></figcaption></figure></div><p><em>Around the time of this post I suffered a severe lower back injury that has pretty much prevented me from working. Paid subscribers received a 1 month extension and billing was fully paused beginning on October 15. In keeping with my policy, subscribers are not paying for what they are not getting. I&#8217;ll provide a further update in a few weeks once I&#8217;ve had an opportunity to treat the injury. My deepest, deepest thanks for all of your patience.</em></p><div><hr></div><p>While OpenAI&#8217;s advanced o1 &#8220;stop and think for a second&#8221; models may be a technical feat, it&#8217;s also increasingly working to enable developers onto its significantly cheaper workhorse models.</p><p>As part of this, OpenAI has to wade into one of modern AI&#8217;s comically difficult problems: <a href="https://www.supervised.news/p/a-rare-opening-against-datadog">evaluating whether the response of a model is &#8220;good&#8221; in some quantitative and automated way</a>. The industry has <a href="https://www.supervised.news/p/vibes-based-evals-and-weights-and">long tried to graduate out of vibes-based evaluation</a>, but as of yet there&#8217;s no immediately reasonable solution. And yet, here we are, with OpenAI trying to do that as <a href="https://openai.com/index/api-model-distillation/">part of its new distillation tool</a>.</p><p>Its distillation tool is one of a series of new developer tools OpenAI unveiled at its DevDay. Its first, which has a much bigger <em>wow</em> factor, is the availability of their real-time advanced voice assistant as <a href="https://openai.com/index/introducing-the-realtime-api/">part of a broader &#8220;real-time&#8221; API</a>. But there&#8217;s also a additional features that focus on mitigating usage costs and increasing performance without having to drop the cost of their APIs.</p><p>Its distillation tool effectively uses larger models to fine-tune a smaller one (like GPT-4o mini). But OpenAI is trying to wrap it into one kind of &#8220;productized&#8221; package that streamlines the process. Developers can used a stored completions feature for exactly what it sounds like, and manage the evaluation and distillation directly through a user interface rather than integrating some complex suite.</p><p>OpenAI is also <a href="https://openai.com/index/api-prompt-caching/">releasing a prompt caching tool</a> that cuts costs for ongoing repeated content within a context window. Added together with its distillation suite, the updates are a roundabout way of finding a way to shave off what developers have to pay to deliver a high performance AI-powered product.</p><p>OpenAI has long run a playbook of being extremely convenient both from a usage and a procurement standpoint, even if you incur a cost penalty by ignoring the alternatives. Enterprises, though, <a href="https://www.supervised.news/p/putting-the-brakes-on-the-ai-hype">have become much more concerned with costs</a> (OpenAI&#8217;s in particular), though. As a result OpenAI&#8217;s convenience advantage has started to become a bit more brittle in the face of a large list of alternative and highly competitive products.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!E-XM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!E-XM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 424w, https://substackcdn.com/image/fetch/$s_!E-XM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 848w, https://substackcdn.com/image/fetch/$s_!E-XM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 1272w, https://substackcdn.com/image/fetch/$s_!E-XM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!E-XM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png" width="1456" height="1012" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1012,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:173969,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!E-XM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 424w, https://substackcdn.com/image/fetch/$s_!E-XM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 848w, https://substackcdn.com/image/fetch/$s_!E-XM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 1272w, https://substackcdn.com/image/fetch/$s_!E-XM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82835c0b-a99c-4df9-ae42-4e819969e871_1468x1020.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So the next step is clearly to find some way to add different layers of efficiency to ease the cost burden on developers and enterprises. It also had to find a way to drop the difficulty of some of the more advanced techniques to get costs under control. That also meant building a product that wades into everyone&#8217;s favorite <s>existential crisis</s> development challenge in the form of model evaluations.</p><p>The realtime API is <em>cool as hell</em> and is going to be useful for a lot of potential product cases. But, setting aside that for now, let&#8217;s get to the more in the weeds bits: how OpenAI is going to shave off costs at a time when it&#8217;s under pressure from all different directions with competing (and cheaper) products.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.supervised.news/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.supervised.news/subscribe?"><span>Subscribe now</span></a></p><h2>Getting into the mess that is model evals</h2><p>There are two direct routes that OpenAI is now deploying to find an optimal point between performance and value for both developers <em>and</em> for OpenAI. One is to collect as many possible efficiencies as you can when a model is actually in use. The other to funnel as many people as you can to a workhorse model&#8212;like Gemini Flash or GPT-4o-mini.</p><p>Its prompt caching tool, where it&#8217;ll offer a 50% discount and faster prompt processing times on input tokens that it&#8217;s already seen recently, is playing a little bit of catch-up. This is becoming a pretty popular feature for multi-turn products that have a lot of information stuffed into a context window, such as large documents, code bases, long customer service interactions, or tools that require more advanced retrieval.</p><p>This feature was already available in Claude and Gemini, though they came in different flavors. Anthropic <a href="https://www.anthropic.com/news/prompt-caching">charges slightly more to write to a cache, but input tokens from that cache are 10% the cost of normal input tokens</a>. Anthropic&#8217;s prompt cache expires after 5 minutes, and that clock restarts every time the cache is used. Gemini&#8217;s context caching meanwhile also offers a substantial discount to normal input prices and <a href="https://ai.google.dev/gemini-api/docs/caching?lang=python">charges an hourly rate for every million tokens cached</a> (with no upper bound on the cache expiring).</p><p>OpenAI&#8212;in its bid to try to simplify the experience as much as possible&#8212;basically sweeps all the complexity under the rug and applies the discount and performance bump right away while the cache window is active. As of today, it&#8217;s automatically applied to its latest models, with developers not having to really do much.</p><p>OpenAI&#8217;s distillation &#8220;suite&#8221; includes enabling developers to store completions in order to generate a large set to fine-tune its models. It also includes that evaluation tool, in which users can use a bunch of off-the-shelf evaluations (like groundedness) or design a custom evaluation. In that way, enterprises can pull in some of their custom evaluations and link them together directly within a single OpenAI interface right on top of completions.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9uyB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9uyB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 424w, https://substackcdn.com/image/fetch/$s_!9uyB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 848w, https://substackcdn.com/image/fetch/$s_!9uyB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 1272w, https://substackcdn.com/image/fetch/$s_!9uyB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9uyB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png" width="1456" height="812" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:812,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2473437,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9uyB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 424w, https://substackcdn.com/image/fetch/$s_!9uyB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 848w, https://substackcdn.com/image/fetch/$s_!9uyB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 1272w, https://substackcdn.com/image/fetch/$s_!9uyB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b440d34-1e1f-49df-bc2c-ea73a8877849_2162x1206.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>
      <p>
          <a href="https://www.supervised.news/p/please-send-openai-your-good-vibes">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[A rare opening against Datadog]]></title><description><![CDATA[Plus: OpenAI's portfolio is getting a little big.]]></description><link>https://www.supervised.news/p/a-rare-opening-against-datadog</link><guid isPermaLink="false">https://www.supervised.news/p/a-rare-opening-against-datadog</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 27 Sep 2024 18:47:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PG9N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PG9N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PG9N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 424w, https://substackcdn.com/image/fetch/$s_!PG9N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 848w, https://substackcdn.com/image/fetch/$s_!PG9N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 1272w, https://substackcdn.com/image/fetch/$s_!PG9N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PG9N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png" width="1456" height="1101" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1101,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5576422,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PG9N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 424w, https://substackcdn.com/image/fetch/$s_!PG9N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 848w, https://substackcdn.com/image/fetch/$s_!PG9N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 1272w, https://substackcdn.com/image/fetch/$s_!PG9N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5897d74a-2d9a-4bf9-9e83-1f7171b36d96_2412x1824.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a purple puppy in a medical lab coat sitting at a desk writing on a piece of paper, the paper is covered with math problems, and there are similar papers taped on the wall in the office, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>This issue was finished on a mobile device so I apologize ahead of time for any typos and formatting errors.</em></p><div><hr></div><h2>OpenAI&#8217;s sprawling portfolio problem</h2><p>While OpenAI&#8217;s latest model, o1, is clearly a massive improvement in performance it&#8217;s also creating a potentially new challenge: product sprawl.</p><p>Its latest products, o1 and o1-mini, essentially gives users and customers a tradeoff: you can wait longer and, for now, spend more, but the output should be much better. It&#8217;s not the kind of API you&#8217;d plug into a call center, but it does fill kind of a new niche for a company that already covers a whole range of niches beyond just a &#8220;one thing that does almost everything well.&#8221;</p><p>OpenAI&#8217;s o1 use case is basically the &#8220;you have to think about it for a second&#8221; problem. Box CEO <a href="https://x.com/levie/status/1835909498000810239">Aaron Levie gives an extremely good enterprise example</a> of having to find a very specific parameter of a contract&#8212;in this case, the date of the <em>final signature</em> on a contract, and thus the date it effectively went live. This is a really crisp &#8220;you had to think about it for a second&#8221; problem: there was an added layer of complexity to parse to an otherwise straightforward, in this case the possibility that people signed the contract on different days.</p><p>That&#8217;s also a kind of <em>newish</em> use case for OpenAI, where historically it might have been a long set of API calls and begging the models to do what they need to do in the form of prompt tuning or retrieval augmented generation (or RAG). Instead, that whole process could&#8212;hypothetically&#8212;get compressed into a single call or two and simplify some architectures. </p><p>But at this point OpenAI has been <a href="https://www.supervised.news/p/openais-silicon-valley-pivot">undergoing a kind of &#8220;SaaS-ification&#8221;</a> where it matures into a real business. A number of executives and co-founders have continued to make their way out the door, like co-founder and chief scientist Ilya Sutskevar leaving to found <a href="https://www.reuters.com/technology/artificial-intelligence/openai-co-founder-sutskevers-new-safety-focused-ai-startup-ssi-raises-1-billion-2024-09-04/">a new safety-focused AI startup that raised $1 billion</a> and <a href="https://x.com/johnschulman2/status/1820610863499509855?lang=en">co-founder John Schulman leaving to join Anthropic</a>. As of this week, that list also <a href="https://www.supervised.news/p/metas-llamas-go-local">includes CTO Mira Murati</a>. </p><p>As OpenAI matures into a real business, it&#8217;s running head-first into the challenge of avoiding product creep at a time when its product portfolio (or in this case, model portfolio) is starting to get more and more complex. Larger product portfolios are inherently more difficult to manage, and the biggest challenge companies at this phase face is keeping it from getting unwieldy and difficult to communicate value to customers. And with the quality of models constantly improving, any small number of distractions could lead to a usurping by one of the other frontier modeling companies.</p><p>So, just quickly, let&#8217;s recap OpenAI&#8217;s now very large portfolio of APIs:</p><ul><li><p><strong>GPT-4o</strong>: a multi-modal, expensive (though not immune to price cuts) model that&#8217;s supposed to be Good and General Purpose.</p></li><li><p><strong>GPT-4o mini</strong>: a less powerful version of GPT-4o that&#8217;s designed to be a successor to its workhorse GPT-3.5 Turbo model to satisfy a large number of simpler use cases.</p></li><li><p><strong>Fine-tuned versions</strong> of the above to provide enterprise-specific needs, though they are fed through an API which might turn off some more security-conscious companies.</p></li><li><p><strong>Batch versions</strong> of the above with a 24 hour completion window, for a 50% price discount.</p></li><li><p><strong>o1</strong>: a model that trades speed and price for quality by allowing it additional time to &#8220;reason&#8221; about an answer. Basically, that &#8220;stop and think about it&#8221; question.</p></li><li><p><strong>o1-mini</strong>: like 4o mini, a smaller version of o1 designed for&#8230; the same &#8220;stop and think about it&#8221; question. But we&#8217;ll just run under the assumption there&#8217;s a set of problems this is really good for.</p></li><li><p><strong>Whisper</strong>: arguably the best speech-to-text model on the market that was most assuredly built for generating training data that OpenAI needs.</p></li><li><p><strong>Text embeddings</strong>: the not-quite-the-best embeddings product whose advantage generally seems to be that it&#8217;s provisioned with other OpenAI products that reduces procurement&#8217;s headaches.</p></li><li><p><strong>Text-to-speech</strong>: an API that you could, hypothetically, slot into something like a call center assuming the latency works. It has both a normal and HD version of the API.</p></li><li><p><strong>Advanced Voice assistant</strong>: a technical marvel of a product that lets you have an active conversation within ChatGPT where the obvious killer &#8220;use case&#8221; still isn&#8217;t super clear.</p></li><li><p><strong>ChatGPT</strong>: OpenAI&#8217;s &#8220;productized&#8221; version of all of the above in one frontend enterprise-friendly wrapper.</p></li></ul><p>Amid all of this, Meta <a href="https://www.supervised.news/p/metas-llamas-go-local">has started to show additional chaos by releasing updated models in almost every single one of those categories</a>&#8212;except they&#8217;re open-ish source and power a very different array of products even though they exist in the same &#8220;bucket.&#8221; With o1, OpenAI is essentially taking another shot at category creation, but also risking a ballooning portfolio.</p><p>Here&#8217;s how all the pricing ends up playing out, and including a handful of competitors as a point of reference:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vYB7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vYB7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 424w, https://substackcdn.com/image/fetch/$s_!vYB7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 848w, https://substackcdn.com/image/fetch/$s_!vYB7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 1272w, https://substackcdn.com/image/fetch/$s_!vYB7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vYB7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png" width="1456" height="1018" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1018,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:224009,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vYB7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 424w, https://substackcdn.com/image/fetch/$s_!vYB7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 848w, https://substackcdn.com/image/fetch/$s_!vYB7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 1272w, https://substackcdn.com/image/fetch/$s_!vYB7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70005815-d87c-41ba-8b08-30992c5aa8c0_1814x1268.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And these costs are probably nowhere close to where things will land after o1 has been out for a while. <a href="https://x.com/OpenAI/status/1835857163765637607">OpenAI has already started incrementally increasing rate limits</a>, with o1-preview recently going up from 30 per week to 50 and o1-mini going from 50 per week to 50 per day.</p><p>While OpenAI&#8217;s appeal has always been in some sweet spot between convenience, price, and performance, its ballooning portfolio certainly poses as much of a challenge as an opportunity. I seriously doubt that the prices will remain this high, as its next GPT model will be ready <em>at some point</em>. But for now it at least gets something into the hands of developers that has an aggressive price tag at a time when it&#8217;s trying to raise a colossal round of funding.</p><p>The challenge here, though, is the same that any maturing company starts to face over time: product creep. While OpenAI technically has &#8220;two&#8221; products in the form of its APIs and ChatGPT, those products all have a ton of branches that serve a very wide number of use cases. The APIs also go well beyond just chat completion and text generation, and include a whole variety of modalities. And its voice product is probably the most awkward part of its portfolio, seemingly a &#8220;wow&#8221; part of ChatGPT.</p><p>And product creep is a Known Problem in startup-land, if you can even call OpenAI a startup any more. As a startup matures, and it&#8217;s user and customer base grows, it has to directionally develop&#8212;anticipating the handful of use cases that satisfy the most customers without building everything for everyone. Or, more succinctly, do a few things but do them well.</p><p>Offering this wide array of use cases gives OpenAI the ability to funnel users to some kind of steady state where it isn&#8217;t necessarily making money, but at least it isn&#8217;t <em>losing</em> money. On the API front that&#8217;s been traditionally pushing users to its workhorse models (particularly fine-tuned versions), but that&#8217;s a little less clear with ChatGPT.</p><h2>ChatGPT and finding efficiencies in inference</h2><p>While the new models are rate-limited in ChatGPT, that is also an extremely important part of its business beyond the APIs. OpenAI COO Brad Lightcap told staff that OpenAI has more than 10 million paying subscribers for ChatGPT and an additional 1 million subscribers for businesses, <a href="https://www.theinformation.com/articles/openai-coo-says-chatgpt-passed-11-million-paying-subscribers?rc=u5devd">according to The Information</a>. (Bloomberg <a href="https://www.bloomberg.com/news/articles/2024-09-05/openai-hits-1-million-paid-users-for-business-version-of-chatgpt?embedded-checkout=true">earlier reported that OpenAI has 1 million paid business users for ChatGPT</a>.)</p><p>We can of course do a lot of napkin math, but would arrive at the same conclusion: ChatGPT&#8217;s enterprise business is, and will be, a massive part of OpenAI&#8217;s business beyond the APIs.</p><p>But it&#8217;s not like the revenue from ChatGPT will just scale up with usage of o1 directly like it would with the API. The cost consideration for actually <em>running</em> it, either via API or within ChatGPT, is probably going to change as pre-training compute resources shift toward inference.</p><p>Fortunately for OpenAI, this is an area where there&#8217;s already a lot happening. One approach that&#8217;s getting a <em>ton</em> of extra attention right now among those in the industry I talk to is Monte Carlo Tree Search as a way to narrow the amount of compute to generate a high-quality result. And distillation, another way of &#8220;shrinking&#8221; the larger models, is also gaining a lot of momentum as interest shifts to managing inference costs of high-performance models.</p><p>&#8220;That&#8217;s the sweet spot, blending strategies from traditional predictive machine learning with years of work that went in there with modern techniques,&#8221; Sri Ambati, co-founder and CEO of model developer and platform <a href="https://h2o.ai/">h2o.ai</a>, told me. &#8220;Tree search is an absolute a genius trick, it&#8217;s very easy low-hanging fruit that&#8217;s combined with the brilliance of LLMs.&#8221;</p><p>The other bit that comes up most often in conversations with experts and sources is <a href="https://noambrown.github.io/">Noam Brown&#8217;s</a> work at OpenAI. Brown is widely considered a premier expert in game theory and many have wondered how his work would be applied to OpenAI&#8217;s products. Brown was also a <a href="https://arxiv.org/pdf/2112.07544">co-author of a paper</a> that in part examined applications of Monte Carlo Tree Search in developing human-like agents.</p><p>The challenge OpenAI has for its APIs is that, based on most enterprises and platforms I talk to lately, costs are either number 1 or 2 on the list of considerations for building an AI app for production use cases. But OpenAI has certainly shown a willingness to drop its prices over time to remain competitive.</p><p>And it&#8217;s also telling that OpenAI&#8217;s ChatGPT enterprise business seems to be the significant driver for its business, making all this development essentially in service of that enterprise kit. The alternative&#8212;building something bespoke internally by using really cheap stuff off the shelf&#8212;is increasingly compelling for companies with way more advanced governance and cost requirements.</p><p>The cynical take here is that OpenAI is trying to say something along the likes of &#8220;hey, see, we&#8217;re still building extremely advanced stuff, don&#8217;t ignore our fundraising calls.&#8221;</p><p>The flip side of that argument, though, is OpenAI&#8212;with what appears to be a <a href="https://www.cnbc.com/2024/09/26/openais-cfo-says-funding-round-should-close-by-next-week-in-letter.html">massive funding round that&#8217;s oversubscribed, according to CNBC</a>&#8212;has already matured into a company better suited to enterprise products. And now OpenAI just has to make sure it isn&#8217;t confusing enterprises on sales calls with a colossal portfolio of models.</p><div><hr></div><h2>An opening in Datadog&#8217;s armor</h2><p>Datadog has enjoyed an incredibly enviable position of power and a relatively pristine track record of launches that cause more than a few companies adjacent to it to be nervous every time something comes out.</p><p>Startups, investors, and developers in the rapidly-burgeoning AI space had long expected Datadog to storm into building tools for evaluating the performance and success of LLMs this year. It was already a considerably difficult space that has been trying to move beyond a &#8220;vibes&#8221; based approach, but if there were a large company best positioned to do it, it would have likely been Datadog.</p><p>But Datadog&#8217;s annual <a href="https://www.datadoghq.com/blog/dash-2024-new-feature-roundup-keynote/">conference came and went in June</a>, and what came instead for those startups (and the investors in them) was a bit of a sigh of relief.</p>
      <p>
          <a href="https://www.supervised.news/p/a-rare-opening-against-datadog">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Meta's Llamas go local]]></title><description><![CDATA[Plus: Notion, one of AI's brightest stars, rolls out a fuller-fledged AI companion, and several additional executives leave OpenAI.]]></description><link>https://www.supervised.news/p/metas-llamas-go-local</link><guid isPermaLink="false">https://www.supervised.news/p/metas-llamas-go-local</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Thu, 26 Sep 2024 17:48:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BJod!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BJod!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BJod!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 424w, https://substackcdn.com/image/fetch/$s_!BJod!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 848w, https://substackcdn.com/image/fetch/$s_!BJod!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 1272w, https://substackcdn.com/image/fetch/$s_!BJod!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BJod!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png" width="1456" height="1099" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1099,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5289399,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BJod!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 424w, https://substackcdn.com/image/fetch/$s_!BJod!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 848w, https://substackcdn.com/image/fetch/$s_!BJod!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 1272w, https://substackcdn.com/image/fetch/$s_!BJod!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef697408-8162-4d0e-8683-5c5fce0a817f_2440x1842.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">an absolutely tiny llama, the size of a fly, sits on top of a home office desk, the desk is 100x the size of the bug-sized llama, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>Existing paid subscribers received a one week extension to account for a missed issue last week while I&#8217;m working through another lower back injury. Thank you everyone for your patience!</em></p><div><hr></div><h2>Meta&#8217;s model paths come together </h2><p>Meta&#8217;s Llama portfolio continues to grow in size, complexity, and nomenclature&#8212;but this time, it&#8217;s going after what&#8217;s quickly becoming one of the most important potential markets in AI: networks of task-specific smaller models.</p><p>While Meta has released larger models in a way to sow a little chaos at the top end of the quality spectrum with OpenAI and Anthropic, it seems to be setting its eyes on sowing <em>more</em> chaos at the lower end of the spectrum where Google&#8217;s Gemma 2 and Microsoft&#8217;s Phi models have become popular. Meta&#8217;s new tiny models, which it <a href="https://ai.meta.com/blog/llama-3-2-connect-2024-vision-edge-mobile-devices/">says in a blog post</a> are designed for edge use cases, fit neatly into one of the kind of four &#8220;bucket&#8221; cases. And Meta is now competing with many of its &#8220;rivals&#8221; in all four tiers.</p><ul><li><p><strong>&#8220;Tiny&#8221; ones that are suited for trivial and straightforward tasks with very little compute</strong>: Gemma 2, Phi-3, Llama 3.2 1B and 3B, and other various open source models. These models are well-suited for cases where you have access to little memory or compute, such as the edge devices Meta suggests.</p></li><li><p><strong>Smaller, more general purpose &#8220;workhorse&#8221; models designed to handle mostly trivial tasks that are still too complex for the &#8220;tiny&#8221; models</strong>: Gemma 2 9B, Llama 3.1 8B, Llama 3.2 11B, Mistral NeMo/Pixtral, and other various open source models. In these cases you&#8217;re less memory-constrained, but cost starts to become a consideration, making these smaller open-ish source models more attractive than APIs like GPT-4o-mini. These could involve summarizing a high volume of inbound documents, like sales calls or customer support tickets.</p></li><li><p><strong>The not-quite-as-good-as-GPT-4-but-still-pretty-good category where tasks start to become a little more ambiguous and open-ended</strong>: Meta&#8217;s 70B (and now 90B) models fit into here, but also you could throw Cohere&#8217;s RAG-focused Command-R+ and the larger versions of the Gemma-series models. You could imagine more complex searches, like across legal documents, that would be well-suited to a GPT-4o but are willing to take a small penalty on performance and ease of use&#8212;such as some increased complexity for fine-tuning&#8212;in exchange for lower costs.</p></li><li><p><strong>The substantially larger and more general purpose versions requiring high-quality responses</strong>: Llama 405B and the available foundation model APIs like Anthropic&#8217;s Sonnet, Gemini Pro, and GPT-4o. These are more suited to extremely high quality conversational experiences like the ones you might find in <a href="https://www.notion.so/product/ai">Notion</a> (which, more on that in a second).</p></li></ul><p>OpenAI is also seemingly trying to create a new category in those buckets with its slower-but-more-powerful o1 series models. In that case, OpenAI is once again flexing its research capabilities, even if that means trying to build out into a new niche. </p><p>But an enormous amount of emphasis from this announcement seems to land on that &#8220;edge device&#8221; part, with Meta off a number of different pieces of hardware it will work on. While that might end up providing some useful experiences on devices, the results from Apple&#8217;s (beta) approach integrated into one of the most prevalent operating systems on the planet <a href="https://www.washingtonpost.com/technology/2024/09/09/iphone-16-apple-intelligence-ai-event-2024/">have been pretty mixed</a>. The upper bound of these tiny models on edge devices&#8212;even augmented with a set of task-specific adaptors&#8212;isn&#8217;t so clear.</p><p>But Meta will also be able to capitalize on the enormous enthusiasm around local device AI inference, which has shown a lot of opportunity around rapid prototyping and development. Meta here gets to once again pounce on a developer community that&#8217;s craving more options to play with, and then just grab the best ideas for its own edge devices like its <a href="https://www.theverge.com/24253908/meta-orion-ar-glasses-demo-mark-zuckerberg-interview">prototype Orion headset</a>. And that starts with finding the absolute bare bones utility of those smaller models.</p><h2>Extending RPA to the edge</h2><p>While these smaller models are designed to work well on edge devices, with some customization on task-specific or company-specific data they excel at simple jobs like classification, entity extraction, or summarization. The emerging <a href="https://www.supervised.news/p/for-apple-its-rpa-all-the-way-down">basic use case for AI has basically been to construct</a> a more advanced version of robotic process automation (or RPA) through the use of smaller models chained together that all independently resolve these more trivial tasks.</p><p>That RPA-oriented &#8220;base case&#8221; has emerged as a pathway to quickly recognize an actual return on investment in AI tooling at a time when we&#8217;re all talking about <a href="https://www.supervised.news/p/about-that-ai-bubble">how we&#8217;re in an enormous AI bubble</a>. And a lot of this was <em>already</em> powered by Meta&#8217;s existing recent smaller Llama 3.1 model, though there has generally been some interest in experimenting with the even smaller ones amongst companies I talk to (particularly the &#8220;tiny&#8221; version of Phi 3).</p><p>Meta&#8217;s existing smaller model (8B) is getting an upgrade to being multimodal <em>and</em> slightly larger (11B), but still has a small enough footprint you could probably comfortably run a compact version of it on a Macbook Pro. A <a href="https://huggingface.co/blog/llama32">Hugging Face blog post indicates a compact version of its vision model takes up around 10GB of GPU RAM during inference</a>. And unsurprisingly, its &#8220;tiny&#8221; models <a href="https://x.com/ollama/status/1839007158865899651">are already available on Ollama</a>, a popular tool for running smaller models locally on devices like a Macbook Pro.</p><p>It&#8217;s also why we&#8217;ve started to see a kind of obsession over <a href="https://www.supervised.news/p/the-point-of-lightning-fast-model">extreme speed when generating results from smaller models</a>&#8212;like in the thousands of tokens per second. Rather than just looking for a single response, fast inference platforms enable companies to quickly chain together task-specific models into a kind of proto-agentic network to, at the end of the chain, accomplish some more difficult task.</p><p>While these kinds of &#8220;buckets&#8221; have been in the making for the past year or so, Meta&#8217;s more direct entry into each with the newest iteration of its Llama 3 models has formalized the boundaries. Rather than constructing the categories based on model size&#8212;which can vary while being similar in performance&#8212;the divisions are based on task complexity. </p><p>Many enterprises I&#8217;ve spoken with interested in actually putting AI-based tooling into production have <a href="https://www.supervised.news/p/putting-the-brakes-on-the-ai-hype">shifted their thinking in how they plan to deploy AI</a>. Expectations have dropped <em>dramatically</em> from a time when executives thought they could throw everything into an OpenAI API and assuming you&#8217;ve completely replaced entire departments, but those lowered expectations have also exposed how practical some of these smaller language models are for actually useful tasks. And hosting and deploying these smaller and more customized models is significantly more attractive now than it was a year ago.</p><p>Part of that is because of the significant cost to relying on APIs like OpenAI and Anthropic for those more powerful frontier models&#8212;even when it comes to their smaller counterparts like GPT-4o-mini. While the price of those API backed models continues to go down (and is, to be clear, a race to the bottom) it&#8217;s still considerably cheaper to host your own custom version of a smaller language model. And in some ways, those smaller models can even be a bit better because they&#8217;re more tailored to specific use cases.</p><p>Going <em>even smaller</em> than the &#8220;practical&#8221; small models exposes a lot of potential new use cases when it comes to edge devices&#8212;which could be phones, but could also be sensors, an array of Raspberry Pis. But when we&#8217;re talking about &#8220;edge&#8221; use cases, one in particular has lately gotten a lot of interest from executives and sources, even if it&#8217;s still just a curiosity: in-browser language model experiences. </p><p><a href="https://www.supervised.news/p/openais-silicon-valley-pivot">In this case, using WebGPU, the &#8220;edge&#8221; device is actually a browser</a>. One example thrown I&#8217;ve seen thrown around before is an in-browser SQL-generating assistant running in tandem with DuckDB, creating a whole kind of new local analytics experience. (There are some technical concerns that come up here around threads and available memory, but again, it&#8217;s an example.)</p><p>Meta, in the meantime, is also benefitting from jumping on the very hyped and exciting&#8212;and still relatively impractical in production&#8212;focus on local model development. <a href="https://github.com/ggerganov/llama.cpp">Llama.cpp</a>, <a href="https://ollama.com/">Ollama</a>, and other tools enabling local model usage has sparked a whole wave of enthusiasm around rapid prototyping and development of potential AI-powered products at a larger scale.</p><p>(Meta, alas, did not give me a heads up on the news or invite me to any of these events. One day!)</p><div><hr></div><h2>Notion tries to do a bit of everything in AI</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_uCq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_uCq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 424w, https://substackcdn.com/image/fetch/$s_!_uCq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 848w, https://substackcdn.com/image/fetch/$s_!_uCq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 1272w, https://substackcdn.com/image/fetch/$s_!_uCq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_uCq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:453579,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/gif&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_uCq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 424w, https://substackcdn.com/image/fetch/$s_!_uCq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 848w, https://substackcdn.com/image/fetch/$s_!_uCq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 1272w, https://substackcdn.com/image/fetch/$s_!_uCq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f8641a-9a3f-4a6f-949f-c0aa9946c201_1920x1080.gif 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On a rather hectic news day with the new Llama-series models and the <a href="https://www.cnbc.com/2024/09/25/openai-cto-mira-murati-announces-shes-leaving-the-company.html">abrupt departure of OpenAI CTO Mira Murati</a>, Notion also rolled out a <a href="https://www.notion.so/product/ai">significantly updated version of its integrated AI tools</a>.</p><p>Notion is considered among most of the sources and industry experts I speak with to be one of the most advanced companies when it comes to actually putting AI tools into production. There&#8217;s a handful of companies routinely come up, but most typically think Notion&#8217;s approach and architecture is one of the most complex and effective ways to utilize everything that&#8217;s available.</p><p>And in that way, a lot of investors, developers, and other startups look at Notion as a kind of &#8220;bellwether&#8221; in AI development&#8212;letting the startup do all the experimentation and product development to see what actually is possible and sticks with customers.</p><p>Its latest update includes connectors to Slack and Google Drive, introducing a new set of unstructured data pipelines, as well as other updates like analyzing files and image attachments, style-specific document generation and editing, and direct selection of knowledge sources.</p><p>&#8220;I&#8217;d say we are small enough and nimble enough that we don&#8217;t have to go through a bunch of layers of approval to try something new out,&#8221; Shir Yehoshua, AI engineering lead at Notion, told me. &#8220;We have a very prototype-y culture. Our general model is to prototype it first and if it works&#8212;whether that&#8217;s an evaluation tool, or an actual LLM provider, a vector DB&#8212;as fast as we possibly can and get very empirical data on how well it works. And also try not to fall into a sunk cost fallacy.&#8221;</p>
      <p>
          <a href="https://www.supervised.news/p/metas-llamas-go-local">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Putting the brakes on the AI hype]]></title><description><![CDATA[Enterprises are taking a more methodical approach when figuring out how to put AI tools into production&#8212;and considering much longer timelines.]]></description><link>https://www.supervised.news/p/putting-the-brakes-on-the-ai-hype</link><guid isPermaLink="false">https://www.supervised.news/p/putting-the-brakes-on-the-ai-hype</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Tue, 10 Sep 2024 18:17:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GyLR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GyLR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GyLR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 424w, https://substackcdn.com/image/fetch/$s_!GyLR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 848w, https://substackcdn.com/image/fetch/$s_!GyLR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 1272w, https://substackcdn.com/image/fetch/$s_!GyLR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GyLR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png" width="1456" height="1100" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1100,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5023211,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GyLR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 424w, https://substackcdn.com/image/fetch/$s_!GyLR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 848w, https://substackcdn.com/image/fetch/$s_!GyLR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 1272w, https://substackcdn.com/image/fetch/$s_!GyLR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3427d7c8-e5ee-4ffc-8686-6f9c09ea7b8b_2432x1838.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a small pixar-style robot in a business suit standing in front of a whiteboard, the whiteboard has a lot of math equations on it, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p>Over the past few weeks I&#8217;ve been asking dozens of industry executives and sources one of the same thing: how have the questions enterprises have been asking about AI tools changed in the last six to nine months as they get more serious about getting projects into production?</p><p>Almost universally the response has been that the tenor of these conversations has changed, especially when talking to enterprises and executives that initially bought into the original AI hype. Or, more specifically:</p><ul><li><p>Those enterprises and executives are still bought in on putting out AI products, and recognize an urgency to find use for them to either remain competitive or to save on costs.</p></li><li><p>They&#8217;re much more educated and more deliberate about the way they are looking to implement these models, and <em>much</em> more discerning when it comes to the cost of operating those tools.</p></li></ul><p>In short, enterprises are getting, well, smarter about what they are looking for when they are evaluating AI-based tools. And it&#8217;s generally confirmed the kind of &#8220;vibe shift&#8221; that started to happen in the past six months. <a href="https://www.supervised.news/p/about-that-ai-bubble">Despite the recent fuss about whether or not we&#8217;re in an AI bubble</a>, the &#8220;hype cycle&#8221; actually ended quite a while ago as companies have started to figure out how these things are going to actually useful <em>now</em>.</p><p><a href="https://www.supervised.news/p/ai-in-august-rbac-is-back-data-as">And we&#8217;ve largely started to see that emerge already</a> in more straightforward&#8212;and boring&#8212;tasks, even if companies aren&#8217;t very loud about it. The era of thinking you could wave a OpenAI-shaped wand and generate completely new lines of business or completely replace an entire class of workers is effectively gone, but at the same time, there&#8217;s enough signal that companies aren&#8217;t just throwing the idea of deploying a generative AI tool into the trash or writing it off as an experiment.</p><p>&#8220;It&#8217;s a lot less frenetic as compared to how it was,&#8221; Brian Raymond, CEO of <a href="https://unstructured.io/">unstructured data ETL provider Unstructured.io</a>, told me. &#8220;Things aren&#8217;t changing every hour or every week from an industry standpoint, we have a little more confidence on where things are going, and there are fewer surprises. We&#8217;re in this unsexy phase, but this is the phase in which most of the value is gonna get created among the organizations that are trying to leverage generative AI. It&#8217;s less rapid experimentation and mind-blown emojis, and it&#8217;s more like, let&#8217;s drive ROI.&#8221;</p><p>All this is pretty far from the days right after the launch of GPT-4 when certain executives at a company would barge into a room shouting &#8220;AI&#8221; and leaving the team scrambling to build <em>something. </em></p><h2>The list of needs for enterprises keeps growing</h2><p><a href="https://www.supervised.news/p/about-that-ai-bubble">While we all talk about whether or not AI is in a bubble</a> for the Very Big Models, what&#8217;s become clearer is how enterprises are evaluating generating near-term ROI through the use of language models. Most companies actively using these tools I&#8217;ve spoken with have done that through <a href="https://www.supervised.news/p/ai-in-april-and-q2-rpa-in-focus-holistic">automating smaller tasks, like summarization and classification, with customized smaller language models</a>. By focusing on these RPA-like tasks with smaller models, they&#8217;re able to save significantly on cost relative to paying for an API product like one from OpenAI or Google.</p><p>These kinds of tasks are augmented with the use of information retrieval techniques like retrieval augmented generation (or RAG),<a href="https://www.supervised.news/p/a-clever-way-to-sidestep-some-of"> which effectively fetches some extra information for a prompt in a given language model</a>. Companies convert their data&#8212;<a href="https://www.supervised.news/p/lakes-catalogs-and-snowflakes-full">which most likely sits in a provider like Snowflake, Databricks, or others</a>&#8212;into a format through a process called embedding. They then make that data readily available on-demand for these prompts. (Most <a href="https://www.supervised.news/p/a-new-whisper-in-the-ai-analytics">developers I&#8217;ve spoken with have also joked</a> that all of these &#8220;new&#8221; techniques that people keep &#8220;discovering,&#8221; <a href="https://www.supervised.news/p/where-graph-databases-live-in-a-future">like RAG augmented with a graph database</a>, are also decades old.)</p><p>The needs of those enterprises are also becoming a lot more sophisticated as time goes on. Rather than just firing stuff into a prompt, enterprises are increasingly looking for governance tooling around these language models, <a href="https://www.supervised.news/p/a-classic-data-problem-is-taking">such as lineage</a> and, seemingly more recently, <a href="https://www.supervised.news/p/ai-in-august-rbac-is-back-data-as">role-based access control (or RBAC</a>). Everything old, it seems, is new again&#8212;just this time in AI.</p>
      <p>
          <a href="https://www.supervised.news/p/putting-the-brakes-on-the-ai-hype">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[AI in August: RBAC is back, data as a product, and something about a bubble]]></title><description><![CDATA[The data engineers are more important than ever these days.]]></description><link>https://www.supervised.news/p/ai-in-august-rbac-is-back-data-as</link><guid isPermaLink="false">https://www.supervised.news/p/ai-in-august-rbac-is-back-data-as</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Thu, 05 Sep 2024 19:47:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!NiD-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NiD-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NiD-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 424w, https://substackcdn.com/image/fetch/$s_!NiD-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 848w, https://substackcdn.com/image/fetch/$s_!NiD-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 1272w, https://substackcdn.com/image/fetch/$s_!NiD-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NiD-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png" width="1456" height="1101" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1101,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6300862,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NiD-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 424w, https://substackcdn.com/image/fetch/$s_!NiD-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 848w, https://substackcdn.com/image/fetch/$s_!NiD-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 1272w, https://substackcdn.com/image/fetch/$s_!NiD-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F581f3e5d-2070-42a0-8725-f69e4b5fa0be_2426x1834.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a small friendly robot staring up at a large tree as the leaves start to change with autumn arriving, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p>The general consensus from sources and experts I talk to is that the universal truth of summer in tech and venture capital held this year: not a lot happened and everyone was on vacation.</p><p>That&#8217;s starting to pick up as we head into the fall, both with a lot of stock prices flying in opposite directions and earnings reports wrapping up. As teams start to trickle back into the office (home or otherwise), we&#8217;re starting to come up on the whole &#8220;AI in prod&#8221; mental deadline that showed up in a lot of companies I&#8217;ve spoken with over the last few months. The failed projects are going in the trash, and the stuff that&#8217;s useful is working its way into roadmaps.</p><p>That kind of &#8220;nearness&#8221; to production, as usual, is showing up in more friction and barriers to getting things out the door. In the past few months, that&#8217;s centered again on enterprise governance requirements and other some other kinds of technical walls&#8212;and a kind of pragmatic shift in the works at a lot of these companies think about their team structure. </p><p>And a lot of arguing about bubbles, which, as usual, is a much more complicated and nuanced situation than it seems at face value. So, with all that said, here&#8217;s what&#8217;s coming up as we head into the fall:</p><ul><li><p><strong>Role-based access control has entered the chat</strong>. Shortened to RBAC (and pronounced are-back), developers and industry executives had added the need to partition out who can get access to what data in an a language model into the very large stack of <em>stuff</em> that&#8217;s needed to get done before something enters production. This is coming up mostly in the context of internal chatbots, but there&#8217;s more to it under the hood.</p></li><li><p><strong>The data engineers are now under the spotlight.</strong> AI (generative or otherwise) is quickly being recognized as a data-centric product. The data engineering teams, traditionally bound by the hip to analysts, now seem to be getting a lot more exposure at larger organizations.</p></li><li><p><strong>It has been (0) days since we have said AI is a bubble</strong>. After more than a year of talking about whether AI is in a bubble, we are again&#8230; talking about whether AI is in a bubble. We&#8217;ll go over this again, but the answer is that it&#8217;s complicated.</p></li></ul><h2>The RBAC wall to get into production</h2><p>One of the biggest potential problems most companies that I talk to are fretting about is whether some random employee will get access to sensitive data that they aren&#8217;t allowed to see. This isn&#8217;t restricted to PII or anything like that&#8212;it could even be in the context of someone who isn&#8217;t on an HR team getting access to salary information they aren&#8217;t supposed to able to view.</p><p>This potential snafu has a lot of names within companies, though most people I talk to throw the term &#8220;leakage&#8221; on it. As a result, there&#8217;s some skittishness around whether to basically feed <em>all</em> of a company&#8217;s data into a custom model (fine-tuning or otherwise), trying to slap some guardrails on it, and hope for the best.</p><p>Well, that&#8217;s not the first step companies usually take when they are looking at deploying something custom that taps company data. Instead, that company data is embedded into a database (such as a vector database, Postgres, or MongoDB) and a model can fetch it for a prompt through a process called retrieval-augmented generation, or RAG. There&#8217;s varying levels of sophistication to it, with <a href="https://www.supervised.news/p/where-graph-databases-live-in-a-future">some companies also using it in conjunction with graph search</a>.</p><p>Each data point you&#8217;d want to retrieve is embedded in a &#8220;chunk,&#8221; or some block of information. The chunks can vary in size and format, ranging from just sentences in an email to full documents, and all that data has to be pre-processed in some form to even get to a point where a developer can embed it and make it available through RAG. But the list of governance requirements is continuously growing as companies get more serious about putting tools on top of language models into production, and determining who can access each chunk of information is now part of the set of requirements.</p><p>This is where role-based access control comes into play by assigning a kind of insulation layer around data points to determine who gets access to what. Some companies have already been <a href="https://www.elastic.co/search-labs/blog/rag-and-rbac-integration">trying to tackle the RBAC problem since earlier this year</a>, but the momentum for it seems to generally be picking up among companies and investors I talk to&#8212;likely as a byproduct of these companies trying to get these products based on language models out the door.</p><p>&#8220;We went from a world where we were focused on data loaders in early 2023, to a world now you have to think about RBAC and bounding boxes so you have traceability and can show your homework,&#8221; Brian Raymond, CEO of unstructured data ETL startup Unstructured.io, told me. &#8220;Regardless of what&#8217;s in that manilla envelope of data that gets written to a vector database, you have to have timestamps, owner, which group it belonged to for RBAC, version history, and a lot of other information. Any time you&#8217;re doing more than a proof-of-concept, it is a complete blocker.&#8221;</p><p>As with the example above, this mostly comes up when talking about companies building internal chatbots that replace sprawling Wikis that are increasingly inaccessible. But it&#8217;s pretty easy to see how it extends out to potential end users for even a customer-facing product&#8212;after all, an employee <em>is</em> a customer of an internal-facing product.</p><p>&#8220;We generate around 30 types of metadata during preprocessing, and that&#8217;s critical when you&#8217;re doing retrieval,&#8221; Raymond told me. &#8220;And we&#8217;re transposing the requirements data engineering teams have been developing over the last ten years onto the generative AI data stack.&#8221;</p><p>Again, guardrailing is <em>one</em> way to do it, but most developers and companies I talk to are increasingly recognizing that they&#8217;ll need something more sophisticated in place. And that leads to another emerging problem in AI, which is&#8230;</p><h2>Someone please check on the data engineers</h2><p>Each of these companies is also increasingly realizing that the problem of getting a generative AI product in production is&#8230; the same problem as getting a regular data centric products into production: high quality proprietary data.</p><p>Data engineers have traditionally built out pipelines that feed into more &#8220;classic&#8221; workflows like analytics. The emergence of Dbt and the whole analytics engineering role started to extend the responsibilities of analysts to handle more data engineering-oriented tasks. But there are two emerging trends within all these companies that are actually trying to move past a proof-of-concept into something that can generate a real return:</p><ul><li><p>Many companies are tasking <em>software engineers</em> with building all of these AI-powered tools, who have to quickly grapple with the importance high fidelity data plays into all of these tools.</p></li></ul>
      <p>
          <a href="https://www.supervised.news/p/ai-in-august-rbac-is-back-data-as">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The point of lightning-fast model inference]]></title><description><![CDATA[We're obsessed with generating thousands of tokens a second for a reason&#8212;and it isn't just to wow end users with text showing up on a screen really fast.]]></description><link>https://www.supervised.news/p/the-point-of-lightning-fast-model</link><guid isPermaLink="false">https://www.supervised.news/p/the-point-of-lightning-fast-model</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Tue, 27 Aug 2024 22:53:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6GkB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6GkB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6GkB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 424w, https://substackcdn.com/image/fetch/$s_!6GkB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 848w, https://substackcdn.com/image/fetch/$s_!6GkB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 1272w, https://substackcdn.com/image/fetch/$s_!6GkB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6GkB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png" width="1456" height="1067" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/eab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1067,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2809528,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6GkB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 424w, https://substackcdn.com/image/fetch/$s_!6GkB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 848w, https://substackcdn.com/image/fetch/$s_!6GkB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 1272w, https://substackcdn.com/image/fetch/$s_!6GkB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feab62297-98df-44b7-9a9d-8571e337739c_1634x1198.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><strong>small robot furiously typing on a typewriter, the typewriter has flames coming out of it like it is spewing letters and words, sunday funny comics aesthetic &#8212; midjourney</strong></figcaption></figure></div><p>When ChatGPT first came out, it started out with&#8212;and popularized&#8212;a fun user experience to go with getting a message back from the early AI product: having the words print out in sequence as they are being generated.</p><p>These tools built on language models don&#8217;t <em>have</em> to do that. You could instead wait for the whole response to get generated, and then show the response. But for one reason or another, that experience largely stuck with standard interfaces for apps built on top of language models&#8212;even as the speed of token production has become so high that you probably can&#8217;t even discern it at this point.</p><p>But while there&#8217;s some threshold where the speed at which tokens are generated just kind of stops being visually noticeable or helpful, there&#8217;s another reason all this is happening under the hood. These responses generated at lightning speeds aren&#8217;t built just for humans&#8212;they&#8217;re built for the bots that they&#8217;ll be talking to in the future.</p><p>&#8220;I don&#8217;t think the interesting work is human-read in the future&#8212;the interesting work is machine-read,&#8221; Andrew Feldman, CEO and co-founder of Cerebras Systems, told me. &#8220;What you&#8217;ll see in the future is concatenations of models, where the output of one is the input the next. That latency stacks. If you wanted to link 6 or 8 of these together, you wait a minute to get an answer. What we know is nobody waits a minute.&#8221;</p><p><a href="https://cerebras.ai/">Cerebras Systems</a>, a developer of custom AI chips (<a href="https://cerebras.ai/product-chip/https://cerebras.ai/product-chip/">that are also </a><em><a href="https://cerebras.ai/product-chip/https://cerebras.ai/product-chip/">colossal</a></em>), is one of the latest companies to step into this blistering speed race with its new product, <a href="https://cerebras.ai/inference">Cerebras Inference</a>. Feldman tells me that you&#8217;ll get speeds above 1,800 tokens per second on the smaller Llama 3.1 8B model, and 450 tokens per second on the larger Llama 3.1 70B model. And Cerebras&#8217; product inferences both models at full 16-bit precision, rather than a compressed 8-bit model that is often the default option&#8212;particularly for the larger models.</p><p>This all becomes relevant in the context of (the current semi-fever dream) of agents&#8212;a collection of models that can operate more autonomously and solve complex tasks by passing them amongst one another until they return a result. Each individual model requires a single call, and a decision on where to route that result, which can hypothetically balloon with the complexity of tasks. And all this doesn&#8217;t even include the possibility of failures in a chain of operations, such as a prompt rejection (such as due to it returning personally identifiable information).</p><p>Cerebras Systems is one of a number of companies looking to exploit architecture alternatives to Nvidia&#8217;s hardware in order to satisfy a very broad set of use cases. And they are trying to demo the lightning speed of all of these operations, much like companies like Groq and SambaNova Systems. Inference <a href="https://www.together.ai/blog/together-ai-partners-with-meta-to-release-meta-llama-3-for-inference-and-fine-tuning">platforms like Together AI</a> and Fireworks also try to push the envelope on tokens per second.</p><p>And while it all moves faster than a human eye can read, it turns out that this blistering speed is just one of what are likely many precursors to the ability to build out networks of models that can satisfy some of the potential dream scenarios for language models.</p><p>&#8220;This new processing power is a game-changer,&#8221; Jonathan Corbin, co-founder and CEO of <a href="https://www.mavenagi.com/">Maven AGI</a>, told me. &#8220;It allows AI agents to handle vast datasets in real-time, make more nuanced decisions, and adapt quickly to new information. We view this as crucial for developing AI agents with human-like understanding and responsiveness.&#8221;</p><h2>The case for agents and speed</h2><p>Right now we&#8217;re a considerable ways off from creating some kind of one-size-fits-all &#8220;agent&#8221; that can perform a task of more arbitrary complexity., That could either be through the use of some model with insane reasoning capabilities or a long, long chain of models strung together. </p><p>But there is already a lot of low-hanging fruit that current, off-the-shelf hardware has the potential to resolve. More specifically, these language models&#8212;particularly the smallest Llama 3.1 8B model&#8212;are able to solve compact tasks that involve processes like classification, summarization, or entity extraction. Each of these excel in customer service use cases, particularly when routing problems and determining whether to rope in a human to resolve it.</p><p>The complete base case of all of this is having a &#8220;robot&#8221; complete each of those tasks in isolation, and then moving it onto the next robot to complete the next incremental step. The classic term for this is, aptly, robotic process automation (or RPA). Except instead of having a bot click around on a website to test something, this process is taking in information and determining what to do with it at a much higher level. These are the same types of tasks companies were using older language models, like BERT, for&#8212;they&#8217;re just substantially more advanced and can handle more complex problems.</p><p>Klarna, the company people reflexively point to when it comes to the potential of dropping AI into customer service, <a href="https://downloads.ctfassets.net/4pxjo1vaz7xk/161qRVqB2B2N8MYwLl9x5A/abcc53df5acfac3fa4bb36d4621db99c/Klarna_Holding_AB_Interim_Report_2024__ENG_.pdf">reiterated in an earnings report</a> that the use of AI was able to perform the work of more than 700 employees, and reduce the average resolution time for a customer service case from 11 minutes to just 2 minutes. It might be boring, but <a href="https://www.supervised.news/p/batch-processing-and-the-rise-of">batch processing is largely where there&#8217;s a lot of value to extract</a>&#8212;and waiting a fifth of the time to resolve some customer service issue seems pretty helpful!</p>
      <p>
          <a href="https://www.supervised.news/p/the-point-of-lightning-fast-model">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The free tokens will continue until business improves]]></title><description><![CDATA[The SaaS-ification of AI continues, with OpenAI and Google ready to go full 2010s to get people on board. Plus, durable execution returns as a topic du jour.]]></description><link>https://www.supervised.news/p/the-free-tokens-will-continue-until</link><guid isPermaLink="false">https://www.supervised.news/p/the-free-tokens-will-continue-until</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 23 Aug 2024 21:47:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!mWpL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mWpL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mWpL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 424w, https://substackcdn.com/image/fetch/$s_!mWpL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 848w, https://substackcdn.com/image/fetch/$s_!mWpL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 1272w, https://substackcdn.com/image/fetch/$s_!mWpL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mWpL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png" width="1456" height="1183" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1183,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2177152,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mWpL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 424w, https://substackcdn.com/image/fetch/$s_!mWpL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 848w, https://substackcdn.com/image/fetch/$s_!mWpL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 1272w, https://substackcdn.com/image/fetch/$s_!mWpL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d22f37a-2f49-4b6e-a399-c0c5bec299fe_1592x1294.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a small friendly robot standing on a massive pile of arcade tokens, the pile of tokens is miles high, sunday funny comics aesthetic &#8212; midjourney</figcaption></figure></div><p><em>Two topics today: first is on all the free stuff we&#8217;re getting to convince me to fine-tune a model; second, why everyone is back to talking about durable execution. Once again, thank you everyone for your patience as I get ramped back up!</em></p><div><hr></div><h2>The free stuff playbook is back in AI</h2><p>When Sam Altman was<a href="https://www.supervised.news/p/with-sam-altman-out-where-does-openai"> briefly removed from OpenAI in November last year</a>, it represented this kind of weird moment in AI where most companies <em>weren&#8217;t</em> led with a very Silicon Valley growth-at-all-costs SaaS-ify everything mentality. Sure, there were a lot of mega-rounds, but one of OpenAI&#8217;s early advantages was that it was able to<a href="https://www.supervised.news/p/openais-silicon-valley-pivot"> undercut the competition with an easy-to-use API</a>.</p><p>Since then, the SaaS-ification of AI seems to have quickly become the norm. Providers are constantly racing each other to see how quickly they can send their prices as close to zero as possible to one-up each other. This is true with the more capable foundation models, but it&#8217;s also <em>particularly</em> true for the workhorse models from those model providers&#8212;the ones that benefit the most from customization on proprietary data.</p><p>And while the basic APIs are easy to simply swap amongst each other&#8212;whether that&#8217;s Google, Anthropic, OpenAI, or an inference platform like Together AI, that you can literally drop in with a few lines of code&#8212;the &#8220;lock-in&#8221; now potentially comes in an alternate form: fine-tuning APIs using proprietary data. </p><p>With fine-tuned models, which are essentially modified to perform well on specific use cases for your business like summarizing documents, the switching cost to a new service isn&#8217;t zero. Rather, the switching cost is both the time, headache, and tokens necessary to re-do a model customization job on a different service. And, just like in the on-demand and Web 2.0 boom, the two largest providers are trying to snag you with a bunch of free stuff.</p><p>OpenAI this week said <a href="https://openai.com/index/gpt-4o-fine-tuning/">it was making fine-tuning available for its GPT-4o</a>. And<a href="https://platform.openai.com/docs/guides/fine-tuning"> while fine-tuning for GPT-4o mini has been available since late July to its higher-tier organizations</a>, it&#8217;s now available for all paid usage tiers, <a href="https://openai.com/index/gpt-4o-fine-tuning/">per the blog post they put out</a>. OpenAI already made 2 million training tokens available per day for GPT-4o mini, and this week it was making 1 million training tokens available per day for GPT-4o. Those free tokens are available through the next month, but also we know exactly how the price wars for Uber, DoorDash, Lyft, and others went with what seemed like endless subsidization. (Sure, that was ZIRP-era, but this time around we have GPU credits too.)</p><p>This is just a little less than two weeks after Google<a href="https://developers.googleblog.com/en/gemini-15-flash-updates-google-ai-studio-gemini-api/"> announced it was making its Gemini Flash 1.5 tuning product available for all developers</a>. Gemini Flash 1.5, in addition to being <em>very cheap</em>, also hands out 1.5 <em>billion</em> free tokens per day to developers. Per Google&#8217;s pricing page, <a href="https://ai.google.dev/pricing">the tuning price for Gemini Flash is free of charge</a>. Google even offers a free tier for its Gemini Pro 1.5, though it&#8217;s considerably lower volume at 1.6 million tokens per day at a much more limited rate (and per its pricing page tuning is unavailable). But the point is Google (and OpenAI) for now can afford this in the first place in the hopes of hosting the portfolios of custom models.</p><p>And while these pay-as-you-go models aren&#8217;t exactly what massive enterprises would adopt&#8212;that would more likely be provisioned usage and contracts with much more aggressive service level agreements&#8212;they serve as an easy and very aggressive onboard to getting stuff into production that can serve as a very enticing carrot for further usage.</p><p>It feels like the industry is quickly coalescing around a handful of companies with the resources to produce these kinds of powerful fine-tuned workhorse models&#8212;and they&#8217;re all trying to one-up each other in how much of a product they can put in front of companies to try to lock them into an ecosystem. We essentially saw all this before, in multiple times like the on-demand era, with companies giving away as much free stuff as they possibly could to eventually convert it to a sustainable business at <em>some point</em>.</p><p>And these more workhorse-focused fine-tuned models are essentially one of the end-games for enterprise focused tasks. Achieving some benefit in fine tuning for tasks like classification or summarization, in the case of Gemini, <a href="https://ai.google.dev/gemini-api/docs/model-tuning">generally requires a number of examples in the low 100s</a>.</p><p>But once you&#8217;ve collected those models, you can&#8217;t move them&#8212;OpenAI has them, and if you want to move them, you&#8217;ll have to re-do the process all over again either through Gemini or some alternate service. This is one of the main selling points of fine-tuning open source models, as you can move around fine-tuned versions of, say, Llama 3.1 8B as needed because they can port to any infrastructure that can run that kind of model.</p><p>And indeed, that&#8217;s also <a href="https://www.supervised.news/p/lakes-catalogs-and-snowflakes-full">part of the threat that Databricks and Snowflake pose to OpenAI (along with Google and company)</a>. If fine-tuning with proprietary data is the way to unlock value&#8212;like building a custom summarization tool for my sales calls&#8212;the button closest to the data with the most optionality is going to be the most valuable. OpenAI faces the uphill battle of courting enterprises with its API that already likely have accounts with all of these abstraction layers and hyperscalers in the first place.</p><p>While we haven&#8217;t seen what it looks like just yet, the next obvious player here will be whatever rabbit Anthropic pulls out of its hat with its next-generation workhorse model. Right now you <a href="https://www.anthropic.com/news/fine-tune-claude-3-haiku">can fine-tune Claude Haiku through Amazon Bedrock</a>, but it is nowhere remotely close to as easy as it is for the pay-as-you-go API approach that both Google and OpenAI offer. But Anthropic also has the benefit of Amazon promoting it to AWS customers, where that data for fine tuning already sits for many companies.</p><p>It also doesn&#8217;t help OpenAI that Cursor, a more flexible AI-enabled IDE where you plug in <em>any</em> provider&#8217;s API key, has a lot of hype among developers and investors I talk to lately. <a href="https://www.supervised.news/p/a-new-whisper-in-the-ai-analytics">I wrote about the company several months ago about how one of Cursor&#8217;s biggest advantages was that you could pop in any code generation API key</a>, making OpenAI even more disposable if a new superior model comes out&#8212;either proprietary or open source. (Cursor&#8217;s <a href="https://www.cursor.com/blog/series-a">parent company Anysphere announced it raised $60 million this week</a>.)</p><p>This constant deluge of free stuff shouldn&#8217;t exactly surprise anyone because all these companies are staffed with the same type of executive that lived through the on-demand era. For OpenAI in particular, both Sam Altman and <a href="https://www.ycombinator.com/blog/welcome-x11">chief operating officer Brad Lightcap</a> hail from Y Combinator, the storied accelerator that spawned both Instacart and Dropbox&#8212;companies that, among others, popularized a &#8220;growth first, business later&#8221; mindset. Meanwhile, though Google has seemed to put together a rather comprehensive stack for AI, it still feels a lot like the same old Google that has been shoving free ad credits into the face of anyone who decides to open a G Suite account.</p><p>The on-demand era, and really Web 2.0 broadly, was often affectionately dubbed &#8220;venture-funded capitalism,&#8221; and came to a halt down when all that free capital dried up. With the duel in on-demand, it was largely around a handful of moat-building exercises: brand affinity, collecting data, wooing drivers (who ended up running both anyway), and altering consumer behavior away from hailing a cab or calling for delivery. For these companies, the last one was really the only one that was truly successful as they all struggled to build out long-term sustainable businesses.&nbsp;</p>
      <p>
          <a href="https://www.supervised.news/p/the-free-tokens-will-continue-until">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[About that AI bubble]]></title><description><![CDATA[AI can be far from achieving its potential, but it can also be really useful right now.]]></description><link>https://www.supervised.news/p/about-that-ai-bubble</link><guid isPermaLink="false">https://www.supervised.news/p/about-that-ai-bubble</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 16 Aug 2024 19:05:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9S_d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9S_d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9S_d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 424w, https://substackcdn.com/image/fetch/$s_!9S_d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 848w, https://substackcdn.com/image/fetch/$s_!9S_d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 1272w, https://substackcdn.com/image/fetch/$s_!9S_d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9S_d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png" width="1456" height="1097" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1097,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4715339,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9S_d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 424w, https://substackcdn.com/image/fetch/$s_!9S_d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 848w, https://substackcdn.com/image/fetch/$s_!9S_d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 1272w, https://substackcdn.com/image/fetch/$s_!9S_d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb2f9273-21cc-4531-8bb0-bcb297d483e9_2444x1842.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a large bubble rising above a large cyberpunk type city, blade runner 2049 aesthetic, deep blue hue, --ar 4:3 &#8212; </figcaption></figure></div><p><em>I had to finish this one on my phone, so apologies ahead of time for formatting or spelling errors!</em></p><div><hr></div><h2>The duality of the AI bubble</h2><p>Much has been said in the last several months about whether or not the bubble for investment in AI is going to pop&#8212;particularly following the acquihires-slash-<a href="https://www.reuters.com/technology/artificial-intelligence/google-hires-characterai-cofounders-licenses-its-models-information-reports-2024-08-02/">acquilicenses</a>-slash-whatever of previously hyped AI startups <a href="https://www.theverge.com/2024/8/2/24212348/google-hires-character-ai-noam-shazeer">Character.AI</a>, <a href="https://www.bloomberg.com/news/articles/2024-03-21/microsoft-to-pay-inflection-ai-650-million-after-scooping-up-most-of-staff?embedded-checkout=true">Inflection</a>, and <a href="https://www.theverge.com/2024/7/1/24190060/amazon-adept-ai-acquisition-playbook-microsoft-inflection">Adept</a>. These are companies that were going after whole generative AI experiences that required massive GPU clusters&#8212;the kinds of products that have sent Nvidia rocketing toward a valuation that at <a href="https://www.reuters.com/technology/nvidia-sets-eye-1-trillion-market-value-2023-05-30/">one point passed $1 trillion.</a></p><p>The list seems to continue to grow every week. One of the most-hyped AI-first products, the Humane Ai Pin, <a href="https://www.theverge.com/2024/8/7/24211339/humane-ai-pin-more-daily-returns-than-sales">has seemingly gone&#8230; not well</a>. The whole industry in and around AI rode a colossal hype wave to the point that companies even adjacent to &#8220;core AI&#8221; were picking up valuations upwards of hundreds of times annual recurring revenue. </p><p>Now, the big question is, where&#8217;s the revenue? Where is the business? Where are the killer apps? Is all this just a complete waste of time with a very limited potential industry that requires colossal upfront investments? Is OpenAI just going to go out of business?</p><p>Well, two things can be true at the same time: the foundation model providers building out various colossi and dreaming of artificial general intelligence aren&#8217;t living up to the hype right now; and AI is actually really, really useful and <a href="https://www.klarna.com/international/press/klarna-ai-assistant-handles-two-thirds-of-customer-service-chats-in-its-first-month/">already realizing returns within some organizations&#8212;just perhaps not in particularly exciting ways</a>.</p><p>Instead, the answer to whether or not we&#8217;re in an AI bubble is probably more disappointing and frustrating: it&#8217;s complicated.</p><h2>The less difficult, and more valuable, use cases for AI right now</h2><p>I&#8217;ve written about those types of use cases before, but let&#8217;s talk about them again: batch processing and robotic process automation (or RPA). These types of use cases are really straightforward and the kinds of things we were doing that pre-date ChatGPT and its neighbors. Summarization, entity extraction, classification, and sentiment analysis are just a few examples&#8212;but the point is that you don&#8217;t need some massively powerful foundation model that costs $10 per million tokens to do it.</p><p>In most of these use cases companies can get away with using models like a customized version of GPT-3.5 Turbo with a system in place to receive relevant information in a specialized database through a process called retrieval augmented generation, or RAG for short. Most of the foundation model providers have these kinds of &#8220;workhorse&#8221; models for a reason&#8212;they accomplish the vast majority of needed tasks and perform at a relatively high level at a relatively low cost.</p><p>This extends to what we&#8217;d consider &#8220;agents&#8221; as well. The kind of one-size-fits-all-self-reasoning-autonomous-replace-us-all digital entity seems not only far away, but in most of these very achievable cases completely unnecessary. There are some potential gains to be had through an increasing ability to self-orchestrate, but the near term also points to networks of agents that are again customized to complete a closed number of tasks before passing them off to the next &#8220;agent.&#8221; </p><p>When we talk about &#8220;AI in prod,&#8221; a which is an umbrella so comically large you could fit a small city under it, the two get bunched together. AGI is the goal to accomplish all these tasks that can&#8230; already be accomplished by what&#8217;s currently on the shelf. Developers I talk to say customer service ticket routing and escalation is a relatively straightforward mechanism through these models&#8212;a very time intensive process that requires a lot of people doing a lot of mundane work. And, to be clear, it doesn&#8217;t even replace these representatives. (One source even found some of these super complicated approaches to fine tuning and deploying small models like Llama 3 overkill, which could easily be handled by a fine-tuned GPT-3.5 Turbo.)</p>
      <p>
          <a href="https://www.supervised.news/p/about-that-ai-bubble">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Revisiting that old Google AI memo]]></title><description><![CDATA[A few things have changed since a Google researcher sounded the alarm on Google's risk to open source AI in a leaked memo last year.]]></description><link>https://www.supervised.news/p/revisiting-that-old-google-ai-memo</link><guid isPermaLink="false">https://www.supervised.news/p/revisiting-that-old-google-ai-memo</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 09 Aug 2024 19:13:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!REoK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!REoK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!REoK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 424w, https://substackcdn.com/image/fetch/$s_!REoK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 848w, https://substackcdn.com/image/fetch/$s_!REoK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 1272w, https://substackcdn.com/image/fetch/$s_!REoK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!REoK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png" width="1456" height="1099" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1099,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6272896,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!REoK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 424w, https://substackcdn.com/image/fetch/$s_!REoK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 848w, https://substackcdn.com/image/fetch/$s_!REoK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 1272w, https://substackcdn.com/image/fetch/$s_!REoK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87b18e7a-476d-4460-8da9-48a2444635e1_2446x1846.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Hope everyone had a productive few weeks! I&#8217;m back and resuming publishing, though slowly ramping back up as I get back in touch with people, so there may be fewer issues for the next two weeks or so. Thanks everyone for you patience!</em></p><div><hr></div><p>A bit more than a year ago at the dawn of the comically short period that we&#8217;ll call &#8220;modern AI,&#8221; a memo from a Google researcher <a href="https://www.semianalysis.com/p/google-we-have-no-moat-and-neither">found its way onto to SemiAnalysis</a> that essentially served as a warning of the risks open source AI presented to foundation model providers like OpenAI and Google. (<a href="https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks">DeepMind CEO Demis Hassabis also confirmed its authenticity to The Verge</a>.)</p><p>The <a href="https://www.supervised.news/p/a-model-in-your-pocket">original Llama model from Meta had recently leaked</a>, and it became pretty clear that the developer community was going to run off with it and develop a lot of new techniques to improve the performance of open source models, <a href="https://www.supervised.news/p/for-apple-its-rpa-all-the-way-down">many of which were actually adopted by Apple for its forthcoming AI suite Apple Intelligence</a>. But at the time the catch phrase was that Google &#8220;had no moat, and neither does OpenAI.&#8221;</p><p>A few things have changed since then! And it seems like as good a time to look at where Google actually sits in this whole mess as its latest version of Gemini Pro, its own foundation model, and its latest micro-model Gemma 2 2B both <a href="https://chat.lmsys.org/?leaderboard">sit atop yet another leaderboard</a> <a href="https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard">respectively</a> for models in their class.</p><p>Google has since run this kind of multi-track approach to AI, one with its own Gemini-series foundation models with the expectation you will run up a million tokens per prompt and a platform with smaller models (including its own Gemma series models) available with Google Vertex. But while it&#8217;s been pretty easy to dunk on Google for not having a real AI strategy as its AI search products seem to have not gone so well, there might be some more subtle signals that Google might not be as floundering as it seems on the surface.</p><p>While OpenAI invests heavily in creating a suite of foundational technology across a wide variety of modalities, it&#8217;s relying on powerful models that require powerful hardware and doesn&#8217;t work with the open source community. Meta, meanwhile, has invested heavily in training a wide variety of open-ish source models. Meta can deploy those open-ish source models its products including the best learning from other users, but carry potential risks of everyone having access to the same base technology it has. And in both situations, they&#8217;re still largely reliant on hardware provided from Nvidia.</p><p>Google has found a way to do both, and has the potential to do so without heavy reliance on external hardware providers. Having a powerful foundational model product effectively gives it a SaaS business to monetize all that work around training and development for current and potential future use cases. (A Google spokesperson noted that while they did use TPUs for internal workloads, they also run internal workloads on GPUs.)</p><p>And ingratiating itself in the developer community gives it a direct line to the rapid experimentation happening there, as well as an opportunity to nudge the development arc of AI in directions that could potentially benefit Google. It&#8217;s much in the same way Meta was able to benefit from the deep learning community largely adopting PyTorch, taking some of the best learnings and applying them to Meta products. </p><p>Google has quietly assembled a kind of comprehensive stack that&#8212;<a href="https://cloud.google.com/nvidia?e=48754805&amp;hl=en">while it obviously has invested in making Nvidia hardware available</a>&#8212;could give it a level of autonomy that many of the other AI developers and providers don&#8217;t necessarily have. At a time where everyone is reliant on Nvidia, the years of development on this immense stack on Google Cloud are paying out, even if it has yet to convert all this into a compelling consumer product. In its most recent operating quarter, Google <a href="https://abc.xyz/assets/19/e4/3dc1d4d6439c81206370167db1bd/2024q2-alphabet-earnings-release.pdf">said its Cloud business for the first time passed $10 billion in quarterly revenue and $1 billion in operating profit</a>. </p><p>That certainly includes hardware with its TPUs, but Google also owns a software stack that companies have been quietly picked up by others, with both xAI and Apple disclosing they used JAX for their <a href="https://x.ai/blog/grok-1.5">development of Grok</a> and <a href="https://machinelearning.apple.com/research/introducing-apple-foundation-models">Apple Intelligence models respectively</a>.</p><p>In that memo, the Google researcher flagged a handful of signals that could threaten Google&#8217;s (and OpenAI&#8217;s) dominance in AI from the open source community, and by extension companies that would use open source technology. But Google has essentially covered the majority of them while also maintaining development and release of a larger, more powerful product at the same time that competes more directly with Anthropic and OpenAI. </p><p>When ChatGPT launched in November 2022, Google was essentially caught flat-footed and it had to scramble to figure <em>something</em> out. We ended up getting a series of poorly-received products like its <a href="https://www.nytimes.com/2024/05/24/technology/google-ai-overview-search.html">attempts at an AI-powered search engine</a>. But setting aside that consumer strategy, Google has clearly created this kind of top-to-bottom developer experience that it can use either internally or serve externally.</p><p>Google was quietly<a href="https://www.businessinsider.com/facebook-pytorch-beat-google-tensorflow-jax-meta-ai-2022-6"> throwing its resources behind JAX for internal use cases several years ago</a>, even before language models became a big focus in modern AI. And while PyTorch still remains a preferred deep learning framework (part of that thanks to Google&#8217;s own histopry with TensorFlow) JAX has seemingly started to make its way into larger organizations. (And let&#8217;s not forget <a href="https://research.google/blog/open-sourcing-bert-state-of-the-art-pre-training-for-natural-language-processing/">Google created practically the original major open source language model with BERT</a>.)</p><p>And while search continues to be Google&#8217;s primary cash engine&#8212;which (in extreme emphasis) <em>could</em> end up disrupted by AI&#8212;its products are all intricately linked to its own developer stack. That was true with its research breakthroughs with MapReduce and TensorFlow, and it seems like it will continue with its development of AI-based language models.</p><h2>What was called correctly in that memo, and what Google has done</h2><p>At the time the original memo leaked, Google had <a href="https://blog.google/technology/ai/bard-google-ai-search-updates/">launched Bard just a few months earlier</a>, and would launch its PaLM 2 model shortly after that while also teasing the existence of Gemini. The former was, well, not great, and the latter demonstrated an emphasis on larger models&#8212;and chasing OpenAI. (And, to be fair, <a href="https://blogs.microsoft.com/blog/2023/02/07/reinventing-search-with-a-new-ai-powered-microsoft-bing-and-edge-your-copilot-for-the-web/">Bing&#8217;s search was also widely panned</a>.)</p><p>The whole sequence of events was comical enough that people were essentially wondering if Google was facing an existential threat as the hype around AI absolutely exploded. That hype led to colossal funding rounds in a lot of emerging companies&#8212;some of which <a href="https://www.reuters.com/technology/microsoft-agreed-pay-inflection-650-mln-while-hiring-its-staff-information-2024-03-21/">have been effectively acquihired</a> (<a href="https://techcrunch.com/2024/08/02/character-ai-ceo-noam-shazeer-returns-to-google/">and acquilicensed, I guess</a>). And Google&#8217;s emphasis on larger models was one of a number of potential problems it was dealing with in the memo, among others, as Meta&#8217;s original and much smaller Llama model captivated the developer ecosystem.</p><p>In rough, broad strokes, this was what the researcher covered back when it leaked in May last year&#8212;much of it which turned out to be directionally where the industry would end up heading leading up to Apple&#8217;s detailing of its own on-device AI suite, Apple intelligence:</p><ul><li><p><strong>On-device model inference</strong>: While on-device turned out to be an extremely popular outcome, this can essentially be abstracted out to low-power inference&#8212;like on a CPU&#8212;which is predicated on quantized versions of smaller models. Google ended up moving into this with Gemini Nano.</p></li><li><p><strong>Fine-tuning for personalization</strong>: Low-rank adaptation, at the time one of a handful of experimental techniques the open source community was running off with, turned out to be a preferred approach for customization&#8212;<a href="https://machinelearning.apple.com/research/introducing-apple-foundation-models">including being deployed by Apple</a>.</p></li><li><p><strong>The quality gap for large foundation models was closing</strong>: If the leaderboardification of AI has taught us anything, it&#8217;s that all these larger models developed by Google, Anthropic, OpenAI, and others are <em>very</em> close to each other in quality and the one-upsmanship doesn&#8217;t look like step function improvements.</p></li><li><p><strong>Focusing on larger foundation models was slowing Google down</strong>: At the time, the hype was around building a ChatGPT competitor that could do a whole lot of stuff all at once. But Google essentially showed it could do a bit of everything and do it pretty well, though it&#8217;s not clear exactly what the kind of payout on the other end will be for its Gemma series models.</p></li><li><p><strong>People might not pay for a restricted model behind an API if free open source models are available</strong>: We&#8217;ll get to this one in a second, but did not end up entirely correct&#8212;though that could still change in the future. But we still did, indeed, see a rapid race to the bottom on price.</p></li></ul><p>At the time, companies were hoovering up as much Nvidia hardware as they possibly could, which led to an overall shortage for GPUs. Meanwhile we started to see projects like Llama.cpp (and eventually Ollama) start to demonstrate the capabilities of smaller versions of models through a process called quantization that were able to run on edge devices like laptops.</p>
      <p>
          <a href="https://www.supervised.news/p/revisiting-that-old-google-ai-memo">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The search for the next class of AI founders]]></title><description><![CDATA[Venture firms are trying to get creative to get in front of the next hottest AI startups. But they have to look in new areas to find them.]]></description><link>https://www.supervised.news/p/the-search-for-the-next-class-of</link><guid isPermaLink="false">https://www.supervised.news/p/the-search-for-the-next-class-of</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Wed, 26 Jun 2024 22:47:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!40zN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!40zN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!40zN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 424w, https://substackcdn.com/image/fetch/$s_!40zN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 848w, https://substackcdn.com/image/fetch/$s_!40zN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 1272w, https://substackcdn.com/image/fetch/$s_!40zN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!40zN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png" width="1456" height="1095" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1095,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5816000,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!40zN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 424w, https://substackcdn.com/image/fetch/$s_!40zN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 848w, https://substackcdn.com/image/fetch/$s_!40zN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 1272w, https://substackcdn.com/image/fetch/$s_!40zN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b193cf7-86e9-4b0d-b730-3631ef01b0f9_2374x1786.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a happy scientist in a lab coat leaning onto a desk tinkering with a robot, the desk is very messy, the scientist's face is close to the robot as he turns a screwdriver on it, there is a school diploma and a graduation cap on the desk as well as a stack of papers that say "OFFER" on it, the robot is small and looks friendly, 45 degree angle camera sunda&#8230;</figcaption></figure></div>
      <p>
          <a href="https://www.supervised.news/p/the-search-for-the-next-class-of">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[For Apple, it's RPA all the way down]]></title><description><![CDATA[Plus: the future of an AI data marketplace continues to potentially lie within Snowflake and Databricks&#8212;if they can convince customers of the value.]]></description><link>https://www.supervised.news/p/for-apple-its-rpa-all-the-way-down</link><guid isPermaLink="false">https://www.supervised.news/p/for-apple-its-rpa-all-the-way-down</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 14 Jun 2024 22:29:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4Rbc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4Rbc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4Rbc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 424w, https://substackcdn.com/image/fetch/$s_!4Rbc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 848w, https://substackcdn.com/image/fetch/$s_!4Rbc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 1272w, https://substackcdn.com/image/fetch/$s_!4Rbc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4Rbc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png" width="1456" height="1096" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1096,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3205387,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4Rbc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 424w, https://substackcdn.com/image/fetch/$s_!4Rbc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 848w, https://substackcdn.com/image/fetch/$s_!4Rbc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 1272w, https://substackcdn.com/image/fetch/$s_!4Rbc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbf4215e7-4e42-49d4-95f5-d3fc35bae6d3_1528x1150.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a small and very advanced, smiling robot sittin in a large sandbox in a playground, there are lots of spherical objects in the sand, several emojis are also lying in the sandbox, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>Good morning and happy Friday everyone! Today we&#8217;re covering a handful of topics including Apple&#8217;s relatively unremarkable A&#8230;</em></p>
      <p>
          <a href="https://www.supervised.news/p/for-apple-its-rpa-all-the-way-down">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Lakes, catalogs, and Snowflake's full court press on AI]]></title><description><![CDATA[Snowflake is converging with its rival, Databricks, faster than ever before. And its new CEO is gearing up for a partial "pivot" into AI.]]></description><link>https://www.supervised.news/p/lakes-catalogs-and-snowflakes-full</link><guid isPermaLink="false">https://www.supervised.news/p/lakes-catalogs-and-snowflakes-full</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Tue, 04 Jun 2024 20:49:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8xZx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8xZx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8xZx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 424w, https://substackcdn.com/image/fetch/$s_!8xZx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 848w, https://substackcdn.com/image/fetch/$s_!8xZx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 1272w, https://substackcdn.com/image/fetch/$s_!8xZx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8xZx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png" width="1456" height="1085" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1085,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4227307,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8xZx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 424w, https://substackcdn.com/image/fetch/$s_!8xZx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 848w, https://substackcdn.com/image/fetch/$s_!8xZx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 1272w, https://substackcdn.com/image/fetch/$s_!8xZx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F764b3c84-8296-418b-b60b-16f68bdddd34_1844x1374.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">two sleek, friendly and smiling robots playing chess with each other, the setting is union square in new york city in a crowded environment, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p>For the majority of its life, Snowflake was synonymous with analytics through its tools as as a data warehouse. It had made some exploratory moves into machine learning as it saw a new rival creeping up, including the <a href="https://www.snowflake.com/blog/snowflake-to-acquire-streamlit/">abrupt $800 million acquisition of the Python developer platform Streamlit</a>, which at the time had practically negligible revenue. </p><p>In the year and a half-ish since the launch of ChatGPT, <em>a lot</em> has changed. <a href="https://www.supervised.news/p/snowflakes-new-leader-is-inheriting">Snowflake has a new CEO, Sridhar Ramaswamy</a>, <a href="https://www.supervised.news/p/what-neevas-quiet-exit-tells-us-about">through its $150 million acquisition of AI search startup Neeva</a>. It&#8217;s announced a series of plays in AI with the <a href="https://investors.snowflake.com/news/news-details/2023/Snowflake-Puts-Industry-Leading-Large-Language-and-AI-Models-in-the-Hands-of-All-Users-with-Snowflake-Cortex/default.aspx">launch of Snowflake Cortex last year</a>. And this week at its Snowflake summit, its announcement largely sat in AI&#8217;s orbit. It has a new catalog tool (Polaris), a new observability suite (Snowflake Trail), a notebook tool, and many others that provide a heavy emphasis on AI.</p><p>What was a database company under its past series of leaders is increasingly a company focusing its efforts on building the developer and operational frameworks for AI. It&#8217;s not exactly a pivot&#8212;Snowflake is still <em>the</em> data warehouse company&#8212;but it it&#8217;s not&#8230; <em>not</em> a pivot.</p><p>People tell me that under its new CEO, Sridhar Ramaswamy, all hands on deck for AI is an understatement at Snowflake. A major sense of urgency that comes down from Ramaswamy&#8212;particularly the risk of missing the AI wave&#8212;is powering a lot of the extreme &#8220;build or buy&#8221; mentality in AI.</p><p>&#8220;[Snowflake&#8217;s new leadership] very, very genuinely believes&#8212;and for that matter, correctly, in my opinion&#8212;that we are essentially in a once in a generation technological inflection,&#8221; Adrian Treuille, head of Snowflake&#8217;s developer platform Streamlit that it acquired for $800 million, told me. &#8220;Five years ago, most machine learning researchers thought we were decades away from talking to decades away from talking to computers&#8230;. And now it's like, ChatGPT, yeah, no a problem. It's amazing. And so I would say that the the thinking about the importance of AI is is genuinely first principles.&#8221;</p><p>That&#8217;s materializing in a lot of forms, this week with the launch of a series of Iceberg- and AI-focused products. But Snowflake is also clearly in dealmaker mode, <a href="https://www.bloomberg.com/news/articles/2024-05-22/snowflake-talks-to-acquire-reka-ai-said-to-fizzle-with-no-deal">most recently evaluating Reka AI for a potential $1 billion</a> per Bloomberg. A month prior to the deal petering out, Snowflake had launched its own large language model, Arctic, as well as a series of embeddings models. And while Databricks <a href="https://tabular.io/blog/tabular-is-joining-databricks/">was the ultimate buyer for Tabular, an Iceberg storage platform</a>, sources tell me Snowflake was aggressively in the mix for the 40-person startup going all the way back to April.</p><p>All this is pretty reasonable to expect for any company in the data abstraction layer among what are effectively considered among sources the three leaders: Snowflake, MongoDB, or Databricks. Enterprise data, which in many cases took <em>years</em> to get onto these platforms, is ripe for usage in modern AI. In fact, <em>most</em> of the value for AI is locked up in that data&#8212;either in the form of accessing it through retrieval augmented generation (RAG) or customizing the models directly with additional data.</p><p>While OpenAI&#8217;s fine-tuning APIs are incredibly straightforward to use, those same enterprises&#8212;especially ones with stricter privacy standards&#8212;aren&#8217;t likely to ship out their data to some API they don&#8217;t have direct control over. And that gives those data abstraction providers the ability to just add a customization or retrieval layer right on top of the data that enterprises have direct control over.</p><p>The difference here is that Snowflake, before the launch of ChatGPT and modern AI, was long considered a laggard in supporting machine learning. It didn&#8217;t support Python development until the launch of Snowpark in 2021. Databricks, meanwhile, had aggressively started in the opposite direction and then inched closer to Snowflake&#8217;s business over time.</p><p>But one of the biggest blockers to getting all these applications into production is having the suite of tooling around them&#8212;lineage, governance, development, observability, and to the extent possible, explainability. Snowflake, with its announcements this week, is bulldozing into all of those categories on its way to trying to be a backbone for AI development.</p><p>At its conference this week, Snowflake has effectively tried to brand itself as an &#8220;AI data cloud.&#8221; It&#8217;s a far cry from its summit in 2022 when one of its marquee announcements was Unistore, which was going after one of the holy grails of database technology in unifying analytical and transactional workloads.</p><p>Snowflake was long considered to be one of the key players in the abstraction layer, which has gone on to become a fundamental part of AI deployment. And with its launches it&#8217;s effectively growing into that role.</p><h2>Lakes, warehouses, and another small startup worth more than $1 billion</h2><p>Arguably the two most significant announcements from Snowflake this week was Polaris, a new open source catalog format built on Iceberg, and its suite of observability tools Snowflake Trail. Catalogs effectively serve as the governance layer of a data store, providing access and logging around data within an organization. And observability&#8212;basically the monitoring of performance of a given product&#8212;is emerging as a key requirement for many larger enterprises in order to graduate out of that proof-of-concept stage for AI.</p><p>Modern AI broadly is built around the premise of data lakes as a backbone for the abstraction layer. You take <em>all</em> unstructured data&#8212;PDFs, images, emails, transcripts, whatever&#8212;pre-process it into something more coherent and accessible, and shove it into a semi-accessible format. It&#8217;s also led to the birth of a whole chain of startups focused on piping that unstructured data into an unstructured data store like unstructured.io, Datavolo, Reducto, and others.</p><p>Snowflake and Databricks have also embraced competing formats for data lakes. While Databricks has its own format (that&#8217;s now open source) in Delta Lake, Snowflake effectively bet on the open source format Iceberg. Its announcements to date have all relied on Iceberg, which has proved to be an extremely popular file format.</p><p>While ongoing for several years now, Snowflake&#8217;s embrace of data lakes is somewhat of a break from its long singular focus on data warehouses. Founded in 2012, Snowflake pounced on the decline in the cost of cloud storage to build out what effectively remade the business intelligence and analytics layer and gave birth to a whole suite of tools to radically improve the quality of analytics. We generally refer to all these tools, including Dbt, Alation, Monte Carlo, and others, as the modern data stack.</p><p>As machine learning emerged as a much-larger-than-anticipated market with techniques and tools maturing on the cusp of the launch of ChatGPT, Snowflake pressed into the world of unstructured data&#8212;and data lakes. In late 2022, Snowflake <a href="https://www.snowflake.com/blog/iceberg-tables-powering-open-standards-with-snowflake-innovations/">launched Iceberg Tables</a>, embracing the open source format that came out of Netflix four years prior. (Databricks, at the time, was already seeing usage of large language models in the form of BERT.)</p><p>Machine learning tooling built on top of data lakes, though, had long been Databricks&#8217; sweet spot. Databricks announced the availability of Delta Lake in 2019, and in 2022 it <a href="https://www.databricks.com/blog/2022/06/30/open-sourcing-all-of-delta-lake.html">announced that its 2.0 version of Delta Lake would be open source</a>. And its 3.0 version, <a href="https://www.databricks.com/blog/announcing-delta-lake-30-new-universal-format-and-liquid-clustering?scid=7018Y000001Fi1mQAC&amp;utm_medium=paid+search&amp;utm_source=google&amp;utm_campaign=17115062609&amp;utm_adgroup=142991766384&amp;utm_content=ebook&amp;utm_offer=delta-lake-running-oreilly&amp;utm_ad=665911854841&amp;utm_term=delta%20lake&amp;gad_source=1&amp;gclid=Cj0KCQjw9vqyBhCKARIsAIIcLMF2ID1V6KIDKSxNrZ9-Fo4hZ_9KLqx6HwQe3s-GClN2SgVU3-IfvFYaAqiWEALw_wcB">launched at last year&#8217;s summit</a>, was also effectively a shot at a universal format for data lake management.</p><p>During all this time, Databricks has long tried to espouse the paradigm of a data lakehouse&#8212;a unified approach to data warehousing and lake management that allows for execution of all tasks, whether that&#8217;s business intelligence or machine learning. And Tabular, a storage platform built on Iceberg, was one of the best positioned tools to grow that paradigm.</p><p>The acquisition of Tabular, though, has been floated for <em>weeks</em>, going back to even April. The most recent number I had heard floated on the Snowflake side </p>
      <p>
          <a href="https://www.supervised.news/p/lakes-catalogs-and-snowflakes-full">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[A new whisper in the AI analytics room]]></title><description><![CDATA[Informatica, of all companies, is creeping back into conversations despite being a legacy vendor. Plus: what happens when RAG isn't enough?]]></description><link>https://www.supervised.news/p/a-new-whisper-in-the-ai-analytics</link><guid isPermaLink="false">https://www.supervised.news/p/a-new-whisper-in-the-ai-analytics</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 31 May 2024 21:56:45 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!KhzM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KhzM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KhzM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 424w, https://substackcdn.com/image/fetch/$s_!KhzM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 848w, https://substackcdn.com/image/fetch/$s_!KhzM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 1272w, https://substackcdn.com/image/fetch/$s_!KhzM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KhzM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png" width="1456" height="1105" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1105,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4035190,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KhzM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 424w, https://substackcdn.com/image/fetch/$s_!KhzM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 848w, https://substackcdn.com/image/fetch/$s_!KhzM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 1272w, https://substackcdn.com/image/fetch/$s_!KhzM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b66d84e-e523-4f51-88f4-3486d9cec443_1824x1384.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a number of small, sleek and happy robots are standing in a laboratory setting that looks like it is out of a Pixar movie, in the center of the lab is a large and much older steampunk robot staring down at the smaller sleek robots, the steampunk robot is waving jovially at the more sleek and advanced robots, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>We have both Snowflake and Databricks&#8217; annual summits coming in the next two weeks with a high volume of events going along with them. As part of that the next few issues will be compressed with multiple topics (like this one) to account for the time sink they are going to cause.</em></p><div><hr></div><h2>Way too many coding models and a Bring Your Own Key paradigm on the horizon</h2><p>One <em>extremely</em> popular emerging tool among developers that consistently ends up in the same echelon as <a href="https://ollama.com/">Ollama</a> (in terms of how fervent their fans are) is <a href="https://cursor.sh/">Cursor</a>, an AI-powered developer tool that&#8217;s essentially a replacement for VSCode. </p><p>There are a lot of reasons Cursor is popular as it <em>heavily</em> streamlines the coding experience with tools like, for example, making a codebase searchable. But there&#8217;s another substantial difference when compared to VSCode: its model selection is more open-ended. </p><p>And going beyond just a tool itself, Curor it also offers a window into a future workflow scenario for one of the most popular use cases for language models: bring your own key. Rather than relying on a GitHub Copilot integration and living in the Microsoft Extended Coding Universe, you could select the best model for your specific problem&#8212;or, maybe at some point, a powerful model run locally.</p><p>Cursor allows you to drop in both your OpenAI and Anthropic keys to swap between each during sessions. And, perhaps more importantly, you can override the OpenAI API to configure it for other endpoints, <a href="https://api.together.xyz/">like ones from Together AI</a>. That means developers can pick the best/cheapest/most directly helpful models for their development environment. And it offers an opportunity for other endpoint providers to sneak directly into development environments. </p><p>And while GPT-4o, Gemini 1.5, and other top APIs will likely capture the majority of code development usage, Cursor gives open-ish source model providers (<em>as in Meta and Mistral</em>) ways to sneak closer into developer environments. Even the coding model ecosystem continues to grow. Meta <a href="https://ai.meta.com/blog/code-llama-large-language-model-coding/">released CodeLlama last year</a>, and Mistral this week released its own coding model, <a href="https://mistral.ai/news/codestral/">Codestral</a>. </p><p>And all these endpoints are, of course, a race to the bottom that&#8217;s only accelerating. Let&#8217;s revisit the pricing chart for the <em>nth</em> time here!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gUQN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gUQN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 424w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 848w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1272w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gUQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp" width="478" height="741.2939560439561" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2258,&quot;width&quot;:1456,&quot;resizeWidth&quot;:478,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gUQN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 424w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 848w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1272w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The process within Cursor itself has a few hurdles to set up&#8212;you have to have access to an endpoint with an API key in the first place, so it&#8217;s not a &#8220;click it and get calls&#8221; product like GitHub Copilot (<a href="https://cursor.sh/pricing">though there is a pro option with Cursor tied to GPT-4</a>). It also requires managing a variety of different keys, which could involve jumping between developer accounts.</p><p>Right now, Anthropic and OpenAI those are two of the best APIs. But a number of others are quickly catching up based on the <a href="https://chat.lmsys.org/">LMSys chatbot arena</a>. That includes Llama 3 variations, but it also includes potentially specialized ones like Cohere&#8217;s Command-R+. </p><p>Cursor effectively offers <em>another</em> avenue for leveling the playing field here. It probably isn&#8217;t the <em>last</em> new development tool to exploit that kind of bring-your-own-key approach. </p><p>But it also feels like an early sign that another of OpenAI&#8217;s biggest advantages&#8212;its deep integration with GitHub Copilot&#8212;has some signs of vulnerability.</p><h2>Informatica gets a second look</h2><p>Earlier this year, <em>The Wall Street Journal</em> reported that Informatica, a legacy vendor for data management, <a href="https://www.wsj.com/tech/salesforce-in-advanced-talks-to-buy-informatica-ba9ec09c">was in advanced talks to be acquired by Salesforce.</a> The <em>Journal</em> reported <a href="https://www.wsj.com/business/deals/salesforces-talks-to-buy-informatica-fizzle-4f6f111a">a little more than a week later</a> that those talks had broken down after being unable to agree on terms.</p><p>While that was a very, uh, <em>loud</em> story for Informatica, the reality is the company has come up more and more lately in the context of developments in AI. Or, in particular, Informatica has found a way to tell a story around its potential AI services. At its conference this month it unveiled a suite of updates and integrations&#8212;<a href="https://finance.yahoo.com/news/informatica-unveils-blueprint-enterprise-generative-130000931.html">particularly with Snowflake&#8217;s AI platform, Cortex. </a></p><p>Informatica is one of <em>many</em> examples of what feels like another growing trend with the emergence of AI tooling for legacy vendors: a second shot at reclaiming ground lost to challenger startups and cloud-native vendors. In the case of Informatica, that&#8217;s the modern data stack, which has picked off nearly every part of its business into independent products.</p><div><hr></div><h6><strong>Sponsored</strong></h6><h4><strong>This week&#8217;s newsletter is brought to you by Felicis</strong></h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="http://felicis.com" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kU-G!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 424w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 848w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1272w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kU-G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png" width="1456" height="366" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:366,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:381012,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;http://felicis.com&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!kU-G!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 424w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 848w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1272w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Felicis has been investing in an impressive variety of <a href="https://www.felicis.com/sectors/ai">AI</a> and <a href="https://www.felicis.com/sectors/infra">infrastructure</a> companies, including: Runway, Weights &amp; Biases, MotherDuck, Supabase, Metaplane, Vannevar Labs, poolside, and Flower Labs. A generalist firm started by Google&#8217;s first product manager, Felicis is known for supporting founders with their <a href="https://www.felicis.com/founder-pledge">Founder&#8217;s Pledge</a>.</p><p><em><a href="https://www.felicis.com/founders-on-felicis">Read what founders say about working with Felicis.</a></em></p><div><hr></div>
      <p>
          <a href="https://www.supervised.news/p/a-new-whisper-in-the-ai-analytics">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[OpenAI's Silicon Valley pivot]]></title><description><![CDATA[What was a research organization is now following dreams of being the next Apple. Plus: Why is everyone suddenly talking about WebGPU again?]]></description><link>https://www.supervised.news/p/openais-silicon-valley-pivot</link><guid isPermaLink="false">https://www.supervised.news/p/openais-silicon-valley-pivot</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Wed, 22 May 2024 22:02:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!p_51!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!p_51!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!p_51!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 424w, https://substackcdn.com/image/fetch/$s_!p_51!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 848w, https://substackcdn.com/image/fetch/$s_!p_51!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 1272w, https://substackcdn.com/image/fetch/$s_!p_51!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!p_51!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp" width="1232" height="928" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:928,&quot;width&quot;:1232,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:165290,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!p_51!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 424w, https://substackcdn.com/image/fetch/$s_!p_51!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 848w, https://substackcdn.com/image/fetch/$s_!p_51!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 1272w, https://substackcdn.com/image/fetch/$s_!p_51!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0308a893-995c-4d58-9591-472e4ad874b8_1232x928.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a startup bro in a wearing a black turtleneck, jeans, and tennis shoes presenting on stage like an Apple keynote with a remote in its hand and a laptop on a pedestal. On the very large screen on stage is a small logo that is spiraled and black and white, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>Since we have a holiday weekend coming up and I&#8217;m still on the other end of a heavy travel week (as well as getting ready for a heavy conference month in June), two columns this week are kind of smashed into one with some additional notes on WebGPU and WebLLM. Enjoy the long weekend!</em></p><div><hr></div><p>In January 2007, Silicon Valley luminary Steve Jobs took the stage to explain a device that was &#8220;a phone, an iPod, and an internet communicator.&#8221; He followed it, to a roaring crowd, by saying, &#8220;are you getting it yet?&#8221; </p><p>Seventeen years later, a handful of employees sat in a room in Silicon Valley speaking with a cheerful synthetic assistant that sounded close enough to Scarlet Johansson&#8217;s voice in the film <em>Her</em> that it might as well have been Johansson. And it was lying in the same device: an iPod, a phone, and an internet communicator. </p><p>But this time around there&#8217;s nothing to &#8220;get,&#8221; as the most iconic Silicon Valley companies have already made casually pulling products out of science fiction&#8212;self-driving cars, buying anything with a press of a button, speaking to friends across the world on demand, and access to nearly the entire corpus of human knowledge&#8212;into reality almost a rite of passage. </p><p>It seems increasingly obvious OpenAI, and its leader Sam Altman, want to become part of the enduring Silicon Valley lore with a truly Silicon Valley product: a feat of engineering that&#8217;s, quite literally, straight out of a movie. And in classic Silicon Valley fashion, there&#8217;s simply the expectation of a number of variable sized bumps along the way. Or, as we used to say, &#8220;move fast and break things.&#8221; </p><p>All this, though, is more or less the logical conclusion of the last six months of OpenAI&#8217;s operations. Altman, <a href="https://www.supervised.news/p/with-sam-altman-out-where-does-openai">ousted by the board in November last year</a>, engineered a return that presaged a pivot from a research firm into an achieve-the-impossible consumer technology company.</p><p>OpenAI has effectively shed its moniker as a research institution in recent weeks not only through its unveiling of that synthetic assistant, but by seemingly moving to an ask-for-forgiveness-not-permission approach to development. Its <a href="https://x.com/ilyasut/status/1790517455628198322">chief scientist Ilya Sutskever left</a>, and Wired reported that<a href="https://www.wired.com/story/openai-superalignment-team-disbanded/"> its Superalignment team has disbanded</a> as well. And this week it is dealing with the <a href="https://www.nbcnews.com/tech/tech-news/scarlett-johansson-shocked-angered-openai-voice-rcna153180">fallout of that same giggling digital persona sounding way too close to Scarlet Johansson</a>, who declined to participate in developing its &#8220;Sky&#8221; voice, <a href="https://www.nbcnews.com/tech/tech-news/scarlett-johansson-shocked-angered-openai-voice-rcna153180">per a statement from NBC</a>.</p><p>OpenAI is now creating one of the kinds of culture-altering company Silicon Valley entrepreneurs dream to achieve&#8212;through whatever means are needed to do so. It&#8217;s why so many people come to San Francisco in the first place, and with Altman at the helm, OpenAI is becoming a quintessential product-led Silicon Valley company.</p><p>Altman has always had very deep ties to Silicon Valley, the San Francisco Bay Area, and San Francisco itself well into the early mobile tech boom. His growing influence in Silicon Valley, before taking over OpenAI, culminated in becoming the president of the industry&#8217;s most prestigious accelerator Y Combinator. You can really <a href="https://www.supervised.news/p/is-openai-about-to-replace-a-bunch">look no further than the large expat contingent of Stripe</a>&#8212;one of Y Combinator&#8217;s most successful companies&#8212;within OpenAI. Airbnb CEO <a href="https://www.theinformation.com/articles/airbnb-ceo-chesky-emerges-as-altman-ally-adviser?rc=u5devd">Brian Chesky&#8217;s, another of Y Combinator&#8217;s greatest success stories, has emerged as one of Altman&#8217;s strongest allies</a>, which you can see in much more detail in an <a href="https://www.theinformation.com/articles/airbnb-ceo-chesky-emerges-as-altman-ally-adviser?rc=u5devd">in-depth report in The Information</a>.</p><p>Altman quipped that its new voice assistant was straight out of &#8220;<em>Her</em>,&#8221; a film that received a smorgasbord of Academy Award nominations&#8212;including winning the best original screenplay in a year where it was competing with films like <em>American Hustle, Twelve Years a Slave</em> (the Best Picture winner that year), and the much-nearer-future <em>Gravity. Her</em>, in addition to being a great film, offered a glimpse at a future with truly humanlike, empathetic synthetic assistants. </p><p>Ever since the film came out, Silicon Valley has desperately tried to make that a reality. But speech technology like Alexa, Cortana, and Siri have long continued to be a dumpster fire and the pathway to that product seemed to just not exist. It was <em>magic</em> in a future that was so far away we couldn&#8217;t even see a timeline away from the limited use cases something like Google Assistant could offer.</p><p>OpenAI&#8217;s unveiling of its latest advanced model, GPT-4o, might as well have been a footnote. Whether or not its voice assistant ends up becoming the ambient companion tech has long dreamed of, OpenAI has also shown that it&#8217;s no longer just a research firm creating an API that can power the next-generation of applications. </p><p>Instead, OpenAI has pivoted to a company chasing the culture-altering step-change in technology that entrepreneurs have long dreamed about. And it has done so not once, but if you include DALL-E 2, potentially <em>three times</em>&#8212;a feat rarely achieved by even the most successful companies in Silicon Valley.</p><h2>Move fast and break things</h2><p>Regardless of what&#8217;s happened in recent weeks, OpenAI&#8217;s voice assistant unveiling was a triumph. The company, hardly a startup any more, needed to create some kind of step-change cultural moment in AI when the vibes were clearly of the &#8220;AI is hitting a wall&#8221; kind. OpenAI also wasn&#8217;t doing itself any favors by running simultaneously one of the most expensive and less performant APIs (on standard benchmarks) for light-to-medium-difficulty problems in GPT-3.5 Turbo.</p><p>The &#8220;vibe&#8221; of its demo was more akin to the release of the iPhone, iPad, or the <a href="https://developers.google.com/search/blog/2010/09/google-instant-impact-on-search-queries">original release of Google&#8217;s we-know-what-you&#8217;re-looking-for Instant Search product</a>.</p><p>OpenAI, in its last few weeks, effectively looks like a company run by a classic Silicon Valley executive. At the time of his ouster, it represented an interesting potential turning point for the industry that is largely flooded with academics. It&#8217;s no surprise that Microsoft, OpenAI&#8217;s largest partner, <a href="https://www.wsj.com/tech/ai/altman-firing-openai-520a3a8c">would work as hard as it could to keep an executive like Altman in place</a>.</p><p>In April 2009, Y Combinator founder Paul Graham <a href="https://paulgraham.com/5founders.html">penned an essay</a> about five of the most interesting founders in the last 30 years that at the time included Altman alongside Silicon Valley legends Steve Jobs, Larry Page, and Sergey Brin. Here&#8217;s part of what he said in the essay:</p><blockquote><p>What I learned from meeting Sama is that the doctrine of the elect applies to startups. It applies way less than most people think: startup investing does not consist of trying to pick winners the way you might in a horse race. But there are a few people with such force of will that they're going to get whatever they want.</p></blockquote><p>That &#8220;force of will&#8221; is the <em>north star</em> of entrepreneurship in Silicon Valley. You don&#8217;t win on ideas, you win on execution. And the founders listed alongside Altman, and if we wanted to include Mark Zuckerberg as well, routinely skirted with the absolute boundary of what was acceptable at the time&#8212;often crossing it just barely before finding a way to avoid the harshest penalties.</p><ul><li><p>In his mission to build an iconic phone, the iPhone 4 brought about the infamous &#8220;<a href="https://www.wsj.com/articles/SB10001424052748704913304575371131458273498">antennagate</a>&#8221; scandal, wherein if you held the phone at the right spots the signal would die out. Apple&#8217;s mea culpa here was, well, free iPhone 4 cases for everyone.</p></li><li><p>While Facebook was routinely the subject of privacy imbroglios, it&#8217;s easy to forget that Meta was <a href="https://techcrunch.com/2012/08/10/facebook-ftc-settlement-12/">under a </a><em><a href="https://techcrunch.com/2012/08/10/facebook-ftc-settlement-12/">twenty-year consent decree</a></em><a href="https://techcrunch.com/2012/08/10/facebook-ftc-settlement-12/"> that the FTC hit it with back in 2012</a>. The FTC found in 2019 that it <a href="https://www.reuters.com/article/idUSKCN1UJ1L9/">had violated that consent decree and slapped them with a $5 billion fine</a>.</p></li><li><p>Meta <em>also</em> <a href="https://www.theatlantic.com/technology/archive/2014/06/everything-we-know-about-facebooks-secret-mood-manipulation-experiment/373648/">came under fire for an experiment that intentionally seeded Facebook&#8217;s News Feed</a> with content to, essentially, see if it altered the mood of its users.</p></li><li><p>Google <em>also</em> <a href="https://techcrunch.com/2012/08/09/google-settles-with-ftc-agrees-to-pay-22-5m-penalty-for-bypassing-safari-privacy-settings/">paid the FTC a fine in 2012 for bypassing Safari&#8217;s privacy settings</a> (for a comically low $22.5 million). Google also paid a $170 million fine <a href="https://www.ftc.gov/news-events/news/press-releases/2019/09/google-youtube-will-pay-record-170-million-alleged-violations-childrens-privacy-law">for violating another privacy law</a> in 2019.</p></li></ul><p>You could walk through the history of the majority of all these iconic tech companies and find these kinds of skirmishes with the norm. That behavior and method of operating distilled down to an early motto at Facebook: move fast and break things. And in AI, one of the strongest &#8220;norms&#8221; is a heavy focus on safety for a technology with <a href="https://www.bloomberg.com/news/newsletters/2024-02-07/how-investigators-solved-the-biden-deepfake-robocall-mystery">an extreme potential for causing chaos</a>, even if the results are a little mixed.</p><p>Altman, born and raised in the classic Silicon Valley mold, is essentially just continuing a long tradition of testing the boundaries as hard as physically possible. You could point to any one of its ongoing engagement with <a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The New York Times over copyright issues</a>, sidelining its Superalignment team, or effectively remaking an iconic sci-fi character <a href="https://www.reuters.com/technology/scarlett-johansson-says-openai-chatbot-voice-eerily-similar-hers-2024-05-21/">without direct involvement with the actress behind her (it?)</a>. </p><p>It&#8217;s not the only voice-cloning scandal we&#8217;ve had, though it is certainly the highest profile as it was effectively part of the brand behind its voice assistant (with Altman literally tweeting the reference to Her). The stakes were, and are, high enough and not just limited to Johansson that it made it into the SAG-AFTRA collective bargaining discussions in the industry.</p><p>And Y Combinator, practically an institution in Silicon Valley, is built to attract founders and teams that are ready to pull out all stops to build a successful startup. With one of its former leaders at the helm, we could only expect these myriad disputes to be just an early signal of what&#8217;s to come as OpenAI looks to remake cultural norms in both tech and beyond.</p><h2>The fine line between enterprise, consumer, and icon</h2><p>But despite the release of ChatGPT, which created one of those cultural <em>moments</em> in tech, OpenAI was largely what you&#8217;d consider an &#8220;enterprise&#8221; business not unlike Stripe. It offered a suite of APIs that would potentially power a next generation of technology products, though for the time being that technology is still looking for its killer use case. But most of the experimentation here is, well, enterprise use cases (<a href="https://www.supervised.news/p/ai-in-april-and-q2-rpa-in-focus-holistic">such as RPA</a>).</p><p>The release of Llama 3, which in many standard benchmarks was highly competitive or more effective than GPT-3.5 Turbo, exposed vulnerabilities to OpenAI&#8217;s core enterprise business. <a href="https://www.supervised.news/p/the-point-of-endpoints">A flurry of startups threw up endpoints</a> that could effectively replace GPT-3.5 Turbo with a few lines of code. OpenAI did what it could do&#8212;<a href="https://www.supervised.news/p/openai-sort-of-covers-one-of-its">with a price cut to GPT-3.5 Turbo and the release of a batch processing API</a>&#8212;but its first-mover and convenience factor advantage was starting to quickly degrade.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gUQN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gUQN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 424w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 848w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1272w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gUQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp" width="576" height="893.2747252747253" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2258,&quot;width&quot;:1456,&quot;resizeWidth&quot;:576,&quot;bytes&quot;:77038,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!gUQN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 424w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 848w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1272w, https://substackcdn.com/image/fetch/$s_!gUQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa52ce17e-f041-4b93-b895-c8fd43f889ef_1456x2258.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Rather than getting a GPT-4 &#8220;Lite&#8221; model to address that problem, we got a semi-equivalent multi-modal model to GPT-4 that is half the price of GPT-4 Turbo. And while that&#8217;s an enormous price cut compared to GPT-4 Turbo (and GPT-4 Classic), it still doesn&#8217;t tackle the workhorse model problem OpenAI now faces. Instead, GPT-4o is more in line with competing with Claude Sonnet and Opus, Cohere&#8217;s Command-R+ (which developers seem to <em>love</em>), and Mistral&#8217;s latest higher-tier mixture-of-experts model and its mistral-large model.</p><p>When we step back and look at the pricing in that light&#8212;and what OpenAI is trying to do here&#8212;it starts to make a little more sense. GPT-4o&#8217;s goals are more aligned with what Gemini 1.5 Pro, Mistral&#8217;s large model, Claude Opus, and to a certain extent what Command-R+ are trying to achieve: a premium tier. Here&#8217;s how the pricing breaks down. (I&#8217;m not going to include GPT-4 Turbo, GPT-4 Classic, or Claude Opus, as they basically break the chart and slightly miss the point.)</p>
      <p>
          <a href="https://www.supervised.news/p/openais-silicon-valley-pivot">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[AI in April (and Q2): RPA in focus, holistic evaluations, and eyes back on Datadog]]></title><description><![CDATA[Plus: OpenAI and Google are doing some stuff next week.]]></description><link>https://www.supervised.news/p/ai-in-april-and-q2-rpa-in-focus-holistic</link><guid isPermaLink="false">https://www.supervised.news/p/ai-in-april-and-q2-rpa-in-focus-holistic</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 10 May 2024 22:54:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TPON!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TPON!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TPON!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 424w, https://substackcdn.com/image/fetch/$s_!TPON!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 848w, https://substackcdn.com/image/fetch/$s_!TPON!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 1272w, https://substackcdn.com/image/fetch/$s_!TPON!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TPON!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png" width="1232" height="928" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b00a4583-306b-4f59-8016-92f092190eab_1232x928.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:928,&quot;width&quot;:1232,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1475355,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TPON!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 424w, https://substackcdn.com/image/fetch/$s_!TPON!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 848w, https://substackcdn.com/image/fetch/$s_!TPON!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 1272w, https://substackcdn.com/image/fetch/$s_!TPON!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb00a4583-306b-4f59-8016-92f092190eab_1232x928.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a friendly-looking pixar-styled robot standing in front of a table that is covered in ants, the ants look like robots and are all carrying leaves, there are NO cables coming out of the robot, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>This issue is going to be a bit compressed and cover multiple topics as I&#8217;m recovering from a <a href="https://twitter.com/mattlynley/status/1787906360824496544">hand injury</a> and a minor illness. I apologize in advance for any typos and whatnot.</em></p><div><hr></div><p>Next week Google and OpenAI are <a href="https://twitter.com/openai/status/1788987793613725786">both going to be making announcements</a> <a href="https://developers.googleblog.com/en/get-ready-for-google-io-program-lineup-revealed/">on their AI products</a>, with OpenAI sliding just ahead of Google with a livestream on Monday.</p><p>Sam Altman on shut down whether this is an announcement around a Perplexity competitor or GPT-5 <a href="https://twitter.com/sama/status/1788989777452408943">on Twitter</a>. But OpenAI says the announcements will be updates for ChatGPT and GPT-4 <em>right</em> before Google&#8217;s developer conference kicks off. And we&#8217;ll probably get <em><a href="https://www.supervised.news/p/is-openai-about-to-replace-a-bunch">another</a></em><a href="https://www.supervised.news/p/is-openai-about-to-replace-a-bunch"> whole new list of startups OpenAI and Google might-but-might-not-but-possibly smush in the process</a>.</p><p>OpenAI obviously has to contend with Google and its continued improvements, as well as Meta&#8217;s largest Llama 3 (400B+) that&#8217;s still in the works. <em>But</em> maybe we&#8217;ll finally see a response to all the pressure OpenAI has faced on its workhorse GPT-3.5 Turbo model, particularly with the release of the Llama 3 models. Llama 3 70B is both <a href="https://www.supervised.news/p/vibes-based-evals-and-weights-and">cheaper and more performant than GPT-3.5 Turbo</a> on standard benchmarks. Cost and simplicity has long been one of OpenAI&#8217;s advantages, but that&#8217;s <a href="https://www.supervised.news/p/a-new-rival-startup-is-emerging-for\">slipping away as more startups offer drop-in replacements for GPT-3.5 Turbo with an API powered by Llama 3</a>.</p><p>What that would look like is unclear, but it wouldn&#8217;t be all that surprising if OpenAI came out with some kind of lightweight GPT-4 or upgraded workhorse model alongside whatever else it has planned. (A hub called the <a href="https://chat.lmsys.org/">LMSys chatbot</a> arena, where users compare results from competing models and rate the better one, <a href="https://twitter.com/sama/status/1787222050589028528">has been exploding with speculation for a series of cheekily-named models</a> <a href="https://twitter.com/sama/status/1785107943664566556">that seemingly suggest they are from OpenAI</a>. )</p><p>Now, with that speculation out of the way, let&#8217;s get to the actually fun stuff for the quarter so far: what everyone&#8217;s arguing about in AI lately. (I&#8217;ll be recapping conversations here from April up til today, even though this is <em>technically</em> supposed to be focused on April.)</p><p>The whole <a href="https://www.supervised.news/p/lets-check-back-in-on-the-vector">retrieval augmented generation (or RAG) pipeline is coalescing into a rather sensible (if not sprawling) chain of startups</a> and tools. And, as usual, language models are getting closer to production&#8212;though the timelines still range from a few quarters to more than a year, depending on who you talk to.</p><p>From everyone I&#8217;ve spoken with lately, the key word this month (and the quarter) seems to be <em>practicality</em>. Not in the sense of whether a product built on a language model can or can&#8217;t be done&#8212;but whether it should be done in the first place. Experiments are giving way to analysis around cost, evaluation, and the simplest places to start to extract value.</p><p>So, with that said, here&#8217;s what everyone is talking about as we head further into the second quarter of the year:</p><ul><li><p><strong>RPA gets a closer look</strong>. Robotic process automation&#8212;basically that work around automatically clicking around on a website&#8212;is increasingly seen as a spot ripe for disruption by language models. In particular, RPA is seen as an initial stepping stone toward a much broader automation network.</p></li><li><p><strong>Long context versus RAG</strong>. There&#8217;s a universal agreement that jamming more relevant information into a prompt yields better results. But now there&#8217;s this emerging <em>debate-ish</em> over whether gargantuan context windows or retrieval augmented generation is the way to go.</p></li><li><p><strong>Evaluations beyond just the benchmarks</strong>. While a lot of these larger models try to one-up each other on performance benchmarks, a lot of companies are now facing a struggle of how to evaluate <em>whether one did the thing they wanted to do</em>. And it turns out that&#8217;s as much of an aesthetic review than it is for basic performance metrics.</p></li><li><p><strong>Eyes are back on Datadog for its next move in AI.</strong> Whispers are starting to trickle in around Datadog&#8217;s intent to enter AI pipelines more formally this year. Like MongoDB with vector search, investors and users are waiting to see what Datadog&#8212;which is already deeply entrenched in enterprises&#8212;does in AI observability and evaluation.</p></li></ul><p>Let&#8217;s dive into each, starting with RPA!</p><h2>RPA gets a closer look as a primitive agent</h2><p>One topic that routinely comes up is whether investing in the infrastructure layer&#8212;the startups like <a href="https://ollama.com/">Ollama</a>, <a href="https://www.langchain.com/">LangChain</a>, <a href="https://www.trychroma.com/">Chroma</a>, or <a href="https://elevenlabs.io/">ElevenLabs</a>&#8212;is saturated and the focus should be going to the app layer on top of all that infrastructure. There are a lot of natural use cases for AI that everyone tends to point to, like talking to legal documents, code generation, customer service chatbots, and so on and so forth.</p><p>Increasingly, though, there&#8217;s another broader focal point emerging on that app layer from experts and investors: robotic process automation, or RPA. And while these automations&#8212;a fancy way of saying a bot acting like a human in some form to complete a simple task&#8212;are currently more discrete, we already have a term for what would come next with substantially more advanced self-orchestrated RPA: agents.</p><p>RPA is seen by a lot of companies and investors as a kind of low-hanging fruit that they can quickly capitalize on, either through the use of smaller models or just calling one of the workhorse endpoints like GPT-3.5 Turbo. Most companies <a href="https://www.klarna.com/international/press/klarna-ai-assistant-handles-two-thirds-of-customer-service-chats-in-its-first-month/">reflexively point to the case study put out from Klarna</a> as what can be accomplished with a language model under the hood just routing requests to specific products, FAQs, or support pages automatically.</p><p>Modern RPA has largely been relegated to manual tasks bots can accomplish autonomously, such as, well, clicking around on a website to figure out if stuff works or how to break an app. The challenge with all that &#8220;clicking around&#8221; is minor changes in workflows and products can eject said bot out of a designated pathway and render it useless (or outright counterproductive).</p><p>This is where language models can potentially plug in here. The extreme version is that all app interactions are through natural language, which you would&#8230; automate with a language model. But in the meantime, language models can potentially filter through customer support tickets for escalation and de-escalation, or fiddle around with websites to find ways to break them.</p><p>UiPath is largely the company with a target on its back here. UiPath <em>did</em> <a href="https://www.uipath.com/newsroom/uipath-unveils-new-family-of-llms-at-ai-summit">announce a family of language models in March this year</a>. But founded in 2005, UiPath effectively created a business that&#8217;s valued more than $10 billion. Meanwhile, Microsof, purveyor of Copilot buttons, has&#8212;shocker&#8212;<a href="https://learn.microsoft.com/en-us/power-platform/release-plan/2023wave1/power-automate/copilot-power-automate">a copilot for its RPA product Power Automate</a>.</p>
      <p>
          <a href="https://www.supervised.news/p/ai-in-april-and-q2-rpa-in-focus-holistic">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Where graph databases live in a future AI data stack]]></title><description><![CDATA[Another once specialized database technology from the previous big data era might find its way into AI products thanks to retrieval augmented generation.]]></description><link>https://www.supervised.news/p/where-graph-databases-live-in-a-future</link><guid isPermaLink="false">https://www.supervised.news/p/where-graph-databases-live-in-a-future</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Wed, 01 May 2024 23:09:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QyWe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QyWe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QyWe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 424w, https://substackcdn.com/image/fetch/$s_!QyWe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 848w, https://substackcdn.com/image/fetch/$s_!QyWe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 1272w, https://substackcdn.com/image/fetch/$s_!QyWe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QyWe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp" width="1232" height="928" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:928,&quot;width&quot;:1232,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:308568,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QyWe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 424w, https://substackcdn.com/image/fetch/$s_!QyWe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 848w, https://substackcdn.com/image/fetch/$s_!QyWe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 1272w, https://substackcdn.com/image/fetch/$s_!QyWe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1b9acd20-4c2f-4c66-89aa-d96cac91ec32_1232x928.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a small friendly pixar-styled robot looking up at a large chart, the chart is connected with multiple dots that stretch out across an entire large wall, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p>If 2024 is supposed to be the year of &#8220;AI in production,&#8221; the first step to getting there is throwing as many ideas at the problem first and hoping for the best on the other side.</p><p>We already saw that with the emergence of retrieval augmented generation, or RAG, which at first seemed like a bandaid but now seems like a permanent fixture. But even those new permanent fixtures have limitations, and there&#8217;s still a bar of quality (and practicality) for all these newer model-powered products to cross.</p><p>Now, another big data technology may be getting a closer look as these kinds of applications move closer to production and demand higher-quality or more accurate responses: graph databases.</p><p>More specifically, some companies are exploring implementing graph search into a RAG setup to improve the relevancy of the results. Like many other once highly specialized technologies, such as vector databases, graph databases potentially offer another layer of relevancy to deliver the right information to a prompt and improve the quality (or accuracy) of the results.</p><p>&#8220;You can think of graph independently and vector independently as one way to store features&#8212;graph captures the connections much better than key-value, and vectors do the semantic capturing much better in a more condense format,&#8221; Naren Narenden, chief scientist at <a href="https://aerospike.com/">Aerospike</a>, a provider of graph and vector databases, told me. &#8220;Independently they work well as new kinds of feature stores, but you can see how they can come together. You can have a RAG application that starts out with semantic search that identifies a few key nodes as a first step, and use the graph to gather the rest of the information.&#8221;</p><p>The idea is pretty straightforward: instead of using just a graph database, or a just vector database, why not both? </p><p>Developers use the same kind of classic RAG approach to retrieve information, but in this case, it&#8217;s retrieving nodes&#8212;a given point with a lot of information connected to other points&#8212;and then running a much more intelligent graph search. </p><p>Graph databases and graph search was already in use in cases where one entity&#8212;say, an e-commerce customer or financial transaction&#8212;has a lot of attributes and is intricately linked to many other concepts. You could see a new transaction show up (which would include location, card number, vendor, amount, and so on) and quickly chase down a lot of information about it to determine if it is an anomaly and should be flagged as fraud. </p><p>That includes a process called traversal, where you search for information by literally move from node to node to find what you&#8217;re looking for. And that&#8217;s largely limited based on where you <em>start</em> <em>on the graph in the first place</em>. But taking a piece of information coming in and using the classic RAG playbook&#8212;embedding it, searching for similar results, and spitting out relevant nodes&#8212;gives you more ideas of <em>where</em> to start on the graph. </p><p>And it turns out it helps potentially alleviate two of the major problems each type has: you have more ideas of where to start search on a graph; and you get a better understanding of relationships and connections than you might get in a vector database.</p><p>&#8220;Vector spaces are completely opaque&#8212;it&#8217;s a bunch of numbers, and human beings can&#8217;t parse that,&#8221; Emil Eifrem, CEO of graph database provider <a href="https://neo4j.com/">Neo4j</a>, told me. &#8220;Graph spaces meanwhile are completely explicit. I show an apple and a tennis ball, and vector similarity search will say they&#8217;re similar but not tell us why. It&#8217;s some dimensions in some kind of latent space out there. In graph space, though, an apple and an orange we see are related explicitly because they&#8217;re both fruit.&#8221;</p><p>For now, it&#8217;s still a two-step process for those providers, and it doesn&#8217;t show up all that often. But as companies move from proof-of-concept to actual production and understand the real performance needs of these apps, anything and everything is up for discussion. And that includes determining whether graph databases fit into a more modern AI data stack.</p><h2>The return of graph databases</h2><p>Perhaps the most well-known startup in graph databases historically is Neo4j, which was founded in 2007. <a href="https://neo4j.com/press-releases/neo4j-announces-seriesf-funding/">Neo4j last raised $325 million at a valuation over $2 billion in June 2021</a>, and <a href="https://neo4j.com/press-releases/2021-company-momentum/">said in January that it had passed $100 million in annual recurring revenue</a>. Neo4j is backed by GV, Greenbridge Venture Partners, Eurazeo, One Peak, and several others. (And ironically, like many older big data startups Neo4j had fallen so far off the radar of most that I talk to that pretty much everyone was curious what they&#8217;re up to.)</p><p>Graph databases have typically found a place in financial services, particularly around fraud prevention, as well as recommendation engines. Each person or transaction has some kind of unique information that is intricately linked to other pieces of information. In the case of a recommendation engine, it could be a user&#8217;s prior purchasing or search behavior, which can change in near real time.</p><p>Vector databases, meanwhile, are well suited for that flood of unstructured information. Lots of different kinds of data sources&#8212;from PDFs to Slack messages or confluence pages&#8212;can all get dumped into the same place for retrieval. Graph thrives on understanding those relationships, and as a result, needs something that falls in the middle.</p><p>Graph databases have certainly not gone away, though they have been traditionally thrown into the bucket of &#8220;very important for some use cases.&#8221; And that&#8217;s still going to be the case going forward, as the technical barrier for implementing this dual setup is pretty high. Still, this month <a href="https://aws.amazon.com/blogs/database/gql-the-iso-standard-for-graphs-has-arrived/">the International Organization for Standardization</a> actually published GQL, a sibling language to SQL designed for graph databases.</p><div><hr></div><h6><strong>Sponsored</strong></h6><h4><strong>This week&#8217;s newsletter is brought to you by Felicis</strong></h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="http://felicis.com" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kU-G!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 424w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 848w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1272w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kU-G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png" width="1456" height="366" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:366,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:381012,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;http://felicis.com&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kU-G!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 424w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 848w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1272w, https://substackcdn.com/image/fetch/$s_!kU-G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd84f0de6-5031-449f-baf5-ed5321f7f768_1822x458.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Felicis has been investing in an impressive variety of <a href="https://www.felicis.com/sectors/ai">AI</a> and <a href="https://www.felicis.com/sectors/infra">infrastructure</a> companies, including: Runway, Weights &amp; Biases, MotherDuck, Supabase, Metaplane, Vannevar Labs, poolside, and Flower Labs. A generalist firm started by Google&#8217;s first product manager, Felicis is known for supporting founders with their <a href="https://www.felicis.com/founder-pledge">Founder&#8217;s Pledge</a>.</p><p><em><a href="https://www.felicis.com/founders-on-felicis">Read what founders say about working with Felicis.</a></em></p><div><hr></div>
      <p>
          <a href="https://www.supervised.news/p/where-graph-databases-live-in-a-future">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Vibes-based evals and Weights & Biases' second act]]></title><description><![CDATA[Plus, a Llama 3 price war is brewing, and what Snowflake's new model means for pre-training.]]></description><link>https://www.supervised.news/p/vibes-based-evals-and-weights-and</link><guid isPermaLink="false">https://www.supervised.news/p/vibes-based-evals-and-weights-and</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Thu, 25 Apr 2024 20:02:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!C7Cy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!C7Cy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!C7Cy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 424w, https://substackcdn.com/image/fetch/$s_!C7Cy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 848w, https://substackcdn.com/image/fetch/$s_!C7Cy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 1272w, https://substackcdn.com/image/fetch/$s_!C7Cy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!C7Cy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp" width="1232" height="928" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:928,&quot;width&quot;:1232,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:387874,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!C7Cy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 424w, https://substackcdn.com/image/fetch/$s_!C7Cy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 848w, https://substackcdn.com/image/fetch/$s_!C7Cy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 1272w, https://substackcdn.com/image/fetch/$s_!C7Cy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F80abd0d4-022a-4b62-ad06-c6df57ddf864_1232x928.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a tiny friendly robot holding a small hammer standing in front of a giant computer screen, the screen takes up an entire wall that dwarfs the robot, the screen has multiple graphs, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>Author&#8217;s note: covering a few things today starting with some notes on Llama 3 for free subscribers, and then a deeper dive into Weights &amp; Biases and its next act and Snowflake&#8217;s latest move into LLMs.</em></p><div><hr></div><h2>OpenAI&#8217;s price advantage is eroding even further</h2><p>Last week, Meta did the Meta thing where it released a family of new Llama models, which are smaller open-ish source models that are trained on <em>a ton amount of data for a</em> <em>really long time</em>. (The largest of the family isn&#8217;t done training.)</p><p>OpenAI already had to contend with a series of products that were highly competitive with GPT-3.5 Turbo, it&#8217;s workhorse model designed for <em>most</em> tasks. Its advantage was it was comically easy to implement, even as other providers <a href="https://www.supervised.news/p/the-point-of-endpoints">started to release drop-in replacements endpoints over the past year that only require a few lines of code</a>.</p><p>As expected, The Llama 3 models that came out are <em>good</em>. There are more than enough technical teardowns out there and I&#8217;ll leave it up to the experts for that analysis. But its middle-of-the-road version, Llama 3 70B, is reaching the point where giving up that kind of consolidated enterprise experience&#8212;where you get a bunch of tools beyond just text completion in a single product&#8212;is actually starting to make sense because <em>it&#8217;s really cheap</em>.</p><p>That combination of <em>pretty</em> <em>cheap</em> and <em>all-in-one</em> has long been OpenAI&#8217;s advantage. You are only paying one company for a lot of stuff. But Llama 3 70B <a href="https://cdn.arstechnica.net/wp-content/uploads/2024/04/llama3_benchmarks.png">pretty much smushes GPT-3.5 Turbo</a> when it comes to most benchmarks (yes, we&#8217;re doing this again), while still being cheaper than its workhorse GPT-3.5 Turbo. </p><p>OpenAI was already staring down <a href="https://www.supervised.news/p/a-new-rival-startup-is-emerging-for">a lot of new &#8220;rivals&#8221;</a> that were basically riding the open source wave to offer really competitive products to GPT-3.5 Turbo thanks to releases from Mistral. Together AI, Fireworks, and others all rolled out endpoints for Mixtral which offered GPT-3.5 Turbo level scores for a fraction of the price.</p><p>But! That still comes with a pretty steep operational tradeoff. Simply chasing endpoints and optimizing for cost runs the risk of having to juggle way too many vendors and create operational headaches even if it&#8217;s shaving off a substantial portion of usage&#8212;especially if teams are just running wild with corporate cards. This is especially true for companies that are shipping data to these API providers that are a little less proven (and less-funded) than OpenAI.</p><p><a href="https://www.supervised.news/p/batch-processing-and-the-rise-of">The majority of AI in production is still batch processing, which focuses on the use of smaller models</a>. While there&#8217;s a class of companies that rely on results that are at a GPT-4 Turbo level of quality (particularly for copilot products), <em>most</em> companies aren&#8217;t there and probably won&#8217;t be there for a while. Those smaller models are usually good enough (particularly if there&#8217;s some light customization on proprietary data).</p><p>OpenAI now has to contend with whether Llama 3 creates both an operational and a <em>behavioral</em> shift. Something that&#8217;s in shooting distance of GPT-4 is now widely available, and it&#8217;s at a comically low price from a lot of providers that have already established themselves as reliable compared to GPT-4 Turbo. Setting aside scores and all that, this is a really tantalizing proposition for companies that have held out on deploying GPT-4 level products due to concerns around cost, latency, security, and reliability. And we haven&#8217;t even gotten to customization of Llama 3.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yBE2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yBE2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 424w, https://substackcdn.com/image/fetch/$s_!yBE2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 848w, https://substackcdn.com/image/fetch/$s_!yBE2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 1272w, https://substackcdn.com/image/fetch/$s_!yBE2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yBE2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png" width="1456" height="2258" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2258,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:187198,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yBE2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 424w, https://substackcdn.com/image/fetch/$s_!yBE2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 848w, https://substackcdn.com/image/fetch/$s_!yBE2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 1272w, https://substackcdn.com/image/fetch/$s_!yBE2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7babac3-8414-4872-8b95-366cc46132e8_1960x3040.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In addition to significantly cheaper and <em>almost good enough</em> endpoints available, most companies at this point understand that hosted versions of all of these&#8212;such as those through Together AI&#8212;push that cost down even further. The barrier to getting dedicated infrastructure is also dropping, <a href="https://www.supervised.news/p/batch-processing-and-the-rise-of">which is </a><em><a href="https://www.supervised.news/p/batch-processing-and-the-rise-of">also</a></em><a href="https://www.supervised.news/p/batch-processing-and-the-rise-of"> manifesting in a proliferation of pre-training</a>.</p><p>It&#8217;ll take a while before we start to see how Llama 3&#8217;s largest model performs in the wild relative to the degree that Meta has handcuffed it. But these benchmarks are the only things that currently present a buyer an apples-to-apples comparison, and they are really favorable to Meta. (The reality is these comparisons are obviously much more complicated, which we&#8217;ll get to in a second.)</p><p>OpenAI still offers what&#8217;s probably the easiest to use fine-tuning API out there with GPT-3.5 Turbo, which when you talk to developers is one of its more underrated products. But all these endpoint companies are incentivized to push customers to dedicated instances, and they&#8217;re going to finesse that pathway to customizing models like Llama 3 to the best that they can.</p><p>Inevitably this all benefits Meta. They get new tools to build into their own products and doesn&#8217;t have to pay another provider. Their intricate institutional knowledge allows them to squeeze even more power out of something like Llama 3 in the same way it can do with PyTorch. And they build up an insane amount of goodwill amongst developers&#8212;which inevitably comes at the expense of more closed developers.</p><p>Companies now have a really clear and <em>cheap</em> pathway to more advanced use cases for language models. What that looks like is still unclear (absent of an explosion in new copilots), but Meta may once again be forcing everyone&#8217;s hands in a more unexpected way.</p><div><hr></div><h2>Weights &amp; Biases makes a bet on software developers</h2><p>Before the launch of ChatGPT, one of the hottest areas in machine learning investing was a category we (arguably haphazardly) referred to as machine learning ops&#8212;or MLops.</p><p>That category comprised a lot of different startups and emerging technologies, such as feature stores with Tecton or frameworks like Anyscale&#8217;s Ray. But perhaps the biggest darling among investors and developers of the MLops category is <a href="https://wandb.ai/site">Weights &amp; Biases</a>, which manages one of the most critical workflow components of machine learning: experiment tracking.</p><p>But Weights &amp; Biases has also turned into a test case for the single biggest questions for startups in MLOps stack: can these startups building tools for machine learning create an &#8220;act two&#8221; for a new generation of technologies powered by language models?</p><p>Weights &amp; Biases is trying to answer that <a href="https://wandb.ai/site/weave">with the launch of Weave</a>, an updated suite of tooling for evaluating and debugging language models in production. And perhaps more importantly, Weave is targeted at the software engineers that are tasked with putting language models into production in some matter&#8212;a departure from its traditional focus on machine learning engineers and classic machine learning.</p><p>&#8220;We don&#8217;t think about machine learning engineers as a subset of software developers, we think about them as something different, and that&#8217;s really helped us,&#8221; Weights &amp; Biases CEO Lukas Biewald told me. &#8220;Now there&#8217;s a kind of software developer that&#8217;s developing generative AI, and they&#8217;re actually having to learn this experimental workflow that&#8217;s brand new to them." It&#8217;s actually not obvious to some software developers that they should even do that.&#8221;</p><p>Weave represents a lot of things, but if we wanted to oversimplify it here, it&#8217;s effectively a framework to debug LLM apps and (more importantly) try to formalize vibes-based evaluations. While developers fight to one-up each other leaderboards with <a href="https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard">benchmarking scores</a> or <a href="https://chat.lmsys.org/?leaderboard">Elo ratings</a>, the reality is the &#8220;success&#8221; of an LLM internally at companies is often determined from eyeballing the results. They look at loosely-defined evaluation scores (sometimes even based off regex), or customer feedback, or even the results themselves, and are like, yeah, looks good to me.</p><p>Language models have inspired a whole new cohort of developers, and many of them are hacker-types that just want to string together a bunch of APIs and tools into something fun. Most language models in production today aren&#8217;t necessarily concerned with complicated processes like model lifecycle management. Instead, you can just drop an API into some frontend skin and, voila, you have an AI-powered app. </p><p>Most in the field I talk to agree that Weights &amp; Biases pretty much owns the experiment tracking market, even with projects like <a href="https://mlflow.org/">MLFlow</a> and alternatives like <a href="https://www.comet.com/site/">CometML</a>. Early last year, <a href="https://www.supervised.news/p/the-second-act-of-the-pre-gpt-ai">Weights &amp; Biases was in conversations for a funding round</a> that would value the startup at $2 billion and had between $20 million and $23 million in annual recurring revenue. (The round, which was opportunistic, didn&#8217;t materialize&#8212;though W&amp;B would raise another strategic round later in the year.)</p><p>And Weights &amp; Biases continues to quietly grow in the category it effectively helped create, even if it isn&#8217;t getting the kind of hype that generative AI application and infrastructure startups are getting. But there are many, many, many more developers than there are machine learning engineers&#8212;and they&#8217;re often the ones in the room when a CEO walks in and says &#8220;put AI in production&#8221; with absolutely no additional guidance or context.</p><p>Weights &amp; Biases is also quickly becoming a test case of whether a highly-successful MLops company (alongside <a href="https://www.supervised.news/p/what-happens-to-fivetran">others in the modern data stack</a>) can grow into a market built around a once-in-a-generation technology. And it&#8217;s now moving aggressively to show that it has a large place workflows built on modern AI tooling.</p><h2>Powering developer workflows in AI</h2><p>Weights &amp; Biases held its first user conference in June last year and it captured the vibe of AI almost perfectly at that time&#8212;incredibly dense, chaotic, a little disorganized (with panels and talks going well into the evening), and overflowing with excitement and energy from attendees despite its practically remote location in the Dogpatch in San Francisco.</p><p>At the time, there were glimpses of what Weights &amp; Biases planned for generative AI with the launch of <a href="https://wandb.ai/site/solutions/llms">tools for aiding prompt engineering</a>. Its tools were perhaps one of the most well-suited to adapt into generative AI use cases as prompt hacking emerged as an important part of building out workflows on top of language models. (One developer with often describes this to me as &#8220;begging the model to do something.&#8221;)</p><p>This year&#8217;s conference was at the Mariott Marquis in San Francisco&#8217;s downtown, practically next door to another conference in the venue. Both more subdued and considerably more formal, it <em>also</em> represented the kind of <em>vibe shift</em> in AI from a limitless technology to something that enterprises were over experimenting with, and were looking to actually use it to drive value for their organizations.</p><p>As machine learning and deep learning took hold in larger companies, Weights &amp; Biases quickly became a go-to platform for managing the lifecycle of machine learning models. That&#8217;s defined by experimentation, testing, iteration, and then (rarely) deployment&#8212;and starting the cycle over again. And <a href="https://medium.com/@l2k/starting-a-second-machine-learning-tools-company-ten-years-later-21a40324d091">one of its earliest </a>customers was a small startup at the time that would go on to become arguably the most important company in modern AI: OpenAI.</p><p>Weights &amp; Biases still carries a lot of weight when it comes to language models too, if only because of its ongoing deep connections with OpenAI. Earlier this month, OpenAI unveiled a slew of updates to its GPT-3.5 Turbo fine-tuning API. One of the biggest updates was third-party platform integration which, no surprise, <a href="https://openai.com/blog/introducing-improvements-to-the-fine-tuning-api-and-expanding-our-custom-models-program">included Weights &amp; Biases as its first integration</a>. Fine-tuning, while it offers a lot of potential benefits, is still considered pretty advanced fro most companies&#8212;and would by then have likely roped in the MLE team.</p>
      <p>
          <a href="https://www.supervised.news/p/vibes-based-evals-and-weights-and">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[OpenAI (sort of) covers one of its blind spots]]></title><description><![CDATA[OpenAI now has a batch processing API. But this time around, it&#8217;s dealing with more than just a handful of startups&#8212;including Snowflake and Databricks.]]></description><link>https://www.supervised.news/p/openai-sort-of-covers-one-of-its</link><guid isPermaLink="false">https://www.supervised.news/p/openai-sort-of-covers-one-of-its</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Tue, 16 Apr 2024 13:01:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!tWDu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tWDu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tWDu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!tWDu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!tWDu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!tWDu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tWDu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:545960,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tWDu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!tWDu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!tWDu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!tWDu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ce75377-eb84-4d3d-8a22-f11d93d7ccbd_1024x1024.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a small friendly robot carrying a very large box full of laptops, monitors, and green notebooks, the box is overflowing, the robot is rushing out of an office, the robot is sweating and smiling awkwardly, sunday funny comics aesthetic &#8212; midjourney</figcaption></figure></div><p><em>Author&#8217;s note: Due to the busy nature of the week with multiple events and the launch of Llama 3, Friday&#8217;s column will be moved to early next week and be a double-issue. </em></p><div><hr></div><p>OpenAI is adding another one of the missing pieces in its toolkit that other providers were quickly running off with&#8212;and this time it handles one of the most important pathways to enterprises actually putting language models in production.</p><p>Last week I&#8217;d noted that <a href="https://www.supervised.news/p/batch-processing-and-the-rise-of">most companies that I&#8217;ve been talking to lately have been using modern language models as part of a batch data processing workflow</a>. OpenAI this week announced the <a href="https://twitter.com/OpenAIDevs/status/1779922566091522492">release of its batch processing API</a>, which enables users to process large amounts of data on a less urgent timeline&#8212;but in exchange you get higher rate limits and <a href="https://openai.com/pricing">half off the standard API price</a>. (And it looks like OpenAI has finally adopted the per-million token pricing structure instead of per-thousand.)</p><p>Companies process large amounts of unstructured data, such as customer support tickets, and execute simple tasks such as classification, sentiment analysis, or summarization, on a longer timeline&#8212;typically using smaller models like Mistral 7B or Llama 13B on timescales that are hours (or even days). It&#8217;s not so dissimilar to what companies were already going with BERT, an early open source language model, and it&#8217;s now more efficient and effective.</p><p>OpenAI, the first mover in making language models broadly available through its GPT-series APIs, had effectively sat on the sidelines while companies like Snowflake and Databricks&#8212;hosting these smaller open source models adjacent to company data&#8212;built up what seems to increasingly be the standard use of language models in enterprises. While you could fire up a smaller Mistral model on top of your data, GPT-3.5 Turbo was still $0.50 per million input tokens and $1.50 per million output after its latest price cut.</p><p>In fact, when you talk to partners, platforms, or the enterprises themselves, batch data processing is the clearest case of &#8220;AI in prod&#8221; you&#8217;re going to get right now. The whole dream of some mythical autonomous agent replacing entire teams has largely faded into the background, and in its place is a bunch of pretty generic and boring use cases that actually return a disproportionate amount of value.</p><p>As is the case with many of those modalities and business models that OpenAI faces increasing competition from startups (and, somewhat less obviously, platform providers), it&#8217;s adding the API to its repertoire in a kind of better-late-than-never approach. But it looks like its own &#8220;flavor&#8221; here is that it&#8217;s offering a cheaper versions of its portfolio of models, rather than just one or two smaller ones.</p><p>There are plenty of reasons why OpenAI would want to do this, least of all to keep up with potential customers turning to others for cheaper batch data processing. But it also offers <a href="https://www.supervised.news/p/the-point-of-endpoints">another interesting opportunity to sate another emerging challenge for companies that manage hordes of GPUs</a>: getting as close to maximum utilization as possible.</p><p>This time around, though, OpenAI isn&#8217;t just dealing with a handful of startups trying to eat away its cost advantage on the edges. It has to contend with the challenge that enterprises handling batch data processing are the same ones that probably have their data already stored somewhere that offers it in some fashion.</p><h2>The appeal of batch data processing</h2><p>When a request that&#8217;s uniquely suited to a language model isn&#8217;t particularly urgent&#8212;like an angry customer on the line&#8212;it turns out you have a lot more language model options at your disposal. You could instead comb through <em>all</em> of those angry customer calls, using a less-powerful model to extract basic insights from them to understand why you product might be broken in the first place. A model with GPT-4 level quality isn&#8217;t really necessary, nor would it be particularly practical from a cost perspective.</p><p>A lot of these problems end up getting routed through centers because they have a kind of "eye test" feel to them, where you couldn't just directly extract the result with a standard machine learning model. Companies like&nbsp;<a href="https://www.linkedin.com/company/snowflake-computing/">Snowflake</a>&nbsp;and&nbsp;<a href="https://www.linkedin.com/company/databricks/">Databricks</a>&nbsp;have been implementing access to open source models like Mistral's, providing those kinds of tools right on top of a company's pool of data.</p><p>Realistically, a company could upgrade up to a more powerful model for batch processing at some multiple of the price of a smaller model like one from Mistral or Meta (or any of its open-available fine-tuned offshoots). With OpenAI&#8217;s batch pricing, it brings it more in line with a company looking at using a Mixtral-type model. And, again, the benefit OpenAI has broadly speaking is that it&#8217;s just incredibly easy to use.</p><p>So we&#8217;re going to look at two different price breakdowns&#8212;one that feels a smidge closer to reality for enterprises with batch processing already in place that are looking at using smaller (especially customized) models, and one that probably better reflects the reality of people that might use the OpenAI batch API. </p><p>As a usual caveat here, <em>most of the value of these smaller models is unlocked through customization</em>. So a raw endpoint-to-endpoint pricing is a little fuzzy as fine-tuned versions are hosted and have different pricing principles (such as GPU-per-hour pricing). OpenAI <em>does</em> offer a fine-tuned GPT-3.5 Turbo, but it&#8217;s on the order of $3 per million input and $6 per million output.</p><p>But, we&#8217;re working with what we have, and the smaller models are already somewhat capable of doing some of these hyper-specific workflow tasks (like less-complicated summarization) before fine-tuning. Let&#8217;s start with the comparisons with some endpoints of Mistral 7B on other providers <a href="https://openai.com/pricing">to OpenAI&#8217;s GPT-3.5 Turbo model at a 50% discount</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YQ3y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YQ3y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 424w, https://substackcdn.com/image/fetch/$s_!YQ3y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 848w, https://substackcdn.com/image/fetch/$s_!YQ3y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 1272w, https://substackcdn.com/image/fetch/$s_!YQ3y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YQ3y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png" width="888" height="544" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:544,&quot;width&quot;:888,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:32873,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YQ3y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 424w, https://substackcdn.com/image/fetch/$s_!YQ3y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 848w, https://substackcdn.com/image/fetch/$s_!YQ3y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 1272w, https://substackcdn.com/image/fetch/$s_!YQ3y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54476cd6-e1e8-41de-8609-69e3cfd9092e_888x544.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As you&#8217;d expect, the pricing here doesn&#8217;t exactly lend itself well for OpenAI. There&#8217;s certainly an upper bound to the performance of these models and some additional complexity that comes with hosting a fine-tuned version, which would bring the performance for specific tasks more in line with what a GPT-series model can do from OpenAI.</p><p>OpenAI, however, is probably looking for a bit of a more favorable comparison here&#8212;with GPT 3.5-Turbo&#8217;s quality more in line with that of Mixtral or Claude Haiku. If we were to look at a comparison amongst those models&#8230;</p>
      <p>
          <a href="https://www.supervised.news/p/openai-sort-of-covers-one-of-its">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Batch processing and the rise of CPUs]]></title><description><![CDATA[Plus: the death of pre-training seems greatly exaggerated.]]></description><link>https://www.supervised.news/p/batch-processing-and-the-rise-of</link><guid isPermaLink="false">https://www.supervised.news/p/batch-processing-and-the-rise-of</guid><dc:creator><![CDATA[Matthew Lynley]]></dc:creator><pubDate>Fri, 12 Apr 2024 21:12:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!n17y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!n17y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!n17y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 424w, https://substackcdn.com/image/fetch/$s_!n17y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 848w, https://substackcdn.com/image/fetch/$s_!n17y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 1272w, https://substackcdn.com/image/fetch/$s_!n17y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!n17y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png" width="1232" height="928" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:928,&quot;width&quot;:1232,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1330395,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!n17y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 424w, https://substackcdn.com/image/fetch/$s_!n17y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 848w, https://substackcdn.com/image/fetch/$s_!n17y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 1272w, https://substackcdn.com/image/fetch/$s_!n17y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72a2cff6-4988-4103-bdb4-a6ece5f39028_1232x928.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">a small sleek friendly pixar-styled robot standing underneath a towering GPU, the GPU is ten times the size of the robot, the robot is holding a power cable and a large book, it's daylight in the middle of a field like the windows XP background, sunday funny comics aesthetic, --ar 4:3 &#8212; midjourney</figcaption></figure></div><p><em>Both issues for the week are coming together in this one! First we&#8217;ll talk about how enterprises are actually using language models in prod, and then how pre-training still seems alive and well.</em></p><div><hr></div><p>When ChatGPT launched, CEOs at pretty much every company you talk to gave a top-down mandate to implement AI into their products.</p><p>And to be clear, at the time, <em>no one</em> knew what that looked like. But the extreme hype around language models&#8212;along with the fact that they, in a lot of ways, worked out of the box&#8212;made them impossible to ignore. The fever dreams of replacing entire work functions were pervasive. And the common refrain was along the lines of, &#8220;if not me, then one of my competitors will do it.&#8221;</p><p><a href="https://www.supervised.news/p/ai-in-august-dwindling-hype-unexpected">That hype cycle </a>has ended, thankfully, and language models specifically have started to find a home within larger companies. But the reality among enterprises and platform providers I talk to, though, is that the use cases aren&#8217;t as epically transformative as we might have originally predicted. </p><p>Actually, they&#8217;re (at face value) pretty straightforward. Instead of creating complex reasoning tasks and full-on automation that replaces entire teams, many enterprises I talk to are either using language models, or plan to use language models, for batch data processing.</p><p>That can include summarization, classification, entity extraction, or even cleaning up data. And it shouldn&#8217;t exactly come as a shock&#8212;<a href="https://www.tensorflow.org/text/tutorials/classify_text_with_bert">that&#8217;s what companies were doing with BERT</a>, <a href="https://research.google/blog/open-sourcing-bert-state-of-the-art-pre-training-for-natural-language-processing/">one of the earliest open source language models</a>, in a lot of use cases. The emergence of the Llama-series models and Mistral&#8217;s models have just jacked up the quality of those results and reduced the barrier to entry to the point that they can be more efficient and (a little more) trustworthy.</p><p>&#8220;All of our customers have this dream of the full-on automation, but the steps to get there feel a little more incremental than a massive shift than all of our chats and emails and SMS handled by something that&#8217;s perhaps a black box,&#8221; Ben Gleitzman, CTO and co-founder of Replicant, a call center automation tool, told me. Replicant is backed by Salesforce Ventures, Norwest, Atomic, and others.</p><p>The &#8220;batch&#8221; part here refers to analyzing large amounts of data on a non-urgent timeline. The tasks here are usually relatively simple and straightforward and not in the kind of &#8220;write an email for me in the style of X&#8221; fashion that require the quality level of, say, a GPT- or Claude-series model. Instead, you&#8217;re more or less on a fact-finding mission within troves of proprietary company data.</p><p>While you could do it through one of OpenAI&#8217;s APIs, open source models like Mistral&#8217;s smaller 7B model, are actually well-suited to the problem&#8212;particularly for enterprises&#8212;because they&#8217;re efficient, cheap, and in many cases are already plugged into data abstraction layers. Databricks and Snowflake <a href="https://www.supervised.news/p/the-point-of-endpoints">have both announced access to Mistral&#8217;s open source models through endpoints</a>, and it&#8217;s an easy jump to a fine-tune because the data&#8217;s already there.</p><p>&#8220;We see a lot of batch processing&nbsp;on Snowflake, from summarization and the sentiment analysis of support tickets, to batch data extraction from SEC filings or the competitive analysis of lost deals,&#8221; Baris Gultekin, Head of AI at Snowflake, told me. &#8220;Large language models increase the productivity of analysts substantially, enabling them to quickly extract insights from text by using something as simple as the English language, all in a cost-effective way across millions of rows of data.&#8221;</p><p>More importantly, these smaller models don&#8217;t necessarily need the extreme power of a cluster of H100 GPUs. It&#8217;s even to the point that Apple&#8217;s <a href="https://twitter.com/awnihannun">Anwi Hannun</a>, who is behind <a href="https://github.com/ml-explore/mlx">Apple&#8217;s excellent MLX Mac development and inference framework</a>, and <a href="https://github.com/ggerganov/llama.cpp">Llama.cpp</a> creator <a href="https://twitter.com/ggerganov?lang=en">Georgi Gerganov</a> <a href="https://twitter.com/ggerganov/status/1775921043858764061">are competing</a> over who can get the highest throughput for Mistral models on a Mac. </p><p>As these models continue to improve, and can run on low-power hardware, they&#8217;re becoming a very attractive option for companies looking to remove a lot of &#8220;eye tests&#8221; in workflows that might be offloaded to support centers. And while we wait for a &#8220;killer&#8221; use case for language models (if one ever emerges), it turns out that some enterprises have already found a reason to implement it&#8212;even if it&#8217;s a bit of a snoozer.</p><h2>Where the advantage of smaller models come in</h2><p>Anyone who&#8217;s worked in sales, research, marketing, or customer service has done their time sitting in on dozens of calls listening to what customers or prospects are saying and asking for, independent of what the data shows. Jeff Bezos gave a particularly good summary of why<a href="https://youtu.be/DcWqzZ3I2cY?feature=shared&amp;t=5628"> on an extensive interview with Lex Friedman</a>:</p><blockquote><p>When the data and anecdotes disagree, the anecdotes are usually right. It doesn&#8217;t mean you just slavishly slavishly follow the anecdotes, it means you go examine the data. It&#8217;s usually not that the data is being mis-collected it&#8217;s usually that you&#8217;re not measuring the right thing.</p></blockquote><p>Now, there are a lot of ways to slice that statement. But <em>one</em> part of it is that understanding what customers need or are looking for comes down to a human with empathy and experience eyeballing something, and getting some insight that can&#8217;t be mechanically extracted with traditional machine learning. Or, at least, something you can&#8217;t extract at a level that&#8217;s fully trustworthy and error-proof.</p><p>While early, modern language models offer a potential chance to capture some of that so-called &#8220;anecdata&#8221; by trying to understand semantically what information is coming in. Many companies that would benefit from using it for batch processing have spent <em>years</em> collecting information from, for example, customer support resolutions. And the results are usually pretty binary: did the customer resolve their process and have a good experience, or not.</p><p>From there, you could extract flags that suggest what direction to send a ticket, or what the largest pain points customers have had recently. We&#8217;ve been trying for years to automate this with machine learning, but the advent of language models, even smaller ones like Mistral 7B, make the process much more approachable due to their flexibility and ease of implementation. And those language models are able to deliver real results in cases where you would want to eyeball data.</p>
      <p>
          <a href="https://www.supervised.news/p/batch-processing-and-the-rise-of">
              Read more
          </a>
      </p>
   ]]></content:encoded></item></channel></rss>