Hi, itโsย Alexandre. Iโm investing in seed & series A European vertical solutions (vSol) which are industry specific solutions aiming to become industry OS and combining dynamics from SaaS, marketplaces and fintechs. Overlooked is a weekly newsletter about venture capital and vSol. Today, Iโm sharing the insightful tech news of August.
Before jumping into the newsletter, Iโm thrilled to tell you that Iโve joined 20VC as Partner, where Iโll be mainly focusing on our Series A investments.
Joining Harry, Kieran and Paul was an easy decision.
At 20VC, we put content at the core of everything we do - this resonates with what Iโve been doing for years with my newsletter on vertical SaaS, Overlooked, and with the French Tech Ecosystem report I publish every year.
At 20VC, we are building a venture platform differentiated through its content (22m monthly listeners) and network (2,800 previous guests).
Harry, Kieran and Paul are incredible human beings and I already know itโs going to be an opportunity of a lifetime to work with them. I felt at home as soon as I met them, and we share similar core values that I know will make us a great team: putting founders first, being humble, working hard and playing as a team.
At 20VC, we aim to back the most ambitious founders in North America and Europe from Pre-Seed to Series A. If that's you, ping me! As a reader of the newsletter, it means that you can expect bigger content projects going forward!
Thursday, Aug. 1st: I listened CCCโs Q2-2024 earnings call. CCC is a vertical SaaS selling solutions to both insurers and car repair shops to digitise the car repair process. - CCC
CCC ONE generates $400m in annual revenues. Itโs CCCโs cloud based estimating and shop management for car repair shops serving both single store an multi store operators. It competes directly with VC-backed players like ShopMonkey or Autoleap.
Customers consolidating their toolstack on CCC. โWe completed the successful, on schedule roll-out of our full suite of auto physical damage or โAPDโ solutions for a top-20 insurer as they transitioned from multiple vendors to the CCC platform.โ
Reaching 30k car repair shops in the US. โWe have also added over 600 new repair facilities so far in 2024. This growth pushed us across a major milestone: we now have over 30,000 repair facilities on the CCC platform.โ
โWe believe that our investments over the last three years have created an exciting pipeline of new solutions with greater breadth and depth than at any time in our history.
Breaking down revenue growth into 3 drivers: (i) cross-sell/upsell from mature products, (ii) new logos, (iii) emerging solutions. โTotal revenue in the second quarter was $232.6 million, up 10% from the prior-year period.
Approximately 7 points of our growth in Q2 was driven by cross-sell, upsell, and adoption of our solutions across our client base, including repair shop upgrades, continued adoption of our digital solutions, and the ongoing strength in Casualty and Parts.
Approximately 3 points of growth came from our new logos โ mostly our repair facilities and parts suppliers.
About 1 point of growth in Q2 came from our Emerging Solutions, mainly Diagnostics, Estimate โ STP and the new, adjacent Casualty solutions.โ
Friday, Aug. 2nd: Tidemark published a benchmark report on vertical SaaS. Congrats to Michael Tan for his outstanding work on this project! - Tidemark
โNearly 40% of respondents offered a fintech product. Fintech products can address key merchant needs while driving ARPA for Vertical SaaS Vendors. For the survey respondents, Payments at 30% was by far the most common fintech expansion product. This was followed by Lending at 11% and Payroll at 9%.โ
When integrated into vertical SaaS platforms, payments exhibit the highest attach rate at 50%, outperforming other financial services such as lending (32%) and payroll (10%).
31% of respondents have launched an AI product and 39% more are planning to launch one in the next 12 months. In terms of pricing, 24% of respondents with an Al product are offering it completely free - while 13% are including it as a premium feature for their premium pricing tier. For direct monetization structures, 46% are doing so via SaaS subscription, 11% via transaction fees, and 7% as a per employee per month structure.
Saturday, Aug. 3rd: Lux published its Q2-2024โs letter to LPs. - Lux Capital
Lux emphasizes its focus on investing in โrebel scientistsโ who challenge conventional thinking. These visionaries are seen as critical to advancing human knowledge and technology, driving innovation in areas like AI, biotechnology, and aerospace.
โNo institution is more important for progress than the laboratory, the epicenter of fundamental shifts in our understanding of the world.โ
โThe federal share of total R&D funding went from around 67% in the 1960s to less than 20% in recent years thanks to expansive spending from the private sector, which now accounts for 75% of R&D. Notably, the private sector is doing more basic research, taking share from universities.โ
While there is immense potential with AI, the sector is also experiencing significant volatility. Lux focuses on companies with defensible technology and long-term potential.
โOver the past two decades, Lux Labs has been an informal initiative that embodies Luxโs commitment to creating companies from scratch in overlooked areas of the hard sciences. Weโve been at the forefront of de novo company creation, turning research from university labs and internally developed theses into over 20 groundbreaking ventures including: Kurion (a nuclear waste treatment company that aided the cleanup of Japanโs Fukushima disaster), Kallyope (in metabolic disease and obesity), Cajal (in neurodegenerative disease), Genocea (in vaccines), Aera (in genetic medicines) and many more.โ
โWe expect this spin-out and start-up strategy to be repeatable for Lux as big tech companies focus on profitability and cut spending on non-core research projects.โ
โTodayโs AI investors are always seeking Yesterdayโs Yes: people want to be invested today where they should've been investing years ago.โ
โThe vertiginous ascent of capital infusion into AI yields a hype-hope continuum, where short-term disappointment may yet yield long-term delight echoing the trajectory of the internet's evolution.โ
โAI isn't a novel distribution platform, which means that many of the gains will accrue to the inertia of incumbency. Take Microsoft, which is using regulatory arbitrage to control OpenAI while also skirting merger scrutiny with Inflection by licensing IP and acqui-hiring its key talent (paying out investors with a modest but not fund-making multiple).โ
โFar from being a democratized asset class, VC is and will remain a rarified ecosystem where only a select cadre of firms consistently access the most promising opportunities. The vast majority of new participants engage in what amounts to a financial fool's errand. We continue to expect the extinction of as many as 30-50% of VC firms.โ
Sunday, Aug. 4th: Procore published its Q2-2024 earnings. Itโs a cloud based construction management software initially focused on streamlining communication among all stakeholders on a construction project. - Procore
Procore is changing its organisation to go from $1bn to $10bn in revenues with general managers in key geographical markets with more leeway to adapt the product and the go-to-market to local specificities.
โI'd like to spend some time today discussing changes that we're making to the business. This year, we are guiding to surpass $1bn in full year revenue. And while this is an important milestone for Procore, it also represents a key inflection point as we move towards becoming a multibillion revenue company.โ
โI'd like to walk through some of the specific changes that we're making to our go-to-market operating model. First, we're moving from a matrix organization structure to a general manager model.โ [โฆ] โGeneral managers will have combined teams reporting into them, which should streamline engagement across the customer life cycle. The general managers roles will cover the following regions: North America, Europe, Australia, New Zealand and the Middle East, which represents our current geographic coverages. In the future, when we elect to move into new countries within these regions, the general managers of these respective regions are going to be better positioned to determine specific strategies and tactics for that particular country.โ [โฆ] โThis creates nimble teams that can quickly address on-the-ground dynamics as they arise. General managers will have the autonomy to respond to the local nature of construction and to tailor and customize their own motions specific to the needs of the markets that they serve.โ
International revenues are growing at a faster pace than total revenues. โTotal revenue in Q2 was $284 million, up 24% year-over-year, and international revenue grew 31% year-over-year.โ [โฆ] โQ2 non-GAAP operating income was $50 million, representing an operating margin of 17.6%โ
Procoreโs Pay adoption is increasing. โMoving on to an update on Pay. Q2 represented only the second full quarter of general availability, and we continue to see strong adoption here. Given the ramp period associated with onboarding Pay, we do not expect this to be a material contributor to our financial results this year, but we are pleased with the progress we have seen thus far and believe this represents a significant opportunity to drive greater cross-sell as we further extend our platform functionality throughout the construction life cycle.โ
Leveraging partners as a go-to-market channel. โIt's a huge opportunity for Procore to capitalize on the opportunity for channel, both in the U.S. and globally.โ
โAchieved a gross revenue retention rate of 94% in the second quarter. Number of organic customers contributing more than $100,000 of annual recurring revenue totaled 2,191 as of June 30, 2024, an increase of 20% year-over-year. Added 152 net new organic customers in the second quarter, ending with a total of 16,750 organic customers.โ
Monday, Aug. 5th: I listened to Sam Altman discuss OpenAI and the future of AI on Lex Fridmanโs podcast. - Lex Fridman
โI think all of these models understand something more about the world model than most of us give them credit for.โ
โOne thing we have to do before releasing the [Sora] system is just get it to work at a level of efficiency that will deliver the scale people are going to want from this so that I donโt want to downplay that.โ
โPeople talk about how many jobs is AI going to do in five years. [โฆ] The way I think about it is not what percent of jobs AI will do, but what percent of tasks will AI do on over one time horizon. So if you think of all of the five-second tasks in the economy, five minute tasks, the five-hour tasks, maybe even the five-day tasks, how many of those can AI do? I think thatโs a way more interesting, impactful, important question than how many jobs AI can do because it is a tool that will work at increasing levels of sophistication and over longer and longer time horizons for more and more tasks and let people operate at a higher level of abstraction.โ
โI expect that the delta between [GPT] 5 and 4 will be the same as between [GPT] 4 and 3 and I think it is our job to live a few years in the future and remember that the tools we have now are going to kind of suck looking backwards at them and thatโs how we make sure the future is better.โ
โOne thing Iโve been using [Chat GPT4] for more recently is sort of like a brainstorming partner. Thereโs a glimmer of something amazing in there.โ โOne of the other things that you can see a very small glimpse of is when [Chat GPT4] can help on longer horizon tasks, break down something in multiple steps, maybe execute some of those steps, search the internet, write code, whatever, put that together. When that works, which is not very often, itโs very magical.โ
โIn some sense, we did have to do two things. We had to invent the underlying technology and then we had to figure out how to make it into a product people would love, which is not just about the actual product work itself, but this whole other step of how you align it and make it useful.โ
โThe thing that I find most interesting is people mostly younger people - who use it as their default start for any kind of knowledge work task.โ โThe most interesting thing to me is the people who just use it as the start of their workflow.โ
โWhat people want is a model that gets to know me and gets more useful to me over time.โ โWe all gain from experience over the course of our lives in varying degrees, and Iโd like my AI agent to gain with that experience too.โ
โI think compute is going to be the currency of the future. I think it will be maybe the most precious commodity in the world, and I think we should be investing heavily to make a lot more compute.โ โIntelligence is going to be more like energy or something like that, where the only thing that I think makes sense to talk about is, at price X, the world will use this much compute, and at price Y, the world will use this much compute. Because if itโs really cheap, Iโll have it reading my email all day, giving me suggestions about what I maybe should think about or work on, and trying to cure cancer, and if itโs really expensive, maybe Iโll only use it, or weโll only use it, to try to cure cancer.โ
โI think the world is going to want a tremendous amount of compute. And thereโs a lot of parts of that that are hard. Energy is the hardest part, building data centers is also hard, the supply chain is hard, and then of course, fabricating enough chips is hard.โ
โWhat are the pros and cons of the competition in the space and competing with Google, Meta, xAI, and others? I think I have a pretty straightforward answer to this that maybe I can think of more nuance later, but the pros seem obvious, which is that we get better products and more innovation faster and cheaper, and all the reasons competition is good. And the con is that I think if weโre not careful, it could lead to an increase in sort of an arms race that Iโm nervous about.โ
โPeople use extremely different definitions for what AGI is. It makes more sense to talk about when weโll build systems that can do capability X or Y or Z, rather than when we fuzzily cross this one mile marker.โ
โI expect that by the end of this decade and possibly somewhat sooner than that, we will have quite capable systems.โ
โI think when a system can significantly increase the rate of scientific discovery in the world, thatโs a huge deal. I believe that most real economic growth comes from scientific and technological progress.โ
Tuesday, Aug. 6th: I read Bessemerโs investment memorandum on beauty vertical SaaS enabled marketplace Mindbody in which Bessemer led a $11m series B in July 2010. - Bessemer
โMindbody launched its SaaS application in 2003 and has become the clear leader in the fitness market with over 8,500 clients, $867k in MRR and $1m in monthly revenues (as of May 31).โ
โWe are impressed with Mindbodyโs position as the de facto standard in the yoga and pilates market (with over 30% share in both). It has gradually expanded into other verticals, including fitness centers, martial arts studios, dance studios, spas, and salons.โ
โThe company has grown from $4.4m in recurring revenue in 2008 to $6.8m in 2009 and is currently growing at a 60% CAGR, on plan to reach $11.4 in 2010.โ
โThe company has acquired customers through a capital-efficient telesales effort, with a gross profit customer acquisition payback of ~13 months.โ
โWe also believe ARPU and margin per account will grow as Mindbody increases its payment processing penetration and starts to drive traffic from its newly-launched consumer-facing offering.โ
โPricing ranges from $29/mth for a single professional facility to $155/mth for facilities with 20+ professionals, with an average price of $75/mth per client.โ
โOur sense is that Mindbodyโs product is successful due to weak competition. Most of Mindbodyโs competitors are on-premise incumbents founded in the mid-1990s that lack the advantages of SaaS. The handful of SaaS competitors are poorly capitalized and sub-scale. Forโ
โAbout 30% of Mindbody subscribers and 70% of new subscribers process payments through Mindbody. This accounts for about 20% of recurring revenues and generates about $25/mth in additional MRR (bringing the total MRR per subscriber to about $100/mth).โ
โMindbody employs an efficient marketing-driven telesales model. The company relies heavily on SEM, outbound cold calling, and tradeshows to drive leads. It subsequently converts those leads by phone with an inside sales team. At a blended customer acquisition cost of ~$1,100, Mindbody typically has a 13-15 month gross profit customer payback period.โ
Wednesday, Aug. 7th: I read a market memorandum from Sapphire summarising the current market momentum about Gen. AI and showcasing revenues generated by Gen. AI companies. - Sapphire
"In fairness, many analysts and market participants have been warning that Generative AI is over-hyped as a technology, at least in its current state of development, and that much of the investment activity across both public and private markets appears, in their opinion, bubble-like. However, despite those calls for caution, the AI-optimist narrative seems to have held the upper hand through 2023 and the first half of 2024, creating an environment where increased AI investment was generally rewarded by investors and championed by public and private company executives. However, as we enter the final third of the year, we believe things have started to change.โ
โWe believe this earnings season seems to have marked a transition point as 1) investors are becoming more pointed in their questions to executives regarding the timelines for returns on the ever-increasing levels of AI infrastructure spend, and 2) executives are more comfortable plainly stating what has been obvious for a long timeโthey are willing to err on the side of over vs. under-investment given the intensely competitive market environment and potentially transformative impact of the technology.โ
โThe one way I think about it is, when you go through a curve like this, the risk of underinvesting is dramatically greater than the risk of overinvesting for us here. Evenโฆ if it turns out we are overinvesting, these are [infrastructures] which are widely useful for us, they have long useful lives, and we can apply it across and we can work through that. But I think not investing to be at the front here, I think, definitely has much more significant downsides.โ - Sundar Pichai (CEO at Google)
โAt this point, I'd rather risk building capacity before it is needed, rather than too late, given the long lead times for spinning up new infra projects. And as we scale these investments, we're of course going to remain committed to operational efficiency across the company.โ - Mark Zuckerberg (CEO at Meta)
On AI startups being acqui-hired by hyperscalers. โSome combination of a
lack of willingness or ability to raise the type of capital required to push through to the next frontier of model development,
recognition that they were operating in rapidly commoditizing areas of the market,
inability to find product-market fit or a clear business model,
challenging hiring and retention dynamics in a hypercompetitive market, and
acknowledgment that the hyperscalers have better capital, compute, and data resources + deep research talent benches.โ
โThe global-scale AI data center build-out has always been good for the startup ecosystem. If the hypercalers are committed to overbuilding, it will further enhance access and reduce costs.โ
Thursday, Aug. 8th: I listened to a 20VCโs podcast interview with David Cahn who is a partner at Sequoia. He invested in AI companies including Hugging Face, Runway, Supabase, Weights & Biases and Replit. - 20VC
โBelieving in AI and believing that infinite CapEx in the next two years are very different concepts. You can believe in AI but also believe that the amount of CapEx in the next 24 months may be difficult to pay back.โ
Tech giants argue that Capex are insignificant because their balance sheet can sustain such spending. They invest in AI defensively to maintain a competitive edge on a new technological wave. โWe had all these promoters coming into AI trying to say that the CapEx doesnโt matter, the budgets donโt matter. What you heard this week from Sundar Pichai (CEO at Google) and Mark Zuckerberg (CEO & cofounder at Meta) is that they understand this is risky. Itโs a risk worth taking, and they have to take it.โ โTheyโre making a strategic calculation: if we donโt make this investment, our competitors are going to get ahead.โ
Tech giants form a powerful oligopoly and will go to great lengths to protect their dominance. โ[Tech giants (Meta, Microsoft, Amazon, Google) are representing] one of the most powerful oligopolies in the history of business that weโre dealing with. Theyโre going to be willing to spend aggressively to protect their oligopoly.โ
Having tech giants investing $600bn in AI Capex is fantastic for startups. โThese big tech companies are producers of compute; startups are consumers of compute. If you believe there is an overproduction of compute and compute prices go down, startups win because they are buying those compute resources at lower costs. Lower cost of compute directly translates to higher gross margin for startups, higher gross margin translates directly to more valuable companies.โ
Tech giants are protecting their cloud golden goose by investing massive Capex in AI. โTech giants are too powerful; theyโre erecting barriers to entry right now. In order to be an AI cloud, you need to be willing to light a bunch of money on fire. That is a pretty significant barrier to entry for new entrants. What you heard from the commentary this week is thatโs not an accident. Thatโs pretty explicitly what theyโre trying to do: protect their golden goose. The cloud business today is a $250 billion business, the same size as the SaaS sector. So of course theyโre going to do everything they can to protect it.โ
โIf models get better and scaling laws hold, we probably need to change the architecture of the data center and new chips.โ
โAmazon announced $50bn of new data centers in the last six months. These data centers are getting built, and they have to hire people to build them.โ
โThe question comes down to whether researchers will come up with research breakthroughs faster than we can build bigger data centers. There are a lot of people in Silicon Valley who say all that matters now is scale. Others believe weโre going to make a breakthrough in reasoning, data usage, and model efficiency. The evidence today favors scaling laws, but we hope for breakthroughs in other areas to drive progress.โ
You need to make your product more useful with AI to be able to charge for AI additional features. โIf you are simply injecting AI to tell investors and for marketing purposes without actual benefit to the consumer, then youโre just adding cost with no benefit, and consumers wonโt pay more.โ
โIn businesses with barriers to entry, you have pricing power. In businesses with low barriers to entry, you donโt. It will be sector by sector. Some sectors will get commoditized and bid down to zero percent gross margin. In industries with structural low gross margins, itโs hard to raise the gross margin. In industries with higher barriers to entry, like those with data moats, those companies have pricing power. If you had pricing power before AI, youโll have it after AI. If you didnโt have it before, you probably wonโt after.โ
โVertical integration between the model layer and the data center matters. Meta, for example, controls its own data centers and is building the models. The vertical integration piece is going to matter as models get bigger and bigger. Separate teams running the data center and building the model wonโt work.โ
โThe evidence of the last couple of years is that in order to play in the big model game, you need a cash machine that cannot be the AI business. Facebook has Instagram, Amazon has AWS, Microsoft has Azure. You need a cash machine to compete.โ
โThe cloud guys are playing defense; Meta is playing offense. Meta can afford to be creative and play for the future, whereas the cloud companies are stuck in a prisonersโ dilemma and have to keep investing.โ
โThe consumer killer use case [for Meta] hasnโt emerged yet, but they have fantastic distribution. If they can unlock that killer use case, theyโll figure out how to monetize it.โ
โAlex Wang from Scale said there are three things that matter on model and AI progression: compute, algorithms, and data. Do you agree, and which do you think is the core constraint?โ โI think compute, models, and data have converged. All the big guys know what theyโre doing and are doing the same thing. Everyoneโs using Scale, and itโs hard to argue that any of the big model companies has a data advantage. Compute is a commodity you pay for, and models have some secret sauce, but scaling laws are what matter.โ โServers, steel, and power are the three things that matter. Servers include chip innovation, steel includes construction, and power includes the energy revolution. The industrial nature of AI is driving these changes.โ
โThe lesson from those companies is to listen to what people do, not what they say. When I called customers on Marqeta, they said itโs too expensive and theyโll build it themselves, but they never churned. The same with UiPath and Snowflake. Customers might say theyโre going to churn, but if theyโre paying for it, itโs for a reason.โ
โThe bar for success is so high that unless the company [you picked as one of the 1-2 investments you do per year] is one of the 50 most important companies in the world, you havenโt done your job well.โ
Friday, Aug. 9th: I listened to a No Priors podcast episode with Elad Gil and Sarah Guo. - No Priors
โDaron Acemoglu (MIT professor) argues that AI will impact less than 5% of all tasks and that the trillion dollars spent on training models is a waste because AI wonโt solve complex problems.โ
โJim Covello (global equity research at GS) contrasts AI with the internet, stating that while the internet disrupted expensive things from the beginning, AI is expensive from the start, and any efficiency gains will be competed away, meaning companies wonโt benefit in the long run.โ
โThereโs a broader misconception about how this technology works and its implications.โ โMany people are stuck in the old AI world, treating it like traditional machine learning without realizing the breakthroughs in Transformer-based models and other architectures that provide different functionalities and features.โ
Enterprise adoption is still in its early days. Elad breaks down AI adoption into three segments: (i) AI vendor tools that I can purchase, (ii) internal AI tools and (iii) customer centric AI tools. Everything that has happened today is mostly AI vendor tools that I can purchase. The big wave of enterprise adoption is still ahead of us.
โAI is opening up markets in unexpected ways through new capabilities, easy access via APIs (no need to do ML ops and ML trainings), and mandates for AI adoption (many companies have a mandate to do AI). These drivers make markets more accessible than theyโve been in decades.โ
On AI-driven buyouts:
โThere are situations where you actually want to do an AI driven buyout because it's better to have somebody take over the asset and dramatically change the cost structure or the leverage per employee using AI.โ
โToday, there are really odd things you can suddenly do just because the buying behavior has shifted. Everybody has an AI mandate and they're willing to consider products that they wouldn't be able to consider without AI.โ
โWith AI driven buyouts, you are shortcutting the change management process for industries that can be like dramatically automated with these new capabilities by just controlling them.โ
โYou are short cutting adoption of technology which in some industries may just not happen very fast. You are short cutting change management. But you are also taking over an asset and you are able to entirely change how a subset of the organisation functions relative to AI (cf. do something similar to what Klarna did in CS). You can cut headcount while drastically improved the quality of the function. There are a lot of industries where a lot of the costs are associated or leverage on time.โ
โYou need both PE expertise and tech-savvy to make AI-driven buyouts work. The early tech-enabled buyouts often failed because they added only a thin software veneer. Successful examples now involve deep integration of AI, significantly altering operations and leveraging AI more than just financial engineering.โ
โThe traditional enabled buyout playbook which largely didn't work is based on a PE person come in and buying a bunch of companies to roll them up and add a thin layer of software on top of it to claim that you have now a tech company.โ Itโs just an arbitrage play on the difference between a single asset and a platform with multiple assets as well as the difference between a PE business and a tech business.
On incubation:
โIncubation makes sense when you have visibility into an opportunity set or a customer set that others do not. Moreover, the expertise set required to make a company work is extremely hard to get organically.โ
โThere's a handful of firms that have actually done very good incubations because they are vertically specialised. They understand the dynamics of their industry. They have proprietary access to customers and product ideas that will take off in their industry.โ
โWith AI the odd thing that's happening is that certain subsets of the market are incredibly crowded with a dozen of companies doing the same thing. Besides, you have these areas where it's really obvious to build stuff and they're just wide open and nobody's doing anything.โ
In the public markets, Elad Gil would focus on investing in 3 categories of companies: (i) those poised to become the next Magnificent 7 compounding high growth over a decade or more (ii) established big tech firms that will effectively capitalize on the AI wave (e.g. Apple replacing all phones thanks to an AI powered phone)? (iii) companies where adding AI offers limited product improvement, making them defensible.
Saturday, Aug. 10th: Ben Evans wrote about the current hype surrounding Gen. AI. He draws parallels to past technology cycles, suggesting that LLMs, much like the Dotcom bubble or early cloud computing, may be experiencing an inflated hype phase. While these models appear to be finished products ready for wide deployment, they lack the refinement necessary for broader, practical applications. - Ben Evans
โThe future can take a while - it took more than 20 years for 20% of US retail to move online. People forget this now, but the iPhone took time as well. The same, of course, applies to the enterprise. If you work in tech, cloud is old and boring and done, but itโs still only a third or so of enterprise workflows 25 year after Marc Benioff tried to persuade people to do software in the browser.โ
โIt does take time to change your habits and ways of thinking around an entirely new kind of tool. Iโve also argued that an LLM by itself is not a product - itโs a technology that can enable a tool or a feature, and it needs to be unbundled or rebundled into new framings, UX and tools to be become useful. That takes even more time.โ
โLast summer Accenture proudly announced that it had already done $300m of โgenerative AIโ work for clientsโฆ and that it had done 300 projects. Even an LLM can divide 300 by 300 - thatโs a lot of pilots, not deployment.โ
โMorgan Stanleyโs latest CIO survey says that 30% of big company CIOs donโt expect to deploy anything before 2026.โ
Monday, Aug. 12th: I listened to a Colossusโ podcast episode with Martin Casado who is GP at a16z. - Colossus
Gen. AI will have a greater impact by driving an explosion of new consumer experiences rather than by solving enterprise automation problems.
โWhen it comes to content creation like any sort of digital creation, the marginal cost has gone to zero.โ
โJust like in the early days of the Internet, I don't think we know what that's going to look like.โ
โOne way to think about this current AI wave is the marginal cost of content creation is going to 0 and to some extent, natural language reasoning is going to 0.โ
โI believe that at least 1/3 of the money made by LLMs are from people that are just playing with a LLM-based app focused on creating an emotional connection (e.g. virtual friend, new style of video game, virtual girlfriend).โ
โA lot of those language reasoning tasks (e.g. analysing a contract with a LLM instead of a law firm) are going to go to these models because they do a pretty good job of the language reasoning piece.โ
โIf you actually look, the progress for AI has always been incredibly steady. It has always been up and to the right. We've solved many problems. We didn't think that you could solve.โ
โPeople use it and then they get bored and it drops off. They have to launch another model that goes up and it drops off. They have to launch another model. And so then you're in this long march, let's call it the waves where you're riding these waves of new models and later on it sucks.โโAt that point, that's when the companies have to find a traditional moat. There's a lot of moats, but that's when you have to create the moat and get off the treadmill. I think if you don't get off the treadmill, this all converges basically to the same model.โ
โI think it's amazing that Facebook releases LLama, but when you release open source without actually having an enterprise sales force and go-to-market engine around it, I think it kind of lands people to do whatever they want, but you don't ever push a product and pull it in the way that an enterprise buyer would want.โ
โMarket annealing is for category creation. If you're in a mature market, just go build something better, people already know how to use it. But in new markets, the customer doesn't know the value that you have, they haven't a clue. They don't know how to use it. They don't have the workflow. So there's a ton of work take you're actually hammering the market and making it so they can accept it.โ
โGen. AI is a new breakthrough, in my opinion. It's got its own set of properties. And those properties are amazing for certain problems and not amazing for other problems. I personally am very excited and attracted to the problem that gets a great fit for the emotion companionship. I think it's a great fit for the creativity problem. I think it's a great fit for the language reasoning problem, it's a great fit for those.โ
Tuesday, Aug. 13th: Microsoft faced a significant challenge when OpenAIโs CEO Sam Altman was unexpectedly fired, leading Satya Nadella to swiftly intervene to stabilize the situation. In response, Microsoft has diversified its AI investments, reducing reliance on OpenAI by partnering with other AI startups (inc. Mistral, Inflection and G42) and developing its own models. - FT
โFor Microsoft and its investors, the incident was a reminder of how central OpenAI had become to its strategy: the growth of artificial intelligence.โ
โIn the eight months since the board dispute, the tech giant has worked to execute an AI strategy independent of Altmanโs start-up. It has diversified its investments and partnerships in generative AI, built its own smaller, cheaper models, and hired aggressively to develop its consumer AI efforts.โ
โThe raft of departures mean that nine of the start-upโs 11 co-founders are currently not working there.โ
โMore controversially, the so-called โacqui-hireโ of Inflection founder Suleyman and most of the start-upโs staff in March set Microsoft on a path to confrontation with its biggest AI partner. The combative former Google DeepMind executive, who left that company having developed a reputation as a bully, was put in charge of a new internal AI unit at Microsoft and tasked with building consumer-facing products that would compete with those from Altmanโs OpenAI.โ
โThe US Federal Trade Commission is probing whether the Inflection deal was structured to circumvent antitrust laws, essentially gutting the smaller company of talent and software, while avoiding the formal scrutiny a full takeover would have brought.โ
โDespite the scrutiny, the Inflection deal has become a model for other tech giants seeking talent. In June, Amazon hired most of the staff at AI-agent start-up Adept and paid $330mn to license its intellectual property. Last week, Google rehired the founder of chatbot maker Character.ai and paid more than $2bn to license its technology and cash out existing investors.โ
โAzure is OpenAIโs exclusive cloud provider and will be paid for whatever computing power it uses. Microsoft is also agnostic about which AI models are used, so long as they are accessed through its cloud.โ
โWe have over 1,600 models available through Azure AIโ.โ.โ.โthe main thing we want is people to be building and using them on Azure.โ
โAzure AI now had 60,000 customers, up more than 60 per cent from a year ago.โ
Wednesday, Aug. 14th: I read Sequoiaโs multiple blog posts on AI. - David Cahn 1, David Cahn 2, David Cahn 3, Sonya Huang
โHow much of this CapEx build out is linked to true end-customer demand, and how much of it is being built in anticipation of future end-customer demand?โ
โDuring historical technology cycles, overbuilding of infrastructure has often incinerated capital, while at the same time unleashing future innovation by bringing down the marginal cost of new product development.โ
โI noticed a big gap between the revenue expectations implied by the AI infrastructure build-out, and actual revenue growth in the AI ecosystem, which is also a proxy for end-user value.โ
โThe supply shortage has subsided:ย Late 2023 was the peak of the GPU supply shortage. Startups were calling VCs, calling anyone that would talk to them, asking for help getting access to GPUs. Today, that concern has been almost entirely eliminated. For most people I speak with, itโs relatively easy to get GPUs now with reasonable lead times.โ
โNvidia reported in Q4 that about half of its data center revenue came from the large cloud providers. Microsoft alone likely represented approximatelyย 22% of Nvidiaโs Q4 revenue. Hyperscale CapEx is reaching historic levels. These investments were a major theme of Big Tech Q1 โ24 earnings, with CEOs effectively telling the market: โWeโre going to invest in GPUs whether you like it or not.โ Stockpiling hardware is not a new phenomenon, and the catalyst for a reset will be once the stockpiles are large enough that demand decreases.โ
โThe Information recently reported that OpenAIโs revenue is nowย $3.4B, up from $1.6B in late 2023.โ โOutside of ChatGPT, how many AI products are consumers really using today? Consider how much value you get from Netflix for $15.49/month or Spotify for $11.99. Long term, AI companies will need to deliver significant value for consumers to continue opening their wallets.โ
โEarlier this year, Nvidia announced their B100 chip, which will haveย 2.5x better performanceย for only 25% more cost. I expect this will lead to a final surge in demand for NVDA chips. The B100 represents a dramatic cost vs. performance improvement over the H100, and there will likely be yet another supply shortage as everyone tries to get their hands on B100s later this year.โ
โGPU computing is increasingly turning into a commodity, metered per hour. Unlike the CPU cloud, which became an oligopoly, new entrants building dedicated AI clouds continue to flood the market. Without a monopoly or oligopoly, high fixed cost + low marginal cost businesses almost always see prices competed down to marginal cost (e.g., airlines).โ
โPerhaps most worryingly, a whisper began to spread within Silicon Valley that generative AI was not actually useful. The products were falling far short of expectations, as evidenced by terrible user retention.โ
โThese applications are different in nature than the first apps out of the gate. They tend to use foundation models as a piece of a more comprehensive solution rather than the entire solution. They introduce new editing interfaces, making the workflows stickier and the outputs better. They are often multi-modal.โ
โWe still believe that there will be a separation between the โapplication layerโ companies and foundation model providers, with model companies specializing in scale and research and application layer companies specializing in product and UI. In reality, that separation hasnโt cleanly happened yet. In fact, the most successful user-facing applications out of the gate have been vertically integrated.โ
โThe first killer apps emerged. Itโs been well documented that ChatGPT was the fastest application to reach 100M MAUโand it did so organically in just 6 weeks. By contrast, Instagram took 2.5 years, WhatsApp took 3.5 years, and YouTube and Facebook took 4 years to reach that level of user demand. But ChatGPT is not an isolated phenomenon. The depth of engagement of Character AI (2 hour average session time), the productivity benefits of Github Copilot (55% more efficient), and the monetization path of Midjourney (hundreds of millions of dollars in revenue) all suggest that the first cohort of killer apps has arrived.โ
โGenerative AIโs biggest problem is not finding use cases or demand or distribution, it is proving value.โ
โAmaraโs Lawโthe phenomenon that we tend to overestimate the effect of a technology in the short run and underestimate the effect in the long runโis running its course.โ
โModel size is the number one driver of performance. As a result, the next generation of models are aiming for a 10x increase in model scale to 300k GPUs. To house one of these models, you need to build an entire new data center. This changes AI in two fundamental ways: First, it changes the lead time between models. If before you could train your model in 6 to 12 months, now you need to add 18 to 24 months of construction time before you can actually start training. Second, it changes the source of maximum competitive advantage. In the new era, construction efficiency may matter more than research breakthroughs.โ
โToday, five companies have arrived at the starting line in this new race toward data center scale-up: Microsoft/OpenAI, Amazon/Anthropic, Google, Meta and xAI.โ
Thursday, Aug. 15th: I listened to a Lex Fridmanโs podcast episode with Yann Lecun. - Lex Fridman
Autoregressive LLMs are not the way weโre going to make progress towards superhuman intelligence for 2 main reasons:
โ(i) The capacity to understand the physical world, (ii) the ability to remember and retrieve things, persistent memory, (iii) the ability to reason, and (iv) the ability to plan. Those are four essential characteristics of intelligent systems or entities, humans, animals. LLMs can do none of those or they can only do them in a very primitive way and they donโt really understand the physical world.โ
โIt seems like an enormous amount of knowledge that LLMs can accumulate, but then you realize itโs really not that much data. [โฆ] Through sensory input, we see a lot more information than we do through language, and that despite our intuition, most of what we learn and most of our knowledge is through our observation and interaction with the real world, not through language.โ
Moravec paradox. โWe have LLMs that can pass the bar exam, but then they canโt learn to drive in 20 hours like any 17-year old, they canโt learn to clear out the dinner table and fill up the dishwasher like any 10-year old can learn in one shot.โ
โThe way LLMs are trained is that you take a piece of text, you remove some of the words in that text, you mask them, you replace them by blank markers, and you train a genetic neural net to predict the words that are missing. If you build this neural net in a particular way so that it can only look at words that are to the left or the one itโs trying to predict, then what you have is a system that basically is trying to predict the next word in a text.โ
โA LLM doesnโt think about its answer really. It retrieves it because itโs accumulated a lot of knowledge. So it can retrieve some things, but itโs going to just spit out one token after the other without planning the answer.โ
โThe world is incredibly more complicated and richer in terms of information than text. Text is discrete, video is high-dimensional and continuous.โ
โAll the systems that attempt to predict missing parts of an image or video from a corrupted version of it, basically, so take an image or a video, corrupt it or transform it in some way, and then try to reconstruct the complete video or image from the corrupted version, and then hope that internally, the system will develop good representations of images that you can use for object recognition, segmentation, whatever it is. That has been essentially a complete failure and it works really well for text.โ
About JEPA (Joint-Embedding Predictive Architecture).
โInstead of training a system to encode the image and then training it to reconstruct the full image from a corrupted version, you take the full image, you take the corrupted or transformed version, you run them both through encoders. Then you train a predictor on top of those encoders to predict the representation of the full input from the representation of the corrupted one.โ
โThe big question is how do you train something like this? And until five years ago or six years ago, we didnโt have particularly good answers for how you train those things.โ
โWhat has changed in the last three, four years is now we have methods that are non-contrastive. So they donโt require those negative contrastive samples of images that we know are different. You turn them on you with images that are different versions or different views of the same thing, and you rely on some other tricks to prevent the system from collapsing. And we have half a dozen different methods for this now.โ
Whatโs the difference between JEPA and generative architectures like LLMs?
โLLMs or vision systems that are trained by reconstruction generate the inputs. They generate the original input that is non-corrupted, non-transformed, so you have to predict all the pixels, and there is a huge amount of resources spent in the system to actually predict all those pixels, all the details.โ
โIn a JEPA, youโre not trying to predict all the pixels, youโre only trying to predict an abstract representation of the inputs. And thatโs much easier in many ways.โ
โWhat the JEPA system, when itโs being trained, is trying to do is extract as much information as possible from the input, but yet only extract information that is relatively easily predictable.โ
โIt allows the system to essentially learn an abstract representation of the world where what can be modeled and predicted is preserved and the rest is viewed as noise and eliminated by the encoder.โ
โIt lifts the level of abstraction of the representation. If you think about this, this is something we do absolutely all the time.โ
โWe were lazy with language because we already got the abstract representation for free.โ
LLMs.
LLMs are exploiting self supervised learnings and LLMs are a good demonstration that they actually work.
โIt allowed us to create systems that understand language, systems that can translate hundreds of languages in any direction, produce summaries, answer questions and produce text.โ
โThe auto regressive trick where you constrain the system to not elaborate a representation of the text from looking at the entire text, but only predicting a word from the words that are come before.โ
โWe just assume that if a system is fluent in manipulating language, then it has all the characteristics of human intelligence, but that impression is false. Weโre really fooled by it.โ
โLLMs are impressive. They can do a lot of useful things. Thereโs a whole industry that is being built around them. Theyโre going to make progress, but there is a lot of things they cannot do, and we have to realize what they cannot do and then figure out how we get there.โ
โThereโs ample evidence that weโre not going to be able to learn good representations of the real world using generative model. So Iโm telling people, everybodyโs talking about generative AI: โif youโre really interested in human level AI, abandon the idea of generative AI.โโ
AI hallucinations
โWhy hallucinations happen from LLMs and to what degree is that a fundamental flaw of LLMs?โ โBecause of the autoregressive prediction, every time a LLM produces a token or a word, there is some level of probability for that word to take you out of the set of reasonable answers.โโIf you assume, that the probability of such error is that those errors are independent across a sequence of tokens being produced. What that means is that every time you produce a token, the probability that you stay within the set of correct answer decreases and it decreases exponentially.โ
โThe way you can correct for this is that you fine tune the system by having it produce answers for all kinds of questions that people might come up with.โ โPeople are people, so a lot of the questions that they have are very similar to each other, so you can probably cover 80% or whatever of questions that people will ask by collecting data and then you fine tune the system to produce good answers for all of those things.โ
โThere is the enormous set of prompts that you have not covered during training.โ
โThereโs a very, very long tail of things that people will ask and you can fine tune the system for the 80% or whatever of the things that most people will ask. And then this long tail is so large that youโre not going to be able to fine tune the system for all the conditions.โ โIn the end, the system ends up being a giant lookup table essentially, which is not really what you want, you want systems that can reason, certainly that can plan.โ
Reasoning in AI
โThe type of reasoning that takes place in LLM is very, very primitive, and the reason you can tell is primitive is because the amount of computation that is spent per token produced is constant.โ
โThis is not the way we work, the way we reason is that when weโre faced with a complex problem or a complex question, we spend more time trying to solve it and answer it because itโs more difficult.โ
โItโs the same difference as the difference between what psychologists call system one and system two in humans. Thatโs system one, so all the things that you do instinctively without really having to deliberately plan and think about it. There is all the tasks where you need to plan. this type of deliberate planning, which uses your internal world model, thatโs system two, this is what LMS currently cannot do.โ
โItโs not going to be a regressive prediction of tokens, itโs going to be more something akin to inference of little variables in what used to be called probabilistic models or graphical models.โ
โAn energy based model can be used to measure to what extent an answer is a good answer for a prompt.โ โitโs much better to do an optimization in continuous space where you can do gradient and descent as opposed to generate tons of things and then select the best, you just iteratively refine your answer to go towards the best, thatโs much more efficient.
Reinforcement Learning with Human Feedback
โI think itโs use should be minimized because itโs incredibly inefficient in terms of samples.โ
โThe way to adjust your word model even in advance is to explore parts of the space where you know that your world model is inaccurate, thatโs called curiosity basically, or play.โ
Friday, Aug. 16th: I listened to Bill Gurley and Brad Gerstner on AI and on the public market volatility. - BG2
โIf we put this in context, since Q1 of 2023, stocks are up 60% after everyone was nervous about a hard landing. Thatโs a huge move in just six quarters, and some stocks like Nvidia have more than doubled. So, whatโs happened this week and last month is more of a run-of-the-mill, healthy consolidation of these big gains.โ
โAs prices go up, risk-reward gets worse unless numbers keep going up, and that hasnโt been the case in 2024. Weโve seen five key forces: taxes, earnings misses or just skinny beats, the potential AI air pocket, and recently, two more emergedโgeopolitical factors like the Japanese carry trade unwind and expanding tensions in the Middle East. These create a toxic brew of uncertainty, which raises the discount rate and lowers multiples on stocks.โ
โWhen you have negative rates, it leads to a carry tradeโborrowing in yen and investing in rate-yielding assets around the world, pocketing the spread. But when the Bank of Japan raised rates, people had to unwind these trades, which meant selling U.S. bonds and stocks around the world, adding to the marketโs volatility.โ
โThe first thing people do when things get tight is defer buying a couchโthatโs what Wayfair is seeing. The second thing is they defer their family trip. When Airbnb, Disney, the cruise lines, and airlines start seeing downturns, you know the consumer is really getting pressured.โ
โIf you look at the S&P, 80% of companies have now reported, and 48% of them had a positive revenue beat. That sounds like a lot, but itโs the lowest level since Q3 2019.โ
โWeโre facing a different tax situation in this election compared to prior ones. The corporate tax cuts were made permanent, but individual tax cuts are set to expire in 2025. If a new administration comes in and those cuts expire, it could be a big risk to the market. The personal income tax reductions from the Trump tax cuts, if reversed, would pull $150bn out of the economy, which could be a significant headwind for the markets.โ
โIf the risks weโre discussing materializeโrecession, geopolitical events, etc.โthe market could correct by another 10-15%, bringing us back to where we were on January 1. Investors need to ask themselves if the world is really that much better than it was at the beginning of the year to justify the current prices.โ
About the recent trend of โtake-unders,โ where companies like Inflection and Character AI are being acquired in non-standard deals.
โFounders like Noam Shazeer at Character.AI might be looking at the daunting amount of capital required to compete with Meta and Google. On the other hand, big tech companies are willing to pay big money for talent but canโt do outright acquisitions due to regulatory concerns. So, theyโre finding creative ways to structure these deals. Iโm not sure how they work exactly, but it seems like investors are earning 2-3x on their money, which is interesting.โ
โThese take-unders seem to be happening to avoid federal regulatory scrutiny. These deals are structured in a way that might involve double taxation risks, and the entity often stays around even after the core team has left. Itโs not clear whether these deals are preferable for venture investors, but theyโre certainly raising eyebrows.โ
Saturday, Aug. 17th: I listened to a Building with AIโs podcast episode with Tomasz Tunguz who is founder and GP at Theory Ventures. - Building with AI
โIโm the founder and general partner of a firm called Theory Ventures. We are a team of nine people focused on data, investing in three different categories: the modern data stack, applications and infrastructure for AI, and blockchains, which we view as databases.โ
โCompared to previous technological waves, there are two major differences with Gen. AI. First, incumbents have an advantage. Startups have had negative time to launch because companies like Google already have 60% of all the PhDs, all the data, and distribution. Second, unlike the internet or mobile apps, which were easier to build and implement, AI is non-deterministic and chaotic. This unpredictability changes product design and the sales process because buyers now have to consider that the AI might behave unpredictably.โ
โWeโll see two waves of AI application development. Weโre in the first peak, where large language models are viewed as the solution itself. But weโre realizing that LLMs alone arenโt enoughโwe need security systems, monitoring systems, classifiers, and ways to eliminate errors and hallucinations to make these systems truly effective.โ
โHistorically those ML systems have been for post hoc analysis (e.g. doing a churn prediction and then using that to inform how the customer success team operates) but increasingly those that data is now being used in production systems (e.g. Netflixโs content recommendation system).โ
โLLMs are unpredictable across versions, with up to 40% of prompts breaking when upgrading from one version to another. This unpredictability makes it difficult to guarantee a consistent user experience. The closest analogy might be search engine quality at Google. Like search engines, LLMs require both machine and human evaluations to assess quality. Every company will need its own set of evaluations, which are highly context-specific. In the future, we may see analytics vendors that create synthetic evaluation tests based on how users engage with AI. This feedback loop between analytics and evaluations will be crucial, but itโs still early days.โ
โTypically, new machine learning innovations start at about 75% accuracy and take 10 to 15 years to improve to 95-98%. Thereโs always an initial over-promise, a collapse, and then a comeback. The key is to match the use case to the risk. For instance, consumers love AI for tasks like generating images, where the range of acceptable outcomes is broad, and the cost of being wrong is low. But it would be a mistake for finance or healthcare professionals to ignore these technologies.โ
โWhat is the framework to identify suitable use cases for LLMs?
Letโs structure this in three ways. First, consider implementation: thereโs co-pilot (sentence completion) and agents (full task automation). Sentence completion can accelerate work by 50% without requiring a high level of trust. Task automation is still early and carries more risk.
Second, consider accuracy. If the system is 75% accurate, is that valuable? In creative tasks like writing, an 80% solution is useful because itโs easier to edit than create. But in areas like credit scoring, 75% accuracy is useless.
Finally, look for rote toilโrepetitive, unattractive work that no one wants to do, especially in industries with demographic shifts or labor shortages. For example, long-haul trucking, BDR roles, and security analysis are all areas ripe for AI automation. The key is finding tasks where an 80% solution is sufficient and the work is highly repetitive.โ
โThe first industries were computer software (like software writing with Copilot), legal, and marketing copy. Now, weโre seeing it in financial analysis, accounting, and security. These are the areas where weโre seeing the most innovation and investment.โ
โMicrosoft is doing an excellent job. Theyโve built a $5bn AI business in 18 months by infusing AI into all their products.โ
โIn the next wave, weโll see a linkage between software and hardware. These models are expensive to run, but your laptop or phone has a powerful GPU. Pushing models to the edge (on-device) will give companies with both software and hardware a huge advantage.โ
โItโs also important for incumbents to support as many model families as possible. Microsoft supports 50 different models, which is crucial because enterprises are starting to build on Azure, perceiving it as the place where the best models are available first. Supporting a broad range of models and fostering strategic relationships with LLM creators will be key.โ
"Gartner published a study showing that customer acquisition costs for software companies have increased 60% over the last five years. This means a small but compounding advantage in customer acquisition costs can give you disproportionate market share.โ
โIf youโre a startup looking to disrupt a specific industry with an AI-native approach, how would you set up to maximize your odds of unseating an incumbent? The next generation of startups needs to offer more than just an AI feature. You need to create some form of intellectual property or competitive advantage thatโs unique, whether itโs proprietary data, better revenue generation, or cost reduction.โ
โInnovation equals invention plus distribution. As the cost of invention goes to zero due to generative AI, distribution becomes much more important.โ
Sunday, Aug. 18th: I listened to a More or Lessโ podcast episode with Vinod Khosla on AI. - More or Less
OpenAIโs partnership with Apple:
โAppleโs AI integration demonstrates that computers will learn humans instead of humans needing to learn computers.โ
โObviously, from an OpenAI perspective, it clearly establishes where OpenAI is in the competition for the best LLMs. Many people wanted that business, and many things were evaluated. Itโs rewarding to see where Apple thinks the best AI is.โ
โI think the bigger thing is how we interface with computers, and Siri is the beginning of what will evolve over time into agentic interfaces for humans.โ
โApple can do a lot on-device with small models.โ โI think what on-device small models do will be different from what the large models do.โ โIf you want human-level intelligence from a model, itโll be a big model. But I donโt think small models substitute for big models. They do some things really well, though. If youโre talking to something with very low latency, you want a shorter path, so thatโd be a small model on-device.โ
โMy advice to all our startups is to ignore the cost of compute because any assumption you make, any dollar you spend on optimizing your software, will be worthless within the year. So forget about it. Rely on competition in the marketplace between Claude, Gemini, and OpenAI to reduce that cost to a point where it doesnโt matter. Below a certain level, it doesnโt matter. If itโs 10% or less of that, it doesnโt matter.โ
โIt is costing an order of magnitude more to train, and thatโs why I donโt think open-source models will be viable because of the training costs. But once youโre trained, you want as broad a set of usage as possible for two reasons. One, you want the most revenue from it, and the lowest-cost models will get the most revenue. But more importantly, thereโs huge data generation for training the next generation of the model.โ
โIโm pretty convinced that in the next few years, revenue isnโt the important metric [for LLMs]. Of course, you donโt want to lose so much money you canโt afford it, but youโre not trying to make tons of money. Youโre trying to get tons of usage and tons of data from that usage to learn to be a better model.โ
โI do think thereโs a lot to be gained in intelligence [for LLMs]โwhether itโs reasoning, probabilistic thinking, or pattern matching. These models have lots of headroom to get better, and I think weโll see stunning advances practically every year.โ
โIf youโre building foundation models and trying to compete with OpenAI and Google, thatโs not a great place to be. LLMs will belong to the large players.โ
โI donโt think weโre anywhere near done with fundamental techniques. Weโre exploiting the Transformer model more and more, but there are other models to be developed. Itโs just that everybodyโs scared to invest in anything other than Transformer models. Iโm very focused on esoteric things.โ
โAt the next level up, the applications, thereโs huge opportunity. One of my big forecasts is that almost all expertise will be free. Whether youโre talking about a primary care doctor, a teacher, a structural engineer, or an oncologist, there are so many areas of expertiseโhundreds, if not thousandsโeach of which will be a very successful venture.โ
How to build a defensible business in the Gen. AI world? โThe idea that one model will do everything is incorrect. Workflows differ in every area.โ
โThereโs room for seed investments, but following what everyone else is doing is a bad strategy. When somebody leaves DeepMind and gets $2 million just because, thatโs not a good strategy.โ
โWe spend a lot of time evaluating whether a business could be railroaded by large models or become obsolete as a feature. About 80% of our seed investments are below a $25m post-money valuation. Occasionally, we step up to something like Adept, even with zero revenue, because we have a certain belief system. About 80-90% of the billion-dollar valuations we pass on, but thereโs 10-20% weโll invest in, like Devon or Replit. Occasionally, weโll make small, token investments for relationships, but weโre very disciplined about it.โ
AI will create a deflationary economy. โIt will be a hugely deflationary economy. If you take a million doctors and reduce them to just 200k and 80% has gone then you're talking about $250bn less spent on healthcare even if you can increase healthcareโs usage.โ
Monday, Aug. 19th: Carta published a report on VC fund performance with data from 1.8k US funds. - Carta
Funds are deploying capital at a slower pace. โFunds in the 2022 vintage have deployed about 43% of their committed capital at the 24 month mark, the lowest share of any analyzed vintage. Prior vintages ranged from 47%-60% after 24 months.โ
Graduation rates are declining. โ30.6% of companies that raised a seed round in Q1 2018 made it to Series A within two years. Only 15.4% of Q1 2022 seed startups did so in the same timeframe.โ
Tuesday, Aug. 20th: The Economist wrote about investors becoming increasingly skeptical about the profitability of AI. - The Economist
โA growing number of investors worry that AI will not deliver the vast profits they seek.โ
โA growing number of observers now question the limitations of LLMs, which power services such as ChatGPT.โ
โAccording to the latest data from the Census Bureau, only 4.8% of American companies useย AIย to produce goods and services, down from a high of 5.4% early this year.โ
โโThe future ofย AIย is just going to be like every other technology. Thereโll be a giant expensive build-out of infrastructure, followed by a huge bust when people realise they donโt really know how to useย AIย productively, followed by a slow revival as they figure it out,โ says Noah Smith, an economics commentator.โ
The โhype cycle,โ a concept from Gartner, suggests that new technologies go through phases of euphoria, disillusionment, and eventual mainstream adoption, but the applicability of this cycle to AI is debatable. Historical examples like railways and the internet show how initial hype led to massive infrastructure build-outs and eventual widespread adoption, but many technologies have not followed this pattern. Data from Gartner and additional analysis suggest that the hype cycle is rare, with only a fifth of breakthrough technologies following this path, and many never recovering after initial enthusiasm fades.
Wednesday, Aug. 21st: I read a portrait on Palmer Lucky who is Andurilโs cofounder and CEO. - Tablemag
After creating Oculus Rift, Palmer Luckey sold the company to Facebook for $2bn but was later fired due to controversial political donations. Palmer Luckeyโs current company, Anduril, develops advanced defense technologies like autonomous drones and AI-powered surveillance systems, now valued at $14bn.
โYou canโt make everyone happy, and if you try, youโll end up with a bland, meaningless existence.โ
โThe future of national security isnโt about building more tanks and planes; itโs about leveraging AI, autonomy, and technology that can outthink the enemy.โ
โDisruption isnโt just about breaking things; itโs about building something better in their place.โ
โInnovation happens when youโre willing to break the rules and challenge the status quo.โ
Thursday, Aug. 22nd: I read Contrary Researchโs deep-dive on Mistral. - Contrary
โMistral AI developsย open-sourceย foundation models. These are an alternative to closed, proprietary AI models like those developed byย OpenAI, Google, and Anthropic. The company also focuses on developing more efficient and cost-effective models, targeting the โperformance-cost frontier.โ
โThe company also offers โoptimizedโย versions of its AI models via its developer platform, for which it chargesย following a usage-based business model.โ
โLa Plateformeย is Mistral AIโs developer platform. This platform leases โoptimizedโ versions of the companyโs models to developers throughย generative endpoints, accessible via an API. Its goal is toย offerย efficient deployment and tailored customization for various use cases.โ
โLe Chat is Mistral AIโs chatbot service, functionally equivalent to OpenAIโs ChatGPT but powered by Mistral AIโs foundation models. Mistral AI also offersย Le Chat Enterprise, a service designed for businesses to โempower your teamโs productivity with self-deployment capacities, and fine-grained moderation mechanisms.โโ
โAs of April 2024, notable Mistral AI customersย includeย Lamini, Arcane, Lindy, Hugging Face, Brave, Cloudflare, Pretto, BNP Paribas, Orange, and MongoDB.โ
Friday, Aug. 23rd: I read a portrait on Alex Karp who is the CEO at Palantir. - NYT, Arny Trezzi
Alex Karp, CEO of Palantir, is an unconventional tech leader with a deep philosophical streak, often contemplating the moral implications of technology.
Palantir, under Karpโs leadership, has grown into a controversial powerhouse, known for its work with government agencies, including the military and intelligence sectors.
Karpโs approach to leadership and technology is marked by a blend of skepticism toward Silicon Valley norms and a commitment to addressing complex global issues.
โIโm not running this company to be loved. Iโm running it to keep the West safe.โ
โIf you want to be liked, this is the wrong place. If you want to solve complex problems, welcome aboard.โ
โI donโt fit the Silicon Valley mold, but maybe thatโs why Palantir works.โ
โIโd rather be misunderstood for doing the right thing than celebrated for following the crowd.โ
โWeโre in the business of providing clarity in a world full of chaos.โ
โThe only solution to stop A.I. abuse is to use A.I.โ
โIn some ways, Alex doesnโt look like a salesperson from central casting you would send to the C.I.A. The formulation I always have is that if youโre trying to sell something to somebody, the basic paradox is you have to be just like them, so they can trust you โ but you have to be very different from them so that they think you have something they donโt have.โ
Saturday, Aug. 24th: Bessemer published its annual Cloud 100 benchmark. - Bessemer
โThe Cloud 100 list value reached $820bn this year, a +25% YoY increase and the highest list value ever in Cloud 100 history (compared with the previous record of $738bn set by the 2022 Cloud 100 cohort).โ
โEven as companies have faced a challenging macroeconomic environment in the past few years, the time it took the average Cloud 100 company to reachย Centaur statusโcrossing $100m ARRโhas decreased consistently since the listโs inception.ย It took the average 2024 Cloud 100 company under eight years to reach Centaur status,ย down from ten years for the inaugural 2016 Cloud 100 cohort. AI companies in particular have shown their ability to scale faster than others, averaging just over six years for this yearโs cohort.โ
โ97% of honourees are forecasted to reach or surpass $100 million of ARR by the end of the year (vs. 95% in 2023).โ
Sunday, Aug. 25th: Matthew Ball wrote about Roblox. The company is evolving from a pure gaming platform to a broader virtual ecosystem integrating new technologies like AI and VR. - Matthew Ball
Robloxโs revenue grew to $2.2bn in 2023, reflecting a 16% YoY increase, highlighting its sustained growth.
The platform saw an increase in DAUs to 65.6m, marking an 18% rise compared to the previous year.
Users spent over 50bn hours on Roblox in 2023, a testament to its high engagement levels.
54% of Roblox users are now over 13, showing the platformโs appeal beyond younger audiences.
Roblox paid out $710m to its developers in 2023, underlining the growing creator economy within the platform.
โOn a monthly basis, Roblox now counts more than 380m users according to RTrack โ 2x as many as PC gaming leader Steam, 3x that of Sonyโs PlayStation, 3x the number of unique annual users of the Nintendo Switch in a year, and 5x as many as have bought an Xbox console in the last decade.โ
โObviously, Roblox has a costs problem. Over the last twelve months it has averaged $138 in costs for every $100 in revenue.โ
โRoblox is increasingly focused on using GenAI to provide real-time player communications (e.g., transcription and translation), as well as power asset and world creation, and operate AI agents and NPCs โ none of which is cheap to create or operate.โ
โGovernments around the world are now forcing Appleโs iOS platform and Googleโs Android to open up their app distribution and monetization models, at least in part (itโs also no coincidence that Epic Games, which operates a rival to Roblox, has been leading various legal efforts against both Apple and Google on exactly this issue). The specifics vary by market, by platform, and sometimes type of transaction. Thereโs not yet reliable data to suggest the degree to which Roblox might save on commissions (at a minimum, they will substitute a 30% fee with a 4% credit card processing expense, but additional iOS/Android fees might also apply), and the more complicated purchasing flow may require discounts and/or lead to fewer transactions, which would harm revenue. Accordingly, the net opportunity may be a reduction from 25% to 20% in the medium term, with a longer-term possibility of something closer to 10%.โ
โIn the average month, fewer than 6% of users and 17% of average DAUs purchase Robux.โ
โSmaller App Store payments, more ARPU, and a bigger ad platform are the linear opportunities for Roblox.โ
Monday, Aug. 26th: Alex Kolicich wrote about AI and the defense industry. - Alex Kolicich
โI've been contemplating how the unique optimism and rapid expansion of AI capabilities we've witnessed over the past year and a half are juxtaposed with danger and uncertainty in the investment world.โ
โAs I mentioned last time, AI valuations are fully priced, but the value capture remains uncertain.โ
โRevenue is being generated, but it's not sufficient. What's uniquely missing is gross margin and defensibility. Many companies inย AI are operating with low gross margins in highly competitive spaces relative to near-term demand.โ
โThe market, however, seems to price the view that cloud software companies will benefit from AI at nearly zero. I've shown this chart many times, but cloud software revenue multiples are trading close to near-term lows. This stands in contrast to a huge boom in AI stocks like NVIDIA, ASML, PLTR, etc.โ
โWhen we discussed investing in the Anduril Series A at 8VC in 2018, there was significant consternation in the tech ecosystem on doing work in defense. It was around this time that Google famously dumped their Project Maven DOD program due to employee unrest, which gives a sense of the strong sentiment against supporting defense innovation in Silicon Valley at the time.โ
โThe Department of Defense (DoD) has moved from often waiting twenty years for a single solution that fit the traditional mold to engaging with the country's most talented technology players in iterative processes.โ
Tuesday, Aug. 27th: Felicis wrote about the physical design software market. - Felicis
โShortly after, the PC was released as well, although it took time for the performance of PC semiconductors to support CAD software. This opened the door for early software companies to only develop software and not worry about hardware development.โ
โFrom 1981 to 1986, six of todayโs oligopolistic design companies were founded: Mentor Graphics (1981), Autodesk (1982), Cadence (1983), Bentley Systems (1984), PTC (1985), and Synopsys (1986). These companies made design software specifically for workstations and/or PCs, creating a cheaper and more flexible alternative to the incumbents.โ
โThe industry hasnโt seen major changes since the 1980s. Companies have switched pricing models and have started to migrate to cloud-based versions of products, but the industry structure has remained mostly constant.โ
โThe moats of the incumbents are so deep and the switching costs have such a high risk that this market has been mostly impenetrable.โ
Wednesday, Aug. 28th: Daniel Kimerling at Deciens wrote a post on alignment in venture capital. It critiques industryโs short-term focus where companies repeatedly raise and burn capital without achieving sustainable growth. It advocates for a long-term investment approach, aligning venture capital with foundersโ and LPsโ interests. The piece suggests that VC firms should prioritize building durable businesses and compounding value over time rather than chasing quick returns through constant fundraising cycles. True alignment and long-term partnerships are proposed as solutions to the current VC industryโs challenges. - Deciens
โBy the time all of the assets are distributed and funds wound down, a venture capital fundโs actual length may very well be 15-20 years.โ
โThe enduring nature of LP capital and the way that VC funds are structured are in stark contrast to the way that most VC-backed companies operate.โ
โThe crux of the problem is that from a first-principles perspective, founders and LPs should be aligned but seemingly arenโt.โ โLPs should want their (underlying) portfolio companies to compound their equity value at high rates of return for many years in tax-advantaged structures while minimizing their capital consumption and existential risk, ideally in an uncorrelated fashion to the rest of their portfolios.โ โFounders, with their lack of diversification, should want to operate their businesses in the same fashion, as it is the only sustainable way to build generational businesses and create associated wealth.โ
โThe behavior of most GPs and non-GP investors today suggests that they have all but stopped caring about generating carry.โ
โFrom where we sit, GPs must utilize the long-term nature of their capital partnerships to their advantage. If we have capital for around 15 years, then we should invest in and help founders build businesses that can compound for all 15 years. This is simply a more prudent, risk-adjusted strategy than burning brightly and hoping that every time a company rolls the fundraising dice, it comes up with double sixes.โ
Thursday, Aug. 29th: Sarah Tavel at Benchmark argues that competition in the development and deployment of LLMs is akin to a high-stakes poker game. - Sarah Tavel
โUnless the scaling hypothesis falters, this is a contest now of โnot blinking firstโ. If youโre a big stack player like Meta, Microsoft, Google, or any of the foundation model pure plays, you have no choice but to go all in โ the prize and power of โwinningโ is too great. If you blink, you are left empty handed, watching someone else count your chips. Itโs likely hundreds of billions will be destroyed, and trillions earned.โ
โMany of the foundation model companies talk about the amount of time it would take a human to do the work as a measure of the power of the LLM. If today LLMs can handle tasks that would have taken a human 30mins to complete, as LLMs progress, they'll be able to handle increasingly complicated tasks that would have taken a human more time.โ
โThe challenge is the amount of capital required to train each successively more sophisticated LLM increases by an order of magnitude, and once a model is leapfrogged by another, the pricing power of the older model quickly falls to zero.โ
โThe economic value you are able to unlock as models become more sophisticated should increaseย significantlyย with each upgrade of the model.โ
โThe prize is theoretically so large, and if a clear winner emerges, their market opportunity so uncapped, you have to keep increasing your bet.โ
Friday, Aug. 30th: I listened to a 20VCโs podcast episode with Dax Dasilva who is the cofounder and CEO at Lightspeed. - 20VC
As a young person, Dax was doing tech support in Appleโs dealerships in Canada. He started to build custom software for these dealerships. Appleโs dealerships were already in Lightspeedโs ICP being a retailer with complex inventory, with high value inventory, with multiple locations and with purchasing across multiple stores and in multiple currencies. At some point, he stopped building custom software for these dealerships and built a unique software for retailers to handle their operations. Lightspeed was born.
โFor the first two years, I coded until four in the morning.โ โWhat helped the product market fit happen is when I wasn't coding, I was with early customers watching them using the software and iterating quickly to adapt to their needs.โ โI wanted it to be physically there with the customer.โ
Lightspeed bootstrapped the business for 7 years before raising external capital. โWhen we did get money, we transitioned our product into the cloud that we released in 2013. We did it at the right time.โ At this point, Lightspeed was doing $10m in sales and cloud based POS competition was heating up with companies like Square or Shopkeep that were also raising venture funding.
Many SaaS solutions are shortcutting a lot of the basic product development in favour of reaching attractive metrics for VCs.
When you play in a intensely and increasingly competitive market, itโs good to create your own blue ocean with a differentiated value proposition (e.g. handling complex operations and payments for Lightspeed).
In the bootstrapping years, working with resellers was really successful. Lightspeed was spending money on resellers conferences in order to attract and train resellers from around the world to resell Ligthspeed. When Lightspeed transitioned to the cloud, it stopped its reseller strategy to do everything direct because the need for a reseller to handle the hardware was less present and because direct and reseller strategies were starting to collude. Today, Lightspeed is trying to rebuild its reseller strategy especially for extremely complex merchants.
โWeโve been a growth story for a long time. Last year, we spent almost the entire fiscal year rolling Lightspeed payments converting 1/3 of our customer base and setting the foundations for profitability. I've spent a lot of time and focus on cutting OpEx. Going forward, I think the market wants to see this profitability, several quarters of software revenue growth as opposed to payments only growth as well as locations growth.โ
โI think we did a couple of really great things with our acquisitions. But what we haven't done is all of the backend integrations. There are a lot of costs that we can pull out. There's a lot of things that we can sunset over time. That work hasn't been done.โ
Saturday, Aug. 31st: Mark Zuckerberg and Daniel Ek argue that Europe should embrace open-source AI to foster innovation, competition, and collaboration. They believe that open-source AI can democratize technology, prevent monopolies, and accelerate advancements while addressing concerns about privacy, security, and ethical standards. They call for regulatory frameworks that support open-source development while ensuring responsible AI use. - The Economist
โA key opportunity for European organisations is through open-sourceย AIโmodels whose weights are released publicly with a permissive licence.โ
โWe believe the next generation of ideas and startups will be built with open-sourceย AI, because it lets developers incorporate the latest innovations at low cost and gives institutions more control over their data. It is the best shot at harnessingย AIย to drive progress and create economic opportunity and security for everyone.โ
โWith more open-source developers than America has, Europe is particularly well placed to make the most of this open-sourceย AIย wave. Yet its fragmented regulatory structure, riddled with inconsistent implementation, is hampering innovation and holding back developers. Instead of clear rules that inform and guide how companies do business across the continent, our industry faces overlapping regulations and inconsistent guidance on how to comply with them. Without urgent changes, European businesses, academics and others risk missing out on the next wave of technology investment and economic-growth opportunities.โ
โIn the short term, delaying the use of data that is routinely used in other regions means the most powerfulย AIย models wonโt reflect the collective knowledge, culture and languages of Europeโand Europeans wonโt get to use the latestย AIย products.โ
โThe stark reality is that laws designed to increase European sovereignty and competitiveness are achieving the opposite.โ
Thanks to Julia for the feedback! ๐ฆ Thanks for reading! See you next week for another issue! ๐