Applying The Lessons Learnt From 2021

Don’t make the mistakes of the past.

With stock markets making new highs this year, it is a good time to look back at the lessons learnt from the collapse of some tech stocks in 2021 and 2022.

Back then, stock markets also reached all-time highs but many tech names collapsed as valuations compressed and growth stalled.

With this in mind, here are some of the key things to be mindful of today as we navigate the stock market.

Don’t celebrate too soon

Investing is a long game. Just because our stocks have risen does not mean we have won the game. The true test of a business’s strength is its enduring ability to keep on growing. Stock prices may also reflect unwarranted short term optimism which does not materialise. 

Make sure to continuously assess the fundamentals of your portfolio companies even (especially) if its stock price has skyrocketed.

Don’t chase stocks

It may be tempting to buy stocks that have risen greatly in a short period of time. Afterall, none of us want to be left out of a massive rally. 

But this fear of missing out can work against us as stocks don’t keep going up forever. Remember that valuations matter and we need to assess if a stock has gone up too much over a short period of time.

In 2021, many stocks rose to unsustainable valuations, only to come crashing down to earth in the next two years. While some have recovered, many still linger up to 90% off their all-time highs.

Sell if valuations don’t make sense

Buy-and-hold is a great strategy when markets are working smoothly and you’ve bought into great growing companies at reasonable valuations. 

But when stock markets are not working well and stock prices rise too high due to unwarranted exuberance, it may be important to look at your sell strategies.

Back in 2021, the stocks of many companies skyrocketed. It was not uncommon to see stocks rise by up to 1,000% in a short period of time.

While some of these companies are undoubtedly growing fast and are resilient, the valuations reached a point where forward returns would likely be depressed. Unsurprisingly, many of these companies’ stocks plunged and have yet to recover.

Growth trends may not continue

It may be tempting to look at a company’s recent revenue and profit growth and assume that it can continue growing at that rate for a long period of time. The reality is that future growth trends may not always mirror the past. This is especially true for companies that have been growing at unsustainably high rates. More often than not, growth will fall back to more normal rates.

The poster boy of the COVID collapse is probably Zoom Communications. The company saw explosive growth, only for its growth rates to flat-line once the pandemic ended.

Besides Zoom, there are numerous other companies that also saw growth decelerate meaningfully as we exited the pandemic era.

These companies unsurprisingly have seen their stock prices collapse.

Look for recurring revenue

Many companies can experience significant upmarkets due to upgrade cycles or loose monetary policy which encourage unsustainable consumer and business spending. However, remember that many companies do experience significant swings in revenue because of the cyclical nature of their end-demand. This may be more true for hardware companies or those that sell big ticket items.

Companies such as Enphase, which sells solar power products such as microinverters, have seen their revenues crater as distributors struggle to clear inventory because of weak end-customer demand.

Bottom line

Although it is nice to see stock prices rise significantly in the past two years, it is important that we remember the key tenets of value investing. The above mistakes are some that many of us have made before.

This time around, let’s try to ensure that we maintain a portfolio of stocks that have good valuations and whose business can continue to thrive in good times and in bad.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Zoom Communications. Holdings are subject to change at any time.

The Latest Thoughts From American Technology Companies On AI (2024 Q3)

A collection of quotes on artificial intelligence, or AI, from the management teams of US-listed technology companies in the 2024 Q3 earnings season.

The way I see it, artificial intelligence (or AI), really leapt into the zeitgeist in late-2022 or early-2023 with the public introduction of DALL-E2 and ChatGPT. Both are provided by OpenAI and are software products that use AI to generate art and writing, respectively (and often at astounding quality). Since then, developments in AI have progressed at a breathtaking pace.

With the latest earnings season for the US stock market – for the third quarter of 2024 – coming to its tail-end, I thought it would be useful to collate some of the interesting commentary I’ve come across in earnings conference calls, from the leaders of technology companies that I follow or have a vested interest in, on the topic of AI and how the technology could impact their industry and the business world writ large. This is an ongoing series. For the older commentary:

With that, here are the latest commentary, in no particular order:

Airbnb (NASDAQ: ABNB)

Airbnb’s management recently introduced a personalised welcome tour of the Airbnb app for first-time users; management sees this personalisation feature as the beginning of a more personalised Airbnb

We also introduced 50 upgrades for guests that make Airbnb a more intuitive and personalized app. And some of the features include a personalized welcome tour of the app for first-time guests, suggest a destination when guests tap the search bar, we’ll recommend locations on their search and booking history, and personalized listing highlights. So when a guest views a listing, we will highlight the details that are relevant to their search, and there are dozens of new features just like these. This is quite literally the beginning of a more personalized Airbnb.

Airbnb’s management is seeing great progress on AI-powered customer service; management sees 3 phases to the deployment of AI-powered customer service, where Phase 1 is Airbnb using AI to answer basic questions from customers, Phase 2 is the AI answering questions from customers in a personalised way, and Phase 3 is the AI taking personalised actions on behalf of customers; management thinks that Airbnb has hired some of the best AI talent to develop AI-powered customer service

We are seeing some really great progress on AI-powered customer service. The way we think about customer service, powered by AI is in 3 phases…

…Phase 1 is just answer basic general questions. We’re rolling out a pilot that can answer basic general questions. Phase 2 is personalization, be able to personalize the questions. Phase 3 is to take action…

…So this is where we think customer service can go enabled by AI, and we’ve hired some of the best people in the world to work on this.

Airbnb is currently in Phase 1 of deploying AI-powered customer service; management thinks that the vast majority of customer chats that are received by Airbnb will be handled directly by AI agents in the future

Phase 1 is the phase we’re in right now. If you were to most — first of all, most of our customer context, we get over 10 million contacts a year. Most of the contacts that we anticipate getting in the coming years aren’t going to be phone calls. They’re going to be chatting through the app. I really personally don’t like calling customer service and having to dial them. I want to be able to chat, and chat AI can intercept. And so we think in the future, the vast majority of our chats are going to be intercepted and handled directly by the AI agent.

An example of the 3rd phase of Airbnb’s AI-powered customer service that management has in mind: An AI agent can help customers to cancel bookings and even make rebookings 

So I’ll give you an example. Let me just give you 1 example. Let’s say I were to contact customer service and I say, “how do I cancel reservation?” In Phase 1, what we’re doing now, the AI agent will answer probably even better than the average customer service agent, how to cancel a reservation. So we’ll take you to how to cancel a reservation step by step. Phase 2 personalization, they’ll say, “hey, Brian, I see you have a reservation coming up in Los Angeles next week. Here’s how you cancel that reservation.” And Phase III is taking action. It would say, “hey, Brian, I see you have a reservation come to Los Angeles. Would you like me to cancel it for you? Just tell me, yes, and I’ll do it for you. I can even handle rebooking.”

Alphabet (NASDAQ: GOOG)

Alphabet’s management thinks Alphabet is positioned to lead in AI because of the company’s full-stack approach of a robust AI infrastructure, world-class research team, and broad user-reach

We are uniquely positioned to lead in the era of AI because of our differentiated full stack approach to AI innovation, and we are now seeing this operate at scale. There’s 3 components: first, a robust AI infrastructure that includes data centers, chips and a global fiber network; second, world-class research teams who are advancing our work with deep technical AI research and who are also building the models that power our efforts. And third, a broad global reach through products and platforms that touch billions of people and customers around the world, creating a virtuous cycle.

Alphabet signed the world’s first corporate agreement for energy from multiple small modular nuclear reactors; the reactors will deliver 500 megawatts of carbon-free power 24/7

We are also making bold clean energy investments, including the world’s first corporate agreement to purchase nuclear energy from multiple small modular reactors, which will enable up to 500 megawatts of new 24/7 carbon-free power.

Since Alphabet began testing AI overviews 18 months ago, the company has reduced the cost to deliver queries by 90% while doubling the size of its Gemini foundation AI model; AI overview has led to users coming to Search more often; AI overview was recently rolled out to 100 new countries and territories and will reach more than 1 billion users on a monthly basis; there’s strong engagement in AI overview, leading to higher overall search usage and user satisfaction, and users are asking longer questions and exploring more websites; the growth driven by AI overviews is increasing over time; the integration of advertising with AI overviews  is performing well; Alphabet is now showing search and shopping ads within AI overviews for mobile users in the USA; management finds that users find ads within AI overviews to be helpful; management expects Search to evolve significant in 2025, driven by advances in AI; management is seeing the monetisation rate on ads within AI overviews to be the same as the broader Search; reminder from management that Google already introduced an answer-machine 10 years ago and management is aware of changing trends in user behaviours in Search

Since we first began testing AI overviews, we have lowered machine cost per query significantly. In 18 months, we reduced cost by more than 90% for these queries through hardware, engineering and technical breakthroughs while doubling the size of our custom Gemini model…

…In Search, recent advancements, including AI overviews, Circle to Search and new features in :ens are transforming the user experience, expanding what people can search for and how they search for it. This leads to users coming to search more often for more of their information needs driving additional search queries. Just this week, AI overview started rolling out to more than 100 new countries and territories. It will now reach more than 1 billion users on a monthly basis. We are seeing strong engagement, which is increasing overall search usage and user satisfaction. People are asking longer and more complex questions and exploring a wide range of websites. What’s particularly exciting is that this growth actually increases over time as people learn that Google can answer more of their questions.

The integration of ads within AI overviews is also performing well, helping people connect with businesses as they search…

…AI overviews, where we have now started showing search and shopping ads within the overview for mobile users in the U.S. As you remember, we’ve already been running ads above and below AI overviews. We’re now seeing that people find ads directly within AI overview is helpful because they can quickly connect with relevant businesses, products and services to take the next step at the exact moment they need…

… So I expect Search to continue to evolve significantly in 2025, both in the search product and in Gemini…

…We recently launched ads within AI overviews on mobile in the U.S. And this really builds on our previous rollout of ads above and below the AI overviews. So overall, for AI overviews, we see monetization at approximately the same rate, which gives us a really strong base on which we can innovate even more…

…[Question] Why doesn’t it make sense to have 2 completely different search experiences one an agent like answers engine; and then two, a link-based more traditional search engine? 

[Answer] In this moment, people are using a lot of buzz words like answer engines and all that stuff. I mean Google started answering questions about 10 years ago in our search product with featured snippets. So look, I think, ultimately, you are serving users. User expectations are constantly evolving. And and we work hard to stay a step ahead, anticipate and stay a step ahead.

Alphabet uses and offers customers both its own TPUs (tensor processing units) and Nvidia GPUs; Alphabet is now on the 6th generation of its TPUs, known as Trillium; LG AI Research reduced its inference processing time by 50% and operating costs by 72% using Google Cloud’s TPUs and GPUs; Alphabet will be one of the first companies to provide Nvidia’s GB 200s at scale; management thinks TPUs have very attractive pricing for its capability

We use and offer our customers a range of AI accelerator options, including multiple classes of NVIDIA GPUs and our own custom-built GPUs. We are now on the sixth generation of TPUs known as Trillium and continue to drive efficiencies and better performance with them…

…Using a combination of our TPUs and GPUs, LG AI research reduced inference processing time for its multimodal model by more than 50% and operating costs by 72%…

…We have a wonderful partnership with NVIDIA. We are excited for the GB 200s and will be one of the first to provide it at scale…

…On your first part of the question on the TPUs. If you look at the flash pricing, we’ve been able to deliver externally, I think and how much more attractive it is compared to other models of that capability.

Usage of Alphabet’s Gemini foundation AI model is in a period of dramatic growth by any measure; improvements to Gemini will soon come; all 7 of Alphabet’s products that have more than 2 billion monthly users each use Gemini models; Gemini is now available on GitHub Copilot; Gemini API calls were up 14x in a 6-month period; Snap saw a 2.5 times increase in engagement with its MyAI chatbot after choosing Gemini to power the chatbot’s user experiences; Gemini’s integration with Android is improving Android; the latest Samsung Galaxy devices’ Android operating system has Gemini Live for users to converse with the Gemini model; Alphabet’s latest Pixel 9 devices have Gemini Nano within; development of the third generation of the Gemini model is progressing well; see Point 23 for how Gemini is helping advertisers

By any measure, token volume, API calls, consumer usage business adoption, usage of the Gemini models is in a period of dramatic growth, and our teams are actively working on performance improvements and new capabilities for our range of models. Stay tuned…

… Today, all 7 of our products and platforms with more than 2 billion monthly users use Gemini models, that includes the latest product to surpass the 2 billion user milestone Google Maps…

…Today, we shared that Gemini is now available on GitHub Copilot with more to come…

…Gemini API calls have grown nearly 14x in a 6-month period. When Snap was looking to power more innovative experiences within their MyAI chatbot, they chose Gemini’s strong multimodal capabilities. Since then, Snap saw over 2.5x as much engagement with MyAI in the United States…

… Gemini’s deep integration is improving Android. For example, Gemini Live lets you have free-flowing conversations with Gemini. People love it. It’s available on Android, including Samsung Galaxy devices. We continue to work closely with them to deliver innovations across their newest devices with much more to come. At Made by Google, we unveiled our latest Pixel 9 series of devices featuring advanced AI models, including Gemini Nano. We have seen strong demand for these devices, and they’ve already received multiple awards…

…We’ve had 2 generations of Gemini model. We are working on the third generation, which is progressing well.

Alphabet’s Project Astra will allow AI to see and reason about the physical world around users, and management aims to ship it as early as 2025

When they’re building out experiences where AI can see and reason about the world around you, Project Astra is a glimpse of that future. We are working to ship experiences like this as early as 2025.

Alphabet is using AI internally to improve coding productivity and efficiency; a quarter of new code at Google is now generated by AI

We’re also using AI internally to improve our coding processes, which is boosting productivity and efficiency. Today, more than 1/4 of all new coated Google is generated by AI, then reviewed and accepted by engineers. This helps our engineers do more and move faster. 

Circle to Search is now available on more than 150 million Android devices; a third of users who have tried Circle to Search now use it weekly; Circle to Search has higher engagement with younger users

Circle to Search is now available on over 150 million Android devices with people using it to shop, translate text and learn more about the world around them. 1/3 of the people who have tried Circle to Search now use it weekly, a testament to its helpfulness and potential…

… For example, with Circle to Search, where we see higher engagement from users aged 18 to 24.  

Lens is now used in over 20 billion visual searches per month; Lens is one of the fastest-growing query types management has seen on Search; management started testing product search on Lens in October and found that shoppers are more likely to engage; management is seeing users use Lens for complex multimodal queries; Alphabet has rolled out shopping ads with Lens visual research results to better connect consumers and businesses.

Lens is now used for over 20 billion visual searches per month. Lens is one of the fastest-growing query types we see on search because of its ability to answer complex multimodal questions and help in product discovery and shopping…

…In early October, we announced product search on Google Lens and in testing this feature, we found that shoppers are more likely to engage with content in this new format. We’re also seeing that people are turning to Lens more often to run complex multimodal quarries, voicing a question or inputting text in addition to a visual. Given these new user behaviors earlier this month, we announced the rollout of shopping ads above and alongside relevant Lens visual search results to help better connect consumers and businesses. 

Customers are using Google Cloud’s AI products in 5 different ways: (1) for AI hardware and software infrastructure; (2) for building and customising AI models with Vertex; (3) for combining Google Cloud’s AI platform with its data platform; (4) for AI-powered cybersecurity solutions; and (5) for building AI agents to improve customer engagement

Customers are using our products in 5 different ways. First, our AI infrastructure. which we differentiate with leading performance driven by storage, compute and software advances as well as leading reliability and a leading number of accelerators…

…Second, our enterprise AI platform, Vertex is used to build and customize the best foundation models from Google and the industry…

…Third, customers use our AI platform together with our data platform, Big Query, because we analyze multimodal data no matter where it is stored with ultra low latency access to Gemini…

…Fourth, our AI-powered cybersecurity solutions Google threat intelligence and security operations are helping customers like BBVA and Deloitte, prevent deduct and respond to cybersecurity threats much faster…

… Fifth, in Q3, we broadened our applications portfolio with the introduction of our new customer engagement suite. It’s designed to improve the customer experience online and in mobile apps as well as in call centers, retail stores and more. 

Waymo is the biggest part of Alphabet’s Other Bets portfolio; Alphabet’s management thinks Waymo is the clear technical leader in autonomous vehicles; Waymo is now serving 150,000 paid rides weekly and driving 1 million fully autonomous miles, and is the first autonomous vehicle company to reach these milestones; Waymo is partnering with Uber and Hyundai to deliver autonomous vehicles to more consumers; Waymo is now on its sixth generation system

I want to highlight Waymo, the biggest part of our portfolio. Waymo is now a clear technical leader within the autonomous vehicle industry and creating a growing commercial opportunity. Over the years, Waymo has been infusing cutting edge AI into its work. Now each week, Waymo is driving more than 1 million fully autonomous miles and serves over 150,000 paid rides, the first time any AV company has reached this kind of mainstream use. Through its expanded network and operations partnership with Uber in Austin and Atlanta, plus a new multiyear partnership with Hyundai, Waymo will bring fully autonomous driving to more people and places. By developing a universal driver, Waymo has multiple paths to market. And with its sixth generation system, Waymo significantly reduced unit costs without compromising safety.

Alphabet’s management finds that AI helps Alphabet better understand consumer-intent and connect consumers with advertisers

AI is expanding our ability to understand intent and connect it to our advertisers. This allows us to connect highly relevant users with the most helpful ad and deliver business impact to our customers.

Advertisers are using Gemini to build and test more creatives at scale; Audi worked with Gemini tools to increase website visits by 80% and increase clicks by 2.7 times

Advertisers now use our Gemini power tools to build and test a larger variety of relevant creators at scale. Audi used our AI tools to generate multiple video image and text assets in different links and orientations out of existing long-form videos. They then fed the newly generated creatives into Demand Gen to drive reach, traffic and booking to their driving experience. The campaign increased website visits by 80% and increased clicks by 2.7x, delivering a lift in their sales. 

Alphabet is offering AI-powered campaigns to help advertisers achieve faster feedback on what is working; DoorDash saw a 15x higher conversion rate at a 50% more efficient cost per action

AI-powered campaigns help advertisers get faster feedback on what creatives workwear and redirect the media buying. Using Demand Gen, DoorDash tested a mix of image and video assets to drive more impact across Google and YouTube’s visually immersive surfaces. They saw a 15x higher conversion rate at a 50% more efficient cost per action when compared to video action campaigns alone. 

Alphabet is using AI to help advertisers better measure their advertising results

This quarter, we extended the availability of our open source marketing mix model, Meridian to more customers, helping to scale measurement of cross-channel budgets to drive better business outcomes.

Alphabet’s big jump capex in 2024 Q3 (was $8.1 billion in 2023 Q3) was mostly for technical infrastructure, in the form of servers and data centers; management expects Alphabet’s 2024 Q4 capex to be similar to what was seen in 2024 Q3; Alphabet announced more than $7 billion in planned data center investments in 2024 Q3, with $6 billion in the USA; management expects further growth in capex in 2025, but not at the same percentage increase seen from 2023 to 2024; the use of TPUs at Alphabet helps to drive efficiencies

With respect to CapEx, our reported CapEx in the third quarter was $13 billion, reflecting investment in our technical infrastructure with the largest component being investment in servers, followed by data centers and networking equipment. Looking ahead, we expect quarterly CapEx in the fourth quarter to be at similar levels to Q3…

…In the third quarter alone, we made announcements of over $7 billion in planned data center investments with nearly $6 billion of that in the U.S…

…As you saw in the quarter, we invested $13 billion in CapEx across the company. And as you think about it, it really is divided into 2 categories. One is our technical infrastructure, and that’s the majority of that $13 billion. And the other one goes into areas such as facilities, the bets and other areas across the company. Within TI, we have investments in servers, which includes both TPUs and GPUs. And then the second categories are data centers and networking equipment. This quarter, approximately 60% of that investments in technical infrastructure went towards servers and about 40% towards data center and networking equipment…

…And as you think about the next quarter and going into next year, as I mentioned in my prepared remarks, we will be investing in Q4 at approximately the same level of what we’ve invested in Q3, approximately $13 billion. And as we think into 2025, we do see an increase coming in 2025, and we will provide more color on that on the Q4 call, likely not the same percent step-up that we saw between ’23 and ’24, but additional increase…

…On your first part of the question on the TPUs. If you look at the flash pricing, we’ve been able to deliver externally, I think and how much more attractive it is compared to other models of that capability. I think probably that gives a good sense of the efficiencies we can generate from our architecture. And so — and we are doing the same that for internal use as well. The models for search while they keep going up in capability we’ve been able to really optimize them for the underlying architecture, and that’s where we are seeing a lot of efficiencies as well.  

Amazon (NASDAQ: AMZN)

Amazon’s management believes that AI will be a big piece of the company’s robotics efforts in its fulfilment network

We continue to innovate in robotics to speed delivery, lower cost to serve, and further improve safety in our fulfillment network…

…We really do believe that AI is going to be a big piece of what we do in our robotics network. We had a number of efforts going on there. We just hired a number of people from an incredibly strong robotics AI organization. And I think that will be a very central part of what we do moving forward, too. 

Amazon’s management sees customers focused on new cloud computing efforts again, and the modernisation of their infrastructure, by migrating to the cloud, is important if they want to work on generative AI at scale

Companies are focused on new efforts again, spending energy on modernizing their infrastructure from on-premises to the cloud. This modernization enables companies to save money, innovate more quickly, and get more productivity from their scarce engineering resources. However, it also allows them to organize their data in the right architecture and environment to do generative AI at scale. It’s much harder to be successful and competitive in generative AI if your data is not in the cloud.

AWS has released nearly twice as many AI features in the last 18 months as other leading cloud providers combined; AWS’s AI business is growing at a triple digit rate at a multi-billion revenue run rate; AWS’s AI business is currently growing more than 3x faster than AWS itself grew when AWS was at a similar stage; management sees AI as an unusually large opportunity

In the last 18 months, AWS has released nearly twice as many machine learning and gen AI features as the other leading cloud providers combined. AWS’ AI business is a multibillion-dollar revenue run rate business that continues to grow at a triple-digit year-over-year percentage and is growing more than 3x faster at this stage of its evolution as AWS itself grew, and we felt like AWS grew pretty quickly…

…It is a really unusually large, maybe once-in-a-lifetime type of opportunity. And I think our customers, the business, and our shareholders will feel good about this long term that we’re aggressively pursuing it.

Amazon has a good relationship with NVIDIA, but management have heard from customers that they want better price performance on their AI workloads, and so AWS developed its own AI chips for training and inference; AWS’s second version of its AI chip for model-training, Trainium 2, will ramp up in the next few weeks; management thinks Trainium 2 have very compelling price performance; management is seeing significant interest in Trainium 2, to the extent they have to increase manufacturing orders much more than originally planned

While we have a deep partnership with NVIDIA, we’ve also heard from customers that they want better price performance on their AI workloads. As customers approach higher scale in their implementations, they realize quickly that AI can get costly. It’s why we’ve invested in our own custom silicon in Trainium for training and Inferentia for inference. The second version of Trainium, Trainium2, is starting to ramp up in the next few weeks and will be very compelling for customers on price performance. We’re seeing significant interest in these chips, and we’ve gone back to our manufacturing partners multiple times to produce much more than we’d originally planned…

…We have a very deep partnership with NVIDIA, we tend to be their lead partner on most of their new chips. We were the first to offer H200s in EC2 instances. And I expect us to have a partnership for a very long time that matters.

Amazon’s management is seeing more model builders standardise on SageMaker,  AWS’s fully-managed AI service; SageMaker’s hyperpod capability helps save model-training time by up to 40%

We also continue to see increasingly more model builders standardize on Amazon SageMaker, our service that makes it much easier to manage your AI data, build models, experiment, and deploy to production. This team continues to add features at a rapid clip punctuated by SageMaker’s unique hyperpod capability, which automatically splits training workloads across more than 1,000 AI accelerators, prevents interruptions by periodically saving checkpoints, and automatically repairing faulty instances from their last saved checkpoint, and saving training time by up to 40%.

Amazon’s management believes Amazon Bedrock, AWS’s AI-models-as-a-service offering for companies that want to leverage existing foundation models for customisation, has the broadest selection of leading foundation models; Bedrock recently added Anthropic’s Claude 3.5 Sonnet model, Meta’s Llama 3.2 models, and more; management is seeing companies use models from different providers within the same application and Bedrock makes it easy to orchestrate the disparate models; Bedrock also helps companies with model-access, prompt engineering, and lowering inference costs

At the middle layer where teams want to leverage an existing foundation model, customized with their data, and then have features to deploy high-quality generative AI applications, Amazon Bedrock has the broadest selection of leading foundation models and most compelling modules for key capabilities like model valuation, guardrails, rag and agents. Recently, we’ve added Anthropic’s Claude 3.5 Sonnet model, Meta’s Llama 3.2 models, Mistral’s Large 2 models and multiple-stability AI models. We also continue to see teams use multiple model types from different model providers and multiple model sizes in the same application.  There’s mucking orchestration required to make this happen. And part of what makes Bedrock so appealing to customers and why it has so much traction is that Bedrock makes this much easier. Customers have many other requests: access to even more models, making prompt management easier, further optimizing inference costs. And our Bedrock team is hard at work making this happen.

Amazon’s management continues to see strong adoption of Amazon Q, Amazon’s generative AI assistant for software development; Amazon Q has the highest reported code acceptance rates in the industry; reminder that Amazon saved $260 million and 4,500 developer years when performing a large Java Development Kit migration through the use of Amazon Q

We’re continuing to see strong adoption of Amazon Q, the most capable generative AI-powered assistant for software development and to leverage your own data. Q has the highest reported code acceptance rates in the industry for multiline code suggestions. The team has added all sorts of capabilities in the last few months, but the very practical use case recently shared where Q Transform saving Amazon’s teams $260 million and 4,500 developer years in migrating over 30,000 applications to new versions of the Java JDK. As excited developers and prompted them to ask how else we could help them with tedious and painful transformations.

Amazon is using generative AI pervasively across its businesses, with hundreds of apps in use or in development; Rufus is a generative AI-powered shopping assistant available in parts of Europe, North America, and India; Amazon is using generative AI to improve personalisation and product-search for consumers when shopping; Project Amelia is an AI system offering tailored business insights to Amazon sellers; Alexa, Amazon’s virtual assistant technology, is being rearchitected with new foundation AI models; the new Kindle Scribe has a built-in AI-powered notebook 

We’re also using generative AI pervasively across Amazon’s other businesses with hundreds of apps in development or launched.

For consumers, we’ve expanded Rufus, our generative AI-powered expert shopping assistant, to the U.K., India, Germany, France, Italy, Spain, and Canada. And in the U.S., we’ve added more personalization, the ability to better narrow customer intent and real-time pricing and deal information. We’ve recently debuted AI shopping guides for consumers, which simplifies product research by using generative AI to pair key factors to consider in a product category with Amazon’s wide selection, making it easier for customers to find the right product for their needs. 

For sellers, we’ve recently launched Project Amelia, an AI system that offers tailored business insights to boost productivity and drive seller growth.

We continue to rearchitect the brain of Alexa with a new set of foundation models that we’ll share with customers in the near future, and we’re increasingly adding more AI into all of our devices. Take the new Kindle Scribe we just announced. The note-taking experience is much more powerful with the new built-in AI-powered notebook, which enables you to quickly summarize pages of notes into concise bullets in a script font that can easily be shared.

Amazon’s management expects capital expenditures of $75 billion for the whole of 2024; most of the capex will be for AWS infrastructure to support demand for AI services; the capex also includes investments in Amazon’s fulfilment and transportation network; management expects capex in 2025 to increase from 2024’s level, with most of the capex for AWS, specifically generative AI; reminder that the faster AWS grows, the faster Amazon needs to invest capital for hardware; many of the assets AWS’s capex is invested in have long, useful lives; management expects to deliver high returns on invested capital with AWS’s generative AI investments; management has a lot of experience, accumulated over the years, in predicting just the right amount of compute capacity to provide for AWS before the generative AI era, and they believe they can do so again for generative AI

Year-to-date capital investments were $51.9 billion. We expect to spend approximately $75 billion in CapEx in 2024. The majority of the spend is to support the growing need for technology infrastructure. This primarily relates to AWS as we invest to support demand for our AI services while also including technology infrastructure to support our North America and international segments. Additionally, we’re continuing to invest in our fulfillment and transportation network to support the growth of the business, improve delivery speeds and lower our cost to serve. This includes investments in same-day delivery facilities, in our inbound network and as well in robotics and automation…

… I’ll take the CapEx part of that. As Brian said in his opening comments, we expect to spend about $75 billion in 2024. I suspect we’ll spend more than that in 2025. And the majority of it is for AWS, and specifically, the increased bumps here are really driven by generative AI…

…The thing to remember about the AWS business is the cash life cycle is such that the faster we grow demand, the faster we have to invest capital in data centers and networking gear and hardware. And of course, in the hardware of AI, the accelerators or the chips are more expensive than the CPU hardware. And so we invest in all of that upfront in advance of when we can monetize it with customers using the resources…

…A lot of these assets are many-year useful life assets. Data centers, for instance, are useful assets for 20 to 30 years…

…I think we’ve proven over time that we can drive enough operating income and free cash flow to make this very successful return on invested capital business. And we expect the same thing will happen here with generative AI…

…One of the least understood parts about AWS over time is that it is a massive logistics challenge. If you think about we have 35-or-so regions around the world, which is an area of the world where we have multiple data centers, and then probably about 130 availability zone through data centers, and then we have thousands of SKUs we have to land in all those facilities. And if you land too little of them, you end up with shortages, which end up in outages for customers. So most don’t end up with too little, they end up with too much. And if you end up with too much, the economics are woefully inefficient. And I think you can see from our economics that we’ve done a pretty good job over time at managing those types of logistics and capacity. And it’s meant that we’ve had to develop very sophisticated models in anticipating how much capacity we need, where, in which SKUs and units.

And so I think that the AI space is, for sure, earlier stage, more fluid and dynamic than our non-AI part of AWS. But it’s also true that people aren’t showing up for 30,000 chips in a day. They’re planning in advance. So we have very significant demand signals giving us an idea about how much we need…

…There are some similarities in the early days here of AI, where the offerings are new and people are very excited about it. It’s moving very quickly and the margins are lower than what I think they will be over time. The same was true with AWS. If you looked at our margins around the time you were citing, in 2010, they were pretty different than they are now. I think as the market matures over time, there are going to be very healthy margins here in the generative AI space.

There are a few hundred million active Alexa devices; management had an initial vision of Alexa being the world’s best personal assistant and they believe now that Alexa’s re-architecture can give it a shot at fulfilling the initial vision

I think we have a really broad number of Alexa devices all over people’s homes and offices and automobiles and hospitality suites. We’ve about 0.5 billion devices out there with a couple of hundred million active end points. And when we first were pursuing Alexa, we had this vision of it being the world’s best personal assistant and people thought that was kind of a crazy idea. And I think if you look at what’s happened in generative AI over the last couple of years, I think you’re kind of missing the boat if you don’t believe that’s going to happen. It absolutely is going to happen. So we have a really broad footprint where we believe if we rearchitect the brains of Alexa with next-generation foundational models, which we’re in the process of doing, we have an opportunity to be the leader in that space.

Amazon’s management believes that AWS’s demand substantially outweighs capacity today; management believes AWS’s rate of growth can improve over time as capacity grows

[Question] On the cloud, are you at all capacity constrained, and will the new Trainium or NVIDIA chips maybe even drive sales growth faster?

[Answer] I believe we have more demand that we could fulfill if we had even more capacity today. I think pretty much everyone today has less capacity than they have demand for, and it’s really primarily chips that are the area where companies could use more supply…

…We’re growing at a very rapid rate and have grown a pretty big business here in the AI space. And it’s early days, but I actually believe that the rate of growth there has a chance to improve over time as we have bigger and bigger capacity.

Apple (NASDAQ: AAPL)

Apple announced Apple Intelligence in June 2024; Apple Intelligence redefines privacy in AI; Apple recently released the first set of Apple Intelligence features in US English for iPhone, iPad, and Mac users, and they include writing tools, an improved version of Siri, a more intelligent Photos App, and notification summaries and priority messages; more Apple Intelligence features will be released in December 2024 and early developer feedback is great; the adoption rate of iOS18 in its first three days is twice as fast as for iOS17, suggesting interest for Apple Intelligence; Apple will release support for additional languages in Apple Intelligence in April 2025

In June, we announced Apple Intelligence, a remarkable personal intelligent system that combines the power of generative models with personal context to deliver intelligence that is incredibly useful and relevant. Apple Intelligence marks the beginning of a new chapter for Apple Innovation and redefines privacy and AI by extending our groundbreaking approach to privacy into the cloud with private cloud compute. Earlier this week, we made the first set of Apple Intelligence features available in U.S. English for iPhone, iPad and Mac users with system-wide writing tools that help you refine your writing, a more natural and conversational Siri, a more intelligent Photos app, including the ability to create movies simply by typing a description, and new ways to prioritize and stay in the moment with notification summaries and priority messages.

And we look forward to additional intelligence features in December with even more powerful writing tools, a new visual intelligence experience that builds on Apple Intelligence and ChatGPT integration as well as localized English in several countries, including the U.K., Australia and Canada. These features have already been provided to developers, and we’re getting great feedback. More features will be rolling out in the coming months as well as support for more languages, and this is just the beginning…

…[Question] I was wondering if you could just expand a little bit on some of the early feedback to Apple Intelligence, both for iOS 18.1 but also the developer beta so far and whether you would attribute Apple Intelligence to any of the strong iPhone performance that we’ve seen to date.

[Answer] We’re getting a lot of positive feedback from developers and customers. And in fact, if you just look at the first 3 days, which is all we have obviously from Monday, the 18.1 adoption is twice as fast as the 17.1 adoption was in the year ago quarter. And so there’s definitely interest out there for Apple Intelligence…

…We started in the — with U.S. English. That started on Monday. There’s another release coming that adds additional features that I had referenced in December in not only U.S. English but also localized for U.K., Australia, Canada, Ireland and New Zealand. And then we will add more languages in April. We haven’t set the specifics yet in terms of the languages, but we’ll add more in April and then more as we step through the year. And so we’re moving just as fast as possible while ensuring quality.

Apple’s management is building the infrastructure to deliver Apple Intelligence, but it does not seem like Apple will need to significantly increase its capex budget from historical norms; management also does not see any significant change to the intensity of research & development (R&D) spending that Apple needs to invest in AI

[Question] Could you just talk a little bit about the CapEx outlook and whether investments in things like private cloud compute could change the historical CapEx range of roughly $10 billion a year?

[Answer] We are rolling out these features, Apple Intelligence features already now. And so we are making all the capacity that is needed available for these features. You will see in our 10-K the amount of CapEx that we’ve incurred during the course of fiscal ’24. And we will — in fiscal ’25, we will continue to make all the investments that are necessary, and of course, the investments in AI-related CapEx will be made…

…[Question] Given how much your tech peers are spending on AI, does this new era of Apple Intelligence actually require Apple to invest more in R&D beyond your current 7% to 8% of sales to capture this opportunity? 

[Answer] We’ve been investing heavily in R&D over the last several years. Our R&D growth has been significant during the last several years. And obviously, as we move through the course of fiscal ’24, we’ve also reallocated some of the existing resources to this new technology, to AI. And so the level of intensity that we’re putting into AI has increased a lot, and you maybe don’t see the full extent of it because we’ve also had some internal reallocation of the base of engineering resources that we have within the company.

Apple’s management thinks the introduction of Apple Intelligence will benefit the entire Apple ecosystem

[Question] I understand Apple Intelligence is a feature on the phone today. But do you think that in the future it could potentially have or benefit the services growth business? Or is that too — are those too bifurcated to even make a call on the — this early in the cycle?

[Answer] Keep in mind that we have released a lot of APIs, and developers will be taking advantage of those APIs. That release has occurred as well, and of course, more are coming. And so I definitely believe that a lot of developers will be taking advantage of Apple Intelligence in a big way. And what that does to services, I’ll not forecast, but I would say that from an ecosystem point of view, I think it will be great for the user and the user experience.

Arista Networks (NYSE: ANET)

Arista Networks’ management is seeing networking for AI gaining a lot of traction; trials that took place in 2023 are becoming pilots in 2024; management expects more production in 2025 and 2026

Networking for AI is gaining a lot of traction as we move from trials in 2023 to more pilots in 2024, connecting to thousands of GPUs, and we expect more production in 2025 and 2026.

AI data traffic is very different from traditional cloud workloads and smooth and consistent data flow is a crucial factor in AI networking

AI traffic differs greatly from cloud workloads in terms of diversity, duration and size of flow. The fidelity of AI traffic flows where the slowest flow matters and one slow flow could slow down the entire job completion time is a crucial factor in networking.

Arista Networks’ management sees the company becoming a pioneer in scale-out Ethernet accelerated networking for large AI workloads; Arista Networks’ new Etherlink portfolio scales well to networks with over 100,000 GPUs and can potentially even handle 1 million GPU clusters; Arista Networks’ latest 77R4 DES platform was launched in close collaboration with Meta Platforms

Our AI centers connect seamlessly from the back end to the front end of compute, storage, WAN and classic cloud networks. Arista is emerging as the a pioneer and scale-out Ethernet accelerated networking for large-scale training and AI workloads. Our new Etherlink portfolio with wire speed 800-gig throughput and non-blocking performance, scales from single tier to efficient 2-tier networks for over 100,000 GPUs, potentially even 1 million AI accelerators with multiple tiers. Our accelerated AI networking portfolio consists of 3 families with over 20 switching products and not just one point switch. At the recent OCP in mid-October 2024, we officially launched a very unique platform that distributed Etherlink 7700 to build 2 tier networks for up to 10,000 GPU clusters. The 77R4 DES platform was developed in close collaboration with Meta. And while it may physically look like and be cable like a 2-tier leaf spine network, DES provides a single-stage forwarding with highly efficient spine fabric, eliminating the need for tuning and encouraging fast failover for large AI accelerator-based clusters. 

Arista Networks’ management believes the company has the broadest set of 800 gigabit per second Ethernet products for AI networks

I’m pleased to report Arista 7700R4 distributed Etherlink switch, the 7800R4 Spine, along with the 7060X6 AI leaf that we announced in June have entered into production providing our customers the broadest set of 800 gigabit per second Ethernet products for their AI networks. Together with 800 gigabit per second parallel optics, our customers are able to connect to 400 gigabit per second GPUs to each port increasing the deployment density over current switching solutions. This broad range of Ethernet platforms allows our customers to optimize density and minimize tiers to best match the requirements of their AI workload.

New AI clusters require high-speed connections to existing backbones

New AI clusters require new high-speed port connections into the existing backbone. These new clusters also increased bandwidth on the backbone to access training data, capture snapshots and deliver results generated by the cluster. This trend is providing increased demand for 7800R3 400-gigabit solution.

Arista Networks’ management sees next-generation AI data centres needing significantly more power while doubling network performance

Next-generation data centers integrating AI will contend with significant increases in power consumption while looking to double network performance.

Arista Networks’ management thinks the adoption of AI networking will rest on specifications that the Ultra Ethernet Consortium (UEC) is expected to soon release; the UEC now has 97 members and Arista Networks is a founding member

Critical to the rapid adoption of AI networking is the Ultra Ethernet consortium specification expected imminently with Arista’s key contributions as a founding member. The UEC ecosystem for AI has evolved to over 97 members.

Arista Networks’ management thinks Ethernet is the only option for open standard space AI networking

In our view, Ethernet is the only long-term viable direction for open standard space AI networking.

Arista Networks’ business growth in 2024 was achieved partly with the help of AI; management is now projecting even more growth in 2025 and is confident of achieving its AI back-end revenue target of US$750 million; the adoption of Arista Networks’ AI back-end products influences the adoption of its front-end AI networking products too; management also expects Arista Networks’ front-end AI networking products to generate around US$750 million in revenue in 2025, but sometimes this gets hard to track; the US$750 million in AI back-end revenue that management expects are brand new for the company

We’ve experienced some pretty amazing growth years with 33.8% growth in ’23 and 2024 appears to be heading at least to 18%, exceeding our prior predictions of 10% to 12%. This is quite a jump in 2024, influenced by faster AI pilots. We are now projecting an annual growth of 15% to 17% next year, translating to approximately $8 billion in 2025 revenue with a healthy expectation of operating margin. Within that $8 billion revenue target, we are quite confident in achieving our campus and AI by back-end networking targets of $750 million each in 2025 that we set way back 1 or 2 years ago. It’s important to recognize though that the back end of AI will influence the front-end AI network and its ratios. This ratio can be anywhere from 30% to 100% and sometimes, we’ve seen it as high as 200% of the back-end network depending on the training requirements. Our comprehensive AI center networking number is therefore likely to be double of our back-end target of $750 million, now aiming for approximately $1.5 billion in 2025…

… I would expect in the back end, any share Arista gets, including that $750 million is incremental. It’s brand new to us. We were never there before…

…I think it all depends on their approach to AI. If they just want to build a back-end cluster and prove something out, then they just look for the highest job training completion and intense training models. And it’s a very narrow use case. But what we’re starting to see more and more, especially with the top 5, like I said, is for every dollar spent in the back end, you could spend 30% more, 100% more, and we’ve even seen a 200% more scenario, which is why our $750 million will carry over to, we believe, next year, another $750 million on front-end traffic that will include AI, but it will include other things as well. It won’t be unique to AI. So I wouldn’t be surprised if that number is anywhere between 30% and 100%, so the average is 100%., which is 2x our back-end number. So feeling pretty good about that. Don’t know how to exactly count that as pure AI, which is why I qualify it by saying increasingly, if you start having inference, training, front end, storage, WAN, classic cloud all come together, the AI — the pure AI number becomes difficult to track.

Arista Networks’ management is stocking up inventory in preparation for a rapid deployment of AI networking products

On the cash front, while we have experienced significant increases in operating cash over the last couple of quarters, we anticipate an increase in working capital requirements in Q4. This is primarily driven by increased inventory in order to respond to the rapid deployment of AI networks and to reduce overall lead times as we move into 2025.

Arista Networks’ management has been surprised by the acceleration of AI pilots by its customers in 2024; management would not be surprised going forward if its AI business grows faster than its classic data center and cloud business (in other words, management would not be surprised if the company’s customers cannibalise some of their classic data center and cloud buildouts for AI)

We were pleasantly surprised with the faster acceleration of AI pilots in 2024. So we definitely see that our large cloud customers are continuing to refresh on the cloud, but are pivoting very aggressively to AI. So it wouldn’t surprise me if we grow faster in AI and faster in campus in the new center markets and slower in our classic markets called that data center and cloud. 

The 4 major AI trials Arista Networks discussed in the 2024 Q1 earnings call have now become 5 trials; 3 of the 5 customers are progressing well and are transitioning from trials to pilots, and they will each grow their GPU clusters from 50,000 to 100,000 in 2025; the customer for the new trial that was started has historically been very focused on Infiniband so management is happy to have won the trial, and management hopes the trail will enter pilot and production in 2025; the last remaining customer is moving slower than management expected with delays in their data center buildout; management has good revenue visibility for 3 of the 5 trials for the next 6-12 months and Arista Networks’ revenue-guide for 2025 does not depend on the remaining 2 trials; a majority of the trials are currently on Arista Networks’ 400-gig products because the customers are waiting for the ecosystem to develop on the 800-gig products, but management expects more adoption of the 800-gig products in 2025; Arista Networks is participating in other smaller AI trials too, but the difference is that management expects the 5 major ones to scale to at least 100,000 GPU clusters 

Arista now believes we’re actually 5 out of 5, not 4 out of 5. We are progressing very well in 4 out of the 5 clusters. 3 of the customers are moving from trials to pilots this year, and we’re expecting those 3 to become 50,000 to 100,000 GPU clusters in 2025. We’re also pleased with the new Ethernet trial in 2024 with our fifth customer. This customer was historically very, very InfiniBand driven. And we are now moving in that particular fifth customer, we are largely in a trial mode in 2024, and we hope to go to pilots and production in 2025. There is one customer who — so 3 are going well. One is starting. The fifth customer is moving slower than we expected. They may get back on their feet. In 2025, they’re awaiting new GPUs, and they’ve got some challenges on power cooling, et cetera. So 3, I would give an A. The fourth one, we’re really glad we won, and we’re getting started and the fifth one, I’d say, steady-state, not quite as great as we would expect them — have expected them to be…

…[Question] I wanted to ask a little bit more about the $750 million in AI for next year. Has your visibility on that improved over the last few months? I wanted to reconcile your comment around the fifth customer not going slower than expected. And it sounds like you’re now in 5 of 5, but wondering if that fifth customer going slower is limiting upside or limiting your visibility there?

[Answer] I think on 3 out of the 5, we have good visibility, at least for the next 6 months, maybe even 12…

…On the fourth one, we are in early trials. We’ve got some improving to do. So let’s see, but we’re not looking for 2025 to be the bang up year on the fourth one. It’s probably 2026. And on the fifth one, we’re a little bit stalled, which may be why we’re being careful about predicting how they’ll do. They may step in nicely in the second half of ’25, in which case, we’ll let you know. But if they don’t, we’re still feeling good about our guide for ’25…

…A majority of the trials and pilots are on 400 because people are still waiting for the ecosystem at 800, including the NICs and the UEC and the packet spring capabilities, et cetera. So while we’re in some early trials on 800, majority of 400 — majority of 2024 is 400 gig. I expect as we go into 2025, we will see a better split between 400 and 800…

… So we’re not saying these 5 are the be-all, end-all, but these are the 5 we predict can go to 100,000 GPUs and more. That’s the way to look at this. So there are the largest AI Titans, if you will. And they can be in the cloud, hyperscaler Titan group, they could be in the Tier 2 as well, by the way, very rarely would they be in a classic enterprise. By the way, we do have at least 10 to 15 trials going on in the classic enterprise too, but they’re much smaller GPU counts, so we don’t talk about it.

Arista Networks’ management sees NVIDIA both as a partner and a competitor in the AI networking market; Arista Networks does see NVIDIA’s Infiniband as a competing solution, but rarely sees NVIDIA’s own Ethernet solution competing; management thinks customers, ranging from those building large GPU clusters to smaller ones, all see Arista Networks as the expert when it comes to AI networking

We view NVIDIA as a good partner. If we didn’t have the ability to connect to their GPUs, we wouldn’t have all this AI networking demand. So thank you, NVIDIA. Thank you, Jensen, for the partnership. Now as you know, NVIDIA sells the full stack and most of the time, it’s with InfiniBand, and with the Mellanox acquisition, they do have some Ethernet capability. We personally do not run into the Ethernet capability very much. We run into it, maybe in 1 or 2 customers. And so generally speaking, Arista has looked upon as the expert there. We have a full portfolio. We have full software. And whether it’s the large scale-out ethernet working customers like the Titans or even the smaller enterprises, we’re seeing a lot of smaller GPU clusters of the enterprise, Arista is looked upon as the expert there. But that’s not to say we’re going to win 100%. We certainly welcome NVIDIA as a partner on the GPU side and a fierce competitor, and we look to compete with them on the Ethernet switching.

The AI back-end market is where Arista Networks natively connects with GPU and where NVIDIA’s Infiniband is the market leader, but Arista Networks’ Ethernet solution is aiming to be the gold standard; for the AI front-end market, Arista Networks’ solutions are the gold standard and management is seeing some customers fail to run their AI application on competing solutions and want to replace them with Arista Networks’ solutions

So since you asked me specifically about AI as opposed to cloud, let me parse this problem into 2 halves, the back end and the front end, right? At the back end, we’re natively connecting to GPUs. And there can be many times, we just don’t say it because somebody just bundles it in the GPU in particular, NVIDIA. And you may remember a year ago, I was saying we’re outside looking in because most of the bundling is happening with InfiniBand…

…So we’ll take all we can get, but we are not claiming to be a market leader there. We’re, in fact, claiming that there are many incumbents there with InfiniBand and smaller versions of Ethernet that Arista is looking to gain more credibility and experience and become the gold standard for the back end.

On the front end, in many ways, we are viewed as the gold standard. So competitively, it’s a much more complex network. You have to build a leaf-spine architecture. John alluded to this, there’s a tremendous amount of scale with L2, L3, EVPN, VXLAN, visibility, telemetry, automation, routing at scale, encryption at scale. And this, what I would call accelerated networking portfolio complements NVIDIA’s accelerated compute portfolio. And compared to all the peers you mentioned, we have the absolute best portfolio of 20 switches and 3 families and the capability and the competitive differentiation is bar none. In fact, I am specifically aware of a couple of situations where the AI applications aren’t even running on some of the industry peers you talked about, and they want to swap theirs for ours. So feeling extremely bullish with the 7800 flagship product, the newly introduced 7700 that we worked closely with Meta, the 7060, this product line running today mostly at 400 gig because a lot of the NIC and the ecosystem isn’t there for 800. But moving forward into 800, this is why John and the team are building the supply chain to get ready for it.

ASML (NASDAQ: ASML)

While ASML’s management has seen the strong performance of AI continue – and expects the performance to continue for some time – other market segments have taken longer to recover than management expected; in the Memory segment, management is seeing limited capacity additions among customers, apart from AI, as the customers embark on technology transition to HBM and DDR5

There have been quite some market dynamics in the past couple of months. Very clearly, the strong performance of AI clearly continues and I think it continues to come with quite some upside. We will also see that in other market segments, it takes longer to recover. Recovery is there, but it’s more gradual than what we anticipated before and it will continue in 2025. That does lead to some customer cautiousness…

…If you look at the Memory business, this customer cautiousness that I talked about, leads to limited capacity additions. While at the same time, we do see a lot of focus and strong demand when it comes to technology transitions and particularly as it is related to High Bandwidth Memory and to DDR5. So again, there anything related to AI is strong, but other than that there are limited capacity additions.

The AI growth-driver is very strong over the long-term and ASML’s management sees that AI is increasing share in ASML’s customers’ business

If you look at the long-term outlook, I believe the growth drivers are still very much intact. The secular growth drivers are clear and they are strong. I think if you look at AI, very, very strong, very clear and undisputed. Taking an increasing share in the business of our customers. So I think that is going very strongly.

ASML’s management is seeing upside on AI because the overall demand for AI applications continues to increase, which has driven a recovery in server demand, but management does not have complete understanding on how the AI market will play out

We also mentioned some upside on AI, because we still believe that the overall demand for those application is there, continue to increase. So if we look at the server demand, we see there a very nice recovery. A lot of that has to do with AI application. So we talk about upside, which also means that the overall dynamic of the market is still playing. And we felt the need to provide an update for next year based on some of the development we have seen. I think in no way we are also saying that there is a complete understanding of how the entire market will continue to play out in the next few months. So I think on the second part of your question, I would say maybe this has not played out fully yet…

…[Question] You would expect to happen then, I guess, to — at some point will happen?

[Answer] Well, I think if everyone — and I think a lot of us still believe in a strong AI demand in the coming years, I think that demand has to be fulfilled. Therefore, yes, I will say mostly, we will see some development also on that front in the coming months.

Datadog (NASDAQ: DDOG)

Datadog’s management is seeing next-gen AI customers want to obtain visibility into their AI usage as they continue experimenting with the technology; around 3,000 customers used at least one of Datadog’s AI integrations at the end of 2024 Q3; management is starting to see Datadog’s LLM (large language model) observability products gain traction as AI experiments start becoming production applications; hundreds of customers are already using LLM observability, and some customers have reduced time spent on investigating LLM issues from days or hours to just minutes; management is seeing customers wanting to use APM (Application Performance Monitoring) alongside LLM observability 

In the next-gen AI space, customers continue to experiment with new AI technologies. And as they do, they want to get visibility into their AI use. At the end of Q3, about 3,000 customers used one or more Datadog AI integrations to send us data about their AI, machine learning and LLM usage. As some of these experiments start turning into production AI applications, we are seeing initial signs of traction for our LLM observability products.

Today, hundreds of customers are using LLM observability with more exploring it every day. And some of our first paying customers have told us that they have cut the time spent investigating LLM latency, errors and quality from days or hours to just minutes. Our customers not only want to understand the performance and cost of their LLM applications, they also want to understand the LLM model performance within the context of their entire application. So they are using APM alongside LLM observability to get fully integrated end-to-end visibility across all their applications and tech stacks

AI-native customers accounted for 6% of Datadog’s ARR in 2024 Q3 (was 6% 2024 Q2); AI-native customers contributed 4 percentage points to Datadog’s year-on-year growth in 2024 Q3, compared to 2 percentage points in 2023 Q3; management has seen a very rapid ramp in usage of Datadog among large customers in the AI-native cohort, and management thinks these customers will optimise cloud and observability usage in the future, while also asking for better terms; management is seeing Datadog’s production-minded LLM observability products being used by real paying customers with real volumes in real production workloads; AI-native companies are model providers or AI infrastructure providers that serve as a proxy for the AI industry

AI native customers who this quarter represented more than 6% of our Q3 ARR, up from more than 4% in Q2 and about 2.5% of our ARR in the year ago quarter. AI native customers contributed about 4 percentage points of year-over-year growth in Q3 versus about 2 percentage points in the year ago quarter. While we believe that adoption of AI will continue to benefit Datadog in the long term, we are mindful that some of the large customers in this cohort have ramped extremely rapidly and that these customers may optimize cloud and observability usage and increase their commitments to us over time with better terms. This may create volatility in our revenue growth in future quarters on the backdrop of long-term volume growth…

…We are seeing our production-minded LLM observability products, for example, being used by real paying customers with real volumes and real applications in real production workloads. So that’s exciting and healthy. I think it’s a great trend for the future…

… We have that group of AI, like smaller — relatively small number of AI companies or AI native companies. Many of them are model providers or infrastructure providers for AI that serve the rest of the industry and they are really a proxy for the future growth of the rest of the industry in AI.

Datadog signed a 7-figure expansion deal with a hyperscaler delivering next-gen AI models; the hyperscaler has its homegrown observability solution, but the solution needs time-consuming customisation and manual configuration; the hyperscaler chose Datadog because Datadog’s platform can scale flexibly

We signed a 7-figure annualized expansion with a division of a hyperscaler delivering next-gen AI models. This customer is very technically capable and already has a homegrown observability solution, which requires time-consuming customization and manual configuration. They will be launching new features for their large language models soon and need a platform that can scale flexibly while supporting proactive incident detection. By expanding the use of Datadog, they expect to efficiently onboard new teams and environments and support the rapidly increasing adoption of the LLMs.

Datadog’s management continues to believe that digital transformation, cloud migration, and AI adoption are long-term growth drivers of Datadog’s business

Overall, we continue to see no change to the multiyear trend towards digital transformation and cloud migration, which we continue to believe are still in early days. We are seeing continued experimentation with new advances such as next-gen AI, and we believe this is one of the many factors that will drive greater use of the cloud and other modern technologies.

Datadog’s management is starting to see more inference AI workloads, but they are still concentrated among API-driven providers and it’s still very early days in terms of customers putting their next-gen AI applications into production; management expects more diversification to occur in the future as more companies enter production with their applications and customise their models 

In terms of the workloads, you’re right that we’re starting to see more inference workloads, but they still tend to be more concentrated across a number of API-driven providers. So there are a few others, both on LLMs and other kinds of models. So this is where I think most of the usage in production at least is today. We expect that to diversify more over time as companies get further into production with their applications and they start to be customizing more on their models…

…We are excited to see what’s happening with the AI innovation as it gets further down the pipe and away from testing and experimenting and more into production applications. And we have some signs that it’s starting to happen. Again, we see that with our LLM observability product. We see that also with some of the workloads we monitor from our customers on the infrastructure side. But I would say it’s still very early days in terms of customers being in production with their next-gen AI applications.

Datadog’s management is seeing a small amount of cloud workloads of companies being cannibalised by their AI initiatives

You’re right that the — where the workloads could have grown maybe instead of growing 20%, they could grow 25%, maybe some of those 5% instead are being invested both in terms of infrastructure budget or innovation — time innovation budget. All that is going into AI, and that’s largely right now in experimentation and model training and that sort of thing. 

Datadog’s Management is working with customers with large inference workloads on how Datadog can be helpful on the GPU profiling side of inference; management is also experimenting with how Datadog can be helpful on side of training; management thinks that in a steady state, 60% of AI workloads will be inference and 40% will be training, so there’s still a lot of value to be found if Datadog can be useful in the training side too

Right now, we’re working with a number of customers that have real-world large inference workloads on how we can help on the GPU profiling side for inference. We’re doing less on the training side, mostly because the training jobs tend to be more bespoke and temporary, and there’s less of an application that’s attached to those that these are just very large clusters of GPUs. So it’s closer to HPC in a way than it is to traditional applications, though we are also experimenting with what we can do there. There is a world where maybe in a durable fashion, 60% of workloads are inference and 40% are training. And if that’s the case, there’s going to be a lot of value to be had by having repeatable training and repeatable tooling for that. So we are also looking into that.

Datadog is not monetising GPU instances as well as CPU instances today, but management thinks that could change in the future

As of today, we really don’t monetize GPU instances all that well compared to the other CPU instances. So GPU instance is many times the cost of a CPU instance, and we charge the same amount for it. That doesn’t have to be the case in the future. If we do things that are particularly interesting and it’s going to have a real impact on — and deliver value and how customers use and make the best of their GPUs and in the end, save money. 

Datadog’s management is seeing Datadog’s AI-native cohort grow faster than its cloud-native cohorts did in the late 2010s and early 2020s

What we’ve seen with cloud native in the late ’10s and early ’20s, where we had these numbers of cloud-native consumer companies that were growing very fast, with 2 differences. The first one is that the AI cohort is growing faster and there are larger individual ACVs [annual contract value] for these customers.

Datadog’s management thinks that workloads on Datadog’s platform could really accelerate when non-AI-native companies start bringing AI applications into production

In terms of the growth of workloads, look, I mean, as we said, we see growth across the customer base pretty much. We see growth of classical workloads in the cloud. We see large growth — very large growth on the AI native side. We think that the one big catalyst for future acceleration will be those AI native applications or those AI applications, I should say, going into production for non-AI native companies for a much broader set of customers than the customers that are deploying these kind of applications to their — in production. And as they do, they will also look less like just large cluster of GPUs and more like traditional applications because the GPU needs a database, it needs [ core ] application in front of it, it needs layers to secure it and authorize it and all the other things. So it’s going to look a lot more like a normal application with some additional more concentrated compute and GPUs.

Datadog’s management does not expect Datadog to make outsized investments in GPU clusters for compute

Unlike many others, we don’t expect at this point to have outsized investments in compute. We’re not building absolutely large GPU clusters.

dLocal (NASDAQ: DLO)

dLocal’s management launched the Smart Requests functionality in 2024 Q3 that improves conversion rates for merchants by 1.22 percentage points on average, which equates to a 1.2% increase in revenue for merchants; Smart Requests relies on localised machine learning models to maximise authorisation rates for merchant

During the quarter, we launched our smart requests functionality, boosting our transaction performance and therefore, improving conversion rates by an average of 1.22 percentage points across the board. It may sound minor, but it isn’t. It actually represents, in practical terms, 1.2% additional revenue to our merchants. Smart requests rely on per country machine learning models that optimize routing and chaining so as to maximize authorization rates for our merchants.

Fiverr (NYSE: FVRR)

Fiverr’s management believes that Fiverr’s next generation of products must empower its community to fully leverage AI, and that the best work will be done in the future by a combination of humans and AI

One thing that became clearer to me in the last year is that with the emergence of GenAI and the promise of AGI, the next generation of products we build must empower our community to fully leverage artificial intelligence. It also became clear to me that in the future, the best work will be done by humans and AI technology together, not humans alone or AI alone.

Fiverr’s management is providing Fiverr’s customers with an AI assistant to help them navigate the company’s platform 

This means that every business that comes to Fiverr will have a world-class AI assistant to help them get things done, from ideation, scoping and briefing to project management and workflow automation. It means that they can seamlessly leverage both human talent and machine intelligence to create the most beautiful results.

Fiverr’s management is building a new search experience on the Fiverr platform for buyers which incorporates Neo, its AI powered smart matching tool; Fiverr has launched Dynamic Matching to allow buyers to put together project briefs with an AI assistant to help them get matched to the most relevant freelancers; these new features have experienced enthusiastic reception in just a few weeks; projects that use these new features are bigger projects than the typical scope of projects on Fiverr

On the buyer side, we are building a new search experience that not only includes more dynamic catalogs but also incorporates Neo, an AI-powered smart matching tool, to help customers match with more contextual information. We launched Dynamic Matching to allow buyers to put together comprehensive briefs with a powerful AI assistant and then get matched with the most relevant freelancer with a tailored proposal…

…Even in the few weeks since we launched these products, we have already seen an enthusiastic reception from our community and promising performance. The projects that come through these products are several times larger than a typical project on Fiverr, and we believe it has a lot more potential down the road as the awareness and trust of these products grow on the platform.

Mastercard (NYSE: MA)

Mastercard acquired Brighterion in 2017 to use AI capabilities for decision intelligence; after boosting the product with generative AI, Mastercard has seen a 20% lift in the product 

One of the more recent ones that we talked about that we invested heavily in using our Brighterion acquisition from back in 2017 to use our AI capabilities is decision intelligence. We’ve now boosted the product with Gen AI and the outcome that we see is tremendous. This is up to a 20% lift that we see.

Meta Platforms (NASDAQ: META)

Meta’s management is seeing rapid adoption of Meta AI and Llama; Meta AI now has more than 500 million monthly actives; Llama token usage has grown exponentially in 2024 so far; Meta released Llama 3.2 in 2024 Q3; the public sector is adopting Llama; management is seeing higher usage of Meta AI as the models improve; Meta AI is built on Llama 3.2; voice functions for Meta AI are now available in English in the USA, Australia, Canada, and New Zealand; image editing through simple text prompts, and the ability to learn about images, are now available in Meta AI in the USA; Meta AI remains on track to be the most-used AI assistant in the world by end-2024; early use cases for Meta AI are for information gathering, help with how-to tasks, explore interests, look for content, and generate images

We’re seeing rapid adoption of Meta AI and Llama, which is quickly becoming a standard across the industry…

…Meta AI now has more than 500 million monthly actives…

…Llama token usage has grown exponentially this year and the more widely that Llama gets adopted and becomes the industry standard the more that the improvements to its quality and efficiency will flow back to all of our products. This quarter, we released Llama 3.2, including the leading small models that run on device and open source multimodal models…

…We’re also working with the public sector to adopt Llama across the U.S. government…

…We’re seeing lifts in usage as we improve our models and have introduced a number of enhancements in recent months to make Meta AI more helpful in engaging. Last month, we began introducing voice, so you can speak with Meta AI more naturally, and it’s now fully available in English to people in the U.S., Australia, Canada and New Zealand. In the U.S., people can now also upload photos to Meta AI to learn more about them, write captions for post and add, remove or change things about their images with a simple text prompt. These are all built with our first multimodal foundation model, Llama 3.2…

…We’re excited about the progress of Meta AI. It’s obviously very early in its journey, but it continues to be on track to be the most used AI assistant in the world by end of year…

… Number of the frequent use cases we’re seeing include information gathering, help with how-to tasks, which is the largest use case. But we also see people using it to go deeper on interests, to look for content on our services, for image generation, that’s also been another pretty popular use case so far.

Meta’s management is seeing AI have a positive impact on nearly all aspects of Meta; improvements to Meta’s AI-driven feed and video recommendations have driven increases in time spent on Facebook this year by 8% and on Instagram by 6%; more than 1 million advertisers are using Meta’s Gen AI tools and advertisers using image generation are enjoying a 7% increase in conversions; management sees plenty of new opportunities for new AI advances to accelerate Meta’s core business, so they want to invest more there

We’re seeing AI have a positive impact on nearly all aspects of our work from our core business engagement and monetization to our long-term road maps for new services and computing platforms…

…Improvements to our AI-driven feed and video recommendations have led to an 8% increase in time spent on Facebook and a 6% increase on Instagram this year alone. More than 1 million advertisers used our Gen AI tools to create more than 15 million ads in the last month. And we estimate that businesses using image generation are seeing a 7% increase in conversions and we believe that there’s a lot more upside here…

…It’s clear that there are a lot of new opportunities to use new AI advances to accelerate our core business that should have strong ROI over the next few years. So I think we should invest more there.

 The development of Llama 4 is progressing well; Llama 4 is being trained on more than 100,000 H100s and it’s the biggest training cluster in the world management is aware of; management expects the smaller Llama 4 models to be ready in early-2025; management thinks Llama 4 will be much faster and will have new modalities, stronger capabilities and reasoning

I’m even more excited about Llama 4, which is now well into its development. We’re training the Llama 4 models on a cluster that is bigger than 100,000 H100s or bigger than anything that I’ve seen reported for what others are doing. I expect that the smaller Llama 4 models will be ready first, and they’ll be ready — we expect sometime early next year. And I think that there are going to be a big deal on several fronts, new modalities, capabilities, stronger reasoning and much faster. 

Meta’s management remains convinced that open source is the way to go for AI development; the more developers use Llama, the more Llama improves in both quality and efficiency; in terms of efficiency, with higher adoption of Llama, management is seeing NVIDIA and AMD optimise their chip designs to better run Llama

It seems pretty clear to me that open source will be the most cost-effective, customizable, trustworthy performance and easiest to use option that is available to developers. And I am proud that Llama is leading the way on this…

…[Question] You said something along the lines of the more standardized Llama becomes the more improvements will flow back to the core meta business. And I guess, could you just dig in a little bit more on that?

[Answer] The improvements to Llama, I’d say come in a couple of flavors. There’s sort of the quality flavor and the efficiency flavor. There are a lot of researchers and independent developers who do work and because Llama is available, they do the work on Llama and they make improvements and then they publish it and it becomes — it’s very easy for us to then incorporate that both back into Llama and into our Meta products like Meta AI or AI Studio or Business AIs because the work — the examples that are being shown are people doing it on our stack.

Perhaps more importantly, is just the efficiency and cost. I mean this stuff is obviously very expensive. When someone figures out a way to run this better if that — if they can run it 20% more effectively, then that will save us a huge amount of money. And that was sort of the experience that we had with open compute and why — part of why we are leaning so much into open source here in the first place, is that we found counterintuitively with open compute that by publishing and sharing the architectures and designs that we had for our compute, the industry standardized around it a bit more. We got some suggestions also that helped us save costs and that just ended up being really valuable for us. Here, one of the big costs is chips — a lot of the infrastructure there. What we’re seeing is that as Llama gets adopted more, you’re seeing folks like NVIDIA and AMD optimize their chips more to run Llama specifically well, which clearly benefits us. 

Meta’s management expects to continue investing seriously into AI infrastructure

Our AI investments continue to require serious infrastructure, and I expect to continue investing significantly there too. We haven’t decided on the final budget yet, but those are some of the directional trends that I’m seeing.

Meta’s management thinks the integration of Meta AI into the Meta Ray-Ban glasses is what truly makes the glasses special; the Meta Ray-Ban glasses can answer questions throughout the day, help wearers remember things, give suggestions to wearers in real-time using multi-modal AI, and translate languages directly into the ear of wearers; management continues to think glasses are the ideal form-factor for AI because glasses lets AI see what you see and hear what you hear; demand for the Meta Ray-Ban glasses continues to be really strong; a recent release of the glasses was sold out almost immediately; Meta has deepened its partnership with EssilorLuxottica to build future generations of the glasses; Meta recently showcased Orion, its first full holographic AR glasses

This quarter, we also had several milestones around Reality Labs and the integration of AI and wearables. Ray-Ban meta glasses are the prime example here. They’re great booking glasses that let you take photos and videos, listen to music and take calls. But what makes them really special is the Meta AI integration. With our new updates, it will be able to not only answer your questions throughout the day, but also help you remember things, give you suggestions as you’re doing things using real-time multi-modal AI and even translate other languages right in your ear for you. I continue to think that glasses are the ideal form factor for AI because you can let your AI see what you see, hear what you hear and talk to you.

Demand for the glasses continues to be very strong. The new clear addition that we released at Connect sold out almost immediately and has been trading online for over $1,000. We’ve deepened our partnership with EssilorLuxottica to build future generations of smart eyewear that deliver both cutting-edge technology and style.

At Connect, we also showed Orion, our first full holographic AR glasses. We’ve been working on this one for about a decade, and it gives you a sense of where this is all going. We’re not too far off from being able to deliver great-looking glasses to let you seamlessly blend the physical and digital worlds so you can feel present with anyone no matter where they are. And we’re starting to see the next computing platform come together and it’s pretty exciting.

Newer scaling laws seen with Meta’s large language models inspired management to develop new ranking model architectures that can learn more effectively from significantly larger data sets; the new ranking model architectures have been deployed to Facebook’s video ranking models, helping to deliver more relevant recommendations; management is exploring the use of the new ranking model architectures on other services and the introduction of cross-surface data to the models, with the view that these moves will unlock more relevant recommendations and lead to better engineering efficiency

Previously, we operated separate ranking and recommendation systems for each of our products because we found that performance did not scale if we expanded the model size and compute power beyond a certain point. However, inspired by the scaling laws we were observing with our large language models, last year, we developed new ranking model architectures capable of learning more effectively from significantly larger data sets.

To start, we have been deploying these new architectures to our Facebook ranking video ranking models, which has enabled us to deliver more relevant recommendations and unlock meaningful gains in launch time. Now we’re exploring whether these new models can unlock similar improvements to recommendations on other services. After that, we will look to introduce cross-surface data to these models, so our systems can learn from what is interesting to someone on one surface of our apps and use it to improve their recommendations on another. This will take time to execute and there are other explorations that we will pursue in parallel. However, over time, we are optimistic that this will unlock more relevant recommendations while also leading to higher engineering efficiency as we operate a smaller number of recommendations.

Meta’s management is using new approaches to AI modelling to allow Meta’s ad systems to consider a person’s sequence of actions before and after seeing an ad, which allow the systems to better predict a person’s response to specific ads; the new approaches to AI modelling have delivered a 2%-4% increase in conversions in tests; Meta is seeing strong user-retention with its generative AI tools for image expansion, background generation, and text generation; Meta has started testing its first generative AI tools for video expansion and image animation and plans to roll them out broadly by early-2025

The second part of improving monetization efficiency is enhancing marketing performance. Similar to organic content ranking, we are finding opportunities to achieve meaningful ads performance gains by adopting new approaches to modeling. For example, we recently deployed new learning and modeling techniques that enable our ad systems to consider the sequence of actions a person takes before and after seeing an ad. Previously, our ad system could only aggregate those actions together without mapping the sequence. This new approach allows our systems to better anticipate how audiences will respond to specific ads. Since we adopted the new models in the first half of this year, we’ve already seen a 2% to 4% increase in conversions based on testing within selected segments…

…Finally, there is continued momentum with our Advantage+ solutions, including our ad creative tools. We’re seeing strong retention with advertisers using our Generative AI-powered image expansion, background generation and text generation tools, and they’re already driving improved performance for advertisers even at this early stage. Earlier this month, we began testing our first video generation features, video expansion and image animation. We expect to make them more broadly available by early next year.

Meta’s management expects to significantly increase Meta’s infrastructure for generative AI while prioritising fungibility

Given the lead time of our longer-term investments, we also continue to maximize our flexibility so that we can react to market developments. Within Reality Labs, this has benefited us as we’ve evolved our road map to respond to the earlier-than-expected success of smart glasses. Within Generative AI, we expect significantly scaling up our infrastructure capacity now while also prioritizing its fungibility will similarly position us well to respond to how the technology and market develop in the years ahead.

Meta’s management continues to develop tools for individuals and businesses to create AI agents easily; management thinks that Meta’s progress with AI agent tools is currently at where Meta was with Meta AI a year ago; management wants the AI agent tools to be widely used in 2025

There are also other new products like that, things around AI Studio. This year, we really focused on rolling out Meta AI as kind of our are kind of single assistant that people can ask any question to, but I think there’s a lot of opportunities that I think we’ll see ramp more over the next year in terms of both consumer and business use cases, for people interacting with a wide variety of different AI agents, consumer ones with AI Studio around whether it’s different creators or kind of different agents that people create for entertainment. Or on the business side, we do want to continue making progress on this vision of making it set any small business or any business over time can with a few clicks stand up in AI agent that can help do customer service and sell things to all of their customers around the world, and I think that’s a huge opportunity. So it’s very broad…

…But I’d say that we’re — today, with AI Studio and business AIs about where we were with Meta AI about a year ago. So I think in the next year, our goal around that is going to be to try to make those pretty widespread use cases, even though there’s going to be a multiyear path to getting kind of the depth of usage and the business results around that we want. 

Meta’s management is not currently sharing quantitative metrics on productivity improvements with the internal use of AI, but management is excited about the internal adoption they are seeing and the future opportunities for doing so

On the use of AI and employee productivity, it’s certainly something that we’re very excited about. I don’t know that we have anything particularly quantitative that we’re sharing right now. I think there are different efficiency opportunities with AI that we’ve been focused on in terms of where we can reduce costs over time and generate savings through increasing internal productivity in areas like coding. For example, it’s early, but we’re seeing a lot of adoption internally of our internal assistant and coding agent, and we continue to make Llama more effective at coding, which should also make this use case increasingly valuable to developers over time.

There are also places where we hope over time that we’ll be able to deploy these tools against a lot of our content moderation efforts to help make the big body of content moderation work that we undertake, to help it make it more efficient and effective for us to do so. And there are lots of other places around the company where I would say we’re relatively early in exploring the way that we can use LLM based tools to make different types of work streams more efficient.

It appears that Meta has achieved more than management expected in terms of developing its own AI infrastructure (in other words, developing its own AI chips)

So I think part of what we’re seeing this year is the infra team is executing quite well. And I think that’s, why over the course of the year, we’ve been able to build out more capacity. I mean going into the year, we had a range for what we thought we could potentially do. And we have been able to do, I think, more than, I think, we’d kind of hoped and expected at the beginning of the year. And while that reflects as higher expenses, it’s actually something that I’m quite happy that the team is executing well on. And I think that will — so that execution makes me somewhat more optimistic that we’re going to be able to keep on building this out at a good pace but that’s part of the whole thing. 

Meta’s management is starting to test the addition of AI-generated or AI-augmented content to users of Instagram and Facebook; management has high confidence that AI-generated and/or AI-augmented content will be an important trend in the future

I think we’re going to add a whole new category of content, which is AI generated or AI summarized content or kind of existing content pulled together by AI in some way. And I think that, that’s going to be just very exciting for the — for Facebook and Instagram and maybe Threads or other kind of feed experiences over time. It’s something that we’re starting to test different things around this. I don’t know if we know exactly what’s going to work really well yet. Some things are promising. I don’t know that this isn’t going to be a big impact on the business in ’25 would be my guess. But I think that there is I have high confidence that over the next several years, this is going to be an important trend and one of the important applications.

Meta’s management is currently focused on the engagement and user-experience of Meta AI; the monetisation of Meta AI will come later

Right now, we’re really focused on making Meta AI as engaging and valuable a consumer experience as possible. Over time, we think there will be a broadening set of queries that people use it for. And I think that the monetization opportunities will exist when over time as we get there. But right now, I would say we are really focused on the consumer experience above all and this is sort of a playbook for us with products that we put out in the world where we really dial in the consumer experience before we focus on what the monetization could look like.

Microsoft (NASDAQ: MSFT)

Microsoft’s AI business is on track to exceed $10 billion in annual revenue run rate in 2024 Q4 after being started for just 2.5 years; it will be the fastest business in the company’s history to do so; Microsoft’s AI business is nearly all inference (see Point 32 for more)

All up, our AI business is on track to surpass an annual revenue run rate of $10 billion next quarter, which will make it the fastest business in our history to reach this milestone…

…We’re excited that only 2.5 years in, our AI business is on track to surpass $10 billion of annual revenue run rate in Q2…

…If you sort of think about the point we even made that this is going to be the fastest growth to $10 billion of any business in our history, it’s all inference, right? 

Azure took share in 2024 Q3 (FY2025 Q1), driven by AI; Azure grew revenue by 33% in 2024 Q3 (was 29% in 2024 Q2), with 12 points of growth from AI services (was 8 points in 2024 Q2); Azure’s AI business has higher demand than available capacity

Azure took share this quarter…. 

… Azure and other cloud services revenue grew 33% and 34% in constant currency, with healthy consumption trends that were in line with expectations. The better-than-expected result was due to the small benefit from in-period revenue recognition noted earlier. Azure growth included roughly 12 points from AI services similar to last quarter. Demand continues to be higher than our available capacity. 

Microsoft’s management thinks Azure offers the broadest selection of AI chips, from Microsoft’s own Maia 100 chip to AMD and NVIDIA’s latest GPUs; Azure is the first cloud provider to offer NVIDIA’s GB200 chips

We are building out our next-generation AI infrastructure, innovating across the full stack to optimize our fleet for AI workloads. We offer the broadest selection of AI accelerators, including our first-party accelerator, Maia 100 as well as the latest GPUs from AMD and NVIDIA. In fact, we are the first cloud to bring up NVIDIA’s Blackwell system with GB200-powered AI servers.

Azure OpenAI usage more than doubled in the past 6 months, as both startups and enterprises move apps from test to production; GE Aerospace used Azure OpenAI to build a digital assistant for its 52,000 employees and in 3 months, the assistant has processed 500,000 internal queries and 200,000 documents; Azure recently added support for OpenAI’s newest o1 family of AI models; Azure AI is offering industry-specific models, including multi-modal models for medical imaging; Azure AI is increasingly an on-ramp for Azure’s data and analytics services, driving acceleration of Azure Cosmos DB and Azure SQL DB hyperscale usage

More broadly with Azure AI, we are building an end-to-end app platform to help customers build their own copilots and agents. Azure OpenAI usage more than doubled over the past 6 months as both digital natives like Grammarly and Harvey as well as established enterprises like Bajaj Finance, Hitachi, KT and LG move apps from test to production. GE Aerospace, for example, used Azure OpenAI to build a new digital assistant for all 52,000 of its employees. In just 3 months, it has been used to conduct over 500,000 internal queries and process more than 200,000 documents…

…This quarter, we added support for OpenAI’s newest model family, o1. We’re also bringing industry-specific models through Azure AI, including a collection of best-in-class multimodal models for medical imaging…

…Azure AI is also increasingly an on-ramp to our data and analytics services. As developers build new AI apps on Azure, we have seen an acceleration of Azure Cosmos DB and Azure SQL DB hyperscale usage as customers like Air India, Novo Nordisk, Telefonica, Toyota Motor North America and Uniper take advantage of capabilities purpose built for AI applications. 

Azure is offering its full catalog of AI models directly within the GitHub developer workflow; GitHub Copilot enterprise customers grew 55% sequentially in 2024 Q3; GitHub Copilot now has agentic workflows, such as Copilot Autofix, which helps users fix code 3x faster than it would take them on their own

And with the GitHub models, we now provide access to our full model catalog directly within the GitHub developer workflow…

… GitHub Copilot is changing the way the world builds software. Copilot enterprise customers increased 55% quarter-over-quarter as companies like AMD and Flutter Entertainment tailor Copilot to their own code base. And we are introducing the next phase of AI code generation, making GitHub Copilot agentic across the developer workflow. GitHub Copilot Workspace is a developer environment, which leverages agents from start to finish so developers can go from spec to plan to code all in natural language. Copilot Autofix is an AI agent that helps developers at companies like Asurion and Auto Group fix vulnerabilities in their code over 3x faster than it would take them on their own. We’re also continuing to build on GitHub’s open platform ethos by making more models available via GitHub Copilot. And we are expanding the reach of GitHub to a new segment of developers introducing GitHub Spark, which enables anyone to build apps in natural language.

Microsoft 365 Copilot has a new Pages feature, which management thinks is the first new digital artefact for the AI age; Pages helps users brainstorm with AI and collaborate with other users; Microsoft 365 Copilot responses are now 2x faster and 3x better; daily users of Microsoft 365 have more than doubled sequentially; Microsoft 365 copilot saves Vodafone employees 3 hours per person per week, and will be rolled out to 68,000 employees; 70% of the Fortune 500 now use Microsoft 365 Copilot; Microsoft 365 copilot is being adopted at a faster rate than any other new Microsoft 365 feature; with Copilot Studio, organisations can build autonomous agents to connect with Microsoft 365 Copilot; more than 10,000 organisations have used Copilot Studio, up 2x sequentially; monthly active users of Copilot across Microsoft’s CRM and ERP portfolio grew 60% sequentially

We launched the next wave of Microsoft 365 Copilot innovation last month, bringing together web, work, and Pages as the new design system for knowledge work. Pages is the first new digital artifact for the AI age, and it’s designed to help you ideate with AI and collaborate with other people. We’ve also made Microsoft 365 Copilot responses 2x faster and improved response quality by nearly 3x. This innovation is driving accelerated usage, and the number of people using Microsoft 365 daily more than doubled quarter-over-quarter. We are also seeing increased adoption from customers in every industry as they use Microsoft 365 Copilot to drive real business value. Vodafone, for example, will roll out Microsoft 365 Copilot to 68,000 employees after a trial showed that, on average, they save 3 hours per person per week. And UBS will deploy 50,000 seats in our largest finserve deal to date. And we continue to see enterprise customers coming back to buy more seats. All up, nearly 70% of the Fortune 500 now use Microsoft 365 Copilot, and customers continue to adopt it at a faster rate than any other new Microsoft 365 suite…

…With Copilot Studio, organizations can build and connect Microsoft 365 Copilot to autonomous agents, which then delegate to Copilot when there is an exception. More than 100,000 organizations from Nsure, Standard Bank and Thomson Reuters to Virgin Money and Zurich Insurance have used Copilot Studio to date, up over 2x quarter-over-quarter…

…Monthly active users of Copilot across our CRM and ERP portfolio increased over 60% quarter-over-quarter. 

Azure is bringing AI to industry-specific workflows; DAX Copilot is used in over 500 healthcare organisations to document more than 1.3 million physician-patient encounters each month; DAX Copilot is growing revenue faster than GitHub Copilot did in its first year

We’re also bringing AI to industry-specific workflows. One year in, DAX Copilot is now documenting over 1.3 million physician-patient encounters each month at over 500 health care organizations like Baptist Medical Group, Baylor Scott & White, Greater Baltimore Medical Center, Novant Health and Overlake Medical Center. It is showing faster revenue growth than GitHub Copilot did in this first year. And new features extend DAX beyond notes, helping physicians automatically draft referrals, after-visit instructions and diagnostic evidence.

LinkedIn’s AI tools help hirers find qualified candidates faster, and hirers who use AI assistant messages see a 44% higher acceptance rate

LinkedIn’s first agent hiring assistant will help hirers find qualified candidates faster by tackling the most time-consuming task. Already hirers who use AI assistant messages see a 44% higher acceptance rate compared to those who don’t. And our hiring business continues to take share.

In September 2024, Microsoft introduced a new AI companion experience – powered by Copilot – that includes voice and vision capabilities, allowing users to browse and converse with Copilot simultaneously

With Copilot, we are seeing the first step towards creating a new AI companion for everyone with new Copilot experience we introduced earlier this month, includes a refreshed design and tone along with improved speed and fluency across the web and mobile. And it includes advanced capabilities like voice and vision that make it more delightful and useful and feel more natural. You can both browse and converse with Copilot simultaneously because Copilot sees what you see. 

Roughly half of Microsoft’s cloud and AI-related capex in 2024 Q3 (FY2025 Q1) are for long-lived assets that will support monetisation over the next 15 years and more, while the other half are for CPUs and GPUs; the capex spend for CPUs and GPUs are made based on demand signals; management will be looking at inference demand to govern the level of AI capex for training; management sees that growth in capex will eventually slow and revenue growth will increase, but how fast that happens will depend on the pace of adoption of AI; the capex that Microsoft has been committing is a sign of management’s commitment to grow together with OpenAI, and to grow Azure beyond OpenAI; Microsoft is currently not interested at all in selling GPUs for companies to train AI models and has turned such business away, and this gives management conviction about the company’s AI-related capex

Capital expenditures including finance leases were $20 billion, in line with expectations, and cash paid for PP&E was $14.9 billion. Roughly half of our cloud and AI-related spend continues to be for long-lived assets that will support monetization over the next 15 years and beyond. The remaining cloud and AI spend is primarily for servers, both CPUs and GPUs, to serve customers based on demand signals…

…The inference demand ultimately will govern how much we invest in training because that’s, I think, at the end of the day, you’re all subject to ultimately demand…

…I think in some ways, it’s helpful to go back to the cloud transition that we worked on over a decade ago, I think, in the early stages. And what you did see and you’ll see us do in the same time is you have to build to meet demand. Unlike the cloud transition, we’re doing it on a global basis in parallel as opposed to sequential given the nature of the demand. And then as long as we continue to see that demand grow, you’re right, the growth in CapEx will slow and the revenue growth will increase. And those 2 things, to your point, get closer and closer together over time. The pace of that entirely depends really on the pace of adoption…

…[Question] How does Microsoft manage the demands on CapEx from helping OpenAI with its scaling ambitions?

[Answer] I’m thrilled with their success and need for supply from Azure and infrastructure and really what it’s meant in terms of being able to also serve other customers for us. It’s important that we continue to invest capital to meet not only their demand signal and needs for compute but also from our broader customers. That’s partially why you’ve seen us committing the amount of capital we’ve seen over the past few quarters, is our commitment to both grow together and for us to continue to grow the Azure platform for customers beyond them…

…One of the things that may not be as evident is that we’re not actually selling raw GPUs for other people to train. In fact, that’s sort of a business we turn away because we have so much demand on inference that we are not taking what I would — in fact, there’s a huge adverse selection problem today where people — it’s just a bunch of tech companies still using VC money to buy a bunch of GPUs. We kind of really are not even participating in most of that because we are literally going to the real demand, which is in the enterprise space or our own products like GitHub Copilot or M365 Copilot. So I feel the quality of our revenue is also pretty superior in that context. And that’s what gives us even the conviction, to even Amy’s answers previously, about our capital spend, is if this was just all about sort of a bunch of people training large models and that was all we got, then that would be ultimately still waiting, to your point, for someone to actually have demand, which is real. And in our case, the good news here is we have a diversified portfolio. We’re seeing real demand across all of that portfolio.

Microsoft’s management continues to expect Azure’s growth to accelerate in FY2025 H2, driven by increase in AI capacity to meet growing demand

In H2, we still expect Azure growth to accelerate from H1 as our capital investments create an increase in available AI capacity to serve more of the growing demand.

Microsoft’s management thinks that the level of supply and demand for AI compute will match up in FY2025 H2

But I feel pretty good that going into the second half of even this fiscal year, that some of that supply/demand will match up…

…I do, as you heard, have confidence, as we get a good influx of supply across the second half of the year particularly on the AI side, that we’ll be better able to do some supply-demand matching and hence, while we’re talking about acceleration in the back half.

Microsoft’s management sees Microsoft’s partnership with OpenAI as having been super beneficial to both parties; Microsoft provides the infrastructure for OpenAI to innovate on models; Microsoft takes OpenAI’s models and innovates further, through post-training of the models, building smaller models, and building products on top of the models; management developed conviction on the OpenAI partnership after seeing products such as GitHub Copilot and DAX Copilot get built; management feels very good about Microsoft’s investment in OpenAI; Microsoft accounts for OpenAI’s financials under the equity method

The partnership for both sides, that’s OpenAI and Microsoft, has been super beneficial. After all, we were the — we effectively sponsored what is one of the most highest-valued private companies today when we invested in them and really took a bet on them and their innovation 4, 5 years ago. And that has led to great success for Microsoft. That’s led to great success for OpenAI. And we continue to build on it, right? So we serve them with world-class infrastructure on which they do their innovation in terms of models, on top of which we innovate on both the model layer with some of the post-training stuff we do as well as some of the small models we build and then, of course, all of the product innovation, right? One of the things that my own sort of conviction of OpenAI and what they were doing came about when I started seeing something like GitHub Copilot as a product get built or DAX Copilot get built or M365 Copilot get built…

… And the same also, I would say, we are investors. We feel very, very good about sort of our investment stake in OpenAI…

…  I would say, just a reminder, this is under the equity method, which means we just take our percentage of losses every quarter. And those losses, of course, are capped by the amount of investment we make in total, which we did talk about in the Q this quarter as being $13 billion. And so over time, that’s just the constraint, and it’s a bit of a mechanical entry. And so I don’t really think about managing that. That’s the investment and acceleration that OpenAI is making in themselves, and we take a percentage of that.

Microsoft’s management sees Copilot as the UI layer for humans to interact with AI; Copilot Studio is used to build AI agents to connect Copilot to other systems of the user’s choice; Copilot Studio can also be used to create autonomous AI agents but these AI agents are not fully autonomous because at some point, they will need to notify a human or require an input and that is where Copilot comes in again

The system we have built is Copilot, Copilot Studio, agents and autonomous agents. You should think of that as the spectrum of things, right? So ultimately, the way we think about how this all comes together is you need humans to be able to interface with AI. So the UI layer for AI is Copilot. You can then use Copilot Studio to extend Copilot. For example, you want to connect it to your CRM system, to your office system, to your HR system. You do that through Copilot Studio by building agents effectively.

You also build autonomous agents. So you can use even — that’s the announcement we made a couple of weeks ago, is you can even use Copilot Studio to build autonomous agents. Now these autonomous agents are working independently, but from time to time, they need to raise an exception, right? So autonomous agents are not fully autonomous because, at some point, they need to either notify someone or have someone input something. And when they need to do that, they need a UI layer, and that’s where, again, it’s Copilot.

So Copilot, Copilot agents built-in Copilot Studio, autonomous agents built in Copilot Studio, that’s the full system, we think, that comes together.

Netflix (NASDAQ: NFLX)

Within entertainment, Netflix’s management thinks the most important question for AI is whether it can help creators produce even better content; the ability of AI to reduce costs in content creation is of secondary importance

 Lots of hype, good and bad, about how AI is going to impact or transform the entertainment industry. I think that the history has been that entertainment and technology have worked hand-in-hand throughout the history of time. And it’s very important, I think, for creators to be very curious about what these new tools are and what they could do. But AI needs to pass a very important test. Actually, can it help make better shows and better films? That is the test and that’s what they got to figure out. But I’ve said this before and I will say it again. We benefit greatly from improving the quality of the movies and the shows much more so than we do from making them a little cheaper. So any tool that can go to enhance the quality, making them better is something that is going to actually help the industry a great deal.

Paycom Software (NYSE: PAYC)

Paycom’s management developed an AI agent internally for the company’s service team to help the team provide even better service; the AI agent improved Paycom’s immediate response rates by 25% without any additional human interaction; the AI agent was built in house; Paycom is using AI in other areas, such as in several existing and upcoming products

Internally, we developed and deployed an AI agent for our service team. This technology utilizes our own knowledge-based semantic search model and enables us to provide service to help our clients more quickly and consistently than ever before.The AI agent continually improves over time and is having an impact on helping our clients achieve even more value out of their relationship with Paycom. By utilizing our own AI agent, we were able to connect our clients to the right solution faster, improving our immediate response rates by 25% without any additional human interaction…

…[Question] Interesting to hear about using AI in the customer service organization. I’m curious if that’s technology that Paycom has built or if you’re using a third party.

[Answer] So that’s internal. We built it ourselves, and we’ve been using it. And so it gets better and better as we mentioned on the call. It’s sped up our process by 25% as far as being able to connect clients to the solution quicker, whether that be a configuration question, a tax question or what have you. And so that’s really been helpful to us, and it continues to do more and more from that perspective…

…[Question] A follow-up on the AI agent or the AI technology that you’re developing. Do you see an opportunity in the future to productize what you’re developing internally, maybe like in your — in future versions of your recruiting product or other products in your platform?

[Answer] I would say this isn’t the only area in which we’re using AI. We have it in several products that we both have released and will be releasing. And so there’s definitely opportunities to monetize AI. As far as this particular solution, it’s really helping us on the back end and helping our client as well. So I think we’re going to see results and benefits from that in other areas of efficiency across the board within our own organization.

Shopify (NASDAQ: SHOP)

Shopify recently enhanced Shopify Flow, a low-code workflow automation app, with a new admin API connector that provides an additional 304 new automation actions

Let’s start with Shopify Flow. A low-code workflow automation app that empowers merchants to build custom automations and help them run their businesses more efficiently. This includes a new automation trigger based on the merchant’s custom data and newly completed admin API connector that provides an additional 304 new actions to use in their automations. And as a result, Flow has become a much more powerful tool, enabling merchants to update products, process customer form submissions, edit orders and so much more.

The Shopify Inbox feature now uses AI to suggest personalised replies for merchants to respond to customer inquiries; half of merchants’ responses are now using the AI-suggested replies; fast customer response helps lift conversion rates for merchants; the replies feature may not seem like a big deal, but it actually helps free up a lot of time for merchants to focus on building products

Within Shopify Inbox, this product now uses AI to suggest replies based on each merchant’s unique store information making it super easy for merchants to respond quickly and accurately to customer inquiries. In fact, on average, merchants are using the Suggest Replies for about half of their responses, edited or not, showing just how effective this feature has become. Replying can quickly boost conversion rates, which means more sales for our merchants and in turn, for Shopify…

…I mentioned suggest replies in Shopify Inbox, which may not seem like a big deal, but it’s a huge deal because it means merchants can spend more of their time focused on the things that they need to be focused on like building our products.  

The Shop App has a new merchant-focused home feed that is powered by machine learning models to increase shopper engagement; the new home feed has led to an 18% increase in sessions where a buyer engaged with a recommendation; management thinks the combination of search with AI will make the search function on the Shop App a lot more relevant and personalised

This quarter, the Shop App launched a new merchant-focused home feed, showcasing the diversity and the richness of brands on Shop. The experience uses new machine learning models to help buyers keep up with the brands they love and discover new brands based on their preferences. These changes have already led to early success with an 18% increase in sessions where a buyer engaged with a recommendation…

…We also think Search and AI together makes the Shop search way more relevant, way more personalized. That is also very compelling.

Essentially every Shopify internal department is using AI to be more productive

Support engineering, sales, finance, just about every department internally is using AI in some way to get more efficient, more productive.

Shopify’s management thinks the integration of AI in search will change how consumers find merchants and products, but Shopify has helped merchants navigate many similar changes before, and Shopify will continue to help merchants navigate the AI-related changes

In terms of where consumers find merchants or find products, yes, AI and search is going to change. But to be clear, this entire flow and discovery process has been changing for many years. It’s the reason that you saw us integrate with places like YouTube or more recently, Roblox or TikTok or Instagram…

…You can rest assured that when consumers shift their buying preferences, their discovery preferences, their search preferences, and they’re looking for great products from great brands, Shopify will ensure that our merchants are able to do so. And that’s the reason even some of the more nuanced or some of the more — as you know, Shopify has an integration to Spotify. Why? Because some merchants that also have very large followings as a musician have massive followings on their artist profile, the fact that so you can now show Shopify products on your artist profile means for that particular segment of merchants, they can easily — they now have a new surface area in which to conduct business. And that’s the same thing when it comes to AI and search. 

Taiwan Semiconductor Manufacturing Company (NYSE: TSM)

TSMC’s management expects TMSC’s business in 2024 Q4 to be supported by strong AI-related demand; management sees very strong AI demand in 2024 H2, leading to higher capacity utilisation rate for TSMC’s leading-edge 3nm and 5nm process technologies; management now expects server AI processors to account for mid-teens percent of TSMC’s total revenue in 2024 (previous expectation was for low-teens percent)

Moving into fourth quarter. We expect our business to continue to be supported by strong demand for our leading-edge process technologies. We continue to observe extremely robust AI-related demand from our customers throughout the second half of 2024, leading to increasing overall capacity utilization rate for our leading-edge 3-nanometer and 5-nanometer process technologies…

…We now forecast the revenue contribution from server AI processors to more than triple this year and account for mid-teens percentage of our total revenue in 2024.

TSMC’s management defines server AI processors as GPUs, AI accelerators, and CPUs for training and inference

At TSMC, we defined server AI processor as GPUs, AI accelerators and CPUs performing training and inference functions and do not including — include networking, edge or on-device AI.

TSMC’s management thinks AI demand is real, based on TSMC’s own experience of using AI and machine learning in its operations; a 1% productivity gain for TSMC is equal to a tangible NT$1 billion return on investment (ROI); management thinks TSMC is not the only company that has benefitted from AI applications

Whether this AI demand is real or not, okay, and my judgment is real, we have talked to our customers all the time, including hyperscaler customers who are building their own chips, and almost every AI innovators is working with TSMC. And so we probably get the deepest and widest look of anyone in this industry. And why I say it’s real? Because we have our real experience. We have using the AI and machine learning in our fab and R&D operations. By using AI, we are able to create more value by driving greater productivity, efficiency, speed, qualities. And think about it, let me use, 1% productivity gain, that was almost equal to about TWD 1 billion to TSMC. And this is a tangible ROI benefit. And I believe we cannot be the only one company that have benefited from this AI application. So I believe a lot of companies right now are using AI and — for their own improving productivity, efficiency and everything.

TSMC’s management thinks AI demand is just at the beginning

[Question] Talk a little bit about what you think about the duration of this current semiconductor up-cycle? Do you think it will continue into the next couple of years? Or are we getting closer to the peak of the cycle?

[Answer] The demand is real and I believe it’s just the beginning of this demand, all right? So one of my key customers said, the demand right now is insane, that it’s just the beginning. It’s [ a form of scientific ] to be engineering, okay? And it will continue for many years.

When TSMC builds fabs to meet AI demand, management has a picture in mind of what the long-term demand picture looks like

[Question] Keen to understand how TSMC gets comfortable with customer demand for AI beyond 2025. And I ask this because it takes a couple of years before you can build a fab, so you need to be taking early — an early view on what does AI look like in 2026, 2027. So how are you specifically cooperating on long-term plans for capacity with these AI customers? And what commitments are these customers giving you?

[Answer]  let me say again that we did talk to a lot of our customers. Almost every AI innovator are working with us and that’s including the hyperscalers. So if you look at the long-term market — long-term structure and market demand profile, I think we have some picture in our mind and we make some judgment, of course, and we work with them on a rolling basis. So how we prepare our capacity, actually, just like Wendell said, we have a disciplined and [ a rollout ] system to plan the appropriate level of capacity. And that — to support our customers’ need, also to maximize our shareholders’ value. That’s what we’re always keeping our mind.

There’s more AI content that goes into the chips in PCs (personal computers) and smartphones; management expects the PC and smartphone business of TSMC to be healthy in the next few years because of AI-related applications 

The unit growth of PC and smartphone is still in the low single digit. But more importantly is the content. The content now we put more AI into that, they are cheap and so the silicon area increased faster than the unit growth. So again, I would like to say that for this PC and the smartphone business, not — is gradually increased and we expect it to be healthy in the next few years because of our AI-related applications.

Advanced packaging is currently a high single-digit percentage of TSMC’s revenue and management expects it to grow faster than TSMC’s overall business over the next 5 years; the margins of advanced packaging are improving, but it’s not at the corporate average level yet

Advanced packaging in the next several years, let’s say, 5 years, will be growing faster than the corporate average. This year, it accounts for about high single digit of our revenue. In terms of margins, yes, it is also improving. However, it’s still — it’s approaching corporate, but not there yet.

Demand for TSMC’s CoWoS (advanced packaging) continues to far exceed supply, even though TSMC has doubled CoWoS capacity compared to a year ago and will double it again

Let me share with you today’s situation is our customer’s demand far exceed our ability to supply. So even we work very hard and increase the capacity by about more than twice, more than 2x as of this year compared with last year and probably double again, but still not enough. And — but anyway, we are working very hard to meet the customers’ requirement.

Tencent (NASDAQ: TCEHY)

Tencent’s management is increasingly seeing tangible benefits from deploying AI across the company’s business; management wants to continue investing in AI; the most significant benefits are in content recommendation and targeting, which directly benefits Tencent’s business and advertising revenue; management also sees AI as a productivity tool, as Tencent’s Copilot is being used by Tencent’s software engineers frequently and is helping them generate efficiency gains; management is trying to incorporate AI into a lot of Tencent’s products, but they think it will take a few more quarters before real use cases show up

We are increasingly seeing a tangible benefit of deploying AI across our products and operations, including marketing services and cloud. And we’ll continue investing in AI technology, tools and solutions that assist users and partners…

…I think that the most significant one right now is actually around content recommendation and at targeting because the AI in — the AI engine in those two use cases are generating a significant amount of additional user time and at the same time, it’s generating a higher incremental targeting rate, response rate for our apps and both of them actually are direct benefits to the business and direct benefit to ad revenue. and both of the video accounts and our performance at revenue actually at scale…

… It’s actually a productivity tool that everybody is using on a frequent basis, for example, our Copilot is being used by our engineers across the board on a very frequent basis, and it’s actually generating efficiency gains for our business. and different businesses, a lot of our products are actually testing our Hunyuan and trying to incorporate AI into the — either the production process, right, so that they would gain efficiency or in the user experience use case so that it can actually make their user experience better. So I would say, right now, we are seeing more and more adoption among all our different products and services. It would take probably a few more quarters for us to see some real use cases at scale. 

Tencent’s management used the company’s foundation AI model, Tencent Hunyuan, to facilitate tagging and categorisation of content and advertising materials; Tencent also upgraded its machine learning platforms to deliver better advertising targeting; marketing services revenue from video accounts was up 60% year-on-year; Mini Programs marketing services revenue had robust growth; Tencent used large language models (LLMs) to improve the relevance of Weixin Search results, leading to higher commercial queries and click-through rates, and consequently, an increase in search revenue of more than 100%

Our Marketing Services revenue grew 17% year-on-year. Strength in games and e-commerce categories outweighed weakness in real estate and food and beverage. The Paris Olympics somewhat cushioned industry-wide weakness in brand ad revenue during the third quarter but this positive factor will be absent in the fourth quarter. We leveraged our foundation model, Tencent Hunyuan to facilitate tagging and categorization of content and ad materials. And we upgraded our machine learning platforms to deliver more accurate ad targeting.

By property, video accounts marketing services revenue increased over 60% year-on-year. As we systematically strengthen transaction capabilities in Weixin, advertisers increasingly utilize our marketing tools to boost their exposure and drive sales conversion. Mini Programs marketing services revenue grew robustly year-on-year as our Mini Games and Mini Dramas provided high-value rewarded video ad inventory and generated incremental closed-loop demand. And for Weixin Search, we utilized large language model capabilities to facilitate understanding of complex queries and content, enhancing the relevance of search results. Commercial queries increased and click-through rate improved, and our search revenue more than doubled year-on-year.

Tencent enjoyed swift year-on-year growth in GPU-focused cloud revenue and this revenue stream is now a teens percentage of Tencent’s infrastructure as a services revenue; Tencent has released Tencent Hunyuan Turbo, the new generation of its foundation AI model, which uses a heterogeneous mixture of experts architecture; compared to the previous generation, Hunyuan Turbo’s training and inference efficiency has doubled while its inference costs has halved; Hunyuan Turbo is ranked first for general capabilities among foundation AI models in China; Tencent has open-sourced Hunyuan models; management sees Tencent’s AI revenue being lesser than US cloud companies because China does not have a large enterprise, SaaS, and startup markets for AI services 

Our cloud revenue from GPUs primarily used for AI grew swiftly year-on-year and now represents a teens percentage of our infrastructure as a services revenue. We released Tencent Hunyuan Turbo, which utilizes a heterogeneous mixture of experts architecture, doubling our training and inference efficiency and halving inference cost versus its predecessor Hunyuan Pro. SuperCLUE ranked Hunyuan Turbo first for general capabilities among domestic peers. Last week, we made the Hunyuan large model and the Hunyuan 3D generation models available on an open-source basis. Our international cloud revenue increased significantly year-on-year. We leveraged domain expertise in areas such as games and live streaming and competitive pricing to win international customers…

…The IAS revenue is now in the teens generated by AI. But having said that, we think the amount of AI revenue is actually less than U.S. cloud companies. And the main reason is because, number one, China doesn’t really have a every big enterprise market. And if you look at the U.S., a lot of enterprises are actually sort of fitted in with AI and the — in testing out how AI can do for their business that they’re actually buying a lot of compute, which is not happening in China yet. There’s a very big SaaS ecosystem in the U.S., which everybody is actually trying to add AI to their functionality and thus charge the customers more. And that SaaS ecosystem is not really that vibrant in China. And thirdly, there are also fewer AI start-ups in China, which are actually buying a lot of compute. So as a result, the AI revenue in China on the cloud side is somewhat sort of at scale for us, but I think it will not be exploding like in the U.S. 

Tencent’s management does not want to embed commercial search results into the company’s AI chatbot, YongBao right now; the current focus for YongBao is on growing usage, not monetisation

[Question] Will you ramp up the Gen AI chatbot, would that eventually embed with the commercial sponsor answer as well?

[Answer] In terms of whether YongBao will embed commercial search results, the answer is no. for the current time, we’re focused on making YongBao be as appealing and attractive to users as it can be and we’re not focused on premature monetization.

Tencent’s management plans to invest in capex for AI, but the amount of investment will be small compared to the companies in the USA

If you look at CapEx, right, we believe we have a progressive CapEx plan, especially given that the development of a cloud business and the advent of AI, but at the same time, it’s measured compared to a lot of the U.S. companies. 

Tencent’s management sees the company’s advertising business being driven by 3 factors, namely consumer spending, Tencent’s ability to utilise AI to continue boosting click-through rates from currently low levels, and deployment of more inventory

In terms of the drivers for 2025, the overall macro environment would obviously be important accelerator or decelerator or neutral force for the aggregate advertising market. And that in turn will be a function primarily of consumer confidence. And consumer spending behavior. Now within that overall environment, our relative performance will be a function of, first of all, our advertising technology and our ability to utilize GPUs, utilize neural networks to continue boosting click-through rates from the current very low levels to higher levels that mechanically translates into more revenue. And then secondly, our deployment of specific inventories, in particular, video accounts, in particular, Weixin Search.

Tesla (NASDAQ: TSLA)

Tesla’s management released FSD v12.5 in 2024 Q3, which has increased data and training compute, and 5x increase in parameter count; Tesla also released Actually Smart Summon (your vehicle will autonomously drive to you in parking lots) and FSD for Cybertruck, which includes end-to-end neural nets for highway driving for the first time; version 13 of FSD is coming soon and it is expected to have a 5-6 fold improvement in miles between interventions compared to version 12.5; over the course of 2024, FSD’s improvement in miles between interventions has been at least 3 orders of magnitude; management expects FSD to become safer than human in 2025 Q2; Tesla vehicles on autopilot have 1 crash per 7 million miles, compared to 1 crash per 700,000 miles for the US average; Tesla has earned $236 million in revenue in 2024 Q3 from the release of FSD for Cybertruck and Actually Smart Summon

In Q3, we released the 12.5 series of FSD (Supervised)1 with improved safety and comfort thanks to increased data and training compute, a 5x increase in parameter count, and other architectural choices that we plan to continue scaling in Q4. We released Actually Smart Summon, which enables your vehicle to autonomously drive to you in parking lots, and FSD (Supervised) to Cybertruck customers, including end-to-end neural nets for highway driving for the first time…

…Version 13 of FSD is going out soon… We expect to see roughly a 5- or 6-fold improvement in miles between interventions compared to 12.5. And actually, looking at the year as whole, the improvement in miles between interventions, we think will be at least 3 orders of magnitude. So that’s a very dramatic improvement in the course of the year, and we expect that trend to continue next year.  The current total expectation, internal expectation for the Tesla FSD having longer miles between interventions [indecipherable] is the second quarter of next year, which means it may end up being in the third quarter but it’s next — it seems extremely likely to be next year…

…miles between critical interventions, mentioned by Elon already made 100x improvement with 12.5 from starting of this year and then with v13 release, we expect to be 1,000x from the beginning, from January of this year on production software. And this came in because of technology improvements going to end-to-end, having higher frame rate, partly also helped by hardware force, more capabilities, so on. And we hope that we continue to scale the neural network, the data, the training compute, et cetera. By Q2 next year, we should cross over the average, even in miles per critical intervention [indiscernible] in that case…

…Our internal estimate is Q2 of next year to be safer than human and then to continue with rapid improvements thereafter…

… So we published Q3 vehicle safety report, which shows 1 crash for every 7 million miles on autopilot that compares with the U.S. average of crash roughly every 700,000 miles. So it’s currently showing a 10x safety improvement relative to the U.S. average…

…We released FSD for Cybertruck and other features like actually small [indiscernible] like Elon talked about in North America, which contributed $326 million of revenues in the quarter. 

Tesla has deployed a 29,000 H100 cluster and expects to have a 50,000 H100 cluster by the end of October 2024, to support FSD and Optimus; Tesla is not training compute-constrained; Tesla’s AI has gotten so good that it now takes a long time to decide which version of the software is better because mistakes happen so infrequently and that is the big bottleneck to Tesla’s AI development; management is being very careful with AI-spending

We deployed and are training ahead of schedule on a 29k H100 cluster at Gigafactory Texas – where we expect to have 50k H100 capacity by the end of October…

…We continue to expand our AI training capacity to accommodate the needs of both FSD and Optimus. We are currently not training compute-constrained. [indiscernible] probably the big limiting factors of the FSD is actually getting so good that it takes us a while to actually find mistakes. And when you start getting to where it can take 10,000 miles to find a mistake, it takes a while to actually figure out which it is, is software A better than software B? It actually takes a while to figure it out because neither 1 of them makes the mistakes, would take a long time to make mistakes. So it’s actually the single biggest limiting factor is how long does it take us to figure out which version is better? Sort of a high-class problem…

… One thing which I’d like to elaborate is that we’re being really judicious on our AI compute spend to and saying how best we can utilize the existing infrastructure before making further investments…

…We still got to take which models are performing better. So the validation network to picking the models because as mentioned the miles between intervention is pretty large. We had to drive a lot of miles going close to. We do have simulation and other ways to get those metrics. Those 2 help, but in the end, that’s a big bottleneck. That’s why we’re not training-compete constrained alone. 

In the 10 October 2024 “We, Robot” event by Tesla, the company had showcased 50 autonomous vehicles, including 20 Cybercabs; the Cybercabs had no steering wheel, brake, or accelerator pedals, so they were truly autonomous

On October 10, we laid out a vision for an autonomous and future that I think is very compelling that the Tesla team did a phenomenal job there with actually giving people an option to experience the future, where you have humanoid robots working among the craft, not with a canned video and a presentation or anything but walking among crowd so he drinks and whatnot. And we had 50 autonomous vehicles. There were 20 Cybercabs but there were an additional 30 Model Ys, operating fully autonomously the entire night, carrying thousands of people with no incidents the entire night…

…Worth emphasizing that the Cybercab had no steering wheel or brake or accelerator panels, meaning there was no way for anyone to intervene manually a unit if they wanted to and the whole night went very smoothly.

Tesla is already offering autonomous ridehailing for Tesla employees in the Bay Area; the ridehailing service currently has a safety driver; Tesla has been testing autonomous ridehailing for some time; Elon Musk expects ridehailing to be rolled out to the public in California and Texas in 2025, and maybe other states in the USA; California has a lot of regulations around ridehailing, but there’s still a regulatory pathway; Tesla actually has passed Federal regulations for ridehailing, but it’s the state level where there are problems

We have for Tesla employees in the Bay Area, we already are offering ridehailing capabilities. So you can actually, with the development app, you can request a ride and it will take you anywhere in the Bay Area. We do have a safety driver for now but it’s not required to do that…

… We’ve been testing it for the good part of the year. And the building blocks that we needed in order to build this functionality and deliver it to production, we’ve been thinking about working on for years…

…So it’s not like we’re just starting to think about this stuff right now while we’re building out the early stages of our ridehailing network. We’ve been thinking about this for quite a long time, and we’re excited to get the functionality out there…

…We do expect to roll out ridehailing in California and Texas next year to the public. Now California is somewhere — there’s quite a long regulatory approval process. I think we should get approval next year but it’s contingent upon regulatory approval. Texas is a lot faster so it’s — we’ll definitely have available in Texas and probably have it available in California, subject to regulatory approval. And then — and maybe some other states actually next year as well, but at least California and Texas…

…[Question] Elon mentioned unsupervised FSD in California and Texas next year. Does that mean regulators have agreed to it in the entire state for existing hardware 3 and 4 vehicles?

[Answer] As I said earlier, California loves regulation… here’s a pathway. Obviously, Waymo operates in California so there’s just a lot of forms and a lot of approvals that are required. I mean, I’d be shocked if we don’t get approved next year, but it’s just not something we totally control. But I think we will get approval next year in California and Texas. And towards the Bay Area, branch out beyond California and Texas…

…I think it’s important to reiterate this like on our certifying a vehicle at the federal level in the U.S. is done by meeting FMVSS regulations. Our vehicles today that are produced there capable to meet all those regulations, the Cybercab regulations. And so the deployment of the vehicle to the road is no limitation, but its limitation is what you said at the state level where they control autonomous vehicle deployment. Some states are relatively easy, as you mentioned, for Texas. It’s other ones have always like California that may take a little longer. The other ones hadn’t set up anything yet. 

Tesla’s management acknowledges that there’s a chance that Tesla vehicles with Hardware Version 3 may not support unsupervised full self-dricing, and if so, Tesla will replace the hardware for those vehicle fleets for free into Hardware Version 4

By some measure, Hardware 4 has really several times the capability of Hardware 3. It’s easier to get things to work with then it takes a lot of effort to sort of squeeze that box analyst hat Hardware 3. And there is some chance that Hardware 3 is — does not achieve the safety level that allows for unsupervised FSD. There is some chance of that. And if that turns out to be the case, we will upgrade those group bought Hardware 3 FSD for free. And we have designed the system to be upgradeable so it’s really just to sort of switch out the computer thing, the camera, the cameras are capable. But we don’t actually know the answers of that. But if it does turn out, we’ll make sure we take care of those who are.

Tesla’s management thinks real-world AI in self-driving cars is different from LLMs (large language models) in that (1) real-world AI requires massive amounts of context that needs to be processed with a small amount of compute power and the way around this limitation is to do massive amounts of training so that the amount of inference that needs to be done is tiny, and (2) it’s difficult to sort out what data coming in from the video feed is important for the training

The nature of real world AI is different from LLM in that you have a massive amount of context. So like the — you’ve got a case of Tesla cameras that [indiscernible] if you include tunnel camera that — so you’ve got some context. And that is then distilled down into a small number of control outputs, whereas it’s like it’s very rare to have, in fact, I’m not sure any LLM out there can do gigabytes of context. And then you’ve got to then process that in the car with a very small amount of compute power. It’s all doable and it’s happening, but it is a different problem than what, say, a Gemini or OpenAI is doing.

And now part of the way you can make up for the fact that the inference computer is quite small, it is by spending a lot of effort on training. And just like a human the way you train on something, the less metal work takes when you try to — when you do it, like when the first time like a driving it absorbs your whole mind. But then as you train more and more on driving then the driving becomes a background task. It doesn’t — it only absorbs a small amount of your mental capacity because you have a lot of training. So we can make up for the fact that the inference computers — it’s tiny compared to a 10-kilowatt bank of GPUs because you’ve got a few hundred watts of inference compute. We can make up that with heavy training.

And then there’s also vast amounts to the actual petabytes of data coming in are tremendous. And then sorting out what training is important, of the vast amounts of video data coming in the feed, what is actually most important for training. That’s also quite difficult.

Tesla’s management thinks Elon Musk’s xAI AI-startup has been helpful to Tesla, but the 2 companies are focused on very different kinds of AI problems 

Well, I should say that xAI has been helpful to Tesla AI quite a few times in terms of things like scaling it, like training, just even like recently in the last week or so, improvements in training, where if you’re doing a big training run and it fails, being able to continue training and to recover from a training run, has been pretty helpful. But there are different problems. xAI actually is working on artificial general intelligence or artificial super intelligence. Tesla is autonomous cars and autonomous robots. There are different problems…

…Yes, Tesla is focused on real-world AI. And I was saying earlier, it is quite a bit different from LLM. But you have massive context in the form of video and some amount of audio, that’s going to be distilled like extremely efficient inference compute. I do think Tesla is the most efficient in the world in terms of inference compute because out of necessity, we have to be very good at efficient inference. We can’t put 10 kilowatts of GPUs in a car. We’ve got a couple of hundred watts. And it’s a pretty well designed Tesla AI chip, but it’s still a couple hundred watts. But there are different problems. I mean, the stuff at xAI. We’re running inference. I mean, it is running inference, answering questions on a 10-kilowatt rack. It’s like you can’t put that in a car. It’s a different problem.

Elon Musk created xAI because he thought there wasn’t a truth-seeking AI company being built

xAI is because I felt there wasn’t there wasn’t a truth-seeking digital super intelligence company out there, like that’s what it came down to. There needed to be a truth-seeking AI company that is very [indiscernible] about being truthful. I’m not saying xAI is perfect, but that is truth, but that is at least the explicit aspiration, even if something is politically incorrect, it would still be truhtful. I think this is very important for AI safety. So I think xAI has been helpful to Tesla and will continue to be helpful to Tesla, but they are very different problems.

There are no other car companies that has a world-class AI and chip-design team like Tesla

And like what other car company has a world-class chip design team? Like zero. What other car company has a world-class AI team like Tesla does? 0. Those were all startups that were created from scratch.

The Trade Desk (NASDAQ: TTD)

The incorporation of AI into Kokai, Trade Desk’s ad-buying platform, is encouraging adoption of Trade Desk by CFOs and CMOs

While there has been a lot of macro focus on the reduction in inflation rates, historic highs for stock market indices and growing indications of a soft landing, that’s not necessarily translating to consumer confidence, which is why CMOs are becoming much more closely aligned with their CFOs. CFOs want more evidence than ever that marketing is working. And for CFOs that doesn’t just mean traditional marketing KPIs. It means growing the top line business. All of our AI and data science injection into Kokai, our latest product release, is encouraging CMOs and CFOs to lean more and more on TTD to deliver real, measured growth…

…When CMOs faced pressure to achieve more with less, they turn to platforms like ours for flexibility, precision and measurable results.

Companies need an AI strategy, and Trade Desk’s AI product, Koa, is a great copilot for advertising traders; Trade Desk has plenty of opportunities in an AI-world because of the data assets it has, and management wants to improve all aspects of the company through AI

Every company needs an AI strategy. Our AI product, Koa, is a great copilot for traders. But this is only the beginning. There are endless possibilities for us as we have 1 of the best data assets on the Internet. The learnings that come from buying the global open Internet outside of walled gardens. To win in this new frontier, we’re looking across our entire suite of products, algorithms and features and asking how they all can be advanced by AI.

Visa (NYSE: V)

For Risk and Identity Solutions within value-added services, Visa wants to acquire Featurespace, an AI payments protection tech company that will enable Visa to enhance fraud prevention tools to clients and protect consumers in real time; Worldline, a Visa partner, will be using Decision Manager to provide businesses with AI-based e-commerce fraud detection abilities; Featurespace is a world leader in providing AI solutions to fight fraud

In Risk and Identity Solutions, we recently announced our intent to acquire Featurespace, a developer of real-time artificial intelligence payments protection technology. It will enable Visa to provide enhanced fraud prevention tools to our clients and protect consumers in real-time across various payment methods.  And Worldline, already a Visa partner and leading European acquirer, will soon be launching an optimized fraud management solution, utilizing Decision Manager to provide businesses with AI-based e-commerce fraud detection capabilities…

…Featurespace is a world leader in providing AI-driven solutions to combat that fraud, to reduce that fraud, to enable our clients and partners to continue to serve their customers in a safe way.

Visa’s management sees AI as being a driver of productivity across multiple functions in the company, and as a differentiator in its products and services

[Question] I just wanted to ask how you see AI playing into the business model. Do you see it more as driving VAS or incremental business model, uplift revenue or cost improvement? Or is it more of a competitive differentiator that will just keep you ahead of your competition?

[Answer] As it relates more broadly to especially generative AI at Visa, I see it really in 2 different buckets. The first is we are adopting it aggressively across our company to drive productivity. And we’ve seen some great results from everywhere to our engineering teams, to our accounting teams, to our sales teams, our client service teams. And we’re still in the early stages of, I think, the very significant impact this will have on the productivity of our business. I also see it as a real differentiator to the products and services that we’re putting in market. You’ve heard me talk about some of the new risk capabilities, risk management capabilities, for example, that we’ve deployed in the account-to-account space, which are all enabled with generative AI. You mentioned Featurespace. We’ve had some really good success in other parts of both our value-added services business and the broader consumer payments business as well. And we’ve got a product pipeline that is very heavily tilted towards some, we think, very exciting generative AI capabilities that hopefully you’ll hear more from us on soon.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Alphabet, Amazon, Apple, ASML, Coupang, Datadog, Fiverr, Mastercard, Meta Platforms, Microsoft, Netflix, Shopify, TSMC, Tesla, and Visa. Holdings are subject to change at any time.

The US Stock Market And US Presidents

History’s verdict on how US stocks have performed under different US presidents

The US presidential election is just a few weeks away. And as usual, large swathes of participants in the US stock market are trying to predict the victor because they think it will have significant consequences on how US stocks perform. I don’t have a crystal ball. But I do have history’s verdict, thanks to excellent research from the US-based wealth management firm, Ritholtz Wealth Management, that I came across recently.

Here’s a table showing the annualised returns of the S&P 500 for each US President, going back to Theodore Roosevelt’s first term in 1901:

Table 1; Source: Ritholtz Wealth Management 

I think the key takeaway from the table is that how the US stock market performs does not depend on what political party the US President belongs to. Republican presidents have presided over bad episodes for US stocks (Herbert Hoover, Richard Nixon, and George W. Bush, for example) as well as fantastic times (Calvin Coolidge, Dwight Eisenhower, and Ronald Reagan, for example). The same goes for Democrat presidents, who have led the country through both poor stock market returns (Woodrow Wilson and Franklin Roosevelt, for example) as well as great gains (Franklin Roosevelt, Lyndon Johnson, and Barack Obama, for example). Presidents do not have that much power over the financial markets. Don’t let politics influence your investing decision-making.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I don’t have a vested interest in any company mentioned. Holdings are subject to change at any time.

More Of The Latest Thoughts From American Technology Companies On AI (2024 Q2)

A collection of quotes on artificial intelligence, or AI, from the management teams of US-listed technology companies in the 2024 Q2 earnings season.

Last month, I published The Latest Thoughts From American Technology Companies On AI (2024 Q2). In it, I shared commentary in earnings conference calls for the second quarter of 2024, from the leaders of technology companies that I follow or have a vested interest in, on the topic of AI and how the technology could impact their industry and the business world writ large. 

A few more technology companies I’m watching hosted earnings conference calls for 2024’s first quarter after I prepared the article. The leaders of these companies also had insights on AI that I think would be useful to share. This is an ongoing series. For the older commentary:

Here they are, in no particular order:

Adobe (NASDAQ: ADBE)

Adobe’s management believes that Adobe’s approach to AI is highly differentiated; the greatest differentiation is at the interface layer, as Adobe is able to rapidly integrate AI across its product portfolio and allow users to realise value

Adobe’s customer-centric approach to AI is highly differentiated across data, models and interfaces…

…Our greatest differentiation comes at the interface layer with our ability to rapidly integrate AI across our industry-leading product portfolio, making it easy for customers of all sizes to adopt and realize value from AI. 

Adobe’s Firefly models are trained on data that allow outputs to be commercially safe and management thinks this feature of being commercially safe is really important to enterprises; Adobe now has Firefly models for imaging, vector, design, and video; Firefly is in a wide range of Adobe products; Firefly has powered more than 12 billion generations since its launch in March 2023 (was 9 billion in 2024 Q1); management’s strategy is to build Firefly models into more streamlined and precise workflows within Adobe’s products; Adobe has Firefly Service APIs for organisations to generate content at scale, and the API calls tripled quarter-on-quarter; Firefly Service APIs are gaining real traction

We train our Firefly models on data that allows us to offer customers a solution designed to be commercially safe. We have now released Firefly models for imaging, vector and design and just previewed a new Firefly video model…

… Firefly-powered features in Adobe Photoshop, Illustrator, Lightroom, and Premier Pro help creators expand upon their natural creativity and accelerate productivity. Adobe Express is a quick and easy create anything application, unlocking creative expression for millions of users. Acrobat AI Assistant helps extract greater value from PDF documents. Adobe Experience Platform AI Assistant empowers brands to automate workflows and generate new audiences and journeys. Adobe GenStudio brings together content and data, integrating high-velocity creative expression with the enterprise activation needed to deliver personalization at scale…

…We have now surpassed 12 billion Firefly-powered generations across Adobe tools…

… Our strategy is to build technology that will create more streamlined and precise workflows within our tools through features like text-to-template in Express, Generative Fill in Photoshop, Generative Recolor in Illustrator, Generative Remove in Lightroom and the upcoming Generative Extend for Video and Premier Pro. We’re exposing the power of our creative tools and the magic of generative AI through Firefly Service APIs so organizations can generate and assemble content at scale…

…The introduction of the new Firefly video model earlier this week at IBC is another important milestone in our journey. Our video model, like the other models in the Firefly family, is built to be commercially safe with fine-grain control and application integration at its core. This will empower editors to realize their creative vision more productively in our video products, including Premier Pro…

…Strong demand for Firefly Services, which provide APIs, tools and services for content generation, editing and assembly, empowering organizations to automate content production while maintaining quality and control. Total API calls tripled quarter over quarter…

…Firefly Services, which is you can think of that also as a consumption model where we have that, it’s off to a really good start. Our ability to give enterprises the ability to automate content, create custom models within enterprises, we’re seeing real traction because it’s a differentiated solution and that it’s designed to be commercially safe…

…One other thing I’d just emphasize there is that the commercial safety is so important to businesses of all sizes, frankly, and that is something that we feel very, very differentiated.

Adobe released significant advancements in AI Assistant across Adobe Acrobat and Reader in 2024 Q2 (FY2024 Q3) and saw 70% sequential growth in AI interactions in AI Assistant; the advancements in AI Assistant include content creation capabilities; Tata Consultancy Services used AI Assistant in Adobe Acrobat to create event summaries of hours of conference videos in minutes; management intends to actively promote subscription plans for Adobe Acrobat and Reader that include generative AI capabilities

For decades, PDF has been the de facto standard for storing unstructured data, resulting in the creation and sharing of trillions of PDFs. The introduction of AI Assistant across Adobe Acrobat and Reader has transformed the way people interact with and extract value from these documents. In Q3, we released significant advancements, including the ability to have conversations across multiple documents and support for different document formats, saving users valuable time and providing important insights. We are thrilled to see this value translate into AI Assistant usage with over 70% quarter-over-quarter growth in AI interactions. 

In addition to consumption, we’re focused on leveraging generative AI to expand content creation in Adobe Acrobat. We’ve integrated Adobe Firefly Image Generation into our edit PDF workflows. We’ve optimized AI Assistant in Acrobat to generate content fit for presentations, e-mails and other forms of communication, and we’re laying the groundwork for richer content creation, including the generation of Adobe Express projects.

The application of this technology across verticals and industries is virtually limitless. Tata Consultancy Services recently used Adobe Premiere Pro to transcribe hours of conference videos and then used AI Assistant in Acrobat to create digestible event summaries in minutes. This allowed them to distribute newsletters on session content to attendees in real-time.

We’re excited to leverage generative AI to add value to content creation and consumption in Acrobat and Reader in the months ahead. Given the early adoption of AI Assistant, we intend to actively promote subscription plans that include generative AI capabilities over legacy perpetual plans that do not.

Adobe GenStudio is integrated across Experience Cloud and Creative Cloud and helps marketers quickly plan, create, store, deliver, and measure marketing content; Vanguard used Adobe GenStudio to increase quality engagement with investors by 176% through one-to-one personalisation, and to enjoy millions in savings

Customers are embracing the opportunity to address their content supply chain challenges with Adobe GenStudio. With native integrations across Experience Cloud and Creative Cloud, GenStudio empowers marketers to quickly plan, create, store, deliver, and measure marketing content and drive greater efficiency in their organizations. Financial services leader Vanguard is creating an integrated content supply chain to serve the strategic goal of deepening their relationships with a broad range of investors. Leveraging the GenStudio solution, Vanguard was able to increase quality engagement by 176% by focusing on one-to-one personalization and to realize millions in savings by improving content velocity and resource allocation with an end-to-end content creation workflow.

Adobe’s management has been very consistent over the past 1-1.5 years in how they have approached AI, and that is, Adobe would be developing a broad set of models for the creative community, and the models would be highly differentiated based on quality, commercial safety, and integrability into Adobe’s product portfolio

I think we’ve been incredibly consistent with what we’ve said, dating back 1 year, 1.5 years ago, where we talked about the fact that we were going to develop the broadest set of models for the creative community. And we were going to differentiate the models based on quality, commercial safety, integratability into our tools and controllability. And as you’ve seen very methodically over the last 18 months, we continue to bring more and more of that innovation to life. And that fundamentally is working as we’ve now started to integrate it much more actively into our base. If you look at it with photography, we now have in our tool, Generative Remove, we have AI-assisted edits in design, we have Generative Pattern, Generative Fill Shape. We have, in Photoshop, we have Gen Remove. We also have Gen Fill, and I can continue on with all the generations, but we’ve also now started to integrate it in Firefly Services for what we’re enabling enterprises to be able to access and use in terms of batch work and through APIs.

Adobe’s management is seeing the accelerated use and consumption of generative AI credits in Adobe’s products play out the way they expected it to; total consumption credits are going up with the introduction of each new generative AI capability 

If you look at sort of how that’s played out, as we talked about, we’re seeing accelerated use and generative credits being consumed because of that deeper integration into all of our tools, and that is playing out as we expected…

…And we do see with every subsequent capability we integrate into the tool, total credits consumed going up. 

Adobe’s management is seeing Adobe enjoying indirect monetisation from the AI features of its products, such as (1) the products having more value and pricing, (2) users being retained better when they use generative AI features, and (3) higher conversion of users when they try out Adobe products

When you look at then how that converts to monetization, first and foremost, we’ve integrated it a lot of that value into our core products with more value and more pricing. We’re also seeing that when people use these generative features, they retain better. We’re also seeing that when people come to Adobe to try our Creative Cloud applications or Express application, they’re able to convert better. And so there are all these ancillary implied benefits that we’re getting. 

For direct monetisation of the AI features in Adobe’s products, management is thinking of (1) instituting caps on generative AI credit consumption, (2) having AI plans with different AI capabilities; but direct monetisation is currently still not the key focus that management has, because they want to focus on proliferation and usage of generative AI across the user base

In terms of direct monetization, what we’ve said in the past is that the current model is around generative credits, which is I think where you’re going with this. And we do see with every subsequent capability we integrate into the tool, total credits consumed going up. Now what we are trying to do as we go forward, we haven’t started instituting the caps yet. And part of this is, as we’ve said all along, we want to really focus our attention on proliferation and usage across our base. We see a lot of users excited about it. It’s some of the most actively used features that we’ve ever released. And we want to avoid the generation anxiety that people feel. But we’re watching very closely as the economy of generative credits evolves, and we’re going to look at instituting those caps at some point when we feel the time is right and/or we’re also looking at other alternative models. What we did with Acrobat AI Assistant has proven to be very effective. And so we’re also considering other opportunities like having standard CC plans that have a core set of generative capabilities but also having premium API — sorry, premium AI plans that will include things more like video and other things.

Adobe’s management thinks Adobe’s generative AI video models are already pretty capable, but they are going to get better over time; management thinks that the real value of generative AI video models is not in their ability to create a video through a description the user gives, but in their ability to extend the video

I don’t know if you had a chance to see some of the videos we put out there integrated directly into premier, also text to video, images to video, more controllability. We have also the ability now to generate not just themes with humans and dogs and organic animals, but all these like overlays and things that creative professionals actually want to work with. And so we’re very excited about the set of things that they can get out of the box that get going. And human faces and things will just continue to get better…

…I spend a couple of hours with our video team. They have just absolutely hit it out of the park. I mean, the work that they have done, which is leveraging the image models with video, and again, I think to David’s point, the integration with Premier, that’s where we’ve always said, it’s the integration of the model and the application that differentiates it. I think when other models first came out, people were like, “Wow, you can describe it.” That’s just such a small part of where the value is. And the real value is, you have a video, you want to extend it. It’s a game changer in terms of what we can do. So really excited about the stuff that we’re doing in video. 

MongoDB (NASDAQ: MDB)

MongoDB’s management sees AI as a longer-term opportunity for MongoDB; management is seeing companies largely still experimenting with AI applications currently; management thinks inference workloads will come, but monetisation of AI apps will take time

AI continues to be an additional long-term opportunity for our business. At the start of the fiscal year, we told you that we didn’t expect AI to be a meaningful tailwind for our business in fiscal year 2025, which has proven accurate. Based on recent peer commentary, it seems that the industry now mostly agrees with this view. Companies are currently focusing their spending on the infrastructure layer of AI and are still largely experimenting with AI applications. Inference workloads will come and should benefit MongoDB greatly in the long run, but we are still very early when the monetization of AI apps will take time. AI demand is a question of when, not if.

MongoDB’s management has been talking to customers and they think MongoDB is the ideal database for AI apps for five reasons: (1) AI workloads involve a wide variety of data types and MongoDB’s document-model database is meant to handle this variety well, thus providing a well-rounded one-stop solution, (2) MongoDB’s database is high-performance and scalable, and allows AI workloads to utilise real-time operational data, (3) MongoDB’s database is integrated with leading app development frameworks and AI platforms, (4) MongoDB’s database has enterprise-grade security and compliance features, and (5) MongoDB’s database can be run anywhere on the customer’s choice; management feels very good about MongoDB’s positioning for AI

Our discussions with customers and partners give us increasing conviction that we are the ideal data layer for AI apps for a number of key reasons.

First, more than any other type of modern workload, AI-driven workloads require the underlying database to be capable of processing queries against rich and complex data structures quickly and efficiently. Our flexible document model is uniquely positioned to help customers build sophisticated AI applications because it is designed to handle different data types, your source data, vector data, metadata and generated data right alongside your live operational data, outdating the need for multiple database systems and complex back-end architectures.

Second, MongoDB offers a high performance and scalable architecture. As the latency of LLMs improve, the value of using real-time operational data for AI apps will become even more important.

Third, we are seamlessly integrated with leading app development frameworks and AI platforms, enabling developers to incorporate MongoDB into their existing workflows while having the flexibility to choose the LLM and other specific tools that best suit their needs.

Fourth, we meet or exceed the security and compliance requirements expected from an enterprise database, including enterprise-grade encryption, authorization and auditability.

Lastly, customers can run MongoDB anywhere, on-premise or as a fully managed service in 1 of the 118 global cloud regions across 3 hyperscalers giving them the flexibility to run workloads to meet — to best meet their application use cases and business needs…

… As the performance of these LLMs and latency of these LLMs increase, accessing real-time data becomes really important like, say, you’re calling and talking to a customer support chatbot, that you want that chatbot to have up-to-date information about that customer so that they can provide the most relevant and accurate information possible…

…I think it’s a quickly evolving space, but we feel very good about our positioning for AI, even though it’s still very early days.

MongoDB’s management sees 3 ways AI can accelerate MongoDB’s business over time: (1) AI will drive the cost of building applications, as all past platform shifts have done, thus leading to more apps and higher demand for databases, (2) MongoDB can be the database of choice for developers building AI applications (see Point 9 on MongoDB’s new AI Applications Program), and (3) MongoDB can help customers modernise their application estate (see Point 10 for more on this opportunity)

We see 3 main opportunities where we believe AI will accelerate our business over time. The first is that the cost of building applications in the world of AI will come down as we’ve seen with every previous platform shift, creating more applications and more data requiring more databases. The second opportunity is for us to be the database of choice for customers building greenfield AI applications…

…The third opportunity is to help customers modernize their legacy application estate. 

MongoDB’s management made the MongoDB AI Applications Program (MAAP) generally available in July 2024; MAAP brings the cloud computing hyperscalers and prominent AI model-building startups into one ecosystem to reduce the complexity and difficulty for MongoDB’s customers when they build new AI applications 

While we see that there’s tremendous amount of interest in and planning for new AI-powered applications, the complexity and fast-moving nature of the AI ecosystem slows customers down. That’s why we launched the MongoDB AI Applications Program, or MAAP, which became generally available to customers last month. MAAP brings together a unique ecosystem, including the 3 major cloud providers, AWS, Azure and GCP as well as Accenture and AI pioneers like Anthropic and Cohere. MAAP offers customers reference architectures and end-to-end technology stack that includes prebuilt integrations, professional services and a unified support system to help customers quickly build and deploy AI applications.

Modernising legacy application estates is a big opportunity, as most of the $80 billion database market is still in legacy relational databases; MongoDB has the Relational Migrator product to help customers migrate from legacy relational databases to the company’s document-model database; management thinks AI can significantly improve the process of modernising legacy applications by helping with understanding legacy code and rewriting them as modern versions; MongoDB launched a few pilots with customers earlier in 2024 to modernise their legacy applications with the help of AI and the results are exciting; the CIO (Chief Information Officer) of an insurance company in the pilots said the modernisation process was the first tangible return he had seen in his AI investments; management thinks it will take time for the modernisation program to contribute meaningful revenue to MongoDB, but they are excited 

Most of the existing $80-billion-plus database industry is built on dated relational architecture. Modernizing legacy applications has always been part of our business, and we have taken steps over the years to simplify and demystify this complex process through partnerships, education and most recently, our Relational Migrator product. AI offers a potential step function improvement, lowering the cost and reducing their time and risk to modernize legacy applications…

…Earlier this year, we launched several pilots with our customers where we work with them to modernize mission-critical applications, leveraging both AI tooling and services. The early results from these pilots are very exciting as our customers are experiencing significant reductions in time and cost of modernization. In particular, we have seen dramatic improvements in time and cost to rewrite application code and generate test suites. We see increasing interest from customers that want to modernize their legacy application estate, including large enterprise customers. As a CIO of one of the world’s largest insurance companies said about our pilot, this is the first tangible return he’s seen on his AI investments. While it’s still early days and generating meaningful revenue from this program will take time, we are excited about the results of our pilots and the growing pipeline of customers eager to modernize their legacy estate…

…Since day one, since our IPO, we’ve been getting customers to migrate off relational to MongoDB. But one of the biggest friction points has been that while it’s easy to move the data, you can map the schema from a relational schema to a document schema and you can automate that, the biggest stumbling block is that the customer has to or some third party has to rewrite the application, which, by definition, creates more costs, more time and in some cases, more risk especially for older apps, where the development teams who built those apps no longer exist. So what’s been compelling about AI is that AI has finally created a shortcut to overcome that big hurdle. And so essentially, you can start basically diagnosing the code, understand the code, recreate a modern version of that code and generate test suites to make sure the new code performs like the old code. So that definitely gets people’s interest because now, all of a sudden, what may take years or multiyears, you can do in a lot less time. And the pilots that we have done, the time and cost savings have been very, very compelling.

That being said, we’re in the very early days. There’s a lot of interest. We have a growing pipeline of customers across, frankly, all parts of the world from North America to EMEA and even the Pac Rim. And so we’re quite excited about the opportunity. But again, I would say it’s very early days.

Delivery Hero, a leading local delivery platform, is using MongoDB Atlas Vector Search to provide AI-powered hypersonalised results to users; Delivery Hero found that MongoDB Atlas Vector Search helped it build solutions for less cost than alternative technologies

Delivery Hero, a long-time MongoDB Atlas customer is the world’s leading local delivery platform, operating in 70-plus countries across 4 continents. Their quick commerce service enables customers to select fresh produce for delivery from local grocery stores. Approximately 10% of the inventory is fast-moving perishable produce that can go quickly out of stock. The company risks losing revenue and increasing customer churn if the customer didn’t have viable alternatives to their first choice. To address these risks, they are now using state-of-the-art AI models and MongoDB Atlas Vector Search to give hyperpersonalized alternatives to customers in real time if items they want to order are out of stock. With the introduction of MongoDB Atlas Vector Search, the data science team recognized that they could build a highly performant, real-time solution more quickly and for less cost than alternative technologies. 

MongoDB’s management believes that general-purpose LLMs (large language models) will win and will use RAG (retrieval augmented generation) as the primary way to combine generally available data to proprietary data; management is seeing advanced RAG use-cases in answering complex questions

There are some questions about LLMs, whether a general-purpose LLM or a fine-tune LLM, what the trade-offs are. Our belief is that given the performance of LLMs, you’re going to see the general purpose LLMs probably win and will use RAG as the predominant approach to marry generally available data with proprietary data. And then you are starting to see things like advanced RAG use cases where you get much more sophisticated ways to ask complex questions, provide more accurate and detailed answers and better adapt to different types of information and queries.

MongoDB’s management is seeing most AI workloads happen in the cloud, but they also see a lot of customers using open-source LLMs and running those workloads locally

We predominantly see most of the AI workloads in the cloud, but there are definitely lots of customers who are looking at using open source LLMs, in particular, things like Llama, and running those workloads locally.

MongoDB’s management believes MongoDB wins against Postgres for winning AI workloads because MongoDB can handle complex data types whereas Postgres, which is a relational – or SQL – database, struggles

MongoDB is designed to handle these different data structures. And I talked about we can help unify metadata, operational data, vector data and generate it all in one platform. Relational databases, and Postgres is one of them, have limitations in terms of what they can — how they can handle different types of data. In fact, when the data gets too large, these relational databases have to do what’s called off-row storage. And it becomes — it creates a performance overhead on these relational platforms. Postgres has this thing called TOAST, which is — stands for The Oversized-Attribute Storage Technique. And it’s basically a way to handle these different data types, but it creates a massive performance overhead. So we believe that we are architecturally far better for these more complex AI workloads than relational databases.

MongoDB’s management is seeing growing adoption of Vector, and Vector is helping attract new customers to MongoDB; an existing Atlas customer, a financial news organisation, migrated from Elastic Search to Atlas Search in order to use MongoDB’s Vector Search capabilities; an European energy company is using Vector Search for a geospatial search application

On Vector, we’re continuing to see growth in adoption, and we see Vector is effective in attracting new customers to the MongoDB platform. A world-renowned financial news organization, which is already running in Atlas, migrated from Elasticsearch to Atlas Search using Search Nodes to take advantage of our Vector Search capabilities to build a site search that combines lexical search with semantic search to find the most relevant articles for user query. And a European energy company built a geospatial search application using Atlas Search and Vector search and the app was built on-prem but — and to clouds to vectorize geospatial data and facilitate research and discovery.

MongoDB’s management is seeing MongoDB’s customers improve their software development productivity with the help of AI, but the rate of improvement is all over the place

[Question] We’ve talked before in the past that AI is just driving a lot of new code, making developers significantly more productive. Have you seen that behavior in any of your existing customers on Atlas where maybe their utilization rate goes up or the number of applications built per customer goes up?

[Answer] A common question I ask our customers when I meet with them in terms of what code generation tools that they’re using and what benefits they’re gaining. The answers tend to be a little bit all over the map. Some people see 10%, 15% productivity improvement. Some people say 20%, 25% productivity improvement. Some people say it helps my senior developers be more productive. Some people say it helps my junior developers become more like senior developers. So the answers tend to be all over the map.

Nvidia (NASDAQ: NVDA)

Nvidia’s Data Center revenue had incredibly strong growth in 2024 Q2, driven by demand for the Hopper GPU computing platform; compute revenue was up by 2.5x while networking revenue was up by 2x

Data Center revenue of $26.3 billion was a record, up 16% sequentially and up 154% year-on-year, driven by strong demand for NVIDIA Hopper, GPU computing and our networking platforms. Compute revenue grew more than 2.5x. Networking revenue grew more than 2x from the last year.

Even as Nvidia is getting ready to launch its Blackwell-architecture GPUs, customers are still buying the Hopper-architecture GPUs; the H200 platform, based on the Hopper architecture, started ramping in 2024 Q2 and offers 40% more memory bandwidth than the H100; management thinks that the reasons why the Hopper-architecture chips still enjoy strong demand despite the imminent arrival of the Blackwell-architecture chips are (1) AI companies need chips today to process data right now, and (2) AI companies are in a race to build the best model and they’re all racing to be the first

Customers continue to accelerate their Hopper architecture purchases while gearing up to adopt Blackwell…

…NVIDIA H200 platform began ramping in Q2, shipping to large CSPs, consumer Internet and enterprise companies. The NVIDIA H200 builds upon the strength of our Hopper architecture and offering over 40% more memory bandwidth compared to the H100…

…The demand for Hopper is really strong. And it’s true, the demand for Blackwell is incredible. There’s a couple of reasons for that. The first reason is, if you just look at the world’s cloud service providers and the amount of GPU capacity they have available, it’s basically none…

…A generative AI company spends the vast majority of their invested capital into infrastructure so that they could use an AI to help them create products. And so these companies need it now. They just simply can’t afford — you just raise money, they want you to put it to use now. You have processing that you have to do. You can’t do it next year. You got to do it today. And so that’s one reason. The second reason for Hopper demand right now is because of the race to the next plateau. The first person to the next plateau gets to introduce some revolutionary level of AI. The second person who gets there is incrementally better or about the same. And so the ability to systematically and consistently race to the next plateau and be the first one there is how you establish leadership…

…We believe our Hopper will continue to grow into the second half. We have many new products for Hopper or existing products for Hopper that we believe will start continuing to ramp in the next quarters, including our Q3 and those new products moving to Q4. So let’s say, Hopper, therefore, versus H1 is a growth opportunity for that. 

Nvidia’s management thinks that the next generation of AI models will need 10-20 times more compute to train

Next-generation models will require 10 to 20x more compute to train with significantly more data. The trend is expected to continue.

Nvidia’s management sees inferencing accounting for 40% of Data Center revenue over the last 4 quarters (was 40% as of 2024 Q1)

Over the trailing 4 quarters, we estimate that inference drove more than 40% of our Data Center revenue.

Nvidia’s management is seeing demand coming from builders of frontier AI models, consumer Internet companies, and companies building generative AI applications for a wide range of use cases

Demand for NVIDIA is coming from frontier model makers, consumer Internet services, and tens of thousands of companies and start-ups building generative AI applications for consumers, advertising, education, enterprise and health care, and robotics. 

Nvidia’s Data Center revenue in China grew sequentially in 2024 Q2, but still remains below the level seen prior to export controls; management expects tough competition in China

Our Data Center revenue in China grew sequentially in Q2 and a significant contributor to our Data Center revenue. As a percentage of total Data Center revenue, it remains below levels seen prior to the imposition of export controls. We continue to expect the China market to be very competitive going forward.

Nvidia has leadership in inference

The latest round of MLPerf inference benchmarks highlighted NVIDIA’s inference leadership, with both NVIDIA Hopper and Blackwell platforms combining to win gold medals on all tests.

Nvidia’s Blackwell family of chips combines GPUs, CPUs, DPUs (data processing units), NVLink, and networking; the GB200 NVL72 system in the Blackwell family links up 72 GPUs to act as 1 GPU and is up to 30 times faster for LLM (large language model) inference workloads; Nvidia has made a change to the Blackwell architecture to improve production yields; Blackwell’s production is expected to ramp in the fourth quarter of 2024; management sees demand for Blackwell exceeding supply by a wide margin up to 2025; there are more than 100 different Blackwell architecture systems; Nvidia’s Blackwell systems come in both air-cooled and liquid-cooled flavours; management expects Nvidia’s Data Center business to grow significantly in 2025 and 2026, powered by the Blackwell system; management sees Blackwell as a step-function improvement over Hopper that delivers 3-5 times more AI throughput than Hopper; Blackwell required 7 one-of-a-kind chips to build; Nvidia designed and optimised the Blackwell system end-to-end

The NVIDIA GB200 NVL72 system with the fifth-generation NVLink enables all 72 GPUs to act as a single GPU and deliver up to 30x faster inference for LLM’s workloads and unlocking the ability to run trillion-parameter models in real time…

…We executed a change to the Blackwell GPU mass to improve production yields. Blackwell production ramp is scheduled to begin in the fourth quarter and continue into fiscal year ’26. In Q4, we expect to get several billion dollars in Blackwell revenue…

Demand for Blackwell platforms is well above supply, and we expect this to continue into next year…

…There are something like 100 different types of Blackwell-based systems that are built that were shown at Computex, and we’re enabling our ecosystem to start sampling those…

…We offer multiple configurations of Blackwell. Blackwell comes in either a Blackwell classic, if you will, that uses the HGX form factor that we pioneered with Volta. I think it was Volta. And so we’ve been shipping the HGX form factor for some time. It is air cooled. The Grace Blackwell is liquid cooled…

…We expect to grow our Data Center business quite significantly next year. Blackwell is going to be a complete game changer for the industry. And Blackwell is going to carry into the following year…

…Blackwall is a step-function leap over Hopper. Blackwell is an AI infrastructure platform, not just the GPU. It also happens to be the name of our GPU, but it’s an AI infrastructure platform. As we reveal more of Blackwell and sample systems to our partners and customers, the extent of Blackwell’s lead becomes clear. The Blackwell vision took nearly 5 years and 7 one-of-a-kind chips to realize: the Gray CPU, the Blackwell dual GPU and a colos package, ConnectX DPU for East-West traffic, BlueField DPU for North-South and storage traffic, NVLink switch for all-to-all GPU communications, and Quantum and Spectrum-X for both InfiniBand and Ethernet can support the massive burst traffic of AI. Blackwell AI factories are building size computers. NVIDIA designed and optimized the Blackwell platform full stack, end-to-end, from chips, systems, networking, even structured cables, power and cooling and mounts of software to make it fast for customers to build AI factories. These are very capital-intensive infrastructures. Customers want to deploy it as soon as they get their hands on the equipment and deliver the best performance and TCO. Blackwell provides 3 to 5x more AI throughput in a power-limited data center than Hopper…

…The Blackwell system lets us connect 144 GPUs in 72 GB200 packages into 1 NVLink domain, with an aggregate NVLink bandwidth of 259 terabytes per second in 1 rack. Just to put that in perspective, that’s about 10x higher than Hopper.  

Nvidia’s Ethernet for AI revenue doubled sequentially; management sees Nvidia’s ethernet product, Spectrum-X, enjoying wide support from the AI ecosystem; Spectrum-X performs 1.6 times better than traditional Ethernet; management plans to launch new Spectrum-X products every year and thinks that Spectrum-X will soon become a multi-billion dollar product

Our Ethernet for AI revenue, which includes our Spectrum-X end-to-end Ethernet platform, doubled sequentially with hundreds of customers adopting our Ethernet offerings. Spectrum-X has broad market support from OEM and ODM partners and is being adopted by CSPs, GPU cloud providers and enterprises, including xAI to connect the largest GPU compute cluster in the world. Spectrum-X supercharges Ethernet for AI processing and delivers 1.6x the performance of traditional Ethernet. We plan to launch new Spectrum-X products every year to support demand for scaling compute clusters from tens of thousands of GPUs today to millions of DPUs in the near future. Spectrum-X is well on track to begin a multibillion-dollar product line within a year.

Japan’s government is working with Nvidia to build an AI supercomputer; Nvidia’s management thinks sovereign AI revenue will be in the low-teens billion-range this year; management is seeing countries want to build their own generative AI that incorporates their own language, culture, and data

Japan’s National Institute of Advanced Industrial Science and Technology is building its AI Bridging Cloud Infrastructure 3.0 supercomputer with NVIDIA. We believe sovereign AI revenue will reach low double-digit billions this year…

…It certainly is a unique and growing opportunity, something that surfaced with generative AI and the desires of countries around the world to have their own generative AI that would be able to incorporate their own language, incorporate their own culture, incorporate their own data in that country.

Most of the Fortune 100 companies are working with Nvidia on AI projects

We are working with most of the Fortune 100 companies on AI initiatives across industries and geographies. 

Nvidia’s management is seeing a range of applications driving the company’s growth; these applications include (1) Amdocs’ smart agent which is reducing customer service costs by 30%, and (2) Wistron’s usage of Nvidia AI Ominiverse to reduce cycle times in its factories by 50%

A range of applications are fueling our growth, including AI-powered chatbots, generative AI copilots and agents to build new, monetizable business applications and enhance employee productivity. Amdocs is using NVIDIA generative AI for their smart agent, transforming the customer experience and reducing customer service costs by 30%. ServiceNow is using NVIDIA for its Now Assist offering, the fastest-growing new product in the company’s history. SAP is using NVIDIA to build Joule copilot. Cohesity is using NVIDIA to build their generative AI agent and lower generative AI development costs. Snowflake, who serves over 3 billion queries a day for over 10,000 enterprise customers, is working with NVIDIA to build copilots. And lastly, Wistron is using NVIDIA AI Omniverse to reduce end-to-end cycle times for their factories by 50%.

Every automobile company that is developing autonomous vehicle technology is working with Nvidia; management thinks that automotive will account for multi-billions in revenue for Nvidia; Nvidia won the Autonomous Brand Challenge at the recent Computer Vision and Pattern Recognition Conference

Every automaker developing autonomous vehicle technology is using NVIDIA in their data centers. Automotive will drive multibillion dollars in revenue across on-prem and cloud consumption and will grow as next-generation AV models require significantly more compute…

…At the Computer Vision and Pattern Recognition Conference, NVIDIA won the Autonomous Brand Challenge in the end-to-end driving at scale category, outperforming more than 400 entries worldwide. 

Nvidia’s management announced Nvidia AI Foundry – a platform for building custom AI models – in 2024 Q2; users of Nvidia AI Foundry are able to customise Meta’s Llama 3.1 foundation AI model; Nvidia AI Foundry is the first platform where users are able to customise an open-source, frontier-level foundation AI model; Accenture is already using Nvidia AI Foundry 

During the quarter, we announced a new NVIDIA AI foundry service to supercharge generative AI for the world’s enterprises with Meta’s Llama 3.1 collection of models… 

…Companies for the first time can leverage the capabilities of an open source, frontier-level model to develop customized AI applications to encode their institutional knowledge into an AI flywheel to automate and accelerate their business. Accenture is the first to adopt the new service to build custom Llama 3.1 models for both its own use and to assist clients seeking to deploy generative AI applications.

Companies from many industries are using NIMs (Nvidia inference microservices) for deployment of generative AI; AT&T saw 70% cost savings and 8 times latency reduction with NIM; there are 150 organisations using NIMs; Nvidia recently announced NIM Agent Blueprints, a catalog of reference AI applications; Nvidia is using NIMs to open the Nvidia Omniverse to new industries

NVIDIA NIMs accelerate and simplify model deployment. Companies across health care, energy, financial services, retail, transportation, and telecommunications are adopting NIMs, including Aramco, Lowes, and Uber. AT&T realized 70% cost savings and 8x latency reduction affter moving into NIMs for generative AI, call transcription and classification. Over 150 partners are embedding NIMs across every layer of the AI ecosystem. 

We announced NIM Agent Blueprints, a catalog of customizable reference applications that include a full suite of software for building and deploying enterprise generative AI applications. With NIM Agent Blueprints, enterprises can refine their AI applications over time, creating a data-driven AI flywheel. The first NIM Agent Blueprints include workloads for customer service, computer-aided drug discovery, and enterprise retrieval augmented generation. Our system integrators, technology solution providers, and system builders are bringing NVIDIA NIM Agent Blueprints to enterprises…

…We announced new NVIDIA USD NIMs and connectors to open Omniverse to new industries and enable developers to incorporate generative AI copilots and agents into USD workloads, accelerating our ability to build highly accurate virtual worlds.

Nvidia’s AI Enterprise software platform is powering Nvidia’s software-related business to approach a $2 billion annual revenue run-rate by the end of this year; management thinks Nvidia AI Enterprise represents great value for customers by providing GPUs at a price of $4,500 per GPU per year; management thinks the TAM (total addressable market) for Nvidia’s AI software business can be significant

NVIDIA NIM and NIM Agent Blueprints are available through the NVIDIA AI Enterprise software platform, which has great momentum. We expect our software, SaaS and support revenue to approach a $2 billion annual run rate exiting this year, with NVIDIA AI Enterprise notably contributing to growth…

…At $4,500 per GPU per year, NVIDIA AI Enterprise is an exceptional value for deploying AI anywhere. And for NVIDIA’s software TAM, it can be significant as the CUDA-compatible GPU installed base grows from millions to tens of millions. 

Computers that contain Nvidia’s RTX chip can deliver up to 1,300 AI TOPS (tera operations per second); there are more than 200 RTX AI computer models from computer manufacturers; there is an installed base of 100 million RTX AI computers; a game called Mecha BREAK is the first game to use Nvidia ACE, a generative AI service for creating digital humans

Every PC with RTX is an AI PC. RTX PCs can deliver up to 1,300 AI tops and there are now over 200 RTX AI laptops designed from leading PC manufacturers. With 600 AI-powered applications and games and an installed base of 100 million devices, RTX is set to revolutionize consumer experiences with generative AI. NVIDIA ACE, a suite of generative AI technologies is available for RTX AI PCs. Mecha BREAK is the first game to use NVIDIA ACE, including our small language model, Nemotron-4 4B, optimized on device inference. 

Foxconn, the largest electronics manufacturer in the world, and Mercedes-Benz, the well-known auto manufacturer, are using Nvidia Omniverse to produce digital twins of their manufacturing plants

The world’s largest electronics manufacturer, Foxconn, is using NVIDIA Omniverse to power digital twins of the physical plants that produce NVIDIA Blackwell systems. And several large global enterprises, including Mercedes-Benz, signed multiyear contracts for NVIDIA Omniverse Cloud to build industrial digital twins of factories.

Many robotics companies are using Nvidia’s AI robot software

Boston Dynamics, BYD Electronics, Figure, Intrinsyc, Siemens, Skilled AI and Teradyne Robotics are using the NVIDIA Isaac robotics platform for autonomous robot arms, humanoids and mobile robots.

Nvidia’s management is seeing some customers save up to 90% in computing costs by transitioning from genera-purpose computing (CPUs) to accelerated computing (GPUs)

We know that accelerated computing, of course, speeds up applications. It also enables you to do computing at a much larger scale, for example, scientific simulations or database processing. But what that translates directly to is lower cost and lower energy consumed. And in fact, this week, there’s a blog that came out that talked about a whole bunch of new libraries that we offer. And that’s really the core of the first platform transition, going from general-purpose computing to accelerated computing. And it’s not unusual to see someone save 90% of their computing cost. And the reason for that is, of course, you just sped up an application 50x. You would expect the computing cost to decline quite significantly.

Nvidia’s management believes that generative AI is a new way to write software and is changing how every layer of computing is done

Generative AI, taking a step back about why it is that we went so deeply into it, is because it’s not just a feature, it’s not just a capability, it’s a fundamental new way of doing software. Instead of human-engineered algorithms, we now have data. We tell the AI, we tell the model, we tell the computer what are the expected answers, what are our previous observations, and then for it to figure out what the algorithm is, what’s the function. It learns a universal — AI is a bit of a universal function approximator and it learns the function. And so you could learn the function of almost anything, and anything that you have that’s predictable, anything that has structure, anything that you have previous examples of. And so now here we are with generative AI. It’s a fundamental new form of computer science. It’s affecting how every layer of computing is done from CPU to GPU, from human-engineered algorithms to machine-learned algorithms, and the type of applications you could now develop and produce is fundamentally remarkable.

Nvidia’s management thinks AI models are still seeing the benefits of scaling

There are several things that are happening in generative AI. So the first thing that’s happening is the frontier models are growing in quite substantial scale. And we’re still all seeing the benefits of scaling.

The amount of compute needed to train an AI model goes up much faster than the size of the model; management thinks the next generation of AI models could require 10-40 times more compute 

Whenever you double the size of a model, you also have to more than double the size of the data set to go train it. And so the amount of flops necessary in order to create that model goes up quadratically. And so it’s not unexpected to see that the next-generation models could take 10x, 20x, 40x more compute than last generation.

Nvidia’s management is seeing more frontier model makers in 2024 than in 2023

Surprisingly, there are more frontier model makers than last year.

Nvidia’s management is seeing advertising-related computing needs shifting from being powered by CPUs to being powered by GPUs and generative AI

The largest systems, largest computing systems in the world today, and you’ve heard me talk about this in the past, which are recommender systems moving from CPUs. It’s now moving from CPUs to generative AI. So recommender systems, ad generation, custom ad generation targeting ads at very large scale and quite hyper-targeting, search and user-generated content, these are all very large-scale applications that have now evolved to generative AI.

Nvidia’s management is seeing generative AI startups generating tens of billions of revenue-opportunities for cloud computing providers

The number of generative AI start-ups is generating tens of billions of dollars of cloud renting opportunities for our cloud partners

Nvidia’s management is seeing that cloud computing providers have zero GPU capacity available because they are using it for internal workloads (such as accelerating data processing) and renting it out to model makers and other AI startups

If you just look at the world’s cloud service providers and the amount of GPU capacity they have available, it’s basically none. And the reason for that is because they’re either being deployed internally for accelerating their own workloads, data processing, for example…

…The second is, of course, the rentals. They’re renting capacity to model makers. They’re renting it to start-up companies. 

Nvidia’s management thinks Nvidia’s GPUs are the only AI GPUs that process and accelerate data; before the advent of generative AI, the number one use case of Nvidia’s GPUs was to accelerate data processing

NVIDIA’s GPUs are the only accelerators on the planet that process and accelerate data. SQL data, Panda’s data, data science toolkits like Panda’s, and the new one, Polar’s, these are the ones that are the most popular data processing platforms in the world, and aside from CPUs which, as I’ve mentioned before, are really running out of steam, NVIDIA’s accelerated computing is really the only way to get boosting performance out of that. And so the #1 use case long before generative AI came along is the migration of applications one after another to accelerated computing.

Nvidia’s management thinks that those who purchase Nvidia AI chips are getting immediate ROI (return on investment) for a few reasons: (1) GPUs are a better way to build data centers compared to CPUs because GPUs save money on data processing compared to CPUs, (2) cloud computing providers who rent out GPUs are able to rent out their GPUs the moment they are built up in the data center because there are many generative AI companies clamouring for the chips, and (3) generative AI improves a company’s own services, which delivers a fast ROI

The people who are investing in NVIDIA infrastructure are getting returns on it right away. It’s the best ROI infrastructure, computing infrastructure investment you can make today. And so one way to think through it, probably the easiest way to think through it is just to go back to first principles. You have $1 trillion worth of general-purpose computing infrastructure. And the question is, do you want to build more of that or not?

And for every $1 billion worth of Juniper CPU-based infrastructure that you stand up, you probably rent it for less than $1 billion. And so because it’s commoditized, there’s already $1 trillion on the ground. What’s the point of getting more? And so the people who are clamoring to get this infrastructure, one, when they build out Hopper-based infrastructure and soon, Blackwell-based infrastructure, they start saving money. That’s tremendous return on investment. And the reason why they start saving money is because data processing saves money, and data processing is probably just a giant part of it already. And so recommender systems save money, so on and so forth, okay? And so you start saving money.

The second thing is everything you stand up are going to get rented because so many companies are being founded to create generative AI. And so your capacity gets rented right away and the return on investment of that is really good.

And then the third reason is your own business. Do you want to either create the next frontier yourself or your own Internet services, benefit from a next-generation ad system or a next-generation recommender system or a next-generation search system? So for your own services, for your own stores, for your own user-generated content, social media platforms, for your own services, generative AI is also a fast ROI.

Nvidia’s management is seeing a significant number of data centers wanting liquid-cooled GPU systems because the use of liquid cooling enables 3-5 times more AI throughput compared to the past, resulting in cheaper TCO (total cost of ownership)

The number of data centers that want to go to liquid cooled is quite significant. And the reason for that is because we can, in a liquid-cooled data center, in any power-limited data center, whatever size of data center you choose, you could install and deploy anywhere from 3 to 5x the AI throughput compared to the past. And so liquid cooling is cheaper. Our TCO is better, and liquid cooling allows you to have the benefit of this capability we call NVLink, which allows us to expand it to 72 Grace Blackwell packages, which has essentially 144 GPUs.

Nvidia does not do the full integration of its GPU systems into a data center because it is not the company’s area of expertise

Our customers hate that we do integration. The supply chain hates us doing integration. They want to do the integration. That’s their value-add. There’s a final design-in, if you will. It’s not quite as simple as shimmying into a data center, but the design fit-in is really complicated. And so the design fit-in, the installation, the bring-up, the repair-and-replace, that entire cycle is done all over the world. And we have a sprawling network of ODM and OEM partners that does this incredibly well.

Nvidia has released many new libraries for CUDA, across a wide variety of use cases, for AI software developers to work with

Accelerated computing starts with CUDA-X libraries. New libraries open new markets for NVIDIA. We released many new libraries, including CUDA-X Accelerated Polars, Pandas and Spark, the leading data science and data processing libraries; CUVI-S for vector databases, this is incredibly hot right now; Ariel and Ciona for 5G wireless base station, a whole world of data centers that we can go into now; Parabricks for gene sequencing and AlphaFold2 for protein structure prediction is now CUDA accelerated.

Nvidia now has 3 networking platforms for GPUs

We now have 3 networking platforms, NVLink for GPU scale-up, Quantum InfiniBand for supercomputing and dedicated AI factories, and Spectrum-X for AI on Ethernet. NVIDIA’s networking footprint is much bigger than before. 

Salesforce (NYSE: CRM)

Agentforce is a new architecture and product that management believes will be fundamental to Salesforce’s AI leadership in the next decade; Salseforce will be introducing Agentforce in its upcoming Dreamforce customer-event; Agentforce is an autonomous AI and management will be getting every attendee at Dreamforce to turn on their own AI agents; Salesforce is already building agents for the company, Workday, and Workday will be Salseforce’s first Agentforce partner; Agentforce allows companies to build custom agents for sales, service, marketing, and commerce; management believes that within a year, most companies will be deploying autonomous AI agents at scale, and these agents will have a big positive impact on companies’ operations; Agentforce is currently management’s singular focus; many companies are already using Agentforce, including one of the world’s largest healthcare companies, which is resolving more than 90% of patient inquiries with Agentforce, and thinks Agentforce is much better than any other competing AI platform; a very large media company is using Agentforce to resolve 90% of employee and consumer issues; management thinks Salesforce is the first company to deploy high-quality enterprise AI agents at scale; Agentforce is not a co-pilot, it is an autonomous agent that is accurate and can be deployed right out of the box; users of Agentforce can do advanced planning and reasoning with minimal input; management sees Agentforce as being a trusted colleague that will complement human users; management sees thousands of companies using Agentforce by January 2025; early trials of Agentforce has showed remarkable success

We’re going to talk about a whole different kind of sales force today, a different kind of architecture and a product that we didn’t even talk about on the last earnings call that is going to be fundamental to our future and a manifestation of our decade of AI leadership, which is Agentforce. Now in just a few weeks, we’re going to kick off Dreamforce, and I hope all of you are planning to be there, the largest AI event in the world with more than 45,000 trailblazers in San Francisco. And this year, Dreamforce is really becoming Agentforce…

…We’re going to show our new Agentforce agents and how we’ve reimagined enterprise software for this new world of autonomous AI. And every customer, I’m going to try to get every customer who comes to Dreamforce to turn agents on while they’re there…

…This idea that you’re not just going to have sales agents and service agents who probably read, heard maybe you saw in CBC, we’re building the agents for Workday and we’re going to be building custom agents for so many of you as well with Agentforce, because it is a development platform as well as this incredible capability to radically extend your sales and service organizations.

So when you arrive at the Dreamforce campus, you’re going to see a big sign outside that says, humans with agents drive customer success together. And that’s because we now so strongly believe the future isn’t about having a sales force or a service force or a marketing force or a commerce force or an analytics force. The future is about also having an Agentforce. And while many customers today don’t yet have agent forces, but they do have sales forces or service forces, I assure you that within a year, we’re all going to have agent forces, and we’re going to have them at scale. And it’s going to radically extend our companies and it’s going to augment our employees, make us more productive. It’s going to turn us into these incredible margin and revenue machines. It’s going to be pretty awesome…

…with this Agentforce platform, we’re making it easy to build these powerful autonomous agents for sales, for service, for marketing, for commerce, automating the entire workflow on their own, embedding agents in the flow of work and getting our customers to the agent future first. And this is our primary goal of our company right now. This is my singular focus…

…We’re going to talk about the customers who have it, customers like OpenTable and Wiley and — and ADP and RBC and so many others who are deploying these agents and running them on top of our Data Cloud and our apps…

…At Dreamforce, you’re going to hear one of the very largest health care companies in the world. It’s got 20 million consumers here in the United States who is resolving more than 90% of all patient inquiries with Agentforce and they’re benchmarking us significantly higher than any other competing AI platform, and that’s based on some incredible AI breakthroughs that we have had at Salesforce…

…One of these very large media companies that we work with, a lot of probably know who have everything, every possible media asset, while they’re just resolving 90% of all of their employee and consumer issues with Agentforce, pretty awesome. So there’s nothing more transformational than agents on the technology horizon that I can see and Salesforce is going to be the first company at scale to deploy enterprise agents and not just any enterprise agents, the highest quality, most accurate agents in the world…

…We’re seeing that breakthrough occur because with our new Agentforce platform, we’re going to make a quantum leap for in AI, and that’s why it wants you all at Dreamforce, because I want you to have your hands on this technology to really understand this. This is not co-pilots…

…These agents are autonomous. They’re able to act with accuracy. They’re able to come right out of the box. They’re able to go right out of the platform…

…These agents don’t require a conversational prompt to take action. You can do advanced planning, reasoning with minimal human input. And the example of this incredible health care company, you’re going to be able to say to the agent, “Hey, I want to look at my labs, I want to understand this. It looks like I need repeat labs. Can you reschedule those for me? It looks like I need to see my doctor, can you schedule that for me? I also want to get an MRI, I want to get this.” And the level of automation that we’re going to be able to provide and unleash the productivity back into these organizations is awesome…

…This is going to be like having these trusted colleagues can handle these time-consuming tasks engaging with these — whether it’s inbound lead or resolving this customer, patient inquiry or whatever it is, this is humans with agents driving customer success together and Agentforce agents can be set up in minutes, easily scalable, work around with the block, any language. And by the beginning of next fiscal year, we will have thousands of customers using this platform. And we will have hand help them to make it successful for them, deploy it. The early trials have been remarkable to see these customers have the success, it has been just awesome…

…We’re just at the beginning of building an Agentforce ecosystem with companies able to build agents on our platform for their workforce and use cases, and we’re excited to have Workday as our first agent force partner.

Salesforce has been able to significantly reduce hallucinations with its AI products, and thus deliver highly accurate results, through the use of new retrieval augmented generation (RAG) techniques

The accuracy of our results, the reduction of hallucinations and the level of capability of AI is unlike anything I think that any of us have ever seen, and we’ve got some incredible new techniques, especially incredible new augmented RAG techniques that are delivering us the capability to deliver this accuracy with our — for our customers.

Salesforce’s management still sees the company as the No.1 AI CRM in the world

Of course, Salesforce is the #1 AI CRM.

 In 2024 Q2, Einstein delivered 25 trillion transactions and 1 trillion workflows; Wyndham is using Einstein to reduce average call times to free up time for service agents for higher-value work

We’re just operating at this incredible scale, delivering 25 trillion Einstein transactions across all of the clouds during the quarter, that’s 25 trillion and more than 1 trillion workflows…

…MuleSoft allows Wyndham to unlock business-critical data from various platforms and onboard future franchisees faster. And with Einstein generated recommended service replies, average call times have been reduced and service agents can focus on higher priority work

Salesforce’s management thinks many of the company’s customers have a misconception about AI in that they need to build and train their own AI models; management is able to use Salesforce’s AI models and resolve issues much better than its customers’ own models; management thinks Salesforce’s AI models have the highest efficacy

I think that there’s a lot of misconceptions about AI with my customers. I have been out there very disappointed with the huge amount of money that so many of these customers have wasted on AI. They are trying to DIY their AI…

…This idea that our customers are going to have to build their own models, train their own models, retrain their own models, retrain them again and I’m meeting with these customers, and they’re so excited when they and they say, “Oh, I built this model, and we’re resolving 10%, 20%, 30%, 40% and — of this or that and whatever. ” And I’m like, really, we’ll take a look at our models and our capability where you don’t have to train or retrain anything and you’re going to get more than 90%. And then they say, wait a minute, how do you do that? And this is a moment where every customer needs to realize you don’t need the DIY your AI. You can use a platform like Salesforce to get the highest efficacy of artificial intelligence, the best capability to fully automate your company, achieve all of your goals and you can do it with professional enterprise software…

…We’re in met with one of the largest CIOs in the world is telling me how excited he was for the B2C part of this business, he built this model and accuracy rates, than I was like, really, let me show you what we’re doing here. And then he said to me, why am I doing this? Why am I not just using your platform? And I said good question. So these customers are spending millions of dollars, but are they really getting the results that they want? It feels like this early days of cloud. It’s just early days of social mobile. Customers feel like they have to DIY it, they don’t need to, they can make it all happen themselves. And you can just see that to deliver this very high-quality capability they can use a deeply integrated platform like Salesforce.

Salesforce’s management is seeing the company’s customers get immediate ROI (return on investment) from deploying AI automation solutions because the solutions can be easily and quickly customised and configured

We’ve created out-of-the-box platform to deliver all of this for them. So this could be service reply recommendations, account summaries, report generation, you’ve seen in Slack, this kind of auto summarization, recaps, all of these amazing things, the level of automation, the amount of code that our team has written, the transformation of our platform in the last 18 months, it’s remarkable. And customers love it because they can take the platform and then all of this generative AI use case customize it for their own needs or configure it using our capability because they’re doing that without writing a line of code. It’s clicks, not code, deploy them in days, not weeks. They’re doing this in months, not years, and you’re getting immediate ROI. 

Salesforce’s management thinks many of the company’s customers are really disappointed with Microsoft Co-pilot because of a lack of accuracy

So many customers are so disappointed in what they bought from Microsoft CoPilots because they’re not getting the accuracy and the response that they want. Microsoft is disappointed so many customers with AI. 

Wiley is achieving double-digit percentage increase in customer satisfaction and deflection rate, and 50% case resolution with the first generation of Agentforce; Royal Bank of Canada and AP are seeing 90% case resolution rates with the second generation of Agentforce; OpenTable is using Agentforce to support its interactions with 60,000 restaurants and 160 million diners

Wiley is a long-standing Salesforce customer. It’s one of our first deployments in the first agent force trial. It’s pretty awesome. And you all know they make textbooks and it’s back-to-school. But maybe you don’t know that Wiley has to like surge their sales and service organization at back-to-school time when everyone’s buying these textbooks. Well, now they can use agents to do that surge. They don’t have to go buy a bunch of gig workers and bring them in. and that age and capacity is so exciting for them. What we saw with Wiley was, this is a quote from them, “we’re seeing double-digit percentage increase in customer satisfaction and deflection rate compared to older technologies and in these early weeks of our busiest season. ” So that was very reassuring to us. that we have the right thing that’s happening. And Wiley has already seen a 50% increase in case resolution. That’s with our first generation of Agentforce.

As I mentioned, the second generation of Agentforce, which we have with customers already, including some of these amazing organizations like Royal Bank of Canada, ADP and others is 90% case resolution. It is an awesome moment in this tech business.

OpenTable is another super great story. You all know they are managing 60,000 restaurants, 160 million diners to support. They’re on Agentforce now. They require that incredible scale to deliver top-notch customer service. That’s why they’re using the product. It’s been awesome to get their results and it can be all kinds of questions resolving basic issues, account activations, reservation management, loyalty point expiration. Agentforce for service can easily answer all of these questions like when do my points expire for a diner asset, a follow-up question like, what about in Mexico? What about — can I make this change? That’s where we’re delivering those incredible moments for OpenTable, giving them this kind of productivity enhancement. 

Agentforce is driving growth in cloud products’ sales for Salesforce

Agentforce for sales, you can imagine extending your sales force with SDRs, BDRs who are agents that are going out and building pipeline for you and generating all kind of demand and even really closing deals. So, this is going to drive sales cloud growth. It already is, service cloud growth. It already is because customers are going to extend their sales and service organizations and become a lot more productive with these agents.

Salesforce will be releasing industry-specific AI agents in the coming months

In the coming months, we’re going to release Agentforce agents for other roles, including industry-specific agents, health agents, as I mentioned. 

Data Cloud provides the foundation for Agentforce because it holds a huge amount of data and metadata; management continues to believe that data is the foundation of AI; Data Cloud federates and connects to all other data clouds of a user to deliver super accurate AI; Data Cloud is Salesforce’s fastest-growing organic product and will be the fastest to hit $1 billion, $5 billion, and $10 billion in revenue; Data Cloud customers were up 130% year-on-year in 2024 Q2; number of Data Cloud customers spending more than $1 million annually have doubled; Data Cloud processed 2.3 quadrillion records in 2024 Q2 (was 2 quadrillion in 2024 Q1); Data Cloud consumption was up 110% year-on-year in 2024 Q2; American Family Insurance is using Data Cloud to create a 360-degree view of customers; Adecco Group is using Data Cloud to create seamless access to information for 27,000 of its employees; Windhma is using Data Cloud to unify profiles of 165 million guest records, many of which are duplicates across multiple sources

This type of performance from our Agentforce platform wouldn’t be possible without Data Cloud. One of the reasons that our agents are so accurate is because of the huge amount of data and metadata that we had. And data is the foundation for every AI transformation. And with Data Cloud, we’re providing a high-performance data lake that brings together all our customer and business data, federating data from external repositories through this credible zero-copy alliance. So customers can use our Data Cloud and then federate and connect to all their other data clouds and then we can bring it all together to deliver the super accurate AI. 

That’s why Data Cloud is absolutely our fastest-growing organic product in history. It will be the fastest product to $1 billion — it’s going to probably be the fastest product of $5 billion, $10 billion. In Q2, the number of paid Data Cloud customers grew 130% year-over-year and the number of customers spending more than $1 million annually have already doubled. In the second quarter alone, and this is amazing, data Cloud processed 2.3 quadrillion records with 110% platform consumption growth year-over-year…

…American Family Insurance with millions of policyholders nationwide is using Data Cloud to consolidate data from multiple sources through our zero-copy partner network, creating a 360-view of the customers, enabling quick segmentation and activating lead data, including their real-time web interactions. The Adecco Group expanded their data cloud in the quarter, a great example of a company leveraging its gold mine of data to gain a unified view of its customers. Connecting all this data means that 27,000 Adecco employees using Salesforce will have seamless access to key information, including financial metrics and job fulfillment status, to help Adecco improve their job fill rate ratio and reduce their cost to serve…

…Wyndham utilizes Data Cloud to unify profiles of 165 million guest records, many of which were duplicates across many sources like Amazon Redshift and the Sabre Reservation System as well as Sales Cloud, Marketing Cloud and Service Cloud. 

Salesforce has rewritten all of its software to be under one unified platform; management thinks building AI agents without a unified platform is risky; the decision to unite all of Salesforce’s software was made 18 months ago with the shift to AI

We’ve automated every customer touch point and now we’re bringing these apps, data and agents together. It’s these 3 levels, and this is in 3 separate pieces of code or 3 different platforms or 3 different systems. This is 1 platform. We’ve rewritten all of our acquisitions, all of our core modules, our Data Cloud and our agents as 1 unified platform, which is how we are delivering not only this incredible functionality but this high level of accuracy and capability. And from this first-hand experience in meeting with these customers around the globe, I can unequivocably tell you that building these agents without a complete integrated platform is like trying to assemble a plane mid-flight, it’s risky chaotic and it’s not likely to succeed…

…With the shift to AI, it just became clear 18 months ago, we need to hit the accelerator pedal and rewrite all these things onto the core platform because customers are going to get this incredible value by having 1 integrated system, and it scales from small companies to extremely large companies. 

Bookings for Salesforce’s AI products was up more than 100% year-on-year in 2024 Q2; Salesforce signed 1,500 AI deals in 2024 Q2; aircraft maker Bombardier is using Salesforce’s AI products to arm sales reps with better information on, and recommendations for, prospects

We’re already accelerating this move from AI hype to AI reality for thousands of customers with amazing capabilities across our entire AI product portfolio. New bookings for these products more than doubled quarter-over-quarter. We signed 1,500 AI deals in Q2 alone. Some of the world’s largest brands are using AI solutions, including Alliant, Bombardier and CMA CGM. Bombardier, the maker of some of the world’s top performing aircraft, is enabling sales reps to sell smarter by consolidating need to know information on prospects in advance of meetings and providing recommendations on how to best engage with them through the Einstein copilot and prompt builder. 

Salesforce has a new team called Salesforce CTOs that will work alongside customers in deploying AI agents

To help our customers navigate this new world, we just launched a new team called Salesforce CTOs. These are deeply technical individuals who work alongside our customers to help them create and execute a plan for every stage of their AI journey to become agent first companies. 

Salesforce sees itself as customer zero for all its AI products, including Agentforce, and it is deploying its own AI products internally with success; 35,000 Salesforce employees are using Einstein as an AI assistant; Salesforce has already used Slack AI to create 500,000 channel summaries since February 2024, saving 3 million hours of work

We’re continuing our own AI journey internally as a Customer Zero of all of our products with great results. We now have 35,000 employees using Einstein as a trusted AI assistant, helping them work smarter and close deals faster. And since we launched Slack AI in February, our employees have created more than 500,000 channels — channel summaries, saving nearly 3 million hours of work. We’ll, of course, deploy Agentforce agents soon in a variety of different roles and tasks to augment, automate and deliver productivity and unmatched experiences for all employees and customers at scale.

Salesforce will be introducing Industry Toolkit at Dreamforce; Industry Toolkit contains more than 100 ready-to-use AI-powered actions; Industry Toolkit can be used with Agentforce 

At Dreamforce, we’re excited to share our new AI tool kit — industry toolkit, which features more than 100 ready-to-use customizable AI-powered actions. All of these actions can be applied to build industry-specific agents with Agentforce.

Salesforce’s management wants to see 1 billion AI agents by FY2026; there are already 200 million agents identified in trials

I’ll just give you my own personal goals. So I’m not giving any guidance here. My goal is that by the end of fiscal year ’26 that we will have 1 billion agents. Already in just looking at the number of consumers identified just in the trials that we have going on, we have like 100 million identified or more. Okay. call it, 200 million. But the funny thing is, of course, it’s only 1 agent. But let’s just think it’s like a manifestation of all these agents talking to all these consumers.

Salesforce already has a long history of selling non-human consumption-based products; with AI agents, management sees pricing on a consumption basis or on a per conversation basis (at $2 per conversation); management thinks AI agents is a very high-margin opportunity

On pricing. When you think about — when you think about apps and you think about humans, because humans use apps, not in all cases. So for example, the Data Cloud is a consumption product. The Commerce Cloud is a consumption product. Of course, the e-mail product, Marketing Cloud is a consumption product. Heroku is a consumption product. So of course, we’ve had non-human consumption-based products for quite a long time at Salesforce…

…When we look at pricing, it will be on a consumption basis. And when we think about that, we think about saying to our customers, and we have, it’s about $2 per conversation. So, that is kind of how we think about it, that we’re going to have a lot of agents out there, even though it’s only 1 agent. It’s a very high margin opportunity, as you can imagine, and we’re going to be reaching — look, you have to think about these agents are like, this is the new website. This is your new phone number. This is how your customers are going to be connecting with you in this new way, and we’re going to be helping our customers to manage these conversations. And it’s probably a per conversational charge as a good way to look at it or we’re selling additional consumption credits like we do with our data cloud. 

Veeva Systems (NYSE: VEEV)

Veeva’s management is seeing the company’s customers appreciate the patience they have displayed in adopting AI; customers started using Veeva’s Vault Direct Data API for AI use cases in 2024 Q2; Vault Direct Data API provides data access 100 times faster than traditional APIs; management thinks that the advantage of providing API access for AI use cases is the possibility of partners developing use cases that management could not even forsee; customers have to pay a fee to turn on Vault Direct Data API and the fee is for covering Veeva’s compute costs; there’s no heavy implementation tasks needed for Vault Direct Data API

When groundbreaking technology like GenAI is first released, it takes time for things to settle and become clearer. That’s starting to happen now. Customers have appreciated our taking the long view on AI and our orientation to tangible value rather than hype. In Q2, our first early customers started using the Vault Direct Data API to power AI and other use cases. The ability to retrieve data 100 times faster than traditional APIs is a major software platform innovation and will be a big enabler of AI that uses data from Vault applications…

… When you make an API like the Direct Data API, you don’t know the innovation you’re unleashing. And that’s the whole point because the data can be consumed so fast and transactionally accurately, use cases that weren’t practical before can become practical. I mean if I step back way back when to designing the first salesforce.com API, I knew it was going to unleash a lot of innovation, and you just don’t know. It’s not predictable, and that’s the good thing…

…[Question] Looking at Vault Direct Data API, how seamless is it for customers to turn it on and start using it? Is it something that needs an implementation? 

[Answer] That is something that’s purchased by the customer, so that is something that is not free for the customers to use. They purchase it. The fee is not that large. It covers our compute cost, that type of thing… 

…After that, no, there’s no implementation. You turn it on, and it’s on. And that’s that.

Veeva’s AI Partner Program is progressing well and has seen 30 AI use cases being developed by 10 partners across Veeva Development Cloud and Veeva Commercial Cloud; the AI use cases in Veeva Commercial Cloud are mostly related to data science while the use cases in Veeva Development Cloud are mostly related to generation of documents and reports; management does not want to compete with the partners that are in the AI Partner Program 

Our AI Partner Program is also progressing well. We now have more than 10 AI partners supporting roughly 30 use cases across R&D and Commercial. We also continue to explore additional AI application opportunities beyond our current AI solutions…

… [Question] You talked about some of the early traction you’re seeing with the AI Partner Program. Can you maybe talk about what are some of the use cases you’ve seen so far?

[Answer] The types of use cases in commercial often have to do with data science. So things like next best action, dynamic targeting, pre-call planning, things like that. And then in R&D, they can be more things like document generation, generate a clinical study report or doing specific medical coding, things like that. So those are the type of use cases…

…In terms of us monitoring that and informing our own road map, I guess there may be some of that. But mostly, that type of innovation really comes from internally our own thinking with our customers. We don’t want to really disrupt our partners, especially when the partners are having customer success. If there’s a major use case that we’re very clear that customers need and for some reason, the ecosystem is not delivering customer success, yes, maybe we might step in there. But I would guess that what we would do would be more holistic, I guess, in some sense and not specifically something a partner would tackle because we’re generally going to have more resources and more ability to sway our own road map than a partner would, and we want to be respectful to the ecosystem.

Zoom Video Communications (NASDAQ: ZM)

Zoom’s management is seeing customers seeking out the AI capabilities of Zoom’s Contact Center packages; Zoom’s management saw the ASP (average selling price) for its Contact Center product double sequentially because of the product’s AI-tier, which comes with higher pricing

We are seeing increased adoption of our advanced Contact Center packages, as customers seek to utilize our AI capabilities to enhance agent performance…

…If you remember, we started with one pricing tier. We eventually added two more and the AI agent is like that Eric was speaking about earlier, is in the highest tier. We actually saw our ASPs for Contact Center almost double quarter-over-quarter because it’s such a premium feature. And when I look at the Q2 deals, the majority of them were purchasing in one of the top 2 tiers, so all of that is contributing to what I would say is not only expansion in terms of seat count but expansion in terms of value being derived from the product.

Zoom’s AI companion uses generative AI to produce meeting summaries, live translations, image generation and more; Zoom AI Companion is now enabled on over 1.2 million accounts; management wIll be upgrading AI companion as Zoom transitions into the 2.0 phase of AI-enabled work; customers really like ZoomAI Companion; Zoom AI Companion is provided at no additional cost; in Zoom meetings, the No.1 use case of Zoom AI Companion is to create meeting summaries; management is constantly improving the quality of Zoom AI Companion’s meeting summaries; customers are giving positive feedback on Zoom AI companion

Today, AI Companion enhances an employee’s capabilities using generative AI to boost productivity through features like meeting summary, chat compose, image generation, live translation and enhanced features in Contact Center. As these features have grown in popularity, we are happy to share that Zoom AI Companion is now enabled on over 1.2 million accounts…

…Our progress broadening Zoom Workplace, building out enhanced AI tools for Contact Center and amassing a large base of AI users sets us up well to transition into the 2.0 phase of AI-enabled work. In this phase, Zoom AI Companion will move beyond enhancing skills to simplifying your workday, providing contextual insights, and performing tasks on your behalf. It will do this by operating across our collaboration platform to ensure your day is interconnected and productive…

…Our customers really like Zoom AI Companion. First of all, it works so well. Secondly, at no additional cost, not like some of other vendors who got to charge the customer a lot. And in our case, this is a part of our package…

… You take a Meeting, for example, right? For sure, the #1 use case like a meeting summary, right? And we keep improving that quality like in the [indiscernible] and or meeting summary are getting better and better. Like in July, we had another upgrade quarter-wise, even better than previous deliveries, right?..

… [Question] One question I had is when you’re looking at Zoom AI Companion, we’ve heard a lot of great things in the field if customers kind of comparing that to other products that are offered out there. Can you kind of remind us about how you guys think about tracking success with the product internally, given that you don’t kind of charge for it directly beyond having millions of people using it?

[Answer] The metrics that we’ve been talking about on here is account activation. So looking at how many — it’s not individual users, it’s actual customer accounts that have activated it… And also they share the stories like how Zoom AI Companion like is very accurate summary, action items are helping their employees’ productivity as well. And yes, a lot of very positive feedback about adopting Zoom AI Companion.

Zoom’s management has intention to monetise AI services for the Contact Center product, but not for Zoom Workplace

[Question] Now that you’re seeing more adoption, Kelly, of Zoom Companion, how do you think about the cost of providing these generative AI features and capabilities? And do you think Zoom could eventually charge on a usage basis for power users of the generally just trying to weigh cost versus revenue opportunities here?

[Answer] I mean when we launched AI Companion, right? So we do not want to charge the customer. However, that’s for the workplace for the business services like a Contact Center, all those new offerings. And I think for sure, we are going to monetize. As I mentioned in the previous earnings calls, new — new solutions or the billing services, AI, I think we are going to charge. They are AI Companion, right? But the workplace and our core you see offering and collaboration offering we do not want to charge. I want to see — I really appreciate our AI team’s great effort, right? And focus on the quality, focus on the cost reduction and so on and forth.

AI services are a drag on Zoom’s margins at the moment (as the company is providing a lot of AI services for free now) but management sees them as important investments for growth

[Question] Just on gross margins, like the impact of generative AI and maybe what you can do to alleviate some of that off there.

[Answer] I mean we’re guiding to 79% for this year, which we will, reflects the prioritization of AI, but also the very strong discipline that we continue to apply. And we are holding to our long-term target for gross margins of 80%. But of course, we think at this point in time, it’s very important to prioritize these investments as they really set us up for future growth.

Zoom’s dev ops team is saving costs for the company to make room for more AI investments

I also want to give a credit to our dev ops team. On the right hand, for sure, we are going to buy more and more GPUs, right? And also leverage that. Our team tried to save the money from other areas, fully automated, and so on and so forth, right? So that’s another way for us to save the cost, right, to make some room for AI.

The regulatory environment for AI in the US and Europe has so far had very little impact on Zoom’s business because Zoom’s management has been adamant and clear that it is not going to use customer’s data to train its AI models
[Question] Are you seeing anything in the broad sweep of AI regulation in the U.S. or Europe that you think can dampen innovation?

[Answer] That’s the reason why we launch AI Companion, we already mentioned, we are not going to use any of our customer data to train our AI models, right? And we take customers data very, very seriously, right? And as a customer, they know that they trust our brand and trust of what we’re doing. And so far, I do not see any impact in terms of like regulation. And again, this AI is moving rapidly, right? So almost the EMEA here and we all look at the potential regulation. But so far, impact actually to us, to our business, I think it’s extremely limited. So like meeting summary, and it’s a very important feature, customer like that. I think we do not use our customer data to train our AI model. And so why not keep using the feature? I think there’s no impact so far.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Adobe, MongoDB, Salesforce, Veeva Systems, and Zoom Video Communications. Holdings are subject to change at any time.

Does News Move The Stock Market?

If we’re constantly looking for news to explain short-term stock price movements, how often can we be right?

A great book I started reading recently is Making Sense of Chaos by economist J. Doyne Farmer. In the book, Farmer discusses his ideas for understanding economies through the lens of complexity science, which is the study of complex adaptive systems. The book referenced an interesting academic finance paper published in 1988 titled What Moves Stock Prices. The paper, authored by David Cutler, James Poterba, and Larry Summers, investigated the influence of news on stock prices.

Farmer described their work as such:

“Cutler, Poterba and Summers began by finding the 100 largest daily fluctuations in the S&P 500 index between 1946 and 1987. They then looked at the New York Times on the day after each move and recorded a summary of the paper’s explanation for the price change. The authors made a subjective judgement as to whether these explanations could plausibly be considered ‘real news’ – or at least real enough to have triggered a sizable change in stock price.”

The largest daily move in the paper’s dataset occurred on 19 October 1987 – now famously known as Black Monday – when the S&P 500 fell by 20.5%. Interestingly, there was no substantial news to explain the collapse. Farmer mentioned in his book:

“The explanations for the 20 per cent drop on October 19, 1987, were ‘worry over dollar decline and rate deficit’ and ‘fear of US not supporting dollar’. Cutler, Poterba and Summers didn’t classify this as news, and I agree. ‘Worry’ and ‘fear’ are subjective statements about the emotional state of the market that have no specific reference to external events.”

Farmer went on to mention:

“Of the dozen largest price fluctuations [shown below], only four were attributed to real news events, a ratio that they found also roughly applied to the largest 100 moves.”

In other words, as I have suspected to be the case for as long as I have been investing, stock prices are indeed more often than not driven by factors outside of the news. I find this to be an important trait of the stock market to know because if we’re constantly looking for news to explain short-term stock price movements, we’re likely to be wrong often, and this can impair our investment decision-making process.

The twelve largest daily price fluctuations in Cutler, Poterba and Summers’ dataset for What Moves Stock Prices:

  1. Date: 19 October 1987
    • Daily change: -20.5%
    • Explanation given: Worry over dollar decline and trade deficit; Fear of US not supporting dollar
  2. Date: 21 October 1987
    • Daily change: 9.1%
    • Explanation given: Interest rates continue to fall; deficit talks in Washington; bargain hunting
  3. Date: 26 October 1987
    • Daily change: -8.3%
    • Explanation given: Fear of budget deficits; margin calls; reaction to falling foreign stocks
  4. Date: 3 September 1946
    • Daily change: -6.7%
    • Explanation given: “… no basic reason for the assault on prices.”
  5. Date: 28 May 1962
    • Daily change:-6.7%
    • Explanation given: Kennedy forces rollback of steel price hike
  6. Date: 26 September 1955:
    • Daily change: – 6.6%
    • Explanation given: Eisenhower suffers heart attack
  7. Date: 26 June 1950:
    • Daily change: -5.4%
    • Explanation given: Outbreak of Korean War
  8. Date: 20 October 1987
    • Daily change: 5.3%
    • Explanation given: Investors looking for “quality stocks”
  9. Date: 9 September 1946
    • Daily change: -5.2%
    • Explanation given: Labor unrest in maritime and trucking industries
  10. Date: 16 October 1987
    • Daily change: -5.2%
    • Explanation given: Fear of trade deficit; fear of higher interest rates; tension with Iran
  11. Date: 27 May 1970
    • Daily change: 5.0%
    • Explanation given: Rumours of change in economic policy; “… the stock surge happened for no fundamental reason”
  12. Date: 11 September 1986
    • Daily change: -4.8%
    • Explanation given: Foreign governments refuse to lower interest rates; crackdown on triple witching announced

Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have no vested interest in any companies mentioned. Holdings are subject to change at any time.

The Buyback Endemic

Buying back stock at unreasonably high valuations is not a good use of capital and can destroy shareholder value.

Buybacks can be a good way for companies to enhance shareholder value. Share buybacks reduce the number of shares outstanding, allowing companies to pay a higher dividend per share in the future.

But not all buybacks are good. Done at the wrong price, buybacks can actually be a bad use of capital. In fact, I have seen so many companies do buybacks recklessly and without consideration of the share price.

The problem probably arises from a few reasons. 

Wrong mindset

First, some executives do not have a good grasp of what buybacks are. Take this statement from Tractor Supply’s management in its 2024 second-quarter earnings report for example:

“The Company repurchased approximately 0.5 million shares of its common stock for $139.2 million and paid quarterly cash dividends totaling $118.5 million, returning a total of $257.7 million of capital to shareholders in the second quarter of 2024.”

The issue with this statement is that it lumps dividends and share repurchases in the same bracket. It also implies that share repurchases are a form of returning capital to shareholders. The truth is that share repurchases is not returning cash to long-term shareholders but only to exiting shareholders. If management mistakes repurchases as capital return, it may lead them to do buybacks regularly, instead of opportunistically.

Although I am singling out Tractor Supply’s management, they are just one out of many management teams that seem to have the wrong mindset when it comes to buybacks.

Incentives

Additionally, executive compensation schemes may encourage management to buy back stock even if it is not the best use of capital. 

For instance, Adobe’s executives have an annual cash remuneration plan that is determined in part by them achieving certain earnings per share goals. This may lead management to buy back stock simply to boost the company’s earnings per share. But doing so when prices are high is not a good use of capital. When Adobe’s stock price is high, it would be better for management to simply return dividends to shareholders – but management may not want to pay dividends as it does not increase the company’s earnings per share.

Again, while I am singling out Adobe’s management, there are numerous other companies that have the same incentive problem.

Tax avoidance

I have noticed that the buyback phenomena is more prevalent in countries where dividends are taxed. 

The US, for instance, seems to have a buyback endemic where companies buy back stock regardless of the price. This may be due to the fact that US investors have to pay a tax on dividends, which makes buybacks a more tax-efficient use of capital for shareholders. On the contrary, Singapore investors do not need to pay taxes on dividends. As such, Singapore companies do not do buybacks as often.

However, simply doing buybacks for tax efficiency reasons without considering the share price can still harm shareholders. Again, management teams need to weigh both the pros and cons of buybacks before conducting them.

Final thoughts

There is no quick fix to this problem but there are some starting points that I believe companies can do to address the issue. 

First, fix the incentives problem. A company’s board of directors need to recognise that incentives that are not structured thoughtfully can encourage reckless buybacks of shares regardless of the share price.

Second, management teams need to educate themselves on how to increase long-term value for shareholders and to understand the difference between buybacks and dividends.

Third, management teams need to understand the implications of taxes properly. Although it is true that taxes can affect shareholders’ total returns when a company pays a dividend, it is only one factor when it comes to shareholder returns. Executive teams need to be coached on these aspects of capital allocation.

Only through proper education and incentives, will the buyback endemic be solved.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Adobe and Tractor Supply. Holdings are subject to change at any time.

The Latest Thoughts From American Technology Companies On AI (2024 Q2)

A collection of quotes on artificial intelligence, or AI, from the management teams of US-listed technology companies in the 2024 Q2 earnings season.

The way I see it, artificial intelligence (or AI), really leapt into the zeitgeist in late-2022 or early-2023 with the public introduction of DALL-E2 and ChatGPT. Both are provided by OpenAI and are software products that use AI to generate art and writing, respectively (and often at astounding quality). Since then, developments in AI have progressed at a breathtaking pace.

With the latest earnings season for the US stock market – for the second quarter of 2024 – coming to its tail-end, I thought it would be useful to collate some of the interesting commentary I’ve come across in earnings conference calls, from the leaders of technology companies that I follow or have a vested interest in, on the topic of AI and how the technology could impact their industry and the business world writ large. This is an ongoing series. For the older commentary:

With that, here are the latest commentary, in no particular order:

Airbnb (NASDAQ: ABNB)

Airbnb’s management is still really excited about AI, but they’ve also realised that it’s going to take a lot longer for applications to change; management sees three layers to AI, namely, the chip, the model, and the application, and while there’s been a lot of innovation on the chip and the model, not much has changed with the applications, especially in e-commerce and travel

ChatGPT launched late November 2022. When it launched, I think we all got like incredibly excited. It was kind of like the moment probably some of us first discovered the Internet or maybe when iPhone was launched. And when it was launched, you had the feeling that everything was going to change. But I think that’s still true. But I think one of the things we’ve learned over the last, say, 18 months or nearly 2 years — 22 months since ChatGPT launched is that it’s going to take a lot longer than people think for applications to change.

If I were to think of AI, I’d probably think about it in 3 layers. You have the chip. You have the model. And you have the applications. There’s been a lot of innovation on the chip. There’s been a lot of innovation on the model. We have a lot of new models, and there’s a prolific rate of improvement in these models. But if you look at your home screen, which of your apps are fundamentally different because of the AI, like fundamentally different because of generative AI? Very little, especially even less in e-commerce or travel. And the reason why is I think it’s just going to take time to develop new AI paradigm. 

Airbnb’s management sees ChatGPT, even though it’s an AI chat software, as an application that could have existed before AI; management thinks what needs to be done is to develop AI applications that are native to the AI models with unique interfaces and no one has done this year; Airbnb is working on an application that will be native to AI models and this will change how users interact with Airbnb, where it becomes much more than a search box; this change in Airbnb will take a few years to develop

ChatGPT [ is an AI model interface that could ] have existed before AI. And so all of our paradigms are pre-AI paradigms. And so what we need to do is we need to actually develop AI applications that are native to the model. No one has done this yet. There’s not been one app that I’m aware of that’s the top 50 app in the app store in the United States that is a fundamentally new paradigm as fundamentally different as a multitouch was to the iPhone in 2008, and we need that interface change. So that’s one of the things that we’re working on. And I do think Airbnb will eventually be much more than a search box where you type a destination, add dates and find a listing. It’s going to be much more of a travel concierge. It’s having a conversation, learning, adapting to you. It’s going to take a number of years to develop this. And so it won’t be in the next year that this will happen. And I think this is probably what most of my tech friends are also saying, is it’s going to just take a bit more time.

Airbnb’s management thinks that having a new AI-driven interface will allow Airbnb to expand into new businesses

But to answer your question on what’s possible, a new interface paradigm would allow us to attach new businesses. So the question is, what permission do we have to go into a business like hotels? Well, today, we have permission because we have a lot of traffic. But if we had a breakthrough interface, we have even more permission because suddenly, we could move top of funnel and not just ask where are you going, but we can point to — we can inspire where you travel. Imagine if we had an index of the world’s communities. We told you we had information about every community, and we can provide the end-to-end trip for you. So there’s a lot of opportunities as we develop new interfaces to cross-sell new more inventory. 

Alphabet (NASDAQ: GOOG)

Google Cloud’s year-to-date AI-related revenue is already in the billions, and its AI infrastructure and solutions are already used by >2 million developers; more than 1.5 million developers are using Gemini, Alphabet’s foundational AI model, across the company’s developers tools

Year-to-date, our AI infrastructure and generative AI solutions for cloud customers have already generated billions in revenues and are being used by more than 2 million developers…

…More than 1.5 million developers are now using Gemini across our developer tools.

Alphabet’s management thinks Alphabet is well-positioned for AI; Alphabet is innovating at every layer of the AI stack, from chips at the bottom to agents at the top

As I spoke about last quarter, we are uniquely well positioned for the AI opportunity ahead. Our research and infrastructure leadership means we can pursue an in-house strategy that enables our product teams to move quickly. Combined with our model building expertise, we are in a strong position to control our destiny as the technology continues to evolve. Importantly, we are innovating at every layer of the AI stack, from chips to agents and beyond, a huge strength.

Alphabet’s management thinks Alphabet is using AI to deliver better responses on Search queries; tests for AI Overviews has showed increase in Search usage and higher user satisfaction; Search users with complex searches keep coming back for AI Overviews; users aged 18-24 have higher engagement when using Search with AI Overviews; Alphabet is prioritising AI-approaches that send traffic to websites; ads that are above or below AI Overviews continue to be valuation; in 2024 Q2, management has doubled the core model size for AI Overviews while improving latency and keeping cost per AI Overviews served flat; management is working on matching the right AI model size to the query’s complexity to improve cost and latency; AI Overviews is rolled out in the USA and will be rolled out to more countries throughout 2024; Alphabet will soon put Search and Shopping ads within the AI Overviews for USA users

With AI, we are delivering better responses on more types of search queries and introducing new ways to search. We are pleased to see the positive trends from our testing continue as we roll out AI Overviews, including increases in Search usage and increased user satisfaction with the results. People who are looking for help with complex topics are engaging more and keep coming back for AI Overviews. And we see even higher engagement from younger users aged 18 to 24 when they use Search with AI Overviews. As we have said, we are continuing to prioritize approaches that send traffic to sites across the web. And we are seeing that ads appearing either above or below AI Overviews continue to provide valuable options for people to take action and connect with businesses…

…Over the past quarter, we have made quality improvements that include doubling the core model size for AI Overviews while at the same time improving latency and keeping cost per AI Overviews served flat. And we are focused on matching the right model size to the complexity of the query in order to minimize impact on cost and latency…

…On the AI Overviews, we are — we have rolled it out in the U.S. And we are — will be, through the course of the year, definitely scaling it up, both to more countries…

…And as you have probably noticed at GML, we announced that soon we’ll actually start testing search and shopping ads in AI Overviews for users in the U.S., and they will have the opportunity to actually appear within the AI Overview in a section clearly labeled as sponsored when they’re relevant to both the quarry and the information in the AI Overview, really giving us the ability to innovate here and take this to the next level.

AI opens up new ways to use Search, such as asking questions by taking a video with Lens; AI Overviews in Lens has led to higher overall visual search usage; Circle to Search is another new way to search, and is available on >100 million Android devices

AI expands the types of queries we are able to address and opens a powerful new ways to Search. Visual search via Lens is one. Soon, you’ll be able to ask questions by taking a video with Lens. And already, we have seen that AI Overviews in Lens leads to an increase in overall visual search usage. Another example is Circle to Search, which is available today on more than 100 million Android devices.

Gemini, now has 4 sizes, each with their own use cases; Gemini comes with a context window of 2 million, the longest of any foundation model to-date; all of Alphabet’s 6 products with more than 2 billion monthly users are using Gemini; through Gemini, users of Google Photos can soon ask questions of their photos and receive answers

Gemini now comes in 4 sizes with each model designed for its own set of use cases. It’s a versatile model family that runs efficiently on everything from data centers to devices. At 2 million tokens, we offer the longest context window of any large-scale foundation model to date, which powers developer use cases that no other model can handle. Gemini is making Google’s own products better. All 6 of our products with more than 2 billion monthly users now use Gemini…

…At I/O, we showed new features coming soon to Gmail and to Google Photos. Soon, you’ll be able to ask Photos questions like, what did I eat at that restaurant in Paris last year?

During Alphabet’s recent developer conference, I/O, management showed their vision of what a universal AI agent could look like

For a glimpse of the future, I hope you saw Project Astra at I/O. It shows multimodal understanding and natural conversational capabilities. We’ve always wanted to build a universal agent, and it’s an early look at how they can be helpful in daily life.

Alphabet has launched Trillium, the sixth-generation of its custom TPU AI accelerator; Trillium has a 5x increase in peak compute performance per chip and a 67% improvement in energy efficiency over TPU v5e

Trillium is the sixth generation of our custom AI accelerator, and it’s our best-performing and most energy-efficient TPU to date. It achieves a near 5x increase in peak compute performance per chip and a 67% more energy efficient compared to TPU v5e.

Google Cloud’s enterprise AI platform, Vertex, is used by Deutsche Bank, Kingfisher, and the US Air Force to build AI agents; Uber and WPP are using Gemini Pro 1.5 and Gemini Flash 1.5 in Vertex for customer experience and marketing; Vertex has broadened support for 3rd-party AI models, including Anthropic’s Claude 3.5 Sonnet, Meta’s Llama, and Mistral’s models

Our enterprise AI platform, Vertex, helps customers such as Deutsche Bank, Kingfisher and the U.S. Air Force build powerful AI agents. Last month, we announced a number of new advances. Uber and WPP are using Gemini Pro 1.5 and Gemini Flash 1.5 in areas like customer experience and marketing. We broadened support for third-party models including Anthropic’s Claude 3.5 Sonnet and open-source models like Gemma 2, Llama and Mistral. 

Google Cloud is the only cloud provider to provide grounding with Google Search; large enterprises such as Moody’s, MSCI, and ZoomInfo are using Google Cloud’s grounding capabilities

We are the only cloud provider to offer grounding with Google Search, and we are expanding grounding capabilities with Moody’s, MSCI, ZoomInfo and more.

Google Cloud’s AI-powered applications are helping it to drive upsells and win new customers; Best Buy and Gordon Food Service are using Google Cloud’s conversational AI platform; Click Therapeutics is using Gemini for Workspace; Wipro is using Gemini Code Assist to speed up software development; MercadoLibre is using BigQuery and Looker for capacity planning and speeding up shipments.

Our AI-powered applications portfolio is helping us win new customers and drive upsell. For example, our conversational AI platform is helping customers like Best Buy and Gordon Food Service. Gemini for Workspace helps Click Therapeutics analyze patient feedback as they build targeted digital treatments. Our AI-powered agents are also helping customers develop better-quality software, find insights from their data and protect their organization against cybersecurity threats using Gemini. Software engineers at Wipro are using Gemini Code Assist to develop, test and document software faster. And data analysts at Mercado Libre are using BigQuery and Looker to optimize capacity planning and fulfill shipments faster.

 In 2024 Q2, Alphabet announced more than 30 new ads features and products to help advertisers leverage AI; Alphabet is applying AI across its advertising products to streamline workflows, enhance asset creation, and improve engagement with consumers; in asset creation, any business using Product Studio can upload an image and enhance it with AI; AI features for consumers such as virtual try-ons in shopping ads are in beta-testing, and feedback shows that virtual try-on gets 60% more high-quality views; advertisers using Alphabet’s AI-powered profit maximisation tools along with Smart Bidding see a 15% increase in profit; Demand Gen, to be rolled out in the coming months, creates high-quality image assets for social marketers and delivers 14% more conversions when paired with Search or Performance Max; Tiffany used Demand Gen and achieved a 2.5% lift in consideration and important customer-actions, and a 5.6x improvement in cost per click compared to social media benchmarks; Alphabet used Demand Gen to create 4,500 ad variations for Pixel 8’s advertising campaigns and delivered twice the clicks per rate at nearly 1/4 of the cost

This quarter, we announced over 30 new ads features and products to help advertisers leverage AI and keep pace with the evolving expectations of customers and users. Across Search, PMax, Demand Gen and retail, we’re applying AI to streamline workflows, enhance creative asset production and provide more engaging experiences for consumers.

Listening to our customers, retailers in particular have welcomed AI-powered features to help scale the depth and breadth of their assets. For example, as part of the new and easier-to-use Merchant Center, we’ve expanded Product Studio with tools that bring the power of Google AI to every business owner. You can upload a product image, prompt the AI with something like feature this product with Paris skyline in the background, and Product Studio will generate campaign-ready assets.

I also hear great feedback from our customers on many of our other new AI-powered features. We’re beta testing virtual try-on in shopping ads and plan to roll it out widely later this year. Feedback shows this feature gets 60% more high-quality views than other images and higher click out to retailer sites. Retailers love it because it drives purchasing decisions and fewer returns.

Our AI-driven profit optimization tools have been expanded to Performance Max and standard shopping campaigns. Advertisers who use profit optimization and Smart Bidding see a 15% uplift in profit on average compared to revenue-only bidding.

Lastly, Demand Gen is rolling out to Display & Video 360 and Search Ads 360 in the coming months with new generative image tools that create stunning, high-quality image assets for social marketers. As we said at GML, when paired with Search or PMax, Demand Gen delivers an average of 14% more conversions…

…Luxury jewelry retailer Tiffany leveraged Demand Gen during the holiday season and saw a 2.5% brand lift in consideration and actions such as adding items to carts and booking appointments. The campaign drove a 5.6x more efficient cost per click compared to social media benchmarks. Our own Google marketing team used Demand Gen to create nearly 4,500 ad variations for Pixel 8 campaign shown across YouTube, Discover and Gmail, delivering twice the clicks per rate at nearly 1/4 of the cost.

Alphabet has used AI to (1) improve broad match performance by 10% in 6 months for advertisers using Smart Bidding, and (2) increase conversions by 25% at similar cost for advertisers who adopt PMax to broad match and Smart Bidding in their Search campaigns

In just 6 months, AI-driven improvements to quality, relevance and language understanding have improved broad match performance by 10% for advertisers using Smart Bidding. Also, advertisers who adopt PMax to broad match and Smart Bidding in their Search campaigns see an average increase of over 25% more conversions of value at a similar cost.

Google Cloud had 29% revenue growth in 2024 Q2 (was 28% in 2024 Q1); operating margin was 11% (was 9% in 2024 Q1 and was 4.9% in 2023 Q2); Google Cloud’s accelerating revenue growth in 2024 Q2 was partly the result of AI demand; GCP’s growth rate is above the growth rate for the overall Google Cloud business

Turning to the Google Cloud segment. Revenues were $10.3 billion for the quarter, up 29%, reflecting, first, significant growth in GCP, which was above growth for Cloud overall and includes an increasing contribution from AI; and second, strong Google Workspace growth, primarily driven by increases in average revenue per seat. Google Cloud delivered operating income of $1.2 billion and an operating margin of 11%…

…[Question] On the cloud acceleration, would you characterize that as new AI demand helping drive that year-to-date? Or is that more of a rebound in just general compute and other demand?

[Answer] There is clearly a benefit as the Cloud team is engaging broadly with customers around the globe with AI-related solutions, AI infrastructure solutions and generative AI solutions. I think we noted that we’re particularly encouraged that the majority of our top 100 customers are already using our generative AI solution. So it is clearly adding to the strength of the business on top of all that they’re doing. And just to be really clear, the results for GCP, the growth rate for GCP is above the growth for Cloud overall.

Alphabet’s big jump capex in 2024 Q2 (was $7.2 billion in 2023 Q2) was mostly for technical infrastructure, in the form of servers and data centers; management continues to expect Alphabet’s quarterly capex for the rest of 2024 to be similar to what was seen in 2024 Q1;

With respect to CapEx, our reported CapEx in the second quarter was $13 billion, once again, driven overwhelmingly by investment in our technical infrastructure with the largest component for servers followed by data centers. Looking ahead, we continue to expect quarterly CapEx throughout the year to be roughly at or above the Q1 CapEx of $12 billion, keeping in mind that the timing of cash payments can cause variability in quarterly reported CapEx.

Alphabet’s management is seeing more tangible use cases for AI in the consumer space compared to the enterprise space; in the consumer space, consumers are engaging with Alphabet’s AI features, but there’s still the question of monetisation; in the enterprise space, a lot of AI models are currently being built and they are converging towards a set of base capabilities; the next wave for the enterprise space will be building applications on top of the models, and there is some traction in some areas, but it’s not widespread yet; management believes value will eventually be unlocked, but it may take time

 I think there is a time curve in terms of taking the underlying technology and translating it into meaningful solutions across the board, both on the consumer and the enterprise side. Definitely, on the consumer side, I’m pleased, as I said in my comments earlier, in terms of how for a product like Search, which is used at that scale over many decades, how we’ve been able to introduce it in a way that it’s additive and enhances overall experience and this positively contributing there. I think across our consumer products, we’ve been able — I think we are seeing progress on the organic side. Obviously, monetization is something that we would have to earn on top of it. The enterprise side, I think we are at a stage where definitely there are a lot of models. I think roughly, the models are all kind of converging towards a set of base capabilities. But I think where the next wave is working to build solutions on top of it. And I think there are pockets, be it coding, be it in customer service, et cetera, where we are seeing some of those use cases are seeing traction, but I still think there is hard work there to completely unlock those…

…But I think we are in this phase where we have to deeply work and make sure on these use cases, on these workflows, we are driving deeper progress on unlocking value, which I’m very bullish will happen. But these things take time. So — but if I were to take a longer-term outlook, I definitely see a big opportunity here. And I think particularly for us, given the extent to which we are investing in AI, our research infrastructure leadership, all of that translates directly. And so I’m pretty excited about the opportunity space ahead.

Alphabet’s management thinks that the risk of underinvesting in AI infrastructure for the cloud business is currently greater than the risk of overinvesting; management thinks that even if Alphabet ends up overinvesting, the infrastructure is still widely useful for internal use cases

[Question] So it looks like from the outside at least, the hyperscaler industry is going from kind of an underbuilt situation this time last year to better meeting the demand with capacity right now to potentially being overbuilt next year if these CapEx growth rates keep up. So do you think that’s a fair characterization? And how are we thinking about the return on invested capital with this AI CapEx cycle?

[Answer] I think the one way I think about it is when we go through a curve like this, the risk of under-investing is dramatically greater than the risk of over-investing for us here, even in scenarios where if it turns out that we are over-investing, we clearly — these are infrastructure which are widely useful for us. They have long useful lives, and we can apply it across, and we can work through that. But I think not investing to be at the front here, I think, definitely has much more significant downside. Having said that, we obsess around every dollar we put in. Our teams are — work super hard. I’m proud of the efficiency work, be it optimization of hardware, software, model deployment across our fleet. All of that is something we spend a lot of time on, and that’s how we think about it.

Amazon (NASDAQ: AMZN)

AWS’s AI business continues to grow dramatically with a multi-billion revenue run rate; management sees AWS’s AI services resonating with customers, who want choice in the AI models and AI chips they use, and AWS is providing them with choices; over the past 18 months, AWS has launched twice as many AI features into general availability than all other major cloud providers combined

Our AI business continues to grow dramatically with a multibillion-dollar revenue run rate despite it being such early days, but we can see in our results and conversations with customers that our unique approach and offerings are resonating with customers. At the heart of this strategy is a firmly held belief, which we’ve had since the beginning of AWS that there is not one tool to rule the world. People don’t want just one database option or one analytics choice or one container type. Developers and companies not only reject it but are suspicious of it. They want multiple options for flexibility and to use the best tool for each job to be done. The same is true in AI. You saw this several years ago when some companies tried to argue that TensorFlow will be the only machine learning framework that mattered and then PyTorch and others overtook it. The same one model or one chip approach dominated the earliest moments of the generative AI boom, but we have a lot of data that suggests this is not what customers want here either, and our AWS team is determined to deliver choice and options for customers…

…During the past 18 months, AWS has launched more than twice as many machine learning and generative AI features into general availability than all the other major cloud providers combined. 

AWS provides NVIDIA chips for AI model builders, but management also hear from customers that they want better price performance and hence AWS developed the Trainium and Inferentia chips for training and inference, respectively; the second version of Trainium is coming later this year and has very compelling price performance; management is seeing significant demand for Trainium and Inferentia; management started building Trainium and Inferentia 5 years ago also because they had the experience of seeing customers wanting better price performance from CPUs; management believes Trainium and Inferentia will generate similarly high ROI as Graviton, Amazon’s custom CPU, does

For those building generative AI models themselves, the cost of compute for training and inference is critical, especially as models get to scale. We have a deep partnership with NVIDIA and the broader selection of NVIDIA instances available, but we’ve heard loud and clear from customers that they relish better price performance. It’s why we’ve invested in our own custom silicon in Trainium for training and Inferentia for inference. And the second versions of those chips, with Trainium coming later this year, are very compelling on price performance. We are seeing significant demand for these chips…

…When we started AWS, we had and still have a very deep partnership with Intel on the generalized CPU space. But what we found from customers is that they — when you find a — an offering that is really high value for you and high return, you don’t actually spend less, even though you’re spending less per unit. You spend less per unit, but it enables you, it frees you up to do so much more inventing and building for your customers. And then when you’re spending more, you actually want better price performance than what you’re getting.

And a lot of times, it’s hard to get that price performance from existing players unless you decide to optimize yourself for what you’re learning from your customers and you push that envelope yourself. And so we built custom silicon in the generalized CPU space with Graviton, which we’re on our fourth model right now. And that has been very successful for customers and for our AWS business, is it saves customers about — up to about 30% to 40% price performance versus the other leading x86 processors that they could use.

And we saw the same trend happening about 5 years ago in the accelerator space in the GPU space, where the products are good, but there was really primarily 1 provider and supply was more scarce than what people wanted. And people — our customers really want improved price performance all the time. And so that’s why we went about building Trainium, which is our training chip, and Inferentia, which is our inference chip, which we’re on second versions of both of those. They will have very compelling relative price performance.

And in a world where it’s hard to get GPUs today, the supply is scarce and all the schedules continue to move over time, customers are quite excited and demanding at a high clip, our custom silicon, and we’re producing it as fast as we can. I think that’s going to have very good return profile just like Graviton has, and I think it will be another differentiating feature around AWS relative to others.

SageMaker, AWS’s fully-managed AI service, helps customers save time and money while they build their AI models; management is seeing model builders standardise on SageMaker

Model builders also desire services that make it much easier to manage the data, construct the models, experiment, deploy to production and achieve high-quality performance, all while saving considerable time and money. That’s what Amazon SageMaker does so well including its most recently launched feature called HyperPod that changes the game and networking performance for large models, and we’re increasingly seeing model builders standardize on SageMaker. 

Amazon Bedrock, AWS’s AI-models-as-a-service offering, caters to companies that want to leverage 3rd-party models and customise with their own data; Bedrock already has tens of thousands of companies using it; Bedrock has the largest selection of models and the best generative AI capabilities in a number of critical areas; Bedrock recently added Anthropic’s Claude 3.5 models, Meta’s new Llama 3.1 models, and Mistral’s new models

While many teams will build their own models, lots of others will leverage somebody else’s frontier model, customize it with their own data, and seek a service that provides broad model selection and great generative AI capabilities. This is what we think of as the middle layer, what Amazon Bedrock does and why Bedrock has tens of thousands of companies using it already. Bedrock has the largest selection of models, the best generative AI capabilities in critical areas like model evaluation, guardrails, RAG and agenting and then makes it easy to switch between different model types and model sizes. Bedrock has recently added Anthropic’s Claude 3.5 models, which are the best performing models on the planet; Meta’s new Llama 3.1 models; and Mistral’s new Large 2 models. And Llama’s and Mistral’s impressive performance benchmarks and open nature are quite compelling to our customers as well.

Amazon’s management is seeing strong adoption of Amazon Q, Amazon’s generative AI assistant for software development; Amazon Q has the highest score and acceptance rate for code suggestions; Amazon Q tests code and outperforms competitors on catching security vulnerabilities; with Amazon Q’s code transformation capabilities, Amazon saved $260 million and 4,500 developer years when performing a large Java Development Kit migration; management thinks Amazon Q can continue to improve and address more use cases  

We’re continuing to see strong adoption of Amazon Q, the most capable generative AI-powered assistant for software development and to leverage your own data. Q has the highest known score and acceptance rate for code suggestions, but it does a lot more than provide code suggestions. It tests code, outperforms all other publicly benchmarkable competitors on catching security vulnerabilities and leads all software development assistance on connecting multiple steps together and applying automatic action.

It also saves development teams time and money on the muck nobody likes to talk about. For instance, when companies decide to upgrade from one version of a framework to another, it takes development teams many months, sometimes years burning valuable opportunity costs and churning developers who hate this tedious though important work. With Q’s code transformation capabilities, Amazon has migrated over 30,000 Java JDK applications in a few months, saving the company $260 million and 4,500 developer years compared to what it would have otherwise cost. That’s a game changer.

And think about how this Q transformation capability might evolve to address other elusive but highly desired migrations. 

Amazon’s management is still very bullish on the medium to long-term impacts of AI, but the progress may not be a straight line; management sees a lot of promise in generative AI being able to improve customer experiences and this is informed by their own experience of using generative AI within Amazon, such as: (1) Rufus, a shopping assistant, improves customers’ shopping decisions, (2) customers can virtually try on apparel, (3) sellers can create new selections with a line or two of text, and (4) better detection of product defects before the products reach customers

We remain very bullish on the medium to long-term impact of AI in every business we know and can imagine. The progress may not be one straight line for companies.

Generative AI especially is quite iterative, and companies have to build muscle around the best way to solve actual customer problems. But we see so much potential to change customer experiences. We see it in how our generative-AI-powered shopping assistant, Rufus, is helping customers make better shopping decisions. We see it in our AI features that allow customers to simulate trying apparel items or changing the buying experience. We see it in our generative AI listing tools enabling sellers to create new selection with a line or 2 of text versus the many forms previously required. We see it in our fulfillment centers across North America, where we’re rolling out Project Private Investigator, which uses a combination of generative AI and computer vision to uncover defects before products reach customers. We see it in how our generative AI is helping our customers discover new music and video. We see it in how it’s making Alexa smart, and we see it in how our custom silicon and services like SageMaker and Bedrock are helping both our internal teams and many thousands of external companies reinvent their customer experiences and businesses. We are investing a lot across the board in AI, and we’ll keep doing so as we like what we’re seeing and what we see ahead of us.

Amazon’s management expects capital expenditures to be higher in 2024 H2 compared to 2024 H1; most of the capex will be for AWS infrastructure in both generative AI and non-generative AI workloads; management has a lot of experience, accumulated over the years, in predicting just the right amount of compute capacity to provide for AWS before the generative AI era, and they believe they can do so again for generative AI; management is investing heavily in AI-related capex because they see a lot of demand and in fact, they would like AWS to have more compute capacity than what it has today

For the first half of the year, CapEx was $30.5 billion. Looking ahead to the rest of 2024, we expect capital investments to be higher in the second half of the year. The majority of the spend will be to support the growing need for AWS infrastructure as we continue to see strong demand in both generative AI and our non-generative AI workloads…

…If you think about the fact that we have about 35 regions and think of a region as multiple — a cluster of multiple data centers and about 110 availability zones, which is roughly equivalent to a data center, sometimes it includes multiple and then if you think about having to land thousands and thousands of SKUs across the 200 AWS services in each of those availability zones at the right quantities, it’s quite difficult. And if you end up actually with too little capacity, then you have service disruptions, which really nobody does because it means companies can’t scale their applications.

So most companies deliver more capacity than they need. However, if you actually deliver too much capacity, the economics are pretty woeful, and you don’t like the returns of the operating income. And I think you can tell from having — we disclosed both our revenue and our operating income in AWS that we’ve learned over time to manage this reasonably well. And we have built models over a long period of time that are algorithmic and sophisticated that land the right amount of capacity. And we’ve done the same thing on the AI side.

Now AI is newer. And it’s true that people take down clumps of capacity in AI that are different sometimes. I mean — but it’s also true that it’s not like a company shows up to do a training cluster asking for a few hundred thousand chips the same day. Like you have a very significant advanced signal when you have customers that want to take down a lot of capacity.

So while the models are more fluid, it’s also true that we’ve built, I think, a lot of muscle and skill over time in building these capacity signals and models, and we also are getting a lot of signal from customers on what they need. I think that it’s — the reality right now is that while we’re investing a significant amount in the AI space and in infrastructure, we would like to have more capacity than we already have today. I mean we have a lot of demand right now, and I think it’s going to be a very, very large business for us.

Companies need to organise their data in specific ways before they can use AI effectively; it’s difficult for companies with on-premise data centers to use AI effectively

It’s quite difficult to be able to do AI effectively if your data is not organized in such a way that you can access that data and run the models on top of them and then build the application. So when we work with customers, and this is true both when we work directly with customers as well as when we work with systems integrator partners, everyone is in a hurry to get going on doing generative AI. And one of the first questions that we ask is show us where your data is, show us what your data lake looks like, show us how you’re going to access that data. And there’s very often work associated with getting your data in the right shape and in the right spot to be able to do generative AI. There — fortunately, because so many companies have done the work to move to the cloud, there’s a number of companies who are ready to take advantage of AI, and that’s where we’ve seen a lot of the growth. But also it’s worth remembering that, again, remember the 90% of the global IT spend being on-premises. There are a lot of companies who have yet to move to the cloud, who will, and the ability to use AI more effectively is going to be one of the many drivers in doing so for them.

Apple (NASDAQ: AAPL)

Apple Intelligence, Apple’s AI technologies embedded in its devices, improves Siri; Apple Intelligence is built on a foundation of privacy and has a ground-breaking approaching to using the cloud, known as Private Cloud Compute, that protects user information; Apple Intelligence is powered by Apple’s custom chips; Apple Intelligence will involve integration with ChatGPT in iPhones, Macs, and iPads; management will continue to invest in AI; because of management’s stance on privacy, Apple Intelligence will maximise the amount of data that is processed directly on people’s devices; Apple Intelligence’s roll out will be staggered; Apple Intelligence’s monetisation appears to involve both the Services business of Apple, and payments from partners

At our Worldwide Developers Conference, we were thrilled to unveil game-changing updates across our platforms, including Apple Intelligence. Apple Intelligence builds on years of innovation and investment in AI and machine learning. It will transform how users interact with technology from Writing Tools to help you express yourself to Image Playground, which gives you the ability to create fun images and communicate in new ways, to powerful tools for summarizing and prioritizing notifications. Siri also becomes more natural, more useful, and more personal than ever. Apple Intelligence is built on a foundation of privacy, both through on-device processing that does not collect users’ data and through Private Cloud Compute, a groundbreaking new approach to using the cloud while protecting users’ information powered by Apple Silicon. We are also integrating ChatGPT into experiences within iPhone, Mac, and iPad, enabling users to draw on a broad base of world knowledge.

We are very excited about Apple Intelligence, and we remain incredibly optimistic about the extraordinary possibilities of AI and its ability to enrich customers’ lives. We will continue to make significant investments in this technology and dedicate ourselves to the innovation that will unlock its full potential…

…We are committed as ever to shipping products that offer the highest standards of privacy for our users. With everything we do, whether it’s offering a browser like Safari that prevents third-parties from tracking you across the Internet, or providing new features like the ability to lock and hide apps, we are determined to keep our users in control of their own data. And we are just as dedicated to ensuring the security of our users’ data. That’s why we work to minimize the amount of data we collect and work to maximize how much is processed directly on people’s devices, a foundational principle that is at the core of all we build, including Apple Intelligence…

…The rollout, as we mentioned in June, sort of we’ve actually started with developers this week. We started with some features of Apple Intelligence, not the complete suite. There are other features like languages beyond U.S. English that will happen over the course of the year, and there are other features that will happen over the course of the year. And ChatGPT is integrated by the end of the calendar year. And so yes, so it is a staggered launch…

…[Question] How should investors think about the monetization models…  in the long term, do you see the Apple Intelligence part, the Services growth from Apple Intelligence being the larger contributor over time? Or do you see these partnerships becoming a larger contributor over time? 

[Answer] The monetization model, I don’t want to get into the terms of the commercial agreements because they’re confidential between the parties, but I see both aspects as being very important. People want both.

Apple is getting its partners to fork out the bill for some of its capex needs for AI cloud compute, so even though its capex will increase over time, it does not seem like the increase may be that high

[Question] Do you see the rollout of these features requiring further increases in R&D or increases in OpEx or CapEx for cloud compute capacity?

[Answer] On the CapEx part, it’s important to remember that we employ a hybrid kind of approach where we do things internally and we have certain partners that we do business with externally where the CapEx would appear in their respective businesses. But yes, I mean, you can expect that we will continue to invest and increase it year-on-year…

…On the CapEx front, as Tim said, we employ a hybrid model. Some of the investments show up on our balance sheet and some other investments show up somewhere else and we pay as we go. But in general, we try to run the company efficiently.

Arista Networks (NYSE: ANET)

Arista Networks recently launched its Etherlink AI platforms that are compatible with the ultra-Ethernet consortium and can lead the migration from Infiniband to Ethernet; the Etherlink AI platforms consist of a portfolio of 800-gig switches and can work with all kinds of GPUs; there are new products in the platform that work well even for very large AI clusters; the Etherlink portfolio is being trialled by customers can support up to be 100,000 XPUs

In June 2024, we launched Arista’s Etherlink AI platforms that are ultra-Ethernet consortium compatible, validating the migration from InfiniBand to Ethernet. This is a rich portfolio of 800-gig products, not just a point product, but in fact, a complete portfolio that is both NIC and GPU agnostic. The AI portfolio consists of the 7060 [indiscernible] switch that supports 64 800-gig or 128 400-gig Ethernet ports with a capacity of 51 terabits per second. The 7800 R4 AI Spine is our fourth generation of Arista’s flagship 7800, offering 100% non-blocking throughput with a proven virtual output queuing architecture. The 7800 R4 supports up to 460 terabits in a single chassis, corresponding to 576800 gigabit Ethernet ports or 1,152400 gigabit port density. The 7700 R4 AI distributed Etherlink Switch is a unique product offering with a massively parallel distributed scheduling and congestion-free traffic spraying fabric. The 7700 represents the first in a new series of ultra-scalable intelligent distributed systems that can deliver the highest consistent throughput for very large AI clusters…

…Our Etherlink portfolio is in the midst of trials and can support up to 100,000 XPUs in a 2-tier design built on our proven and differentiated extensible OS.

Arista Networks had a recent AI enterprise win with a Tier 2 cloud provider to provide Ethernet fabrics for its fleet of NVIDIA H100 GPUs; the cloud provider was using a legacy networking vendor that could not scale

The first example is an AI enterprise win with a large Tier 2 cloud provider which has been heavily investing in GPUs to increase their revenue and penetrate new markets. Their senior leadership wanted to be less reliant on traditional core services and work with Arista on new, reliable and scalable Ethernet fabrics. Their environment consisted of new NVIDIA H100s. However, it was being connected to their legacy networking vendor, which resulted in them having significant performance and scale issues with their AI applications. The goal of our customer engagement was to refresh the front-end network to alleviate these issues. Our technical partnership resulted in deploying a 2-step migration path to alleviate the current issues using 400-gig 7080s, eventually migrating them to an 800-gig AI Ethernet link in the future. 

Arista Networks’ management is once again seeing the network becoming the computer as AI training models require a lossless network to connect every AI accelerator in a cluster to one another; Arista Networks’ AI networking solutions also connect trained AI models to end users and other systems

I am reminded of the 1980s when Sun [Microsystems] for declaring the network is the computer. Well, 40 years later, we’re seeing the same cycle come true again with the collective nature of AI training models mandating a lossless highly available network to seamlessly connect every AI accelerator in the cluster to one another for peak job completion times. Our AI networks also connect trained models to end users and other multi-tenant systems in the front-end data center, such as storage, enabling the AI system to become more than the sum of its parts.

Arista Networks’ management think that data centers will evolve to be holistic AI centers, where the network will be the epicenter; management thinks that AI centers will need a foundational data architecture; Arista Networks has an AI agent within its EOS (Extensible Operating System) that can connect to NVIDIA’s Bluefield NICs (network interface cards), with more NICs to be added in the future

We believe data centers are evolving to holistic AI centers, where the network is the epicenter of AI management for acceleration of applications, compute, storage and the wide area network. AI centers need a foundational data architecture to deal with the multimodal AI data sets that run on our differentiated EOS network data systems. Arista showcased the technology demonstration of our EOS-based AI agent that can directly connect on the NIC itself or alternatively, inside the host. By connecting into adjacent Arista switches to continuously keep up with the current state, send telemetry or receive configuration updates, we have demonstrated the network working holistically with network interface cards such as NVIDIA Bluefield and we expect to add more NICs in the future.

Arista Networks’ management thinks that as GPUs increase in speed, the dependency on the network for higher throughput increases

I think as the GPUs get faster and faster, obviously, the dependency on the network for higher throughput is clearly related.

The 4 major AI trials Arista Networks discussed in the 2024 Q1 earnings call are all going well and ar removing into pilots these year

[Question] Last quarter, you had mentioned kind of 4 major AI trials that you guys were a part of…  any update on where those 4 AI trials stand or what the current count of AI trials is currently?

[Answer] All 4 trials are largely in what I call Cloud and AI Titans. A couple of them could be classified as specialty providers as well, depending on how they end up. But those 4 are going very well. They started out as largely trials. They’re now moving into pilots this year, most of them. 

Arista Networks has tens of smaller customers who are starting to do AI pilots with the company that typically involve a few hundred GPUs; these customers go to Arista Networks for AI trials because they want best-of-breed reliability and performance

We have tens of smaller customers who are starting to do AI pilots…

…They’re about to build an AI cluster. It’s a reasonably small size, not classified in thousands or 10 thousands. But you’ve got to start somewhere. So they started about a few hundred GPUs, would you say?…

…The AI cloud we talked about, they tend to be smaller, but it’s a representation of the confidence the customer has. They may be using other GPUs, servers, et cetera. But when it comes to the mission critical networks, they’ve recognized the importance of best-of-breed reliability, availability, performance, no loss and the familiarity with the data center is naturally leading to pilots and trials on the AI side with us.

Arista Networks’ management classifies its TAM (total addressable market) within AI as how much of Infiniband will move to Ethernet and it’s far larger than the AI-related revenue of $750 million that management has guided for in 2025

The TAM is far greater than the $750 million we’ve signed up for. And remember, that’s early years. But that can consist of our data center TAM. Our AI TAM, which we count in a more narrow fashion as how much of InfiniBand will move to Ethernet on the back end. We don’t count the AI TAM that’s already in the front end, which is part and parcel of our data center.

Arista Networks’ management continues to see its large customers preferring to spend on AI, but is also seeing classic cloud continue to be an important part of its business and they believe the demand for classic cloud infrastructure will eventually rebound once the AI models are more established

We saw that last year. We saw that there was a lot of pivot going on from the classic cloud, as I like to call it, to the AI in terms of spend. And we continue to see favorable preferences to AI spend in many of our large cloud customers. Having said that, at the same time, simultaneously, we are going through a refresh cycle where many of these customers are moving from 100 to 200 or 200 to 400 gig. So while we think AI will grow faster than cloud, we’re betting on classic cloud continuing to be an important aspect of our contributions…

… I would say there’s such a heavy bias towards — in the Cloud Titans towards training and super training and the bigger and better the GPUs, the billion parameters, the OpenAI, ChatGPT and [indiscernible] that you’re absolutely right that at some level, the classic cloud, what you call traditional, I’m still calling classic, is a little bit neglected last year and this year. Having said that, I think once the training models are established, I believe this will come back, and it will sort of be a vicious cycle that feeds on each other. But at the moment, we’re seeing more activity on the AI and more moderate activity on the cloud.

Arista Networks’ management thinks that as AI networking moves towards Ethernet, it will be difficult to distinguish between front-end and back-end networks

It’s going to become difficult to distinguish the back end from the front end when they all move to Ethernet. For this AI center, as we call it, is going to be a conglomeration of both the front and the back. So if I were to fast forward 3, 4 years from now, I think the AI center is a supercenter of both the front end and the back end. So we’ll be able to track it as long as there’s GPUs and strictly training use cases. But if I were to fast forward, I think there may be many more edge use cases, many more inference use cases and many more small-scale training use cases which will make that distinction difficult to make.

Arista Networks’ management sees NVIDIA more as a friend than a competitor despite NVIDIA trying to compete with the company with the Spectrum-X switches; management rarely sees Spectrum-X as a competing technology in the deals Arista Networks is working on; management feels good about Arista Networks’ win rate

[Question] If you’re seeing Spectrum-X from NVIDIA? And if so, how you’re doing against it?

[Answer] When you say competitive environment, it’s complicated with NVIDIA because we really consider them a friend on the GPUs as well as the mix, so not quite a competitor. But absolutely, we will compete with them on the Spectrum switch. We have not seen the Spectrum except in one customer where it was bundled. But otherwise, we feel pretty good about our win rate and our success for a number of reasons, great software, portfolio of products and architecture that has proven performance, visibility features, management capabilities, high availability. And so I think it’s fair to say that if a customer were bundling with their GPUs, then we wouldn’t see it. If a customer were looking for best of breed, we absolutely see it and win it.

When designing GPU clusters for AI, a network design-centric approach has to be taken

If you look at an AI network design, you can look at it through 2 lenses, just through the compute, in which case you look at scale up and you look at it strictly through how many processes there are. But when we look at an AI network design, it’s a number of GPUs or XTUs per workload. Distribution and location of these GPUs are important. And whether the cluster has multiple tenants and how it’s divvied up between the host, the memory, the storage and the wide area plays a role, and the optimization to make on the applications for the collective communication libraries for specific workloads, levels of resilience, how much redundancy you want to put in, active, link base, load balancing, types of visibility. So the metrics are just getting more and more. There are many more commutations in combination. But it all starts with number of GPUs, performance and billions of parameters. Because the training models are definitely centered around job completion time. But then there’s multiple concentric circles of additional things we have to add to that network design. All this to say, a network design-centric approach has to be taken for these GPU clusters. Otherwise, you end up being very siloed

Arista Networks’ management is seeing huge clusters of GPUs – in the tens of thousands to hundreds of thousands – being deployed in 2025

Let me just remind you of how we are approaching 2024, including Q4, right? Last year, trials. So small, it was not material. This year, we’re definitely going into pilots. Some of the GPUs, and you’ve seen this in public blogs published by some of our customers have already gone from tens of thousands to 24,000 and are heading towards 50,000 GPUs. Next year, I think there will be many of them heading into tens of thousands aiming for 100,000 GPUs. So I see next year as more promising.

ASML (NASDAQ: ASML)

ASML’s management sees no change to the company’s outlook for 2024 from what was mentioned in the 2023 Q4 earnings call and 2024 Q1 earnings call, with AI-related applications still driving demand

Our outlook for the full year 2024 has not changed. We expect a revenue similar to last year. As indicated before, and based on our current guidance, the second half of the year is expected to be significantly higher than the first half. This is in line with the industry’s continued recovery from the downturn. Our guidance on market segments is similar to what we’ve stated in previous quarters…

……We currently see strong developments in AI driving most of the industry recovery and growth, ahead of other end market segments.

ASML’s management sees AI driving the majority of recovery in the semiconductor industry in both Logic and Memory chips; AI’s positive effects on semiconductor industry demand will start showing up in 2025 and management expects that to continue into 2026; Memory chips used in AI require high-bandwidth memory and so have higher density of DRAM; ASML’s management sees other non-AI segments as being behind in terms of recovery, but they do expect recovery eventually

We currently see strong developments in AI driving most of the industry recovery and growth, ahead of other end market segments…

… I think AI is driving, I would say, right now, the biggest part of the recovery. This is true for Logic. This is true for Memory. Roger just commented on Logic. I think you know that for high-bandwidth memory, those products drive more demand, more of a wafer demand because we are looking basically at a higher density of DRAM on those products. And we look at something that, of course, will take course over several months. So we started to see the positive effect of that for 2025. We expect that to continue into 2026, both for Memory and for Logic. And at some point of time, I also mentioned that maybe the other segments are a bit behind in terms of recovery.

So a lot of the capacity today, either Logic or DRAM capacity will be [indiscernible] those AI product. As the other segments recover, we also expect potentially some capacity to be needed there. 

ASML’s management thinks DRAM for AI memory chips will continue to see an increasing use of EUV lithography at each technology node; management also see opportunity for DRAM to use High-NA EUV lithography systems in 2025 or 2026

On DRAM, so I think there also, I think I’ll be very consistent with the information we have shared with you previously. So we see on there an increase of EUV use on every node. I think this is a trend that continue at least in the foreseeable future. Of course, it’s always more difficult to make forecast on nodes or technology that are still being defined by a customer. But that logic is still in place. I think you have seen also in DRAM that at this point of time, all customers are using EUV in production. I think the last customer was very public about that recently.

ASML’s management is not seeing much revenue made on AI at the moment, but it’s still seeing a lot of investment made for AI and these investments require a lot of semiconductor manufacturing capacity

I think what we have seen with AI is a major investment from many companies in supercomputer and the ability basically to train model. What we still miss in AI, I think, is the emergence of end product. So I think today, there’s not much revenue made on AI. There’s just a lot of investment. What we see is that still that investment require a lot of capacity. I think you have seen some of our customers announcing also more capacity to be built before 2028.

Coupang (NYSE: CPNG)

Coupang’s Product Commerce segment had sequential and year-on-year improvement in gross profit in 2024 Q2, driven partly by the use of AI technologies

Product Commerce gross profit increased 26% year-over-year to over $1.9 billion, and a record gross profit margin of 30.3%. This represents a 310 basis points improvement over last year and 200 basis points over last quarter. Our margin improvement this quarter was driven by strong growth rates in categories with higher margin composition, as well as higher efficiencies across operations, including benefits from greater utilization of automation and technology, including AI. We also continue to benefit from further optimization in our supply chain, and the scaling of margin accretive offerings.

Datadog (NASDAQ: DDOG)

Datadog’s management classifies digital natives as SMBs and mid-market companies, and within digital natives, the AI natives are inflecting in usage growth that others are not

I would add that the digital natives are largely SMB and mid-market, they’re not enterprise. And even when you look at the digital native, there’s two stories, depending on whether you talk about the AI natives or the others. The AI natives are inflecting in a way that the others are not at this point. So today, we see this higher growth from AI natives and from traditional enterprises. And stable growth, but not accelerating, from the rest of the pack.  

Datadog’s management has announced general availability of LLM Observability for generative AI for companies to monitor, troubleshoot, and secure LLM (large language model) applications; WHOOP and AppFolio are two early adopters of LLM Observability; it’s still very early days for the LLM Observability product; management thinks a good proxy for the future demand for LLM Observability is the growth of the model providers and the AI-native companies; management expects the LLM market to change a lot over time because it’s still nascent; in order of LLMs to work, they need to be connected to other applications and it’s at that point where management thinks the LLMs need observability; customers that are currently using LLM Observability also use Datadog for the rest of their technology stack and it does not make sense for the customers to operate their LLM applications in isolation

In the next-gen AI space, we announced the general availability of LLM Observability, which application developers and machine learning engineers to efficiently monitor, troubleshoot and secure LLM applications. With LLM Observability, companies can accelerate the deployment of AI applications into production environments and reliably operate and scale them…

… It’s still early. We do see customers that are going increasingly into production, and we have a few of those. I mean, we named a couple as early customers of LLM Observability. I think the two we named were WHOOP, the fitness band; and AppFolio. And we see many more that are lining up and then are going to do that. But in the grand scheme of things, looking at the whole market, it’s still very early. I would say the best proxy you can get from the future demand there is the growth of the model providers and the AI natives because they tend to be the ones that currently are being used to provide AI functionality into other applications and largely in production environment. And so I always said they are the harbinger of what’s to come…

… [Question] When people are thinking about bringing on LLMs into their organization, do they want the observability product in place already? Or are they testing out LLMs and then bringing you on after the fact?

[Answer] We expect this market to change a lot over time because it is far from being mature. And so a lot of the things that might happen today in a certain way might happen 2 years in a very, very different form. That being said, the way it works typically is customers build applications using developer tools, and there’s a whole industry that has emerged around developer tools for — and playgrounds and things like that for LLM. And so they use not one, but 100 different things to do that, which is fairly similar to what you might find on the IDE side or code editor side for the more traditional development, which is lots of different, very fragmented environment on that side. When they start connecting the LLM to the rest of the application, then they start to need like visibility that includes the other components because the LLM doesn’t work in a vacuum, it’s plugged into a front end. It works with authentication and security. It works with — connects to other system databases in other services to get the data. And at that point, they need it to be integrated with the rest of the observability. For the customers that use our LLM Observability product, they use us for the rest — all the rest of their stack. And it would make absolutely no sense for them to operate their LLM in isolation completely separately and not have the visibility across the whole applications. So it’s — at that point, it’s a no-brainer that they need everything to be integrated in production.    

Datadog’s management has expanded Bits AI, Datadog’s AI copilot, with new capabilities, such as the ability to perform autonomous investigations

We also expanded Bits AI with new capabilities. As a reminder, Bits AI is a Datadog built-in AI copilot. In addition to being able to summarize incidents and answer questions, we previewed at DASH, the ability for Bits AI to operate as an agent and perform autonomous investigations. With this capability, this AI proactively surfaces key information and performs complex tasks such as investigating alerts and coordinating — response.

Datadog’s management is hearing from all of Datadog’s customers that they are ramping experiments with AI with the goal of delivering business value with the technology; currently, 2,500 Datadog customers are using one or more of Datadog’s AI integrations for visibility into their use of AI; AI-native customers accounted for 4% of Datadog’s ARR in June 2024 (was 3.5% 2024 Q1); management thinks the percentage of ARR from AI-native customers will lose its relevance over time as AI usage becomes more widespread

Taking a step back and looking at our customer base, we continue to see a lot of excitement around AI technologies. All customers are telling us that they are leveling up on AI and ramping experimentations with the goal of delivering additional business value with AI. And we can see them doing this. Today, about 2,500 customers use one or more of our AI integrations to get visibility into their increasing use of AI. We also continue to grow our business with AI-native customers. which increased to over 4% of our ARR in June. We see this as a sign of the continued expansion of this ecosystem and of the value of using Datadog to monitor the product environment. I will note that over time, we think this metric will become less relevant as AI usage and production broadens beyond this group of customers.

Datadog’s management recently announced Toto, Datadog’s first foundational model for time-series forecasting; Toto delivered state-of-the-art performance on all 11 benchmarks; Toto’s capabilities come from the quality of Datadog’s training dataset; management sees Toto’s existence as evidence of the company’s ability to train, build, and incorporate AI models into its platform

We announced Toto, our first foundational model for time-series forecasting, which delivered state-of-the-art performance on all 11 benchmarks. In addition to the technical innovations devised by our research team, TOTO derives its record performance from the quality of our training dataset and points to our unique ability to train, build and incorporate AI models into a platform that will meaningfully improve operations for our customers.

Datadog’s management continues to believe that digital transformation, cloud migration, and AI adoption are long-term growth drivers of Datadog’s business

Overall, we continue to see no change to the multiyear trend towards digital transformation and cloud migration. We are seeing continued experimentation with new technologies, including next-gen AI, and we believe this is just one of the many factors that will drive greater use of the cloud and next-gen infrastructure.

Datadog’s management thinks the emergence of AI has led to large enterprises realising they need to be on the cloud sooner rather later; management sees a lot of growth in the cloud migration of enterprises as it’s really early in their transition

Some of the strengths we see today has to do with the fact that, to serve their — in part to — the emergence of AI has reaffirmed for them the need to go to the cloud sooner rather than later. So they can build the right kind of applications, they have the right kind of data available to give those applications…

…I’d point you to the numbers we shared, I think, 2 quarters ago in terms of our enterprise penetration and the average size of our contracts with enterprises, which are still fairly small. Like there’s a lot of runway there. And the growth of those accounts is not predicated on the growth of the enterprise themselves. They’re still early in their transformation.

Fiverr (NYSE: FVRR)

Fiverr’s management is deepening the integration of Neo, the company’s AI assistant, into its marketplace experience; management realised that not everyone wants the outright chatbot experience on its marketplace, so Neo only pops up when friction arises to provide guidance for buyers who are navigating Fiverr’s catalogue of talent; management wants Neo to be a personal assistant throughout the Fiverr purchasing experience and also answer buyers’ questions

The second theme of our Summer Product Release is deepening the integration of Neo, Fiverr’s AI tool throughout the market-based experience. As Gen-AI applications quickly shift consumers’ Internet behavior and expectations, we want to stay ahead of the curve to build a more personable experience on Fiverr. At the same time, tests and data in the past 6 months have shown that not everyone prepares the outright chatbot experience when it comes to shopping. So, our strategy for Neo is to incorporate it as an assistance throughout the funnel to help customers when friction arises. For search, Neo provides the guidance you need to navigate Fiverr’s massive catalog of services and talent. And it is trained to understand customers’ past transactions and preference to provide the most relevant recommendations. When it comes to project briefing, having Neo is like having a strategist by your side. It transforms customers’ ideas into a structured brief document that not only looks good, but also delivers better business results. Neo can also help customers write more detailed reviews faster by generating content based on transactions and providing language assistance…

The experimentation that we’ve done with Neo as a personal assistant within the inbox, which is the — which was the first version of doing it, taught us a lot about how our customers are actually using it and how it improves the conversion in briefing. It allows buyers to complete, and it leads to higher conversion as a result. And so, the idea here is that we’re graduating Neo to get out of the inbox and essentially being integrated in all of our experience. Right now, it’s being rolled out gradually because we want to test its accuracy and performance. But essentially, you can fund it as a personal assistant throughout the experience. So, it allows customers to search better, to be more accurate about their needs, and as a result get much higher quality match.

But it also has awareness about where it exists. So, if you’re looking at a specific page, you can ask questions about that page. So, it helps people make decisions and get to what they’re looking for better. The same goes with the integration in briefing. If customers have a brief premade then they can just upload it, and we help make that brief even better. But if they don’t, then the technology that is behind Neo actually helps them write a better, more accurate brief and again, as a result of that, get matched with a much more specific cohort of potential talent that can do the job.

Fiverr’s management continues to believe that AI will be a multiyear tailwind for the company and that AI will have a net positive impact on the company’s business; the deterioration seen in the simple services categories has improved, for whatever reasons (unsure if it’s a one-off event from low base, as management also spoke about the low-base effect); around 20% of Fiverr’s GMV comes from simple jobs 

We are in the early innings of unleashing the full potential of AI in our marketplace, and we believe it will be a multiyear tailwind for us to drive product innovation and growth…

…We also see AI continuing to have a net positive impact on our business. It is important to note that we are starting to see stabilizing and improving trends in simple services…

Now several quarters in, we are actually seeing that in our — we’re seeing this in our data. So, for example, writing and translation as a vertical is the vertical with the biggest exposure to AI impact. In Q2, we’re actually seeing traffic in that vertical improved 10 percentage points in terms of year-over-year growth rate compared to Q1…

That said, with us now opening professions catalog and hourly contracts this will open up new funnels and create growth opportunities, especially for complex services categories. And remember that we have over 700 categories. So, our exposure to specific categories is relatively low and seasonal trends in category spend are a regular thing in our line of business…

…When we think about the overall mix complex is in the mid-30s of GMV and simple is about 20%.

Mastercard (NYSE: MA)

Mastercard’s management intends to further embed AI into Mastercard’s value-added services, particularly in data analytics, fraud, and cybersecurity, because they are seeing companies asking for these solutions; the embedding of AI into the value-added services portfolio does not involve changing the existing portfolio, but augmenting them with a higher weightage to AI

We will also enhance and expand our value-added services, such as in data analytics, fraud and cybersecurity particularly as we further embed AI into our products and services…

…It’s pretty clear that on the services side, as far as the areas of focus are concerned, we continue to be guided by underlying strong secular trends, and one of that is for really any of our corporate partners and B2B partners that they want to make sense of their enterprise data and make better decisions. And how do we do that? We do that by leveraging our artificial intelligence solutions, our set of assistants, a set of fine-tuning, how they could have more personalized suggestions to their end consumers, et cetera, et cetera. That’s one part, help our customers make better decisions, not changing, but very specific solutions with a higher weightage to AI.

And then on the security side and the cybersecurity side, all of this data has to be kept safe. We kept saying that for years. That’s a strong secular trend in itself and making sure that we fine-tune our solutions here. We’ve got to move faster because the bad guys are also moving faster, and they have the similar technology tools at their hand now. So leveraging artificial intelligence, an example I gave last quarter around Decision Intelligence Pro, that’s predicting what is the next card that might be frauded, before it actually happens. Those kind of solutions provide significant lift to our customers in terms of preventing fraud, obviously giving peace of mind to their consumers and overall helping our business, and it’s a close link to our payments — underlying payments business.

Mastercard has been using AI technology successfully for the better part of a decade, in areas such as fraud prevention; management thinks generative AI gives the opportunity for Mastercard to understand more data faster; management has used generative AI to create artificial data sets to train Mastercard’s discriminative AI models; management has also used generative AI to build a new product, such as Decision Intelligence Pro; Decision Intelligence Pro brings a 20% improvement in fraud prediction; management believes that generative AI will increase in penetration within Mastercard’s fraud and cybersecurity products 

 AI isn’t actually anything new for us. So we’ve — for the better part of a decade, we’ve been using AI. This is a discrete machine learning technology to really predict where is the next problem, and analyze data of — that we have and the data that our customers have to prevent fraud. So that’s been very successful.

As far as generative AI is concerned, evolving technology here, there’s obviously an opportunity for us to understand more data in a quicker way. And we have used that initially to train our AI models, our discriminative AI models using generative AI to create artificial data set. So that was the first step. And then we went into putting out a new set of products. I mentioned Decision Intelligence Pro. Decision Intelligence is a product that we’ve had for a long time, machine learning driven that was predicting fraud outcomes and now we’re using more data sets to — that are externally available, stolen card data and so forth, to understand where fraud vulnerabilities might be. The lift is tremendous, 20%, we see in terms of effectiveness out of that product. So we start to see demand for the whole reason on the vulnerabilities that I talked about…

…I believe that the penetration of generative AI and our fraud and cybersecurity product set will only expand. 

Mercado Libre (NASDAQ: MELI)

MercadoLibre has been putting a lot of resources into AI and generative AI; management sees many ways AI can help the commerce business, such as producing better ways for consumers to look at product reviews, enhance product pictures, generate seller-responses when sellers are unable to, and improve the product search experience for consumers; MercadoLibre has 16,000 developers and they are using AI to improve productivity; MercadoLibre is using AI inc customer support to respond more cost-effectively and more accurately

We have been — put a lot of resources into AI and GenAI throughout the company, really. We don’t have a centralized department of AI, but all of our different business units…

… On the commerce side, obviously, we are using AI to help us with recommendations, as you mentioned, but more important than that on reviews, for instance, that in the past, you have to — if you were to review a product, you have to go through many different views, now we can consolidate that into a more efficient way of communicating the qualities, the prospects of a particular product pictures, as you know, our pictures that publish might not be the quality that we are expecting from our merchants, and we can improve those with answers from sellers is another good example in the past, if you were to buy something at 2 AM in the morning, you’ll have to wait until the next day to get an answer that obviously affected significantly the conversion of the product. Now we can respond right away with using GenAI models…

…On the developer side, we have 16,000 developers, which are also using AI tools to improve productivity and that also generating some improvements and efficiencies in the way we deploy products throughout the company. And I think 1 of the most important projects that we have is on CX, customer experience and customer support by which we are also applying AI tools that will help us to not only respond more efficiently in terms of cost, but also be more accurate in terms of the way we manage those issues. These are some examples, but there are many others…

… You asked about search and where you’re seeing technical and bedding to power search that technical — turn search into something more semantic. So it’s easier to try to send the users to what they’re looking for.

Meta Platforms (NASDAQ: META)

Meta’s AI work continues to improve quality of recommendations on Facebook and Instagram, and drives engagement; the more general recommendation models Meta develops, the better the content recommendations get; Meta rolled out a unified video recommendation service across Facebook in 2024 Q2 for Reels, longer videos, and Live; Meta’s unified AI systems had already increased engagement on Facebook Reels more than Meta’s shift from using CPUs to GPUs; management wants to eventually have a single, unified AI recommendation system for all kinds of content across Meta’s social apps; the unified video recommendation service has encouraging early results, and management expects the relevance of video recommendations to increase

Across Facebook and Instagram, advances in AI continue to improve the quality of recommendations and drive engagement. And we keep finding that as we develop more general recommendation models, content recommendations get better. This quarter we rolled out our full-screen video player and unified video recommendation service across Facebook — bringing Reels, longer videos, and Live into a single experience. This has allowed us to extend our unified AI systems, which had already increased engagement on Facebook Reels more than our initial move from CPUs to GPUs did. Over time, I’d like to see us move towards a single, unified recommendation system that powers all of the content including things like People You May Know across all of our surfaces. We’re not there, so there’s still upside — and we’re making good progress here…

…On Facebook, we are seeing encouraging early results from the global roll-out of our unified video player and ranking systems in June. This initiative allows us to bring all video types on Facebook into one viewing experience, which we expect will unlock additional growth opportunities for short-form video as we increasingly mix shorter videos into the overall base of Facebook video engagement. We expect the relevance of video recommendations will continue to increase as we benefit from unifying video ranking across Facebook and integrating our next generation recommendation systems. These have already shown promising gains since we began using the new systems to support Facebook Reels recommendations last year. We expect to expand these new systems to support more surfaces beyond Facebook video over the course of this year and next year

In the past, advertisers would tell Meta the specific audience they wanted to reach, but over time, Meta could predict the interested-audience better than the advertisers could, even though the advertisers still needed to come up with collateral; management thinks that AI will generate personalised collateral for advertisers in the coming years and all the advertiser needs to do is to tell Meta a business objective and a budget, and Meta will handle everything else; Meta’s first generative AI ad features, such as image expansion and text generation, were used by more than 1 million advertisers in June 2024; Meta rolled out full image generation capabilities in Advantage+ in May 2024

It used to be that advertisers came to us with a specific audience they wanted to reach — like a certain age group, geography, or interests. Eventually we got to the point where our ads system could better predict who would be interested than the advertisers could themselves. But today advertisers still need to develop creative themselves. In the coming years, AI will be able to generate creative for advertisers as well — and will also be able to personalize it as people see it. Over the long term, advertisers will basically just be able to tell us a business objective and a budget, and we’re going to go do the rest for them. We’re going to get there incrementally over time, but I think this is going to be a very big deal…

…We’ve seen promising early results since introducing our first generative AI ad features – image expansion, background generation, and text generation – with more than one million advertisers using at least one of these solutions in the past month. In May, we began rolling out full image generation capabilities into Advantage+ creative, and we’re already seeing improved performance from advertisers using the tool. 

Meta’s management thinks that Meta AI, the company’s AI assistant feature, will be the most used AI assistant by end-2024; Meta AI is improving in intelligence and features quickly, and seems on track to be an important service; Meta AI’s current use cases include searching for information, role-playing difficult conversations, and creating images, but new use cases are likely to emerge; Meta AI has been used for billions of queries thus far; Meta AI has helped with WhatsApp retention and engagement; India has become the largest market for Meta AI; Meta AI is now available in 20 countries and 8 languages; management thinks that people who bet on the early indicators of Meta tend to do pretty well, and Meta AI is one of those early indicators that are signalling well; management wants to build a lot more functionality into Meta AI, but that will take a few years

Last quarter we started broadly rolling out our assistant, Meta AI, and it is on track to achieve our goal of becoming the most used AI assistant by the end of the year. We have an exciting roadmap ahead of things that we want to add, but the bottom line here is that Meta AI feels on track to be an important service and it’s improving quickly both in intelligence and features. Some of the use cases are utilitarian, like searching for information or role-playing difficult conversations before you have them with another person, and other uses are more creative, like the new Imagine Yourself feature that lets you create images of yourself doing whatever you want in whatever style you want. And part of the beauty of AI is that it’s general, so we’re still uncovering the wide range of use cases that it’s valuable for…

…People have used Meta AI for billions of queries since we first introduced it. We’re seeing particularly promising signs on WhatsApp in terms of retention and engagement, which has coincided with India becoming our largest market for Meta AI usage. You can now use Meta AI in over 20 countries and eight languages, and in the US we’re rolling out new features like Imagine edit, which allows people to edit images they generate with Meta AI…

… I think that the people who bet on those early indicators tend to do pretty well, which is why I wanted to share in my comments the early indicator that we had on Meta AI, which is, I mean, look, it’s early…

…I was talking before about we have the initial usage trends around Meta AI but there’s a lot more that we want to add, things like commerce and you can just go vertical by vertical and build out specific functionality to make it useful in all these different areas are eventually, I think, what we’re going to need to do to make this just as — to fulfill the potential around just being the ideal AI assistant for people. So it’s a long road map. I don’t think that this stuff is going to get finished in the next couple of quarters or anything like that. But this is part of what’s going to happen over the next few years as we build something that will, I think, just be a very widely used service. So I’m quite excited about that.

Meta’s management recently launched AI Studio, which allows anyone to create AIs that people can interact with; AI Studio is useful for creators who want to engage more with their communities, but can also be useful for anyone who wants to build their own AI agents, including businesses; management thinks every business in the future will have its own AI agent for customer interactions that drives sales and reduces costs; management expects Business AI agents to dramatically accelerate Meta’s business messaging revenue when the feature reaches scale

This week we launched AI Studio, which lets anyone create AIs to interact with across our apps. I think that creators are especially going to find this quite valuable. There are millions of creators across our apps — and these are people who want to engage more with their communities and their communities want to engage more with them — but there are only so many hours in the day. So now they’re going to be able to use AI Studio to create AI agents that can channel them to chat with their community, answer people’s questions, create content, and more. So I’m quite excited about this. But this goes beyond creators too. Anyone is going to be able to build their own AIs based on their interests or different topics that they are going to be able to engage with or share with their friends.

Business AIs are the other big piece here. We’re still in alpha testing with more and more businesses. The feedback we’re getting is positive so far. Over time I think that just like every business has a website, social media presence, and an email address, in the future I think that every business is also going to have an AI agent that their customers can interact with. Our goal is to make it easy for every small business, and eventually every business, to pull all their content and catalog into an AI agent that drives sales and saves them money. When this is working at scale, I expect it to dramatically accelerate our business messaging revenue.

The Llama family of foundation models is the engine that powers all of Meta’s AI-related work; in 2024 Q2, Meta released Llama 3.1, the first frontier-level open source model, and other new and industry-leading small and medium models; the Llama 3.1 405B model has better cost performance compared to leading closed models; management thinks Llama 3.1 will mark an inflection point for open source AI becoming the industry standard; Meta is already working on Llama 4 and management is aiming for it to be the most advanced foundation AI model when released in 2025; the Llama models are well-supported by the entire cloud computing ecosystem

The engine that powers all these new experiences is the Llama family of foundation models. This quarter we released Llama 3.1, which includes the first frontier-level open source model, as well as new and industry-leading small and medium-sized models. The 405B model has better cost performance relative to the leading closed models, and because it’s open, it is immediately the best choice for fine-tuning and distilling your own custom models of whatever size you need. I think we’re going to look back at Llama 3.1 as an inflection point in the industry where open source AI started to become the industry standard, just like Linux is…

…We’re already starting to work on Llama 4, which we’re aiming to be the most advanced in the industry next year…

… Part of what we’re doing is working closely with AWS, I think, especially did great work for this release. Other companies like Databricks, NVIDIA, of course, other big players like Microsoft with Azure, and Google Cloud, they’re all supporting this. And we want developers to be able to get it anywhere. I think that’s one of the advantages of an open source model like Llama is — it’s not like you’re locked into 1 cloud that offers that model, whether it’s Microsoft with OpenAI or Google with Gemini or whatever it is, you can take this and use it everywhere and we want to encourage that. So I’m quite excited about that.

Meta’s management is planning for the AI compute needs of the company for the next several years; management thinks the compute requirements for training Llama 4 will likely be 10x that of Llama 3, and future models will require even more; given long lead times to build compute capacity, management would rather risk overbuilding than being too late in realising there’s a shortfall; even as Meta builds compute capacity, management still remains focused on cost efficiency

We’re planning for the compute clusters and data we’ll need for the next several years. The amount of compute needed to train Llama 4 will likely be almost 10x more than what we used to train Llama 3 — and future models will continue to grow beyond that. It’s hard to predict how this will trend multiple generations out into the future, but at this point I’d rather risk building capacity before it is needed, rather than too late, given the long lead times for spinning up new infra projects. And as we scale these investments, we’re of course going to remain committed to operational efficiency across the company…

A few years ago, management thought holographic AR (augmented reality) technology would be ready before smart AI, but the reverse has happened; regardless, Meta is still well positioned for this reverse order; because of AI, Meta’s smart glasses continue to be a bigger hit than management expected and supply cannot keep up with demand; Meta will continue to partner EssilorLuxottica for the long term to build its smart glasses

A few years ago I would have predicted that holographic AR would be possible before smart AI, but now it looks like those technologies will actually be ready in the opposite order. We’re well-positioned for that because of the Reality Labs investments that we’ve already made. Ray-Ban Meta glasses continue to be a bigger hit sooner than we expected — thanks in part to AI. Demand is still outpacing our ability to build them, but I’m hopeful we’ll be able to meet demand soon. EssilorLuxottica has been a great partner to work with on this, and we’re excited to team up with them to build future generations of AI glasses as we continue to build our long term partnership.

AI is playing an increasingly important role in improving Meta’s marketing performance; the AI-powered Meta Lattice ad ranking architecture continued to drive ad performance and efficiency gains in 2024 Q2; Advantage+ Shopping campaigns are driving 22% higher return on ad spend for US advertisers; advertiser adoption of Meta’s advertising automation tools continue to expand; Meta has continued to increase the capabilities of Advantage+, such as expanding conversion types, and helping advertisers automatically select which ad format to serve after they upload multiple images and videos; Meta rolled out full image generation capabilities in Advantage+ in May 2024

The second part of improving monetization efficiency is enhancing marketing performance. We continue to be pleased with progress here, with AI playing an increasingly central role. We’re improving ad delivery by adopting more sophisticated modeling techniques made possible by AI advancements, including our Meta Lattice ad ranking architecture, which continued to provide ad performance and efficiency gains in the second quarter. We’re also making it easier for advertisers to maximize ad performance and automate more of their campaign set up with our Advantage+ suite of solutions. We’re seeing these tools continue to unlock performance gains, with a study conducted this year demonstrating 22% higher return on ad spend for US advertisers after they adopted Advantage+ Shopping campaigns. Advertiser adoption of these tools continues to expand, and we’re adding new capabilities to make them even more useful. For example, this quarter we introduced Flexible Format to Advantage+ Shopping, which allows advertisers to upload multiple images and videos in a single ad that we can select from and automatically determine which format to serve, in order to yield the best performance. We have also now expanded the list of conversions that businesses can optimize for using Advantage+ Shopping to include an additional 10 conversion types, including objectives like “add to cart”…

…In May, we began rolling out full image generation capabilities into Advantage+ creative, and we’re already seeing improved performance from advertisers using the tool. 

Monetisation for Meta’s AI products such as Meta AI or AI Studio will take years because management is following the same playbook they have had for years, which is to start a product, then take time to scale the product to a billion users before monetising; Meta’s management is a little different from other companies in terms of how they think about the time needed to monetise products

We have a relatively long business cycle of starting a new product, scaling it to something that reaches 1 billion people or more and only then really focusing on monetizing at scale. So realistically, for things like Meta AI or AI Studio, I mean, these are things that I think will increase engagement in our products and have other benefits that will improve the business and engagement in the near term. But before we’re really talking about monetization of any of those things by themselves, I mean, I don’t think that anyone should be surprised that I would expect that, that will be years, right?…

…And I think that, that’s something that is a little bit different about Meta in the way we build consumer products and the business around them than a lot of other companies that ship something and start selling it and making revenue from it immediately. So I think that’s something that our investors and folks thinking about analyzing the business, if needed, to always grapple with is all these new products, we ship them and then there’s a multiyear time horizon between scaling them and then scaling them into not just consumer experiences but very large businesses.

Meta’s ongoing capex investments in AI infrastructure is informed by the strong returns management has seen and expect to achieve in the future; management expects the returns from generative AI to take some time to appear, but they see signification monetisation opportunities that could be unlocked through the AI investments; Meta’s capital expenditures for AI infrastructure are done with flexibility in mind so that AI training capacity can also be redirected to generative AI inference and its ranking and recommendation systems, if needed; management is focused on improving cost efficiency of its AI workloads over time; Meta’s AI capex come in 2 buckets, core AI and generative AI (genAI), which are built to be fungible if needed; the core AI bucket is much more mature in driving revenue for Meta and management takes an ROI (return on investment) approach; the gen AI bucket is much earlier in revenue-generation-maturity but is expected to open up new revenue opportunities over time to deliver that ROI; it’s difficult for management to plan for Meta’s long-term capex trajectory

Our ongoing investment in core AI capacity is informed by the strong returns we’ve seen, and expect to deliver in the future, as we advance the relevance of recommended content and ads on our platforms. While we expect the returns from generative AI to come in over a longer period of time, we are mapping these investments against the significant monetization opportunities that we expect to be unlocked across customized ad creative, business messaging, a leading AI assistant, and organic content generation. As we scale generative AI training capacity to advance our foundation models, we will continue to build our infrastructure in a way that provides us with flexibility in how we use it over time. This will allow us to direct training capacity to gen AI inference, or to our core ranking and recommendation work when we expect that doing so would be more valuable. We will also continue our focus on improving the cost efficiency of our workloads over time…

… I would broadly characterize our AI investments into 2 buckets: core AI and gen AI. And the 2 are really at different stages as it relates to driving revenue for our businesses and our ability to measure returns. On our core AI work, we continue to take a very ROI-based approach to our investment here. We’re still seeing strong returns as improvements to both engagement and ad performance have translated into revenue gains, and it makes sense for us to continue investing here. Gen AI is where we’re much earlier, as Mark just mentioned in his comments. We don’t expect our gen AI products to be a meaningful driver of revenue in ’24. But we do expect that they’re going to open up new revenue opportunities over time that will enable us to generate a solid return off of our investment while we’re also open sourcing subsequent generations of Llama. And we’ve talked about the 4 primary areas that we’re focused here on the gen AI opportunities to enhance the core ads business, to help us grow in business messaging, the opportunities around Meta AI, and the opportunities to grow core engagement over time.

The other thing I would say is, we’re continuing to build our AI infrastructure with fungibility in mind so that we can flex capacity where we think it will be put to best use. The infrastructure that we build for gen AI training can also be used for gen AI inference. We can also use it for ranking and recommendations by making certain modifications like adding general compute and storage. And we’re also employing a strategy of staging our data center sites at various phases of development, which allows us to flex up to meet more demand and less lead time if needed while limiting how much spend we’re committing to in the outer years…

…We haven’t really shared an outlook sort of on the longer-term CapEx trajectory. In part, infrastructure is an extraordinarily dynamic planning area for us right now. We’re continuing to work through what the scope of the gen AI road maps will look like over that time. Our expectation, obviously again, is that we are going to significantly increase our investments in AI infrastructure next year, and we’ll give further guidance as appropriate. But we are building all of that CapEx, again with the factors in mind that I talked about previously, thinking about both how to build it flexibly so we can deploy to core AI and gen AI use cases as needed…

… There’s sort of a whole host of use cases for the life of any individual data center ranging from gen AI training at its outset to potentially supporting gen AI inference to being used for core ads and content ranking and recommendation and also thinking through the implications, too, of what kinds of servers we might use to support those different types of use cases.

Microsoft (NASDAQ: MSFT)

Microsoft’s management sees the AI platform shift as involving both knowledge and capital-intensive investments, similar to the Cloud platform shift; as Microsoft goes through the AI platform shift, management is focused on product innovation, and using customer demand signals and time to value to manage the cost structure dynamically

 I want to offer some broader perspective on the AI platform shift. Similar to the Cloud, this transition involves both knowledge and capital-intensive investments. And as we go through this shift, we are focused on 2 fundamental things. First, driving innovation across a product portfolio that spans infrastructure and applications, so as to ensure that we are maximizing our opportunity while in parallel, continuing to scale our cloud business and prioritizing fundamentals, starting with security. Second, using customer demand signal and time to value to manage our cost structure dynamically and generate durable long-term operating leverage.

Azure’s share gains accelerated in FY2024 (fiscal year ended 30 June 2024), driven by AI; Azure grew revenue by 29% in 2024 Q2 (was 31% in 2024 Q1), with 8 points of growth from AI services (was 7 points in 2024 Q1); Azure’s AI business has higher demand than available capacity; 50% of Azure AI users are also using a data meter within Azure, which is excellent for Azure

Starting with Azure. Our share gains accelerated this year driven by AI…

…Azure and other cloud services revenue grew 29% and 30% in constant currency, in line with expectations and consistent with Q3 when adjusting for leap year. Azure growth included 8 points from AI services, where demand remained higher than our available capacity…

…AI doesn’t sit on its own, right? So it’s just for — we have a concept of design wins in Azure. So in fact, 50% of the folks who are using Azure AI are also using a data meter. That’s very exciting to us because the most important thing in Azure is to win workloads in the enterprise. And that is starting to happen. And these are generational things once they get going with you. So that’s, I think, how we think about it at least when I look at what’s happening on our demand side. 

Azure added new AI accelerators from both AMD and NVIDIA, and its own in-house Azure Maia chips; Azure also introduced its own Cobalt 100 CPUs

We added new AI accelerators from AMD and NVIDIA as well as our own first-party silicon Azure Maia and we introduced new Cobalt 100, which provides best-in-class performance for customers like Elastic, MongoDB, Siemens, Snowflake and Teradata.

Azure AI offers the most diverse selection of models for customers; Azure AI now has 60,000 customers and average spend per customer continues to grow; Azure OpenAI started to provide access to GPT-4o and GPT-4o Mini in 2024 Q2; Azure OpenAI is being used by companies from diverse industries; Phi-3 within Azure AI offers small language models that are already being used by a wide range of companies; Models as a Service within Azure AI offers access to third-party models including open-sourced models and it is being used by a diverse range of large companies; paid Models as a Service customers doubled sequentially

With Azure AI, we are building out the app server for the AI wave providing access to the most diverse selection of models to meet customers’ unique cost, latency and design considerations. All up, we now have over 60,000 Azure AI customers up nearly 60% year-over-year and average spend per customer continues to grow.  Azure OpenAI service provides access to best-in-class frontier models, including as of this quarter GPT-4o and GPT-4o mini. It’s being used by leading companies in every industry, including H&R Block, Suzuki, Swiss Re, Telstra as well as digital natives like Freshworks, Meesho and Zomato. With Phi-3, we offer a family of powerful small language models, which are being used by companies like BlackRock, Emirates, Epic, ITC, Navy Federal Credit Union and others. And with Models as a Service, we provide API access to third-party models, including as of last week, the latest from Cohere, Meta and Mistral. The number of paid Models as a Service customers more than doubled quarter-over-quarter, and we are seeing increased usage by leaders in every industry from Adobe and Bridgestone to Novo Nordisk and Palantir.

Microsoft Fabric, an AI-powered data platform, now has more than 14,000 customers (was more than 11,000 in 2024 Q1)

Microsoft Fabric, our AI-powered next-generation data platform, now has over 14,000 paid customers, including leaders in every industry from Accenture and Kroger to Rockwell Automation and Zeiss, up 20% quarter-over-quarter. And this quarter, we introduced new first of their kind, real-time intelligence capabilities in Fabric, so customers can unlock insights on high-volume, time-sensitive data.

GitHub Copilot is the most widely adopted AI-powered developer tool; 77,000 organisations have adopted GitHub Copilot in just over 2 years since its general availability and the number of organisations is up 180% from a year ago; GitHub Copilot is driving GitHub’s overall growth; GitHub’s annual revenue run rate is $2 billion and Copilot accounted for more than 40% of GitHub’s revenue growth in FY2024; GitHub Copilot alone is already a larger business than the entire GitHub when Microsoft acquired it in 2018

GitHub Copilot is by far the most widely adopted AI power developer tool. Just over 2 years since its general availability, more than 77,000 organizations from BBVA, FedEx and H&M to Infosys and Paytm have adopted Copilot up 180% year-over-year…

…Copilot is driving GitHub growth all up. GitHub annual revenue run rate is now $2 billion. Copilot accounted for over 40% of GitHub revenue growth this year and is already a larger business than all of GitHub was when we acquired it.

More than 480,000 organisations have used AI-features within Microsoft’s Power Platform (was more than 330,000 in 2024 Q1), and Power Platform has 48 million monthly active users (was 25 million in 2024 Q1)

We are also integrating generative AI across Power Platform, enabling anyone to use natural language to create apps, automate workflows or build a website. To date, over 480,000 organizations have used AI-powered capabilities in Power Platform, up 45% quarter-over-quarter. In total, we now have 48 million monthly active users of Power Platform, up 40% year-over-year.

The number of Copilot for Microsoft 365 users doubled sequentially; Copilot for Microsoft 365 customers increased 60% sequentially; number of customers for Copilot for Microsoft 365 with more than 10,000 seats doubled sequentially; Copilot Studio customers can build custom Copilots for agentic work; 50,000 organisations have used Copilot Studio

Copilot for Microsoft 365 is becoming a daily habit for knowledge workers as it transforms work, workflow and work artifacts. The number of people who use Copilot daily at work nearly doubled quarter-over-quarter as they use it to complete tasks faster, hold more effective meetings and automate business workflows and processes. Copilot customers increased more than 60% quarter-over-quarter. Feedback has been positive with majority of enterprise customers coming back to purchase more seats, all up the number of customers with more than 10,000 seats more than doubled quarter-over-quarter, including Capital Group, Disney, Dow, Kyndryl, Novartis, and EY alone will deploy Copilot to 150,000 of its employees and we are going further adding agent capabilities to Copilot. New Team Copilot can facilitate meetings and create an assigned task. And with Copilot Studio customers can extend Copilot for Microsoft 365 and build custom Copilots that proactively respond to data and events using their own first and third-party business data. To date, 50,000 organizations from Carnival Corporation, Cognizant and Eaton to KPMG, Majesco and McKinsey have used Copilot Studio, up over 70% quarter-over-quarter.

DAX Copilot has been purchased by more than 400 healthcare organisations to-date, up 40% sequentially; the number of AI-generated clinical reports have tripled

With DAX Copilot, more than 400 health care organizations, including Community Health Network, Intermountain, Northwestern Memorial Healthcare and Ohio State University Wexner Medical Center have purchased DAX Copilot to date, up 40% quarter-over-quarter and the number of AI-generated clinical reports more than tripled.

Microsoft introduced a new category of Copilot+ PCs in 2024 Q2; the Copilot+ PCs have a new system architecture design to deliver breakthrough AI experiences; early reviews are promising

When it comes to devices, we introduced our new category of Copilot+ PCs this quarter. They are the fastest, most intelligent Windows PCs ever. They include a new system architecture designed to deliver best-in-class performance and breakthrough AI experiences. We are delighted by early reviews, and we are looking forward to the introduction of more Copilot+ PCs powered by all of our silicon and OEM partners in the coming months.

More than 1,000 paid customers used Copilot for security ; Microsoft now has 1.2 million security customers and over 800,000 of them use 4 or more workloads, up 25% from a year ago

Over 1,000 paid customers used Copilot for security, including Alaska Airlines, Oregon State University, Petrofac, Wipro, WTW, and we are also securing customers’ AI deployments with updates to Defender and Purview. All up, we now have 1.2 million security customers over 800,000, including Dell Technologies, Deutsche Telekom, TomTom use 4 or more workloads, up 25% year-over-year. 

Combined revenue of Bing, Edge, and Copilot was up 19% year-on-year and management said Bing and Edge took share; management is applying generative AI to Bing to test a new generative search experience, whose aim is to create dynamic responses while still driving clicks to publishers

We are ensuring that Bing, Edge and Copilot collectively are driving more engagement and value to end users, publishers and advertisers. Our overall revenue ex-TAC increased 19% year-over-year and we again took share across Bing and Edge. We continue to apply Generative AI to pioneer new approaches to how people search and browse. Just last week, we announced we are testing a new generative search experience, which creates a dynamic response to users’ query while maintaining click share to publishers. 

Copilot for the web has created more than 12 billion images and did more than 13 billion chats to-date, up 150% since the start of 2024

We continue to drive record engagement with Copilot for the web, consumers have used Copilot to create over 12 billion images and conduct 13 billion chats to date, up 150% since the start of the calendar year.

Microsoft is using AI in its Performance Max advertising tool to create and optimise ads for advertisers, increasing their advertising ROI (return on investment)

We are helping advertisers increase their ROI, too. We have seen positive response to Performance Max, which uses AI to dynamically create and optimize ads and Copilot and Microsoft ad platform helps marketers create campaigns and troubleshoot using natural language.

Microsoft’s capex in 2024 Q2 (FY2024 Q4) and the whole of FY2024 are basically for AI and cloud, and it can be split roughly 50-50 into (1) data centers and (2) servers consisting of GPU/CPUs; management sees the capex for the data centers as providing support for monetisation over the next 15-plus years; the capex for GPUs and CPUs are driven by demand signals; the demand signals that management is seeing include Microsoft 365 Copilot demand, GitHub Copilot demand, and Azure AI growth; Microsoft can be spending on the data centres first, because they have long lead times, without spending on the GPUs and CPUs if the demand signals no longer persist, moreover, revenue growth will not be affected by the throttling of GPU/CPU spending; part of the capex is for AI training, but management will be scaling training only if they see demand; the capex on the data centres itself is really flexible because Microsoft has built a consistent architecture for its technological infrastructure

Capital expenditures, including finance leases, were $19 billion, in line with expectations and cash paid for PP&E was $13.9 billion. Cloud and AI-related spend represents nearly all of our total capital expenditures. Within that, roughly half is for infrastructure needs where we continue to build and lease data centers that will support monetization over the next 15 years and beyond. The remaining Cloud and AI-related spend is primarily for servers, both CPUs and GPUs to serve customers based on demand signals. For the full fiscal year, the mix of our Cloud and AI-related spend was similar to Q4…

…So when I think about what’s happening with M365 Copilot as perhaps the best Office 365 or M365 suite we have had, the fact that we’re getting recurring customers, so our customers coming back buying more seats. So GitHub Copilot now being bigger than even GitHub when we bought it. What’s happening in the contact center with Dynamics. So I would say — and obviously, the Azure AI growth, that’s the first place we look at. That then drives bulk of the CapEx spend, basically, that’s the demand signal because you got to remember, even in the capital spend, there is land and there is data center build, but 60-plus percent is the kit, that only will be bought for inferencing and everything else if there is demand signal, right? So that’s, I think, the key way to think about capital cycle even. The asset, as Amy said, is a long-term asset, which is land and the data center, which, by the way, we don’t even construct things fully, we can even have things which are semi-constructive, we call Kohl’s shelves and so on. So we know how to manage our CapEx spend to build out a long-term asset and a lot of the hydration of the kit happens when we have the demand signal. 

There is definitely spend for training. Even there, of course, we will only be scaling training as we see the demand accrue in any given period in time…

…Being able to maybe share a little more about that when we talked about roughly half of FY ’24’s total capital expense as well as half of Q4’s expense, it’s really on land and build and finance leases, and those things really will be monetized over 15 years and beyond. And they’re incredibly flexible because we’ve built a consistent architecture, first with the Commercial Cloud and second with the Azure Stack for AI, regardless of whether the demand is at the platform layer or at the app layer or through third parties and partners or, frankly, our first-party SaaS, it uses the same infrastructure. So it’s a long-lived flexible assets…

…Could we see sort of consistent revenue growth without maybe what you would say is more of this sort of elevated capital expense number or something that continues to accelerate. And the answer to that is yes because there’s 2 different pieces, right? You’re seeing half of this go toward long-term builds that Satya mentioned, the pace at which we fill those builds with CPUs or GPUs will be demand-driven. And so if we see differences in demand signal, we can throttle that investment on the CPU side, which we’ve done for I guess, a long time at this point, as I reflect, and we’ll use all that same learning and demand signal understanding to do the same thing on the GPU side. And so you’re right that you could see relatively consistent revenue patterns and yet see these inconsistencies and capital spend quarter-to-quarter…

…We think about it in terms of what’s the total percentage of cost that goes into each line item, land which obviously has a very different duration and a very different lead time. So those are the other 2 considerations. We think about lead time and duration of the asset. Land, network, construction, the system or the kit and then the ongoing cost. And so if you think about it that way, then you know how to even adjust, if you will, the capital spend based on demand signal.

For Azure’s expected growth of 28%-29% in 2024 Q3 (FY2025 Q1), management expects consumption trends from 2024 Q2 (FY2024 Q4) to continue through FY2025 H1 and the consumption trends include capacity-constrained AI-demand as well as non-AI growth; management expects Azure’s growth to accelerate in FY2025 H2, driven by increase in AI capacity to meet growing demand

 In Azure, we expect Q1 revenue growth to be 28% to 29% in constant currency. Growth will continue to be driven by our consumption business, inclusive of AI, which is growing faster than total Azure. We expect the consumption trends from Q4 to continue through the first half of the year. This includes both AI demand impacted by capacity constraints and non-AI growth trends similar to June. Growth in our per user business will continue to moderate. And in H2, we expect Azure growth to accelerate as our capital investments create an increase in available AI capacity to serve more of the growing demand…

… Capacity constraints, particularly on AI and Azure will remain in Q4 and will remain in H1. 

When Microsoft transitioned to the cloud (in the late 2000s and early 2010s), it was rolled out geography by geography, whereas this current AI platform shift is done globally straight away; Microsoft’s consistent technological infrastructure helps its current AI platform shift achieve faster margin improvement compared to the shift to cloud

You can see what we’re doing and focused on is building out this network in parallel across the globe. Because when we did this last transition, the first transition to the Cloud, which seems a long time ago sometimes. It rolled out quite differently. We rolled out more geo by geo and this one because we have demand on a global basis, we are doing it on a global basis, which is important. We have large customers in every geo… 

…[Question] With Cloud, it took time for margins to improve. It looks like with AI, it’s happening quicker. Can you give us a sense of how you think about the margin impact near term and long term from all the investment on AI?

[Answer] To answer the second half of your question on margin improvement, looking different than it did through the last cloud cycle. That’s primarily for a reason I’ve mentioned a couple of times. We have a consistent platform. So — because we’re building to on Azure AI stack, we don’t have to have multiple infrastructure investments. We’re making one. We’re using that internally first party, and that’s what we’re using with customers to build on as well as ISVs. So it does, in fact, make margins start off better and obviously scale consistently.

Management sees generative AI as fundamentally just being software, and it is translating into growth for Microsoft’s SaaS (software-as-a-service) products; management sees the growth in the usage of Microsoft’s software products as a healthy sign of AI adoption

[Question] How should we think about what it’s going to take for GenAI to become more real across the industry and for it to become more visible within your SaaS offerings?

[Answer] At the end of the day, GenAI is just software. So it is really translating into fundamentally growth on what has been our M365 SaaS offering with a newer offering that is the Copilot SaaS offering, which today is on a growth rate that’s faster than any other previous generation of software we launched as a suite in M365. That’s, I think, the best way to describe it. I mean the numbers I think we shared even this quarter are indicative of this, Mark. So if you look at it, we have both the landing of the seats itself quarter-over-quarter that is growing 60%, right? That’s a pretty good healthy sign. The most healthy sign for me is the fact that customers are coming back there. That is the same customers with whom we landed the seats coming back and buying more seats. And then the number of customers with 10,000-plus seats doubled, right? It’s 2x quarter-over-quarter. That, to me, is a healthy SaaS core business.

Microsoft has dealt with AI capacity constraints by working with third parties who are happy to help Microsoft extend the Azure platform

We’ve talked about now for quite a few quarters, we are constrained on AI capacity. And because of that, actually, we’ve, to your point, have signed up with third parties to help us as we are behind with some leases on AI capacity. We’ve done that with partners who are happy to help us extend the Azure platform, to be able to serve this Azure AI demand. 

Netflix (NASDAQ: NFLX)

Netflix has been using AI (artificial intelligence) and ML (machine learning) for many years to improve the content discovery experience and drive more engagement, and management thinks GenAI (generative AI) has great potential to improve these efforts; but it’s also important ultimately for Netflix to have great content

 We’ve been using similar technologies, AI and ML, for many years to improve the discovery experience and drive more engagement through those improvements. We think that generative AI has tremendous potential to improve our recommendations and discovery systems even further. We want to make it even easier for people to find an amazing story that’s just perfect for them in that moment. But I think it’s also worth noting that the key to our success stacks, right, it’s quality at all levels. So it’s great movies, it’s great TV shows, it’s great games, it’s great live events, and a great and constantly improving recommendation system that helps unlock all of that value for all of those stories.

Management is unsure how AI will specifically impact content creation, but they think AI will result in a great set of creator tools, as there has been a long history of technology improving the content creation process; management thinks that when it comes to content creation, great story-telling is still the most important thing, even as content creators experiment with AI

But I think it’s also worth noting that the key to our success stacks, right, it’s quality at all levels. So it’s great movies, it’s great TV shows, it’s great games, it’s great live events, and a great and constantly improving recommendation system that helps unlock all of that value for all of those stories. nd one thing that’s sure, if you look back over 100 years of entertainment, you can see how great technology and great entertainment work hand in hand to build great, big businesses. You can look no further than animation. Animation didn’t get cheaper, it got better in the move from hand-drawn to CG animation. And more people work in animation today than ever in history. So I’m pretty sure that there’s a better business and a bigger business in making content 10% better than it is making it 50% cheaper…

…I think that shows and movies, they win with the audience when they connect. It’s in the beauty of the writing. It’s in the chemistry of the actors. It’s in the plot, the surprise and the plot twist, all those things…

….So my point is they’re looking to connect. So we have to focus on the quality of the storytelling. There’s a lot of filmmakers and a lot of producers experimenting with AI today. They’re super excited about how useful a tool it can be. And we got to see how that develops before we can make any meaningful predictions on what it means for anybody. But our goal remains unchanged, which is telling great stories.

Nu Holdings (NYSE: NU)

Nu Holdings made a recent acquisition of Hyperlane, a provider of AI solutions in the financial services space; Hyperlane’s AI platform has improved the performance of even Nu Holdings’ most advanced machine learning models when utilising a foundation model focused on financial services that used Nu Holdings’ own unstructured data

I wanted to highlight our recently announced acquisition of Hyperplane. Hyperplane is a Silicon Valley-based leader in AI power solutions for the financial services space. As we tested Hyperplane’s platform on our vast amount of data, we were impressed by the opportunity to meaningfully improve performance of even our most advanced machine learning models by using a financial services focused foundation model that included our own unstructured data. We’re very excited to welcome the Hyperplane team on board and see them as a key part of our AI strategy in the foreseeable future. 

Shopify (NASDAQ: SHOP)

Shopify’s management believes the company can continue to post operating leverage, partly through the internal use of AI to drive productivity

We believe that we can continue to drive operating leverage through 4 key things: disciplined growth in headcount, which we have kept essentially flat for 5 quarters and where we expect we can keep head count growth well below revenue growth; strategic returns-based marketing to support and sustain our long-term revenue growth; internal use of AI and automation to drive productivity; and leveraging and continuing to enhance our internally-built GSD and Shopify OS systems, which allow us to smartly aim the product development work and size the team for maximum impact and efficiency.

Taiwan Semiconductor Manufacturing Company (NYSE: TSM)

TSMC’s capital expenditure is always in anticipation of growth in future years; capex for 2024 is now expected to be US$30 billion to US$32 billion (2023’s capex was US$30.4 billion), up at the low-end from commentary given in the 2024 Q1 earnings call; most of TSMC’s capex are for advanced process technologies; management sees strong structural AI-related demand and is willing to invest to support its customers

Every year, our CapEx is spent in anticipation of the growth that will follow in the future years, and our CapEx and capacity planning is always based on the long-term market demand profile. As the strong structural AI-related demand continues, we continue to invest to support our customers’ growth. We are narrowing the range of our 2024 capital budget to be between USD 30 billion and USD 32 billion as compared to USD 28 million to USD 32 billion previously. Between 70% and 80% of the capital budget will be allocated for advanced process technologies. About 10% to 20% will be spent for specialty technologies, and about 10% will be spent for advanced packaging, testing, mass-making and others. At TSMC, a higher level of capital expenditures is always correlated with the higher growth opportunities in the following years. 

TSMC’s management is seeing a continuation of a strong surge in AI-related demand, which supports structural demand for energy-efficient computing

The continued surge in AI-related demand supports a strong structural demand for energy-efficient computing.

TSMC’s management sees TSMC as a key enabler of AI; management has a disciplined framework, consisting of both a top-down and bottoms-up approach, to plan its capacity buildout; management is not going to make the same kind of mistake it made in 2021 and 2022 when planning its capacity; management has spent a lot of effort studying AI-demand for its capacity-planning and has also asked its customer (likely referring to Nvidia) to be more realistic; management has been testing out AI within TSMC and have found it to be very useful, so management thinks AI demand is real; TSMC has been buying chips from its customer (likely referring to Nvidia)

 As a key enabler of AI applications, the value of our technology position is increasing as customers rely on TSMC to provide the most advanced process and packaging technology at scale in the most efficient and cost-effective manner. As such, TSMC employs a disciplined framework to address the structural increase in the long-term market demand profile underpinned by the industry megatrend of AI, HPC and 5G. We work closely with our customers to plan our capacity. We also have a rigorous and robust system that evaluates and judges market demand from both a top-down and bottom-up approach to determine the appropriate capacity to build…

… [Question] Now looking at GenAI, obviously, the technology has lots of great potential, but a new technology also have lots of volatilities where you start to ramp. And so how are we managing the volatilities of the demand? Why do you think this time around it is different versus COVID period?

[Answer] I thought I explained that our capacity premium process, right, and the investment, we have — I put a wording of discipline. That means we are not going to repeat the same kind of mistake that we have in 2021, 2022. Now this time, again, we look at the overall very big demand forecast for my customer. And so I look at it into actually the whole company with many people now examining and study that really is AI is so used for will be used by a lot of people or not. And we test ourself first inside TSMC, we are using AI, we are using machine learning skill to improve our productivity, and we found out it’s very useful. And so I also in the line to buy my customer’s product, and we have to form in the line, like I cannot privilege here, I’m sorry, but it’s useful.

And so I believe that this time, AI’s demand is more real than 2 or 3 years ago. At that timing it is because people were afraid of a shortage, and so automotive, everything, you name it, they are all in shortage. This time, AI alone only AI alone, it will be a very useful tool for the human being to improve all the productivity in our daily life, be it in medical industry or in any product, manufacturing industry or autonomous driving, everything you need AI. And so I believe it’s more real. But even with that, we also have a top-down bottom-up approach and discuss with our customers and ask them to be more realistic. I don’t want to repeat the same kind of mistake 2 or 3 years ago, and that’s what we are doing right now.

TSMC’s management sees N2, N2P, and A16 as the technologies that will enable TSMC to capture growth opportunities in the years ahead; TSMC’s AI customers are migrating aggressively from N-1 to leading edge nodes, and management is seeing a lot of customers wanting to move into N2, N2P, and A16 quickly, but capacity is very tight and will only loosen in the next year or two years

We believe N2, N2P, N16 and its derivatives will further extend our technology leadership position and enable TSMC to capture the growth opportunities well into the future…

…[Question] We’re hearing that AI chipmakers are looking to migrate more aggressively from N-1 to the leading edge, particularly due to backside power because they’re trying to lower their power budgets going forward. So my question, can you support this move?

[Answer] You are right. All the people want to move into kind of a power-efficient mode. And so they are looking for the more advanced technology so that they can save power consumption. And so a lot of my customers want to move into N2, N2P, A16 quickly. We are working very hard to build the capacity to support them. Today, it’s a little bit tight, not a little bit, actually, today is very tight. I hope in next year or the next 2 years, we can build enough capacity to support this kind of demand. 

TSMC’s management is seeing such high demand for AI-accelerator and CoWoS packaging that supply is so tight; management is hopeful that a balance between demand and supply can be met in 2025 or 2026; it appears that TSMC will be doubling CoWoS capacity again in 2025; CoWoS (or advanced packaging) used to have much lower gross margin than the corporate average, but it is now approaching the corporate average; TSMC is working with its OSAT (outsourced semiconductor assembly and test) partners to expand its CoWoS capacity

[Question] How do you think about supply/demand balance for AI accelerator and CoWoS advanced packaging capacity?

[Answer]  I also tried to reach the supply and demand balance, but I cannot today. The demand is so high. I had to work very hard to meet by customers demand. We continue to increase. I hope sometime in 2025 or 2026, I can reach the balance… The supply continues to be very tight all the way to probably 2025 and hope it can be eased in 2026. That’s today’s situation…

…[Question] Are you going to double your capacity again next year for CoWoS?

[Answer] The last time I say that this year, I doubled it, right, more than double, okay? So next year, if I say double it, probably, I will answer your question again next year, and say more than double, okay? We’re working very hard, as I said, wherever we can, whenever we can…

…For advanced packaging, the gross margin used to be much lower than the corporate average. Now it’s approaching corporate average. We are improving it that’s because of scale of the economics, and we put a lot of effort to reduce our cost. So gross margin is greatly improving in these 2 years…

… I just answered the question whether the CoWoS capacity is enough or not? Is not enough. And in great shortage, and that limited my customers’ growth. So we are working with our OSAT partner and trying to give more capacity to my customer so that they can grow here.

TSMC’s smartphone customers have been using InFO (Integrated Fan-Out) technologies but as they start building edge-AI devices, they are starting to use 3DIC (Three Dimensional Integrated Circuit) and SoIC (System on Integrated Chip) technologies

[Question] In regards to advanced packaging with more and more customers working on edge AI devices without — well, being overly specific, but what does it mean or the implication for advanced packaging solutions that we expect in the next 2 years to see these edge AI customers start to use SoIC or 3DIC particularly smartphone? Will they still be using info? Or will they also consider these solutions as well.

[Answer] As my customer moving into 2-nanometer or A16, they all need to probably take in the approach of chiplets. So once you use your chiplets, you have to use in advanced packaging technologies. On the edge AI, for those kind of smartphone customer, as compared with the HPC customers, HPC is moving faster because of bandwidth concerns, latency of footprint or all those kind of thing. For smartphone customer, they need to pay more attention to the footprint as well as the functionality increase. So you observe my big customers taking the info first and then for a few years, nobody catch it up. They are catching up okay? 

TSMC’s management is seeing a lot of customers wanting to put AI functionality into edge devices; this will increase dye sizes by 5% to 10%, but so far there’s no spike in unit growth of the devices; management thinks the unit growth will happen a few years later as the AI functionalities start to stimulate demand for replacement of older devices

[Question] For silicon content, recall a few years back when 5G just started to ramp you used to provide the silicon content expectations of 5G high-end and mid-end and low-end smartphones, so I wonder at this point of time, if you have any estimates for AI for smartphone going to next 2, 3 years?

[Answer] AI is so hard. So that’s right now everybody — all my customers want to put the AI functionality into the edge devices and so the dye size will be increased, okay? How much? I mean it’s different from my customer-to-customers product. But basically, probably 5% to 10% dye size increase will be a general rule. Unit growth, not yet, okay? Because we did not see kind of unit growth suddenly increased, but we expect this AI functionality was stimulated some of the demand to stimulate the replacement to be shorter. So in terms of unit growth that in a few years later, probably 2 years later, you will start to see a big increase in the edge device that’s a smartphone and the PC.

AI chips have larger die sizes, so TSMC’s management thinks there’s a need to adopt fan-out panel-level packaging eventually, but the technology is currently not mature enough and will need 2-3 years to attain that maturity

[Question] We also see the bigger footprint of the AI chips. So while there are quite some activities about fan-out panel-level packaging. So do you think that, that solution will be mentioned in the mid- to long run? Or does TSMC have any plan to do the related investment?

[Answer] We are looking at this as kind of a panel level fan-out technology. But the maturity today is not yet, so I — personally, I will think it’s about at least 3 years later, okay? In this, within these 3 years, we don’t have any very solid solution for a dye size bigger than 10x of the radical size. Today, we support our customer all the way to 5x, 6x chip size. I’m talking about the [ fuel ] size, the big [indiscernible] size. 2 years later, I believe the panel fan-out will be — start to be introduced and we are working on it.

Tencent (NASDAQ: TCEHY)

Tencent’s advertising business is benefitting from better click through rates driven by AI; management sees AI technology increasing advertising conversion rates by 10%

We are benefiting from deployment of neural network artificial intelligence on a GPU infrastructure to boost the click-through rate on our advertising inventory…

…And at the same time, on the ad recommendation end, if we can actually increase conversion by 10%, right, that’s sort of pretty modest improvement. The revenue actually grows quite a bit, right? So I think that’s areas in which we are leveraging AI to deliver material and tangible commercial results.

Tencent’s AI-related external revenue is growing, and the company recently launched 3 AI-powered solutions for enterprises, namely image generation engine, video generation engine, and knowledge engine

Tencent Meeting deepened its adoption and monetization, especially in the pharmaceutical manufacturing and retail sectors. We’re generating increasing AI-related external revenue from customers utilizing our high-performance computing infrastructure, such as GPUs and our model library services. We’re generating increasing AI-related external revenue from customers utilizing our high-performance computing infrastructure, such as GPUs and our model library services. We recently launched 3 AI-powered platform solutions for enterprises, image generation engine and video generation engine, which are pretty useful for advertisers creating ad content; as well as knowledge engine, which is particularly useful for finance, education and retail-related services, deploying customer service chat bots.

Tencent’s operating capex in 2024 Q2 was up 144% year-on-year because of investments in GPUs and CPUs; non-operating capex was up 53% year-on-year, driven by construction, but down 80% sequentially

Operating CapEx was RMB 7.2 billion, up 144% year-on-year driven by investment in GPU and CPU servers.  Non-operating CapEx was RMB 1.5 billion, up 53% year-on-year, driven by construction and progress. On a quarter-on-quarter basis, non-operating CapEx was down 80% from the high base in the prior quarter. As a result, total CapEx was RMB 8.7 billion, up 121% year-on-year.

Tencent’s management thinks of AI as more than just large language models

We look at AI as a more complete suite than just large language model. There are the neural networks, machine learning-based recommendation engines, which we use for content recommendation, video recommendation as well as the talking in the ads and content use case, which is already delivering very good result.

Tencent has delivered better content to users through the use of AI

If you take Video Accounts as an example, by using AI, we actually are able to deliver better content and that generates more use time — a pretty big part of the growth in terms of the Video Accounts user time. It’s actually driven by better targeting, better recommendation and that’s in turn driven by AI.

Tencent’s management thinks AI can improve PVE (player vs environment) games by making the computer smarter

In the area of games, we’re actually using AI to bridge the gap between PVE and PVP, right? So when you have games, which allow people to play against other players, but at the same time, sometimes you actually want to create a game mode in which a player actually play against the machine, right? Then — in the past, the machine is actually quite dumb, right? And with AI, we can actually make the machine play like a real player. And we can actually sort of have it to play a varying levels of skills and make the user experience and the gameplay very fun.

Tencent’s management’s focus with LLMs is to improve the technology; Tencent has already built a MOE (mixture of experts) architecture model, which is one of the top AI models in the Chinese language; Tencent is deploying its LLM in Yuanbao, an app launched to allow users to interact with its LLM; Tencent’s LLM is improving search results and Yuanbao is getting positive feedback; when Yuanbao improves, management will increase promotional resources to increase the user base; management also wants to incorporate Yuanbao into different parts of its ecosystem

Now in terms of LLM, the key thing for us is actually improving the technology. And as we shared before, we have already built an MOE architecture model, which is performing as one of the top models in China. And when compared with international models on Chinese language, I think we are at that top of the pack. And we are deploying our LLM in Yuanbao, which is an app that we have launched which allowed users to interact with our large language model in multiple ways. And one way is enhanced search functionality so that users can actually ask a question. And based on search results, we can actually provide a very direct answer to the questions that our users pose and we have rolled it out to a large enough sample size to get user feedback and the feedback so far has been quite positive…

…Over time, Yuanbao, when it gets to a certain level of quality, then we’re going to increase our promotional resources and try to get more users into the app. And at the same time, when it gets to an even better level of expertise, then we can actually start incorporating it into different parts of our ecosystem. We have a lot of apps which actually has got interaction use cases, which we can leverage our generative AI technology.

Renting out GPUs for AI workloads is a big business in China too, but it’s to a smaller extent when compared to what’s happening in the USA; Tencent’s management is seeing very fast growth in demand for GPU-rentals for AI needs partly because the growth is happening off a low base; the demand for GPU-rentals is partially cannibalising the demand for CPUs

Clearly, for the U.S. hyperscale Cloud providers, renting out GPUs to other companies with AI requirements has become a very big business. In China, the same trend is evident, but to a lesser extent because you don’t have the same multitude of extremely well-funded start-ups trying to build large language models on their own in China. There are many small companies, but they’re capitalized for $1 billion, $2 billion. They’re not capitalized at $10 billion or $90 billion, other way that some of the giant U.S. VC-funded start-ups are now capitalized in the space. And it’s also a somewhat challenging economic environment. Now that said, we have seen that within our Cloud, the demand from customers for renting GPUs for their own AI needs has been growing very swiftly. The percentage growth rates are very fast, but they’re very fast partly because it’s a low base. And also partly because, while some of that demand for renting GPUs in the Cloud is incremental, some of it is replacing demands that would otherwise have existed anyway for renting CPUs in the Cloud. And so while the business of GPU provision is doing very well, the business of CPU processing is more flat because the incremental demand is for GPU, not CPU.

Tesla (NASDAQ: TSLA)

Tesla has made a lot of progress with full self-driving in Q2; a new version, version 12.5, of the autonomous software has just started to be rolled out; version 12.5 of the FSD (full self-driving) software is a step-change improvement in supervised full self-driving; management thinks that most people still do not know how good version 12.5 is; as Tesla increases the miles between intervention, the system can transition from supervised full self-driving to unsupervised full self-driving; management would be shocked if Tesla cannot achieve unsupervised full self-driving next year, but they also note that they have been overly optimistic on the timeline for self-driving; management believes that Tesla will be able to get regulatory approval for unsupervised full self-driving once it shows the rate of accidents is less than human driving; self-driving capabilities of Tesla vehicles outside of North America are far behind those of Tesla vehicles in North America; management is asking for regulatory approval of Tesla supervised full self-driving in Europe, China, and other countries, and the approvals, which are expected before end-2024, will be a driver of demand for Tesla vehicles; FSD uptake is still low despite some increase after a recent price reduction

Regarding full self-driving and Robotaxi, we’ve made a lot of progress with full self-driving in Q2. And with version 12.5 beginning rollout, we think customers will experience a step change improvement in how well supervised full self-driving works. Version 12.5 has 5x the parameters of 12.4 and finally merged the highway and city stacks. So the highway stack at this point is pretty old. So often the issues people encounter are on the highway. But with 12.5, we finally merged the 2 stacks. I still find that most people actually don’t know how good the system is. And I would encourage anyone to understand the system better to simply try it out and let the car drive you around…

…And as we increase the miles between intervention, it will transition from supervised full self-driving to unsupervised full self-driving, and we can unlock massive potential [ in the fleet ]…

…I guess that, that’s really just a question of when can we expect the first — or when can we do unsupervised full self-driving. It’s difficult, obviously, my predictions on this have been overly optimistic in the past. So I mean, based on the current trend, it seems as though we should get miles between interventions to be high enough that — to be far enough in excess of humans that you could do unsupervised possibly by the end of this year. I would be shocked if we cannot do it next year. So next year seems highly probable to me based on quite simply plus the points of the curve of miles between intervention. That trend exceeds the humans for sure next year, so yes…

So it’s this capability. I think in our experience, once we demonstrate that something is safe enough or significantly safer than human, we find that regulators are supportive of deployment of that capability. It’s difficult to argue with — if you have got a large number of — if you’ve got billions of miles that show that in the future unsupervised FSD is safer than human, what regulator could really stand in the way of that. They’re morally obligated to approve. So I don’t think regulatory approval will be a limiting factor. I should also say that the self-driving capabilities that are deployed outside of North America are far behind that in North America. So with Version 12.5, and maybe 12.6, but pretty soon, we will ask for regulatory approval of the Tesla supervised FSD in Europe, China and other countries. And I think we’re likely to receive that before the end of the year. There will be a helpful demand driver in those regions…

[Question] You mentioned that FSD take rates were up materially after you reduced the price. Is there any way you can help us quantify what that means exactly?

[Answer] We’ve shared that how — that we’ve seen a meaningful increase. I don’t want to get into specifics because we started from a low base, but we are seeing encouraging results. 

Tesla will unveil its robotaxi product on 10th of October, after postponing it for a few months; the current plan is for robotaxis to be produced in Tesla’s headquarters at Giga Texas; management’s aim is to have a robotaxi fleet that’s made up of both Tesla-owned vehicles and consumer-owned vehicles, and consumers can rent out their cars, just like renting out their apartments for Airbnb; Tesla has a clause with every vehicle purchase that Tesla vehicles can only be used in the Tesla fleet and not in any 3rd-party autonomy fleet; management believes that once unsupervised full self-driving is available, most people will rent out their Tesla vehicles, so the Tesla robotaxi service will achieve instant scale given the existing number of Teslas on the road

We postponed the sort of robotaxi product unveil by a couple of months where it’s shifted to 10/10, to the 10th of October. And this is because I wanted to make some important changes that I think would improve the vehicle — the sort of — the Robotaxi — the thing — the main thing that we’re going to show…

…And I should say that the Cybertaxi or Robotaxi will be locally produced here at our headquarters at Giga Texas… 

This would just be the Tesla network. You just literally open the Tesla app and summon a car and we send a car to pick you up and take you somewhere. And our — we will have a fleet that’s on the order of 7 million [ vehicle autonomy ] soon. In the U.S. it will be over 10 million and over 20 million. This is in that scale. And the car is able to operate 24/7 unlike the human drivers. So the capability to — like this basically instant scale with a software update. And now this is for a customer-owned fleet. So you can think of that as being a bit like Airbnb, like you can choose to allow your car to be used by the fleet or cancel that and bring it back. It will be used by the fleet all the time, can be used by the fleet some of the time and then Tesla will take a share in the revenue with the customer…

…And there’s an important clause we’ve put in every Tesla purchase, which is that the Tesla vehicles can only be used in the Tesla fleet. They cannot be used by a third party for autonomy…

…[Question] Do you think that scales like progressively, so you can start in a city with just a handful of cars. Then you grow the number of cars over time? Or do you think there is like a critical mass you need to get to, to be able to offer like a service that is of competitive quality compared to what like Uber would be typically delivering already?

[Answer] I guess I’m not — I’m not conveying this correctly. The entire Tesla fleet basically becomes active. This is obviously — maybe there’s some number of people who don’t want their car to earn money. But I think most people will. It’s instant scale.

Tesla is nearing completion of the South expansion of Giga Texas, which is Tesla’s largest training cluster of GPUs to-date; there was a story earlier this year that Tesla sent its new H100 AI chip deliveries to Elon Musk’s other entities but this happened only because Tesla had no place to house the chips at that point in time; Tesla now has a place for the chips because of the South expansion of Giga Texas

We’re also nearing completion of the South expansion of Giga Texas, which will house our largest training cluster to date. So it will be an incremental 50,000 H100s, plus 20,000 of our hardware for AI5, Tesla AI computer…

…I mean I think you’re referring to a very — like an old article regarding GPUs. I think that’s like 6 or 7 months old. Tesla simply had no place to turn them on. So it would have been a waste of Tesla Capital because we would just have to order H100s and have no place to turn them on. So I was just – this wasn’t a let’s pick xAI over Tesla. There was no — the Tesla test centers were full. There was no place to actually put them. The — we’ve been working 24/7 to complete the South extension on the Tesla [indiscernible] Texas. That self extension is what will house the 50,000 H100s, and we’re beginning to move the certain H100 server racks in place there. But we really needed — we needed that to complete basically. You can’t just order compute — order GPUs and turn them on, you need a data center. So I want to be clear, that was in Tesla’s interest, not contrary to Tesla’s interest. Does Tesla no good to have GPUs that it can’t turn on. That South extension is able to take GPUs, which is really just this week. We are moving the GPUs in there and we’ll bring them online.

The Optimus robot is already performing tasks in Tesla’s factory; management expects to start limited production of Optimus in early 2025; early production is for Tesla’s consumption, and management expects a few thousand robots in Tesla’s factories by end-2025; management expects Optimus to enter high-volume production in 2026 and to release Optimus to external customers by then; management believes that Optimus will be the biggest revenue contributor to Tesla in the future, with an estimated total addressable market of 20 billion units of Optimus robots; management thinks Tesla has all the ingredients to build large scale, generalised humanoid robots 

With Optimus, Optimus is already performing tasks in our factory. And we expect to have Optimist production Version 1 and limited production starting early next year. This will be for Tesla consumption. It’s just better for us to iron out the issues ourselves. But we expect to have several thousand Optimus robots produced and doing useful things by the end of next year in the Tesla factories. And then in 2026, ramping up production quite a bit. And at that point, we’ll be providing Optimus robots to outside customers. There will be a production Version 2 of Optimus…

I mean, as I said a few times, I think the long-term value of Optimus will exceed that of everything else that Tesla combined. So it’s simply just never considered the usefulness, utility of a humanoid robot that can do pretty much anything you asked of it. II think everyone on earth is going to want one. There are 8 billion people on earth. So it’s 8 billion right there. Then you’ve got all of the industrial uses, which is probably at least as much, if not, way more. So I suspect that the long term demand for general purpose humanoid robots is in excess of 20 billion units. And Tesla has the most advanced humanoid robot in the world and is also very good at manufacturing, which these other companies are not. And we’ve got a lot of experience with — the most experienced — we’re the word leaders in [ Real World AI ]. So we have all of the ingredients. I think we’re unique in having all of the ingredients necessary for large scale, high utility, generalized humanoid robots.

Management expects capex to be over US$10 billion in 2024 (was US$8.9 billion in 2023) because of spending on the AI GPU cluster

On the CapEx front, while we saw a sequential decline in Q2, we still expect the year to be over $10 billion in CapEx as we increase our spend to bring a 50 GPU cluster on luck. This new center will immensely increase our capabilities to scale FSD and other AI initiatives. 

Tesla will continue working on its own AI GPU called Dojo to reduce reliance on NVIDIA, and also because NVIDIA’s supply for GPUs is so tight; management sees a path where Dojo’s chips can be competitive with NVIDIA’s

So Dojo, I should preface this by saying I’m incredibly impressed by NVIDIA’s execution and the capability of their hardware. And what we are seeing is that the demand for NVIDIA hardware is so high that it’s often difficult to get the GPUs. And there just seems this — I guess I’m quite concerned about actually being able to get steady out NVIDIA GPUs and when we want them. And I think this therefore requires that we put a lot more effort on Dojo in order to have — in order to ensure that we’ve got the training capability that we need. So we are going to double down on Dojo and we do see a path to being competitive with NVIDIA with Dojo. And I think we kind of have no choice because the demand for NVIDIA is so high and it’s obviously their obligation essentially to raise the price of GPUs to whatever the market will bear, which is very high. So I think we’ve really got to make Dojo work and we will.

Tesla is learning from Elon Musk’s AI startup, xAI; Musk is aware that Tesla needs shareholder approval before the company can invest in xAI, but he thinks it’s a good idea; Musk sees opportunities to integrate xAI’s foundation model, Grok, into Tesla’s software; Musk found that some engineers are only interested in working on AGI (artificial general intelligence) and they would have gone to other AI startups if Musk was not working on xAI since they would not have chosen Tesla anyway

Tesla is learning quite a bit from xAI. It’s been actually helpful in advancing full self-driving and in building up the new Tesla data center. With — regarding investing in xAI, I think, we need to have a shareholder approval of any such investment. But I’m certainly supportive of that if shareholders are, the group — probably, I think we need a vote on that. And I think there are opportunities to integrate Grok into Tesla’s software, yes…

…With regard to xAI, there are a few that only want to work on AGI. So what I was finding was that when trying to recruit people to Tesla, they were only interested in working on AGI and not on Tesla’s specific problems and they want to start — do a start-up. So it was a case of either they go to a startup or — and I am involved or they do a start-up and I am not involved. Those are the 2 choices. This wasn’t they would come to Tesla. They were not going to come to Tesla under any circumstances…

…I tried to recruit them to Tesla, including to say, like, you can work on AGI if you want and they refused. Only then was xAI created.

Management still thinks Tesla can rent out latent AI inferencing compute for general computing purposes from its fleet of vehicles (and perhaps humanoid robots) in the future

Just distributed compute. It seems like a pretty obvious thing to do. I think where the distributed compute becomes interesting is with next-generation Tesla AI truck, which is hardware viable, what we’re calling AI5, which is from the standpoint of inference capability comparable to B200 and [ a bit of ] B200. And we’re aiming to have that in production at the end of next year and scale production in ’26. So it just seems like if you’ve got autonomous vehicles that are operating for 50 or 60 hours a week, there’s 168 hours in a week. So we have somewhere above, I think, 100 neural net computing. I think we need a better word than GPU because GPU means graphics processing unit. So there’s a 100 hours plus per week of AI compute, AI [ first ] compute from the fleet in the vehicles and probably some percentage from humanoid robots. That it would make sense to do distributed inference. And if there’s a fleet of at some point, 100 million vehicles with AI5 and beyond, AI6 and 7 and what not and there are maybe billions of humanoid robots. That is just a staggering amount of inference compute that could be used for general purpose computing. Doesn’t have to use it for the humanoid robot or for the car.

Management believes that Waymo’s approach to autonomous vehicles is a localised solution that requires high-density mapping and is thus quite fragile compared to Tesla’s approach

I mean our solution is a generalized solution like what everybody else has. You could see if Waymo has [ one of it ], they have very localized solution that requires high-density mapping. It’s not — it’s quite fragile. So their ability to expand, I believe, is limited. Our solution is a general solution that works anywhere. It would even work on a different earth. So if you [ branded ] a new earth, it would work on new earth…

…in terms of regulatory approval, the vehicles are governed by FMVSS in U.S., which is the same across all 50 states. The road rules are the same across all 50 states. So creating a generalized solution gives us the best opportunity to deploy in all 50 states reasonably. Of course, there are state and even local municipal level regulations that may apply to being a transportation company or deploying taxis. But as far as getting the vehicle on the road, that’s all federal and that’s very much in line with what Elon was suggesting of the data and the vehicle itself…

…To add to the technology point, the end-to-end network basically makes no assumption about the location. Like you could add data from different countries and it just like performs equally well there. That’s like almost close to 0, U.S. specific code in there. It’s all just the data that comes from the U.S.

Visa (NYSE: V)

Visa’s management is investing in AI, particularly generative AI (genAI), because the company has use-cases for the technology in areas such as fraud reduction and productivity improvement; management is very optimistic about the positive impact that generative AI can have 

First of all, to frame it is we are all in on GenAI at Visa as we’ve been all in on predictive AI for more than a decade. We’re applying it in 2 broad-based different ways. One is sort of adopting across the company to drive productivity and we’re seeing real results there. We’re seeing great results, great adoption, great productivity increases from technology to accounting to sales all across the company. The second is applying generative AI to enhance the entire payment ecosystem. And to the latter part of your question, absolutely. I guess I’d give you one set of examples or some of the risk tools and capabilities that we’ve been deploying in the market. I mentioned the risk products that we’re using on RTP and account-to-account payments. That is an opportunity to reduce fraud, both for merchants and for issuers. I think I mentioned on a previous call, we have our Visa Provisioning Intelligence Service, which is using artificial intelligence to help predict token provisioning fraud before it happens. That also is a benefit to both issuers and merchants. And the list goes on. So we are very optimistic about the positive impact that generative AI can have, not just on our own productivity but on our ability to help drive increased sales and lower fraud across the ecosystem.

Wix (NASDAQ: WIX)

Wix’s management continues to improve the company’s AI capabilities; Wix has released 17 AI business assistants to-date; the AI business assistants support a wide range of use cases and Wix has already received positive feedback on them; Wix will be releasing dozens more AI assistant later in 2024; the 17 business assistants are all customer-facing but the assistants can play one of two roles, (1) be a question-and-answer AI assistant, and (2) be an assistant that executes actions; the AI business assistants rarely hallucinate; management wants to add these AI assistants everywhere in the Wix product suite

We continue to build up our suite of AI capabilities as a result of the numerous AI initiatives and work streams across Wix. Last quarter, we introduced our plan to embed AI assistance across our platform and products. I’m excited to share that we have released 17 AI business assistants so far to date. These assistants span a wide range of use cases to support users with minimal hands-on support, thus streamlining their experience. These conversational AI assistants act as a right-hand aid for users to guide them through the entire life cycle of ideating, creating and managing their online presence. Our offering includes an analytics assistant that can help Wix users find the data they need without having to search through dozens of reports, and an assistant that helps users create events through a conversational chat. We have already received positive feedback on this first set of AI assistants with dozens more set to launch later this year…

…how many of the 17 are customer-facing? And the answer is all of them. The concept is that we are currently — we build a platform in which it is easier for us to build an AI assistant. And then that enable us to develop 2 kinds of different assistants. The first one would be a question-and-answer AI assistant, so if you have a product like booking, how do I add a staff member to my yoga studio, right? And so you can actually talk to the AI and ask questions, get answered, and ask question, get answer, as you would do with the normal human being. And then we see a great result in that in terms of how customers quickly find the answers. Hallucinations are very small, the percentage, probably similar to what a human would do or not even better…

…The other thing that we are doing is that you can ask questions and you can have the AI do things for you. So this is the second kind. And for example, if you go to our analytics, you see that you can actually start asking questions and get the reports done for you automatically by the AI. So this is an AI that activates other agents in order to give you answers or do actions for you. How do I make an event that is a wedding event? What not? And then it will do — analyze [ VP ]. But if you want to create an event which is selling tickets for a concert, it will define that, willing to work with you on that. So those kind of things streamline and reduce a lot of friction from the customer…

…We’re going to add those kind of assistants in pretty much everywhere that we can on Wix. 

Wix’s management launched AI creation capabilities for its mobile app builder in June 2024, which enables users to create and edit iOS and Android apps through a cha 

We launched AI creation capabilities for our mobile app builder in June. This new solution enables users to create and edit iOS or Android apps through an AI chat experience. Once AI understands the user’s goals, intent and desired aesthetic, our technology generates a branded app that can be customized and managed from the App Editor.

Wix’s management recently released new AI features to help users with content-generation

We also recently released a suite of new AI features designed to help users identify relevant topics for blogs as well as generate outlined content and images for their target audience. With this new experience, users can swiftly turn ideas into new ready articles, significantly reducing the time and effort required to create engaging content, and ultimately, changing the blog creation experience.

Wix’s management sees both Self Creators and Partners having excellent engagement with Wix’s AI tools; management expects Wix’s AI tools to be a competitive advantage and significant driver of future growth; Wix’s AI tools continue to drive user conversion; Wix released its first AI product all the way back in 2016 and management saw that the AI functionality had very high adoption and drove dramatic improvement in user conversion; the latest version of the AI product, released earlier this year, had the same effect; Wix’s AI agents are having measurable positive impact on engagement; management thinks that their 7-8 years of experience with releasing AI technology is helping them integrate AI into Wix’s product suite in a highly intuitive way

Both Self Creators and Partners continue to show excellent engagement with our AI tools. As we expand the breadth of our AI technology, we expect it to continue to be a competitive advantage for us as well as a significant driver of growth going forward…

… Our AI tools continue to drive user conversion…

…Released ADI, the first AI product — GenAI product, actually created website right in the end of 2016. And since then, we’ve seen that by exposing users to AI functionality as part of the natural progression in the product life cycle, we get very high adoption, obviously using those kind of tools and results that can improve. And for ADI, we show that we improved the conversion dramatically. The new version that came earlier this year did it again. And we are seeing that a lot of the agents that we have now, AI agents, when they start to pick up more user interactions and more user conversations, again, create measurable effect. So I’m very optimistic. I think that our experience in releasing AI technology, right, which is almost, what, 8 years now — 7 years now, is helping us understand how to integrate them into the product in a way that actually mixed user interact with them and that they feel natural and don’t feel like you’re stepping out of what you’re doing to do something else and then coming back. And I think that creates a big difference. So yes, I’m very optimistic on the potential that we’re going to see a continuation of the improvement.

There is a big difference between what an agency and a Self Creators need from AI. So for me, if I want to design a website, and I’m not a designer, I want AI to help me design it because English is not my first language and I’m not writing so well in Hebrew as well, right? So I would love AI to also help me write great text and generate images.

When you’re an agency, you probably know how to design and you have your system of design and how things should look like. So you don’t need that. You probably need a little bit to help with the text, but other things, like the image editing, right, and the content recomposition create tremendous value. And then the other things that — in addition to that, for example, a great designer not necessarily know how to configure things to work in a responsive way on different screen resolutions, and we have an AI to do that. So we are utilizing those kind of technologies to streamline the agency’s experience and work and efficiency in a way that is significant to them. I think we have some ideas on how to make it even more significant going forward.

Wix’s management thinks there’s a long way to go before AI technology will make agencies become obsolete by having the computer know automatically what website you want to build and get it fully functioning, so agencies will still be an important business for Wix for many years

In theory, if you can just one day talk to a computer and get the full website functioning that knows exactly what should be there and that it’s easy to update then maybe some of the agency’s business will disappear. But there is a long way until we get to something similar to that. And I think the majority of businesses in the case that they need a website, they want somebody to be responsible for it, somebody that know how to activate the tools and use them and utilize them, and that’s why they go to agencies because they have a professional that understand how to take care of all of their business needs. And there’s a lot of those, right from SEO to how do you write things correctly in order to get the right shipping rules, and there’s a ton of things. So I think that where there’s a long way for AI to go before it can successfully replace good agencies. 

Unless, of course, you are a self-creator by nature, which is a lot of most of our customers, and you want to create your website, you can control it and you can do those things and you can change it. So I think the difference is in the user type and user intent and not necessarily in technology, which I believe means that both will continue to grow, agencies and Self Creators.

Wix’s management is seeing that the newer users who join Wix are those who use more AI tools to automate website creation as compared to earlier users; the presence of Wix’s AI tools opens up new types of customers for Wix

One of the qualification that you needed to have in order to be able to use Wix in the past was to know how to design to some level, to know how to write text to some level and to trust yourself that you’re good enough to do it, right? And then — so most of our users feel that they know how to do those things. And naturally, they will use less AI because they think they can just do it. And I think we are now opening to users that don’t feel that, right? They don’t expect themselves to know how to do those things and expect us to have the tools to — AI tools to automate it for them. So we are already seeing some of this gap, and I believe that this will continue to grow. And essentially, we are opening Wix to be more useful to more new types of customers.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Alphabet, Amazon, Apple, ASML, Coupang, Datadog, Fiverr, Mastercard, Meta Platforms, Microsoft, Netflix, Nu Holdings, Shopify, TSMC, Tesla, Visa, and Wix. Holdings are subject to change at any time.

What The USA’s Largest Bank Thinks About The State Of The Country’s Economy In Q2 2024

Insights from JPMorgan Chase’s management on the health of American consumers and businesses in the second quarter of 2024.

JPMorgan Chase (NYSE: JPM) is currently the largest bank in the USA by total assets. Because of this status, JPMorgan is naturally able to feel the pulse of the country’s economy. The bank’s latest earnings conference call – for the second quarter of 2024 – was held three weeks ago and contained useful insights on the state of American consumers and businesses. The bottom-line is this: The US economy is stronger than what many would have thought a few years ago given the current monetary conditions, but there are signs of weakness such as slightly higher unemployment and slower GDP growth; at the same time,  inflation and interest rates may stay higher than the market expects, and the Fed’s quantitative tightening may have unpredictable consequences.

What’s shown between the two horizontal lines below are quotes from JPMorgan’s management team that I picked up from the call.


1. Broader financial market conditions suggest a benign economic outlook, but JPMorgan’s management continue to be vigilant about potential tail risks; management is concerned about inflation and interest rates staying higher than the market expects, and the effects of the Federal Reserve’s quantitative tightening

While market valuations and credit spreads seem to reflect a rather benign economic outlook, we continue to be vigilant about potential tail risks. These tail risks are the same ones that we have mentioned before. The geopolitical situation remains complex and potentially the most dangerous since World War II — though its outcome and effect on the global economy remain unknown. Next, there has been some progress bringing inflation down, but there are still multiple inflationary forces in front of us: large fiscal deficits, infrastructure needs, restructuring of trade and remilitarization of the world. Therefore, inflation and interest rates may stay higher than the market expects. And finally, we still do not know the full effects of quantitative tightening on this scale.

2. Net charge-offs (effectively bad loans that JPMorgan can’t recover) rose from US$1.4 billion a year ago, mostly because of card-related credit losses that are normalising to historical norms

Credit costs were $3.1 billion, reflecting net charge-offs of $2.2 billion and a net reserve build of $821 million. Net charge-offs were up $820 million year-on-year, predominantly driven by Card…

…I still feel like when it comes to Card charge-offs and delinquencies, there’s just not much to see there. It’s still — it’s normalization, not deterioration. It’s in line with expectations. 

3. JPMorgan’s credit card outstanding loans was up double-digits

Card outstandings were up 12% due to strong account acquisition and the continued normalization of revolve.

4. Auto originations are down

In auto, originations were $10.8 billion, down 10%, coming off strong originations from a year ago while continuing to maintain healthy margins. 

5. JPMorgan’s investment banking fees had strong growth in 2024 Q2, partly because of favourable market conditions; management is cautiously optimistic about the level of appetite that companies have for capital markets activity, but headwinds persist 

This quarter, IB fees were up 50% year-on-year, and we ranked #1, with year-to-date wallet share of 9.5%. In advisory, fees were up 45% primarily driven by the closing of a few large deals and a weak prior year quarter. Underwriting fees were up meaningfully, with equity up 56% and debt up 51%, benefiting from favorable market conditions. In terms of the outlook, we’re pleased with both the year-on-year and sequential improvement in the quarter. We remain cautiously optimistic about the pipeline, although many of the same headwinds are still in effect. It’s also worth noting that pull-forward refinancing activity was a meaningful contributor to the strong performance in the first half of the year…

…In terms of dialogue and engagement, it’s definitely elevated. So I would say the dialogue on ECM [Equity Capital Markets] s elevated and the dialogue on M&A is quite robust as well. So all of those are good things that encourage us and make us hopeful that we could be seeing sort of a better trend in this space. But there are some important caveats.

So on the DCM [Debt Capital Markets] side, yes, we made pull-forward comments in the first quarter, but we still feel that this second quarter still reflects a bunch of pull-forward, and therefore, we’re reasonably cautious about the second half of the year. Importantly, a lot of the activity is refinancing activity as opposed to, for example, acquisition finance. So the fact that M&A remains still relatively muted in terms of actual deals has knock-on effects on DCM as well. And when a higher percentage of the wallet is refi-ed, then the pull-forward risk becomes a little bit higher.

On ECM, if you look at it kind of [ at a removed ], you might ask the question, given the performance of the overall indices, you would think it would be a really booming environment for IPOs, for example. And while it’s improving, it’s not quite as good as you would otherwise expect. And that’s driven by a variety of factors, including the fact that, as has been widely discussed, that extent to which the performance of the large industries is driven by like a few stocks, the sort of mid-cap tech growth space and other spaces that would typically be driving IPOs have had much more muted performance. Also, a lot of the private capital that was raised a couple of years ago was raised at pretty high valuations. And so in some cases, people looking at IPOs could be looking at down rounds, that’s an issue. And while secondary market performance of IPOs has improved meaningfully, in some cases, people still have concerns about that. So those are a little bit of overhang on that space. I think we can hope that over time that fades away and the trend gets a bit more robust.

And yes, on the advisory side, the regulatory overhang is there, remains there. And so we’ll just have to see how that plays out.

6. Management is seeing muted demand for new loans from companies as current economic conditions make them cautious

Demand for new loans remains muted as middle market and large corporate clients remain somewhat cautious due to the economic environment and revolver utilization continues to be below pre-pandemic levels. 

7. Demand for loans in the commercial real estate (CRE) market is muted

In CRE, higher rates continue to suppress both loan origination and payoff activity.

8. Lower income cohorts are facing a little more pressure than higher income cohorts because even though the US economy is stronger than what many would have thought a few years ago given the current monetary conditions, there is currently slightly higher unemployment and slower GDP growth

As I say, we always look quite closely inside the cohort, inside the income cohorts. And when you look in there, specifically, for example, on spend patterns, you can see a little bit of evidence of behavior that’s consistent with a little bit of weakness in the lower-income segments, where you see a little bit of rotation of the spend out of discretionary into nondiscretionary. But the effects are really quite subtle, and in my mind, definitely entirely consistent with the type of economic environment that we’re seeing, which, while very strong and certainly a lot stronger than anyone would have thought given the tightness of monetary conditions, say, like they’ve been predicting it a couple of years ago or whatever, you are seeing slightly higher unemployment, you are seeing moderating GDP growth. And so it’s not entirely surprising that you’re seeing a tiny bit of weakness in some pockets of spend. 

9. The increase in nonaccrual loans in the Corporate & Investment Bank business is not a broader sign of cracks happening in the business

[Question] I know your numbers are still quite low, but in the Corporate & Investment Bank, you had about a $500 million pickup in nonaccrual loans. Can you share with us what are you seeing in C&I? Are there any early signs of cracks or anything?

[Answer] I think the short answer is no, we’re not really seeing early signs of cracks in C&I. I mean, yes, I agree with you like the C&I charge-off rate has been very, very low for a long time. I think we emphasized that at last year’s Investor Day. If I remember correctly, I think the C&I charge-off rate [ over the preceding ] 10 years was something like literally 0. So that is clearly very low by historical standards. And while we take a lot of pride in that number and I think it reflects the discipline in our underwriting process and the strength of our credit culture across bankers and the risk team, that’s not — we don’t actually run that franchise to like a 0 loss expectation. So you have to assume there will be some upward pressure on that. But in any given quarter, the C&I numbers tend to be quite lumpy and quite idiosyncratic. So I don’t think that anything in the current quarter’s results is indicative of anything broader and I haven’t heard anyone internally talk that way, I would say.

10. Management is unwilling to lower their standards for risk-taking just because it has excess capital because they think it makes sense to be patient now given their current assessment of economic risk

And of course, for the rest of the loan space, the last thing that we’re going to do is have the excess capital mean that we lean in to lending that is not inside our risk appetite or inside our credit box, especially in a world where spreads are quite compressed and terms are under pressure. So there’s always a balance between capital deployment and assessing economic risk rationally. And frankly, that is, in some sense, a microcosm of the larger challenge that we have right now. When I talked about if there was ever a moment where the opportunity cost of not deploying the capital relative to how attractive the opportunities outside the walls of the company are, now would be it in terms of being patient. That’s a little bit one example of what I was referring to.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I don’t have a vested interest in any company mentioned. Holdings are subject to change at any time.

Why It’s So Difficult To Tell When The Stock Market Will Peak (Revised)

Many investors think that it’s easy to figure out when stocks will hit a peak. But it’s actually really tough to tell when a bear market would happen.

Note: This article is a copy of Why It’s So Difficult To Tell When The Stock Market Will Peak that I published more than four years ago on 21 February 2020. With the US stock market at new all-time highs, I thought it would be great to revisit this piece. The content in the paragraphs and table near the end of the article have been revised to include the latest valuation and returns data. 

Here’s a common misconception I’ve noticed that investors have about the stock market: They think that it’s easy to figure out when stocks will hit a peak. Unfortunately, that’s not an easy task at all.

In a 2017 Bloomberg article, investor Ben Carlson showed the level of various financial data that were found at the start of each of the 15 bear markets that US stocks have experienced since World War II:

Source: Ben Carlson

The financial data that Carlson presented include valuations for US stocks (the trailing P/E ratio,  the cyclically adjusted P/E ratio, and the dividend yield), interest rates (the 10 year treasury yield), and the inflation rate. These are major things that the financial media and many investors pay attention to. (The cyclically-adjusted P/E ratio is calculated by dividing a stock’s price with the 10-year average of its inflation-adjusted earnings.)

But these numbers are not useful in helping us determine when stocks will peak. Bear markets have started when valuations, interest rates, and inflation were high as well as low. This is why it’s so tough to tell when stocks will fall. 

None of the above is meant to say that we should ignore valuations or other important financial data. For instance, the starting valuation for stocks does have a heavy say on their eventual long-term return. This is shown in the chart below. It uses data from economist Robert Shiller on the S&P 500 from 1871 to June 2024 and shows the returns of the index against its starting valuation for 10-year holding periods. It’s clear that the S&P 500 has historically produced higher returns when it was cheap compared to when it was expensive.

Source: Robert Shiller data; my calculations

But even then, the dispersion in 10-year returns for the S&P 500 can be huge for a given valuation level. Right now, the S&P 500 has a cyclically-adjusted P/E ratio of around 35. The table below shows the 10-year annual returns that the index has historically produced whenever it had a CAPE ratio of more than 30.

Source: Robert Shiller data; my calculations

If it’s so hard for us to tell when bear markets will occur, what can we do as investors? It’s simple: We can stay invested. Despite the occurrence of numerous bear markets since World War II, the US stock market has still increased by 532,413% (after dividends) from 1945 to June 2024. That’s a solid return of 11.4% per year. Yes, bear markets will hurt psychologically. But we can lessen the pain significantly if we think of them as an admission fee for worthwhile long-term returns instead of a fine by the market-gods. 


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I currently have no vested interest in any company mentioned. Holdings are subject to change at any time.

No, Dividends Are Great

Dividends are the fruits of our investments and are what makes investing in companies so profitable.

In recent times, large American technology companies such as Meta Platforms, Salesforce, and Alphabet have initiated a dividend.

It’s easy to imagine that their shareholders would be pleased about it, but this isn’t always the case. Some shareholders are actually disappointed about the dividend announcements. They think that the companies have nowhere else to invest their capital and are thus returning it to their shareholders. In other words, they think that the companies’ growth potential have stalled.

But I see things differently. Dividends are ultimately what we, as shareholders, invest in a company for. Long-term shareholders are here to earn a cash stream from investing in companies. This is akin to building your own business which generates profits which you can cash out and enjoy. As such, dividends are the fruits of our investment.

And just because a company has started paying a dividend does not mean it can’t grow its earnings. Just look at some of the dividend aristocrats that have grown their earnings over a long span of time. There are many companies that can generate high returns on invested capital. This means that they can pay out a high proportion of their earnings as dividends and still continue to grow.

Dividends can compound too

For investors who don’t want to spend the dividend a company is paying, they can put that dividend to use by reinvesting it.

When a company is not paying a dividend, its shareholders have to rely on management to invest the company’s profits. When there’s a dividend, shareholders can invest the dividend in a way that they believe give them the highest risk-adjusted return available. Moreover, a company’s management team may not be the best capital allocators around – in such a case, when the company generates excess cash, management may invest it in a way that does not generate good returns. When a company pays a dividend, shareholders can make their own decisions and do not have to rely on management’s capital allocation skills.

And if you think the company was better off buying back shares, you can simply buy shares of that company with your dividends. This will have a similar effect to share buybacks as it will increase your stake in the company.

What’s the catch?

Dividends have some downsides though. 

Compared to buybacks, reinvesting dividends to buy more shares may be slightly less effective as shareholders may have to pay tax on those dividends. For example, Singapore-based investors who buy US stocks have to pay a 30% withholding tax on all US-company dividends.

The other downside is there’s more work for shareholders. If management was reinvesting prudently and not paying dividends, shareholders wouldn’t need to make a decision. But with dividends, shareholders have to decide where and when to reinvest that dividend. This said, it does give shareholders more options and opens up possibilities of where the dividend can be invested, instead of just relying on management. To me, I would happily take this tradeoff.

Don’t fret

Dividends are good. It’s funny that I even need to say this.

Dividends are the fruits of our investments and are what makes investing in companies so profitable. Without it, we will just be traders of companies, and not investors.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I currently have a vested interest in Alphabet, Meta Platforms, and Salesforce. Holdings are subject to change at any time.