The Latest Thoughts From American Technology Companies On AI (2024 Q4) – Part 2

A collection of quotes on artificial intelligence, or AI, from the management teams of US-listed technology companies in the 2024 Q4 earnings season.

The way I see it, artificial intelligence (or AI), really leapt into the zeitgeist in late-2022 or early-2023 with the public introduction of DALL-E2 and ChatGPT. Both are provided by OpenAI and are software products that use AI to generate art and writing, respectively (and often at astounding quality). Since then, developments in AI have progressed at a breathtaking pace.

With the latest earnings season for the US stock market – for the fourth quarter of 2024 – coming to its tail-end, I thought it would be useful to collate some of the interesting commentary I’ve come across in earnings conference calls, from the leaders of technology companies that I follow or have a vested interest in, on the topic of AI and how the technology could impact their industry and the business world writ large. This is an ongoing series. For the older commentary:

I’ve split the latest commentary into two parts for the sake of brevity. This is Part 2, and you can Part 1 here. With that, I’ll let the management teams take the stand… 

Microsoft (NASDAQ: MSFT)

Microsoft’s management is seeing enterprises move to enterprise-wide AI deployments 

Enterprises are beginning to move from proof of concepts to enterprise-wide deployments to unlock the full ROI of AI. 

Microsoft’s AI business has surpassed an annual revenue run rate of $13 billion, up 175% year-on-year; Microsoft’s AI business did better than expected because of Azure, Microsoft Copilot (within Copilot, price per seat was a strength and still retains good signal for value)

Our AI business has now surpassed an annual revenue run rate of $13 billion, up 175% year-over-year…

…[Question] Can you give more color on what drove the far larger-than-expected Microsoft AI revenue? We talked a bit about the Azure AI component of it. But can you give more color on that? And our estimates are that the Copilot was much bigger than we had expected and growing much faster. Any more details on the breakdown of what that Microsoft AI beat would be great.

[Answer] A couple of pieces to that, which you correctly identified, number one is the Azure component we just talked about. And the second piece, you’re right, Microsoft Copilot was better. And what was important about that, we saw strength both in seats, both new seats and expansion seats, as Satya talked about. And usage doesn’t directly impact revenue, but of course, indirectly does as people get more and more value added. And also price per seat was actually quite good. We still have a good signal for value.

Microsoft’s management is seeing AI scaling laws continue to show up in both pre-training and inference-time compute, and both phenomena have been observed internally at Microsoft for years; management has seen gains of 2x in price performance for each new hardware generation, and 10x for each new model generation

AI scaling laws continue to compound across both pretraining and inference time compute. We ourselves have been seeing significant efficiency gains in both training and inference for years now. On inference, we have typically seen more than 2x price performance gain for every hardware generation and more than 10x for every model generation due to software optimizations. 

Microsoft’s management is balancing across training and inference in the buildout of Microsoft’s AI capacity; the buildout going forward will be governed by revenue growth and capability growth; Microsoft’s Azure data center capacity is expanding in line with both near-term and long-term demand signals; Azure has more than doubled its capacity in the last 3 years, and added a record amount of capacity in 2024; Microsoft’s data centres uses both in-house as well as 3rd-party chips

Much as we have done with the commercial cloud, we are focused on continuously scaling our fleet globally and maintaining the right balance across training and inference as well as geo distribution. From now on, it’s a more continuous cycle governed by both revenue growth and capability growth thanks to the compounding effects of software-driven AI scaling laws and Moore’s Law…

…Azure is the infrastructure layer for AI. We continue to expand our data center capacity in line with both near-term and long-term demand signals. We have more than doubled our overall data center capacity in the last 3 years, and we have added more capacity last year than any other year in our history. Our data centers, networks, racks and silicon are all coming together as a complete system to drive new efficiencies to power both the cloud workloads of today and the next-generation AI workloads. We continue to take advantage of Moore’s Law and refresh our fleet as evidenced by our support of the latest from AMD, Intel, NVIDIA, as well as our first-party silicon innovation from Maia, Cobalt, Boost and HSM.

Microsoft’s management is seeing growth in raw storage, database services, and app platform services as AI apps scale, with an example being Azure OpenAI apps that run on Azure databases and Azure App Services

We are seeing new AI-driven data patterns emerge. If you look underneath ChatGPT or Copilot or enterprise AI apps, you see the growth of raw storage, database services and app platform services as these workloads scale. The number of Azure OpenAI apps running on Azure databases and Azure App Services more than doubled year-over-year, driving significant growth in adoption across SQL, Hyperscale and Cosmos DB.

OpenAI has made a new large Azure commitment; OpenAI’s APIs run exclusively on Azure; management is still very happy with the OpenAI partnership; Microsoft has ROFR (right of first refusal) on OpenAI’s Stargate project

As we shared last week, we are thrilled OpenAI has made a new large Azure commitment…

… And with OpenAI’s APIs exclusively running on Azure, customers can count on us to get access to the world’s leading models…

…[Question] I wanted to ask you about the Stargate news and the announced changes in the OpenAI relationship last week. It seems that most of your investors have interpreted this as Microsoft, for sure, remaining very committed to OpenAI’s success, but electing to take more of a backseat in terms of funding OpenAI’s future training CapEx needs. I was hoping you might frame your strategic decision here around Stargate.

[Answer] We remain very happy with the partnership with OpenAI. And as you saw, they have committed in a big way to Azure. And even in the bookings, what we recognized is just the first tranche of it. And so you’ll see, given the ROFR we have, more benefits of that even into the future. 

Microsoft’s management thinks Azure AI Foundry has best-in-class tooling run times for users to build AI agents and access thousands of AI models; Azure AI Foundry already has 200,000 monthly active users after just 2 months; the models available on Azure AI Foundry include DeepSeek’s R1 model, and more than 30 industry-specific models from partners; Microsoft’s Phi family of SLMs (small language model) has over 20 million downloads

Azure AI Foundry features best-in-class tooling run times to build agents, multi-agent apps, AIOps, API access to thousands of models. Two months in, we already have more than 200,000 monthly active users, and we are well positioned with our support of both OpenAI’s leading models and the best selection of open source models and SLMs. DeepSeek’s R1 launched today via the model catalog on Foundry and GitHub with automated red teaming, content safety integration and security scanning. Our Phi family of SLM has now been downloaded over 20 million times. And we also have more than 30 models from partners like Bayer, PAYG AI, Rockwell Automation, Siemens to address industry-specific use cases.

Microsoft’s management thinks Microsoft 365 Copilot is the UI (user interface) for AI; management is seeing accelerated adoption of Microsoft 365 Copilot across all deal sizes; majority of Microsoft 365 Copilot customers purchase more seats over time; daily users of Copilot more than doubled sequentially in 2024 Q4, while usage intensity grew 60% sequentially; more than 160,000 organisations have used Copilot Studio, creating more than 400,000 custom agents in 2024 Q4, uo 2x sequentially; Microsoft’s data cloud drives Copilot as the UI for AI; management is seeing Copilot plus AI agents disrupting business applications; the initial seats for Copilot were for departments that could see immediate productivity benefits, but the use of Copilot then spreads across the enterprise

Microsoft 365 Copilot is the UI for AI. It helps supercharge employee productivity and provides access to a swarm of intelligent agents to streamline employee workflow. We are seeing accelerated customer adoption across all deal sizes as we win new Microsoft 365 Copilot customers and see the majority of existing enterprise customers come back to purchase more seats. When you look at customers who purchased Copilot during the first quarter of availability, they have expanded their seat collectively by more than 10x over the past 18 months. To share just one example, Novartis has added thousands of seats each quarter over the past year and now have 40,000 seats. Barclays, Carrier Group, Pearson and University of Miami all purchased 10,000 or more seats this quarter. And overall, the number of people who use Copilot daily, again, more than doubled quarter-over-quarter. Employees are also engaging with Copilot more than ever. Usage intensity increased more than 60% quarter-over-quarter and we are expanding our TAM with Copilot Chat, which was announced earlier this month. Copilot Chat, along with Copilot Studio, is now available to every employee to start using agents right in the flow of work…

…More than 160,000 organizations have already used for Copilot Studio, and they collectively created more than 400,000 custom agents in the last 3 months alone, up over 2x quarter-over-quarter…

…What is driving Copilot as the UI for AI as well as our momentum with agents is our rich data cloud, which is the world’s largest source of organizational knowledge. Billions of e-mails, documents and chats, hundreds of millions of Teams meetings and millions of SharePoint sites are added each day. This is the enterprise knowledge cloud. It is growing fast, up over 25% year-over-year…

…What we are seeing is Copilot plus agents disrupting business applications, and we are leaning into this. With Dynamics 365, we took share as organizations like Ecolab, Lenovo, RTX, TotalEnergies and Wyzant switched to our AI-powered apps from legacy providers…

…[Question] Great to hear about the strength you’re seeing in Copilot… Would love to get some color on just the common use cases that you’re seeing that give you that confidence that, that will ramp into monetization later.

[Answer] I think the initial sort of set of seats were for places where there’s more belief in immediate productivity, a sales team, in finance or in supply chain where there is a lot of, like, for example, SharePoint grounded data that you want to be able to use in conjunction with web data and have it produce results that are beneficial. But then what’s happening very much like what we have seen in these previous generation productivity things is that people collaborate across functions, across roles, right? For example, even in my own daily habit, it’s I go to chat, I use Work tab and get results, and then I immediately share using Pages with colleagues. I sort of call it think with AI and work with people. And that pattern then requires you to make it more of a standard issue across the enterprise. And so that’s what we’re seeing.

Azure grew revenue by 31% in 2024 Q4 (was 33% in 2024 Q3), with 13 points of growth from AI services (was 12 points in 2024 Q3); Azure AI services was up 157% year-on-year, with demand continuing to be higher than capacity;  Azure’s non-AI business had weaker-than-expected growth because of go-to-market execution challenges

Azure and other cloud services revenue grew 31%. Azure growth included 13 points from AI services, which grew 157% year-over-year, and was ahead of expectations even as demand continued to be higher than our available capacity. Growth in our non-AI services was slightly lower than expected due to go-to-market execution challenges, particularly with our customers that we primarily reach through our scale motions as we balance driving near-term non-AI consumption with AI growth.

For Azure’s expected growth of 31%-32% in 2025 Q1 (FY2025 Q3), management expects  contribution from AI services to grow from increased AI capacity coming online; management expects Azure’s non-AI services to still post healthy growth, but there are still impacts from execution challenges; management expects Azure to no longer be capacity-constrained by the end of FY2025 (2025 Q2); Azure’s capacity constraint has been in power and space

In Azure, we expect Q3 revenue growth to be between 31% and 32% in constant currency driven by strong demand for our portfolio of services. As we shared in October, the contribution from our AI services will grow from increased AI capacity coming online. In non-AI services, healthy growth continues, although we expect ongoing impact through H2 as we work to address the execution challenges noted earlier. And while we expect to be AI capacity constrained in Q3, by the end of FY ’25, we should be roughly in line with near-term demand given our significant capital investments…

…When I talk about being capacity constrained, it takes two things. You have to have space, which I generally call long-lived assets, right? That’s the infrastructure and the land and then you have to have kits. We’re continuing, and you’ve seen that’s why our spend has pivoted this way, to be in the long-lived investment. We have been short power and space. And so as you see those investments land that we’ve made over the past 3 years, we get closer to that balance by the end of this year.

More than half of Microsoft’s cloud and AI-related capex in 2024 Q4 (FY2025 Q2) are for long-lived assets that will support monetisation over the next 15 years and more, while the other half are for CPUs and GPUs; management expects Microsoft’s capex in 2025 Q1 (FY2025 Q3) and 2025 Q2 (FY2025 Q4) to be at similar levels as 2024 Q4 (FY2025 Q2); FY2026’s capex will grow at a lower rate than in FY2025; the mix of spend in FY2026 will shift to short-lived assets in FY2026; Microsoft’s long-lived infrastructure investments are fungible; the long-lived assets are land; the presence of Moore’s Law means that management does not want to invest too much capex in any one year because the hardware and software will become much better in just 1 year; management thinks Microsoft’s AI infrastructure should be continuously upgraded to take advantage of Moore’s Law; Microsoft’s AI capex growth going forward will be tagged to customer contract delivery; the fungibility of Microsoft’s AI infrastructure investments relates to not just inference (which is the primary use case), but also training, post training, and running the commercial cloud business

More than half of our cloud and AI-related spend was on long-lived assets that will support monetization over the next 15 years and beyond. The remaining cloud and AI spend was primarily for servers, both CPUs and GPUs, to serve customers based on demand signals, including our customer contracted backlog…

…Next, capital expenditures. We expect quarterly spend in Q3 and Q4 to remain at similar levels as our Q2 spend. In FY ’26, we expect to continue investing against strong demand signals, including customer contracted backlog we need to deliver against across the entirety of our Microsoft Cloud. However, the growth rate will be lower than FY ’25 and the mix of spend will begin to shift back to short-lived assets, which are more correlated to revenue growth. As a reminder, our long-lived infrastructure investments are fungible, enabling us to remain agile as we meet customer demand globally across our Microsoft Cloud, including AI workloads…

…When I talk about being capacity constrained, it takes two things. You have to have space, which I generally call long-lived assets, right? That’s the infrastructure and the land and then you have to have kits. We’re continuing, and you’ve seen that’s why our spend has pivoted this way, to be in the long-lived investment. We have been short power and space. And so as you see those investments land that we’ve made over the past 3 years, we get closer to that balance by the end of this year…

…You don’t want to buy too much of anything at one time because, in Moore’s Law, every year is going to give you 2x, your optimization is going to give you 10x. You want to continuously upgrade the fleet, modernize the fleet, age the fleet and, at the end of the day, have the right ratio of monetization and demand-driven monetization to what you think of as the training expense…

…I do think the way I want everyone to internalize it is that the CapEx growth is going through that cycle pivot, which is far more correlated to customer contract delivery, no matter who the end customer is…

…  the other thing that’s sometimes missing is when we say fungible, we mean not just the primary use, which we’ve always talked about, which is inference. But there is some training, post training, which is a key component. And then they’re just running the commercial cloud, which at every layer and every modern AI app that’s going to be built will be required. It will be required to be distributed, and it will be required to be global. And all of those things are really important because it then means you’re the most efficient. And so the investment you see us make in CapEx, you’re right, the front end has been this sort of infrastructure build that lets us really catch up not just on the AI infrastructure we needed, but think about that as the building itself, data centers, but also some of the catch-up we need to do on the commercial cloud side. And then you’ll see the pivot to more CPU and GPU. 

Microsoft’s management thinks DeepSeek had real innovations, but those are going to be commoditized and become broadly used; management thinks that innovations in AI that reduce the cost of inference will drive more consumption and more apps being developed, and make AI more ubiquitous, which are all positive forces for Microsoft

I think DeepSeek has had some real innovations. And that is some of the things that even OpenAI found in ’01. And so we are going to — obviously, now that all gets commoditized and it’s going to get broadly used. And the big beneficiaries of any software cycle like that is the customers, right? Because at the end of the day, if you think about it, right, what was the big lesson learned from client server to cloud? More people bought servers, except it was called cloud. And so when token prices fall, inference computing prices fall, that means people can consume more, and there will be more apps written. And it’s interesting to see that when I referenced these models that are pretty powerful, it’s unimaginable to think that here we are in sort of beginning of ’25, where on the PC, you can run a model that required pretty massive cloud infrastructure. So that type of optimization means AI will be much more ubiquitous. And so therefore, for a hyperscaler like us, a PC platform provider like us, this is all good news as far as I’m concerned.

Microsoft has been reducing prices of GPT models over the years through inference optimizations

We are working super hard on all the software optimizations, right? I mean, just not the software optimizations that come because of what DeepSeek has done, but all the work we have done to, for example, reduce the prices of GPT models over the years in partnership with OpenAI. In fact, we did a lot of the work on the inference optimizations on it, and that’s been key to driving, right?

Microsoft’s management is aware that launching a frontier AI model that is too expensive to serve is useless

One of the key things to note in AI is you just don’t launch the frontier model, but if it’s too expensive to serve, it’s no good, right? It won’t generate any demand.

Microsoft’s management is seeing many different AI models being used for any one application; management thinks that there will always be a combination of different models used in any one application

What you’re seeing is effectively lots of models that get used in any application, right? When you look underneath even a Copilot or a GitHub Copilot or what have you, you already see lots of many different models. You build models. You fine-tune models. You distill models. Some of them are models that you distill into an open source model. So there’s going to be a combination…

….There’s a temporality to it, right? What you start with as a given COGS profile doesn’t need to be the end because you continuously optimize for latency and COGS and putting in different models.

NVIDIA (NASDAQ: NVDA)

NVIDIA’s Data Center revenue again had incredibly strong growth in 2024 Q4, driven by demand for the Hopper GPU computing platform and the ramping of the Blackwell GPU platform 

In the fourth quarter, Data Center revenue of $35.6 billion was a record, up 16% sequentially and 93% year-on-year, as the Blackwell ramp commenced and Hopper 200 continued sequential growth. 

Blackwell’s sales exceeded management’s expectations and is the fastest product ramp in NVIDIA’s history; it is common for Blackwell clusters to start with 100,000 GPUs or more and NVIDIA has started shipping for multiple such clusters; management architected Blackwell for inference; Blackwell has 25x higher token throughput and 20x lower cost for AI reasoning models compared to the Hopper 100; Blackwell has a NVLink domain that handles the growing complexity of inference at scale; management is seeing great demand for Blackwell for inference, with many of the early GB200 (GB200 is based on the Blackwell family of GPUs) deployments earmarked for inference; management expects NVIDIA’s gross margin to decline slightly initially as the Blackwell family ramps, before rebounding; management expects a significant ramp of Blackwell in 2025 Q1; the Blackwell Ultra, the next generation of GPUs within the Blackwell family, is slated for introduction in 2025 H2; the system architecture between Blackwell and Blackwell Ultra is exactly the same

In Q4, Blackwell sales exceeded our expectations. We delivered $11 billion of Blackwell revenue to meet strong demand. This is the fastest product ramp in our company’s history, unprecedented in its speed and scale…

…With Blackwell, it will be common for these clusters to start with 100,000 GPUs or more. Shipments have already started for multiple infrastructures of this size…

…Blackwell was architected for reasoning AI inference. Blackwell supercharges reasoning AI models with up to 25x higher token throughput and 20x lower cost versus Hopper 100. Its revolutionary transformer engine is built for LLM and mixture of experts inference. And its NVLink domain delivers 14x the throughput of PCIe Gen 5, ensuring the response time, throughput and cost efficiency needed to tackle the growing complexity of inference at scale…

…Blackwell has great demand for inference. Many of the early GB200 deployments are earmarked for inference, a first for a new architecture…

…As Blackwell ramps, we expect gross margins to be in the low 70s. Initially, we are focused on expediting the manufacturing of Blackwell systems to meet strong customer demand as they race to build out Blackwell infrastructure. When fully ramped, we have many opportunities to improve the cost and gross margin will improve and return to the mid-70s, late this fiscal year…

…Continuing with its strong demand, we expect a significant ramp of Blackwell in Q1…

…Blackwell Ultra is second half…

…The next train is on an annual rhythm and Blackwell Ultra with new networking, new memories and of course, new processors, and all of that is coming online…

…This time between Blackwell and Blackwell Ultra, the system architecture is exactly the same. It’s a lot harder going from Hopper to Blackwell because we went from an NVLink 8 system to a NVLink 72-based system. So the chassis, the architecture of the system, the hardware, the power delivery, all of that had to change. This was quite a challenging transition. But the next transition will slot right in. Blackwell Ultra will slot right in.

NVIDIA’s management sees post-training and model customisation has demanding orders of magnitude more compute than pre-training

The scale of post-training and model customization is massive and can collectively demand orders of magnitude, more compute than pretraining.

NVIDIA’s management is seeing accelerating demand for NVIDIA GPUs for inference, driven by test-time scaling and new reasoning models; management thinks reasoning models require 100x more compute per task than one-shot inference models; management is hopeful that future generation of reasoning models will require millions of times more compute; management is seeing that the vast majority of NVIDIA’s compute today is inference

Our inference demand is accelerating, driven by test-time scaling and new reasoning models like OpenAI o3, DeepSeek-R1 and Grok 3. Long thinking reasoning AI can require 100x more compute per task compared to one-shot inferences…

…. The amount of tokens generated, the amount of inference compute needed is already 100x more than the one-shot examples and the one-shot capabilities of large language models in the beginning. And that’s just the beginning. This is just the beginning. The idea that the next generation could have thousands times and even hopefully, extremely thoughtful and simulation-based and search-based models that could be hundreds of thousands, millions of times more compute than today is in our future…

……The vast majority of our compute today is actually inference and Blackwell takes all of that to a new level.

Companies such as ServiceNow, Perplexity, Microsoft, and Meta are using NVIDIA’s software and GPUs to achieve lower costs and/or better performance with their inference workloads

ServiceNow tripled inference throughput and cut costs by 66% using NVIDIA TensorRT for its screenshot feature. Perplexity sees 435 million monthly queries and reduced its inference costs 3x with NVIDIA Triton Inference Server and TensorRT-LLM. Microsoft Bing achieved a 5x speed up at major TCO savings for Visual Search across billions of images with NVIDIA TensorRT and acceleration libraries…

…Meta’s cutting-edge Andromeda advertising engine runs on NVIDIA’s Grace Hopper Superchip serving vast quantities of ads across Instagram, Facebook applications. Andromeda harnesses Grace Hopper’s fast interconnect and large memory to boost inference throughput by 3x, enhanced ad personalization and deliver meaningful jumps in monetization and ROI.

NVIDIA has driven a 200x reduction in inference costs in the last 2 years

We’re driven to a 200x reduction in inference costs in just the last 2 years.

Large cloud service providers (CSPs) were half of NVIDIA’s Data Centre revenue in 2024 Q4, and up nearly 2x year-on-year; large CSPs were the first to stand up Blackwell systems

In Q4, large CSPs represented about half of our data center revenue, and these sales increased nearly 2x year-on-year. Large CSPs were some of the first to stand up Blackwell with Azure, GCP, AWS and OCI bringing GB200 systems to cloud regions around the world to meet surging customer demand for AI. 

Regional clouds increased as a percentage of NVIDIA’s Data Center revenue in 2024 Q4, driven by AI data center build outs globally; management is seeing countries across the world building AI ecosystems

Regional cloud hosting NVIDIA GPUs increased as a percentage of data center revenue, reflecting continued AI factory build-outs globally and rapidly rising demand for AI reasoning models and agents where we’ve launched a 100,000 GB200 cluster-based incidents with NVLink Switch and Quantum-2 InfiniBand…

…Countries across the globe are building their AI ecosystems and demand for compute infrastructure is surging. France’s EUR 200 billion AI investment and the EU’s EUR 200 billion InvestAI initiatives offer a glimpse into the build-out to set redefined global AI infrastructure in the coming years.

NVIDIA’s revenue from consumer internet companies tripled year-on-year in 2024 Q4

Consumer Internet revenue grew 3x year-on-year, driven by an expanding set of generative AI and deep learning use cases. These include recommender systems, vision-language understanding, synthetic data generation, search and agentic AI.

NVIDIA’s revenue from enterprises nearly doubled year-on-year in 2024 Q4, partly with the help of agentic AI demand

Enterprise revenue increased nearly 2x year on accelerating demand for model fine-tuning, RAG and agentic AI workflows and GPU accelerated data processing.

NVIDIA’s management has introduced NIMs (NVIDIA Inference Microservices) focused on AI agents and leading AI agent platform providers are using these tools

We introduced NVIDIA Llama Nemotron model family NIMs to help developers create and deploy AI agents across a range of applications, including customer support, fraud detection and product supply chain and inventory management. Leading AI agent platform providers, including SAP and ServiceNow are among the first to use new models.

Healthcare companies are using NVIDIA’s AI products to power healthcare innovation

Health care leaders, IQVIA, Illumina and Mayo Clinic as well as ARC Institute are using NVIDIA AI to speed drug discovery, enhance genomic research and pioneer advanced health care services with generative and agentic AI.

Hyundai will be using NVIDIA’s technologies for the development of AVs (autonomous vehicles); NVIDIA’s automotive revenue had strong growth year-on-year and sequentially in 2024 Q4, driven by ramp in AVs; automotive companies such as Toyota, Aurora, and Continental are working with NVIDIA to deploy AV technologies; NVIDIA’s AV platform has passed 2 of the automotive industry’s foremost authorities for safety and cybersecurity

 At CES, Hyundai Motor Group announced it is adopting NVIDIA technologies to accelerate AV and robotics development and smart factory initiatives…

…Now moving to Automotive. Revenue was a record $570 million, up 27% sequentially and up 103% year-on-year…

…Strong growth was driven by the continued ramp in autonomous vehicles, including cars and robotaxis. At CES, we announced Toyota, the world’s largest auto maker will build its next-generation vehicles on NVIDIA Orin running the safety certified NVIDIA DriveOS. We announced Aurora and Continental will deploy driverless trucks at scale powered by NVIDIA DRIVE Thor. Finally, our end-to-end autonomous vehicle platform NVIDIA DRIVE Hyperion has passed industry safety assessments like TÜV SÜD and TÜV Rheinland, 2 of the industry’s foremost authorities for automotive-grade safety and cybersecurity. NVIDIA is the first AV platform to receive a comprehensive set of third-party assessments.

NVIDIA’s management has introduced the NVIDIA Cosmos World Foundation Model platform for the continued development of autonomous robots; Uber is one of the first major technology companies to adopt the NVIDIA Cosmos World Foundation Model platform

At CES, we announced the NVIDIA Cosmos World Foundation Model Platform. Just as language, foundation models have revolutionized language AI, Cosmos is a physical AI to revolutionize robotics. Leading robotics and automotive companies, including ridesharing giant Uber, are among the first to adopt the platform.

As a percentage of total Data Center revenue, NVIDIA’s Data Center revenue in China is well below levels seen prior to the US government’s export controls; management expects the Chinese market to be very competitive

Now as a percentage of total data center revenue, data center sales in China remained well below levels seen on the onset of export controls. Absent any change in regulations, we believe that China shipments will remain roughly at the current percentage. The market in China for data center solutions remains very competitive.

NVIDIA’s networking revenue declined sequentially in 2024 Q4, but the networking-attach-rate to GPUs remains robust at 75%; NVIDIA is transitioning to NVLink 72 with Spectrum-X (Spectrum-X is NVIDIA’s Ethernet networking solution); management expects networking revenue to resume growing in 2025 Q1; management sees AI requiring a new class of networking, for which the company’s NVLink, Quantum Infiniband, and Spectrum-X networking solutions are able to provide; large AI data centers, including OpenAI’s Stargate project, will be using Spectrum X

Networking revenue declined 3% sequentially. Our networking attached to GPU compute systems is robust at over 75%. We are transitioning from small NVLink 8 with InfiniBand to large NVLink 72 with Spectrum-X. Spectrum-X and NVLink Switch revenue increased and represents a major new growth vector. We expect networking to return to growth in Q1. AI requires a new class of networking. NVIDIA offers NVLink Switch systems for scale-up compute. For scale out, we offer Quantum InfiniBand for HPC supercomputers and Spectrum-X for Ethernet environments. Spectrum-X enhances the Ethernet for AI computing and has been a huge success. Microsoft Azure, OCI, CoreWeave and others are building large AI factories with Spectrum-X. The first Stargate data centers will use Spectrum-X. Yesterday, Cisco announced integrating Spectrum-X into their networking portfolio to help enterprises build AI infrastructure. With its large enterprise footprint and global reach, Cisco will bring NVIDIA Ethernet to every industry.

NVIDIA’s management is seeing 3 scaling laws at play in the development of AI models, namely pre-training scaling, post-training scaling, and test-time compute scaling

There are now multiple scaling laws. There’s the pre-training scaling law, and that’s going to continue to scale because we have multimodality, we have data that came from reasoning that are now used to do pretraining. And then the second is post-training scaling law, using reinforcement learning human feedback, reinforcement learning AI feedback, reinforcement learning, verifiable rewards. The amount of computation you use for post training is actually higher than pretraining. And it’s kind of sensible in the sense that you could, while you’re using reinforcement learning, generate an enormous amount of synthetic data or synthetically generated tokens. AI models are basically generating tokens to train AI models. And that’s post-training. And the third part, this is the part that you mentioned is test-time compute or reasoning, long thinking, inference scaling. They’re all basically the same ideas. And there you have a chain of thought, you’ve search.

NVIDIA’s management thinks the popularity of NVIDIA’s GPUs stems from its fungibility across all kinds of AI model architectures and use cases; NVIDIA’s management thinks that NVIDIA GPUs have an advantage over the ASIC (application-specific integrated circuit) AI chips developed by others because of (1) the general-purpose nature of NVIDIA GPUs, (2) NVIDIA’s rapid product development roadmap, (3) the software stack developed for NVIDIA GPUs that is incredibly hard to replicate

The question is how do you design such an architecture? Some of it — some of the models are auto regressive. Some of the models are diffusion-based. Some of it — some of the times you want your data center to have disaggregated inference. Sometimes it is compacted. And so it’s hard to figure out what is the best configuration of a data center, which is the reason why NVIDIA’s architecture is so popular. We run every model. We are great at training…

…When you have a data center that allows you to configure and use your data center based on are you doing more pretraining now, post training now or scaling out your inference, our architecture is fungible and easy to use in all of those different ways. And so we’re seeing, in fact, much, much more concentration of a unified architecture than ever before…

…[Question] We heard a lot about custom ASICs. Can you kind of speak to the balance between custom ASIC and merchant GPU?

[Answer] We build very different things than ASICs, in some ways, completely different in some areas we intercept. We’re different in several ways. One, NVIDIA’S architecture is general whether you’re — you’ve optimized for auto regressive models or diffusion-based models or vision-based models or multimodal models or text models. We’re great in all of it. We’re great at all of it because our software stack is so — our architecture is flexible, our software stack ecosystem is so rich that we’re the initial target of most exciting innovations and algorithms. And so by definition, we’re much, much more general than narrow…

…The third thing I would say is that our performance and our rhythm is so incredibly fast. Remember that these data centers are always fixed in size. They’re fixed in size or they’re fixed in power. And if our performance per watt is anywhere from 2x to 4x to 8x, which is not unusual, it translates directly to revenues. And so if you have a 100-megawatt data center, if the performance or the throughput in that 100-megawatt or the gigawatt data center is 4x or 8x higher, your revenues for that gigawatt data center is 8x higher. And the reason that is so different than data centers of the past is because AI factories are directly monetizable through its tokens generated. And so the token throughput of our architecture being so incredibly fast is just incredibly valuable to all of the companies that are building these things for revenue generation reasons and capturing the fast ROI…

…The last thing that I would say is the software stack is incredibly hard. Building an ASIC is no different than what we do. We build a new architecture. And the ecosystem that sits on top of our architecture is 10x more complex today than it was 2 years ago. And that’s fairly obvious because the amount of software that the world is building on top of architecture is growing exponentially and AI is advancing very quickly. So bringing that whole ecosystem on top of multiple chips is hard.

NVIDIA’s management thinks that only consumer AI and search currently have well-developed AI use cases, and the next wave will be agentic AI, robotics, and sovereign AI

We’ve really only tapped consumer AI and search and some amount of consumer generative AI, advertising, recommenders, kind of the early days of software. The next wave is coming, agentic AI for enterprise, physical AI for robotics and sovereign AI as different regions build out their AI for their own ecosystems. And so each one of these are barely off the ground, and we can see them.

NVIDIA’s management sees the upcoming Rubin family of GPUs as being a big step-up from the Blackwell family

The next transition will slot right in. Blackwell Ultra will slot right in. We’ve also already revealed and been working very closely with all of our partners on the click after that. And the click after that is called Vera Rubin and all of our partners are getting up to speed on the transition of that and so preparing for that transition. And again, we’re going to provide a big, huge step-up.

NVIDIA’s management sees AI as having the opportunity to address a larger part of the world’s GDP than any other technology has ever had

No technology has ever had the opportunity to address a larger part of the world’s GDP than AI. No software tool ever has. And so this is now a software tool that can address a much larger part of the world’s GDP more than any time in history.

NVIDIA’s management sees customers still actively using older families of NVIDIA GPUs because of the high level of programmability that CUDA has

People are still using Voltas and Pascals and Amperes. And the reason for that is because there are always things that — because CUDA is so programmable you could use it — one of the major use cases right now is data processing and data curation. You find a circumstance that an AI model is not very good at. You present that circumstance to a vision language model, let’s say, it’s a car. You present that circumstance to a vision language model. The vision language model actually looks at the circumstances and said, “This is what happened and I wasn’t very good at it.” You then take that response — the prompt and you go and prompt an AI model to go find in your whole lake of data, other circumstances like that, whatever that circumstance was. And then you use an AI to do domain randomization and generate a whole bunch of other examples. And then from that, you can go train the model. And so you could use the Amperes to go and do data processing and data curation and machine learning-based search. And then you create the training data set, which you then present to your Hopper systems for training. And so each one of these architectures are completely — they’re all CUDA-compatible and so everything runs on everything. But if you have infrastructure in place, then you can put the less intensive workloads onto the installed base of the past. All of our GPUs are very well employed.

Paycom Software (NYSE: PAYC)

Paycom’s management rolled out an AI agent six months ago to its service team, and has seen higher immediate response rates to clients and eliminated service tickets by 25% from a year ago; Paycom’s AI agent is driving internal efficiencies, higher client satisfaction, and higher Net Promoter Scores

Paycom’s AI agent, which was rolled out to our service team 6 months ago, utilizes our own knowledge-based semantic search model to provide faster responses and help our clients more quickly and consistently than ever before. As responses continuously improve over time, our client interactions become more valuable, and we connect them faster to the right solution. As a result, we are seeing improved immediate response rates and have eliminated service tickets by over 25% compared to a year ago…

…With automations like AI agent, we are realizing internal efficiencies, driving increasing client satisfaction and seeing higher Net Promoter Scores.

PayPal (NASDAQ: PYPL)

One of the focus areas for PayPal’s management in 2025 will be on raising efficiency with the help of AI 

Fourth is efficiency and effectiveness. In 2024, we reduced headcount by 10%. We made deliberate investments in AI and automation, which are critical to our future. This year, we are prioritizing the use of AI to improve the customer experience and drive efficiency and effectiveness within PayPal.

PayPal’s management sees AI being a huge opportunity for PayPal given the company’s volume of data; PayPal is using AI on its customer facing side to more efficiently process customer support cases and interactions with customers (PayPal Assistant has been rolled out and it has cut down phone calls and active events for PayPal); PayPal is using AI to personalise the commerce journey for consumers; PayPal is also using AI for back-office productivity and risk decisions

[Question] The ability to use AI for more operating efficiency. And are those initiatives that are requiring some incremental investment near term? Or are you already seeing sort of a positive ROI from that?

[Answer] AI is opening up a huge opportunity for us. First, at our scale, we saw 26 billion transactions on our platform last year. We have a massive data set that we are actively working and investing in to be able to drive our effectiveness and efficiency…

First, on the customer-facing side, we’re leveraging AI to really become more efficient in our support cases and how we interact with our customers. We see tens of millions of support cases every year, and we’ve rolled out our PayPal Assistant, which is now really cutting down phone calls and active events that we have. 

We also are leveraging AI to personalize the commerce journey, and so working with our merchants to be able to understand and create this really magical experience for consumers. When they show up at a checkout, it’s not just a static button anymore. This really can become a dynamic, personalized button that starts to understand the profile of the consumer, the journey that they’ve been on, perhaps across merchants, and be able to enable a reward or a cash-back offer in the moment or even a Buy Now, Pay Later offer in a dynamic experience…

In addition, we also are looking at our back office and ensuring that not just on the engineering and employee productivity side, but also in things like our risk decisions. We see billions and billions of risk decisions that often, to be honest, we’re very manual in the past. We’re now leveraging AI to be able to understand globally what are the nature of these risk decisions and how do we automate these across both risk models as well as even just ensuring that customers get the right response at the right time in an automated fashion.

Salesforce (NYSE: CRM)

Salesforce ended 2024 (FY2025) with $900 million in Data Cloud and AI ARR (annual recurring revenue), up 120% from a year ago; management has never seen products grow at this rate before, especially Agentforce

We ended this year with $900 million in Data Cloud and AI ARR. It grew 120% year-over-year. We’ve never seen products grow at these levels, especially Agentforce.

Salesforce’s management thinks building digital labour (AI agents) is a much bigger market than just building software

I’m sure you saw those ARK slides that got released over the weekend where she said that she thought this digital labor revolution, which is really like kind of what we’re in here now, this digital labor revolution, this looks like it’s anywhere from a few trillion to $12 trillion. I mean, I kind of agree with her. I think this is much, much bigger than software. I mean, for the last 25 years, we’ve been doing software to help our customers manage their data. That’s very exciting. I think building software that kind of prints and deploys digital workers is more exciting.

Salesforce’s unified platform, under one piece of code, combining customer data and an agentic platform, is what gives Agentforce its accuracy; Agentforce already has 3,000 paying customers just 90 days after going live; management thinks Agentforce is unique in the agentic capabilities it is delivering; Salesforce is Customer Zero for Agentforce; Agentforce has already resolved 380,000 service requests for Salesforce, with an 84% resolution rate, and just 2% of requests require human escalation; Agentforce has accelerated Salesforce’s sales-quoting cycles by 75% and increased AE (account executive) capacity by 7%; Agentforce is helping Salesforce engage more than 50 leads per day, freeing up the sales team for higher-value conversations; management wants every Salesforce customer to be using Agentforce; Data Cloud is at the heart of Agentforce; management is seeing customers across every industry deploying Agentforce; management thinks Salesforce’s agentic technology works better than many other providers, and that other providers are just whitewashing their technology with the “agent” label; Agentforce is driving growth across Salesforce’s portfolio; Salesfroce has prebuilt 170 specialised Agentforce industry skills; Agentforce’s 3,000 customers come from a diverse set of industries

Our formula now really for our customers is this idea that we have these incredible Customer 360 apps. We have this incredible Data Cloud, and this incredible agentic platform. These are the 3 layers. But that it is a deeply unified platform, it’s a deeply unified platform, it’s just one piece of code, that’s what makes it so unique in this market…

…It’s this idea that it’s a deeply unified platform with one piece of code all wrapped in a beautiful layer of trust. And that’s what gives Agentforce this incredible accuracy that we’re seeing…

…Just 90 days after it went live, we’ve already have 3,000 paying Agentforce customers who are experiencing unprecedented levels of productivity, efficiency and cost and cost savings. No one else is delivering at this level of capability…

…We’re seeing some amazing results on Salesforce as Customer Zero for Agentforce. Our digital labor force is resolving tens of thousands of customer service inquiries, freeing our human employees to focus on the most nuanced issues and customer relationships. We’re seeing tremendous momentum and success stories emerge as we execute our vision to make every company, every single company, every customer of ours, an Agentforce company, that is, we want every customer to be an Agentforce customer…

…We also continued phenomenal growth with Data Cloud this year, which is the heart of Agentforce. Data Cloud is the fuel that powers Agentforce and our customers are investing in it…

…We’re seeing customers deploy Agentforce across every industry…

…You got to be aware of the false agent because the false agent is out there where people can use the word agent or they kind of — they’re trying to whitewash all the agent, the thing, everywhere. But the reality is there is the real agents and there are the false agents, and we’re very fortunate to have the real stuff going on here. So we’ve got a lot more groundbreaking AI innovation coming…

…Today, we’re live on Agentforce across service and sales, our business technology organization, customer support and more. And the results are phenomenal. Since launching on our Salesforce help portal in October, Agentforce has autonomously handled 380,000 service requests, achieving an incredible 84% resolution rate and only 2% of the requests require human escalation. And we’re using Agentforce for quoting, accelerating our quoting cycles by more than 75%. In Q4, we increased our AE [account executive] capacity while still driving productivity up 7% year-over-year. Agentforce is transforming how we do outbound prospecting, already engaging more than 50 leads per day with personalized outreach and timely follow-ups, freeing up our teams to focus on high-value conversation. Our reps are participating in thousands of sales coaching training sessions each month…

…Agentforce is revolutionizing how our customers work by bringing AI-powered insights and actions directly into the workflows across the Customer 360 applications. This is driving strong growth across our portfolio. Sales Cloud and Service Cloud both achieved double-digit growth again in Q4. We’re seeing fantastic momentum with Slack, with customers like ZoomInfo, Remarkable and MIMIT Health using Agentforce and Slack to boost productivity…

…We’ve prebuilt over 170 specialized Agentforce industry skills and a team of 400 specialists, supporting transformations across sectors and geographies…

…We closed more than 3,000 paid Agentforce deals in the quarter. As customers continue to harness the value of AI deeply embedded across our unified platform, it is no surprise that these customers average nearly 4 clouds. And these customers came from a diverse set of industries with more than half in technology, manufacturing, financial services and HLS.

Lennar, the USA’s largest homebuilder, has been a Salesforce customer for 8 years and it is deploying Agentforce to fulfill their management’s vision of selling all kinds of new products; jewelry company, Pandora, an existing Salesforce customer, is deploying Agentforce with the aim of handling 30%-60% of its service cases with Agentforce; pharmaceutical giant Pfizer is using Agentforce to augment its sales teams; Singapore-based airline, Singapore Airlines, is now a customer of Agentforce and wants to deliver service through it; Goodyear is using Agentforce to automate and increase the effectiveness of its sales efforts; Accenture is using Agentforce to coach its sales team and expects to achieve higher win rates; Deloitte is using Agentforce and expects to achieve significant productivity gains

We’ve been working with Lennar, the nation’s largest homebuilder. And most of you know Lennar is really an incredible company, and they’ve been a customer of ours for about 8 years…

…You probably know Stuart Miller, Jon Jaffe, amazing CEOs. And those co-CEOs called me and said, “Listen, these guys have done a hackathon around Agentforce. We’ve got 5 use cases. We see incredible opportunities on our margin, incredible opportunities in our revenue. And do you have our back if we’re going to deploy this?” And we said, “Absolutely. We’ve deployed it ourselves,” which is the best evidence that this is real. And they are just incredible, their vision as a homebuilder providing 24/7 support, sales leads through all their digital channels. They’re able to sell all kinds of new products. I think they’re going to sell mortgages and insurance and all kinds of things to their customers. And the cool thing is they’re using our sales product, our service product, marketing, MuleSoft, Slack, Tableau, they use everything. But they are able to leverage it all together by realizing that just by turning it on, they get this incredible Agentforce capability…

…I don’t know how many of you know about Pandora. If you’ve been to a shopping center, you will see the Pandora store. You walk in, they have this gorgeous jewelry. They have these cool charm bracelets. They have amazing products. And if you know their CEO, Alex, he’s absolutely phenomenal…

…They’re in 100 countries. They employ 37,000 people worldwide. And Alex has this great vision to augment their employees with digital labor. And this idea that whether you’re on their website or in their store, or whatever it is, that they’re going to be able to do so much more with Agentforce. They already use — first of all, they already use Commerce Cloud. So if you’ve been to pandora.com and bought their products — and if you have it, by the way, it’s completely worthwhile. It’s great. And you can experience our Commerce Cloud, but it’s deeply integrated with our Service Cloud, with Data Cloud. It’s the one unified platform approach. And now they’re just flipping the switch, turning agents on, and they’re planning to deliver 30% to 60% of their service cases with Agentforce. That is awesome. And I really love Alex’s vision of what’s possible….

…The last customer I really want to hit on, which I’m so excited about, is Pfizer. And Albert is an incredible CEO. They are doing unbelievable things. They’ve been a tremendous customer. But now they’re really going all in on our Life Sciences Cloud…

…And with Agentforce, sales agents, for example, with Pfizer, that’s — they’ve got 20,000 customer-facing employees and customer-facing folks. That is just a radical extension for them with agents…

…I’m sure a lot of you — like, I have flown in Singapore Air. You know what? It’s a great airline. The CEO, Goh, is amazing. And he has a huge vision that also came out of Dreamforce, where — they’ve already delivered probably the best service of any airline in the world — they want to deliver it through agents. So whether you’re doing it with service or sales or marketing or commerce or all the different things that Singapore Air is doing with us, you’re going to be able to do this right on Singapore Air…

…Goodyear is partnering with us on their transformation, using Agentforce to automate and increase the effectiveness of their sales efforts. With Agentforce for Field Service, Goodyear will be able to reduce repair time by assisting technicians with answers to vehicle-related questions and autonomously scheduling field tech appointments…

…Accenture is using Agentforce Sales Coach, which provides personalized coaching and recommendations for sales teams, which is expected to lead to higher win rates. And Deloitte is projecting significant productivity gains and saved workforce hours as they roll out Agentforce over the next few years.

Salesforce’s management expects modest revenue contribution from Agentforce in 2025 (FY2026); contribution from Agentforce is expected to be more meaningful in 2026 (FY2027)

Starting with full fiscal year ’26. We expect revenue of $40.5 billion to $40.9 billion, growth of approximately 7% to 8% year-over-year in nominal and constant currency. And for subscription and support revenue, we expect growth of approximately 9% year-over-year in constant currency…

…On Agentforce, we are incredibly excited about the customer momentum we are seeing. However, the adoption cycle is still early as we focus on deployment with our customers. As a result, we are assuming a modest contribution to revenue in fiscal ’26. We expect the momentum to build throughout the year, driving a more meaningful contribution in fiscal ’27.

Salesforce has long had a mix of per-seat and consumption pricing models; for now, Agentforce is a consumption product, but management sees Agentforce evolving to a mix of per-seat and consumption pricing models; there was a customer that bought Agentforce in 2024 Q4 (FY2025 Q4) along with other Salesforce products and the customer signed a $7 million Agentforce contract and a $13 million contract for the other products; based on early days of engagement with Agentforce customers, management sees significant future upside to Salesforce’s pricing structure; Agentforce’s pricing will also take into account whether Agentforce will bring other human-based clouds into the customer; Agentforce is currently creating some halo around Salesforce’s other products

We’ve kind of started the company out with the per user pricing model, and that’s about humans. We price per human, so you’re kind of pricing per human. And then we have products, though, that are also in the consumption world as well. And of course, those started in the early days, things like our sandboxes, even things like our Commerce Cloud, even our e-mail marketing product, our Marketing Cloud. These are consumption-based products we’ve had for years…

…Now we have these kind of products that are for agents also, and agents are also a consumption model. So when we look at our Data Cloud, for example, that’s a consumption product. Agentforce is a consumption product. But it’s going to be a mix. It’s going to be a mix between what’s going on with our customers with how many humans do they have and then how many agents are they deploying…

…In the quarter, we did a large transaction with a large telecommunications company… we’re rebuilding this telecommunications company. So it’s Sales Cloud, it’s Service Cloud, it’s Marketing Cloud. It’s all of our core clouds, but then also it’s Agentforce. And the Agentforce component, I think, was maybe $7 million in the transaction. So she was buying $7 million of Agentforce. She bought $13 million in our products for humans, and I think that was about $20 million in total…

…We will probably move into the near future from conversations as we price most of our initial deals to universal credit. It will allow our customers far more flexibility in the way they transact with us. But we see this as a significant upside to our pricing structures going forward. And that’s what we’ve seen in the early days with our engagement with customers…

…Here’s a transaction that you’re doing, let’s say, a customer comes in, they’re very interested in building an agentic layer on their company, is that bringing other human-based clouds along with it?…

…[Question] Is Agentforce having a bit of a halo effect around some of your other products, meaning, as we are on the journey to get more monetization from Agentforce, are you seeing pickups or at least higher activity levels in some of your other products?

[Answer] That’s exactly right. And we’re seeing it in the way that our customers are using our technology, new ideas, new workflows, new engagements. We talked about Lennar as an example, their ability to handle leads after hours that they weren’t able to get back to or respond to in a quick time frame are now able to touch and engage with those leads. And that, of course, flows into their Salesforce automation system. And so we are seeing this halo effect with our core technology. It is making every single one of our core apps better as they deliver intelligence, underpinning these applications.

Salesforce’s management sees the combination of apps, data, and agents as the winning combination in an AI-world; management disputes Microsoft’s narrative that software apps will become a dumb database layer in an AI-dominated world, because it is the combination of apps, data, and agents that is important

I don’t know any company that’s 100% agents. I don’t know of any company that doesn’t need automation for its humans. I don’t know of any company that doesn’t need a data cloud where it needs a consistent common data repository for all of its agents to gain their intelligence. And I don’t know of any company that’s not going to need an agentic layer. And that idea of having apps, data and agents, I think, is going to be the winning combination…

…[Question] As part of that shift to agentic technology, there’s been a lot of debate about the SaaS technology and the business model. The SaaS tech stack that you built and pioneered, how does that fit into the agentic world? Is there a risk that SaaS just becomes a CRUD database?

[Answer] I’ve heard that Microsoft narrative, too. So I watched the podcast you watched, and that’s a very interesting idea. Here’s how I look at it, which is, I believe there is kind of a holy trinity here of AI CRM, which is the apps, the data and the agents. And these three things have to kind of work together. And I kind of put my money where our mouth is where we kind of built it and we delivered it. And you can see the 380,000 conversations that we had as point of evidence here in the last 90 days on our service and with a very high resolution rate of 84%. You can go to help.salesforce.com, and you can see that today.

Now Microsoft has had Copilot available for, I think, about 2 years or more than 2 years. And I know that they’re the reseller of OpenAI and they’ve invested, they kind of repackaged this ChatGPT, whatever. But where on their side are they delivering agents? Where in their company have they done this? Are they a best practice? Because I think that while they can say such a thing, do they have humans and agents working together to create customer success? Are they rebalancing their workforce with humans and agents? I think that it’s a very interesting point that, yes, the agentic layer is very important, but it doesn’t operate by itself. It operates with data, with a Data Cloud that has to be federated through your company, to all your data sources. And humans, we’re still here

Salesforce’s management is seeing Agentforce deliver tremendous efficiency in Salesforce’s customer support function that they may rebalance some customer-support roles into other roles; management is currently seeing AI coding tools improve the productivity of Salesforce’s engineering team by 30% and thinks even more productivity can be found; management will not be expanding Salesforce’s engineering team this year, but will grow the sales team

We really are seeing tremendous efficiency with help.salesforce.com. So we may see the opportunity to rebalance some of those folks into sales and marketing and other functions…

…We definitely have seen a lot of efficiency with engineering and with some of the new tools that I’ve seen, especially some of these high-performance coding tools. One of the key members of my staff who’s here in the room with us has just showed me one of his new examples of what we’re able to do with these coding tools, pretty awesome. And we’re not going to hire any new engineers this year. We’re seeing 30% productivity increase on engineering. And we’re going to really continue to ride that up…

…We’re going to grow sales pretty dramatically this year. Brian has got a big vision for how to grow the sales organization. probably another 10% to 20%, I hope, this year because we’re seeing incredible levels of demand.

Salesforce’s management thinks that AI agents is one of the catalysts to drive GDP growth

So if you want productivity to go up and you want GDP to grow up and you want growth, I think that digital labor is going to be one of the catalysts to make that happen.

Shopify (NASDAQ: SHOP)

Shopify launched its first AI-powered search integration with Perplexity in 2024

Last year, we… launched our first AI-powered search integration with Perplexity, enabling new ways for buyers to find merchants.

One of Shopify’s management’s focus areas in 2025 is to continue embracing AI by investing more in Sidekick and other AI capabilities that help merchants launch and grow faster; management wants to shift Shopify towards producing goal-oriented software; management believes Shopify is well-positioned as a leader for commerce in an AI-driven world

We will continue to embrace the transformative potential of AI. This technology is not just a part of the future, it is redefining it. We’ve anticipated this. So we’re already transforming Shopify into a platform where users and machines work seamlessly together. We plan to deepen our investment in Sidekick and other AI capabilities to help not just brand-new merchants to launch, but also to help larger merchants scale faster and drive greater productivity. Our efforts to shift towards more goal-oriented software will further help to streamline operations and improve decision-making. This focus on embracing new ways of thinking and working positions us not only as the platform of choice today, but also as a leader for commerce in the AI-driven era with a relentless focus on cutting-edge technology.

Shopify’s management believes Shopify will be one of the major net beneficiaries in the AI era as the company is leveraging AI really well, such as its partnerships with Perplexity and OpenAI

I actually think Shopify will very much be one of the major net beneficiaries in this new AI era. I think we are widely recognized as one of the best companies that foster long-term partnership. And so when it comes to partnership in AI, whether it’s Perplexity, where we’re now powering their search results with incredible product across the Shopify product catalog or OpenAI where we’re using — we have a direct set of their APIs to help us internally, we are really leveraging it as best as we can.

In terms of utilising AI, Shopify’s management sees 2 angles; the 1st angle is Shopify using AI to help merchants with mundane tasks and allow merchants to focus only on the things they excel at; the 2nd angle is Shopify using AI internally to make developers and customer-support teams more effective (with customer-support teams, Shopify is using AI to handle low-quality conversations with customers)

[Question] A question in regards to AI and the use of AI internally. Over the last year or so, you’ve made significant investments. Where are you seeing it operationally having the most impact? And then what has been the magnitude of productivity gains that you’ve seen?

[Answer] We think about it in sort of 2 ways. The first is from a merchant perspective, how can we make our merchants way more successful, get them to do things faster, more effectively. So things like Sidekick or media editor or Shopify Inbox, Semantic Search, Sidekick, these are things that now — that every merchant should want when they’re not just getting started, but also scaling their business. And those are things that are only available from Shopify. So we’re trying to make some of the more mundane tasks far more easy to do and get them to focus on things that only they can — only the merchants can do. And I think that’s an important aspect of what Shopify will bring…

…Internally, however, this is where it gets really interesting, because not only can we use it to make our developers more effective, but also, if you think about our support organization, now we can ensure that our support team is actually having very high-quality conversations with merchants, whereas a lot of low-quality conversations, things like configuring a domain or a C name or a user name and password issue, that can be handled really elegantly by AI.

Taiwan Semiconductor Manufacturing Company (NYSE: TSM)

TSMC’s AI accelerators revenue more than tripled in 2024 and was mid-teens percent of overall revenue in 2024, but management expects AI accelerators revenue to double in 2025; management sees really strong AI-related demand in 2025

Revenue from AI accelerators, which we now define as AI GPU, AI ASICs and HBM controller for AI training and inference in the data center, accounted for close to mid-teens percent of our total revenue in 2024. Even after more than tripling in 2024, we forecast our revenue from AI accelerator to double in 2025 as the strong surge in AI-related demand continues…

…[Question] Try to get a bit more clarity on the cloud growth for 2025. I think, longer term, without a doubt, the technology definitely has lots of potential for demand opportunities, but I think — if we look at 2025 and 2026, I think there could be increasing uncertainties coming from maybe [indiscernible] spending, macro or even some of the supply chain challenges. And so I understand the management just provided a pretty good guidance for this year for sales to double. And so if you look at that number, do you think there is still more upside than downside as we go through 2025?

[Answer] I certainly hope there is upside, but I hope I get — my team can supply enough capacity to support it. Does that give you enough hint? 

TSMC’s management saw a mixed year of recovery for the global semiconductor industry in 2024 with strong AI-related demand but mild recovery in other areas

2024 was a mixed year of recovery for the global semiconductor industry. AI-related demand was strong, while other applications saw only a very mild recovery as macroeconomics condition weighed on consumer sentiment and the end market demand. 

TSMC’s management expects mid-40% revenue CAGR from AI accelerators in the 5-years starting from 2024 (previous forecast was for 50% CAGR for the 5-years starting from 2024, but off a lower base); management expects AI accelerators to be the strongest growth driver for TSMC’s overall HPC  platform and overall revenue over the next few years

Underpinned by our technology leadership and broader customer base, we now forecast the revenue growth from AI accelerators to approach a mid-40% CAGR for the 5-year period starting off the already higher base of 2024. We expect AI accelerators to be the strongest driver of our HPC platform growth and the largest contributor in terms of our overall incremental revenue growth in the next several years.

TSMC’s management expects 20% revenue CAGR in USD terms in the 5-years starting from 2024, driven by growth across all its platforms; management thinks that in the next few years, TSMC’s smartphone and PC end-markets will have higher silicon content and faster replacement cycle, driven by AI-related demand, which will in turn drive robust demand for TSMC’s chip manufacturing service; the AI-related demand in the smartphone and PC end-markets are related to edge-AI

Looking ahead, as the world’s most reliable and effective capacity provider, TSMC is playing a critical and integral role in the global semiconductor industry. With our technology leadership, manufacturing excellence and customer trust, we are well positioned to address the growth from the industry megatrend of 5G, AI and HPC with our differentiated technologies. For the 5-year period starting from 2024, we expect our long-term revenue growth to approach a 20% CAGR in U.S. dollar term, fueled by all 4 of our growth platform, which are smartphone, HPC, IoT and automotive…

…[Question] I believe that 20% starting from a very — already very high base in 2024 is a really good long-term objective but just wondering that, aside from the strong AI demand, what’s your view on the traditionals, applications like PC and the smartphone, growth and particularly for this year.

[Answer] This year is still mild growth for PC and smartphone, but everything is AI related, all right, so you can start to see why we have confidence to give you a close to 20% CAGR in the next 5 years. AI: You look at a smartphone. They will put AI functionality inside, and not only that. So the silicon content will be increased. In addition to that, actually the replacement cycle will be shortened. And also they need to go into the very advanced technology because of, if you want to put a lot of functionality inside a small chip, you need a much more advanced technology to put those [indiscernible]. Put all together, that even smartphone, the unit growth is almost low single digit, but then the silicon and the replacement cycle and the technology migration, that give us more growth than just unit growth; similar reason for PC…

…On the edge AI, in our observation, we found out that our customers start to put up more neural processing inside. And so we estimated the 5% to 10% more silicon being used. [ Can it be ] every year 5% to 10%? Definitely it is no, right? So they will move to next node, the technology migration. That’s also to TSMC’s advantage. Not only that, I also say that, the replacement cycle, I think it will be shortened because of, when you have a new toy that — with AI functionality inside it, everybody want replacing, replace their smartphone, replace their PCs. And [ I count that one ] much more than the — a mere 5% increase.

TSMC’s upcoming A16 process technology is best suited for specific HPC (high-performance computing) products, which means it is best suited for AI-related workloads

We will also introduce A16 featuring Super Power Rail or SPR as separate offering. TSMC’s SPR is a innovative, best-in-class backside power delivery solution that is first in the industry to incorporate a novel backside metal scheme that preserves gate density and device width flexibility to maximize product benefit. Compared with N2P, A16 provide a further 8% to 10% speed improvement at the same power or 15% to 20% power improvement at the same speed, and additional 7% to 10% chip density gain. A16 is the best suitable for specific HPC product with a complex signal route and dense power delivery network. Volume production is scheduled for second half 2026.

TSMC’s management thinks that the US government’s latest AI restrictions will only have a minimal impact on the company’s business

[Question] Overnight, the U.S. seems to put a new framework on restricting China’s AI business, right? So I’m wondering whether that will create some business impact to your China business.

[Answer] We don’t have all analysis yet, but the first look is not significantly. It’s manageable. So that meaning that, my customers who are being restricted [ or something ], we are applying for the special permit for them. And we believe that we have confidence that they will get some permission, so long as they are not in the AI area, okay, especially automotive industry. Or even you talk about crypto mining, yes.

TSMC’s management does not want to reveal the level of demand for AI-related ASICs (application-specific integrated circuits) from the cloud hyperscalers, but they are confident that the demand is real, and that the cloud hyperscalers will be working with TSMC as they all need leading-edge technology for their AI-related ASICs

[Question] Broadcom’s CEO recently laid out a large SAM for AI hyperscalers building out custom silicon. I think he was talking about million clusters from each of the customers he has in the next 2 or 3 years. What’s TSMC’s perspective on all this? 

[Answer] I’m not going to answer the question of the specific number, but let me assure you that, whether it’s ASIC or it’s graphic, they all need a very leading-edge technology. And they’re all working with TSMC, okay, so — and the second one is, is the demand real. Was — is — as a number that my customers said. I will say that the demand is very strong.

AI makes up all of the current demand for CoWoS (chip on wafer on substrate) capacity that TSMC’s management is seeing, but they think non-AI-related demand for CoWoS will come in the near future from CPUs and servers; there are rumours of a cut in orders for CoWoS, but management is not seeing any cuts; it appears that HBM (high bandwidth memory) is the key constraint on AI demand, instead of CoWoS; CoWoS was over 8% of TSMC’s revenue in 2024 and will be over 10% in 2025; CoWoS gross margin is better than before, but still lower than the corporate average

[Question] When can we see non-AI application such as server, smartphone or anything else can be — can start to adopt CoWoS capacity in case there is any fluctuation in the AI demand?

[Answer] Today is all AI focused. And we have a very tight capacity and cannot even meet customers’ need, but whether other products will adopt this kind of CoWoS approach, they will. It’s coming and we know that it’s coming.

[Question] When?

[Answer] It’s coming… On the CPU and on the server chip. Let me give you a hint…

…[Question] About your CoWoS and SoIC capacity ramp. Can you give us more color this year? Because recently there seemed to be a lot of market noises. Some add orders. Some cut orders, so I would like to see your view on the CoWoS ramp.

[Answer] That’s a rumor. I assure you. We are working very hard to meet the requirement of my customers’ demand, so “cut the order,” that won’t happen. We actually continue to increase, so we are — again I will say that. We are working very hard to increase the capacity…

…[Question] A question on AI demand. Is there a scenario where HBM is more of a constraint on the demand, rather than CoWoS which seems to be the biggest constraint at the moment? 

[Answer] I don’t comment on other supplier, but I know that we have a very tight capacity to support the AI demand. I don’t want to say I’m the bottleneck. TSMC, always working very hard with customer to meet their requirement…

…[Question] So we have observed an increasing margin of advanced packaging. Could you remind us the CoWoS contribution of last year? And do you expect the margin to kind of approach the corporate average or even exceed it after the so-called — the value reflection this year?

[Answer] Overall speaking, advanced packaging accounted for over 8% of revenue last year. And it will account for over 10% this year. In terms of gross margins, it is better. It is better than before but still below the corporate average. 

AI makes up all of the current demand for SoIC (system on integrated chips) that TSMC’s management is seeing, but they think non-AI-related demand for SoIC will come in the future

Today, SoIC’s demand is still focused on AI applications, okay? For PC or for other area, it’s coming but not right now.

Tesla (NASDAQ: TSLA)

Tesla’s management thinks Tesla’s FSD (Full Self Driving) technology has grown up a lot in the past few years; management thinks that car-use can grow from 10 hours per week to 55 hours per week with autonomous vehicles; autonomous vehicles can be used for both cargo and people delivery; FSD currently works very well in the USA , and will soon work well everywhere else; the constraint Tesla is currently experiencing with autonomous vehicles is in battery packs; FSD makes traffic commuting safer; FSD is currently on Version 13, and management believes Version 14 will have a significant step-improvement; Tesla has launched the Cortex training cluster at Gigafactory Austin, and it has played a big role in advancing FSD; Tesla will launch unsupervised FSD in June 2025 in Austin; Tesla already has thousands of its cars driving autonomously daily in its factories in Fremont and Texas, and Tesla will soon do that in Austin and elsewhere in the world; Tesla’s solution for autonomous vehicles is a generalised AI solution which does not need high-precision maps; Tesla’s unsupervised FSD work outside of Austin even when it’s launched only in June 2025 in Austin, but management just wants to be cautious; management thinks Tesla will release unsupervised FSD in many parts of the USA by end-2025; management’s safety-standard for FSD is for it to be far, far, far superior to humans; management thinks Tesla will have unsupervised FSD in almost every market this year

For a lot of people, like their experience of Tesla autonomy is like if it’s even a year old, if it’s even 2 years old, it’s like meeting someone when they’re like a toddler and thinking that they’re going to be a toddler forever. But obviously not going to be a toddler forever. They grow up. But if their last experience was like, “Oh, FSD was a toddler.” It’s like, well, it’s grown up now. Have you seen it? It’s like walks and talks…

…My #1 recommendation for anyone who doubts is simply try it. Have you tried it? When is the last time you tried it? And the only people who are skeptical, the only people who are skeptical are those who have not tried it.

So a car goes — a passenger car typically has only about 10 hours of utility per week out of 168, a very small percentage. Once that car is autonomous, my rough estimate is that it is in use for at least 1/3 of the hours per week, so call it, 50, maybe 55 hours of the week. . And it can be used for both cargo delivery and people delivery…

That same asset, the thing that — these things that already exist with no incremental cost change, just a software update, now have 5x or more the utility than they currently have. I think this will be the largest asset value increase in human history…

…So look, the reality of autonomy is upon us. And I repeat my advice, try driving the car or let it drive you. So now it works very well in the U.S., but of course, it will, over time, work just as well everywhere else…

…Our current constraint is battery packs this year but we’re working on addressing that constraint. And I think we will make progress in addressing that constraint…

…So a bit more on full self-driving. Our Q4 vehicle safety report shows continued year-over-year improvement in safety for vehicles. So the safety numbers, if somebody has supervised full self-driving turn on or not, the safety differences are gigantic…

…People have seen the immense improvement with version 13, and with incremental versions in version 13 and then version 14 is going to be yet another step beyond that, that is very significant. We launched the Cortex training cluster at Gigafactory Austin, which was a significant contributor to FSD advancement…

…We’re going to be launching unsupervised full self-driving as a paid service in Austin in June. So I talked to the team. We feel confident in being able to do an initial launch of unsupervised, no one in the car, full self-driving in Austin in June…

…We already have Teslas operating autonomously unsupervised full self-driving at our factory in Fremont, and we’ll soon be doing that at our factory in Texas. So thousands of cars every day are driving with no one in them at our Fremont factory in California, and we’ll soon be doing that in Austin and then elsewhere in the world with the rest of our factories, which is pretty cool. And the cars aren’t just driving to exactly the same spot because, obviously, it all — [ implied ] at the same spot. The cars are actually programmed with where — with what lane they need to park into to be picked up for delivery. So they drive from the factory end of line to their destination parking spot and to be picked up for delivery to customers and then doing this reliably every day, thousands of times a day. It’s pretty cool…

…Our solution is a generalized AI solution. It does not require high-precision maps of locality. So we just want to be cautious. It’s not that it doesn’t work beyond Austin. In fact, it does. We just want to be — put our toe in the water, make sure everything is okay, then put a few more toes in the water, then put a foot in the water with safety of the general public as and those in the car as our top priority…

…I think we will most likely release unsupervised FSD in many regions of the country of the U.S. by the end of this year…

…We’re looking for a safety level that is significantly above the average human driver. So it’s not anywhere like much safer, not like a little bit safer than human, way safer than human. So the standard has to be very high because the moment there’s any kind of accident with an autonomous car, that immediately gets worldwide headlines, even though about 40,000 people die every year in car accidents in the U.S., and most of them don’t even get mentioned anywhere. But if somebody [ scrapes a shed ] within autonomous car, it’s headline news…

…But I think we’ll have unsupervised FSD in almost every market this year, limited simply by regulatory issues, not technical capability. 

Tesla’s management thinks the compute needed for Optimus will be 10x that of autonomous vehicles, even though a humanoid robot has 1,000x more uses than an autonomous vehicle; management has seen the cost of training Optimus (or AI, in general) dropping dramatically over time; management thinks Optimus can produce $10 trillion in revenue, and that will still make the training needs of $500 billion in compute a good investment; management realises their revenue projections for Optimus sound insane, but they believe in it (sounds like a startup founder trying to get funding from VCs); it’s impossible for management to predict the exact timing for Optimus because everything about the robot has to be designed and built from the ground up by Tesla (nothing could be bought off-the-shelf by Tesla), but management thinks Tesla will build a few thousand Optimus robots by end-2025, and that these robots will be doing useful work in Tesla’s factories in the same timeframe; management’s goal is to ramp up Optimus production at a far faster rate than anything has ever been ramped; Optimus can even do delicate things such as play the piano; Optimus is still not design-locked for production; Tesla might be able to deliver Optimus to external clients by 2026 H2; management is confident that at scale, Optimus will be cheaper to produce than a car

The training needs for Optimus, our Optimus humanoid robot are probably at least ultimately 10x what is needed for the car, at least to get to the full range of useful role. You can say like how many different roles are there for a humanoid robot versus a car? A humanoid robot has probably 1,000x more uses and more complex things than in a car. That doesn’t mean the training scales by 1,000 but it’s probably 10x…

…It doesn’t mean like — or Tesla’s going to spend like $500 billion in training compute because we will obviously train Optimus to do enough tasks to match the output of Optimus robots. And obviously, the cost of training is dropping dramatically with time. But it is — it’s one of those things where I think long-term, Optimus will be — Optimus has the potential to be north of $10 trillion in revenue, like it’s really bananas. So that you can obviously afford a lot of training compute in that situation. In fact, even $500 billion training compute in that situation will be quite a good deal…

…With regard to Optimus, obviously, I’m making these revenue predictions that sound absolutely insane, I realize that. But they are — I think they will prove to be accurate…

…There’s a lot of uncertainty on the exact timing because it’s not like a train arriving at the station for Optimus. We are designing the train and the station and in real time while also building the tracks. And sort of like, why didn’t the train arrive exactly at 12:05? And like we’re literally designing the train and the tracks and the station in real-time while like how can we predict this thing with absolute precision? It’s impossible. The normal internal plan calls for roughly 10,000 Optimus robots to be built this year. Will we succeed in building 10,000 exactly by the end of December this year? Probably not. But will we succeed in making several thousand? Yes, I think we will. Will those several thousand Optimus robots be doing useful things by the end of the year? Yes, I’m confident they will do useful things…

…Our goal is to run Optimus production faster than maybe anything has ever been ramped, meaning like aspirationally in order of magnitude, ramp per year. Now if we aspire to an order of magnitude ramp per year, perhaps, we only end up with a half order of magnitude per year. But that’s the kind of growth that we’re talking about. It doesn’t take very many years before we’re making 100 million of these things a year, if you go up by let’s say, a factor by 5x per year…

This is an entirely new supply chain, it’s entirely new technology. There’s nothing off the shelf to use. We tried desperately with Optimus to use any existing motors, any actuators, sensors. Nothing worked for a humanoid robot at any price. We had to design everything from physics-first principles to work for humanoid robot and with the most sophisticated hand that has ever been made before by far. Optimus will be also able to play the piano and be able to thread a needle. I mean this is the level of precision no one has been able to achieve…

…Optimus is not design-locked. So when I say like we’re designing the train as it’s going — we’re redesigning the train as it’s going down the tracks while redesigning the tracks and the train stations…

…I think probably with version 2, it is a very rough guess because there’s so much uncertainty here, very rough guess that we start delivering Optimus robots to companies that are outside of Tesla in maybe the second half of next year, something like that…

I’m confident at 1 million units a year, that the production cost of Optimus will be less than $20,000. If you compare the complexity of Optimus to the complexity of a car, so just the total mass and complexity of Optimus is much less than a car.

The buildout of Cortex accelerated the rollout of FSD Version 13; Tesla has invested $5 billion so far in total AI-related capex

The build-out of Cortex was accelerated because of the role — actually accelerate the rollout of FSD Version 13. Our cumulative AI-related CapEx, including infrastructure, so far has been approximately $5 billion. 

Tesla’s management is seeing significant interest from some car manufacturers in licensing Tesla’s FSD technology; management thinks that car manufacturers without FSD technology will go bust; management will only entertain situations where the volume would be very high

What we’re seeing is at this point, significant interest from a number of major car companies about licensing Tesla full self-driving technology…

…We’re only going to entertain situations where the volume would be very high. Otherwise, it’s just not worth the complexity. And we will not burden our engineering team with laborious discussions with other engineering teams until we obviously have unsupervised full self-driving working throughout the United States. I think the interest level from other manufacturers to license FSD will be extremely high once it is obvious that unless you have FSD, you’re dead.

Compared to Version 13, Version 14 of FSD will have a larger model size, longer context length, more memory, more driving-context, and more data on tricky corner cases

[Question] What technical breakthroughs will define V14 of FSD, given that V13 already covered photon to control? 

[Answer] So continuing to scale the model size a lot. We scale a bunch in V13 but then there’s still room to grow. So we’re going to continue to scale the model size. We’re going to increase the context length even more. The memory sort of like limited right now. We’re going to increase the amount of memory [indiscernible] minutes of context for driving. They’re going to add audio and emergency vehicles better. Add like data of the tricky corner cases that we get from the entire fleet, any interventions or any kind of like user intervention. We just add to the data set. So scaling in basically every access, training compute, [ asset ] size, model size, model context and also all the reinforcement learning objectives.

Tesla has difficulties training AI models for autonomous vehicles in China because the country previously did not allow Tesla to transfer training videos outside of China while the US government did not allow Tesla to do training in China; a workaround Tesla did was to train on publicly available videos of streets in China; Tesla also had to build a simulator for its AI models to train on bus lanes in China because they are complicated 

In China, which is a gigantic market, we do have some challenges because they weren’t — they currently allow us to transfer training video outside of China. And then the U.S. government won’t let us do training in China. So we’re in a bit of a buying there. It’s like a bit of a quandary. So what we’re really solving then is by literally looking at videos of streets in China that are available on the Internet to understand and then feeding that into our video training so that publicly available video of street signs and traffic rules in China can be used for training and then also putting it in a very accurate simulator. And so it will train using SIM for bus lanes in China. Like bus lanes in China, by the way, one of our biggest challenges in making FSD work in China is the bus lanes are very complicated. And there’s like literally like hours of the day that you’re allowed to be there and not be there. And then if you accidently go in at a bus lane at the wrong time, you get an automatic ticket instantly. And so it was kind of a big deal, bus lanes in China. So we put that into our simulator train on that, the car has to know what time of the day it is, read the sign. We’ll get this solved.

Elon Musk knows LiDAR technology really well because he built a LiDAR system for SpaceX that is in-use at the moment, but he thinks LiDAR is simply the wrong technology to use for autonomous vehicles because it has issues, and because humans are driving vehicles simply with our eyes and our biological neural nets

[Question] You’ve said in the past about LiDAR, for EVs, that LiDAR is a crutch, a fool’s errand. I think you even told me once, even if it was free, you’d say you wouldn’t use it. Do you still feel that way?

[Answer] Obviously humans drive without shooting lasers out of their eyes. I mean unless you’re superman. But like humans drive just with passive visual — humans drive with eyes and a neural net — and a brain neural net, sort of biological — so the digital equivalent of eyes and a brain are cameras and digital neural nets or AI. So that’s — the entire road system was designed for passive optical neural nets. That’s how the whole real system was designed and what everyone is expecting, that’s how we expect other cars to behave. So therefore, that is very obviously a solution for full self-driving as a generalized — but the generalized solution for full self-driving as opposed to the very specific neighborhood-by-neighborhood solution, which is very difficult to maintain, which is what our competitors are doing…

…LiDAR has a lot of issues. Like SpaceX Dragon docks with the space station using LiDAR, that’s a program that I personally spearheaded. I don’t have some fundamental bizarre dislike of LiDAR. It’s simply the wrong solution for driving cars on roads…

…I literally designed and built our own red LiDAR. I oversaw the project, the engineering thing. It was my decision to use LiDAR on Dragon. And I oversaw that engineering project directly. So I’m like we literally designed and made a LiDAR to dock with the space station. If I thought it was the right solution for cars, I would do that, but it isn’t.

The Trade Desk (NASDAQ: TTD)

Trade Desk’s management continues to invest in AI and thinks that AI is game-changing for forecasting and insights on identity and measurement; Trade Desk’s AI efforts started in 2017 with Koa, but management sees much bigger opportunities today; management is asking every development team in Trade Desk to look for opportunities to introduce AI into Trade Desk’s platform; there are already hundreds of AI enhancements to Trade Desk’s platform that have been shipped or are going to be shipped in 2025

AI is providing next-level performance in targeting and optimization, but it is also particularly game-changing in forecasting and identity and measurement. We continue to look at our technology stack and ask, where can we inject AI and enhance our product and client outcomes? Over and over again, we are finding new opportunities to make AI investments…

…We started our ML and AI efforts in 2017 with the launch of Koa, but today, the opportunities are much bigger. We’re asking every scrum inside of our company to look for opportunities to inject AI into our platform. Hundreds of enhancements recently shipped and coming in 2025 would not be possible without AI. We must keep the pedal to the metal, not to chest them on stages, which everyone else seems to be doing, but instead to produce results and win share.

Wix (NASDAQ: WIX)

Wix’s AI Website Builder was launched in 2024 and has driven stronger conversion and purchase behaviour from users; more than 1 million sites have been created and published with AI Website Builder; most new Wix users today are creating their websites through Wix’s AI tools and AI Website Builder and these users have higher rates of converting to paid subscribers

2024 was also the year of AI innovation. In addition to the significant number of AI tools introduced, we notably launched our AI Website Builder, the new generation of our previous AI site builder introduced in 2016. The new AI Website Builder continues to drive demonstrably stronger conversion and purchase behavior…

…Over 1 million sites have been created and published with the Website Builder…

…Most new users today are creating their websites through our AI-powered onboarding process and Website Builder which is leading to a meaningful increase in conversion of free users to paid subscriptions, particularly among Self Creators.

Wix’s management launched Wix’s first directly monetised AI product – AI Site Chat – in December 2024; AI Site Chat will help businesses converse with customers round the clock; users of AI Site Chat have free limited access, with an option to pay for additional usage; AI Site Chat’s preliminary results look very promising

In December, we also rolled out our first directly monetized AI product – the AI Site Chat…

…The AI Site-Chat was launched mid-December to Wix users in English, providing businesses with the ability to connect with visitors 24/7, answer their questions, and provide relevant information in real time, even when business owners are unavailable. By enhancing availability and engagement on their websites, the feature empowers businesses to meet the needs of their customers around the clock, ultimately improving the customer experience and driving potential sales. Users have free limited access with the option to upgrade to premium plans for additional usage…

…So if you’re a Wix customer, you can now install a chat, AI-powered chat on your website, and this will handle customer requests, product inquiries and support request. And from — and again, it’s very early in days and the preliminary results, but it looks very promising. 

AI agents and assistants are an important part of management’s product roadmap for Wix in 2025; Wix is testing (1) an AI assistant for its Wix Business Manager dashboard, and (2) Marketing Agent, a directly monetizable AI agent that helps users accomplish marketing tasks; Marketing Agent is the first of a number of specialised AI agents management will roll out in 2025; management intends to test monetisation opportunities with the new AI agents

AI remains a major part of our 2025 product roadmap with particular focus on AI-powered agents and assistants…

… Currently, we are testing our AI Assistant within the Wix Business Manager as well as our AI Marketing Agent.

The AI Assistant in the Wix Business Manager is a seamlessly integrated chat interface within the dashboard. Acting as a trusted aide, this assistant guides users through their management journey by providing answers to questions and valuable insights about their site. With its comprehensive knowledge, the AI Assistant empowers users to better understand and leverage available resources, assisting with site operations and business tasks. For instance, it can suggest content options, address support inquiries, and analyze analytics—all from a single entry point.

The AI Marketing Agent helps businesses to market themselves online by proactively generating tailored marketing plans that align with users’ goals and target audiences. By analyzing data from their website, the AI delivers personalized strategies to enhance SEO, create engaging content, manage social media, run email campaigns and optimize paid advertising—all with minimal effort from the user. This solution not only simplifies marketing but also drives Wix’s monetization strategy, seamlessly guiding users toward high-impact paid advertising and premium marketing solutions. As businesses invest in growing their online presence, Wix benefits through a share of ad spend and premium feature adoption—fueling both user success and revenue growth.

We will continue to release and optimize specialized AI agents that assist our users in building the online presence they envision. We are exploring various monetization strategies as we fully roll out these agents and adoption increases.

Wix’s management is seeing Wix’s gross margin improve because of AI integration in customer care

Creative Subscriptions non-GAAP gross margin improved to 85% in Q4’24 and to 84% for the full year 2024, up from 82% in 2023. Business Solutions non-GAAP gross margin increased to 32% in Q4’24 and to slightly above 30% for the full year 2024. Continued gross margin expansion is the product of multiple years of cost structure optimization and efficiencies from AI integration across our Customer Care operations.

Wix’s management believes the opportunity for Wix in the AI era is bigger than what came before

There’s a lot of discussions about a lot of theories about it. But I really believe that the opportunity there is bigger than anything else because what we have today are going to continue to dramatically evolve into something that is probably more powerful and more enabling for small businesses to be successful. Overall, the Internet has a tendency to do it every 10 years or so, right, in the ’90s, the Internet started and became HTML, then it became images and then later on videos and then it became mobile, right? And I think they became interactive, everything become an application, kind of an application. And I think how website will look at the AI universe is the next step, and I think there’s a lot of exciting things we can offer our users there.

Visa (NYSE: V)

Visa is an early adopter of AI and management continues to drive adoption; Visa has seen material gains in engineering productivity; Visa has deployed AI in many functions, such as analytics, sales, finance, and marketing

We were very early adopters of artificial intelligence, and we continue to drive hard at the adoption of generative AI as we have for the last couple of years. So we’ve been working to embed AI and AI tooling into our company’s operations, I guess, broadly. We’ve seen material gains in productivity, particularly in our engineering teams. We’ve deployed AI tooling in client services, sales, finance, marketing, really everywhere across the company. And we were a very early adopter of applied AI in the analytics and modeling space, very early by like decades, we’ve been using AI in that space. So our data science and risk management teams have, at this point, decades of applied experience with AI, and they’re aggressively adopting the current generations of AI technology to enhance both our internal and our market-facing predictive and detective modeling capabilities. Our product teams are also aggressively adopting gen AI to build and ship new products.

Zoom Communications (NASDAQ: ZM)

Zoom AI Companion’s monthly active users (MAUs) grew 68% quarter-on-quarter; management has added new agentic AI capabilities to Zoom AI Companion; management will launch the Custom AI Companion add-on in April 2025; management will launch AI Companion for clinicians in March 2025; Zoom AI Companion is added into a low-end Zoom subscription plan at no added cost, and customers do not want to leave their subscriptions because of the added benefit of Zoom AI Companion; Zoom will be monetising Zoom AI Companion from April 2025 onwards through the Custom AI Companion add-on; the Custom AI Companion add-on would be $12 a seat when it’s launched in April 2025 and management thinks this price would provide a really compelling TCO (total cost of ownership) for customers; management thinks Custom AI Companion would have a bigger impact on Zoom’s revenue in 2026 (FY2027) than in 2025 (FY2026); see Point 28 for use cases for Custom AI Companion

Growth in monthly active users of Zoom AI Companion has accelerated to 68% quarter-over-quarter, demonstrating the real value AI is providing customers…

As part of AI Companion 2.0, we added advanced agentic capabilities, including memory, reasoning, orchestration and a seamless integration with Microsoft and Google services. In April, we’re launching Custom AI Companion add-on to automate workplace tasks through custom agents. This will personalize AI to fit customer needs, connect with their existing data, and work seamlessly with their third-party tools. We’re also enhancing Zoom Workplace for Clinicians with an upgraded AI Companion that will enable clinical note-taking capabilities and specialized medical features for healthcare providers starting in March…

…If you look at our low SMB customer online buyers, AI Companion is part of that at no additional cost, made our service very sticky and also the customers give a very basic example, like meeting summary, right? It works so well, more and more customers follow the value…

For high end, for sure, and we understand that today’s AI Companion and additional cost we cannot monetize. However, in April, we are going to announce the customized Companion for interested customers. We can monetize…

…[Question] So in April, when the AI customization, the AI Companion becomes available, I think it’s $11 or $12 a seat. Can you maybe help us understand how you’re thinking about like what’s the real use case?

[Answer] In regards to your question about what are sort of the assumptions or what’s the targeting in our [ head ] with the $12 Custom AI Companion SKU. I would say, starting with enterprise customers, obviously, the easiest place to sort of pounce on them is our own customer base and talk about that, but certainly not just limited to that. But we’ll be probably giving a lot more, I would say, at Enterprise Connect, which you can see on the thing there. But I would say we’ve assumed some degree of monetization in FY ’26, I think you’ll see more of it in ’27. And we think that the $12 price point is going to be a really compelling TCO story for our customers, it’s differentiated from what others in the market are pricing now. 

The Zoom Virtual Agent feature will soon be able to handle complex tasks

Zoom Virtual Agent will soon expand reasoning abilities to handle complex tasks while maintaining conversational context for more natural and helpful outcomes.

Zoom’s management believes Zoom is uniquely positioned to win in agentic AI for a few reasons, including Zoom having exception context of users’ ongoing conversations, and Zoom’s federated AI approach where the company can use the best models for each task

We’re uniquely positioned to succeed in agentic AI for several reasons:

● Zoom is a system of engagement for our users with recent information in ongoing conversations. This exceptional context along with user engagement allows us to drive greater value for customers.

● Our federated AI approach lets us combine the best models for each task. We can use specialized small language models where appropriate, while leveraging larger models for more complex reasoning – driving both quality and cost efficiency

Zoom’s management is seeing large businesses want to use Zoom because of the AI features of its products

You take a Contact Center, for example, why we are winning? Because a lot of AI features like AI Expert Assist. AI, a lot of features built into our quality management and so on and so forth. 

Zoom’s management sees Zoom’s AI business services as a great way to monetise AI

You take a Contact Center, for example, why we are winning? Because a lot of AI features like AI Expert Assist. AI, a lot of features built into our quality management and so on and so forth. But all those business services, that’s another great way for us to monetize AI.

Zoom’s management thinks Zoom’s cost of ownership with AI is lower than what competitors are offering

And I look at our AI Companion, all those AI Companion core features today at no additional cost, right? And customer really like it because of the quality, they’re getting better and better every quarter and very useful, right? Not like some other competitors, right? They talk about their AI strategy and when customers realize that, wow, it’s very expensive. And the total cost of ownership is not getting better because cost of the value is not [ great ], but also it’s not [ free ] and they always try to increase price.

A good example of a use case for Custom AI Companion

[Question] So in April, when the AI customization, the AI Companion becomes available, I think it’s $11 or $12 a seat. Can you maybe help us understand how you’re thinking about like what’s the real use case?

[Answer] So regarding the Custom AI Combined on use cases, high levels, we give a customer ability to customize their needs. I’ll give a few examples. One feature like we have a Zoom Service Call video clip, and we are going to support the standard template, right? How to support every customer? They have a customized template for each of the users, and this is a part of combining AI Studio, right? And also all kinds of third-party integration, right? And they like they prefer, right, some of those kind of sort of third-party application integration. With their data, with the knowledge, whether the [ big scenery ], a lot of things, right? Each company is different, they would not customized, so we can leverage our combining studio to work together with the customer to support their needs and also at same time commodities.

Zoom’s management expects the cost from AI usage to increase and so that will impact Zoom’s margins in the future, but management is also building efficiencies to offset the higher cost of AI

[Question] As we think about a shift more towards AI contribution, aren’t we shifting more towards a consumption model rather than a seat model over time, why wouldn’t we see margin compression longer term?

[Answer] Around how to think about margins and business models and why we don’t see compression. And what I would say is that — what we expect to see is similar to what you saw in FY ’25, which is we’re seeing obvious increase in cost from AI.  And that we have an ongoing methodical kind of efficiency list to offset, and we certainly expect that broadly to continue into FY ’26. So I think we feel good about our ability to kind of moderate that. There’s other things we do more holistically where we can offset stuff that’s maybe not in AI in our margins, things like [ colos ], et cetera, that we’ve talked about previously. 


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Microsoft, Paycom Software, PayPal, Salesforce, Shopify, TSMC, Tesla, The Trade Desk, Wix, Visa, and Zoom. Holdings are subject to change at any time.

The Latest Thoughts From American Technology Companies On AI (2024 Q4) – Part 1

A collection of quotes on artificial intelligence, or AI, from the management teams of US-listed technology companies in the 2024 Q4 earnings season.

The way I see it, artificial intelligence (or AI), really leapt into the zeitgeist in late-2022 or early-2023 with the public introduction of DALL-E2 and ChatGPT. Both are provided by OpenAI and are software products that use AI to generate art and writing, respectively (and often at astounding quality). Since then, developments in AI have progressed at a breathtaking pace.

With the latest earnings season for the US stock market – for the fourth quarter of 2024 – coming to its tail-end, I thought it would be useful to collate some of the interesting commentary I’ve come across in earnings conference calls, from the leaders of technology companies that I follow or have a vested interest in, on the topic of AI and how the technology could impact their industry and the business world writ large. This is an ongoing series. For the older commentary:

I’ve split the latest commentary into two parts for the sake of brevity. This is Part 1, and you can Part 2 here. With that, I’ll let the management teams take the stand… 

Airbnb (NASDAQ: ABNB)

Airbnb’s management thinks AI is early and has yet to fundamentally change the travel market for any of the large travel platforms; most travel companies are starting with AI on trip planning but management thinks AI is still too early for trip planning

Here’s what I think about AI. I think it’s still really early. It’s probably similar to like the mid- to late ’90s for the Internet. So I think it’s going to have a profound impact on travel, but I don’t think it’s yet fundamentally changed for any of the large travel platforms…

…So most companies, what they’re actually doing is they’re doing integrations of these other platforms on trip planning. But the trip planning, it’s still early. I don’t think it’s quite bit ready for prime time.

Airbnb’s management is starting with AI in customer service; Airbnb will roll out AI-powered customer support later in 2025; management thinks AI can provide great customer support partly because it can speak all languages 24/7 and read thousands of pages of documents; management will eventually graduate the customer-support AI into a travel and living concierge; management thinks AI can help improve efficiency at Airbnb in customer service

We’re actually starting with customer service. So later this year, we’re going to be rolling out, as part of our Summer Release, AI-powered customer support. As you imagine, we get millions of contacts every year. AI can do an incredible job of customer service. It can speak every language 24/7. It can read a corpus of thousands of pages of documents. And so we’re starting with customer support. And over the coming years, what we’re going to do is we’re going to take the AI-powered customer service agent, and we’re going to bring it into essentially Airbnb search to eventually graduate to be a travel and living concierge…

…[Question] With respect to the AI, I appreciate your answer with respect to outward-looking and how it might change the landscape. What do you think the potential is internally to apply AI for efficiencies inside the company and create an additional layer of potential margin efficiency and/or free cash flow conversion in the years ahead?

[Answer] There’s like a couple like efficiencies that you could imagine at Airbnb. One is obviously customer service. I think that’s like one of the biggest ones. I’ve kind of already covered that, but I think that’s like a massive change for Airbnb.

Airbnb’s management thinks that AI models are getting cheaper and are starting to be commoditised

I think it’s a really exciting time in the space because you’ve seen like with DeepSeek and more competition with models is models are getting cheaper or nearly free. They’re getting faster and they’re getting more intelligent. And they are, for all intent and purpose, starting to get commoditized.

Airbnb’s management thinks that a lot of the value from AI is going to accrue to platforms, and they want Airbnb to be the platform for travel and living that will reap most of the value from AI

What I think that means is a lot of value is going to accrue to the platform. And ultimately, I think the best platform, the best applications are going to be the ones that like most accrue the value from AI. And I think we’re going to be the one to do that with traveling and living.

Airbnb’s management thinks AI can help improve efficiency at Airbnb in engineering productivity; in the short-term, the improvement in engineering productivity has not been material; over the next few years, management thinks AI can drive a 30% increase in engineering productivity at Airbnb; over the long-term, management thinks there can an order of magnitude more productivity; management think younger, more innovative companies, could benefit from AI more than incumbent enterprises

[Question] With respect to the AI, I appreciate your answer with respect to outward-looking and how it might change the landscape. What do you think the potential is internally to apply AI for efficiencies inside the company and create an additional layer of potential margin efficiency and/or free cash flow conversion in the years ahead?

[Answer] The other, I assume, you refer to is essentially engineering productivity. We are seeing some productivity gains. I’ve talked to a lot of other tech CEOs, and here’s what I’ve heard talking to other like tech CEOs. Most of them haven’t seen a material like change in engineering productivity. Most of the engineers are using AI tools. They’re seeing some productivity. I don’t think it’s flowing to like a fundamental step-change in productivity yet. I think a lot of us believe in some kind of medium term of a few years, you could easily see like a 30% increase in technology and engineering productivity. And then, of course, beyond that, I mean, I think it could be like an order of magnitude more productivity because — but that’s going to be like down the road. And I think that’s going to be something that almost all companies benefit from. I think the kind of younger, more innovative, startup-like companies might benefit a little bit more because they’ll have engineers who are more likely to adopt the tools.

Alphabet (NASDAQ: GOOG)

AI Overviews in Search is now available in more than 100 countries; AI Overviews drive higher user satisfaction and search usage; Google’s Gemini model is being used in AI Overviews; with AI Overviews, usage growth of Search is growing over time, especially with younger users; management recently launched ads in AI Overviews and AI Overviews is currently monetising at nearly the same rate as Google Search; Google Search has continued to perform well in this AI age, as overall usage has continued to grow, with stronger growth seen in AI Overviews across all segments

In Search, AI overviews are now available in more than 100 countries. They continue to drive higher satisfaction and search usage…

…That includes Search where Gemini is pairing our AI overviews. People use search more with AI overviews and usage growth increases over-time as people learn that they can ask new types of questions. This behavior is even more pronounced with younger users who really appreciate the speed and efficiency of this new format…

…We’ve already started testing Gemini 2.0 in AI overviews and plan to roll it out more broadly later in the year…

…We recently launched the ads within AI Overviews on mobile in the U.S., which builds on our previous rollout of ads above and below. And as I talked about before, for the AI Overviews, overall, we actually see monetization at approximately the same rate, which I think really gives us a strong base on which we can innovate even more…

…On Search usage, overall, our metrics are healthy. We are continuing to see growth in Search on a year-on-year basis in terms of overall usage. Of course, within that, AI Overviews has seen stronger growth, particularly across all segments of users, including younger users, so it’s being well received. But overall, I think through this AI moment, I think Search is continuing to perform well.

Circle to Search is now available on more than 200 million Android devices; Circle to Search is opening new Search use cases; Circle to Search is popular with younger users; Circle to Search is used to start more than 10% of searches among users who have tried it before

Circle to Search is now available on over 200 million Android devices…

…Circle to Search is driving additional Search use and opening up even more types of questions. This feature is also popular among younger users. Those who have tried Circle to Search before now use it to start more than 10% of their searches…

…In Search, we’re seeing people increasingly ask entirely new questions using their voice, camera or in ways that were not possible before, like with Circle to search.

Alphabet’s management believes Google has a unique infrastructure advantage in AI because the company has developed each component of its technology stack; Alphabet broke ground on 11 new cloud regions and data center campuses in 2024, and announced plans for 7 new subsea cable projects; Google data centers now deliver 4x more computing power per unit of electricity compared to 5 years ago

We have a unique advantage because we develop every component of our technology stack, including hardware, compilers, models and products. This approach allows us to drive efficiencies at every level from training and serving to develop our productivity. In 2024, we broke ground on 11 new cloud regions and data center campuses in places like South Carolina, Indiana, Missouri and around the world.

We also announced plans for seven new subsea cable projects, strengthening global connectivity. Our leading infrastructure is also among the world’s most efficient. Google data centers deliver nearly four times more computing power per unit of electricity compared to just five years ago.

Google Cloud customers consume 8x more compute capacity for training and inference compared to 18 months ago; first-time commitments to Google Cloud more than doubled in 2024; Google Cloud closed a few deals in 2024 worth more than $1 billion each; Google Cloud’s AI hypercomputer utilises both GPUs (graphics processing units) and TPUs (tensor processing units), and has helped Wayfair improve performance and scalability by 25%; Google saw strong uptake of Trillium, its 6th generation TPU, in 2024 Q4; Trillium is 4x better in training and has 3x higher inference throughput than the 5th generation TPU; Google Cloud is offering NVIDIA’s H200 GPUs to customers; Google Cloud is the first cloud provider to provide NVIDIA’s Blackwell GPUs; the capex for Google Cloud is mostly for Google’s own self-designed data centers and TPUs (tensor processing units)

Today, cloud customers consume more than eight times the compute capacity for training and inferencing compared to 18 months ago…

…In 2024, the number of first-time commitments more than double compared to 2023…

…Last year, we closed several strategic deals over $1 billion and the number of deals over $250 million doubled from the prior year…

…We continue to see strong growth across our broad portfolio of AI-powered cloud solutions. It begins with our AI hypercomputer, which delivers leading performance and cost across both GPUs and TPUs. These advantages help Citadel with modeling markets and training and enabled Wayfair to modernize its platform, improving performance and scalability by nearly 25%. 

In Q4, we saw strong uptake of Trillium, our sixth-generation TPU, which delivers four times better training performance and three times greater inference throughput compared to the previous generation. We also continue our strong relationship with NVIDIA. We recently delivered their H200 based platforms to customers. And just last week, we were the first to announce a customer running on the highly-anticipated Blackwell platform…

……Our strategy is mostly to rely our own self-design and build data centers. So, they’re industry-leading in terms of both cost and power efficiency at scale. We have our own customized TPUs. They’re customized for our own workload, so they do deliver outstanding the superior performance and capex efficiency. So, we’re going to be looking at all that when we make decisions as to how we’re going to progress capital investments throughout the coming years.

Google launched an experimental version of its Gemini 2.0 Flash model in December 2024, but the model will be generally available for developers and customers; Google debuted its experimental Gemini 2.0 Flash Thinking model in late-2024 and it has gathered extremely positive reviews; Google is working on even better thinking models; Gemini 2.0’s advances in multimodality and native tool use helps Google build a universal AI assistant; an example of this universal assistant can be seen in Deep Research; Deep Research was launched in Gemini Advanced in December and is being rolled out to Android users globally; the consumer Gemini app debuted on iOS in November 2024 and has seen great product momentum; Project Mariner and Project Astra are AI agent products currently being tested and they will appear in the Gemini app sometime in 2025; Gemini and Google’s video and image generation models consistently excel in industry leaderboards and benchmarks; 4.4 million developers are using Gemini models today, double from just six months ago; Google has 7 products with over 2 billion users each, and all 7 products use Gemini; all Google Workspace business and enterprise customers were recently given access to all of Gemini’s AI capabilities; Gemini 2.0 Flash is one of the most capable models people can access for free; management’s current intention for monetisation of Gemini is through subscriptions and improving the user experience, but they have an eye on advertising revenues

In December, we unveiled Gemini 2.0, our most capable AI model yet, built for the agent era. We launched an experimental version of Gemini 2.0 Flash, our workhorse model with low-latency and enhanced performance. Flash has already rolled-out to the Gemini app, and tomorrow we are making 2.0 Flash generally available for developers and customers, along with other model updates…

…Late last year, we also debuted our experimental Gemini 2.0 Flash Thinking model. The progress to scale thinking has been super-fast and the review so-far have been extremely positive. We are working on even better thinking models and look-forward to sharing those with the developer community soon.

Gemini 2.0’s advances in multimodality and native tool use enable us to build new agents that bring us closer to our vision of a universal assistant. One early example is deep research. It uses agent capabilities to explore complex topics on your behalf and give key findings along with sources. It launched in Gemini Advanced in December and is rolling out to Android users all over the world.

We are seeing great product momentum with our consumer Gemini app, which debuted on iOS last November.

…We have opened up trusted tester access to a handful of research prototypes, including Project Mariner, which can understand and reason across information on a browser screen to complete tasks and Project Astra. We expect to bring features from both to the Gemini app later this year…

…Veo2, our state-of-the-art video generation model and Imagine3, our highest-quality text image model. These generative media models as well as Gemini consistently top industry leaderboards and score top marks across industry benchmarks. That’s why more than 4.4 million developers are using our Gemini models today, double the number from just six months ago…

…We have seven products and platforms with over 2 billion users and all are using Gemini…

…We recently gave all Google Workspace business and enterprise customers access to all of our powerful Gemini AI capabilities to help boost their productivity…

…2.0 Flash. I mean, I think that’s one of the most capable models you can access at the free tier…

…[Question] How should we think about the future monetization opportunity of Gemini? Today, it’s really a premium subscription offering or a free offering. Over time, do you see an ad component?

[Answer] On the monetization side, obviously, for now, we are focused on a free tier and subscriptions. But obviously, as you’ve seen in Google over time, we always want to lead with user experience. And we do have very good ideas for native ad concepts, but you’ll see us lead with the user experience. And — but I do think we’re always committed to making the products work and reach billions of users at scale. And advertising has been a great aspect of that strategy. And so, just like you’ve seen with YouTube, we’ll give people options over time. But for this year, I think you’ll see us be focused on the subscription direction.

Google Cloud’s AI developer platform, Vertex AI, had a 5x year-on-year increase in customers in 2024 Q4; Vertex AI offers more than 200 foundation models form Google; Vertex AI’s usage grow 20x in 2024

Our AI developer platform, Vertex AI, saw a 5x increase in customers year-over-year with brands like, International and WPP building new applications and benefiting from our 200 plus foundation models. Vertex usage increased 20x during 2024 with particularly strong developer adoption of Gemini Flash, Gemini 2.0, and most recently VEO.

Alphabet’s management will be introducing Veo2, Google’s video generation model, for creators in Youtube in 2025; advertisers around the world can now promote Youtube creator videos and ad campaigns across all AI-powered campaign types and Google ads

Expanding on our state-of-the-art video generation model, we announced Veo2, which creates incredibly high-quality video in a wide range of subjects and styles. It’s been inspiring to see how people are experimenting with it. We’ll make it available to creators on YouTube in the coming months…

…. All advertisers globally can now promote YouTube creator videos and ad campaigns across all AI-powered campaign types and Google Ads, and creators can tag partners in their brand videos…

Alphabet’s management announced the first beta of Android 16 in January 2025; there will be deeper Gemini integration for the new Samsung Galaxy S25 smartphone series; Alphabet has announced Android XR, the first Android platform for the Gemini era

Last month, we announced the first beta of Android 16 plus new Android updates, including a deeper Gemini integration coming to the new Samsung Galaxy S25 series. We also recently-announced Android XR, the first Android platform built for the Gemini era. Created with Samsung and Qualcomm, Android XR is designed to power an ecosystem of next-generation extended reality devices like headsets and glasses.

Waymo is now serving more than 150,000 trips per week (was 150,000 in 2024 Q3); Waymo is expanding in new markets in the USA this year and in 2026; Waymo will soon be launched in Tokyo; Waymo is developing its 6th-gen driver, which will significantly reduce hardware costs

It’s now averaging over 150,000 trips each week and growing. Looking ahead, Waymo will be expanding its network and operations partnerships to open up new markets, including Austin and Atlanta this year and Miami next year. And in the coming weeks, Waymo One vehicles will arrive in Tokyo for their first international road trip. We are also developing the sixth-generation Waymo driver, which will significantly lower hardware costs.

Alphabet’s management introduced a new Google shopping experience, infused with AI, in 2024 Q4, and there was 13% more daily active users in Google shopping in December 2024 compared to a year ago; the new Google Shopping experience helps users speed up their shopping research

Google is already present in over half of journeys where a new brand, product or retailer are discovered by offering new ways for people to search, we’re expanding commercial opportunities for our advertisers…

…Retail was particularly strong this holiday season, especially on Black Friday and Cyber Monday, which each generated over $1 billion in ad revenue. Interestingly, despite the U.S. holiday shopping season being the shortest since 2019, retail sales began much earlier in October, causing the season to extend longer than anticipated.

People shop more than 1 billion times a day across Google. Last quarter, we introduced a reinvented Google shopping experience, rebuilt from the ground up with AI. This December saw roughly 13% more daily active users in Google shopping in the U.S., compared to the same period in 2023…

…The new Google Shopping experiences specifically to your question, users to really intelligently show the most relevant products, helping to speed up and simplify your research. You get an AI-generated brief with top things to consider for your search plus maybe products that meet your needs. So, shoppers very often want low prices. So, the new page not only includes like deal-finding tools like price comparison, price insights, price tracking throughout. But it’s also a new and dedicated personalized deals page, which can browse deals for you, and all this is really built on the backbone of AI.

Shoppers can now take a photo and use Lens to quickly find information about the product; Lens is now used in over 20 billion visual searches per month (was over 20 billion in 2024 Q3); majority of Lens searches are incremental

Shoppers can now take a photo of a product and using Lens quickly find information about the product, reviews, similar products and where they can get it for a great price. Lens is used for over 20 billion visual search queries every month and the majority of these searches are incremental.

Alphabet’s management continues to infuse AI capabilities into Google’s advertising business; Petco used Demand Gen campaigns to achieve a 275% increase in return on ad spend and a 74% increase in click-through rates compared to social benchmarks; Youtube Select Creator Takeovers is now generally available in the US and will be rolled out across the world; PMax was recently strengthened with new controls and easier reporting functions; Event Ticket Center used PMax and saw a 5x increase in production of creative assets, driving a 300% increase in conversions compared to using manual assets; Meridian, Google’s marketing mix model, was recently made generally available and it delivers 17% higher return on advertising spend on Youtube compared to manual campaigns

We continue investing in AI capabilities across media buying, creative and measurement. As I’ve said before, we believe that AI will revolutionize every part of the marketing value chain.

And over the past quarter, we’ve seen how our customers are increasingly focusing on optimizing the use of AI. As an example, [ Petco ], used Demand Gen campaigns across targeting, creative generation and bidding to find new pet parent audiences across YouTube. They achieved a 275% higher return on ad spend and a 74% higher click-through rate than their social benchmarks.

On media buying, we made YouTube Select Creator Takeovers generally available in the U.S. and will be expanding to more markets this year. Creators know their audience the best and creator takeovers help businesses connect with consumers through authentic and relevant content.

Looking at Creative, we introduced new controls and made reporting easier in PMax, helping customers better understand and reinvest into their best-performing assets. Using asset generation in PMax, Event Ticket Center achieved a 5x increase in production of creative assets saving time and effort. They also increased conversions by 300% compared to the previous period when they used manual assets…

…Last week, we made Meridian, our marketing mix model, generally available for customers, helping more business reinvest into creative and media buying strategies that they know work. Based on the Nielsen meta analysis of marketing mix models, on average, Google AI-powered video campaigns on YouTube delivered 17% higher return on advertising spend than manual campaigns.

Sephora used Demand Gen Shorts-only channel for advertising that drove an 82% increase in searches for Sephora Holiday

Sephora used demand gen Shorts-only channel to boost traffic and brand searches for the holiday gift guide campaign and leverage greater collaborations to find the best gift. This drove an 82% relative uplift in searches for Sephora holiday.

Citi is using Google Cloud for its generative AI initiatives across customer service, document summarisation, and search

Another expanding partnership is with Citi, who is modernizing its technology infrastructure with Google Cloud to transform employee and customer experiences. Using Google Cloud, it will improve its digital products, streamline employee workflows and use advanced high-performance computing to enable millions of daily computations. This partnership also fuels Citi’s generate AI initiatives across customer service, document summarization and search to reduce manual processing.

Google Cloud had 30% revenue growth in 2024 Q4 (was 35% in 2024 Q3) driven by growth in core GCP products, AI infrastructure, and generative AI solutions; operating margin was 17.5% (was 17% in 2024 Q3 and was 9.4% in 2023 Q4); GCP grew at a much higher rate than Google Cloud overall; Google Cloud had more AI demand than capacity in 2024 Q4; management is thinking about Google’s capital intensity, but they want to invest because they are seeing strong AI demand both internally and externally; the capex Google is making can be repurposed across its different businesses

Turning to the Google Cloud segment, which continued to deliver very strong results this quarter. Revenue increased by 30% to $12 billion in the fourth quarter, reflecting growth in GCP, across core GCP products, AI infrastructure, and generative AI solutions. Once again, GCP grew at a rate that was much higher than cloud overall. Healthy Google Workspace growth was primarily driven by increase in average revenue per seat. Google Cloud operating income increased to $2.1 billion and operating margin increased from 9.4% to 17.5%…

…We do see and have been seeing very strong demand for our AI products in the fourth quarter in 2024. And we exited the year with more demand than we had available capacity. So, we are in a tight supply demand situation, working very hard to bring more capacity online…

…[Question] How do you think about long-term capital intensity for this business?

[Answer] On the first one, certainly, we’re looking ahead, but we’re managing very responsibly. It was a very rigorous, even internal governance process, looking at how do we allocate the capacity and what would we need to support the customer demand externally, but also across the Google — the Alphabet business. And as you’ve seen in the comment I’ve just made on Cloud, we do have demand that exceeds our available capacity. So, we’ll be working hard to address that and make sure we bring more capacity online. We do have the benefit of having a very broad business, and we can repurpose capacity, whether it’s through Google Services or Google Cloud to support, as I said, whether it’s search or GDM, or Google Cloud customers, we can do that in a more efficient manner.

Alphabet’s management thinks Google’s AI models are in the lead when compared to DeepSeek’s, and this is because of Google’s full-stack development

If you look at one of the areas in which the Gemini model shines is the Pareto frontier of cost, performance, and latency. And if you look at all three attributes, I think we are — we lead this period of frontier. And I would say both our 2.0 Flash models, our 2.0 Flash thinking models, they are some of the most efficient models out there, including comparing to DeepSeek’s V3 and R1. And I think a lot of it is our strength of the full stack development, end-to-end optimization, our obsession with cost per query.

Alphabet’s management has seen the proportion of AI spend on inference growing over the last 3 years when compared to training; management thinks reasoning AI models will accelerate this trend

A couple of things I would say are if you look at the trajectory over the past three years, the proportion of the spend toward inference compared to training has been increasing, which is good because, obviously, inferences to support businesses with good ROIC…

…I think the reasoning models, if anything, accelerates that trend because it’s obviously scaling upon inference dimension as well.

Alphabet’s management thinks that AI agents and Google Search are not competing in a zero-sum game

[Question] With your own project Mariner efforts and a competitor’s recent launch, it seems there’s suddenly really strong momentum on AI consumer agents and kind of catching up to that old Google Duplex Vision. I think when you look a few years ahead, where do you see consumer agents going? And really, what does it mean to Google Search outside of Lens? Is there room for both to flourish?

[Answer] Gemini 2.0 was definitely built with the view of enabling more agentic use cases. And so, I actually — we are definitely seeing progress inside. And I think we’ll be able to do more agentic experiences for our users. Look, I actually think all of this expands the opportunity space. I think it — historically, we’ve had information use cases, but now you can have — you can act on your information needs in a much deeper way. It’s always been our vision when we have talked about Google Assistant, etc. So, I think the opportunity space expands. I think there’s plenty of it, feels very far from a zero-sum game. There’s plenty of room, I think, for many new types of use cases to flourish. And I think for us, we have a clear sense of additional use cases we can start to tackle for our users in Google Search.

Alphabet’s management has been passing on cost differentiations arising from Google Cloud’s end-to-end stack approach to customers

Part of the reason we have taken the end-to-end stack approach is so that we can definitely drive a strong differentiation in end-to-end optimizing and not only on a cost but on a latency basis, on a performance basis. Be it the Pareto frontier we mentioned, and I think our full stack approach and our TPU efforts all play give a meaningful advantage. And we plan — you already see that. I know you asked about the cost, but it’s effectively captured when we price outside, we pass on the differentiation. 

Amazon (NASDAQ: AMZN)

AWS grew 19% year-on-year in 2024 Q4, and is now at a US$115 billion annualised revenue run rate; management expects lumpiness in AWS’s growth in the next few years, but is incredibly optimistic about AWS’s growth; management thinks the future will be one where (a) every app is infused with generative AI that has inference as a core building block, and (b) companies will have AI agents accomplishing tasks and interacting with each other; management believes this future will be built on the cloud, and mostly on AWS; the shift by enterprises from on-premises to the cloud, which is a non-AI activity, continues for AWS; AWS continues to innovate in non-AI areas; AWS’s growth in 2024 Q4 was driven by both generative AI and non-generative AI offerings; AWS had a massive 48% year-on-year jump in operating income in 2024 Q4, helped partly by an increase in estimated useful life of servers that started in 2024; management sees AWS being capable of faster growth today if not for supply constraints; the constraints relate to (1) chips from 3rd-party partners (most likely referring to NVIDIA), (2) AWS’s own Trainium chips, (3) power for data centers, and (4) other supply chain components; management sees the AWS constraints starting to relax in 2025 H2; AWS’s AI services come with lower margins right now, but management thinks the AI-related margin will over time be on par with the non-AI margin

In Q4, AWS grew 19% year-over-year and now has a $115 billion annualized revenue run rate. AWS is a reasonably large business by most folks’ standards. And though we expect growth will be lumpy over the next few years as enterprise adoption cycles, capacity considerations and technology advancements impact timing, it’s hard to overstate how optimistic we are about what lies ahead for AWS’ customers and business…

…While it may be hard for some to fathom a world where virtually every app has generative AI infused in it, with inference being a core building block just like compute, storage and database, and most companies having their own agents that accomplish various tasks and interact with one another, this is the world we’re thinking about all the time. And we continue to believe that this world will mostly be built on top of the cloud with the largest portion of it on AWS…

…While AI continues to be a compelling new driver in the business, we haven’t lost our focus on core modernization of companies’ technology infrastructure from on-premises to the cloud. We signed new AWS agreements with companies, including Intuit, PayPal, Norwegian Cruise Line Holdings, Northrop Grumman, The Guardian Life Insurance Company of America, Reddit, Japan Airlines, Baker Hughes, The Hertz Corporation, Redfin, Chime Financial, Asana, and many others. Consistent customer feedback from our recent AWS re:Invent gathering was appreciation that we’re still inventing rapidly in non-AI key infrastructure areas like storage, compute, database and analytics…

…During the fourth quarter, we continued to see growth in both generative AI and non-generative AI offerings as companies turn their attention to newer initiatives, bring more workloads to the cloud, restart or accelerate existing migrations from on-premise to the cloud, and tap into the power of generative AI…

…AWS reported operating income of $10.6 billion, an increase of $3.5 billion year-over-year. This is a result of strong growth, innovation in our software and infrastructure to drive efficiencies, and continued focus on cost control across the business. As we’ve said in the past, we expect AWS operating margins to fluctuate over time driven in part by the level of investments we’re making. Additionally, we increased the estimated useful life of our servers starting in 2024, which contributed approximately 200 basis points to the AWS margin increase year-over-year in Q4…

……It is true that we could be growing faster, if not for some of the constraints on capacity. And they come in the form of, I would say, chips from our third-party partners, come a little bit slower than before with a lot of midstream changes that take a little bit of time to get the hardware actually yielding the percentage-healthy and high-quality servers we expect. It comes with our own big new launch of our own hardware and our own chips and Trainium2, which we just went to general availability at re:Invent, but the majority of the volume is coming in really over the next couple of quarters, the next few months. It comes in the form of power constraints where I think the world is still constrained on power from where I think we all believe we could serve customers if we were unconstrained. There are some components in the supply chain, like motherboards, too, that are a little bit short in supply for various types of servers…

…I predict those constraints really start to relax in the second half of ’25…

…At the stage we’re in right now, AI is still early stage. It does come originally with lower margins and a heavy investment load as we’ve talked about. And in the short term, over time, that should be a headwind on margins. But over the long term, we feel the margins will be comparable in non-AI business as well.

Amazon’s management sees NVIDIA being an important partner of AWS for a long time; management does not see many large-scale generative AI apps existing right now; when generative AI apps reach scale, their costs to operate can rise very quickly, and management believes this will drive customers to demand better price performance from chips, which is why AWS built its custom AI chips; Trainium 2, AWS’s custom AI chip, was launched in December 2024; EC2 instances powered by Trainium 2 is 30%-40% more price performant than instances powered by other GPUs; important technology companies such as Adobe, Databricks, and Qualcomm have seen impressive results after testing Trainium 2; Anthropic is building its future frontier models on Trainium 2; AWS is collaborating with Anthropic on Project Rainier, which is a cluster of a few hundred thousand Trainium 2 chips that have 5x the exaflops Anthropic used to train its current set of models; management is already 

Most AI compute has been driven by NVIDIA chips, and we obviously have a deep partnership with NVIDIA and will for as long as we can see into the future. However, there aren’t that many generative AI applications of large scale yet. And when you get there, as we have with apps like Alexa and Rufus, cost can get steep quickly. Customers want better price performance and it’s why we built our own custom AI silicon. Trainium2 just launched at our AWS re:Invent Conference in December. And EC2 instances with these chips are typically 30% to 40% more price performant than other current GPU-powered instances available. That’s very compelling at scale. Several technically-capable companies like Adobe, Databricks, Poolside and Qualcomm have seen impressive results in early testing of Trainium2. It’s also why you’re seeing Anthropic build their future frontier models on Trainium2. We’re collaborating with Anthropic to build Project Rainier, a cluster of Trainium2 UltraServers containing hundreds of thousands of Trainium2 chips. This cluster is going to be 5x the number of exaflops as the cluster that Anthropic used to train their current leading set of cloud models. We’re already hard at work on Trainium3, which we expect to preview late in ’25 and defining Trainium4 thereafter.

Building outstanding performant chips that deliver leading price performance has become a core strength of AWS’, starting with our Nitro and Graviton chips in our core business and now extending to Trainium and AI and something unique to AWS relative to other competing cloud providers.

Amazon’s management has seen Amazon SageMaker AI, AWS’s fully-managed AI service, become the go-to service for AI model builders; SageMaker’s HyperPod automatically splits training workloads across many AI accelerators and prevents interruptions, saving training time up tp 40%; management recently released new features for SageMaker, such as the ability to prioritise which workloads to receive capacity when budgets are reached; the latest version of SageMaker is able to integrate all of AWS’s data analytics and AI services into one surface

I won’t spend a lot of time in these comments on Amazon SageMaker AI, which has become the go-to service for AI model builders to manage their AI data, build models, experiment and deploy these models, except to say that SageMaker’s HyperPod capability, which automatically splits training workloads across many AI accelerators, prevents interruptions by periodically saving checkpoints, and automatically repairing faulty instances from their last saved checkpoint and saving training time by up to 40%. It continues to be a differentiator, received several new compelling capabilities at re:Invent, including the ability to manage costs at a cluster level and prioritize which workloads should receive capacity when budgets are reached, and is increasingly being adopted by model builders…

…There were several key launches customers were abuzz about, including Amazon Aurora DSQL, our new serverless distributed SQL database that enables applications with the highest availability, strong consistency, PostgreS compatibility and 4x faster reads and writes compared to other popular distributed SQL databases; Amazon S3 tables, which make S3 the first cloud object store with fully managed support for Apache Iceberg for faster analytics; Amazon S3 Metadata, which automatically generates queryable metadata, simplifying data discovery, business analytics, and real-time inference to help customers unlock the value of their data in S3; and the next generation of Amazon SageMaker, which brings together all of the data analytics services and AI services into one interface to do analytics and AI more easily at scale.

Amazon Bedrock is AWS’s fully-managed service for developers to build generative AI applications by leverage on frontier models; management recently introduced more than 100 popular emerging models on Bedrock, including DeepSeek’s R1 models; management recently introduced new features to Bedrock to help customers lower cost and latency in inference workloads; management is seeing Bedrock resonate strongly with customers; management recently released Amazon’s own Nova family of frontier models on Bedrock; customers are starting to experiment with DeepSeek’s models

Amazon Bedrock is our fully managed service that offers the broadest choice of high-performing foundation models with the most compelling set of features that make it easy to build a high-quality generative AI application. We continue to iterate quickly on Bedrock announcing Luma AI poolside and over 100 other popular emerging models to Bedrock at re:Invent. In short order, we also just added DeepSeek’s R1 models to Bedrock and SageMaker…

…We delivered several compelling new Bedrock features at re:Invent, including prompt caching, intelligent prompt routing and model distillation, all of which help customers achieve lower cost and latency in their inference. Like SageMaker AI, Bedrock is growing quickly and resonating strongly with customers…

…We also just launched Amazon’s own family of frontier models in Bedrock called Nova…

…We moved so quickly to make sure that DeepSeek was available both in Bedrock and in SageMaker faster than you saw from others. And we already have customers starting to experiment with that.

The Nova family has comparable intelligence with other leading AI models, but also offers lower latency and price, and integration with important Bedrock features; many large enterprises, including Palantir, Deloitte, and SAP, are already using Nova

We also just launched Amazon’s own family of frontier models in Bedrock called Nova. These models compare favorably in intelligence against the leading models in the world but offer lower latency; lower price, about 75% lower than other models in Bedrock; and are integrated with key Bedrock features like fine-tuning, model distillation, knowledge bases of RAG and agentic capabilities. Thousands of AWS customers are already taking advantage of the capabilities and price performance of Amazon Nova models, including Palantir, Deloitte, SAP, Dentsu, Fortinet, Trellix, and Robinhood, and we’ve just gotten started.

Amazon’s management still thinks Amazon Q is the most capable AI-powered software development assistant; early testing shows that Amazon Q can now shorten a multi-year mainframe migration by 50%

Amazon Q is the most capable generative AI-powered assistant for software development and to leverage your own data…

…We obliged with our recent deliveries of Q Transformations that enable moves from Windows.NET applications to Linux, VMware to EC2, and accelerates mainframe migrations. Early customer testing indicates that Q can turn what was going to be a multiyear effort to do a mainframe migration into a multi-quarter effort, cutting by more than 50% the time to migrate mainframes. This is a big deal and these transformations are good examples of practical AI.

Amazon’s management expects capital expenditures of around US$105 billion for the whole of 2025 (was around $75 billion in 2024); the capex in 2025 will be for AWS as well as the retail business, but will primarily be for AWS’s AI infrastructure; reminder that the faster AWS grows, the faster Amazon needs to invest capital for hardware; management will only spend on capex if they see significant signals of demand; management thinks AI is a once-in-a-lifetime business opportunity, and that it’s a good sign on the long-term growth opportunities AWS has when capex is expanding

Capital investments were $26.3 billion in the fourth quarter, and we think that run rate will be reasonably representative of our 2025 capital investment rate. Similar to 2024, the majority of the spend will be to support the growing need for technology infrastructure. This primarily relates to AWS, including to support demand for our AI services, as well as tech infrastructure to support our North America and International segments. Additionally, we’re continuing to invest in capacity for our fulfillment and transportation network to support future growth. We’re also investing in same-day delivery facilities and our inbound network as well as robotics and automation to improve delivery speeds and to lower our cost to serve. These capital investments will support growth for many years to come…

…The vast majority of that CapEx spend is on AI for AWS. The way that AWS business works and the way the cash cycle works is that the faster we grow, the more CapEx we end up spending because we have to procure data center and hardware and chips and networking gear ahead of when we’re able to monetize it. We don’t procure it unless we see significant signals of demand. And so when AWS is expanding its CapEx, particularly in what we think is one of these once-in-a-lifetime type of business opportunities like AI represents, I think it’s actually quite a good sign, medium to long term, for the AWS business…

…We also have CapEx that we’re spending this year in our Stores business, really with an aim towards trying to continue to improve the delivery speed and our cost to serve. And so you’ll see us expanding the number of same-day facilities from where we are right now. You’ll also see us expand the number of delivery stations that we have in rural areas so we can get items to people who live in rural areas much more quickly, and then a pretty significant investment as well on robotics and automation so we can take our cost to serve down and continue to improve our productivity.

Amazon’s management completed a useful life study for its servers and network equipment in 2024 Q4 and has decreased the useful life estimate; management early retired some servers and network equipment in 2024 Q4; the decrease in useful life estimate and the early retirement will lower Amazon’s operating income, primarily in the AWS segment

In Q4, we completed a useful life study for our servers and network equipment, and observed an increased pace of technology development, particularly in the area of artificial intelligence and machine learning. As a result, we’re decreasing the useful life for a subset of our servers and network equipment from 6 years to 5 years, beginning in January 2025. We anticipate this will decrease full year 2025 operating income by approximately $700 million. In addition, we also early retired a subset of our servers and network equipment. We recorded a Q4 2024 expense of approximately $920 million from accelerated depreciation and related charges and expect this will also decrease full year 2025 operating income by approximately $600 million. Both of these server and network equipment useful life changes primarily impact our AWS segment.

Amazon’s management sees AI as the biggest opportunity since cloud and the internet

From our perspective, we think virtually every application that we know of today is going to be reinvented with AI inside of it and with inference being a core building block, just like compute and storage and database. If you believe that, plus altogether new experiences that we’ve only dreamed about are going to actually be available to us with AI, AI represents, for sure, the biggest opportunity since cloud and probably the biggest technology shift and opportunity in business since the Internet.

Amazon’s management has been impressed with DeepSeek’s innovations

I think like many others, we were impressed with what DeepSeek has done, I think in part impressed with some of the training techniques, primarily in flipping the sequencing of reinforcement learning being earlier and without the human-in-the-loop. We thought that was interesting ahead of the supervised fine-tuning. We also thought some of the inference optimizations they did were also quite interesting

Amazon’s management’s core belief remains that generative AI apps will use multiple models and different customers will use different AI models for different workloads 

You have a core belief like we do that virtually all the big generative AI apps are going to use multiple model types, and different customers are going to use different models for different types of workloads.

Amazon’s management thinks that the cheaper AI inference becomes, the more inference spending there will be; management believes that the cost of AI inference will fall substantially over time

Sometimes people make the assumptions that if you’re able to decrease the cost of any type of technology component, in this case, we’re really talking about inference, that somehow it’s going to lead to less total spend in technology. And we have never seen that to be the case. We did the same thing in the cloud where we launched AWS in 2006, where we offered S3 object storage for $0.15 a gigabyte and compute for $0.10 an hour, which, of course, is much lower now many years later, people thought that people would spend a lot less money on infrastructure technology. And what happens is companies will spend a lot less per unit of infrastructure, and that is very, very useful for their businesses, but then they get excited about what else they could build that they always thought was cost prohibitive before, and they usually end up spending a lot more in total on technology once you make the per unit cost less. And I think that is very much what’s going to happen here in AI, which is the cost of inference will substantially come down. What you heard in the last couple of weeks, DeepSeek is a piece of it, but everybody is working on this. I believe the cost of inference will meaningfully come down. I think it will make it much easier for companies to be able to infuse all their applications with inference and with generative AI.

Amazon’s management currently sees 2 main ways that companies are getting value out of AI; the 1st way is through productivity and cost savings, and it is the lowest-hanging fruit; the 2nd way is by building new experiences

There’s kind of two macro buckets of how we see people, both ourselves inside Amazon as well as other companies using AWS, how we see them getting value out of AI today. The first macro bucket, I would say, is really around productivity and cost savings. And in many ways, this is the lowest-hanging fruit in AI…

….I’d say the other big macro bucket are really altogether new experiences.

Amazon has built a chatbot with generative AI and it has lifted customer satisfaction by 500 basis points; Amazon has built a generative AI application for 3rd-party sellers to easily fill up their product detail pages; Amazon has built generative AI applications for inventory management that improve inventory forecasting by 10% and regional predictions by 20%; the brains of Amazon’s robotics are infused with generative AI

If you look at customer service and you look at the chatbot that we’ve built, we completely rearchitected it with generative AI. It’s delivering. It already had pretty high satisfaction. It’s delivering 500 basis points better satisfaction from customers with the new generative AI-infused chatbot.

If you look at our millions of third-party selling partners, one of their biggest pain points is, because we put a high premium on really organizing our marketplace so that it’s easy to find things, there’s a bunch of different fields you have to fill out when you’re creating a new product detail page, but we’ve built a generative AI application for them where they can either fill in just a couple of lines of text or take a picture of an image or point to a URL, and the generative AI app will fill in most of the rest of the information they have to fill out, which speeds up getting selection on the website and easier for sellers.

If you look at how we do inventory management and trying to understand what inventory we need, at what facility, at what time, the generative AI applications we’ve built there have led to 10% better forecasting on our part and 20% better regional predictions.

In our robotics, we were just talking about the brains in a lot of those robotics are generative AI-infused that do things like tell the robotic claw what’s in a bin, what it should pick up, how it should move it, where it should place it in the other bin that it’s sitting next to. So it’s really in the brains of most of our robotics.

Amazon’s Rufus is an AI-infused shopping assistant that is growing significantly; users can take a picture of a product with Amazon Lens and have the service surface the exact item through the use of AI; Amazon is using AI to know the relative sizing of clothes and shoes from different brands so that it can recommend the right sizes to shoppers; Amazon is using AI to improve the viewing experience of sporting events; Rufus provides a significant improvement to the shopping experience for shoppers and management expects the usage of Rufus to increase throughout 2025

You see lots of those in our retail business, ranging from Rufus, which is our AI-infused shopping assistant, which continues to grow very significantly; to things like Amazon Lens, where you can take a picture of a product that’s in front of you, you check it out in the app, you can find it in the little box at the top, you take a picture of an item in front of you, and it uses computer vision and generative AI to pull up the exact item in search result; to things like sizing, where we basically have taken the catalogs of all these different clothing manufacturers and then compare them against one another so we know which brands tend to run big or small relative to each other. So when you come to buy a pair of shoes, for instance, it can recommend what size you need; to even what we’re doing in Thursday Night Football, where we’re using generative AI for really inventive features like it sends alerts where we predict which players are going to put quarterback or defensive vulnerabilities, where we were able to show viewers what area of the field is vulnerable…

…I do think that Rufus, if you look at how it impacts the customer experience and if you actually use it month-to-month, it continues to get better and better. If you’re buying something and you’re on our product detail page, our product detail pages provide so much information that sometimes it’s hard, if you’re trying to find something quickly, to scroll through and find that little piece of information. And so we have so many customers now who just use Rufus to help them find a quick fact about a product. They also use Rufus to figure out how to summarize customer reviews so they don’t have to read 100 customer reviews to get a sense of what people think about that product. If you look at the personalization, really, most prominently today, your ability to go into Rufus and ask what’s happened to an order or what did I just order or can you pull up for me this item that I ordered 2 months ago, the personalization keeps getting much better. And so we expect throughout 2025, that the number of occasions where you’re not sure what you want to buy and you want help from Rufus are going to continue to increase and be more and more helpful to customers.

Amazon has around 1,000 generative AI applications that it has built or is building

We’ve got about 1,000 different generative AI applications we’ve either built or in the process of building right now.

Apple (NASDAQ: AAPL)

Apple Intelligence was first released in the USA in October 2024, with more features and countries introduced in December 2024; Apple Intelligence will be rolled out to even more countries in April 2025; management sees Apple Intelligence as a breakthrough for privacy in AI; SAP is using Apple Intelligence in the USA to improve the employee as well as customer experience; the Apple Intelligence features that people are using include Writing Tools, Image Playground, Genmoji, Visual Intelligence, Clean Up, and more; management has found Apple Intelligence’s email summarisation feature to be very useful; management thinks that different users will find their own “killer feature” within Apple Intelligence

In October, we released the first set of Apple Intelligence features in U.S. English for iPhone, iPad and Mac, and we rolled out more features and expanded to more countries in December.

Now users can discover the benefits of these new features in the things they do every day. They can use Writing Tools to help find just the right words, create fun and unique images with Image Playground and Genmoji, handle daily tasks and seek out information with a more natural and conversational Siri, create movies of their memories with a simple prompt and touch up their photos with Clean Up. We introduced visual intelligence with Camera Control to help users instantly learn about their surroundings. Users can also seamlessly access ChatGPT across iOS, iPadOS and macOS.

And we were excited to recently begin our international expansion with Apple Intelligence now available in Australia, Canada, New Zealand, South Africa and the U.K. We’re working hard to take Apple Intelligence even further. In April, we’re bringing Apple Intelligence to more languages, including French, German, Italian, Portuguese, Spanish, Japanese, Korean and simplified Chinese as well as localized English to Singapore and India. And we’ll continue to roll out more features in the future, including an even more capable Siri.

Apple Intelligence builds on years of innovations we’ve made across hardware and software to transform how users experience our products. Apple Intelligence also empowers users by delivering personal context that’s relevant to them. And importantly, Apple Intelligence is a breakthrough for privacy in AI with innovations like Private Cloud Compute, which extends the industry-leading security and privacy of Apple devices into the cloud…

…We’re excited to see leading enterprises such as SAP leverage Apple Intelligence in the U.S. with features like Writing Tools, summarize and priority notifications to enhance both their employee and customer experiences…

…In terms of the features that people are using, they’re using all of the ones that I had referenced in my opening comments, from Writing Tools to Image Playground and Genmoji, to visual intelligence and more. And so we see all of those being used. Clean Up is another one that is popular, and people love seeing that one demoed in the stores as well…

…I know from my own personal experience, once you start using the features, you can’t imagine not using them anymore. I know I get hundreds of e-mails a day, and the summarization function is so important…

…[Question] Do you guys see the upgraded Siri expected in April as something that will, let’s say, be the killer application among the suite of features that you have announced in Apple Intelligence?

[Answer] I think the killer feature is different for different people. But I think for most, they’re going to find that they’re going to use many of the features every day. And certainly, one of those is the — is Siri, and that will be coming over the next several months.

Many customers are excited about the iPhone 16 because of Apple Intelligence; the iPhone 16’s year-on-year performance was stronger in countries where Apple Intelligence was available compared to countries where Apple Intelligence was not available

Our iPhone 16 lineup takes the smartphone experience to the next level in so many ways, and Apple Intelligence is one of many reasons why customers are excited…

…We did see that the markets where we had rolled out Apple Intelligence, that the year-over-year performance on the iPhone 16 family was stronger than those where Apple Intelligence was not available…

Apple’s management thinks the developments in the AI industry brought on by DeepSeek’s emergence is a positive for Apple

[Question] There’s a perception that you’re a big beneficiary of lower cost of compute. And I was wondering if you could give your worldly perspective here on the DeepSeek situation.

[Answer] In general, I think innovation that drives efficiency is a good thing. And that’s what you see in that model. Our tight integration of silicon and software, I think, will continue to serve us very well.

Arista Networks (NYSE: ANET)

Cloud and AI titans were a significant contributor to Arista Networks’ revenue in 2024; management considers Oracle an AI titan too

Now shifting to annual sector revenue for 2024. Our cloud and AI titans contributed significantly at approximately 48%, keeping in mind that Oracle is a new member of this category.

Arista Networks’ core cloud AI and data center products are built off its extensible OS (operating system) and goes up to 800 gigabit Ethernet speeds

Our core cloud AI and data center products are built off a highly differentiated, extensible OS stack and is successfully deployed across 10, 25, 100, 200, 400 and 800 gigabit Ethernet speeds. It delivers power efficiency, high availability, automation and agility as the data centers demand, insatiable bandwidth capacity and network speeds for both front-end and back-end storage, compute and AI zones.

Arista Networks’ management expects the company’s 800 gigabit Ethernet switch to emerge as an AI back-end cluster in 2025

We expect 800 gigabit Ethernet to emerge as an AI back-end cluster in 2025.

Arista Networks’ management is still optimistic that AI revenues will reach $1.5 billion in 2025, including $750 million in AI back-end clusters; the $750 million in revenue from AI back-end clusters will have a major helping hand from 3 of the 5 major AI trials Arista Networks is working on that are rolling out a cumulative 100,000 GPUs in 2025 (see more below)

We remain optimistic about achieving our AI revenue goal of $1.5 billion in AI centers, which includes the $750 million in AI back-end clusters in 2025…

…[Question] You are reiterating $750 million AI back-end sales this year despite the stalled or the fifth customer. Can you talk about where is the upside coming from this year? Is it broad-based or 1 or 2 customers?

[Answer] We’re well on our way and 3 customers deploying a cumulative of 100,000 GPUs is going to help us with that number this year. And as we increased our guidance to $8.2 billion, I think we’re going to see momentum both in AI, cloud and enterprises. I’m not ready to break it down and tell you which where. I think we’ll see — we’ll know that much better in the second half. But Chantelle and I feel confident that we can definitely do the $8.2 billion that we historically don’t call out so early in the year. So having visibility if that helps.

Arista Networks is building some of the world’s greatest Arista AI centers at production scale and it’s involved with both the back-end clusters and front-end networks; Arista Networks’ management sees the data traffic flow of AI workloads as having significant differences from traditional cloud workloads and Arista AI centers can seamlessly connect to the front end compute storage with its backend Ethernet portfolio; Arista’s AI networking portfolio consists of 3 families and over 20 Etherlink switches

Networking for AI is also gaining traction as we move into 2025, building some of the world’s greatest Arista AI centers at production scale. These are constructed with both back-end clusters and front-end networks…

…The fidelity of the AI traffic differs greatly from cloud workloads in terms of diversity, duration and size of flow. Just one slow flow can flow the entire job completion time for a training workload. Therefore, Arista AI centers seamlessly connect to the front end of compute storage WAN and classic cloud networks with our back-end Arista Etherlink portfolio. This AI accelerated networking portfolio consists of 3 families and over 20 Etherlink switches, not just 1 point switch.

Arista Networks’ management’s AI for Networking strategy is doing well and it includes software that have superior AI ops

Our AI for networking strategy is also doing well, and it’s about curating the data for higher-level network functions. We instrument our customer’s networks with our published subscribed state Foundation with our software called Network Data Lake to deliver proactive, predictive and prescriptive platforms that have superior AI ops with A care support and product functions.

Arista Networks’ management is still committed to 4 of the 5 major AI trials that they have been discussing in recent earnings calls; the remaining AI trial is still stalled and the customer is not a Cloud Titan and is waiting for funding; 3 of the 4 trials that are active are expected to roll out a cumulative 100,000 GPUs in 2025 and they are all waiting for the next-generation NVIDIA GPU; Arista Networks’ management expects to do very well on the back-end with those 3 trials; the remaining trial of the 4 active trials is migrating from Infiniband to Ethernet to test the viability of Ethernet, and Arista Networks’ management expects to enter production in 2026

I want to say Arista is still committed to 4 out of our 5 AI clusters that I mentioned in prior calls, but just one is a little bit stalled. It is not a Cloud Titan. They are awaiting GPUs and some funding too, I think. So I hope they’ll come back next year, but for this year, we won’t talk about them. But the remaining 4, let me spend some — jgive you some color, 3 out of the 4 customers are expected to this year rolled out a cumulative of 100,000 GPUs. So we’re going to do very well with 3 of them on the back end. And you can imagine, they’re all pretty much one major NVIDIA class of GPU — it’s — they will be waiting for the next generation of GPUs. But independent of that, we’ll be rolling out fairly large numbers. On the fourth one, we are migrating right now from InfiniBand to proving that Ethernet is a viable solution, so we’re still — they’ve historically been InfiniBand. And so we’re still in pilot and we expect to go into production next year. We’re doing very well in 4 out of 4, the Fifth one installed and 3 out of the 4 expected to be 100,000 GPUs this year.

Arista Networks thinks the market for AI networking is large enough that there will be enough room for both the company and other whitebox networking manufacturers; management also thinks Arista Networks’ products have significant differentiation from whitebox products, especially in the AI spine in a typical leaf-spine network, because Arista Networks’ products can automatically provide an alternate connection when a GPU in the network is in trouble

[Question] Can you maybe share your perspective that when it comes to AI network especially the back-end networks, how do you see the mix evolving white box versus OEM solution?

[Answer] This TAM is so huge and so large. We will always coexist with white boxes and operating systems that are non-EOS, much like Apple coexists on the iPhone with other phones of different types. When you look at the back end of an AI cluster, there are typically 2 components, the AI lead and the AI spine. The AI lead connects to the GPUs and therefore, is the first, if you will, point of connection. And the AI spine aggregates all of these AI leads. Almost in all the back-end examples we’ve seen, the AI spine is generally 100% Arista-branded EOS. You’ve got to do an awful lot of routing, scale, features, capabilities that are very rich that would be difficult to do in any other environment. The AI leads can vary. So for example, the — let’s take the example of the 5 customers I mentioned a lot, 3 out of the 5 are all EOS in the [indiscernible] spine. 2 out of the 5 are kind of hybrids. Some of them have some form of SONic or FBOSS. And as you know, we co-develop with them and coexist in a number of use cases where it’s a real hybrid combination of EOS and an open OS. So for most part, I’d just like to say that white box and Arista will coexist and will provide different strokes for different folks…

…A lot of our deployments right now is 400 and 800 gig, and you see a tremendous amount of differentiation, not only like I explained to you in scale and routing features, but cost and load balancing, AI visibility and analytics at real time, personal queuing, congestion control, visibility and most importantly, smart system upgrade because you sort of want your GPUs to come down because you don’t have the right software to accelerate so that the network provides the ideal foundation that if the GPU is in trouble, we can automatically give a different connection and an alternate connection. So tremendous amount of differentiation there and even more valid in a GPU which costs typically 5x as much as a CPU…

…When you’re buying these expensive GPUs that cost $25,000, they’re like diamonds, right? You’re not going to string a diamond on a piece of thread. So first thing I want to say is you need a mission-critical network, whether you want to call it white box, blue box, EOS or some other software, you’ve got to have mission-critical functions, analytics, visibility, high availability, et cetera. As I mentioned, and I want to reiterate, they’re also typically a leaf spine network. And I have yet to see an AI spine deployment that is not EOS-based. I’m not saying it can’t happen or won’t happen. But in all 5 major installations, the benefit of our EOS features for high availability for routing, for VXLAN, for telemetry, our customers really see that. And the 7800 is the flagship AI spine product that we have been deploying last year, this year and in the future. Coming soon, of course, is also the product we jointly engineered with Meta, which is the distributed [Ecolink] switch. And that is also an example of a product that provides that kind of leaf spine combination, both with FBOSS and EOS options in it. So in my view, it’s difficult to imagine a highly resilient system without Arista EOS in AI or non-AI use cases.

On the leaf, you can cut corners. You can go with smaller buffers, you may have a smaller installation. So I can imagine that some people will want to experiment and do experiment in smaller configurations with non-EOS. But again, to do that, you have to have a fairly large staff to build the operations for it. So that’s also a critical element. So unless you’re a large Cloud Titan customer, you’re less likely to take that chance because you don’t have the staff.

Arista Networks’ management is seeing strong demand from its Cloud Titan customers

Speaking specifically to Meta, we are obviously in a number of use cases in Meta. Keep in mind that our 2024 Meta numbers is influenced by more of their 2023 CapEx, and that was Meta’s year of efficiency where their CapEx was down 15% to 20%. So you’re probably seeing some correlation between their CapEx being down and our revenue numbers being slightly lower in ’24. In general, I would just say all our cloud titans are performing well in demand, and we shouldn’t confuse that with timing of our shipments. And I fully expect Microsoft and Meta to be greater than 10% customers in a strong manner in 2025 as well. Specific to the others we added in, they’re not 10% customers, but they’re doing very well, and we’re happy with their cloud and AI use cases.

Arista Networks’ management thinks the emergence of DeepSeek will lead to AI development evolving from back-end training that’s concentrated in a handful of users, to being distributed more widely across CPUs and GPUs; management also thinks DeepSeek’s emergence is a positive for Arista Networks because DeepSeek’s innovations can drive the AI industry towards a new class of CPUs, GPUs, AI accelerators and Arista Networks is able to scale up network for all kinds of XPUs

DeepSeek certainly deep fixed many stocks, but I actually see this as a positive because I think you’re now going to see a new class of CPUs, GPUs, AI accelerators and where you can have substantial efficiency gains that go beyond training. So that could be some sort of inference or mixture of experts or reasoning and which lowers the token count and therefore, the cost. So what I like about all these different options is Arista can scale up network for all kinds of XPUs and accelerators. And I think the eye-opening thing here for all of our experts who are building all these engineering models is there are many different types and training isn’t the only one. So I think this is a nice evolution of how AI will not just be a back-end training only limited to 5 customers type phenomenon, but will become more and more distributed across a range of CPUs and GPUs.

Arista Networks’ management thinks hyper-scale GPU clusters, such as Project Stargate, will drive the development of vertical rack integration in the next few years and Andy Bechtolsheim, an Arista Networks co-founder, is personally involved in these projects

If you look at how we have classically approached GPUs and connected libraries, we’ve largely looked at it as 2 separate building blocks. There’s the vendor who provides the GPUs and then there’s us who provides the scale-out networking. But when you look at Stargate and projects like this, I think you’ll start to see more of a vertical rack integration where the processor, the scale up, the scale out and all of the software to provide a single point of control and visibility starts to come more and more together. This is not a 2025 phenomenon, but definitely in ’26 and ’27, you’re going to see a new class of AI accelerators for — and a new class of training and inference, which is extremely different than the current more pluggable label type of version. So we’re very optimistic about it.

Andy Bechtolsheim is personally involved in the design of a number of these next-generation projects and the need for this type of shall we say, pushing Moore’s Law of improvements in density of performance that we saw in the 2000s is coming back, and you can boost more and more performance per XPU, which means you have to boost the network scale from 800 gig to 1.16.

Arista Networks’ management sees a $70 billion total addressable market in 2028, of which roughly a third is related to AI

[Question] If you can talk to the $70 billion TAM number for 2028, how much is AI?

[Answer] On the $70 billion TAM in 2028, I would roughly say 1/3 is AI, 1/3 is data center and cloud and 1/3 is campus and enterprise. And obviously, absorbed into that is routing and security and observability. I’m not calling them out separately for the purpose of this discussion.

Arista Networks’ management sees co-packaged optics (CPO) as having weak adoption compared to co-packaged copper (CPC) because CPO has been experiencing field failures

Co-packaged optics is not a new idea. It’s been around 10 to 20 years. So the fundamental reason, let’s go through why co-packaged optics has had a relatively weak adoption so far is because of field failures and most of it is still in proof of concept today. So going back to networking, the most important attribute of a network switch is reliability and troubleshooting. And once you solder a co-packaged optics on a PCB, you lose some of that flexibility and you don’t get the serviceability and manufacturing. That’s been the problem. Now a number of alternatives are emerging, and we’re a big fan of co-packaged copper as well as pluggable optics that can complement this like linear drive or LTO as we call it.

Now we also see that if co-packaged optics improves some of the metrics it has right now. For example, it has a higher channel count than the industry standard of 8-channel pluggable optics, but we can do higher channel pluggable optics as well. So some of these things improve, we can see that both CPC and CPO will be important technologies at 224 gig or even 448 gig. But so far, our customers have preferred a LEGO approach that they can mix and match pluggable switches and pluggable optics and haven’t committed to soldering them on the PCB. And we feel that will change only if CPO gets better and more reliable. And I think CPC can be a nice alternative to that.

Arista Networks’ management is seeing customers start moving towards actual use-cases for AI, but the customers are saying that these AI projects take time to implement

For the AI perspective, speaking with the customers, it’s great to move from kind of a theory to more specific conversation, and you’re seeing that in the banks and some of the higher tier Global 2000, Fortune 500 companies. And so they’re moving from theory to actual use cases they’re speaking to. And the way they describe it is it takes a bit of time. They’re working mostly with cloud service providers at the beginning, kind of doing some training and then they’re deciding whether they bring that on-prem and inference. So they’re making those decisions.

Arista Networks’ management is seeing a new class of Tier 2 specialty AI cloud providers emerge

We are seeing a new class of Tier 2 specialty cloud providers emerge that want to provide AI as a service and want to be differentiated there. And there’s a whole lot of funding, grant money, real money going in there. So service providers, too early to call. But Neo clouds and specialty providers, yes, we’re seeing lots of examples of that.

The advent of AI has accelerated the speed-transitions in networking data switches, but there’s still going to be a long runway for Arista Networks’ 400 gig and 800 gig products, with 1.6 tera products being deployed in a measured way

The speed transitions because of AI are certainly getting faster. It used to take when we went from 200 gig, for example, at Meta or 100 gig in some of our Cloud Titans to 400, that speed transition typically took 3 to 4, maybe even 5 years, right? In AI, we see that cycle being almost every 2 years…

…2024 was the year of real 400 gig. ’25 and ’26, I would say, is more 800 gig. And I really see 1.6T coming into the picture because we don’t have chips yet, maybe in what do you say, John, late ’26 and real production maybe in ’27. So there’s a lot of talk and hype on it, just like I remember talk and hype on 400 gig 5 years ago. But I think realistically, you’re going to see a long runway for 400 and 800 gig. Now as we get into 1.6T, part of the reason I think it’s going to be measured and thoughtful is many of our customers are still awaiting their own AI accelerators or NVIDIA GPUs, which with liquid cooling that would actually push that kind of bandwidth. So new GPUs will require new bandwidth, and that’s going to push it out a year or 2.

Arista Networks’ management sees a future where the market share between NVIDIA GPUs and custom AI accelerators (ASICs) is roughly evenly-split, but Arista Networks’ products will be GPU-agnostic

[Question] There’s been a lot of discussion over the last few months between the general purpose GPU clusters from NVIDIA and then the custom ASIC solutions from some of your popular customers. I guess just in your view, over the longer term, does Arista’s opportunity differ across these 2 chip types?

[Answer] I think I’ve always said this, you guys often spoke about NVIDIA as a competitor. And I don’t see it that way. I see that — thank you, NVIDIA. Thank you, Jensen, for the GPUs because that gives us an opportunity to connect to them, and that’s been a predominant market for us. As we move forward, we see not only that we connect to them, but we can connect to AMD GPUs and built in in-house AI accelerators. So a lot of them are in active development or in early stages. NVIDIA is the dominant market share holder with probably 80%, 90%. But if you ask me to guess what it would look like 2, 3 years from now, I think it could be 50-50. So Arista could be the scale-out network for all types of accelerators. We’ll be GPU agnostic. And I think there’ll be less opportunity to bundle by specific vendors and more opportunity for customers to choose best-of-breed. 

ASML (NASDAQ: ASML)

AI will be the biggest driver of ASML’s growth and management sees customers benefiting very strongly from it; management thinks ASML will hit the upper end of the revenue guidance range for 2025 if its customers can bring on additional AI-related capacity during the year, but there are also risks that could result in only the lower end of the guidance coming true

We see total revenue for 2025 between €30 billion and €35 billion and the gross margin between 51% and 53%. AI is the clear driver. I think we started to see that last year. In fact, at this point, we really believe that AI is creating a shift in the market and we have seen customers benefiting from it very strongly…

…If AI demand continues to be strong and customers are successful in bringing on additional capacity online to support that demand, there is potential opportunity towards the upper end of our range. On the other hand, there are also risks related to customers and geopolitics that could drive results towards the lower end of the range.

ASML’s management is still very positive on the long-term outlook for ASML, with AI being a driver for growth; management expects AI to create a shift in ASML’s end-market products towards more HPC (high performance computing) and HBM (high bandwidth memory), which requires more advanced logic and DRAM, which in turn needs more critical lithography exposures

I think our view on the long term is still, I would say, very positive…

…Looking longer term, overall the semiconductor market remains strong with artificial intelligence creating growth but also a shift in market dynamics as I highlighted earlier. These dynamics will lead to a shift in the mix of end market products towards more HPC and HBM which requires more advanced logic and DRAM. For ASML, we anticipate that an increased number of critical lithography exposures for these advanced Logic and Memory processes will drive increasing demand for ASML products and services. As a result, we see a 2030 revenue opportunity between 44 billion euros and 60 billion euros with gross margins expected between 56 percent and 60 percent, as we presented in Investor Day 2024.

ASML’s management is seeing aggressive capacity addition among some DRAM memory customers because of demand for high bandwidth memory (HBM), but apart from HBM, other DRAM memory customers have a slower recovery

 I think that high-bandwidth memory is driving today, I would say, also an aggressive capacity addition, at least for some of the customer. I think on the normal DRAM, I would say, my comment is similar to the one on mobile [ photology ] before. I think there are also nothing spectacular, but there is some recovery, which also called for more capacity. So that’s why we still see DRAM pretty strong in 2025.

Datadog (NASDAQ: DDOG)

Datadog launched LLM Observability in 2024; management continues to see increased interest in next-gen AI; 3,500 Datadog customers at the end of 2024 Q4 used 1 or more Datadog AI integrations; when it comes to AI inference, management is seeing most customers using a 3rd-party AI model through an API or a 3rd-party inference platform, and these customers want to observe whether the model is doing the right thing, and this need is what LLM Observability is serving; management is seeing very few customers running the full AI inference stack currently, but they think this could happen soon and it would be an exciting development

We launched LLM observability, in general availability to help customers evaluate, safely deploy and manage their models in production, and we continue to see increased interest in next-gen AI. At the end of Q4, about 3,500 customers use 1 or more Datadog AI integrations to send this data about their machine learning, AI, and LLM usage…

…On the inference side, the — mostly still what customers do is they use a third-party model either through an API or through a third-party inference platform. And what they’re interested in is measuring whether that model is doing the right thing. And that’s what we serve right now with LLM observability, for example, as well, we see quite a bit of adoption that does not come largely from the AI-native companies. So that’s what we see today.

In terms of operating the inference stack fully and how we see relatively few customers with that yet, we think that’s something that’s going to come next. And by the way, we’re very excited by the developments we see in the space. So it looks like there is many, many different options that are going to be viable for running your AI inference. There’s a very healthy set of commercial API-gated services. There’s models that you can install in the open source. There are models in the open source today that are rivalling in quality with the best closed API models. So we think the ecosystem is developing into a rich diversification that will allow customers to have a diversity of modalities for using AI, which is exciting. 

AI-native customers accounted for 6% of Datadog’s ARR in 2024 Q4 (was 6% 2024 Q3); AI-native customers contributed 5 percentage points to Datadog’s year-on-year growth in 2024 Q4, compared to 3 percentage points in 2023 Q4; among customers in the AI-native cohort, management has seen optimisation of usage and volume discounts related to contracts in 2024 Q4, and management thinks these customers will continue to optimise cloud and observability usage in the future; the dynamics with the AI-native cohort that happened in 2024 Q4 was inline with management’s expectations

We continue to see robust contribution from AI native customers who represented about 6% of Q4 ARR roughly the same as the quarter — as last quarter and up from about 3% of ARR in the year-ago quarter. AI native customers contributed about 5 percentage points of year-over-year revenue growth in Q4 versus 4 points in the last quarter and about 3 points in the year-ago quarter. So we saw strong growth from AI native customers in Q4. We believe that adoption of AI will continue to benefit Datadog in the long term. Meanwhile, we did see some optimization and volume discounts related to contract renewals in Q4. We remain mindful that we may see volatility in our revenue growth on the backdrop of long-term volume growth from this cohort as customers renew with us on different terms, and as they may choose to optimize cloud and observability usage. ..

… [Question] I’m trying to understand if the AI usage and commits are kind of on the same trajectory that they were on or whether you feel that there are some oscillations there.

[Answer] What happened during the quarter is pretty much what we thought would happen when we discussed it in the last earnings call. When you look at the AI cohort, we definitely saw some renewals with higher commit, better terms and optimization usage all at the same time, which is fairly typical, which typically happens with larger end customers in particular is at the time of renewal, customers are going to trying and optimize what they can. They’re going to get better prices from us, up their commitments and we might see a flat or down a month or quarter after that, with a still sharp growth from the year before and growth to come in the year to come. So that’s what we typically see. When you look at the cohort as a whole, even with that significant renewal optimization and better unit economics this quarter is wholly stable, this quarter as a whole is stable quarter-to-quarter in its revenue and it’s growing a lot from the quarter before, even with all that.

Datadog’s management sees some emerging opportunities in Infrastructure Monitoring that are related to the usage of GPUs (Graphics Processing Units), but the opportunities will only make sense if there is broad usage of GPUs by a large number of customers, which is not happening today

There’s a number of new use cases that are emerging that are related to infrastructure that we might want to cover. Again, we — when I say they’re emerging, they’re actually emerging, like we still have to see what the actual need is from a large number of customers. I’m talking in particular about infrastructure concerns around GPU management, GPU optimization, like there’s quite a lot going on there that we can potentially do. But we — for that, we need to see broad usage of the raw GPUs by a large number of customers as opposed to usage by a smaller number of native customers, which is mostly what we still see today.

Datadog’s management thinks it’s hard to tell where AI agents can be the most beneficial for Datadog’s observability platform because it’s still a very nascent field and management has observed that things change really quickly; when management built LLM Observability, the initial use cases were for AI chatbots and RAG (retrieval augmented generation), but now the use cases are starting to shift towards AI agents

[Question] Just curious, when we think about agents, which parts of the core observability platform that you think are most relevant or going to be most beneficial to your business as you start to monitor those?

[Answer] It’s a bit hard to tell because it’s a very nascent field. So my guess is in a year if we probably look different from what it looks like today. Just like this year, it looks very different from what it looks like last year. What we do see, though, is that — so when we built — we started building our LLM Observability product, most of the use cases we saw there from customers were chatbot in nature or RAG in nature, trying to access information and return the information. Now we see more and more customers building agents on top of that and sending the data from their agents. So we definitely see a growing trend there of adoption and the LLM Observability product is a good level of abstraction, at least for the current iteration of these agents to get them. So that’s what we can see today.

Datadog’s management sees AI touching many different areas of Datadog, such as how software is being written and deployed, how customer-support is improved, and more

What’s fascinating about the current evolution of AI, in particular, is that it touches a lot of the different areas of the business. The first area for company like ours the first area to be transformed is really the way software is being built. What engineers use, how they write software, how they debug software, how do they also operate systems. And part of that is outside tooling we’re using for writing software. Part of that is dogfooding, or new products for incident resolution and that sort of thing. So that’s the first area. There’s a number of other areas that are going to see large improvements in productivity. Typically, everything that has to do with supporting customers, helping with onboarding and helping troubleshoot issues like all of that is in acceleration. In the end, we expect to see improvements everywhere, from front office to back office.

Fiverr (NYSE: FVRR)

Fiverr’s management launched Fiverr Go in February 2025, an open platform for personalised AI tools designed to give creators full control over their creative processes and rights; Fiverr Go enables freelancers to build personalised AI models (there was a presentation on this recently) without having to know AI engineering; Fiverr Go is designed to be personalised for the creator, so the creator becomes more important compared to the AI technology; Fiverr Go is generative AI technology with human accountability (will be interesting to see if Fiverr Go is popular; people can create designs/images with other AI models, so customers who use Fiverr Go are those who need the special features that Fiverr Go offers); Fiverr Go generates content that is good enough for mission critical business tasks, unlike what’s commonly happening with other AI-generated content; Fiverr Go is no different from a direct order from the freelancer themself, except it is faster and easier for buyers; Fiverr Go has personalised AI assistants for freelancers; Fiverr Go has an open developer platform for 3rd-party developers to build generative AI apps

Fiverr Go is an open platform for personalized AI tools that include the personalized AI assistant and the AI creation model. Different from other AI platforms that often exploit human creativity without proper attribution or compensation, Fiverr Go is uniquely designed to reshape this power dynamic by giving creators full control over their creative process and rights. It also enables freelancers to build personalized AI models without the need to collect training data sets or understand AI engineering, thanks to Fiverr’s unparalleled foundation of over 6.5 billion interactions and nearly 150 million transactions on the marketplace and most importantly, it allows freelancers to become a one-person production house, making more money while focusing on the things that matter. By giving freelancers control over configuration, pricing and creative rights and leveling the playing field of implementing AI technology, Fiverr Go ensures that creators remain at the center of the creative economy. It decisively turned the power dynamic between humans and AI towards the human side…

For customers, Fiverr Go is also fundamentally different from other AI platforms. It is GenAI with human accountability. AI results often feel unreliable, generic and very hard to edit. What is good enough for a simple question and answer on ChatGPT does not cut it for business mission-critical tasks. In fact, many customers come to Fiverr today with AI-generated content because they miss the confidence that comes from another human eye and talent, helping them perfect the results for their needs. Fiverr Go eliminates all of this friction and frustration. Every delivery on Fiverr Go is backed by the full faith of the creator behind it with an included revision as the freelancer defines. This means that the quality and service you get from Fiverr Go is no different from a direct order from the freelancers themselves, except for a faster, easier and more delightful experience. The personalized AI assistant on Fiverr Go can communicate with potential clients when the freelancer is away or busy, handle routine tasks and provide actionable business insights, effectively becoming the creator’s business partner. It often feels smarter than an average human assistant because it’s equipped with all the history of how the freelancer works as well as knowledge of trends and best practices on the Fiverr marketplace…

…We’ve also announced an open developer platform on Fiverr Go to allow AI specialists and model developers to build generative AI applications across any discipline. We provide them with an ecosystem to collaborate with domain experts on Fiverr and the ability to leverage Fiverr’s massive data assets so that these applications can solve real-world problems and most important of all, Fiverr provides them an avenue to generate revenue from those applications through our marketplace…

…So from our experience with AI, what we come to learn is that a lot of the creation process using AI is very random and take you through figuring out what are the best tools because there’s thousands of different options around AI. And each one operates slightly differently. And you need to master each one of them. And you need to become a prompt engineer. And then editing is extremely, extremely hard. Plus you don’t get the feedback that comes from working with a human being that can actually look at the creation from a human eye and give you a sense if this is actually capturing what you’re trying to do. It allows us or allows freelancers to design their own model in a way that rewards them but remains extremely accurate to their style, allowing customers to get the results they expect to get because they see the portfolio of their freelancer, like the style of writing or design or singing or narration, and they can get exactly this. So we think that, that combination and that confidence that comes from the fact that the creator itself is always there.

The AI personal assistant in Fiverr Go can help to respond to customer questions based on individual freelancers; the first 3 minutes after a buyer writes to a freelancer is the most crucial time for conversion and this is where the AI assistant can help; there are already thousands of AI assistants running on Fiverr Go, converting customers

Fiverr Go is actually a tool for conversion. That’s the entire idea because we know that customers these days expect instant responses and instant results. And as a result of that, we designed those 2 tools, the AI personal assistant, which is able to answer customer questions immediately even if the freelancer is away or busy. We know that the first 3 minutes after a customer writes to a freelancer are the most crucial time for conversion and this is why we designed this tool. And this tool is essentially encapsulating the entire knowledge of the freelancer and basing itself on it, being able to address any possible question and bring it to conversion…

…It’s fresh from yesterday, but we have many thousands of assistants running on the system, converting customers already, which is an amazing sign.

Fiverr Go is a creator tool that can create works based off freelancers’ style and allows customers to get highly-accurate samples of a freelancers’ work to lower friction in selecting freelancers

When we think about the creation model, the creation model allows customers to get the confidence that this is the freelancer, this is the style that they’re looking for, because now instead of asking a freelancer for samples, waiting for it, causing the freelancer to essentially work for free, they can get those samples right away. Now the quality of these samples is just mind-blowing. The level of accuracy that these samples produce are exact match with the style of the freelancer, which gives the customer the confidence that if they played and liked it, this is the type of freelancer that they should engage with.

The Fiverr Go open developer platform is essentially an app store for AI apps; the open developer platform allows developers to train AI models on Fiverr’s transactional data set, which is probably the largest dataset of its kind in existence

Now what we’re doing with this is actually we’re opening up the Go platform to outside developers. Think about it as an app store in essence. So what we’re doing is we’re allowing them to develop models, APIs, workflows, but then train those models on probably the biggest transactional data set in existence today that we hold so that they can actually help us build models that freelancers can join — can enjoy from. And we believe that by doing so and giving those developers incentives to do so because every time their app is going to be used for a transaction, they’re going to make money out of it.

Fiverr Go’s take rate will be the same for now and management will learn as they go

[Question] Would your take rate be different in Fiverr Go?

[Answer] For now, the take rate remains the same for Go. And as we roll it out and as we see usage, we will figure out what to do or what’s the right thing to do. For now, we treat it as a normal transaction with the same take rate.

Mastercard (NYSE: MA)

Mastercard closed the Recorded Future acquisition in 2024 Q4 (Recorded Future provides AI-powered solutions for real-time visibility into potential threats related to fraud); Recorded Future has been deploying AI for over a decade, just like Mastercard has; Recorded Future uses AI to analyse threat data across the entire Internet; the acquisition of Recorded Future improves Mastercard’s cybersecurity solutions

Our diverse capabilities in payments and services and solutions, including the acquisition of Recorded Future this quarter set us apart…

…Recorded Future is the world’s largest threat intelligence company with more than 1,900 customers across 75 countries. Customers include over 50% of the Fortune 100 and government agencies in 45 countries, including more than half of the G20. We’ve been deploying AI at scale for well over a decade, so has Recorded Future. They leverage AI-powered insights to analyze threat data from every corner of the Internet and customers gain real-time visibility and actionable insights to proactively reduce risks. We now have an even more robust set of powerful intelligence, identity, dispute, fraud and scan prevention solutions. Together, these uniquely differentiated technologies will enable us to create smarter models, distribute these capabilities more broadly and help our customers anticipate threats before cyber-attacks can take place. That means better protection for governments, businesses, banks, consumers the entire ecosystem and well beyond the payment transactions. We’re also leveraging our distribution at scale to deepen market penetration of our services and solutions.

Meta Platforms (NASDAQ: META)

Meta’s management expects Meta AI to be the leading AI assistant in 2025, reaching more than 1 billion people; Meta AI is already the most-used AI assistant in the world with more than 700 million monthly actives; management believes Meta AI is at a scale that allows it to develop a durable long-term advantage; management has an exciting road map for Meta AI in 2025 that focuses on personalisation; management does not believe that there’s going to be only one big AI that is the same for everyone; there are some fun surprises for Meta AI in 2025 that management has up their sleeves; Meta AI can now remember certain details of people’s prior queries; management sees a few possible paths for Meta AI’s monetisation, but their focus right now is just on building a great user experience; WhatsApp has the strongest Meta AI usage, followed by Facebook; people are using Meta AI on WhatsApp for informational, educational, and emotional purposes 

 In AI, I expect that this is going to be the year when a highly intelligent and personalized AI assistant reaches more than 1 billion people, and I expect Meta AI to be that leading AI assistant. Meta AI is already used by more people than any other assistant. And once a service reaches that kind of scale, it usually develops a durable long-term advantage.

We have a really exciting road map for this year with a unique vision focused on personalization. We believe that people don’t all want to use the same AI. People want their AI to be personalized to their context, their interests, their personality, their culture, and how they think about the world. I don’t think that there’s just going to be one big AI that everyone uses that does the same thing. People are going to get to choose how their AI works and what it looks like for them. I continue to think that this is going to be one of the most transformative products that we’ve made, and we have some fun surprises that I think people are going to like this year…

… Meta AI usage continues to scale with more than 700 million monthly actives. We’re now introducing updates that will enable Meta AI to deliver more personalized and relevant responses by remembering certain details from people’s prior queries and considering what they engage with on Facebook and Instagram to develop better intuition for their interest and preferences…

…Our initial focus for Meta AI is really about building a great consumer experience, and that’s frankly, where all of our energies are kind of directed to right now. There will, I think, be pretty clear monetization opportunities over time, including paid recommendations and including a premium offering, but really not where we are focused in terms of the development of Meta AI today…

…WhatsApp continues to see the strongest Meta AI usage across our family of apps. People there are using it most frequently for information seeking and educational queries along with emotional support use cases. Most of the WhatsApp engagement is in one-on-one threads, though we see some usage in group messaging. And on Facebook, which is the second largest driver of Meta AI engagement, we’re seeing strong engagement from our feed deep dives integration that lets people ask Meta AI questions about the content that is recommended to that. So across, I would say, all query types, we continue to see signs that Meta AI is helping people leverage our apps for new use cases. We talked about information gathering, social interaction and communication Lots of people use it for humor and casual conversation. They use it for writing and editing research recommendations. 

Meta’s management thinks Llama will become the most advanced and widely-used AI model in 2025; Llama 4 is making great progress; Meta has a reasoning Llama model in the works; management’s goal for Llama 4 is for it be the leading AI model; Llama 4 is built to be multi-modal and agentic; management expects Llama 4 to unlock a lot of new use cases

I think this will very well be the year when Llama and open-source become the most advanced and widely used AI models as well. Llama 4 is making great progress in training, Llama 4 Mini is doing with pretraining and our reasoning models and larger model are looking good too. 

Our goal with Llama 3 was to make open source competitive with closed models. And our goal for Llama 4 is to lead. Llama 4 will be natively multimodal. It’s an omni model, and it will have agenetic capabilities. So it’s going to be novel, and it’s going to unlock a lot of new use cases.

Meta’s management thinks it will be possible in 2025 to build an AI engineering agent that is as capable as a human mid-level software engineer; management believes that the company that builds this AI engineering agent first will have a meaningful advantage in advancing AI research; Meta already has internal AI coding assistants, powered by Llama; management has no plan to release the AI engineer as an external product anytime soon, but sees the potential for it in the longer-term; management does not expect the AI engineer to be extremely widely deployed in 2025, with the dramatic changes happening in 2026 and beyond

I also expect that 2025 will be the year when it becomes possible to build an AI engineering agent that has coding and problem-solving abilities of around a good mid-level engineer. And this is going to be a profound milestone and potentially one of the most important innovations in history, like as well as over time, potentially a very large market. Whichever company builds this first, I think it’s going to have a meaningful advantage in deploying it to advance their AI research and shape the field…

…As part of our efficiency focus over the past 2 years, we’ve made significant improvements in our internal processes and developer tools and introduce new tools like our AI-powered coding assistant, which is helping our engineers write code more quickly. Looking forward, we expect that the continuous advancements in Llama’s coding capabilities will provide even greater leverage to our engineers, and we are focused on expanding its capabilities to not only assist our engineers in writing and reviewing our code but to also begin generating code changes to automate tool updates and improve the quality of our code base…

…And then the AI engineer piece, I’m really excited about it. I mean, I don’t know that that’s going to be an external product anytime soon. But I think for what we’re working on, our goal is to advance AI research and advance our own development internally. And I think it’s just going to be a very profound thing. So I mean that’s something that I think will show up through making our products better over time. But — and then as that works, there will potentially be a market opportunity down the road. But I mean, for now and this year, we’re really — I think this is — I don’t think you’re going to see this year like an AI engineer that is extremely widely deployed, changing all of development. I think this is going to be the year where that really starts to become possible and lays the groundwork for a much more dramatic change in ’26 and beyond.

The Ray-Ban Meta AI glasses are a big hit so far but management thinks 2025 will be the pivotal year to determine if the AI glasses can be on a path towards being the next computing platform and selling hundreds of millions, or more, units; management continues to think that glasses are the perfect form factor for AI; management is optimistic about AI glasses, but there’s still uncertainty about the long-term trajectory

Our Ray-Ban Meta AI glasses are a real hit. And this will be the year when we understand the trajectory for AI glasses as a category. Many breakout products in the history of consumer electronics have sold 5 million to 10 million units and they’re third generation. This will be a defining year that determines if we’re on a path towards many hundreds of millions and eventually billions of AI glasses and glasses being the next computing platform like we’ve been talking about for some time or if this is just going to be a longer grind. But it’s great overall to see people recognizing that these glasses are the perfect form factor for AI as well as just great stylish glasses…

…There are a lot of people in the world who have glasses. It’s kind of hard for me to imagine that a decade or more from now, all the glasses aren’t going to basically be AI glasses as well as a lot of people who don’t wear glasses today, finding that to be a useful thing. So I’m incredibly optimistic about this…

…But look, the Ray-Ban Metas were hit. We still don’t know what the long-term trajectory for this is going to be. And I think we’re going to learn a lot this year. 

Meta will bring ~1 gigawatt of AI data center capacity online in 2025 and is building an AI data center that is at least 2 gigawatts in capacity; management intends to fund the investments through revenue growth that is driven by its AI advances; most of Meta’s new headcount growth will go towards its AI infrastructure and AI advances; management expects compute will be very important for the opportunities they want Meta to pursue; management is simultaneously growing Meta’s capacity and increasing the efficiency of its workloads; Meta is increasing the useful lives of its non-AI and AI servers to 5.5 years (from 4-5 years previously), which will lead to lower depreciation expenses per year; Meta started deploying its own MTIA (Meta Training and Inference Accelerator) AI chips in 2024 for inference workloads; management expects to ramp up MTIA usage for inference in 2025 and for training workloads in 2026; management will continue to buy third-party AI chips (likely referring to NVIDIA), but wants to use in-house chips for unique workloads; management thinks MTIA helps Meta achieve greater compute efficiency and performance per cost and power; management has been thinking about the balance of compute used in pre-training versus inference, but this does not mean that Meta will need less compute; management thinks that inference-time compute (or test-time compute) scaling will help Meta deliver a higher quality of service and that Meta has a strong business model to support the delivery of inference-time compute scaling; management believes that investing heavily in AI infrastructure is still going to be a strategic advantage over time, but it’s possible the reverse may be true in the future; management thinks it’s too early to tell what the long-run capacity intensity will look like

I announced last week that we expect to bring online almost a gigawatt of capacity this year. And we’re building a 2 gigawatt and potentially bigger AI data center that is so big that it will cover a significant part of Manhattan if we were placed there. We’re planning to fund all of this by, at the same time, investing aggressively in initiatives that use these AI advances to increase revenue growth…

…That’s what a lot of our new headcount growth is going towards and how well we execute on this will also determine our financial trajectory over the next few years…

…We expect compute will be central to many of the opportunities we’re pursuing as we advance the capabilities of Llama, drive increased usage of generative AI products and features across our platform and fuel core ads and organic engagement initiatives. We’re working to meet the growing capacity needs for these services by both scaling our infrastructure footprint and increasing the efficiency of our workloads…

…Our expectation going forward is that we’ll be able to use both our non-AI and AI [indiscernible] servers for a longer period of time before replacing them, which we estimate will be approximately 5.5 years. This will deliver savings in annual CapEx and resulting depreciation expense, which is already included in our guidance.

Finally, we’re pursuing cost efficiencies by deploying our custom MTIA silicon in areas where we can achieve a lower cost of compute by optimizing the chip to our unique workloads. In 2024, we started deploying MTIA to our ranking and recommendation inference workloads for ads and organic content. We expect to further ramp adoption of MTIA for these use cases throughout 2025, before extending our custom silicon efforts to training workloads for ranking and recommendations next year…

…We expect that we are continuing to purchase third-party silicon from leading providers in the industry. And we are certainly committed to those long-standing partnerships, but we’re also very invested in developing our own custom silicon for unique workloads, where off-the-shelf silicon isn’t necessarily optimal and specifically because we’re able to optimize the full stack to achieve greater compute efficiency and performance per cost and power because our workloads might require a different mix of memory versus network, bandwidth versus compute and so we can optimize that really to the specific needs of our different types of workloads.

Right now, the in-house MTIA program is focused on supporting our core ranking and recommendation inference workloads. We started adopting MTIA in the first half of 2024 for core ranking and recommendations in [indiscernible]. We’ll continue ramping adoption for those workloads over the course of 2025 as we use it for both incremental capacity and to replace some GPU-based servers when they reach the end of their useful lives. Next year, we’re hoping to expand MTIA to support some of our core AI training workloads and over time, some of our Gen AI use cases…

…There’s already sort of a debate around how much of the compute infrastructure that we’re using is going to go towards pretraining versus as you get more of these reasoning time models or reasoning models where you get more of the intelligence by putting more of the compute into inference, whether just will mix shift how we use our compute infrastructure towards that. That was already something that I think a lot of the — the other labs and ourselves were starting to think more about and already seemed pretty likely even before this, that — like of all the compute that we’re using, that the largest pieces aren’t necessarily going to go towards pre-training. But that doesn’t mean that you need less compute because one of the new properties that’s emerged is the ability to apply more compute at inference time in order to generate a higher level of intelligence and a higher quality of service, which means that as a company that has a strong business model to support this, I think that’s generally an advantage that we’re now going to be able to provide a higher quality of service than others who don’t necessarily have the business model to support it on a sustainable basis…

…I continue to think that investing very heavily in CapEx and infra is going to be a strategic advantage over time. It’s possible that we’ll learn otherwise at some point, but I just think it’s way too early to call that…

…I think it is really too early to determine what long-run capital intensity is going to look like. There are so many different factors. The pace of advancement in underlying models, how efficient can they be? What is the adoption and use case of our Gen AI products, what performance gains come from next-generation hardware innovations, both our own and third party and then ultimately, what monetization or other efficiency gains our AI investments unlock. 

In 2024 H2, Meta introduced a new machine learning system for ads ranking, in partnership with Nvidia, named Andromeda; Andromeda has enabled a 10,000x increase in the complexity of AI models Meta uses for ads retrieval, driving an 8% increase in quality of ads that people see; Andromeda can process large volumes of ads and positions Meta well for a future where advertisers use the company’s generative AI tools to create and test more ads

In the second half of 2024, we introduced an innovative new machine learning system in partnership with NVIDIA called Andromeda. This more efficient system enabled a 10,000x increase in the complexity of models we use for ads retrieval, which is the part of the ranking process where we narrow down a pool of tens of millions of ads to the few thousand we consider showing someone. The increase in model complexity is enabling us to run far more sophisticated prediction models to better personalize which ads we show someone. This has driven an 8% increase in the quality of ads that people see on objectives we’ve tested. Andromeda’s ability to efficiently process larger volumes of ads also positions us well for the future as advertisers use our generative AI tools to create and test more ads.

Advantage+ has surpassed a $20 billion annual revenue run rate and grew 70% year-on-year in 2024 Q4; Advantage+ will now be turned on by default for all campaigns that optimise for sales, app, or lead objectives; more than 4 million advertisers are now using at least one of Advantage+’s generative AI ad creative tools, up from 1 million six months ago; Meta’s first video generation tool, released in October, already has hundreds of thousands of advertisers using it monthly

 Adoption of Advantage+ shopping campaigns continues to scale with revenues surpassing a $20 billion annual run rate and growing 70% year-over-year in Q4. Given the strong performance and interest we’re seeing in Advantage+ shopping and our other end-to-end solutions, we’re testing a new streamlined campaign creation flow. So advertisers no longer need to choose between running a manual or Advantage+ sales or app campaign. In this new setup, all campaigns optimizing for sales, app or lead objectives will have Advantage+ turned on from the beginning. This will allow more advertisers to take advantage of the performance Advantage+ offers while still having the ability to further customize aspects of their campaigns when they need to. We plan to expand to more advertisers in the coming months before fully rolling it out later in the year.

Advantage+ Creative is another area where we’re seeing momentum. More than 4 million advertisers are now using at least one of our generative AI ad creative tools, up from 1 million six months ago. There has been significant early adoption of our first video generation tool that we rolled out in October, Image Animation, with hundreds of thousands of advertisers already using it monthly.

Meta’s management thinks the emergence of DeepSeek makes it even more likely for a global open source standard for AI models to develop; the presence of DeepSeek also makes management think it’s important that the open source standard be made in America and that it ’s even more important for Meta to focus on building open source AI models; Meta is learning from DeepSeek’s innovations in building AI models; management currently does not have a strong opinion on how Meta’s capex plans for AI infrastructure will change because of the recent news with DeepSeek 

I also just think in light of some of the recent news, the new competitor DeepSeek from China, I think it also just puts — it’s one of the things that we’re talking about is there’s going to be an open source standard globally. And I think for our kind of national advantage, it’s important that it’s an American standard. So we take that seriously, and we want to build the AI system that people around the world are using and I think that if anything, some of the recent news has only strengthened our conviction that this is the right thing for us to be focused on…

…I can start on the DeepSeek question. I think there’s a number of novel things that they did that I think we’re still digesting. And there are a number of things that they have advances that we will hope to implement in our systems. And that’s part of the nature of how this works, whether it’s a Chinese competitor or not…

…It’s probably too early to really have a strong opinion on what this means for the trajectory around infrastructure and CapEx and things like that. There are a bunch of trends that are happening here all at once.

Meta’s capex in 2025 is going to grow across servers, data centers, and networking; within each of servers, data centers, and networking, management expects growth in both AI and non-AI capex; management expects most of the AI-related capex in 2025 to be directed specifically towards Meta’s core AI infrastructure, but the infrastructure Meta is building can support both AI and non-AI workloads, and the GPU servers purchased can be used for both generative AI and core AI purposes

[Question] As we think about the $60 billion to $65 billion CapEx this year, does the composition change much from last year when you talked about servers as the largest part followed by data centers and networking equipment. And how should we think about that mix between like training and inference

[Answer] We certainly expect that 2025 CapEx is going to grow across all 3 of those components you described.

Servers will be the biggest growth driver that remains the largest portion of our overall CapEx budget. We expect both growth in AI capacity as we support our gen AI efforts and continue to invest meaningfully in core AI, but we are also expecting growth in non-AI capacity as we invest in the core business, including to support a higher base of engagement and to refresh our existing servers.

On the data center side, we’re anticipating higher data center spend in 2025 to be driven by build-outs of our large training clusters and our higher power density data centers that are entering the core construction phase. We’re expecting to use that capacity primarily for core AI and non-AI use cases.

On the networking side, we expect networking spend to grow in ’25 as we build higher-capacity networks to accommodate the growth in non-AI and core AI-related traffic along with our large Gen AI training clusters. We’re also investing in fiber to handle future cross-region training traffic.

And then in terms of the breakdown for core versus Gen AI use cases, we’re expecting total infrastructure spend within each of Gen AI, non-AI and core AI to increase in ’25 with the majority of our CapEx directed to our core business with some caveat that, that is — that’s not easy to measure perfectly as the data centers we’re building can support AI or non-AI workloads and the GPU-based servers, we procure for gen AI can be repurposed for core AI use cases and so on and so forth.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Alphabet, Amazon, Apple, ASML, Coupang, Datadog, Fiverr, Mastercard, and Meta Platforms. Holdings are subject to change at any time.

Insights From Warren Buffett’s 2024 Shareholder’s Letter

There’s much to learn from Warren Buffett’s latest letter, including his thoughts on the P/C (property and casualty) insurance industry, and how to think about shares in public-listed as well as private companies.

One document I always look forward to reading around this time of the year is Warren Buffett’s annual Berkshire Hathaway shareholder’s letter. Over the weekend, Buffett published the 2024 edition and here are some of my favourite insights from it that I wish to document and share. 

Without further ado (emphases are Buffett’s)…

Could the use of the word “mistakes” frequently in company reports be a signal to find great investing opportunities?

During the 2019-23 period, I have used the words “mistake” or “error” 16 times in my letters to you. Many other huge companies have never used either word over that span. Amazon, I should acknowledge, made some brutally candid observations in its 2021 letter. Elsewhere, it has generally been happy talk and pictures.

I have also been a director of large public companies at which “mistake” or “wrong” were forbidden words at board meetings or analyst calls. That taboo, implying managerial perfection, always made me nervous (though, at times, there could be legal issues that make limited discussion advisable. We live in a very litigious society.)   

It’s hard to strike a bad deal when you’re dealing with a great person, even when the deal is vaguely-worded

Let me pause to tell you the remarkable story of Pete Liegl, a man unknown to most Berkshire shareholders but one who contributed many billions to their aggregate wealth. Pete died in November, still working at 80.

 I first heard of Forest River – the Indiana company Pete founded and managed – on June 21, 2005. On that day I received a letter from an intermediary detailing relevant data about the company, a recreational vehicle (“RV”) manufacturer…

…I did some checking with RV dealers, liked what I learned and arranged a June 28th meeting in Omaha…

…Pete next mentioned that he owned some real estate that was leased to Forest River and had not been covered in the June 21 letter. Within a few minutes, we arrived at a price for those assets as I expressed no need for appraisal by Berkshire but would simply accept his valuation.

Then we arrived at the other point that needed clarity. I asked Pete what his compensation should be, adding that whatever he said, I would accept. (This, I should add, is not an approach I recommend for general use.)

Pete paused as his wife, daughter and I leaned forward. Then he surprised us: “Well, I looked at Berkshire’s proxy statement and I wouldn’t want to make more than my boss, so pay me $100,000 per year.” After I picked myself off the floor, Pete added: “But we will earn X (he named a number) this year, and I would like an annual bonus of 10% of any earnings above what the company is now delivering.” I replied: “OK Pete, but if Forest River makes any significant acquisitions we will make an appropriate adjustment for the additional capital thus employed.” I didn’t define “appropriate” or “significant,” but those vague terms never caused a problem.

The four of us then went to dinner at Omaha’s Happy Hollow Club and lived happily ever after. During the next 19 years, Pete shot the lights out. No competitor came close to his performance.   

A handful of great decisions can wash away a multitude of mistakes, and then some

Our experience is that a single winning decision can make a breathtaking difference over time. (Think GEICO as a business decision, Ajit Jain as a managerial decision and my luck in finding Charlie Munger as a one-of-a-kind partner, personal advisor and steadfast friend.) Mistakes fade away; winners can forever blossom. 

A person’s educational background has no bearing on his/her ability

One further point in our CEO selections: I never look at where a candidate has gone to school. Never!…

…Not long ago, I met – by phone – Jessica Toonkel, whose step-grandfather, Ben Rosner, long ago ran a business for Charlie and me. Ben was a retailing genius and, in preparing for this report, I checked with Jessica to confirm Ben’s schooling, which I remembered as limited. Jessica’s reply: “Ben never went past 6th grade.”    

Insurance companies are going to face a staggering environmental catastrophe event someday, but insurance companies can still do well if they price their policies appropriately

In general, property-casualty (“P/C”) insurance pricing strengthened during 2024, reflecting a major increase in damage from convective storms. Climate change may have been announcing its arrival. However, no “monster” event occurred during 2024. Someday, any day, a truly staggering insurance loss will occur – and there is no guarantee that there will be only one per annum…

…We are not deterred by the dramatic and growing loss payments sustained by our activities. (As I write this, think wildfires.) It’s our job to price to absorb these and unemotionally take our lumps when surprises develop. It’s also our job to contest “runaway” verdicts, spurious litigation and outright fraudulent behavior.  

EBITDA (earnings before interest, taxes, depreciation, and amortisation) is not a good measure of a company’s profitability

Here’s a breakdown of the 2023-24 earnings as we see them. All calculations are after depreciation, amortization and income tax. EBITDA, a flawed favorite of Wall Street, is not for us.  

A company can pay massive tax bills and yet be immensely valuable

To be precise, Berkshire last year made four payments to the IRS that totaled $26.8 billion. That’s about 5% of what all of corporate America paid. (In addition, we paid sizable amounts for income taxes to foreign governments and to 44 states.)…

…For sixty years, Berkshire shareholders endorsed continuous reinvestment and that enabled the company to build its taxable income. Cash income-tax payments to the U.S. Treasury, miniscule in the first decade, now aggregate more than $101 billion . . . and counting.   

Shares of public-listed companies and private companies should be seen as the same kind of asset class – ownership stakes in businesses

Berkshire’s equity activity is ambidextrous. In one hand we own control of many businesses, holding at least 80% of the investee’s shares. Generally, we own 100%. These 189 subsidiaries have similarities to marketable common stocks but are far from identical. The collection is worth many hundreds of billions and includes a few rare gems, many good-but-far-from-fabulous businesses and some laggards that have been disappointments… 

…In the other hand, we own a small percentage of a dozen or so very large and highly profitable businesses with household names such as Apple, American Express, Coca-Cola and Moody’s. Many of these companies earn very high returns on the net tangible equity required for their operations…

…We are impartial in our choice of equity vehicles, investing in either variety based upon where we can best deploy your (and my family’s) savings…

…Despite what some commentators currently view as an extraordinary cash position at Berkshire, the great majority of your money remains in equities. That preference won’t change. While our ownership in marketable equities moved downward last year from $354 billion to $272 billion, the value of our non-quoted controlled equities increased somewhat and remains far greater than the value of the marketable portfolio. 

Berkshire Hathaway has a great majority of its capital invested in equities, despite what may seem to be otherwise on the surface (this is related to the point above on how shares of public-listed companies and private companies should be both seen as equities); Berkshire will always be deploying the lion’s share of its capital in equities

Despite what some commentators currently view as an extraordinary cash position at Berkshire, the great majority of your money remains in equities. That preference won’t change. While our ownership in marketable equities moved downward last year from $354 billion to $272 billion, the value of our non-quoted controlled equities increased somewhat and remains far greater than the value of the marketable portfolio.

Berkshire shareholders can rest assured that we will forever deploy a substantial majority of their money in equities – mostly American equities although many of these will have international operations of significance. Berkshire will never prefer ownership of cash-equivalent assets over the ownership of good businesses, whether controlled or only partially owned. 

Good businesses will still succeed even if a government bungles its fiscal policy (i.e. government spending), but it’s still really important for a government to maintain a stable currency 

Paper money can see its value evaporate if fiscal folly prevails. In some countries, this reckless practice has become habitual, and, in our country’s short history, the U.S. has come close to the edge. Fixed-coupon bonds provide no protection against runaway currency.

Businesses, as well as individuals with desired talents, however, will usually find a way to cope with monetary instability as long as their goods or services are desired by the country’s citizenry…

…So thank you, Uncle Sam. Someday your nieces and nephews at Berkshire hope to send you even larger payments than we did in 2024. Spend it wisely. Take care of the many who, for no fault of their own, get the short straws in life. They deserve better. And never forget that we need you to maintain a stable currency and that result requires both wisdom and vigilance on your part. 

Capitalism is still a force for good

One way or another, the sensible – better yet imaginative – deployment of savings by citizens is required to propel an ever-growing societal output of desired goods and services. This system is called capitalism. It has its faults and abuses – in certain respects more egregious now than ever – but it also can work wonders unmatched by other economic systems.

P/C (property and casualty) insurance companies often do not know the true cost of providing their services until much later; the act of pricing insurance policies is part art and part science, and requires a cautious (pessimistic?) mindset

When writing P/C insurance, we receive payment upfront and much later learn what our product has cost us – sometimes a moment of truth that is delayed as much as 30 or more years. (We are still making substantial payments on asbestos exposures that occurred 50 or more years ago.)

This mode of operations has the desirable effect of giving P/C insurers cash before they incur most expenses but carries with it the risk that the company can be losing money – sometimes mountains of money – before the CEO and directors realize what is happening.

Certain lines of insurance minimize this mismatch, such as crop insurance or hail damage in which losses are quickly reported, evaluated and paid. Other lines, however, can lead to executive and shareholder bliss as the company is going broke. Think coverages such as medical malpractice or product liability. In “long-tail” lines, a P/C insurer may report large but fictitious profits to its owners and regulators for many years – even decades…

…Properly pricing P/C insurance is part art, part science and is definitely not a business for optimists. Mike Goldberg, the Berkshire executive who recruited Ajit, said it best: “We want our underwriters to daily come to work nervous, but not paralyzed.”   

There are forms of executive compensation that are one-sided in favour of the executives, which can create distorted incentives

Greg, our directors and I all have a very large investment in Berkshire in relation to any compensation we receive. We do not use options or other one-sided forms of compensation; if you lose money, so do we. This approach encourages caution but does not ensure foresight.

Economic growth in a country is a necessary ingredient for the P/C insurance industry to grow

P/C insurance growth is dependent on increased economic risk. No risk – no need for insurance.

Think back only 135 years when the world had no autos, trucks or airplanes. Now there are 300 million vehicles in the U.S. alone, a massive fleet causing huge damage daily

It’s important for an insurance company to know when to shrink its business (when policies are priced inadequately)

No private insurer has the willingness to take on the amount of risk that Berkshire can provide. At times, this advantage can be important. But we also need to shrink when prices are inadequate. We must never write inadequately-priced policies in order to stay in the game. That policy is corporate suicide. 

A P/C insurance company that does not depend on reinsurers has a material cost advantage

All things considered, we like the P/C insurance business. Berkshire can financially and psychologically handle extreme losses without blinking. We are also not dependent on reinsurers and that gives us a material and enduring cost advantage. 

Good investments can be made by just assessing a company’s financial records and buying when prices are low; it can make sense to invest in foreign countries even without having a view on future foreign exchange rates 

It’s been almost six years since Berkshire began purchasing shares in five Japanese companies that very successfully operate in a manner somewhat similar to Berkshire itself. The five are (alphabetically) ITOCHU, Marubeni, Mitsubishi, Mitsui and Sumitomo…

…Berkshire made its first purchases involving the five in July 2019. We simply looked at their financial records and were amazed at the low prices of their stocks…

…At yearend, Berkshire’s aggregate cost (in dollars) was $13.8 billion and the market value of our holdings totaled $23.5 billion…

…Greg and I have no view on future foreign exchange rates and therefore seek a position approximating currency-neutrality.

Berkshire Hathaway will be investing in Japan for a very long time

A small but important exception to our U.S.-based focus is our growing investment in Japan…

…Our holdings of the five are for the very long term, and we are committed to supporting their boards of directors. From the start, we also agreed to keep Berkshire’s holdings below 10% of each company’s shares. But, as we approached this limit, the five companies agreed to moderately relax the ceiling. Over time, you will likely see Berkshire’s ownership of all five increase somewhat. 

It can make sense to borrow to invest for the long haul in foreign countries if you can fix your interest payments at low rates

Meanwhile, Berkshire has consistently – but not pursuant to any formula – increased its yen-denominated borrowings. All are at fixed rates, no “floaters.” Greg and I have no view on future foreign exchange rates and therefore seek a position approximating currency-neutrality…

…We like the current math of our yen-balanced strategy as well. As I write this, the annual dividend income expected from the Japanese investments in 2025 will total about $812 million and the interest cost of our yen-denominated debt will be about $135 million. 


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I currently have a vested interest in Amazon and Apple. Holdings are subject to change at any time.

Potential Bargains In A Niche Corner Of The US Stock Market

Small community banks in the USA undergoing a change in ownership structure could be interesting to look at

I first came across a niche corner of the US stock market known as thrift conversions in January 2024. Upon further research over the subsequent months, I realised it could be an interesting hunting ground for potential bargains. 

For the purpose of this article, thrifts, which have roots in the USA tracing back to the early 19th century, are small community banks in the country that are mutually owned by their depositors. The mutual ownership structure means that these thrifts have no shareholders. As a result, a thrift’s depositors – despite being owners – have no legal way to access its economics. In the 1970s, regulations were introduced to allow thrifts to convert their ownership structure (hence the term “thrift conversions”) and become public-listed companies with shareholders. Today, there are two main ways for thrifts to convert:

  • The first is a standard conversion, where a thrift converts fully into a public-listed entity at one go.
  • The second is a two-step conversion. In the first-step, a thrift converts only a minority interest in itself into a public-listed entity and thus still has a partial mutual ownership structure. In the second-step, a thrift that has undergone the first-step conversion process goes on to convert fully into a public-listed entity. As far as we know, there’s no time limit for a thrift that has undergone the first-step conversion to partake in the second-step of the process.

Subsequently in this article, I will be using the word “conversion”, or other forms of the same word, to refer only to the standard conversion, unless otherwise stated.

A thrift conversion can be thought of as a thrift undergoing an initial public offering (IPO). During a conversion, the incentives of a thrift’s management and those of its would-be shareholders are highly aligned. In the process, a thrift offers shares to management and depositors first; if there’s insufficient demand, the thrift will then offer shares to outsiders. Importantly, management would be buying the thrift’s shares during the conversion at the same price as other would-be shareholders (the other would-be shareholders are the depositors and outsiders; as a reminder, prior to a conversion, a thrift has no shareholders1). This means it’s very likely that management wants a thrift’s shares to have as cheap a valuation as possible during the conversion. Moreover, new capital that’s raised from management and would-be shareholders in the conversion goes directly to the thrift’s coffers. This new capital adds to the thrift’s equity (calculated by deducting the thrift’s liabilities from its assets) that it has built from the profits it has accumulated over time from providing banking services. These features mean that a thrift often becomes a full public-listed entity at a low valuation while having a high equity-to-assets ratio. It’s worth noting that a thrift can conduct share buybacks and sell itself to other financial institutions after the one-year and three-year marks, respectively, from its conversion.2

Investor Jim Royal’s comprehensive book on thrift conversions (referring to both standard and two-step conversions), aptly titled The Zen of Thrift Conversions, referenced a 2016 study by investment bank Piper Jaffray. The study showed that since 1982, thrifts that became full public-listed entities did so at an average price-to-tangible book (P/TB) ratio of just 0.75. After becoming public-listed entities, thrifts tend to continue trading at low P/TB ratios. This is because they also tend to have very low returns on equity – a consequence of them having a high equity-to-assets ratio after their conversion – and a bank with a low return on equity deserves to trade at a low P/TB ratio. But the chronically low P/TB ratio is why thrift conversions could be a fertile space for bargains.

Assuming that converted thrifts have low P/TB ratios of less than 1, those that conduct share buybacks increase their tangible book value per share over time even when they have low returns on equity. Moreover, as mentioned earlier, converted thrifts tend to have high equity-to-asset ratios, which means they have overcapitalised balance sheets and thus have plenty of excess capital to buy back shares without harming their financial health. To top it off, the 2016 study from Piper Jaffray also showed that since 1982, 70% of thrifts were acquired after the third anniversary of them becoming full public-listed entities and these thrifts were acquired at an average P/TB ratio of 1.43 (the median time between them becoming fully public and them being acquired was five years).

The growth in a converted thrift’s tangible book value per share from buybacks, and the potential increase in its P/TB ratio when acquired, could result in a strong annualised return for an investor. For example, consider a thrift conversion with the following traits:

  1. It has a return on equity of 3% in each year;
  2. It has a P/TB ratio that consistently hovers at 0.7;
  3. It buys back 5% of its outstanding shares annually for four years after the first anniversary of its conversion, and;
  4. It gets acquired at a P/TB ratio of 1.4 five years after its conversion

Such a thrift will generate a handsome annualised return of 20% over five years. Investing in the thrift on the third-anniversary of its conversion – when the thrift can legally sell itself to other financial institutions – will result in an even more impressive annualised return of 52% when the thrift’s acquired4. There are also past examples of converted thrifts that go on to produce impressive gains even without being acquired. In his book Beating The Street, Peter Lynch, the famed ex-manager of the Fidelity Magellan Fund, shared many examples. Here’s a sample (emphasis is mine):

“In 1991, 16 mutual thrifts and savings banks came public. Two were taken over at more than four times the offering price, and of the remaining 14, the worst is up 87 percent in value. All the rest have doubled or better, and there are four triples, one 7-bagger, and one 10-bagger. Imagine making 10 times your money in 32 months by investing in Magna Bancorp, Inc., of Hattiesburg, Mississippi.”

But not every thrift conversion leads to a happy ending. Table 1 below shows some pertinent figures of Mid-Southern Bancorp, a thrift which produced a pedestrian return from its second-step conversion in July 2018 to its acquisition by Beacon Credit Union in January 2024.

Table 1

There are a few important things I look out for in thrift conversions5:

  • The equity-to-assets ratio: The higher the better, as it signifies an over-capitalised and strong balance sheet, and would make a thrift look attractive to a would-be acquirer
  • The P/TB ratio: The lower the better, as a P/TB ratio that is materially below 1 will (a) make share buybacks a value-enhancing activity for a thrift’s shareholders, and (b) enhance the potential return for us as investors
  • Share buybacks: The more buybacks that happen at a P/TB ratio below 1, the better, as it is not only value-enhancing, but also indicates that management has a good understanding of capital allocation
  • Non-performing assets as a percentage of total assets: The lower the better, as it signifies a thrift that is conducting its banking business conservatively
  • Net income: If the play is for a potential acquisition of a thrift, we want to avoid a chronically loss-making thrift as consistent losses indicate risky lending practices, but the amount of net income earned by the thrift is not important because an acquirer would be improving the thrift’s operations; if the play is for a thrift to generate strong returns for investors from its underlying business growth, then we would want to see a history of growth in net income and at least a decent return on equity (say, 8% or higher)
  • Change in control provisions: This relates to payouts that a thrift’s management can receive upon being acquired and such information can typically be found in a thrift’s DEF 14-A filing; if management can receive a nice payout when a thrift is acquired, management is incentivised to sanction a sale
  • Management’s compensation: The annual compensation of a thrift’s management should not be high relative to the monetary value of management’s ownership stakes in the thrift

Expanding on the last point of what I look out for, I’ve seen cases of fully-public thrifts with poor long-term business results have management teams with high compensation and relatively low dollar-amounts in ownership stakes. In such cases, I think there’s a low possibility of these thrifts being acquired in a reasonable amount of time to maximise shareholder value because it’s lucrative for the management teams to entrench their positions.

If any of you reading this letter is interested to have deeper conversations about investing in thrifts, please reach out, I would love to engage.

1. Thrifts that undertake the two-step conversion process would have no shareholders prior to the first-step conversion. After the first-step conversion is completed and before the second-step conversion commences, these thrifts would have shareholders who own only a minority economic interest in them.  

2. Thrifts that decide to participate in the second-step of the two-step conversion process after completing the first step can begin share buybacks after the first anniversary of the second-step; they can also be acquired on the third anniversary. 

3. Why would a converted thrift (referring to both standard conversions and two-step conversions) be an attractive acquisition target and be acquired at a premium to its tangible book value? This is because the acquirer of a converted thrift can easily cut significant costs and make more efficient use of the thrift’s overcapitalised balance sheet; this means an acquirer can pay a premium to book value (i.e. a P/TB ratio of more than 1) for a converted thrift and still end up with a good deal. 

4. The potential return of a thrift that has completed the second-step of the two-step conversion process is identical to a thrift that has completed the standard conversion, ceteris paribus. This is because the former has the same important features as the latter, such as the low valuation, the over-capitalised balance sheet, and the possibility of being acquired by other financial institutions at a premium to tangible book value. 

5. What I look out for in a thrift that has completed the standard conversion is the same as what I look out for in a thrift that has completed the second-step of the two-step conversion.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I currently have no vested interest in any company mentioned. Holdings are subject to change at any time.

Great Stocks Can Come From The Worst Industries

A gleaming diamond can be found amongst a lump of coal for those with the ability to spot a true bargain.

I’ve long been sector-agnostic when it comes to the companies I’m interested in because I believe that great companies – and thus, great stocks – can come from anywhere. 

My belief was formed because of something I learnt more than a dozen years ago about the US-based airline, Southwest Airlines. Ned Davis Research was tasked by CNN’s MONEY Magazine in 2002 to find the five US-listed stocks with the highest returns over the past 30 years. The winner was Southwest Airlines with its annualised return of 26% from 1972 to 2002; a $1,000 investment at the start of the period would have become $1 million by the end. What is noteworthy here is airlines were widely regarded back then as businesses with horrendous economics. In Berkshire Hathaway’s 2007 annual shareholders’ letter, Warren Buffett wrote (emphasis is mine): 

The worst sort of business is one that grows rapidly, requires significant capital to engender the growth, and then earns little or no money. Think airlines. Here a durable competitive advantage has proven elusive ever since the days of the Wright Brothers. Indeed, if a farsighted capitalist had been present at Kitty Hawk, he would have done his successors a huge favor by shooting Orville down.

The airline industry’s demand for capital ever since that first flight has been insatiable. Investors have poured money into a bottomless pit, attracted by growth when they should have been repelled by it. And I, to my shame, participated in this foolishness when I had Berkshire buy U.S. Air preferred stock in 1989. As the ink was drying on our check, the company went into a tailspin, and before long our preferred dividend was no longer being paid. But we then got very lucky. In one of the recurrent, but always misguided, bursts of optimism for airlines, we were actually able to sell our shares in 1998 for a hefty gain. In the decade following our sale, the company went bankrupt. Twice.” 

And yet, it was an airline that topped the charts in 2002 for the best-performing US stock in the past 30 years. The timeframe of 30 years is also sufficiently long, such that Southwest Airlines’ gains had to be the result of its business’s excellent long-term performance, and not some fortunate short-term hiccup in the fortunes of its business or its stock price.

A recent study from the highly-regarded investment researcher Michael Maubossin, titled Measuring the Moat: Assessing the Magnitude and Sustainability of Value Creation, bolsters my belief. He found that differences in the return on invested capital (ROIC) between industries is lower than the differences in ROICs of companies within industries. In Mauboussin’s data-set, the industry with the highest median ROIC from 1963 to 2023 is Personal Care Products at around 18%. But within Personal Care Products, the companies have ROICs ranging from a low of around 5% to a high of around 40%. Meanwhile, the Wireless Telecom Services industry has one of the lowest median ROICs at around 1%. Yet, the companies within have ROICs ranging from just below 40% to deeply negative figures. Said another way, the best company in a poor industry (Wireless Telecom Services) still has an excellent business that performs significantly better than the median company in a great industry (Personal Care Products)

I continue to believe that excellent investing opportunities can be found everywhere, so I will, for the foreseeable future, remain sector-agnostic. Sometimes, a gleaming diamond can be found amongst a lump of coal for those with the ability to spot a true bargain.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I currently have no vested interest in any company mentioned. Holdings are subject to change at any time.

What The USA’s Largest Bank Thinks About The State Of The Country’s Economy In Q4 2024

Insights from JPMorgan Chase’s management on the health of American consumers and businesses in the fourth quarter of 2024.

JPMorgan Chase (NYSE: JPM) is currently the largest bank in the USA by total assets. Because of this status, JPMorgan is naturally able to feel the pulse of the country’s economy. The bank’s latest earnings conference call – for the fourth quarter of 2024 – was held earlier this week and contained useful insights on the state of American consumers and businesses. The bottom-line is this: the US economy remains resilient, but two significant risks remain, namely, persistent inflation and dangerous geopolitical conditions 

What’s shown between the two horizontal lines below are quotes from JPMorgan’s management team that I picked up from the call.


1. The US economy remains resilient, with low unemployment and healthy consumer spending; businesses are now more optimistic about the economy

The U.S. economy has been resilient. Unemployment remains relatively low, and consumer spending stayed healthy, including during the holiday season. Businesses are more optimistic about the economy, and they are encouraged by expectations for a more pro-growth agenda and improved collaboration between government and business.

2. Management sees two significant risks, namely, persistent inflation, and the most dangerous geopolitical conditions since World War II; management thinks a high level of optimism is embedded in asset prices; management is focused on being prepared for a wide range of scenarios

Two significant risks remain. Ongoing and future spending requirements will likely be inflationary, and therefore, inflation may persist for some time. Additionally, geopolitical conditions remain the most dangerous and complicated since World War II…

…We think it’s important to acknowledge the tension in the risks and uncertainties in the environment and the degree of optimism embedded in asset prices and expectations. In that context, we remain upbeat about the strength of the franchise, but we are focused on being prepared for a wide range of scenarios.

3. Net charge-offs for the whole bank (effectively bad loans that JPMorgan can’t recover) rose from US$2.2 billion a year ago; Consumer & Community Banking’s net charge-offs rose by US$0.4 billion from a year ago

Credit costs were $2.6 billion, reflecting net charge-offs of $2.4 billion and a net reserve of $267 million…

…In terms of credit performance this quarter, credit costs were $2.6 billion, reflecting net charge-offs of $2.1 billion, up $428 million year-on-year driven by card. The net reserve build was $557 million predominantly driven by higher card revolving balances.

4. JPMorgan’s credit card outstanding loans was up double-digits; management expects card loans to grow in 2025, but at a slower pace than in 2024

Card outstandings were up 11% due to strong account acquisition and revolvers…

… We expect healthy card loan growth again this year but below the 12% pace we saw in 2024 as tailwinds from revolver normalization are largely behind us. 

5. Auto originations were up

In auto, originations were $10.6 billion, up 7%, reflecting higher lease volume on robust new vehicle inventory. 

6. JPMorgan’s investment banking fees had strong growth in 2024 Q4, with strong growth in debt underwriting and equity underwriting fees, signalling higher appetite for capital-markets activity from companies; management is optimistic about companies’ enthusiasm towards capital markets activities

IB fees were up 49% year-on-year, and we ranked #1 with wallet share of 9.3% for 2024. Advisory fees were up 41%, benefiting from large deals and share growth in a number of key sectors. Underwriting fees were up meaningfully with debt up 56% and equity up 54% primarily driven by favorable market conditions. In terms of the outlook for the overall Investment Banking wallet, in light of the positive momentum, we remain optimistic about our pipeline. 

7. Management is seeing companies paydown bank loans and is not seeing loan growth, but the lack of loan growth is not necessarily a negative thing, as it involves companies having wide access to capital markets

Global Corporate and Investment Banking loans were down 2% quarter-on-quarter driven by paydowns and lower short-term financing, primarily offset by originations. In Commercial Banking, middle market loans were also down 2% driven by paydowns, predominantly offset by new originations. And commercial real estate loans were flat as new originations were offset by paydowns…

…I think given the significant improvement in business sentiment and the general optimism out there, you might have expected to see some big open loan growth. We are not really seeing that. I don’t particularly think that’s a negative. I think it’s probably explained by a combination of wide open capital markets and so many of the larger corporates accessing the capital markets and healthy balance sheets in small businesses and maybe some residual caution. And maybe there are some pockets in some industries where some aspects of the policy uncertainty that we might be facing are making them a little bit more cautious than they otherwise would be about what they’re executing in the near term. But we’ll see what the new year brings. The current optimism starts getting tested with reality one way or the other.

8. Management is incorporating interest rate cuts in 2025

We expect 2025 NII ex Markets to be approximately $90 billion. Going through the drivers, as usual, the outlook assumes that rates follow the forward curve. It’s worth noting that the NII decrease is driven by both the cut expected in 2025 and the impact of the 100 basis points of cuts in the back half of 2024. 

9. Management expects credit card net charge-offs in 2025 of 3.6%, up from 3.34% in 2024

On credit, we expect the 2025 card net charge-off rate to be in line with our previous guidance of approximately 3.6%.

10. Management has extra capital for JPMorgan as they think there’s a good chance the bank can deploy the capital at better prices in the future, but they’re not increasing the size of the extra capital

The way we’re thinking about it right now is that we feel very comfortable with the notion that it makes sense for us to have a nice store of extra capital in light of the current environment. We believe there is a good chance that there will be a moment where we get to deploy it at better levels essentially in whatever way than the current opportunities would suggest. And so that feels like a correct kind of strategic and financial decision for us. Having said that, having studied it quite extensively over the last 6 months and have all these debates you would expect, we’ve concluded that we do have enough. We have not [indiscernible]. And given that, we would like to not have the excess grow from here.

11. The mortgage market for housing looks poor given the high interest rates there

You know well the state of the mortgage market given rates. 

12. Management thinks that the biggest sources of risk to the credit market are unemployment and stagflation

Just the biggest driver of credit has been and always will be unemployment, both on the consumer side and it feeds into the corporate side. It feeds into mortgages, subprime, credit card. So really it’s your forecast of unemployment. You have to make your own, which will determine that over time. And so the second thing you said vulnerabilities. It’s unemployment, but the worst case would be stagflation. High rates with higher unemployment will drive higher credit losses literally across the board. I’m not — we’re not predicting that, but you just ask for the vulnerabilities. That’s the vulnerabilities.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I don’t have a vested interest in any company mentioned. Holdings are subject to change at any time.

Where Are US Stocks Headed In 2025?

Beware of bad forecasts (and the most honest forecast you can find)

We’re at the start of a new year in 2025, and there has been a deluge of forecasts in recent weeks for where the US stock market will end the year. 

For me, the most important thing to know about the forecasts is just how often they have been right. Unfortunately, the collective forecasting track record of even the USA’s biggest banks and investment firms have been poor.

Morgan Housel once studied their annual forecasts for the S&P 500 – a widely followed index for the US stock market – from 2000 to 2014. He found that a simple assumption of the S&P 500 going up by 9% a year (the 9% figure was chosen because it represented the index’s long-term annualised return) was more accurate than the forecasts provided by the banks and investment firms; the former was off by an average of 14.1 percentage points per year while the latter was off by 14.7 percentage points per year.

When thinking about the future return of stocks, Housel once wrote that it can be boiled down simply to the “dividend yield + earnings growth +/- change in the earnings multiple (valuations).” I agree, it really is that simple. The dividend yield and earnings growth can be estimated with a reasonable level of accuracy. What’s tricky here is the change in the earnings multiple. Housel explained:

“Earnings multiples reflect people’s feelings about the future. And there’s just no way to know what people are going to think about the future in the future. How could you?”

To compound the problem, over short periods of time, such as a year, it’s the change in the earnings multiple that has an overwhelming impact on how stock prices move. In Housel’s dataset when he was looking at market forecasts, 2002 was a year with one of the largest declines for the S&P 500 – it fell by more than 20%. According to data from economist and Nobel Laureate Robert Shiller, the S&P 500’s earnings actually grew by 12% in 2002. It was the decline in the index’s earnings multiple by 30% from 46 to 33 that led to the sharp drop in its price during the year. The forecasters were predicting that the S&P 500 would increase by a mid-teens percentage in price in 2002, which was close to the index’s earnings growth for the year – I believe what the forecasters failed to anticipate was the sharp drop in the earnings multiple. 

If you really need a forecast for where the US stock market will end up in 2025, check out the table below. It shows where the S&P 500 will be given various assumptions for its earnings growth and its earnings multiple. For reference, the index ended the year at a price level of 5,882 with a price-to-earnings (P/E) ratio of 28. If the S&P 500’s earnings fell by 20% in 2025 and the P/E ratio shrank to 5, we’d be looking at a price level of 840 and a disastrous 86% price decline; if earnings growth was 20%, and the P/E ratio expanded to 40, we’d be looking at a price level of 10,083, and a handsome gain of 71%. 

The table contains a wide range of outcomes. But it’s possible for the S&P 500’s actual performance in 2025 to exceed the boundaries of the table. It’s hard to say where the limits are when it comes to the feelings of market participants. Nonetheless, of all the forecasts you’ve seen and are going to see about the US stock market for 2025, I’m confident the table in this article will be the most honest forecast you can find.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I currently have no vested interest in any companies mentioned. Holdings are subject to change at any time. 

6 Things I’m Certain Will Happen In The Financial Markets In 2025

There are so many things that can happen, but here are six things that I’m certain will happen in the financial markets in 2025.

There are no guarantees in the world of investing… or are there? Here are six things I’m certain will happen in 2025.

1. There will be something huge to worry about in the financial markets.

Peter Lynch is the legendary manager of the Fidelity Magellan Fund who earned a 29% annual return during his 13-year tenure from 1977 to 1990. He once said:

“There is always something to worry about. Avoid weekend thinking and ignore the latest dire predictions of the newscasters. Sell a stock because the company’s fundamentals deteriorate, not because the sky is falling.”

Imagine a year in which all the following happened: (1) The US enters a recession; (2) the US goes to war in the Middle East; and (3) the price of oil doubles in three months. Scary? Well, there’s no need to imagine: They all happened in 1990. And what about the S&P 500? It has increased by more than 1,600% from the start of 1990 to today, even without counting dividends.

There will always be things to worry about. But that doesn’t mean we shouldn’t invest.

2. Individual stocks will be volatile.

From 1997 to today, the maximum peak-to-trough decline in each year for Amazon.com’s stock price ranged from 12.6% to 83.0%. In other words, Amazon’s stock price had suffered a double-digit fall every single year for 27 years. Meanwhile, the same Amazon stock price had climbed by an astonishing 233,924% (from US$0.098 to more than US$229) over the same period. 

If you’re investing in individual stocks, be prepared for a wild ride. Volatility is a feature of the stock market – it’s not a sign that things are broken. 

3. US-China relations will either remain status quo, intensify, or blow over.

“Seriously!?” I can hear your thoughts. But I’m stating the obvious for a good reason: We should not let our views on geopolitical events dictate our investment actions. Don’t just take my words for it. Warren Buffett himself said so. In his 1994 Berkshire Hathaway shareholders’ letter, Buffett wrote (emphases are mine):

“We will continue to ignore political and economic forecasts, which are an expensive distraction for many investors and businessmen. 

Thirty years ago, no one could have foreseen the huge expansion of the Vietnam War, wage and price controls, two oil shocks, the resignation of a president, the dissolution of the Soviet Union, a one-day drop in the Dow of 508 points, or treasury bill yields fluctuating between 2.8% and 17.4%.

But, surprise – none of these blockbuster events made the slightest dent in Ben Graham’s investment principles. Nor did they render unsound the negotiated purchases of fine businesses at sensible prices. 

Imagine the cost to us, then, if we had let a fear of unknowns cause us to defer or alter the deployment of capital. Indeed, we have usually made our best purchases when apprehensions about some macro event were at a peak. Fear is the foe of the faddist, but the friend of the fundamentalist.

A different set of major shocks is sure to occur in the next 30 years. We will neither try to predict these nor to profit from them. If we can identify businesses similar to those we have purchased in the past, external surprises will have little effect on our long-term results.”

From 1994 to the third quarter of 2024, Berkshire Hathaway’s book value per share, a proxy for the company’s intrinsic value – albeit a flawed measure – grew by 13.5% annually. Buffett’s disciplined focus on long-term business fundamentals – while ignoring the distractions of political and economic forecasts – has worked out just fine.

4. Interest rates will move in one of three ways: Sideways, up, or down.

“Again, Captain Obvious!?” Please bear with me. There is a good reason why I’m stating the obvious again.

Much ado has been made about what central banks have been doing, and would do, with their respective economies’ benchmark interest rates. This is because of the theoretical link between interest rates and stock prices.

Stocks and other asset classes (bonds, cash, real estate etc.) are constantly competing for capital. In theory, when interest rates are high, the valuation of stocks should be low, since the alternative to stocks – bonds – are providing a good return. On the other hand, when interest rates are low, the valuation of stocks should be high, since the alternative – again, bonds – are providing a poor return. 

But what does reality say? Here’re important historical data on the actual relationship between interest rates and stocks in the US:  

  • Rising interest rates have been met with rising valuations. According to data from economist and Nobel Laureate Robert Shiller, the US 10-year Treasury yield was 2.3% at the start of 1950. By September 1981, it had risen to 15.3%, the highest rate recorded in Shiller’s dataset. In that same period, the S&P 500’s price-to-earnings (P/E) ratio moved from 7 to 8. In other words, the P/E ratio for the S&P 500 increased slightly despite the huge jump in interest rates. It’s worth noting too that the S&P 500’s P/E ratio of 7 at the start of 1950 was not a result of earnings that were temporarily inflated. Yes, there’s cherry picking with the dates. For example, if I had chosen January 1946 as the starting point, when the US 10-year Treasury yield was 2.2% and the P/E ratio for the S&P 500 was 19, then it would be a case of valuations falling alongside rising interest rates. But this goes to show that while interest rates have a role to play in the movement of stocks, it is far from the only thing that matters.
  • Stocks have climbed in rising interest rate environments. In a September 2022 piece, Ben Carlson, Director of Institutional Asset Management at Ritholtz Wealth Management, showed that the S&P 500 climbed by 21% annually from 1954 to 1964 even when the yield on 3-month Treasury bills (a good proxy for the Fed Funds rate, which is the key interest rate set by the Federal Reserve) surged from around 1.2% to 4.4% in the same period. In the 1960s, the yield on the 3-month Treasury bill doubled from just over 4% to 8%, but US stocks still rose by 7.7% per year. And then in the 1970s, rates climbed from 8% to 12% and the S&P 500 still produced an annual return of nearly 6%.
  • Stocks have done poorly in both high and low interest rate environments, and have also done well in both high and low interest rate environments. Carlson published an article in February 2023 that looked at how the US stock market performed in different interest rate regimes. It turns out there’s no clear link between the two. In the 1950s, the 3-month Treasury bill (which is effectively a risk-free investment, since it’s a US government bond with one of the shortest maturities around) had a low average yield of 2.0%; US stocks returned 19.5% annually back then, a phenomenal gain. In the 2000s, US stocks fell by 1.0% per year when the average yield on the 3-month Treasury bill was 2.7%. Meanwhile, a blockbuster 17.3% annualised return in US stocks in the 1980s was accompanied by a high average yield of 8.8% for the 3-month Treasury bill. In the 1970s, the 3-month Treasury bill yielded a high average of 6.3% while US stocks returned just 5.9% per year. 
  • A cut in interest rates by the Federal Reserve is not guaranteed to be a good or bad event for stocks. Josh Brown, CEO of Ritholtz Wealth Management, shared fantastic data in an August 2024 article on how US stocks have performed in the past when the Federal Reserve lowered interest rates. His data, in the form of a chart, goes back to 1957 and I reproduced them in tabular format in Table 1; it shows how US stocks did in the next 12 months following a rate cut, as well as whether a recession occurred in the same window. I also split the data in Table 1 according to whether a recession had occurred shortly after a rate cut, since eight of the 21 past rate-cut cycles from the Federal Reserve since 1957 took place without an impending recession. Table 2 shows the same data as Table 1 but for rate cuts with a recession; Table 3 is for rate cuts without a recession. What the data show is that US stocks have historically done well, on average, in the 12 months following a rate-cut. The overall record, seen in Table 1, is an average 12-month forward return of 9%. When a recession happened shortly after a rate-cut, the average 12-month forward return is 8%; when a recession did not happen shortly after a rate-cut, the average 12-month forward return is 12%. A recession is not necessarily bad for stocks. As Table 2 shows, US stocks have historically delivered an average return of 8% over the next 12 months after rate cuts that came with impending recessions. It’s not a guarantee that stocks will produce good returns in the 12 months after a rate cut even if a recession does not occur, as can be seen from the August 1976 episode in Table 3.
Table 1; Source: Josh Brown
Table 2; Source: Josh Brown
Table 3; Source: Josh Brown

It turns out that the actual relationship between interest rates and stocks is not as clear-cut as theory suggests. There’s an important lesson here, in that one-factor analysis in finance – “if A happens, then B will occur” – should be largely avoided because clear-cut relationships are rarely seen.  

I also think that time that’s spent watching central banks’ decisions regarding interest rates will be better spent studying business fundamentals. The quality of a company’s business and the growth opportunities it has matter far more to its stock price over the long run than interest rates. 

Sears is a case in point. In the 1980s, the US-based company was the dominant retailer in the country. Morgan Housel wrote in a blog post, Common Plots of Economic History :

“Sears was the largest retailer in the world, housed in the tallest building in the world, employing one of the largest workforces.

“No one has to tell you you’ve come to the right place. The look of merchandising authority is complete and unmistakable,” The New York Times wrote of Sears in 1983.

Sears was so good at retailing that in the 1970s and ‘80s it ventured into other areas, like finance. It owned Allstate Insurance, Discover credit card, the Dean Witter brokerage for your stocks and Coldwell Banker brokerage for your house.”

US long-term interest rates fell dramatically from around 15% in the early-to-mid 1980s to around 3% in 2018. But Sears filed for bankruptcy in October 2018, leaving its shareholders with an empty bag. In his blog post mentioned earlier, Housel also wrote:

“Growing income inequality pushed consumers to either bargain or luxury goods, leaving Sears in the shrinking middle. Competition from Wal-Mart and Target – younger and hungrier – took off.

By the late 2000s Sears was a shell of its former self. “YES, WE ARE OPEN” a sign outside my local Sears read – a reminder to customers who had all but written it off.” 

If you’re investing for the long run, there are far more important things to watch than interest rates.

5. There will be investors who are itching to make wholesale changes to their investment portfolios for 2025.

Ofer Azar is a behavioural economist. He once studied more than 300 penalty kicks in professional football (or soccer) games. The goalkeepers who jumped left made a save 14.2% of the time while those who jumped right had a 12.6% success rate. Those who stayed in the centre of the goal saved a penalty 33.3% of the time.

Interestingly, only 6% of the keepers whom Azar studied chose to stay put in the centre. Azar concluded that the keepers’ moves highlight the action bias in us, where we think doing something is better than doing nothing. 

The bias can manifest in investing too, where we develop the urge to do something to our portfolios, especially during periods of volatility. We should guard against the action bias. This is because doing nothing to our portfolios is often better than doing something. I have two great examples. 

First is a paper published by finance professors Brad Barber and Terry Odean in 2000. They analysed the trading records of more than 66,000 US households over a five-year period from 1991 to 1996. They found out that the most frequent traders generated the lowest returns – and the difference is stark. The average household earned 16.4% per year for the timeframe under study but the active traders only made 11.4% per year.

Second, finance professor Jeremy Siegel discovered something fascinating in the mid-2000s. In an interview with Wharton, Siegel said:

“If you bought the original S&P 500 stocks, and held them until today—simple buy and hold, reinvesting dividends—you outpaced the S&P 500 index itself, which adds about 20 new stocks every year and has added almost 1,000 new stocks since its inception in 1957.”

Doing nothing beats doing something. 

6. There are nearly 8.2 billion individuals in the world today, and the vast majority of us will wake up every morning wanting to improve the world and our own lot in life.

This motivation is ultimately what fuels the global economy and financial markets. There are miscreants who appear occasionally to mess things up, but we should have faith in the collective positivity of humankind. We should have faith in us. The idiots’ mess will be temporary.

Mother Nature threw us a major global health threat in 2020 with COVID-19. But we – mankind – managed to build a vaccine against the disease in record time; Moderna even managed to design its vaccine in just two days. This is a great example of the ingenuity of humanity at work.

To me, investing in stocks is the same as having the long-term view that we humans are always striving, collectively, to improve the world.      

A final word

This article is a little cheeky, because it describes incredibly obvious things, such as “interest rates will move in one of three ways: sideways, up, or down.” But I wrote it in the way I did for a good reason. A lot of seemingly important things in finance – things with outcomes that financial market participants obsess over and try to predict – actually turn out to be mostly inconsequential for long-term investors. Keep this in mind when you read the next “X Things That Will Happen To Stocks in 2025” article.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I currently have a vested interest in Amazon. Holdings are subject to change at any time. 

More Of The Latest Thoughts From American Technology Companies On AI (2024 Q3)

A collection of quotes on artificial intelligence, or AI, from the management teams of US-listed technology companies in the 2024 Q3 earnings season.

Last month, I published The Latest Thoughts From American Technology Companies On AI (2024 Q3). In it, I shared commentary in earnings conference calls for the third quarter of 2024, from the leaders of technology companies that I follow or have a vested interest in, on the topic of AI and how the technology could impact their industry and the business world writ large. 

A few more technology companies I’m watching hosted earnings conference calls for 2024’s third quarter after I prepared the article. The leaders of these companies also had insights on AI that I think would be useful to share. This is an ongoing series. For the older commentary:

Here they are, in no particular order:

Adobe (NASDAQ: ADBE)

Adobe’s management introduced multiple generative AI models in the Firefly family in 2024 and now has a generative video model; Adobe’s generative AI models are designed to be safe for commercial usage; the Firefly models are integrated across Adobe’s software products, which brings value to creative professionals across the world; Firefly has powered 16 billion generations (12 billion in 2024 Q2) since its launch in March 2023 and each month in 2024 Q3 has set a new record in generations; the new Firefly video model is in limited beta, but has already gathered massive customer interest (the model has driven a 70% increase in Premier Pro beta users since its introduction) and will be generally available in early-2025; recent improvements to the Firefly models include 4x faster image generation; enterprises such as Tapestry and Pepsi are using Firefly Services to scale content production; Firefly is the foundation of Adobe’s AI-related innovation; management is using Firefly to drive top-of-funnel user-acquisition for Adobe

2024 was also a transformative year of product innovation, where we delivered foundational technology platforms. We introduced multiple generative AI models in the Adobe Firefly family, including imaging, vector design and, most recently, video. Adobe now has a comprehensive set of generative AI models designed to be commercially safe for creative content, offering unprecedented levels of output quality and user control in our applications…

…The deep integration of Firefly across our flagship applications in Creative Cloud, Document Cloud, and Experience Cloud is driving record customer adoption and usage. Firefly-powered generations across our tools surpassed 16 billion, with every month this past quarter setting a new record…

…We have made major strides with our generative AI models with the introduction of Firefly Image Model 3 enhancements to our vector models, richer design models, and the all-new Firefly Video Model. These models are incredibly powerful on their own and their deep integration into our tools like Lightroom, Photoshop, Premiere, InDesign and Express have brought incredible value to millions of creative professionals around the world…

…The launch of the Firefly Video Model and its unique integration in Premier Pro and limited public beta garnered massive customer interest, and we look forward to making it more broadly available in early 2025.  This feature drove a 70% increase in the number of Premier Pro beta users since it was introduced at MAX. Enhancements to Firefly image, vector, and design models include 4x faster image generation and new capabilities integrated into Photoshop, Illustrator, Premiere Pro and Adobe Express…

…Firefly Services adoption continued to ramp as enterprises such as Pepsi and Tapestry use it to scale content production, given the robust APIs and ease of creating custom models that are designed to be commercially safe…

…This year, we introduced Firefly Services. That’s been — that’s off to a great start. We have a lot of customers that are using that. A couple we talked about on the call include Tapestry. They’re using it for scaled content production. Pepsi, for their Gatorade brand, is enabling their customers to personalize any merchandise that they’re buying in particular, starting with Gatorade bottles. And these have been very, very productive for them, and we are seeing this leveraged by a host of other companies for everything from localization at scale to personalization at scale to user engagement or just raw content production at scale as well…

…You’re exactly right in terms of Firefly is a platform and a foundation that we’re leveraging across many different products. As we talked about, everything from Express and Lightroom and even in Acrobat on mobile for a broad-based but then also in our core Creative products, Photoshop, Illustrator, Premiere. And as we’ve alluded to a number of times on this call, with the introduction of video, even a stand-alone offer for Firefly that we think will be more valuable from a tiering perspective there. And then into Firefly Services through APIs in connection to GenStudio. So we are looking at leveraging the power of this AI foundation in all the activities…

…We see that when we invest in mobile and web, we are getting some very positive signals in terms of user adoption and user conversion rate. So we’re using Firefly very actively to do that.

Adobe’s management has combined content and data in Adobe GenStudio to integrate content creation with marketing, leading to an end-to-end content supply chain solution; the Adobe GenStudio portfolio has a new addition in Adobe GenStudio for Performance Marketing, which has seen strong customer demand since becoming generally available recently; management is expanding the go-to-market teams to sell GenStudio solutions that cut across the Digital Media and Digital Experience segments and early success has been found, with management expecting acceleration in this pipeline throughout FY2025 and beyond

We set the stage to drive an AI content revolution by bringing content and data together in Adobe GenStudio integrating high-velocity creative expression with enterprise activation. The release of Adobe GenStudio for performance marketing integrates Creative Cloud, Express, and Experience Cloud and extends our end-to-end content supply chain solution, empowering freelancers, agencies, and enterprises to accelerate the delivery of content, advertising and marketing campaigns…

…We have brought our Creative and Experience Clouds together through the introduction of Firefly Services and GenStudio, addressing the growing need for scaled content production in enterprises…

… GenStudio enables agencies and enterprises to unlock new levels of creativity and efficiency across content creation and production, workflow and planning, asset management, delivery and activation and reporting and insights. 

Adobe GenStudio for Performance Marketing is a great addition to the GenStudio portfolio, offering an integrated application to create paid social ads, display ads, banners, and marketing e-mails by leveraging preapproved on-brand content. It brings together creative teams that define the foundational requirements of a brand, including guidelines around brand voice, channels, and images with marketing teams that need to deliver numerous content variations with speed and agility. We are seeing strong customer demand for Adobe GenStudio for Performance Marketing since its general availability at MAX…

… We’re expanding our enterprise go-to-market teams to sell these integrated solutions that cut across Digital Media and Digital Experience globally under the new GenStudio umbrella. We have seen early success for this strategy that included Express and Firefly Services in Q4. As we enable our worldwide field organization in Q1, we anticipate acceleration of this pipeline throughout the rest of the year and beyond.

Adobe’s management introduced AI Assistant in Acrobat and Reader in FY2024; users of AI Assistant completed their document-tasks 4x faster on average; AI Assistant is now available across desktop, web, and mobile; management introduced specialised AI for specific document-types and tasks in 2024 Q3 (FY2024 Q4); management saw AI Assistant conversations double sequentially in 2024 Q3; AI Assistant is off to an incredibly strong start and management sees it continuing to accelerate; AI Assistant allows users to have conversations with multiple documents, some of which are not even PDFs, and it turns Acrobat into a general-purpose productivity platform; the rollout of AI Assistant in more languages and documents gives Acrobat’s growth more durability

We took a major step forward in FY ’24 with the introduction of AI Assistant in Acrobat and Reader. AI Assistant and other AI features like Liquid Mode and Firefly are accelerating productivity through faster insights, smarter document editing and integrated image generation. A recent productivity study found that users leveraging AI Assistant completed their document-related tasks 4x faster on average. AI Assistant is now available in Acrobat across desktop, web, and mobile and integrated into our Edge, Chrome, and Microsoft Teams extensions. In Q4, we continued to extend its value with specialized AI for contracts and scanned documents, support for additional languages, and the ability to analyze larger documents…

… We saw AI Assistant conversations double quarter-over-quarter, driving deeper customer value…

… AI Assistant for Acrobat is off to an incredibly strong start and we see it continuing to accelerate…

…One of the big things that I think has been unlocked this year is moving, not just by looking at a PDF that you happen to be viewing, but being able to look at and have a conversation with multiple documents, some of which don’t even have to be PDF. So that transition and that gives us the ability to really take Acrobat and make it more of a general purpose productivity platform…

…The thing I’ll add to that is the durability of that, to your point, in languages, as we roll that out in languages, as we roll it out across multiple documents and as we roll it out in enterprises and B2B specifically. So again, significant headroom in terms of the innovation agenda of how Acrobat can be made even more meaningful as a knowledge tool within the enterprise.  

Adobe’s management will soon introduce a new higher-priced Firefly offering that includes the video models; management thinks the higher-priced Firefly offering will help to increase ARPU (average revenue per user); management sees video generation as a high-value activity, which gives Adobe the ability to introduce higher subscription tiers that come with video generation; management sees consumption of AI services adding to Adobe’s ARR (annual recurring revenue) in 2 ways in FY2025, namely, (1) pure consumption-based pricing, and (2) consumption leading to a higher pricing-tier; management has learnt from pricing experiments for AI services and found that the right model for Adobe is a combination of access to features and usage-limits

We will soon introduce a new higher-priced Firefly offering that includes our video models as a comprehensive AI solution for creative professionals. This will allow us to monetize new users, provide additional value to existing customers, and increase ARPU…

…Video generation is a much higher-value activity than image generation. And as a result, it gives us the ability to start to tier Creative Cloud more actively there…

…You’re going to see “consumption” add to ARR in 2 or maybe 3 ways more so in ’25 than in ’24. The first, and David alluded to this, is if you have a video offering and that video offering, that will be a pure consumption pricing associated with it. I think the second is in GenStudio and for enterprises and what they are seeing. With respect to Firefly Services, which, again, I think David touched on how much momentum we are seeing in that business. So that is, in effect, a consumption business as it relates to the enterprise so I think that will also continue to increase. And then I think you’ll see us with perhaps more premium price offering. So the intention is that consumption is what’s driving the increased ARR, but it may be as a result of a tier in the pricing rather than a consumption model where people actually have to monitor it. So it’s just another way, much like AI Assistant is of monetizing it, but it’s not like we’re going to be tracking every single generation for the user, it will just be at a different tier…

… What we’ve done over the last year, there’s been a bit of experimentation, obviously, in the core Creative applications. We’ve done the generative credits model. What we saw with Acrobat was this idea of a separate package and a separate SKU that created a tier that people were able to access the feature through. And as we learn from all of these, we think, as Shantanu had mentioned earlier, that the right tiering model for us is going to be a combination of feature, access to certain features and usage limits on it. So the higher the tier, the more features you get and the more usage you get of it.

The Adobe Experience Platform (AEP) AI Assistant helps marketers automate tasks and generate new audiences and journeys

Adobe Experience Platform AI Assistant empowers marketers to automate tasks and generate new audiences and journeys. Adobe Experience Manager generates variations, provides dynamic and personalized content creation natively through AEM, enabling customers to deliver more compelling and engaging experiences on their websites.

Adobe’s management thinks there are 3 foundational differences in the company’s AI models and what the rest are doing, namely, (1) commercially safe models, (2) incredible control of the models, and (3) the integration of the models into products

The foundational difference between what we do and what everyone else does in the market really comes down to 3 things: one is commercially safe, the way we train the models; two is the incredible control we bake into the model; and three is the integration that we make with these models into our products, increasingly, of course, in our CC flagship applications but also in Express and Legroom and these kinds of applications but also in Anil’s DX products as well. So that set of things is a critical part of the foundation and a durable differentiator for us as we go forward.

Adobe’s management is seeing that users are onboarded to products faster when using generative AI capabilities; management is seeing that users who use generative AI features have higher retention rates

We are seeing in the core Creative business, when people try something like Photoshop, the onboarding experience is faster to success because of the use of generative AI and generative capabilities. So you’ll start to see us continuing to drive more proliferation of those capabilities earlier in the user journeys, and that has been proven very productive. But we also noticed that more people use generative AI. Again, we’ve always had good retention rates, but the more people use generative AI, the longer they retain as well. 

MongoDB (NASDAQ: MDB)

MongoDB’s management is seeing a lot of large customers want to run workloads, even AI workloads, in on-premise format

We definitely see lots of large customers who are very, very committed to running workloads on-prem. We even see some customers want who are on to run AI workloads on-prem…

… I think you have some customers who are very committed to running a big part of the estate on-prem. So by definition, then if they’re going to build an AI workload, it has to be run on-prem, which means that they also need access to GPUs, and they’re doing that. And then other customers are leveraging basically renting GPUs from the cloud providers and building their own AI workloads.    

MongoDB’s initiative to accelerate legacy app modernisation with AI (Relational Migrator) has seen a 50% reduction in the cost to modernise in its early days; customer interest in this initiative is exceeding management’s expectations; management expects modernisation projects to include large services engagements and MongoDB is increasing its professional services delivery capabilities; management is building new tools to accelerate future monetisation of service engagements; management has growing confidence that the monetisation of modernisation capabilities will be a significant growth driver for MongoDB in the long term; there are a confluence of events, including the emergence of generative AI to significantly reduce the time needed for migration of databases, that make the modernisation opportunity attractive for MongoDB; the buildout of MongoDB’s professional services capabilities will impact the company’s gross margin

We are optimistic about the opportunity to accelerate legacy app modernization using AI and are investing more in this area. As you recall, we ran a few successful pilots earlier in this year, demonstrating that AI tooling combined with professional services and our relational migrator product, can significantly reduce the time, cost and risk of migrating legacy applications on to MongoDB. While it’s early days, we have observed a more than 50% reduction in the cost to modernize. On the back of these strong early results, additional customer interest is exceeding our expectations. 

Large enterprises in every industry and geography are experiencing acute pain from their legacy infrastructure and are eager for more agile performance and cost-effective solutions. Not only our customers excited to engage with us, they also want to focus on some of the most important applications in their enterprise further demonstrating the level of interest and size of the long-term opportunity.

As relational applications encompass a wide variety of database types, programming languages, versions and other customer-specific variables, we expect modernization projects to continue to include meaningful services engagements in the short and medium term. Consequently, we are increasing our professional services delivery capabilities, both directly and through partners. In the long run, we expect to automate and simplify large parts of the modernization process. To that end, we are leveraging the learnings from early service engagements to develop new tools to accelerate future monetization efforts. Although it’s early days and scaling our legacy app monetization capabilities will take time, we have increased conviction that this motion will significantly add to our growth in the long term…

…We’re so excited about the opportunity to go after legacy applications is that, one, it seems like there’s a confluence of events happening. One is that the increasing cost and tax of supporting and managing these legacy apps are just going up enough. Second, for many customers who are in regulated industries, the regulators are calling their the fact that they’re running on these legacy apps a systemic risk, so they can no longer kick the can down the road. Third, also because they no longer kick the can around, some vendors are going end of life, So they have to make a decision to migrate those applications to a more modern tech stack. Fourth, because Gen AI is so predicated on data and to build a competitive advantage, you need to leverage your proprietary data. People want to access that data and be able to do so easily. And so that’s another reason for them to want to modernize…

…we always could help them very easily move the data and map the schema from a relational schema to a document schema. The hardest part was essentially rewriting the application. Now with the advent of GenAI, you can now significantly reduce the time. One, you can use GenAI to analyze the existing code. Two, you can use GenAI to reverse engineer tests to test what the code does. And then three, you can use GenAI to build new code and then use this test to ensure that the new code produce the same results as the old code. And so all that time and effort is suddenly cut in a meaningful way…

…We’re really building out that capacity in order to meet the demand that we’re seeing relative to the opportunity. We’re calling it in particular because it has a gross margin impact because that’s where that will typically show up. 

MongoDB’s management thinks that the company’s database is uniquely suited for the query-rich and complex data structures commonly found in AI applications; AI-powered recommendation systems have to consider complex data structures, beyond a customer’s purchase history; MongoDB’s database unifies source data, metadata, operational data and vector data in all 1 platform, providing a better developer experience; management thinks MongoDB is well-positioned for AI agents because AI agents that perform tasks need to interact with complex data structures, and MongoDB’s database is well-suited for this

MongoDB is uniquely equipped to query-rich and complex data structures typical of AI applications. The ability of a database to query-rich and complex data structures is crucial because AI applications often rely on highly detailed, interrelated and nuanced data to make accurate predictions and decisions. For example, a recommendation system doesn’t just analyze a single customer’s purchase but also considers their browsing history, peer group behavior and product categories requiring a database that can query and ensuring these complex data structures. In addition, MongoDB’s architecture unified source data, metadata, operational data and vector data in all 1 platform, updating the need for multiple database systems and complex back-end architectures. This enables a more compelling developer experience than any other alternative…

…When you think about agents, there’s jobs, there’s sorry, there’s a job, this project and then this task. Right now, the agents that are being rolled out are really focused on task, like, say, something from Sierra or some other companies that are rolling out agents. But you’re right, what they deem to do is to deal with being able to create a rich and complex data structure.

Now why is this important for in AI is that AI models don’t just look at isolated data points, but they need to understand relationships, hierarchies and patterns within the data. They need to be able to essentially get real-time insights. For example, if you have a chat bot where someone’s clearing customers kind of trying to get some update on the order they placed 5 minutes ago because they may have not gotten any confirmation, your chatbot needs to be able to deal with real-time information. You need to be able to deal with basically handling very advanced use cases, understanding like do things like fraud detection, to understand behaviors on supply chain, you need to understand intricate data relationships. All these things are consistent with MongoDB offers. And so we believe that at the end of the day, we are well positioned to handle this.

And the other thing that I would say is that we’ve embedded in a very natural way, search and vector search. So we’re just not an OLTP [online transaction processing] database. We do tech search and vector search, and that’s all one experience and no other platform offers that, and we think we have a real advantage. 

In the AI market, MongoDB’s management is seeing most customers still being in the experimental stage, but the number of AI apps in production is increasing; MongoDB has thousands of AI apps on its platform, but only a small number have achieved enterprise-scale; there’s one AI app on MongoDB’s platform that has grown 10x since the start of 2024 and is a 7-figure workload today; management believes that as AI technology matures, there will be more AI apps that attain product-market fit but it’s difficult to predict when this will happen; management remains confident that MongoDB will capture its share of successful AI applications, as MongoDB is popular with developers building sophisticated AI apps; there are no compelling AI models for smartphones at the moment because phones do not have sufficient computing power

From what we see in the AI market today, most customers are still in the experimental stage as they work to understand the effectiveness of the underlying tech stack and build early proof-of-concept applications. However, we are seeing an increasing number of AI apps in production. Today, we have thousands of AI apps on our platform.  What we don’t yet see is many of these apps actually achieving meaningful product-market fit and therefore, significant traction. In fact, as you take a step back and look at the entire universe of AI apps, a very small percentage of them have achieved the type of scale that we commonly see with enterprise-specific applications. We do have some AI apps that are growing quickly, including one that is already a 7-figure workload that has grown 10x since the beginning of the year.

Similar to prior platform shifts as the usefulness of AI tech improves and becomes more cost-effective we will see the emergence of many more AI apps that do nail product market fit, but it’s difficult to predict when that will happen more broadly. We remain confident that we will capture our fair share of these successful AI applications as we see that our platform is popular with developers building more sophisticated AI use cases…

…Today, we don’t have a very compelling model designed for our phones, right? Because today, the phones don’t have the computing horsepower to run complex models. So you don’t see a ton of very, very successful consumer apps besides, say, ChatGPT or Claude.

MongoDB’s management is building enterprise-grade Atlas Vector Search functionality into the company’s platform so that MongoDB will be in an even better position to win AI opportunities; management is bringing vector search into MongoDB’s community and EA (Enterprise Advance, which is the company’s non-Atlas business) offerings

We continue investing in our product capabilities, including enterprise-grade Atlas Vector Search functionality to build on this momentum and even better position MongoDB to capture the AI opportunity. In addition, as previously announced, we are bringing search and vector service to our community and EA offerings, leveraging our run-anywhere competitive advantage in the world of AI…

…We are investing in our what we call our EA business. First, we’re starting by investing with Search and Vector Search and a community product. That does a couple of things for us. One, whenever anyone starts with MongoDB with the open source product, they need get all the benefits of that complete and highly integrated platform. Two, those capabilities will then migrate to EA. So EA for us is an investment strategy.

MongoDB’s management is expanding the MongoDB AI Applications Program (MAAP); the MAAP has signed on new partners, including with Meta; management expects more of the MAAP workloads to happen on Atlas initially

We are expanding our MongoDB AI Applications program, or MAAP, which helps enterprise customers build and bring AI applications into production by providing them with reference architectures, integrations with leading tech providers and coordinated services and support. Last week, we announced a new cohort of partners, including McKinsey, Confluent, CapGemini and Instructure as well as the collaboration with Meta to enable developers to build arenrich applications on MongoDB using Llama…

…[Question] On the MAAP program, are most of those workloads going to wind up in Atlas? Or will that be a healthy combination of EA and Atlas?

[Answer] I think it’s, again, early days. I would say — I would probably say more on the side of Atlas than EA in the early days. I think once we introduce Search and Vector Search into the EA product, you’ll see more of that on-prem. Obviously, people can use MongoDB for AI workloads using other technologies as well in conjunction with MongoDB for on-prem AI use cases. But I would say you’re probably going to see that happen first in Atlas.

Tealbook consolidated from Postgres, PG Vector, and Elastic Search to MongoDB; Tealbook has seen cost efficiencies and increased scalability with Atlas Vector Search for its application that uses generative AI to collect, verify and enrich supplier data across various sources

Tealbook, a supplier intelligence platform migrated from [ Postgres ], [ PG Vector ] and Elastic Search to MongoDB to eliminate technical debt and consolidate their tech stack. The company experienced workload isolation and scalability issues in PG vector, and we’re concerned with the search index inconsistencies, which were all resolved with the migration to MongoDB. With Atlas Vector search and dedicated search notes, Tealbook has realized improved cost efficiency and increase scalability for the supplier data platform, an application that uses GenAI to collect, verify and enrich supplier data across various sources.

MongoDB’s partnerships with all 3 major cloud providers – AWS, Azure, and GCP – for AI workloads are going well; management expects the cloud providers to bundle their own AI-focused database offerings with their other AI offerings, but management also thinks the cloud providers realise that MongoDB has a better offering and it’s better to partner with the company

With AWS, as you said, they just had their Reinventure last week. It remains very, very strong. We closed a ton of deals this past quarter, some of them very, very large deals. We’re doing integrations to some of the new products like Q and Bedrock and the engagement in the field has been really strong.

On Azure, I think we — as I’ve shared in the past, we start off with a little bit of a slower start. But in the words of the person who runs their partner leadership, the Azure MongoDB relationship has never been stronger. — we closed a large number of deals, we’re part of what’s called the Azure-native IC service program and have a bunch of deep integrations with Azure, including Fabric, Power BI, Visual Studio, Symantec Kernel and Azure OpenAI studio. And we’re also one of Azure’s largest marketplace partners.

And GCP does — we’ve actually seen some uptick in terms of co-sales that we’ve done this past quarter. GCP made some comp changes where that were favorable to working with MongoDB that we saw some results in the field and we’re focused on closing a handful of large deals with GCP in Q4. So in general, I would say things are going quite well.

And then in terms of, I guess, implying your question was like the hyperscalers and are they potentially bundling things along with their AI offerings? I mean, candidly, since day 1, the hyperscalers have been bundling their database offerings with every offering that they have. And that’s been their pretty predominant strategy. And we’ve — I think we’ve executed well against strategy because databases are not like a by-the-way decision. It’s an important decision. And I think the hyperscalers are seeing our performance and realize it’s better to partner with us. And as I said, customers understand the importance of the data layer, especially by our applications. And so the partnership across all 3 hyperscalers is strong.

A new MongoDB AI-related capability called Atlas Search Nodes is seeing very high demand; Atlas Search is being used by one of the world’s largest banks to provide a Google-like Search experience on payments data for customers; an AI-powered accounting software provider is using Atlas Search to allow end-users to perform ad-hoc analysis

On search, we introduced a new capability called Atlas Search nodes, which where you can asymmetrically scale your search nodes because if you have a search intensive use case, you don’t have to scale all your nodes because that have become quite expensive. And we’ve seen that this kind of groundbreaking capability really well received. The demand is quite high. And because customers like they can tune the configuration to the unique needs of their search requirements.

One of the world’s largest banks is using Atlas Search to provide like a Google-like search experience on payments data for massive corporate customers. So there’s a customer-facing application, and so performance and scalability are critical. A leading provider of AI-powered accounting software uses Atlas Search to Power’s invoice analytics future, which allows end users on finance teams to perform ad hoc analysis and easily find past due invoices and voices that contain errors.

Vector Search is only in its first full year of being generally available; uptake of Vector Search has been very high; MongoDB released a feature on Atlas Vector Search in 2024 Q3 that reduces memory requirements by up to 96% and this helps Atlas Vector Search support larger vector workloads at a better price-performance ratio; a multinational news organisation used Vector Search to create a generative AI tool to help producers and journalists sift through vast quantities of information; a security firm is using Vector Search for AI fraud; a global media company replaced Elastic Search with Vector Search for a user-recommendation engine

On Vector Search, again, and it’s been our kind of our first full year since going generally available and the product uptake has been actually very, very high. In Q3, we released quantization for Atlas Vector Search, which reduces the memory requirements by up to 96%, allowing us to support larger Vector workloads with vastly improved price performance.

For example, a multinational news organization created a GenAI powered tool designed to help producers and journalists efficiently search, summarize and verify information from vast and varied data sources. A leading security firm is using Atlas Vector certified AI fraud and a leading global media company replaced elastic search with hybrid search and vector search use case for a user recommendation engine that’s built to suggest that’s building to suggest articles to end users.

MongoDB’s management thinks the industry is still in the very early days of shifting towards AI applications

I do think we’re in the very, very early days. They’re still learning experimenting…  I think as people get more sophisticated with AI as the AI technology matures and becomes more and more useful, I think applications will — you’ll start seeing these applications take off. I kind of chuckle that today, I see more senior leaders bragging about the chips they are using versus the Appstore building. So it just tells you that we’re still in the very, very early days of this big platform shift.

Nvidia (NASDAQ: NVDA)

Nvidia’s Data Center revenue again had incredibly strong growth in 2024 Q3, driven by demand for the Hopper GPU computing platform; Nvidia’s H200 sales achieved the fastest ramp in the company’s history

Another record was achieved in Data Center. Revenue of $30.8 billion, up 17% sequential and up 112% year-on-year. NVIDIA Hopper demand is exceptional, and sequentially, NVIDIA H200 sales increased significantly to double-digit billions, the fastest prod ramp in our company’s history.

Nvidia’s H200 product has 2x faster inference speed, and 50% lower total cost of ownership (TCO)

The H200 delivers up to 2x faster inference performance and up to 50% improved TCO. 

Cloud service providers (CSPs) were half of Nvidia’s Data Centre revenue in 2024 Q3, and up more than 2x year-on-year; CSPs are installing tens of thousands of GPUs to meet rising demand for AI training and inference; Nvidia Cloud Instances with H200s are now available, or soon-to-be-available, in the major CSPs

Cloud service providers were approximately half of our Data Center sales with revenue increasing more than 2x year-on-year. CSPs deployed NVIDIA H200 infrastructure and high-speed networking with installations scaling to tens of thousands of GPUs to grow their business and serve rapidly rising demand for AI training and inference workloads. NVIDIA H200-powered cloud instances are now available from AWS, CoreWeave and Microsoft Azure with Google Cloud and OCI coming soon.

North America, India, and Asia Pacific regions are ramping up Nvidia Cloud Instances and sovereign clouds; management is seeing an increase in momentum of sovereign AI initiatives; India’s CSPs are building data centers containing tens of thousands of GPUs and increasing GPU deployments by 10x in 2024 compared to a year ago; Softbank is building Japan’s most powerful AI supercomputer with Nvidia’s hardware 

Alongside significant growth from our large CSPs, NVIDIA GPU regional cloud revenue jumped 2x year-on-year as North America, India, and Asia Pacific regions ramped NVIDIA Cloud instances and sovereign cloud build-outs…

…Our sovereign AI initiatives continue to gather momentum as countries embrace NVIDIA accelerated computing for a new industrial revolution powered by AI. India’s leading CSPs include product communications and Yotta Data Services are building AI factories for tens of thousands of NVIDIA GPUs. By year-end, they will have boosted NVIDIA GPU deployments in the country by nearly 10x…

…In Japan, SoftBank is building the nation’s most powerful AI supercomputer with NVIDIA DGX Blackwell and Quantum InfiniBand. SoftBank is also partnering with NVIDIA to transform the telecommunications network into a distributed AI network with NVIDIA AI Aerial and AI-RAN platform that can process both 5G RAN on AI on CUDA.

Nvidia’s revenue from consumer internet companies more than doubled year-on-year in 2024 Q3

Consumer Internet revenue more than doubled year-on-year as companies scaled their NVIDIA Hopper infrastructure to support next-generation AI models, training, multimodal and agentic AI, deep learning recommender engines, and generative AI inference and content creation workloads. 

Nvidia’s management sees Nvidia as the largest inference platform in the world; Nvidia’s management is seeing inference really starting to scale up for the company; models that are trained on previous generations of Nvidia chips inference really well on those chips; management thinks that as Blackwell proliferates in the AI industry, it will leave behind a large installed base of infrastructure for inference; management’s dream is that plenty of AI inference happens across the world; management thinks that inference is hard because it needs high accuracy, high throughput, and low latency

NVIDIA’s Ampere and Hopper infrastructures are fueling inference revenue growth for customers. NVIDIA is the largest inference platform in the world. Our large installed base and rich software ecosystem encourage developers to optimize for NVIDIA and deliver continued performance and TCO improvements…

…We’re seeing inference really starting to scale up for our company. We are the largest inference platform in the world today because our installed base is so large. And everything that was trained on Amperes and Hoppers inference incredibly on Amperes and Hoppers. And as we move to Blackwells for training foundation models, it leads behind it a large installed base of extraordinary infrastructure for inference. And so we’re seeing inference demand go up…

… Our hopes and dreams is that someday, the world does a ton of inference. And that’s when AI has really exceeded is when every single company is doing inference inside their companies for the marketing department and forecasting department and supply chain group and their legal department and engineering, of course, and coding of course. And so we hope that every company is doing inference 24/7…

…Inference is super hard. And the reason why inference is super hard is because you need the accuracy to be high on the one hand. You need the throughput to be high so that the cost could be as low as possible, but you also need the latency to be low. And computers that are high-throughput as well as low latency is incredibly hard to build. 

Nvidia’s management has driven a 5x improvement in Hopper inference throughput in 1 year via advancements in the company’s software; Hopper’s inference performance is set to increase by a further 2.4x shortly because of NIM (Nvidia Inference Microservices)

Rapid advancements in NVIDIA software algorithms boosted Hopper inference throughput by an incredible 5x in 1 year and cut time to first token by 5x. Our upcoming release of NVIDIA NIM will boost Hopper inference performance by an additional 2.4x. 

Nvidia’s Blackwell family of chips is now in full production; Nvidia shipped 13,000 Blackwell samples to customers in 2024 Q3; the Blackwell family comes with a wide variety of customisable configurations; management sees all Nvidia customers wanting to be first to market with the Blackwell family; management sees staggering demand for Blackwell, with Oracle announcing the world’s first zetta-scale cluster with more than 131,000 Blackwell GPUs, and Microsoft being the first CSP to offer private-preview Blackwell instances; Blackwell is dominating GPU benchmarks; Blackwell performs 2.2x better than Hopper and is also 4x cheaper; Blackwell with NVLink Switch delivered up to a 30x improvement in inference speed; Nvidia’s management expects the company’s gross margin to decline slightly initially as the Blackwell family ramps, before rebounding; Blackwell’s production is in full-steam ahead and Nvidia will deliver more Blackwells in 2024 Q4 than expected; demand for Blackwell exceeds supply

Blackwell is in full production after a successfully executed mask change. We shipped 13,000 GPU samples to customers in the third quarter, including one of the first Blackwell DGX engineering samples to OpenAI. Blackwell is a full stack, full infrastructure, AI data center scale system with customizable configurations needed to address a diverse and growing AI market from x86 to ARM, training to inferencing GPUs, InfiniBand to Ethernet switches, and NVLink and from liquid cooled to air cooled. 

Every customer is racing to be the first to market. Blackwell is now in the hands of all of our major partners, and they are working to bring up their data centers. We are integrating Blackwell systems into the diverse data center configurations of our customers. Blackwell demand is staggering, and we are racing to scale supply to meet the incredible demand customers are placing on us. Customers are gearing up to deploy Blackwell at scale. Oracle announced the world’s first zetta-scale AI cloud computing clusters that can scale to over 131,000 Blackwell GPUs to help enterprises train and deploy some of the most demanding next-generation AI models. Yesterday, Microsoft announced they will be the first CSP to offer, in private preview, Blackwell-based cloud instances powered by NVIDIA GB200 and Quantum InfiniBand.

Last week, Blackwell made its debut on the most recent round of MLPerf training results, sweeping the per GPU benchmarks and delivering a 2.2x leap in performance over Hopper. The results also demonstrate our relentless pursuit to drive down the cost of compute. The 64 Blackwell GPUs are required to run the GPT-3 benchmark compared to 256 H100s or a 4x reduction in cost. NVIDIA Blackwell architecture with NVLink Switch enables up to 30x faster inference performance and a new level of inference scaling, throughput and response time that is excellent for running new reasoning inference applications like OpenAI’s o1 model…

…As Blackwell ramps, we expect gross margins to moderate to the low 70s. When fully ramped, we expect Blackwell margins to be in the mid-70s…

… Blackwell production is in full steam. In fact, as Colette mentioned earlier, we will deliver this quarter more Blackwells than we had previously estimated…

…It is the case that demand exceeds our supply. And that’s expected as we’re in the beginnings of this generative AI revolution as we all know…

…In terms of how much Blackwell total systems will ship this quarter, which is measured in billions, the ramp is incredible…

…[Question] Do you think it’s a fair assumption to think NVIDIA could recover to kind of mid-70s gross margin in the back half of calendar ’25?

[Answer] Yes, I think it is a reasonable assumption or goal for us to do, but we’ll just have to see how that mix of ramp goes. But yes, it is definitely possible.  

Nvidia’s management is seeing that hundreds of AI-native companies are already delivering AI services and thousands of AI-native startups are building new services

Hundreds of AI-native companies are already delivering AI services with great success. Though Google, Meta, Microsoft, and OpenAI are the headliners, Anthropic, Perplexity, Mistral, Adobe Firefly, Runway, Midjourney, Lightricks, Harvey, Codeium, Cursor and the Bridge are seeing great success while thousands of AI-native start-ups are building new services. 

Nvidia’s management is seeing large enterprises build copilots and AI agents with Nvidia AI; management sees the potential for billions of AI agents being deployed in the years ahead; Accenture has an internal AI agent use case that reduces steps in marketing campaigns by 25%-35%

Industry leaders are using NVIDIA AI to build Copilots and agents. Working with NVIDIA, Cadence, Cloudera, Cohesity, NetApp, Nutanix, Salesforce, SAP and ServiceNow are racing to accelerate development of these applications with the potential for billions of agents to be deployed in the coming years…

… Accenture with over 770,000 employees, is leveraging NVIDIA-powered agentic AI applications internally, including 1 case that cuts manual steps in marketing campaigns by 25% to 35%.

Nearly 1,000 companies are using NIM (Nvidia Inference Microservices); management expects the Nvidia AI Enterprise platform’s revenue in 2024 to be double that from 2023; Nvidia’s software, service, and support revenue now has an annualised revenue run rate of $1.5 billion and management expects the run rate to end 2024 at more than $2 billion

Nearly 1,000 companies are using NVIDIA NIM, and the speed of its uptake is evident in NVIDIA AI enterprise monetization. We expect NVIDIA AI enterprise full year revenue to increase over 2x from last year and our pipeline continues to build. Overall, our software, service and support revenue is annualizing at $1.5 billion, and we expect to exit this year annualizing at over $2 billion.

Nvidia’s management is seeing an acceleration in industrial AI and robotics; Foxconn is using Nvidia Omniverse to improve the performance of its factories, and Foxconn’s management expects a reduction of over 30% in annual kilowatt hour usage in Foxconn’s Mexico facility

Industrial AI and robotics are accelerating. This is triggered by breakthroughs in physical AI, foundation models that understand the physical world, like NVIDIA NeMo for enterprise AI agents. We built NVIDIA Omniverse for developers to build, train, and operate industrial AI and robotics…

…Foxconn, the world’s largest electronics manufacturer, is using digital twins and industrial AI built on NVIDIA Omniverse to speed the bring-up of its Blackwell factories and drive new levels of efficiency. In its Mexico facility alone, Foxconn expects to reduce — a reduction of over 30% in annual kilowatt hour usage.

Nvidia saw sequential growth in Data Center revenue in China because of export of compliant Hopper products; management expects the Chinese market to be very competitive

Our data center revenue in China grew sequentially due to shipments of export-compliant Hopper products to industries…

…We expect the market in China to remain very competitive going forward. We will continue to comply with export controls while serving our customers.

Nvidia’s networking revenue declined sequentially, but there was sequential growth in Infiniband and Ethernet switches, Smart NICs (network interface controllers), and BlueField DPUs; management expects sequential growth in networking revenue in 2024 Q4; management is seeing CSPs adopting Infiniband for Hopper clusters; Nvidia’s Spectrum-X Ethernet for AI revenue was up 3x year-on-year in 2024 Q3; xAI used Spectrum-X for its 100,000 Hopper GPU cluster and achieved zero application latency degradation and maintained 95% data throughput, compared to 60% for Ethernet

Areas of sequential revenue growth include InfiniBand and Ethernet switches, SmartNICs and BlueField DPUs. Though networking revenue was sequentially down, networking demand is strong and growing, and we anticipate sequential growth in Q4. CSPs and supercomputing centers are using and adopting the NVIDIA InfiniBand platform to power new H200 clusters.

NVIDIA Spectrum-X Ethernet for AI revenue increased over 3x year-on-year. And our pipeline continues to build with multiple CSPs and consumer Internet companies planning large cluster deployments. Traditional Ethernet was not designed for AI. NVIDIA Spectrum-X uniquely leverages technology previously exclusive to InfiniBand to enable customers to achieve massive scale of their GPU compute. Utilizing Spectrum-X, xAI’s Colossus 100,000 Hopper supercomputer experienced 0 application latency degradation and maintained 95% data throughput versus 60% for traditional Ethernet…

…Our ability to sell our networking with many of our systems that we are doing in data center is continuing to grow and do quite well. So this quarter is just a slight dip down and we’re going to be right back up in terms of growing. We’re getting ready for Blackwell and more and more systems that will be using not only our existing networking but also the networking that is going to be incorporated in a lot of these large systems we are providing them to.

Nvidia has begun shipping new GeForce RTX AI PCs

We began shipping new GeForce RTX AI PC with up to 321 AI FLOPS from ASUS and MSI with Microsoft’s Copilot+ capabilities anticipated in Q4. These machines harness the power of RTX ray tracing and AI technologies to supercharge gaming, photo, and video editing, image generation and coding.

Nvidia’s Automotive revenue had strong growth year-on-year and sequentially in 2024 Q3, driven by self-driving brands of Nvidia Orin; Volvo’s electric SUV will be powered by Nvidia Orin

Moving to Automotive. Revenue was a record $449 million, up 30% sequentially and up 72% year-on-year. Strong growth was driven by self-driving brands of NVIDIA Orin and robust end market demand for NAVs. Volvo Cars is rolling out its fully electric SUV built on NVIDIA Orin and DriveOS.

Nvidia’s management thinks pre-training scaling of foundation AI models is intact, but it’s not enough; another way of scaling has emerged, which is inference-time scaling; management thinks that the new ways of scaling has resulted in great demand for Nvidia’s chips, but for now, most of Nvidia’s chips are used in pre-training 

Foundation model pretraining scaling is intact and it’s continuing. As you know, this is an empirical law, not a fundamental physical law. But the evidence is that it continues to scale. What we’re learning, however, is that it’s not enough, that we’ve now discovered 2 other ways to scale. One is post-training scaling. Of course, the first generation of post-training was reinforcement learning human feedback, but now we have reinforcement learning AI feedback and all forms of synthetic data generated data that assists in post-training scaling. And one of the biggest events and one of the most exciting developments is Strawberry, ChatGPT o1, OpenAI’s o1, which does inference time scaling, what’s called test time scaling. The longer it thinks, the better and higher-quality answer it produces. And it considers approaches like chain of thought and multi-path planning and all kinds of techniques necessary to reflect and so on and so forth…

… we now have 3 ways of scaling and we’re seeing all 3 ways of scaling. And as a result of that, the demand for our infrastructure is really great. You see now that at the tail end of the last generation of foundation models were at about 100,000 Hoppers. The next generation starts at 100,000 Blackwells. And so that kind of gives you a sense of where the industry is moving with respect to pretraining scaling, post-training scaling, and then now very importantly, inference time scaling…

…[Question] Today, how much of the compute goes into each of these buckets? How much for the pretraining? How much for the reinforcement? And how much into inference today?

[Answer] Today, it’s vastly in pretraining a foundation model because, as you know, post-training, the new technologies are just coming online. And whatever you could do in pretraining and post-training, you would try to do so that the inference cost could be as low as possible for everyone. However, there are only so many things that you could do a priority. And so you’ll always have to do on-the-spot thinking and in context thinking and a reflection. And so I think that the fact that all 3 are scaling is actually very sensible based on where we are. And in the area foundation model, now we have multimodality foundation models and the amount of petabytes video that these foundation models are going to be trained on, it’s incredible. And so my expectation is that for the foreseeable future, we’re going to be scaling pretraining, post-training as well as inference time scaling and which is the reason why I think we’re going to need more and more compute.  

Nvidia’s management thinks the company generates the greatest possible revenue for its customers because its products has much better performance per watt

Most data centers are now 100 megawatts to several hundred megawatts, and we’re planning on gigawatt data centers, it doesn’t really matter how large the data centers are. The power is limited. And when you’re in the power-limited data center, the best — the highest performance per watt translates directly into the highest revenues for our partners. And so on the one hand, our annual road map reduces cost. But on the other hand, because our perf per watt is so good compared to anything out there, we generate for our customers the greatest possible revenues. 

Nvidia’s management sees Hopper demand continuing through 2025

Hopper demand will continue through next year, surely the first several quarters of the next year. 

Nvidia’s management sees 2 fundamental shifts in computing happening today: (1) the movement from code that runs on CPUs to neural networks that run on GPUs and (2) the production of AI from data centres; the fundamental shifts will drive a $1 trillion modernisation of data centres globally

We are really at the beginnings of 2 fundamental shifts in computing that is really quite significant. The first is moving from coding that runs on CPUs to machine learning that creates neural networks that runs on GPUs. And that fundamental shift from coding to machine learning is widespread at this point. There are no companies who are not going to do machine learning. And so machine learning is also what enables generative AI. And so on the one hand, the first thing that’s happening is $1 trillion worth of computing systems and data centers around the world is now being modernized for machine learning.

On the other hand, secondarily, I guess, is that on top of these systems are going to be — we’re going to be creating a new type of capability called AI. And when we say generative AI, we’re essentially saying that these data centers are really AI factories. They’re generating something. Just like we generate electricity, we’re now going to be generating AI. And if the number of customers is large, just as the number of consumers of electricity is large, these generators are going to be running 24/7. And today, many AI services are running 24/7, just like an AI factory. And so we’re going to see this new type of system come online, and I call it an AI factory because that’s really as close to what it is. It’s unlike a data center of the past.

Nvidia’s management does not see any digestion happening for GPUs until the world’s data centre infrastructure is modernised

[Question] My main question, historically, when we have seen hardware deployment cycles, they have inevitably included some digestion along the way. When do you think we get to that phase? Or is it just too premature to discuss that because you’re just at the start of Blackwell?

[Answer] I believe that there will be no digestion until we modernize $1 trillion with the data centers.

Okta (NASDAQ: OKTA)

Okta AI is really starting to help newer Okta products

Second thing is that we have Okta AI, which we talked a lot about a couple of years ago, and we continue to work on that. And it’s really starting to help these new products like identity threat protection with Okta AI. The model inside of identity threat protection and how that works is AI is a big part of the product functionality. 

Okta’s management sees the need for authentication for AI agents and has a product called Auth for Gen AI; management thinks authentication of AI agents could be a new area of growth for Okta; management sees the pricing for Auth for Gen AI as driven by a fee per monthly active machine

Some really interesting new areas are we have something we talked about at Oktane called Auth for Gen AI, which is basically authentication platform for agents. Everyone is very excited about agents, as they should be. I mean, we used to call them bots, right? 4, 5 years ago, they’re called bots. Now they’re called agents, like what’s the big deal? How different is it? Well, you can interact with them natural languages and they can do a lot more with these models. So now it’s like bots are real in real time. But the problem is all of these bots and all of these platforms to build bots, they have the equivalent of the monitor sticky notes with passwords on them, they have the equivalent of that inside the bot. So there’s no protocol for single sign-on for bots. They have like stored passwords in the bot. And if that bot gets hacked, guess what? You signed up for that bot and it has access to your calendar and has access to your travel booking and it has access to your company e-mail and your company data, that’s gone because the hacker is going to get all those passwords out there. So Auth for Gen AI automates that and make sure you can have a secure protocol to build a bot around. And so that’s a really interesting area. It’s very new. We just announced it and all these agent frameworks and so forth are new…

… Auth for GenAI, it’s basically like — think about it as a firm machine authentication. So every time — we have this feature called machine-to-machine, which does a similar thing today, and you pay basically by the monthly active machine.

Salesforce (NYSE: CRM)

Salesforce’s management thinks Salesforce is at the edge of the rise of digital labour, which are autonomous AI agents; management thinks the TAM (total addressable market) for digital labour is much larger than the data management market that Salesforce was previously in; management thinks Salesforce is the largest supplier of digital labour right from the get-go; Salesforce’s AgentForce service went into production in 2024 Q3 and Salesforce has already delivered 200 AgentForce deals with more to come; management has never seen anything like AgentForce; management sees AgentForce as the next evolution of Salesforce; management thinks AgentForce will help companies scale productivity independent of workforce growth; management sees AgentForce AI agents manifesting as robots that will supplement human labour; management sees AgentForce, together with robots, as a driving force for future global economic growth even with a stagnant labour force; AgentForce is already delivering tangible value to customers; Salesforce’s customers recently built 10,000 AI agents with AgentForce in 3 days, and thousands more AI agents have been built since then; large enterprises across various industries are building AI agents with AgentForce; management sees AgentForce unlocking a whole new level of operational efficiency; management will be delivering AgentForce 2.0 in December this year

We’re really at the edge of a revolutionary transformation. This is really the rise of digital labor. Now for the last — I would say for the last 25 years at Salesforce, and we’ve been helping companies to manage and share their information…

…But now we’ve really created a whole new market, a new TAM, a TAM that is so much bigger and so much more exciting than the data management market that it’s hard to get our head completely around. This is the market for digital labor. And Salesforce has become, right out of the gate here, the largest supplier of digital labor and this is just the beginning. And it’s all powered by these autonomous AI agents…

…With Salesforce, agent force, we’re not just imagining this future. We’re already delivering it. And you so know that in the last week of the quarter, Agentforce went production. We delivered 200 deals, and our pipeline is incredible for future transactions. We can talk about that with you on the call, but we’ve never seen anything like it. We don’t know how to characterize it. This is really a moment where productivity is no longer tied to workforce growth, but through this intelligent technology that can be scaled without limits. And Agentforce represents this next evolution of Salesforce. This is a platform now, Salesforce as a platform or AI agents work alongside humans in a digital workforce that amplifies and augments human capabilities and delivers with unrivaled speed…

…On top of the agenetic layer, we’ll soon see a robotic layer as well where these agents will manifest into robots…

…These agents are not tools. They are becoming collaborators. They’re working 24/7 to analyze data, make decisions, take action, and we can all start to picture this enterprise managing millions of customer interactions daily with Agentforce seamlessly resolving issues, processing transactions, anticipating customer needs, freeing up humans to focus on the strategic initiatives and building meaningful relationships. And this is going to evolve into customers that we have, whether it could be a large hospital or a large hotel where not only are the agents working 24/7, but robots are also working side-by-side with humans, robots manifestations of agents this idea that it’s all happening before our eyes and that this isn’t just some far-off future. It’s happening right now…

…For decades, economic growth dependent on expanding the human workforce. It was all about getting more labor. But with labor and with the labor force stagnating globally, Agentforce is unlocking a new path forward. It’s a new level of growth for the world and for our GPT and businesses no longer need to choose between scale and efficiency with agents, they can achieve both…

…Our customers are already experiencing this transformation. Agentforce is deflecting service cases and resolving issues, processing, qualifying leads, helping close more deals, creating optimizing marketing campaigns, all at an unprecedented scale, 24/7…

…What was remarkable was the huge thirst that our customers had for this and how they built more than 10,000 agents in 3 days. And I think you know that we then unleashed a world tour of that program, and we have now built thousands and thousands of more agents in these world tours all over the world…

…So companies like FedEx, [indiscernible], Accenture, Ace Hardware, IBM, RBC Wealth Management and many more are now building their digital labor forces on the Salesforce platform with Agentforce. So that is the largest and most important companies in the world across all geographies, across all industries are now building and delivering agents…

…While these legacy chatbots have handled these basic tasks like password resets and other basic mundane things, Agentforce is really unlocking an entirely new level of digital intelligence and operational efficiency at this incredible scale…

…I want to invite all of you to join us for the launch of Agentforce 2.0. And it is incredible what you are going to see the advancements in the technology already are amazing and accuracy and the ability to deliver additional value. And we hope that you’re going to join us in San Francisco. This is going to happen on December 17. You’ll see Agentforce 2.0 for the first time,

Salesforce is customer-zero for AgentForce and the service is live on Salesforce’s help-website; AgentForce is handling 60 million sessions and 2 millions support cases annually on the help-website; the introduction of AgentForce in Salesforce’s help-website has allowed management to rebalance headcount into growth-areas; users of Salesforce’s help-website will experience very high levels of accuracy because AgentForce is grounded with the huge repository of internal and customer data that Salesforce has; management sees Salesforce’s data as a huge competitive advantage for AgentForce; AgentForce can today quickly deliver personalised insights to users of Salesforce’s help-website and hand off users to support engineers for further help; management thinks AgentForce will deflect between a quarter and half of annual case volume; Salesforce is also using AgentForce internally to engage prospects and hand off prospects to SDR (sales development representative) team

We pride ourselves on being customer [ 0 ] for all of our products, and Agentforce is no exception. We’re excited to share that Agentforce is now live on help.salesforce.com…

… Our help portal, help.salesforce.com, which is now live. This portal, this is our primary support mechanism for our customers. It lets them authenticate in, it then becomes grounded with the agent, and that Help portal already is handling 60 million sessions and more than 2 million support cases every year. Now that is 100% on Agentforce…

…From a human resource point of view, where we can really start to look at how are we going to rebalance our headcount into areas that now are fully automated and to into areas that are critical for us to grow like distribution…

…Now when you use help.salesforce.com, especially as authenticated users, as I mentioned, you’re going to see this incredible level of accuracy and responsiveness and you’re going to see remarkably low hallucinogenic performance whether for solving simple queries or navigating complex service issues because Agentforce is not just grounded in our Salesforce data and metadata including the repository of 740,000 documents and 17 languages, it’s also grounded in each customer’s data, their purchases, returns, that data it’s that 200 petabytes or through 200 to 300 petabytes of Salesforce data that we have that gives us this kind of, I would say, almost unfair advantage with Agentforce because our agents are going to be more accurate in the least hallucinogenic of any because they have access to this incredible capability. And Agentforce can instantly reason over this vast amounts of data, deliver precise personalizing [indiscernible] with citations in seconds, and Agentforce can seamlessly hand off to support engineers, delivering them complete summary and recommendation as well. And you can all try this today. This isn’t some fantasy land future idea this is today reality…

…We expect that our own transformation with Agentforce on help.salesforce.com and in many other areas of our company, it is going to deflect between a quarter and half of annual case volume and in optimistic cases, probably much, much more of that…

…We’re also deploying Agentforce to engage our prospects on salesforce.com, answering their questions 24/7 as well as handing them off to our SDR team. You can see it for yourself and test it out on our home page. We’ll use our new Agentforce SDR agent to further automate top of funnel activities when gatherings leads, lead data for providing education and qualifying prospects and booking meetings.

Salesforce’s management thinks AgentForce is much better than Microsoft’s AI Copilots

I just want to compare and contrast that against other companies who say they are doing enterprise AI. You can look at even Microsoft. We all know about Copilot, it’s been out, it’s been touted now for a couple of years. We’ve heard about CoPilot. We’ve seen the demo. In many ways, it’s just repackaged ChatGPT. You can really see the difference where Salesforce now can operate its company on our platform. And I don’t think you’re going to find that on Microsoft’s website, are you?

Vivint is using AgentForce for customer support and for technician scheduling, payment requests, and more; Adecco is using AgentForce to improve the handling of job applicants (Adecco receives 300 million job applications annually); Wiley is resolving cases 40% faster with AgentForce; Heathrow Airport is using AgentForce to respond to thousands of travelers instantly, accurately, and simultaneously; SharkNinja is using AgentForce for personalised 24/7 customer support in 28 geographies; Accenture is using AgentForce to improve deal quality and boost bid coverage by 75%

One of them is the smart home security provider, Vivint. They’ve struggled with this high volume of support calls, a high churn rate for service reps. It’s a common story. But now using the Agentforce, Vivint has created a digital support staff to autonomously provide support through their app, their website, troubleshooting, a broad variety of issues across all their customer touch points. And in addition, Vivint is planning to utilize Agentforce to further automate technician scheduling, payment request, proactive issue resolution, the use of device telemetry because Agentforce is across the entire sales force product line and including Slack…

…Another great customer example that’s already incredible to work they’ve already done to get this running and going in their company Adecco, the world’s leading provider of talent solutions, handling 300 million job applications annually, but historically, they have just not been able to go through or respond in a timely way, of course, to the vast majority of applications that they’re gating, but now the Agentforce is going to operate an incredible scale, sorting through the millions of resumes, 24/7 matching candidates to opportunities proactively prequalifying them for recruiters. And in addition, Agentforce can also assess candidates helping them to refine their resumes, giving them a better chance of qualifying for a role…

…Wiley, an early adopter, is resolving cases over 40% faster with Agentforce than their previous chat bot. Heathrow Airport, one of the busiest airports in the world will be able to respond to thousands of travelers inquiries instantly, accurately and simultaneously. SharkNinja, a new logo in the quarter, chose Agentforce and Commerce Cloud to deliver 24/7 personalized support for customers across 28 international markets and unifying its service operations…

…Accenture chose Agentforce to streamline sales operations and enhance bid management for its 52,000 global sellers. By integrating sales coach and custom AI agents, Agentforce is improving deal quality and targeting a 75% boost in bid coverage. 

College Possible is using AgentForce to build virtual college counsellors as there’s a shortage of labour (for example, California has just 1 counsellor for every 500 students); College Possible built its virtual counsellors with AgentForce in under a week – basically like flipping a switch – because it has been accumulating all its data in Salesforce for years

Another powerful example is a nonprofit, College Possible. College Possible matches eligible students with counselors to help them navigate and become ready for college. And in California, for example, the statewide average stands at slightly over 1 counselor for every 500 students. It just isn’t enough. Where are we going to get all that labor…

…We’re going to get it from Agentforce. This means the vast majority of students are not getting the help they need, and now they are going to get the help they need.

College Possible creates a virtual counselor built on Agentforce in under a week. They already had all the data. They have the metadata, they already knew the students. They already had all of the capabilities built into their whole Salesforce application. It was just a flip of a switch…

…  But why? It’s because all of the work and the data and the capability that College Possible has put into Salesforce over the years and years that they had it. It’s not the week that it took to get them to turn it on. They have done a lot of work.

Salesforce’s management’s initiative to have all of the company’s apps be rewritten into a single core platform is called More Core; the More Core initiative also involves Salesforce’s Data Cloud, which is important for AI to work; Salesforce is now layering the AI agent layer on top of More Core, and management sees this combination as a complete AI system for enterprises that also differentiates Salesforce’s AgentForce product

Over the last few years, we’ve really aggressively invested in integrating all of our apps on a single core platform with shared services for security workflow user interfaces more. We’ve been rewriting all of our acquisitions into that common area. We’re really looking at how do we take all of our applications and all of our acquisitions, everything and delivered into one consistent platform, we call that More Core internally inside Salesforce. And when you look at that More Core initiative, I don’t think there’s anyone who delivers this comprehensive platform, sales, service, marketing, commerce, analytics, Slack, all of it as one piece of code. And then now deeply integrated in that 1 piece of code is also our data cloud. That is a key part of our strategy, which continues to have this phenomenal momentum as well to help customers unify and federate with zero-copy data access across all their data and metadata, which is crucial for AI to work.

And now that third layer is really opening up for us, which is this agenetic layer. We have built this agenetic layer that takes advantage of all the investments in Salesforce for our customers and made it in our platform. It’s really these 3 layers. And in these 3 layers that form a complete AI system for enterprises and really uniquely differentiate Salesforce uniquely differentiate Agentforce from every other AI platform that this is one piece of code. This isn’t like 3 systems. It’s not a bunch of different apps all running independently. This is all one piece of code. That’s why it works so well, by the way, because it is 1 platform.

Salesforce’s management thinks jobs and roles within Salesforce will change because of AI, especially AI agents

The transformation is not without challenges. Jobs are going to evolve, roles are going to shift and businesses will need to adapt. And listen, at Salesforce, jobs are going to evolve and roles will shift and businesses will need to adapt as well. We’re all going to need to rebalance our workforce. This is the agents take on more of the workforce.

Salesforce’s management is hearing that a large customer of Salesforce is targeting 25% more efficiency with AI

This morning, I was on the phone with one of our large customers, and they were telling me how they’re targeting inside their company, 25% more efficiency with artificial intelligence.

Salesforce signed more than 2,000 AI deals in 2024 Q3 (FY2025 Q3), and number of AI deals that are over $1 million more than tripled year-on-year; 75% of Salesforce’s AgentForce deals, and 9 of Salesforce’s top 10 deals, in 2024 Q3 involved Salesforce’s global partners; more than 80,000 system integrators have completed AgentForce training; hundreds of ISVs (independent software vendors) and partners are building and selling AI agents; Salesforce has a new AgentForce partner network that allows customers to deploy customised AI agents using trusted 3rd-party extensions from Salesforce App Exchange; Salesforce’s partnership with AWS Marketplace is progression well as transactions doubled sequentially in 2024 Q3, with 10 deals exceeding $1 million

In Q3, the number of wins greater than $1 million with AI more than tripled year-over-year. and we signed more than 2,000 AI deals, including more than the 200 Agentforce wins that Marc shared…

…We’re also seeing amazing Agentforce energy across the ecosystem with our global partners involved in 75% of our Q3 Agentforce deals and 9 of our top 10 wins in the quarter. Over 80,000 system integrators have completed Agentforce training and hundreds of ISVs and technology partners are building and selling agents…

… We continue to unlock customer spend through new channels, including the Agentforce partner network that launched at Dreamforce, which allows customers to customize and deploy specialized agents using trusted third-party extensions from Salesforce App Exchange. And AWS Marketplace continues to be a growth driver. Our Q3 transactions doubled quarter-over-quarter with 10 deals exceeding $1 million. 

Veeva Systems (NYSE: VEEV)

Veeva Vault CRM has a number of new innovations coming, including two AI capabilities that will be available in late-2025 at no additional charge; one of the AI capabilities leverages Apple Intelligence; Vault CRM’s CRM Bot AI application will see Vault CRM be hooked onto customers’ own large language models, and Veeva will not be incurring compute costs

We just had our European Commercial Summit in Madrid where we announced a number of new innovations coming in Vault CRM, including two new AI capabilities – CRM Bot and Voice Control. CRM Bot is a GenAI assistant in Vault CRM. Voice Control is a voice interface for Vault CRM, leveraging Apple Intelligence. Both are included in Vault CRM for no additional charge and are planned for availability in late 2025…

…For the CRM Bot, that’s where we will hook our CRM system into the customers’ own large language model that they’re running. And that’s where we will not charge for, and we will not incur compute cost…

Veeva has a new AI application, MLR Bot, for Vault PromoMats within Commercial Cloud; MLR Bot helps perform checks on content with a Veeva-hosted large language model (LLM); MLR Bot will be available in late-2025 and will be charged separately; management has been thinking about MLR Bot for some time; management is seeing a lot of excitement over MLR Bot; management is still working through the details of the monetisation of MLR Bot; MLR Bot’s LLM will be from one of the big tech providers but it will be Veeva who’s the one paying for the compute 

We also announced MLR Bot, an AI application in Vault PromoMats to perform content quality and compliance checks with a Veeva-hosted large language model. Planned for availability in late 2025, MLR Bot will require a separate license…

… So I was at our Europe Summit event where we announced MLR Bot, something we’ve been thinking about and evaluating for some time…

…So there’s a lot of excitement. This is a really core process for life sciences companies. So a lot of excitement there…

…In terms of sizing and the monetization, we’re still working through the details on that, but there’s a ton of excitement from our existing customers. We look forward to getting some early customers started on that as we go into next year…

…MLR Bot, we will charge for, and that’s where we will host and run a large language model. Not our own large language model, right? We’ll use one from the big tech providers, but we will be paying for the compute power for that, and so we’ll be charging for that.

CRM Bot, Voice Control, and MLR Bot are part of Veeva’s management’s overall AI strategy to provide AI applications with tangible value; another part of the AI strategy involves opening up data for customers to power all forms of AI; management’s current thinking is to charge for AI applications if Veeva is responsible for paying compute costs

These innovations are part of our overall AI strategy to deliver specific AI applications that provide tangible value and enable customers and partners with the AI Partner Program, as well as the Vault Direct Data API, for the data needed to power all forms of AI…

… So where we have to use significant compute power, we will most likely charge. And where we don’t, we most likely won’t.

Wix (NASDAQ: WIX)

More than 50% of new Wix users are using the company’s AI-powered onboarding process which was launched nearly a year ago; users who onboard using Wix’s AI process are 50% more likely to start selling on Wix and are more likely to become paid subscribers; the AI-powered onboarding process is leading to a 13% uplift in conversion rate for the most recent Self-Creator cohort; the AI website builder is free but it helps with conversions to paid subscribers

Almost one year ago, we launched our AI website builder, which is now available in 20 languages and has been a game changer in our user onboarding strategy. Today, more than 50% of new users are choosing to create their online presence through our AI-powered onboarding process. The tool is resonating particularly well with small businesses and entrepreneurs as paid subscriptions originated from this AI-powered onboarding are 50% more likely to have a business vertical attached and significantly more likely to start selling on Wix by streamlining the website building process while offering a powerful and tailored commerce-enablement solution…

…Cash in our most recent self-created cohort showed a 13% uplift in conversion rate from our AI onboarding tool…

…[Question] A lot of the commentary seems that today, AI Website Builder is helping on conversion. I wanted to ask about specifically, is there an opportunity to directly monetize the AI products within the kind of core website design funnel?

[Answer] So I think that the way we monetize, of course, during the buildup phase of the website, is by making it easier. And our customers are happy with their websites, of course, we convert better. So I don’t think there is any better way to monetize than that, right? The more users finish the website, the better the website, the higher conversion and the high monetization. 

Wix now has 29 AI assistants to support users

Earlier this year, we spoke about our plan to embed AI assistance across our platform and we’re continuing to push that initiative forward. We now have a total of 29 assistants, spanning a wide range of use cases to support users and to service customers throughout their online journeys.

Wix has a number of AI products that are launching in the next few months that are unlike anything in the market and they will be the first AI products that Wix will be monetising directly

We have a number of AI products coming in the next few months that are unlike anything in the market today. These products will transform the way merchants manage their businesses, redefine how users interact with their customers and enhance the content creation experience. Importantly, these will also be the first AI products we plan to monetize directly. We are on the edge of unforeseen innovation, and I’m looking forward to the positive impact it will have on our users.

Zoom Communications (NASDAQ: ZM)

Zoom’s management has a new vision for Zoom, the AI-first Work Platform for Human Connection

In early October, we hosted Zoomtopia, our annual customer and innovation event, and it was an amazing opportunity to showcase all that we have been working on for our customers. We had a record-breaking virtual attendance, and unveiled our new vision, AI-first Work Platform for Human Connection. This update marks an exciting milestone as we extend our strength as a unified communication and collaboration platform into becoming an AI-first work platform. Our goal is to empower customers to navigate today’s work challenges, streamline information, prioritizing tasks and making smarter use of time.

Management has released AI Companion 2.0, which is an agentic AI technology; AI Companion 2.0 is able to see a broader window of context and gather information from internal and external sources; Zoom AI Companion monthly active users grew 59% sequentially in 2024 Q3; Zoom has over 4 million accounts that have enabled AI Companion; management thinks customers really like Zoom AI Companion; customer feedback for AI Companion has been extremely positive; management does not intend to charge customers for AI Companion

At Zoomtopia, we took meaningful steps towards that vision with the release of AI Companion 2.0…

…This release builds upon the awesome quality of Zoom AI Companion 1.0 across features like Meeting Summary, Meeting Query and Smart Compose, and brings it together in a way that evolves beyond task-specific AI towards agentic AI. This major update allows the AI Companion to see a broader window of context, synthesize the information from internal and external sources, and orchestrate action across the platform. AI Companion 2.0 raises the bar for AI and demonstrates to customers that we understand their needs…

…We saw progress towards our AI-first vision with Zoom AI Companion monthly active users growing 59% quarter-over-quarter…

…At Zoomtopia, we mentioned that there are over 4 million accounts who are already enabled AI Companion. Given the quality, ease of use and no additional cost, the customer really like Zoom AI Companion…

…Feedback from our customers at Zoomtopia Zoom AI Companion 2.0 were extremely positive because, first of all, they look at our innovation, the speed, right? And the — a lot of features built into the AI Companion 2.0, again, at no additional cost, right? At the same time, Enterprise customers also want to have some flexibility. That’s why we also introduced customized AI Companion and also AI Companion Studio. And that will be available first half of next year and also we can monetize…

…We are not going to charge the customer for AI Companion, at no additional cost

Zscaler is using Zoom AI Companion to improve productivity across the whole company; large enterprises such as HSBC and Exxon Mobil are also using Zoom AI Companion

Praniti Lakhwara, CIO of Zscaler, provided a great example of how Zoom AI Companion helped democratize AI and enhance productivity across the organization, without sacrificing security and privacy. And it wasn’t just Zscaler. the RealReal, HSBC, ExxonMobil and Lake Flato Architects shared similar stories about Zoom’s secure, easy-to-use solutions, helping them thrive in the age of AI and flexible work.

Zoom’s management recently introduced a road map of AI products that expands Zoom’s market opportunity; Custom AI Companion add-on, including paid add-ons for healthcare and education, will be released in 2025 H1; management built the monetisable parts of AI Companion after gathering customer feedback 

Building on our vision for democratizing AI, we introduced a road map of TAM-expanding AI products that create additional business value through customization, personalization and alignment to specific industries or use cases. 

 Custom AI Companion add-on, which will be released in the first half of next year, aims to meet our customers where they are in their AI journey by plugging into knowledge bases, integrating with third-party apps and personalizing experiences like custom AI avatars and AI coaching. Additionally, we announced that we’ll also have Custom AI Companion paid add-ons for health care and education available as early as the first quarter of next year…

…The reason why we introduced the Customized AI Companion or AI Companion Studio because, a few quarters ago — and we talked to many Enterprise customers. They shared with us feedback, right? So they like AI Companion. Also, they want to make sure, hey, some customers, they already build their own AI large language model. How to [ federate ] that into our federated AI approach. And some customers, they have very large content, like a knowledge base, how to connect with that. Some customers, they have with other beginning systems, right, like a ServiceNow, Atlassian and Workday, a lot of Box and HubSpot, how to connect those data sources, right? And also even from an employee perspective, right, they won’t have a customized avatar, like AI to — as a personal culture as well. So meaning those customers, they have customized requirements. To support those customer requirements, we need to make sure we have AI infrastructure and technology ready, right? That’s the reason why we introduced the AI Companion, the Customized AI Companion. The goal is really working together with integrated customers to tailored for each Enterprise customer. That’s the reason why it’s not free.

I think the feedback from Zoomtopia is very positive because, again, those features are not built by our — just the several product managers, engineers think about let’s build that. We already solicited feedback from our Enterprise content before, those features that I think can truly satisfy their needs.

Zoon’s management thinks that Zoom is very well-positioned because it is providing AI-powered tools to customers at no additional cost, unlike other competitors

Given our strength on the quality plus at no additional cost, Zoom is much better positioned. In particular, customers look at all the vendors when they try to consult and look at — again, the AI cost is not small, right? You look at some of the competitors, per user per month, $30, right? And look at Zoom, better quality at no additional cost. That’s the reason why it comes with a total cost of ownership. Customers look at Zoom, I think, much better positioned…

…Again, almost every business, they subscribe to multiple software services. If each software service vendors they are going to charge the customer with AI, guess what, every business is — they have to spend more. That’s the reason why they trust Zoom, and I think we are much better positioned.

Zoom’s management is seeing some customers find new budgets to invest in AI, whereas some customers are reallocating budgets from other areas towards AI

Every company, I think now they are all thinking about where they should allocate the budget, right? Where should they get more money or fund, right, to support AI? I think every company is different. And some internal customers, and they have a new budget. Some customers, they consolidated into the few vendors and some customers, they just want to say, hey, maybe actually save the money from other areas and to shift the budget towards embracing AI.

Zoom’s management thinks Zoom will need to continue investing in AI, but they are not worried about the costs because the AI features will be monetised

Look at AI, right? So we have to invest more, right? And I think a few areas, right? One is look at our Zoom Workplace platform, right? We have to [ invent ] more talent, deploy more GPUs and also use more of the cloud, basically GPUs, as well as we keep improving the AI quality and innovate on AI features. That’s for Workplace. And at the same time, we are going to introduce the customized AI Companion, also AI Studio next year. Not only do we offer the free service for AI Companion, but those Enterprise customization certainly can help us in terms of monetization. At the same time, we leverage the technology we build for the workplace, apply that to the Contact Center, like Zoom Virtual Agent, right, and also some other Contact Center features. We can share the same AI infrastructure and also a lot of technology components and also can be shared with Zoom Contact Center.

Where AI Companion is not free, the Contact Center is different, right? We also can monetize. Essentially, we build the same common AI infrastructure architecture and Workplace — Customized AI Companion, we can monetize. Contact Center, also, we can monetize. I think more and more — and like today, you see you keep investing more and more, and soon, we can also monetize more as well. That’s why I think we do not worry about the cost in the long run at all, I mean, the AI investment because with the monetization coming in, certainly can help us more. So, so far, we feel very comfortable.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Adobe, Alphabet (parent of Google and GCP), Amazon (parent of AWS), Meta Platforms, Microsoft, MongoDB, Okta, Salesforce, Veeva Systems, Wix, and Zoom Video Communications. Holdings are subject to change at any time.

The Latest Thoughts From American Technology Companies On AI (2024 Q3)

A collection of quotes on artificial intelligence, or AI, from the management teams of US-listed technology companies in the 2024 Q3 earnings season.

The way I see it, artificial intelligence (or AI), really leapt into the zeitgeist in late-2022 or early-2023 with the public introduction of DALL-E2 and ChatGPT. Both are provided by OpenAI and are software products that use AI to generate art and writing, respectively (and often at astounding quality). Since then, developments in AI have progressed at a breathtaking pace.

With the latest earnings season for the US stock market – for the third quarter of 2024 – coming to its tail-end, I thought it would be useful to collate some of the interesting commentary I’ve come across in earnings conference calls, from the leaders of technology companies that I follow or have a vested interest in, on the topic of AI and how the technology could impact their industry and the business world writ large. This is an ongoing series. For the older commentary:

With that, here are the latest commentary, in no particular order:

Airbnb (NASDAQ: ABNB)

Airbnb’s management recently introduced a personalised welcome tour of the Airbnb app for first-time users; management sees this personalisation feature as the beginning of a more personalised Airbnb

We also introduced 50 upgrades for guests that make Airbnb a more intuitive and personalized app. And some of the features include a personalized welcome tour of the app for first-time guests, suggest a destination when guests tap the search bar, we’ll recommend locations on their search and booking history, and personalized listing highlights. So when a guest views a listing, we will highlight the details that are relevant to their search, and there are dozens of new features just like these. This is quite literally the beginning of a more personalized Airbnb.

Airbnb’s management is seeing great progress on AI-powered customer service; management sees 3 phases to the deployment of AI-powered customer service, where Phase 1 is Airbnb using AI to answer basic questions from customers, Phase 2 is the AI answering questions from customers in a personalised way, and Phase 3 is the AI taking personalised actions on behalf of customers; management thinks that Airbnb has hired some of the best AI talent to develop AI-powered customer service

We are seeing some really great progress on AI-powered customer service. The way we think about customer service, powered by AI is in 3 phases…

…Phase 1 is just answer basic general questions. We’re rolling out a pilot that can answer basic general questions. Phase 2 is personalization, be able to personalize the questions. Phase 3 is to take action…

…So this is where we think customer service can go enabled by AI, and we’ve hired some of the best people in the world to work on this.

Airbnb is currently in Phase 1 of deploying AI-powered customer service; management thinks that the vast majority of customer chats that are received by Airbnb will be handled directly by AI agents in the future

Phase 1 is the phase we’re in right now. If you were to most — first of all, most of our customer context, we get over 10 million contacts a year. Most of the contacts that we anticipate getting in the coming years aren’t going to be phone calls. They’re going to be chatting through the app. I really personally don’t like calling customer service and having to dial them. I want to be able to chat, and chat AI can intercept. And so we think in the future, the vast majority of our chats are going to be intercepted and handled directly by the AI agent.

An example of the 3rd phase of Airbnb’s AI-powered customer service that management has in mind: An AI agent can help customers to cancel bookings and even make rebookings 

So I’ll give you an example. Let me just give you 1 example. Let’s say I were to contact customer service and I say, “how do I cancel reservation?” In Phase 1, what we’re doing now, the AI agent will answer probably even better than the average customer service agent, how to cancel a reservation. So we’ll take you to how to cancel a reservation step by step. Phase 2 personalization, they’ll say, “hey, Brian, I see you have a reservation coming up in Los Angeles next week. Here’s how you cancel that reservation.” And Phase III is taking action. It would say, “hey, Brian, I see you have a reservation come to Los Angeles. Would you like me to cancel it for you? Just tell me, yes, and I’ll do it for you. I can even handle rebooking.”

Alphabet (NASDAQ: GOOG)

Alphabet’s management thinks Alphabet is positioned to lead in AI because of the company’s full-stack approach of a robust AI infrastructure, world-class research team, and broad user-reach

We are uniquely positioned to lead in the era of AI because of our differentiated full stack approach to AI innovation, and we are now seeing this operate at scale. There’s 3 components: first, a robust AI infrastructure that includes data centers, chips and a global fiber network; second, world-class research teams who are advancing our work with deep technical AI research and who are also building the models that power our efforts. And third, a broad global reach through products and platforms that touch billions of people and customers around the world, creating a virtuous cycle.

Alphabet signed the world’s first corporate agreement for energy from multiple small modular nuclear reactors; the reactors will deliver 500 megawatts of carbon-free power 24/7

We are also making bold clean energy investments, including the world’s first corporate agreement to purchase nuclear energy from multiple small modular reactors, which will enable up to 500 megawatts of new 24/7 carbon-free power.

Since Alphabet began testing AI overviews 18 months ago, the company has reduced the cost to deliver queries by 90% while doubling the size of its Gemini foundation AI model; AI overview has led to users coming to Search more often; AI overview was recently rolled out to 100 new countries and territories and will reach more than 1 billion users on a monthly basis; there’s strong engagement in AI overview, leading to higher overall search usage and user satisfaction, and users are asking longer questions and exploring more websites; the growth driven by AI overviews is increasing over time; the integration of advertising with AI overviews  is performing well; Alphabet is now showing search and shopping ads within AI overviews for mobile users in the USA; management finds that users find ads within AI overviews to be helpful; management expects Search to evolve significant in 2025, driven by advances in AI; management is seeing the monetisation rate on ads within AI overviews to be the same as the broader Search; reminder from management that Google already introduced an answer-machine 10 years ago and management is aware of changing trends in user behaviours in Search

Since we first began testing AI overviews, we have lowered machine cost per query significantly. In 18 months, we reduced cost by more than 90% for these queries through hardware, engineering and technical breakthroughs while doubling the size of our custom Gemini model…

…In Search, recent advancements, including AI overviews, Circle to Search and new features in :ens are transforming the user experience, expanding what people can search for and how they search for it. This leads to users coming to search more often for more of their information needs driving additional search queries. Just this week, AI overview started rolling out to more than 100 new countries and territories. It will now reach more than 1 billion users on a monthly basis. We are seeing strong engagement, which is increasing overall search usage and user satisfaction. People are asking longer and more complex questions and exploring a wide range of websites. What’s particularly exciting is that this growth actually increases over time as people learn that Google can answer more of their questions.

The integration of ads within AI overviews is also performing well, helping people connect with businesses as they search…

…AI overviews, where we have now started showing search and shopping ads within the overview for mobile users in the U.S. As you remember, we’ve already been running ads above and below AI overviews. We’re now seeing that people find ads directly within AI overview is helpful because they can quickly connect with relevant businesses, products and services to take the next step at the exact moment they need…

… So I expect Search to continue to evolve significantly in 2025, both in the search product and in Gemini…

…We recently launched ads within AI overviews on mobile in the U.S. And this really builds on our previous rollout of ads above and below the AI overviews. So overall, for AI overviews, we see monetization at approximately the same rate, which gives us a really strong base on which we can innovate even more…

…[Question] Why doesn’t it make sense to have 2 completely different search experiences one an agent like answers engine; and then two, a link-based more traditional search engine? 

[Answer] In this moment, people are using a lot of buzz words like answer engines and all that stuff. I mean Google started answering questions about 10 years ago in our search product with featured snippets. So look, I think, ultimately, you are serving users. User expectations are constantly evolving. And and we work hard to stay a step ahead, anticipate and stay a step ahead.

Alphabet uses and offers customers both its own TPUs (tensor processing units) and Nvidia GPUs; Alphabet is now on the 6th generation of its TPUs, known as Trillium; LG AI Research reduced its inference processing time by 50% and operating costs by 72% using Google Cloud’s TPUs and GPUs; Alphabet will be one of the first companies to provide Nvidia’s GB 200s at scale; management thinks TPUs have very attractive pricing for its capability

We use and offer our customers a range of AI accelerator options, including multiple classes of NVIDIA GPUs and our own custom-built GPUs. We are now on the sixth generation of TPUs known as Trillium and continue to drive efficiencies and better performance with them…

…Using a combination of our TPUs and GPUs, LG AI research reduced inference processing time for its multimodal model by more than 50% and operating costs by 72%…

…We have a wonderful partnership with NVIDIA. We are excited for the GB 200s and will be one of the first to provide it at scale…

…On your first part of the question on the TPUs. If you look at the flash pricing, we’ve been able to deliver externally, I think and how much more attractive it is compared to other models of that capability.

Usage of Alphabet’s Gemini foundation AI model is in a period of dramatic growth by any measure; improvements to Gemini will soon come; all 7 of Alphabet’s products that have more than 2 billion monthly users each use Gemini models; Gemini is now available on GitHub Copilot; Gemini API calls were up 14x in a 6-month period; Snap saw a 2.5 times increase in engagement with its MyAI chatbot after choosing Gemini to power the chatbot’s user experiences; Gemini’s integration with Android is improving Android; the latest Samsung Galaxy devices’ Android operating system has Gemini Live for users to converse with the Gemini model; Alphabet’s latest Pixel 9 devices have Gemini Nano within; development of the third generation of the Gemini model is progressing well; see Point 23 for how Gemini is helping advertisers

By any measure, token volume, API calls, consumer usage business adoption, usage of the Gemini models is in a period of dramatic growth, and our teams are actively working on performance improvements and new capabilities for our range of models. Stay tuned…

… Today, all 7 of our products and platforms with more than 2 billion monthly users use Gemini models, that includes the latest product to surpass the 2 billion user milestone Google Maps…

…Today, we shared that Gemini is now available on GitHub Copilot with more to come…

…Gemini API calls have grown nearly 14x in a 6-month period. When Snap was looking to power more innovative experiences within their MyAI chatbot, they chose Gemini’s strong multimodal capabilities. Since then, Snap saw over 2.5x as much engagement with MyAI in the United States…

… Gemini’s deep integration is improving Android. For example, Gemini Live lets you have free-flowing conversations with Gemini. People love it. It’s available on Android, including Samsung Galaxy devices. We continue to work closely with them to deliver innovations across their newest devices with much more to come. At Made by Google, we unveiled our latest Pixel 9 series of devices featuring advanced AI models, including Gemini Nano. We have seen strong demand for these devices, and they’ve already received multiple awards…

…We’ve had 2 generations of Gemini model. We are working on the third generation, which is progressing well.

Alphabet’s Project Astra will allow AI to see and reason about the physical world around users, and management aims to ship it as early as 2025

When they’re building out experiences where AI can see and reason about the world around you, Project Astra is a glimpse of that future. We are working to ship experiences like this as early as 2025.

Alphabet is using AI internally to improve coding productivity and efficiency; a quarter of new code at Google is now generated by AI

We’re also using AI internally to improve our coding processes, which is boosting productivity and efficiency. Today, more than 1/4 of all new coated Google is generated by AI, then reviewed and accepted by engineers. This helps our engineers do more and move faster. 

Circle to Search is now available on more than 150 million Android devices; a third of users who have tried Circle to Search now use it weekly; Circle to Search has higher engagement with younger users

Circle to Search is now available on over 150 million Android devices with people using it to shop, translate text and learn more about the world around them. 1/3 of the people who have tried Circle to Search now use it weekly, a testament to its helpfulness and potential…

… For example, with Circle to Search, where we see higher engagement from users aged 18 to 24.  

Lens is now used in over 20 billion visual searches per month; Lens is one of the fastest-growing query types management has seen on Search; management started testing product search on Lens in October and found that shoppers are more likely to engage; management is seeing users use Lens for complex multimodal queries; Alphabet has rolled out shopping ads with Lens visual research results to better connect consumers and businesses.

Lens is now used for over 20 billion visual searches per month. Lens is one of the fastest-growing query types we see on search because of its ability to answer complex multimodal questions and help in product discovery and shopping…

…In early October, we announced product search on Google Lens and in testing this feature, we found that shoppers are more likely to engage with content in this new format. We’re also seeing that people are turning to Lens more often to run complex multimodal quarries, voicing a question or inputting text in addition to a visual. Given these new user behaviors earlier this month, we announced the rollout of shopping ads above and alongside relevant Lens visual search results to help better connect consumers and businesses. 

Customers are using Google Cloud’s AI products in 5 different ways: (1) for AI hardware and software infrastructure; (2) for building and customising AI models with Vertex; (3) for combining Google Cloud’s AI platform with its data platform; (4) for AI-powered cybersecurity solutions; and (5) for building AI agents to improve customer engagement

Customers are using our products in 5 different ways. First, our AI infrastructure. which we differentiate with leading performance driven by storage, compute and software advances as well as leading reliability and a leading number of accelerators…

…Second, our enterprise AI platform, Vertex is used to build and customize the best foundation models from Google and the industry…

…Third, customers use our AI platform together with our data platform, Big Query, because we analyze multimodal data no matter where it is stored with ultra low latency access to Gemini…

…Fourth, our AI-powered cybersecurity solutions Google threat intelligence and security operations are helping customers like BBVA and Deloitte, prevent deduct and respond to cybersecurity threats much faster…

… Fifth, in Q3, we broadened our applications portfolio with the introduction of our new customer engagement suite. It’s designed to improve the customer experience online and in mobile apps as well as in call centers, retail stores and more. 

Waymo is the biggest part of Alphabet’s Other Bets portfolio; Alphabet’s management thinks Waymo is the clear technical leader in autonomous vehicles; Waymo is now serving 150,000 paid rides weekly and driving 1 million fully autonomous miles, and is the first autonomous vehicle company to reach these milestones; Waymo is partnering with Uber and Hyundai to deliver autonomous vehicles to more consumers; Waymo is now on its sixth generation system

I want to highlight Waymo, the biggest part of our portfolio. Waymo is now a clear technical leader within the autonomous vehicle industry and creating a growing commercial opportunity. Over the years, Waymo has been infusing cutting edge AI into its work. Now each week, Waymo is driving more than 1 million fully autonomous miles and serves over 150,000 paid rides, the first time any AV company has reached this kind of mainstream use. Through its expanded network and operations partnership with Uber in Austin and Atlanta, plus a new multiyear partnership with Hyundai, Waymo will bring fully autonomous driving to more people and places. By developing a universal driver, Waymo has multiple paths to market. And with its sixth generation system, Waymo significantly reduced unit costs without compromising safety.

Alphabet’s management finds that AI helps Alphabet better understand consumer-intent and connect consumers with advertisers

AI is expanding our ability to understand intent and connect it to our advertisers. This allows us to connect highly relevant users with the most helpful ad and deliver business impact to our customers.

Advertisers are using Gemini to build and test more creatives at scale; Audi worked with Gemini tools to increase website visits by 80% and increase clicks by 2.7 times

Advertisers now use our Gemini power tools to build and test a larger variety of relevant creators at scale. Audi used our AI tools to generate multiple video image and text assets in different links and orientations out of existing long-form videos. They then fed the newly generated creatives into Demand Gen to drive reach, traffic and booking to their driving experience. The campaign increased website visits by 80% and increased clicks by 2.7x, delivering a lift in their sales. 

Alphabet is offering AI-powered campaigns to help advertisers achieve faster feedback on what is working; DoorDash saw a 15x higher conversion rate at a 50% more efficient cost per action

AI-powered campaigns help advertisers get faster feedback on what creatives workwear and redirect the media buying. Using Demand Gen, DoorDash tested a mix of image and video assets to drive more impact across Google and YouTube’s visually immersive surfaces. They saw a 15x higher conversion rate at a 50% more efficient cost per action when compared to video action campaigns alone. 

Alphabet is using AI to help advertisers better measure their advertising results

This quarter, we extended the availability of our open source marketing mix model, Meridian to more customers, helping to scale measurement of cross-channel budgets to drive better business outcomes.

Alphabet’s big jump capex in 2024 Q3 (was $8.1 billion in 2023 Q3) was mostly for technical infrastructure, in the form of servers and data centers; management expects Alphabet’s 2024 Q4 capex to be similar to what was seen in 2024 Q3; Alphabet announced more than $7 billion in planned data center investments in 2024 Q3, with $6 billion in the USA; management expects further growth in capex in 2025, but not at the same percentage increase seen from 2023 to 2024; the use of TPUs at Alphabet helps to drive efficiencies

With respect to CapEx, our reported CapEx in the third quarter was $13 billion, reflecting investment in our technical infrastructure with the largest component being investment in servers, followed by data centers and networking equipment. Looking ahead, we expect quarterly CapEx in the fourth quarter to be at similar levels to Q3…

…In the third quarter alone, we made announcements of over $7 billion in planned data center investments with nearly $6 billion of that in the U.S…

…As you saw in the quarter, we invested $13 billion in CapEx across the company. And as you think about it, it really is divided into 2 categories. One is our technical infrastructure, and that’s the majority of that $13 billion. And the other one goes into areas such as facilities, the bets and other areas across the company. Within TI, we have investments in servers, which includes both TPUs and GPUs. And then the second categories are data centers and networking equipment. This quarter, approximately 60% of that investments in technical infrastructure went towards servers and about 40% towards data center and networking equipment…

…And as you think about the next quarter and going into next year, as I mentioned in my prepared remarks, we will be investing in Q4 at approximately the same level of what we’ve invested in Q3, approximately $13 billion. And as we think into 2025, we do see an increase coming in 2025, and we will provide more color on that on the Q4 call, likely not the same percent step-up that we saw between ’23 and ’24, but additional increase…

…On your first part of the question on the TPUs. If you look at the flash pricing, we’ve been able to deliver externally, I think and how much more attractive it is compared to other models of that capability. I think probably that gives a good sense of the efficiencies we can generate from our architecture. And so — and we are doing the same that for internal use as well. The models for search while they keep going up in capability we’ve been able to really optimize them for the underlying architecture, and that’s where we are seeing a lot of efficiencies as well.  

Amazon (NASDAQ: AMZN)

Amazon’s management believes that AI will be a big piece of the company’s robotics efforts in its fulfilment network

We continue to innovate in robotics to speed delivery, lower cost to serve, and further improve safety in our fulfillment network…

…We really do believe that AI is going to be a big piece of what we do in our robotics network. We had a number of efforts going on there. We just hired a number of people from an incredibly strong robotics AI organization. And I think that will be a very central part of what we do moving forward, too. 

Amazon’s management sees customers focused on new cloud computing efforts again, and the modernisation of their infrastructure, by migrating to the cloud, is important if they want to work on generative AI at scale

Companies are focused on new efforts again, spending energy on modernizing their infrastructure from on-premises to the cloud. This modernization enables companies to save money, innovate more quickly, and get more productivity from their scarce engineering resources. However, it also allows them to organize their data in the right architecture and environment to do generative AI at scale. It’s much harder to be successful and competitive in generative AI if your data is not in the cloud.

AWS has released nearly twice as many AI features in the last 18 months as other leading cloud providers combined; AWS’s AI business is growing at a triple digit rate at a multi-billion revenue run rate; AWS’s AI business is currently growing more than 3x faster than AWS itself grew when AWS was at a similar stage; management sees AI as an unusually large opportunity

In the last 18 months, AWS has released nearly twice as many machine learning and gen AI features as the other leading cloud providers combined. AWS’ AI business is a multibillion-dollar revenue run rate business that continues to grow at a triple-digit year-over-year percentage and is growing more than 3x faster at this stage of its evolution as AWS itself grew, and we felt like AWS grew pretty quickly…

…It is a really unusually large, maybe once-in-a-lifetime type of opportunity. And I think our customers, the business, and our shareholders will feel good about this long term that we’re aggressively pursuing it.

Amazon has a good relationship with NVIDIA, but management have heard from customers that they want better price performance on their AI workloads, and so AWS developed its own AI chips for training and inference; AWS’s second version of its AI chip for model-training, Trainium 2, will ramp up in the next few weeks; management thinks Trainium 2 have very compelling price performance; management is seeing significant interest in Trainium 2, to the extent they have to increase manufacturing orders much more than originally planned

While we have a deep partnership with NVIDIA, we’ve also heard from customers that they want better price performance on their AI workloads. As customers approach higher scale in their implementations, they realize quickly that AI can get costly. It’s why we’ve invested in our own custom silicon in Trainium for training and Inferentia for inference. The second version of Trainium, Trainium2, is starting to ramp up in the next few weeks and will be very compelling for customers on price performance. We’re seeing significant interest in these chips, and we’ve gone back to our manufacturing partners multiple times to produce much more than we’d originally planned…

…We have a very deep partnership with NVIDIA, we tend to be their lead partner on most of their new chips. We were the first to offer H200s in EC2 instances. And I expect us to have a partnership for a very long time that matters.

Amazon’s management is seeing more model builders standardise on SageMaker,  AWS’s fully-managed AI service; SageMaker’s hyperpod capability helps save model-training time by up to 40%

We also continue to see increasingly more model builders standardize on Amazon SageMaker, our service that makes it much easier to manage your AI data, build models, experiment, and deploy to production. This team continues to add features at a rapid clip punctuated by SageMaker’s unique hyperpod capability, which automatically splits training workloads across more than 1,000 AI accelerators, prevents interruptions by periodically saving checkpoints, and automatically repairing faulty instances from their last saved checkpoint, and saving training time by up to 40%.

Amazon’s management believes Amazon Bedrock, AWS’s AI-models-as-a-service offering for companies that want to leverage existing foundation models for customisation, has the broadest selection of leading foundation models; Bedrock recently added Anthropic’s Claude 3.5 Sonnet model, Meta’s Llama 3.2 models, and more; management is seeing companies use models from different providers within the same application and Bedrock makes it easy to orchestrate the disparate models; Bedrock also helps companies with model-access, prompt engineering, and lowering inference costs

At the middle layer where teams want to leverage an existing foundation model, customized with their data, and then have features to deploy high-quality generative AI applications, Amazon Bedrock has the broadest selection of leading foundation models and most compelling modules for key capabilities like model valuation, guardrails, rag and agents. Recently, we’ve added Anthropic’s Claude 3.5 Sonnet model, Meta’s Llama 3.2 models, Mistral’s Large 2 models and multiple-stability AI models. We also continue to see teams use multiple model types from different model providers and multiple model sizes in the same application.  There’s mucking orchestration required to make this happen. And part of what makes Bedrock so appealing to customers and why it has so much traction is that Bedrock makes this much easier. Customers have many other requests: access to even more models, making prompt management easier, further optimizing inference costs. And our Bedrock team is hard at work making this happen.

Amazon’s management continues to see strong adoption of Amazon Q, Amazon’s generative AI assistant for software development; Amazon Q has the highest reported code acceptance rates in the industry; reminder that Amazon saved $260 million and 4,500 developer years when performing a large Java Development Kit migration through the use of Amazon Q

We’re continuing to see strong adoption of Amazon Q, the most capable generative AI-powered assistant for software development and to leverage your own data. Q has the highest reported code acceptance rates in the industry for multiline code suggestions. The team has added all sorts of capabilities in the last few months, but the very practical use case recently shared where Q Transform saving Amazon’s teams $260 million and 4,500 developer years in migrating over 30,000 applications to new versions of the Java JDK. As excited developers and prompted them to ask how else we could help them with tedious and painful transformations.

Amazon is using generative AI pervasively across its businesses, with hundreds of apps in use or in development; Rufus is a generative AI-powered shopping assistant available in parts of Europe, North America, and India; Amazon is using generative AI to improve personalisation and product-search for consumers when shopping; Project Amelia is an AI system offering tailored business insights to Amazon sellers; Alexa, Amazon’s virtual assistant technology, is being rearchitected with new foundation AI models; the new Kindle Scribe has a built-in AI-powered notebook 

We’re also using generative AI pervasively across Amazon’s other businesses with hundreds of apps in development or launched.

For consumers, we’ve expanded Rufus, our generative AI-powered expert shopping assistant, to the U.K., India, Germany, France, Italy, Spain, and Canada. And in the U.S., we’ve added more personalization, the ability to better narrow customer intent and real-time pricing and deal information. We’ve recently debuted AI shopping guides for consumers, which simplifies product research by using generative AI to pair key factors to consider in a product category with Amazon’s wide selection, making it easier for customers to find the right product for their needs. 

For sellers, we’ve recently launched Project Amelia, an AI system that offers tailored business insights to boost productivity and drive seller growth.

We continue to rearchitect the brain of Alexa with a new set of foundation models that we’ll share with customers in the near future, and we’re increasingly adding more AI into all of our devices. Take the new Kindle Scribe we just announced. The note-taking experience is much more powerful with the new built-in AI-powered notebook, which enables you to quickly summarize pages of notes into concise bullets in a script font that can easily be shared.

Amazon’s management expects capital expenditures of $75 billion for the whole of 2024; most of the capex will be for AWS infrastructure to support demand for AI services; the capex also includes investments in Amazon’s fulfilment and transportation network; management expects capex in 2025 to increase from 2024’s level, with most of the capex for AWS, specifically generative AI; reminder that the faster AWS grows, the faster Amazon needs to invest capital for hardware; many of the assets AWS’s capex is invested in have long, useful lives; management expects to deliver high returns on invested capital with AWS’s generative AI investments; management has a lot of experience, accumulated over the years, in predicting just the right amount of compute capacity to provide for AWS before the generative AI era, and they believe they can do so again for generative AI

Year-to-date capital investments were $51.9 billion. We expect to spend approximately $75 billion in CapEx in 2024. The majority of the spend is to support the growing need for technology infrastructure. This primarily relates to AWS as we invest to support demand for our AI services while also including technology infrastructure to support our North America and international segments. Additionally, we’re continuing to invest in our fulfillment and transportation network to support the growth of the business, improve delivery speeds and lower our cost to serve. This includes investments in same-day delivery facilities, in our inbound network and as well in robotics and automation…

… I’ll take the CapEx part of that. As Brian said in his opening comments, we expect to spend about $75 billion in 2024. I suspect we’ll spend more than that in 2025. And the majority of it is for AWS, and specifically, the increased bumps here are really driven by generative AI…

…The thing to remember about the AWS business is the cash life cycle is such that the faster we grow demand, the faster we have to invest capital in data centers and networking gear and hardware. And of course, in the hardware of AI, the accelerators or the chips are more expensive than the CPU hardware. And so we invest in all of that upfront in advance of when we can monetize it with customers using the resources…

…A lot of these assets are many-year useful life assets. Data centers, for instance, are useful assets for 20 to 30 years…

…I think we’ve proven over time that we can drive enough operating income and free cash flow to make this very successful return on invested capital business. And we expect the same thing will happen here with generative AI…

…One of the least understood parts about AWS over time is that it is a massive logistics challenge. If you think about we have 35-or-so regions around the world, which is an area of the world where we have multiple data centers, and then probably about 130 availability zone through data centers, and then we have thousands of SKUs we have to land in all those facilities. And if you land too little of them, you end up with shortages, which end up in outages for customers. So most don’t end up with too little, they end up with too much. And if you end up with too much, the economics are woefully inefficient. And I think you can see from our economics that we’ve done a pretty good job over time at managing those types of logistics and capacity. And it’s meant that we’ve had to develop very sophisticated models in anticipating how much capacity we need, where, in which SKUs and units.

And so I think that the AI space is, for sure, earlier stage, more fluid and dynamic than our non-AI part of AWS. But it’s also true that people aren’t showing up for 30,000 chips in a day. They’re planning in advance. So we have very significant demand signals giving us an idea about how much we need…

…There are some similarities in the early days here of AI, where the offerings are new and people are very excited about it. It’s moving very quickly and the margins are lower than what I think they will be over time. The same was true with AWS. If you looked at our margins around the time you were citing, in 2010, they were pretty different than they are now. I think as the market matures over time, there are going to be very healthy margins here in the generative AI space.

There are a few hundred million active Alexa devices; management had an initial vision of Alexa being the world’s best personal assistant and they believe now that Alexa’s re-architecture can give it a shot at fulfilling the initial vision

I think we have a really broad number of Alexa devices all over people’s homes and offices and automobiles and hospitality suites. We’ve about 0.5 billion devices out there with a couple of hundred million active end points. And when we first were pursuing Alexa, we had this vision of it being the world’s best personal assistant and people thought that was kind of a crazy idea. And I think if you look at what’s happened in generative AI over the last couple of years, I think you’re kind of missing the boat if you don’t believe that’s going to happen. It absolutely is going to happen. So we have a really broad footprint where we believe if we rearchitect the brains of Alexa with next-generation foundational models, which we’re in the process of doing, we have an opportunity to be the leader in that space.

Amazon’s management believes that AWS’s demand substantially outweighs capacity today; management believes AWS’s rate of growth can improve over time as capacity grows

[Question] On the cloud, are you at all capacity constrained, and will the new Trainium or NVIDIA chips maybe even drive sales growth faster?

[Answer] I believe we have more demand that we could fulfill if we had even more capacity today. I think pretty much everyone today has less capacity than they have demand for, and it’s really primarily chips that are the area where companies could use more supply…

…We’re growing at a very rapid rate and have grown a pretty big business here in the AI space. And it’s early days, but I actually believe that the rate of growth there has a chance to improve over time as we have bigger and bigger capacity.

Apple (NASDAQ: AAPL)

Apple announced Apple Intelligence in June 2024; Apple Intelligence redefines privacy in AI; Apple recently released the first set of Apple Intelligence features in US English for iPhone, iPad, and Mac users, and they include writing tools, an improved version of Siri, a more intelligent Photos App, and notification summaries and priority messages; more Apple Intelligence features will be released in December 2024 and early developer feedback is great; the adoption rate of iOS18 in its first three days is twice as fast as for iOS17, suggesting interest for Apple Intelligence; Apple will release support for additional languages in Apple Intelligence in April 2025

In June, we announced Apple Intelligence, a remarkable personal intelligent system that combines the power of generative models with personal context to deliver intelligence that is incredibly useful and relevant. Apple Intelligence marks the beginning of a new chapter for Apple Innovation and redefines privacy and AI by extending our groundbreaking approach to privacy into the cloud with private cloud compute. Earlier this week, we made the first set of Apple Intelligence features available in U.S. English for iPhone, iPad and Mac users with system-wide writing tools that help you refine your writing, a more natural and conversational Siri, a more intelligent Photos app, including the ability to create movies simply by typing a description, and new ways to prioritize and stay in the moment with notification summaries and priority messages.

And we look forward to additional intelligence features in December with even more powerful writing tools, a new visual intelligence experience that builds on Apple Intelligence and ChatGPT integration as well as localized English in several countries, including the U.K., Australia and Canada. These features have already been provided to developers, and we’re getting great feedback. More features will be rolling out in the coming months as well as support for more languages, and this is just the beginning…

…[Question] I was wondering if you could just expand a little bit on some of the early feedback to Apple Intelligence, both for iOS 18.1 but also the developer beta so far and whether you would attribute Apple Intelligence to any of the strong iPhone performance that we’ve seen to date.

[Answer] We’re getting a lot of positive feedback from developers and customers. And in fact, if you just look at the first 3 days, which is all we have obviously from Monday, the 18.1 adoption is twice as fast as the 17.1 adoption was in the year ago quarter. And so there’s definitely interest out there for Apple Intelligence…

…We started in the — with U.S. English. That started on Monday. There’s another release coming that adds additional features that I had referenced in December in not only U.S. English but also localized for U.K., Australia, Canada, Ireland and New Zealand. And then we will add more languages in April. We haven’t set the specifics yet in terms of the languages, but we’ll add more in April and then more as we step through the year. And so we’re moving just as fast as possible while ensuring quality.

Apple’s management is building the infrastructure to deliver Apple Intelligence, but it does not seem like Apple will need to significantly increase its capex budget from historical norms; management also does not see any significant change to the intensity of research & development (R&D) spending that Apple needs to invest in AI

[Question] Could you just talk a little bit about the CapEx outlook and whether investments in things like private cloud compute could change the historical CapEx range of roughly $10 billion a year?

[Answer] We are rolling out these features, Apple Intelligence features already now. And so we are making all the capacity that is needed available for these features. You will see in our 10-K the amount of CapEx that we’ve incurred during the course of fiscal ’24. And we will — in fiscal ’25, we will continue to make all the investments that are necessary, and of course, the investments in AI-related CapEx will be made…

…[Question] Given how much your tech peers are spending on AI, does this new era of Apple Intelligence actually require Apple to invest more in R&D beyond your current 7% to 8% of sales to capture this opportunity? 

[Answer] We’ve been investing heavily in R&D over the last several years. Our R&D growth has been significant during the last several years. And obviously, as we move through the course of fiscal ’24, we’ve also reallocated some of the existing resources to this new technology, to AI. And so the level of intensity that we’re putting into AI has increased a lot, and you maybe don’t see the full extent of it because we’ve also had some internal reallocation of the base of engineering resources that we have within the company.

Apple’s management thinks the introduction of Apple Intelligence will benefit the entire Apple ecosystem

[Question] I understand Apple Intelligence is a feature on the phone today. But do you think that in the future it could potentially have or benefit the services growth business? Or is that too — are those too bifurcated to even make a call on the — this early in the cycle?

[Answer] Keep in mind that we have released a lot of APIs, and developers will be taking advantage of those APIs. That release has occurred as well, and of course, more are coming. And so I definitely believe that a lot of developers will be taking advantage of Apple Intelligence in a big way. And what that does to services, I’ll not forecast, but I would say that from an ecosystem point of view, I think it will be great for the user and the user experience.

Arista Networks (NYSE: ANET)

Arista Networks’ management is seeing networking for AI gaining a lot of traction; trials that took place in 2023 are becoming pilots in 2024; management expects more production in 2025 and 2026

Networking for AI is gaining a lot of traction as we move from trials in 2023 to more pilots in 2024, connecting to thousands of GPUs, and we expect more production in 2025 and 2026.

AI data traffic is very different from traditional cloud workloads and smooth and consistent data flow is a crucial factor in AI networking

AI traffic differs greatly from cloud workloads in terms of diversity, duration and size of flow. The fidelity of AI traffic flows where the slowest flow matters and one slow flow could slow down the entire job completion time is a crucial factor in networking.

Arista Networks’ management sees the company becoming a pioneer in scale-out Ethernet accelerated networking for large AI workloads; Arista Networks’ new Etherlink portfolio scales well to networks with over 100,000 GPUs and can potentially even handle 1 million GPU clusters; Arista Networks’ latest 77R4 DES platform was launched in close collaboration with Meta Platforms

Our AI centers connect seamlessly from the back end to the front end of compute, storage, WAN and classic cloud networks. Arista is emerging as the a pioneer and scale-out Ethernet accelerated networking for large-scale training and AI workloads. Our new Etherlink portfolio with wire speed 800-gig throughput and non-blocking performance, scales from single tier to efficient 2-tier networks for over 100,000 GPUs, potentially even 1 million AI accelerators with multiple tiers. Our accelerated AI networking portfolio consists of 3 families with over 20 switching products and not just one point switch. At the recent OCP in mid-October 2024, we officially launched a very unique platform that distributed Etherlink 7700 to build 2 tier networks for up to 10,000 GPU clusters. The 77R4 DES platform was developed in close collaboration with Meta. And while it may physically look like and be cable like a 2-tier leaf spine network, DES provides a single-stage forwarding with highly efficient spine fabric, eliminating the need for tuning and encouraging fast failover for large AI accelerator-based clusters. 

Arista Networks’ management believes the company has the broadest set of 800 gigabit per second Ethernet products for AI networks

I’m pleased to report Arista 7700R4 distributed Etherlink switch, the 7800R4 Spine, along with the 7060X6 AI leaf that we announced in June have entered into production providing our customers the broadest set of 800 gigabit per second Ethernet products for their AI networks. Together with 800 gigabit per second parallel optics, our customers are able to connect to 400 gigabit per second GPUs to each port increasing the deployment density over current switching solutions. This broad range of Ethernet platforms allows our customers to optimize density and minimize tiers to best match the requirements of their AI workload.

New AI clusters require high-speed connections to existing backbones

New AI clusters require new high-speed port connections into the existing backbone. These new clusters also increased bandwidth on the backbone to access training data, capture snapshots and deliver results generated by the cluster. This trend is providing increased demand for 7800R3 400-gigabit solution.

Arista Networks’ management sees next-generation AI data centres needing significantly more power while doubling network performance

Next-generation data centers integrating AI will contend with significant increases in power consumption while looking to double network performance.

Arista Networks’ management thinks the adoption of AI networking will rest on specifications that the Ultra Ethernet Consortium (UEC) is expected to soon release; the UEC now has 97 members and Arista Networks is a founding member

Critical to the rapid adoption of AI networking is the Ultra Ethernet consortium specification expected imminently with Arista’s key contributions as a founding member. The UEC ecosystem for AI has evolved to over 97 members.

Arista Networks’ management thinks Ethernet is the only option for open standard space AI networking

In our view, Ethernet is the only long-term viable direction for open standard space AI networking.

Arista Networks’ business growth in 2024 was achieved partly with the help of AI; management is now projecting even more growth in 2025 and is confident of achieving its AI back-end revenue target of US$750 million; the adoption of Arista Networks’ AI back-end products influences the adoption of its front-end AI networking products too; management also expects Arista Networks’ front-end AI networking products to generate around US$750 million in revenue in 2025, but sometimes this gets hard to track; the US$750 million in AI back-end revenue that management expects are brand new for the company

We’ve experienced some pretty amazing growth years with 33.8% growth in ’23 and 2024 appears to be heading at least to 18%, exceeding our prior predictions of 10% to 12%. This is quite a jump in 2024, influenced by faster AI pilots. We are now projecting an annual growth of 15% to 17% next year, translating to approximately $8 billion in 2025 revenue with a healthy expectation of operating margin. Within that $8 billion revenue target, we are quite confident in achieving our campus and AI by back-end networking targets of $750 million each in 2025 that we set way back 1 or 2 years ago. It’s important to recognize though that the back end of AI will influence the front-end AI network and its ratios. This ratio can be anywhere from 30% to 100% and sometimes, we’ve seen it as high as 200% of the back-end network depending on the training requirements. Our comprehensive AI center networking number is therefore likely to be double of our back-end target of $750 million, now aiming for approximately $1.5 billion in 2025…

… I would expect in the back end, any share Arista gets, including that $750 million is incremental. It’s brand new to us. We were never there before…

…I think it all depends on their approach to AI. If they just want to build a back-end cluster and prove something out, then they just look for the highest job training completion and intense training models. And it’s a very narrow use case. But what we’re starting to see more and more, especially with the top 5, like I said, is for every dollar spent in the back end, you could spend 30% more, 100% more, and we’ve even seen a 200% more scenario, which is why our $750 million will carry over to, we believe, next year, another $750 million on front-end traffic that will include AI, but it will include other things as well. It won’t be unique to AI. So I wouldn’t be surprised if that number is anywhere between 30% and 100%, so the average is 100%., which is 2x our back-end number. So feeling pretty good about that. Don’t know how to exactly count that as pure AI, which is why I qualify it by saying increasingly, if you start having inference, training, front end, storage, WAN, classic cloud all come together, the AI — the pure AI number becomes difficult to track.

Arista Networks’ management is stocking up inventory in preparation for a rapid deployment of AI networking products

On the cash front, while we have experienced significant increases in operating cash over the last couple of quarters, we anticipate an increase in working capital requirements in Q4. This is primarily driven by increased inventory in order to respond to the rapid deployment of AI networks and to reduce overall lead times as we move into 2025.

Arista Networks’ management has been surprised by the acceleration of AI pilots by its customers in 2024; management would not be surprised going forward if its AI business grows faster than its classic data center and cloud business (in other words, management would not be surprised if the company’s customers cannibalise some of their classic data center and cloud buildouts for AI)

We were pleasantly surprised with the faster acceleration of AI pilots in 2024. So we definitely see that our large cloud customers are continuing to refresh on the cloud, but are pivoting very aggressively to AI. So it wouldn’t surprise me if we grow faster in AI and faster in campus in the new center markets and slower in our classic markets called that data center and cloud. 

The 4 major AI trials Arista Networks discussed in the 2024 Q1 earnings call have now become 5 trials; 3 of the 5 customers are progressing well and are transitioning from trials to pilots, and they will each grow their GPU clusters from 50,000 to 100,000 in 2025; the customer for the new trial that was started has historically been very focused on Infiniband so management is happy to have won the trial, and management hopes the trail will enter pilot and production in 2025; the last remaining customer is moving slower than management expected with delays in their data center buildout; management has good revenue visibility for 3 of the 5 trials for the next 6-12 months and Arista Networks’ revenue-guide for 2025 does not depend on the remaining 2 trials; a majority of the trials are currently on Arista Networks’ 400-gig products because the customers are waiting for the ecosystem to develop on the 800-gig products, but management expects more adoption of the 800-gig products in 2025; Arista Networks is participating in other smaller AI trials too, but the difference is that management expects the 5 major ones to scale to at least 100,000 GPU clusters 

Arista now believes we’re actually 5 out of 5, not 4 out of 5. We are progressing very well in 4 out of the 5 clusters. 3 of the customers are moving from trials to pilots this year, and we’re expecting those 3 to become 50,000 to 100,000 GPU clusters in 2025. We’re also pleased with the new Ethernet trial in 2024 with our fifth customer. This customer was historically very, very InfiniBand driven. And we are now moving in that particular fifth customer, we are largely in a trial mode in 2024, and we hope to go to pilots and production in 2025. There is one customer who — so 3 are going well. One is starting. The fifth customer is moving slower than we expected. They may get back on their feet. In 2025, they’re awaiting new GPUs, and they’ve got some challenges on power cooling, et cetera. So 3, I would give an A. The fourth one, we’re really glad we won, and we’re getting started and the fifth one, I’d say, steady-state, not quite as great as we would expect them — have expected them to be…

…[Question] I wanted to ask a little bit more about the $750 million in AI for next year. Has your visibility on that improved over the last few months? I wanted to reconcile your comment around the fifth customer not going slower than expected. And it sounds like you’re now in 5 of 5, but wondering if that fifth customer going slower is limiting upside or limiting your visibility there?

[Answer] I think on 3 out of the 5, we have good visibility, at least for the next 6 months, maybe even 12…

…On the fourth one, we are in early trials. We’ve got some improving to do. So let’s see, but we’re not looking for 2025 to be the bang up year on the fourth one. It’s probably 2026. And on the fifth one, we’re a little bit stalled, which may be why we’re being careful about predicting how they’ll do. They may step in nicely in the second half of ’25, in which case, we’ll let you know. But if they don’t, we’re still feeling good about our guide for ’25…

…A majority of the trials and pilots are on 400 because people are still waiting for the ecosystem at 800, including the NICs and the UEC and the packet spring capabilities, et cetera. So while we’re in some early trials on 800, majority of 400 — majority of 2024 is 400 gig. I expect as we go into 2025, we will see a better split between 400 and 800…

… So we’re not saying these 5 are the be-all, end-all, but these are the 5 we predict can go to 100,000 GPUs and more. That’s the way to look at this. So there are the largest AI Titans, if you will. And they can be in the cloud, hyperscaler Titan group, they could be in the Tier 2 as well, by the way, very rarely would they be in a classic enterprise. By the way, we do have at least 10 to 15 trials going on in the classic enterprise too, but they’re much smaller GPU counts, so we don’t talk about it.

Arista Networks’ management sees NVIDIA both as a partner and a competitor in the AI networking market; Arista Networks does see NVIDIA’s Infiniband as a competing solution, but rarely sees NVIDIA’s own Ethernet solution competing; management thinks customers, ranging from those building large GPU clusters to smaller ones, all see Arista Networks as the expert when it comes to AI networking

We view NVIDIA as a good partner. If we didn’t have the ability to connect to their GPUs, we wouldn’t have all this AI networking demand. So thank you, NVIDIA. Thank you, Jensen, for the partnership. Now as you know, NVIDIA sells the full stack and most of the time, it’s with InfiniBand, and with the Mellanox acquisition, they do have some Ethernet capability. We personally do not run into the Ethernet capability very much. We run into it, maybe in 1 or 2 customers. And so generally speaking, Arista has looked upon as the expert there. We have a full portfolio. We have full software. And whether it’s the large scale-out ethernet working customers like the Titans or even the smaller enterprises, we’re seeing a lot of smaller GPU clusters of the enterprise, Arista is looked upon as the expert there. But that’s not to say we’re going to win 100%. We certainly welcome NVIDIA as a partner on the GPU side and a fierce competitor, and we look to compete with them on the Ethernet switching.

The AI back-end market is where Arista Networks natively connects with GPU and where NVIDIA’s Infiniband is the market leader, but Arista Networks’ Ethernet solution is aiming to be the gold standard; for the AI front-end market, Arista Networks’ solutions are the gold standard and management is seeing some customers fail to run their AI application on competing solutions and want to replace them with Arista Networks’ solutions

So since you asked me specifically about AI as opposed to cloud, let me parse this problem into 2 halves, the back end and the front end, right? At the back end, we’re natively connecting to GPUs. And there can be many times, we just don’t say it because somebody just bundles it in the GPU in particular, NVIDIA. And you may remember a year ago, I was saying we’re outside looking in because most of the bundling is happening with InfiniBand…

…So we’ll take all we can get, but we are not claiming to be a market leader there. We’re, in fact, claiming that there are many incumbents there with InfiniBand and smaller versions of Ethernet that Arista is looking to gain more credibility and experience and become the gold standard for the back end.

On the front end, in many ways, we are viewed as the gold standard. So competitively, it’s a much more complex network. You have to build a leaf-spine architecture. John alluded to this, there’s a tremendous amount of scale with L2, L3, EVPN, VXLAN, visibility, telemetry, automation, routing at scale, encryption at scale. And this, what I would call accelerated networking portfolio complements NVIDIA’s accelerated compute portfolio. And compared to all the peers you mentioned, we have the absolute best portfolio of 20 switches and 3 families and the capability and the competitive differentiation is bar none. In fact, I am specifically aware of a couple of situations where the AI applications aren’t even running on some of the industry peers you talked about, and they want to swap theirs for ours. So feeling extremely bullish with the 7800 flagship product, the newly introduced 7700 that we worked closely with Meta, the 7060, this product line running today mostly at 400 gig because a lot of the NIC and the ecosystem isn’t there for 800. But moving forward into 800, this is why John and the team are building the supply chain to get ready for it.

ASML (NASDAQ: ASML)

While ASML’s management has seen the strong performance of AI continue – and expects the performance to continue for some time – other market segments have taken longer to recover than management expected; in the Memory segment, management is seeing limited capacity additions among customers, apart from AI, as the customers embark on technology transition to HBM and DDR5

There have been quite some market dynamics in the past couple of months. Very clearly, the strong performance of AI clearly continues and I think it continues to come with quite some upside. We will also see that in other market segments, it takes longer to recover. Recovery is there, but it’s more gradual than what we anticipated before and it will continue in 2025. That does lead to some customer cautiousness…

…If you look at the Memory business, this customer cautiousness that I talked about, leads to limited capacity additions. While at the same time, we do see a lot of focus and strong demand when it comes to technology transitions and particularly as it is related to High Bandwidth Memory and to DDR5. So again, there anything related to AI is strong, but other than that there are limited capacity additions.

The AI growth-driver is very strong over the long-term and ASML’s management sees that AI is increasing share in ASML’s customers’ business

If you look at the long-term outlook, I believe the growth drivers are still very much intact. The secular growth drivers are clear and they are strong. I think if you look at AI, very, very strong, very clear and undisputed. Taking an increasing share in the business of our customers. So I think that is going very strongly.

ASML’s management is seeing upside on AI because the overall demand for AI applications continues to increase, which has driven a recovery in server demand, but management does not have complete understanding on how the AI market will play out

We also mentioned some upside on AI, because we still believe that the overall demand for those application is there, continue to increase. So if we look at the server demand, we see there a very nice recovery. A lot of that has to do with AI application. So we talk about upside, which also means that the overall dynamic of the market is still playing. And we felt the need to provide an update for next year based on some of the development we have seen. I think in no way we are also saying that there is a complete understanding of how the entire market will continue to play out in the next few months. So I think on the second part of your question, I would say maybe this has not played out fully yet…

…[Question] You would expect to happen then, I guess, to — at some point will happen?

[Answer] Well, I think if everyone — and I think a lot of us still believe in a strong AI demand in the coming years, I think that demand has to be fulfilled. Therefore, yes, I will say mostly, we will see some development also on that front in the coming months.

Datadog (NASDAQ: DDOG)

Datadog’s management is seeing next-gen AI customers want to obtain visibility into their AI usage as they continue experimenting with the technology; around 3,000 customers used at least one of Datadog’s AI integrations at the end of 2024 Q3; management is starting to see Datadog’s LLM (large language model) observability products gain traction as AI experiments start becoming production applications; hundreds of customers are already using LLM observability, and some customers have reduced time spent on investigating LLM issues from days or hours to just minutes; management is seeing customers wanting to use APM (Application Performance Monitoring) alongside LLM observability 

In the next-gen AI space, customers continue to experiment with new AI technologies. And as they do, they want to get visibility into their AI use. At the end of Q3, about 3,000 customers used one or more Datadog AI integrations to send us data about their AI, machine learning and LLM usage. As some of these experiments start turning into production AI applications, we are seeing initial signs of traction for our LLM observability products.

Today, hundreds of customers are using LLM observability with more exploring it every day. And some of our first paying customers have told us that they have cut the time spent investigating LLM latency, errors and quality from days or hours to just minutes. Our customers not only want to understand the performance and cost of their LLM applications, they also want to understand the LLM model performance within the context of their entire application. So they are using APM alongside LLM observability to get fully integrated end-to-end visibility across all their applications and tech stacks

AI-native customers accounted for 6% of Datadog’s ARR in 2024 Q3 (was 6% 2024 Q2); AI-native customers contributed 4 percentage points to Datadog’s year-on-year growth in 2024 Q3, compared to 2 percentage points in 2023 Q3; management has seen a very rapid ramp in usage of Datadog among large customers in the AI-native cohort, and management thinks these customers will optimise cloud and observability usage in the future, while also asking for better terms; management is seeing Datadog’s production-minded LLM observability products being used by real paying customers with real volumes in real production workloads; AI-native companies are model providers or AI infrastructure providers that serve as a proxy for the AI industry

AI native customers who this quarter represented more than 6% of our Q3 ARR, up from more than 4% in Q2 and about 2.5% of our ARR in the year ago quarter. AI native customers contributed about 4 percentage points of year-over-year growth in Q3 versus about 2 percentage points in the year ago quarter. While we believe that adoption of AI will continue to benefit Datadog in the long term, we are mindful that some of the large customers in this cohort have ramped extremely rapidly and that these customers may optimize cloud and observability usage and increase their commitments to us over time with better terms. This may create volatility in our revenue growth in future quarters on the backdrop of long-term volume growth…

…We are seeing our production-minded LLM observability products, for example, being used by real paying customers with real volumes and real applications in real production workloads. So that’s exciting and healthy. I think it’s a great trend for the future…

… We have that group of AI, like smaller — relatively small number of AI companies or AI native companies. Many of them are model providers or infrastructure providers for AI that serve the rest of the industry and they are really a proxy for the future growth of the rest of the industry in AI.

Datadog signed a 7-figure expansion deal with a hyperscaler delivering next-gen AI models; the hyperscaler has its homegrown observability solution, but the solution needs time-consuming customisation and manual configuration; the hyperscaler chose Datadog because Datadog’s platform can scale flexibly

We signed a 7-figure annualized expansion with a division of a hyperscaler delivering next-gen AI models. This customer is very technically capable and already has a homegrown observability solution, which requires time-consuming customization and manual configuration. They will be launching new features for their large language models soon and need a platform that can scale flexibly while supporting proactive incident detection. By expanding the use of Datadog, they expect to efficiently onboard new teams and environments and support the rapidly increasing adoption of the LLMs.

Datadog’s management continues to believe that digital transformation, cloud migration, and AI adoption are long-term growth drivers of Datadog’s business

Overall, we continue to see no change to the multiyear trend towards digital transformation and cloud migration, which we continue to believe are still in early days. We are seeing continued experimentation with new advances such as next-gen AI, and we believe this is one of the many factors that will drive greater use of the cloud and other modern technologies.

Datadog’s management is starting to see more inference AI workloads, but they are still concentrated among API-driven providers and it’s still very early days in terms of customers putting their next-gen AI applications into production; management expects more diversification to occur in the future as more companies enter production with their applications and customise their models 

In terms of the workloads, you’re right that we’re starting to see more inference workloads, but they still tend to be more concentrated across a number of API-driven providers. So there are a few others, both on LLMs and other kinds of models. So this is where I think most of the usage in production at least is today. We expect that to diversify more over time as companies get further into production with their applications and they start to be customizing more on their models…

…We are excited to see what’s happening with the AI innovation as it gets further down the pipe and away from testing and experimenting and more into production applications. And we have some signs that it’s starting to happen. Again, we see that with our LLM observability product. We see that also with some of the workloads we monitor from our customers on the infrastructure side. But I would say it’s still very early days in terms of customers being in production with their next-gen AI applications.

Datadog’s management is seeing a small amount of cloud workloads of companies being cannibalised by their AI initiatives

You’re right that the — where the workloads could have grown maybe instead of growing 20%, they could grow 25%, maybe some of those 5% instead are being invested both in terms of infrastructure budget or innovation — time innovation budget. All that is going into AI, and that’s largely right now in experimentation and model training and that sort of thing. 

Datadog’s Management is working with customers with large inference workloads on how Datadog can be helpful on the GPU profiling side of inference; management is also experimenting with how Datadog can be helpful on side of training; management thinks that in a steady state, 60% of AI workloads will be inference and 40% will be training, so there’s still a lot of value to be found if Datadog can be useful in the training side too

Right now, we’re working with a number of customers that have real-world large inference workloads on how we can help on the GPU profiling side for inference. We’re doing less on the training side, mostly because the training jobs tend to be more bespoke and temporary, and there’s less of an application that’s attached to those that these are just very large clusters of GPUs. So it’s closer to HPC in a way than it is to traditional applications, though we are also experimenting with what we can do there. There is a world where maybe in a durable fashion, 60% of workloads are inference and 40% are training. And if that’s the case, there’s going to be a lot of value to be had by having repeatable training and repeatable tooling for that. So we are also looking into that.

Datadog is not monetising GPU instances as well as CPU instances today, but management thinks that could change in the future

As of today, we really don’t monetize GPU instances all that well compared to the other CPU instances. So GPU instance is many times the cost of a CPU instance, and we charge the same amount for it. That doesn’t have to be the case in the future. If we do things that are particularly interesting and it’s going to have a real impact on — and deliver value and how customers use and make the best of their GPUs and in the end, save money. 

Datadog’s management is seeing Datadog’s AI-native cohort grow faster than its cloud-native cohorts did in the late 2010s and early 2020s

What we’ve seen with cloud native in the late ’10s and early ’20s, where we had these numbers of cloud-native consumer companies that were growing very fast, with 2 differences. The first one is that the AI cohort is growing faster and there are larger individual ACVs [annual contract value] for these customers.

Datadog’s management thinks that workloads on Datadog’s platform could really accelerate when non-AI-native companies start bringing AI applications into production

In terms of the growth of workloads, look, I mean, as we said, we see growth across the customer base pretty much. We see growth of classical workloads in the cloud. We see large growth — very large growth on the AI native side. We think that the one big catalyst for future acceleration will be those AI native applications or those AI applications, I should say, going into production for non-AI native companies for a much broader set of customers than the customers that are deploying these kind of applications to their — in production. And as they do, they will also look less like just large cluster of GPUs and more like traditional applications because the GPU needs a database, it needs [ core ] application in front of it, it needs layers to secure it and authorize it and all the other things. So it’s going to look a lot more like a normal application with some additional more concentrated compute and GPUs.

Datadog’s management does not expect Datadog to make outsized investments in GPU clusters for compute

Unlike many others, we don’t expect at this point to have outsized investments in compute. We’re not building absolutely large GPU clusters.

dLocal (NASDAQ: DLO)

dLocal’s management launched the Smart Requests functionality in 2024 Q3 that improves conversion rates for merchants by 1.22 percentage points on average, which equates to a 1.2% increase in revenue for merchants; Smart Requests relies on localised machine learning models to maximise authorisation rates for merchant

During the quarter, we launched our smart requests functionality, boosting our transaction performance and therefore, improving conversion rates by an average of 1.22 percentage points across the board. It may sound minor, but it isn’t. It actually represents, in practical terms, 1.2% additional revenue to our merchants. Smart requests rely on per country machine learning models that optimize routing and chaining so as to maximize authorization rates for our merchants.

Fiverr (NYSE: FVRR)

Fiverr’s management believes that Fiverr’s next generation of products must empower its community to fully leverage AI, and that the best work will be done in the future by a combination of humans and AI

One thing that became clearer to me in the last year is that with the emergence of GenAI and the promise of AGI, the next generation of products we build must empower our community to fully leverage artificial intelligence. It also became clear to me that in the future, the best work will be done by humans and AI technology together, not humans alone or AI alone.

Fiverr’s management is providing Fiverr’s customers with an AI assistant to help them navigate the company’s platform 

This means that every business that comes to Fiverr will have a world-class AI assistant to help them get things done, from ideation, scoping and briefing to project management and workflow automation. It means that they can seamlessly leverage both human talent and machine intelligence to create the most beautiful results.

Fiverr’s management is building a new search experience on the Fiverr platform for buyers which incorporates Neo, its AI powered smart matching tool; Fiverr has launched Dynamic Matching to allow buyers to put together project briefs with an AI assistant to help them get matched to the most relevant freelancers; these new features have experienced enthusiastic reception in just a few weeks; projects that use these new features are bigger projects than the typical scope of projects on Fiverr

On the buyer side, we are building a new search experience that not only includes more dynamic catalogs but also incorporates Neo, an AI-powered smart matching tool, to help customers match with more contextual information. We launched Dynamic Matching to allow buyers to put together comprehensive briefs with a powerful AI assistant and then get matched with the most relevant freelancer with a tailored proposal…

…Even in the few weeks since we launched these products, we have already seen an enthusiastic reception from our community and promising performance. The projects that come through these products are several times larger than a typical project on Fiverr, and we believe it has a lot more potential down the road as the awareness and trust of these products grow on the platform.

Mastercard (NYSE: MA)

Mastercard acquired Brighterion in 2017 to use AI capabilities for decision intelligence; after boosting the product with generative AI, Mastercard has seen a 20% lift in the product 

One of the more recent ones that we talked about that we invested heavily in using our Brighterion acquisition from back in 2017 to use our AI capabilities is decision intelligence. We’ve now boosted the product with Gen AI and the outcome that we see is tremendous. This is up to a 20% lift that we see.

Meta Platforms (NASDAQ: META)

Meta’s management is seeing rapid adoption of Meta AI and Llama; Meta AI now has more than 500 million monthly actives; Llama token usage has grown exponentially in 2024 so far; Meta released Llama 3.2 in 2024 Q3; the public sector is adopting Llama; management is seeing higher usage of Meta AI as the models improve; Meta AI is built on Llama 3.2; voice functions for Meta AI are now available in English in the USA, Australia, Canada, and New Zealand; image editing through simple text prompts, and the ability to learn about images, are now available in Meta AI in the USA; Meta AI remains on track to be the most-used AI assistant in the world by end-2024; early use cases for Meta AI are for information gathering, help with how-to tasks, explore interests, look for content, and generate images

We’re seeing rapid adoption of Meta AI and Llama, which is quickly becoming a standard across the industry…

…Meta AI now has more than 500 million monthly actives…

…Llama token usage has grown exponentially this year and the more widely that Llama gets adopted and becomes the industry standard the more that the improvements to its quality and efficiency will flow back to all of our products. This quarter, we released Llama 3.2, including the leading small models that run on device and open source multimodal models…

…We’re also working with the public sector to adopt Llama across the U.S. government…

…We’re seeing lifts in usage as we improve our models and have introduced a number of enhancements in recent months to make Meta AI more helpful in engaging. Last month, we began introducing voice, so you can speak with Meta AI more naturally, and it’s now fully available in English to people in the U.S., Australia, Canada and New Zealand. In the U.S., people can now also upload photos to Meta AI to learn more about them, write captions for post and add, remove or change things about their images with a simple text prompt. These are all built with our first multimodal foundation model, Llama 3.2…

…We’re excited about the progress of Meta AI. It’s obviously very early in its journey, but it continues to be on track to be the most used AI assistant in the world by end of year…

… Number of the frequent use cases we’re seeing include information gathering, help with how-to tasks, which is the largest use case. But we also see people using it to go deeper on interests, to look for content on our services, for image generation, that’s also been another pretty popular use case so far.

Meta’s management is seeing AI have a positive impact on nearly all aspects of Meta; improvements to Meta’s AI-driven feed and video recommendations have driven increases in time spent on Facebook this year by 8% and on Instagram by 6%; more than 1 million advertisers are using Meta’s Gen AI tools and advertisers using image generation are enjoying a 7% increase in conversions; management sees plenty of new opportunities for new AI advances to accelerate Meta’s core business, so they want to invest more there

We’re seeing AI have a positive impact on nearly all aspects of our work from our core business engagement and monetization to our long-term road maps for new services and computing platforms…

…Improvements to our AI-driven feed and video recommendations have led to an 8% increase in time spent on Facebook and a 6% increase on Instagram this year alone. More than 1 million advertisers used our Gen AI tools to create more than 15 million ads in the last month. And we estimate that businesses using image generation are seeing a 7% increase in conversions and we believe that there’s a lot more upside here…

…It’s clear that there are a lot of new opportunities to use new AI advances to accelerate our core business that should have strong ROI over the next few years. So I think we should invest more there.

 The development of Llama 4 is progressing well; Llama 4 is being trained on more than 100,000 H100s and it’s the biggest training cluster in the world management is aware of; management expects the smaller Llama 4 models to be ready in early-2025; management thinks Llama 4 will be much faster and will have new modalities, stronger capabilities and reasoning

I’m even more excited about Llama 4, which is now well into its development. We’re training the Llama 4 models on a cluster that is bigger than 100,000 H100s or bigger than anything that I’ve seen reported for what others are doing. I expect that the smaller Llama 4 models will be ready first, and they’ll be ready — we expect sometime early next year. And I think that there are going to be a big deal on several fronts, new modalities, capabilities, stronger reasoning and much faster. 

Meta’s management remains convinced that open source is the way to go for AI development; the more developers use Llama, the more Llama improves in both quality and efficiency; in terms of efficiency, with higher adoption of Llama, management is seeing NVIDIA and AMD optimise their chip designs to better run Llama

It seems pretty clear to me that open source will be the most cost-effective, customizable, trustworthy performance and easiest to use option that is available to developers. And I am proud that Llama is leading the way on this…

…[Question] You said something along the lines of the more standardized Llama becomes the more improvements will flow back to the core meta business. And I guess, could you just dig in a little bit more on that?

[Answer] The improvements to Llama, I’d say come in a couple of flavors. There’s sort of the quality flavor and the efficiency flavor. There are a lot of researchers and independent developers who do work and because Llama is available, they do the work on Llama and they make improvements and then they publish it and it becomes — it’s very easy for us to then incorporate that both back into Llama and into our Meta products like Meta AI or AI Studio or Business AIs because the work — the examples that are being shown are people doing it on our stack.

Perhaps more importantly, is just the efficiency and cost. I mean this stuff is obviously very expensive. When someone figures out a way to run this better if that — if they can run it 20% more effectively, then that will save us a huge amount of money. And that was sort of the experience that we had with open compute and why — part of why we are leaning so much into open source here in the first place, is that we found counterintuitively with open compute that by publishing and sharing the architectures and designs that we had for our compute, the industry standardized around it a bit more. We got some suggestions also that helped us save costs and that just ended up being really valuable for us. Here, one of the big costs is chips — a lot of the infrastructure there. What we’re seeing is that as Llama gets adopted more, you’re seeing folks like NVIDIA and AMD optimize their chips more to run Llama specifically well, which clearly benefits us. 

Meta’s management expects to continue investing seriously into AI infrastructure

Our AI investments continue to require serious infrastructure, and I expect to continue investing significantly there too. We haven’t decided on the final budget yet, but those are some of the directional trends that I’m seeing.

Meta’s management thinks the integration of Meta AI into the Meta Ray-Ban glasses is what truly makes the glasses special; the Meta Ray-Ban glasses can answer questions throughout the day, help wearers remember things, give suggestions to wearers in real-time using multi-modal AI, and translate languages directly into the ear of wearers; management continues to think glasses are the ideal form-factor for AI because glasses lets AI see what you see and hear what you hear; demand for the Meta Ray-Ban glasses continues to be really strong; a recent release of the glasses was sold out almost immediately; Meta has deepened its partnership with EssilorLuxottica to build future generations of the glasses; Meta recently showcased Orion, its first full holographic AR glasses

This quarter, we also had several milestones around Reality Labs and the integration of AI and wearables. Ray-Ban meta glasses are the prime example here. They’re great booking glasses that let you take photos and videos, listen to music and take calls. But what makes them really special is the Meta AI integration. With our new updates, it will be able to not only answer your questions throughout the day, but also help you remember things, give you suggestions as you’re doing things using real-time multi-modal AI and even translate other languages right in your ear for you. I continue to think that glasses are the ideal form factor for AI because you can let your AI see what you see, hear what you hear and talk to you.

Demand for the glasses continues to be very strong. The new clear addition that we released at Connect sold out almost immediately and has been trading online for over $1,000. We’ve deepened our partnership with EssilorLuxottica to build future generations of smart eyewear that deliver both cutting-edge technology and style.

At Connect, we also showed Orion, our first full holographic AR glasses. We’ve been working on this one for about a decade, and it gives you a sense of where this is all going. We’re not too far off from being able to deliver great-looking glasses to let you seamlessly blend the physical and digital worlds so you can feel present with anyone no matter where they are. And we’re starting to see the next computing platform come together and it’s pretty exciting.

Newer scaling laws seen with Meta’s large language models inspired management to develop new ranking model architectures that can learn more effectively from significantly larger data sets; the new ranking model architectures have been deployed to Facebook’s video ranking models, helping to deliver more relevant recommendations; management is exploring the use of the new ranking model architectures on other services and the introduction of cross-surface data to the models, with the view that these moves will unlock more relevant recommendations and lead to better engineering efficiency

Previously, we operated separate ranking and recommendation systems for each of our products because we found that performance did not scale if we expanded the model size and compute power beyond a certain point. However, inspired by the scaling laws we were observing with our large language models, last year, we developed new ranking model architectures capable of learning more effectively from significantly larger data sets.

To start, we have been deploying these new architectures to our Facebook ranking video ranking models, which has enabled us to deliver more relevant recommendations and unlock meaningful gains in launch time. Now we’re exploring whether these new models can unlock similar improvements to recommendations on other services. After that, we will look to introduce cross-surface data to these models, so our systems can learn from what is interesting to someone on one surface of our apps and use it to improve their recommendations on another. This will take time to execute and there are other explorations that we will pursue in parallel. However, over time, we are optimistic that this will unlock more relevant recommendations while also leading to higher engineering efficiency as we operate a smaller number of recommendations.

Meta’s management is using new approaches to AI modelling to allow Meta’s ad systems to consider a person’s sequence of actions before and after seeing an ad, which allow the systems to better predict a person’s response to specific ads; the new approaches to AI modelling have delivered a 2%-4% increase in conversions in tests; Meta is seeing strong user-retention with its generative AI tools for image expansion, background generation, and text generation; Meta has started testing its first generative AI tools for video expansion and image animation and plans to roll them out broadly by early-2025

The second part of improving monetization efficiency is enhancing marketing performance. Similar to organic content ranking, we are finding opportunities to achieve meaningful ads performance gains by adopting new approaches to modeling. For example, we recently deployed new learning and modeling techniques that enable our ad systems to consider the sequence of actions a person takes before and after seeing an ad. Previously, our ad system could only aggregate those actions together without mapping the sequence. This new approach allows our systems to better anticipate how audiences will respond to specific ads. Since we adopted the new models in the first half of this year, we’ve already seen a 2% to 4% increase in conversions based on testing within selected segments…

…Finally, there is continued momentum with our Advantage+ solutions, including our ad creative tools. We’re seeing strong retention with advertisers using our Generative AI-powered image expansion, background generation and text generation tools, and they’re already driving improved performance for advertisers even at this early stage. Earlier this month, we began testing our first video generation features, video expansion and image animation. We expect to make them more broadly available by early next year.

Meta’s management expects to significantly increase Meta’s infrastructure for generative AI while prioritising fungibility

Given the lead time of our longer-term investments, we also continue to maximize our flexibility so that we can react to market developments. Within Reality Labs, this has benefited us as we’ve evolved our road map to respond to the earlier-than-expected success of smart glasses. Within Generative AI, we expect significantly scaling up our infrastructure capacity now while also prioritizing its fungibility will similarly position us well to respond to how the technology and market develop in the years ahead.

Meta’s management continues to develop tools for individuals and businesses to create AI agents easily; management thinks that Meta’s progress with AI agent tools is currently at where Meta was with Meta AI a year ago; management wants the AI agent tools to be widely used in 2025

There are also other new products like that, things around AI Studio. This year, we really focused on rolling out Meta AI as kind of our are kind of single assistant that people can ask any question to, but I think there’s a lot of opportunities that I think we’ll see ramp more over the next year in terms of both consumer and business use cases, for people interacting with a wide variety of different AI agents, consumer ones with AI Studio around whether it’s different creators or kind of different agents that people create for entertainment. Or on the business side, we do want to continue making progress on this vision of making it set any small business or any business over time can with a few clicks stand up in AI agent that can help do customer service and sell things to all of their customers around the world, and I think that’s a huge opportunity. So it’s very broad…

…But I’d say that we’re — today, with AI Studio and business AIs about where we were with Meta AI about a year ago. So I think in the next year, our goal around that is going to be to try to make those pretty widespread use cases, even though there’s going to be a multiyear path to getting kind of the depth of usage and the business results around that we want. 

Meta’s management is not currently sharing quantitative metrics on productivity improvements with the internal use of AI, but management is excited about the internal adoption they are seeing and the future opportunities for doing so

On the use of AI and employee productivity, it’s certainly something that we’re very excited about. I don’t know that we have anything particularly quantitative that we’re sharing right now. I think there are different efficiency opportunities with AI that we’ve been focused on in terms of where we can reduce costs over time and generate savings through increasing internal productivity in areas like coding. For example, it’s early, but we’re seeing a lot of adoption internally of our internal assistant and coding agent, and we continue to make Llama more effective at coding, which should also make this use case increasingly valuable to developers over time.

There are also places where we hope over time that we’ll be able to deploy these tools against a lot of our content moderation efforts to help make the big body of content moderation work that we undertake, to help it make it more efficient and effective for us to do so. And there are lots of other places around the company where I would say we’re relatively early in exploring the way that we can use LLM based tools to make different types of work streams more efficient.

It appears that Meta has achieved more than management expected in terms of developing its own AI infrastructure (in other words, developing its own AI chips)

So I think part of what we’re seeing this year is the infra team is executing quite well. And I think that’s, why over the course of the year, we’ve been able to build out more capacity. I mean going into the year, we had a range for what we thought we could potentially do. And we have been able to do, I think, more than, I think, we’d kind of hoped and expected at the beginning of the year. And while that reflects as higher expenses, it’s actually something that I’m quite happy that the team is executing well on. And I think that will — so that execution makes me somewhat more optimistic that we’re going to be able to keep on building this out at a good pace but that’s part of the whole thing. 

Meta’s management is starting to test the addition of AI-generated or AI-augmented content to users of Instagram and Facebook; management has high confidence that AI-generated and/or AI-augmented content will be an important trend in the future

I think we’re going to add a whole new category of content, which is AI generated or AI summarized content or kind of existing content pulled together by AI in some way. And I think that, that’s going to be just very exciting for the — for Facebook and Instagram and maybe Threads or other kind of feed experiences over time. It’s something that we’re starting to test different things around this. I don’t know if we know exactly what’s going to work really well yet. Some things are promising. I don’t know that this isn’t going to be a big impact on the business in ’25 would be my guess. But I think that there is I have high confidence that over the next several years, this is going to be an important trend and one of the important applications.

Meta’s management is currently focused on the engagement and user-experience of Meta AI; the monetisation of Meta AI will come later

Right now, we’re really focused on making Meta AI as engaging and valuable a consumer experience as possible. Over time, we think there will be a broadening set of queries that people use it for. And I think that the monetization opportunities will exist when over time as we get there. But right now, I would say we are really focused on the consumer experience above all and this is sort of a playbook for us with products that we put out in the world where we really dial in the consumer experience before we focus on what the monetization could look like.

Microsoft (NASDAQ: MSFT)

Microsoft’s AI business is on track to exceed $10 billion in annual revenue run rate in 2024 Q4 after being started for just 2.5 years; it will be the fastest business in the company’s history to do so; Microsoft’s AI business is nearly all inference (see Point 32 for more)

All up, our AI business is on track to surpass an annual revenue run rate of $10 billion next quarter, which will make it the fastest business in our history to reach this milestone…

…We’re excited that only 2.5 years in, our AI business is on track to surpass $10 billion of annual revenue run rate in Q2…

…If you sort of think about the point we even made that this is going to be the fastest growth to $10 billion of any business in our history, it’s all inference, right? 

Azure took share in 2024 Q3 (FY2025 Q1), driven by AI; Azure grew revenue by 33% in 2024 Q3 (was 29% in 2024 Q2), with 12 points of growth from AI services (was 8 points in 2024 Q2); Azure’s AI business has higher demand than available capacity

Azure took share this quarter…. 

… Azure and other cloud services revenue grew 33% and 34% in constant currency, with healthy consumption trends that were in line with expectations. The better-than-expected result was due to the small benefit from in-period revenue recognition noted earlier. Azure growth included roughly 12 points from AI services similar to last quarter. Demand continues to be higher than our available capacity. 

Microsoft’s management thinks Azure offers the broadest selection of AI chips, from Microsoft’s own Maia 100 chip to AMD and NVIDIA’s latest GPUs; Azure is the first cloud provider to offer NVIDIA’s GB200 chips

We are building out our next-generation AI infrastructure, innovating across the full stack to optimize our fleet for AI workloads. We offer the broadest selection of AI accelerators, including our first-party accelerator, Maia 100 as well as the latest GPUs from AMD and NVIDIA. In fact, we are the first cloud to bring up NVIDIA’s Blackwell system with GB200-powered AI servers.

Azure OpenAI usage more than doubled in the past 6 months, as both startups and enterprises move apps from test to production; GE Aerospace used Azure OpenAI to build a digital assistant for its 52,000 employees and in 3 months, the assistant has processed 500,000 internal queries and 200,000 documents; Azure recently added support for OpenAI’s newest o1 family of AI models; Azure AI is offering industry-specific models, including multi-modal models for medical imaging; Azure AI is increasingly an on-ramp for Azure’s data and analytics services, driving acceleration of Azure Cosmos DB and Azure SQL DB hyperscale usage

More broadly with Azure AI, we are building an end-to-end app platform to help customers build their own copilots and agents. Azure OpenAI usage more than doubled over the past 6 months as both digital natives like Grammarly and Harvey as well as established enterprises like Bajaj Finance, Hitachi, KT and LG move apps from test to production. GE Aerospace, for example, used Azure OpenAI to build a new digital assistant for all 52,000 of its employees. In just 3 months, it has been used to conduct over 500,000 internal queries and process more than 200,000 documents…

…This quarter, we added support for OpenAI’s newest model family, o1. We’re also bringing industry-specific models through Azure AI, including a collection of best-in-class multimodal models for medical imaging…

…Azure AI is also increasingly an on-ramp to our data and analytics services. As developers build new AI apps on Azure, we have seen an acceleration of Azure Cosmos DB and Azure SQL DB hyperscale usage as customers like Air India, Novo Nordisk, Telefonica, Toyota Motor North America and Uniper take advantage of capabilities purpose built for AI applications. 

Azure is offering its full catalog of AI models directly within the GitHub developer workflow; GitHub Copilot enterprise customers grew 55% sequentially in 2024 Q3; GitHub Copilot now has agentic workflows, such as Copilot Autofix, which helps users fix code 3x faster than it would take them on their own

And with the GitHub models, we now provide access to our full model catalog directly within the GitHub developer workflow…

… GitHub Copilot is changing the way the world builds software. Copilot enterprise customers increased 55% quarter-over-quarter as companies like AMD and Flutter Entertainment tailor Copilot to their own code base. And we are introducing the next phase of AI code generation, making GitHub Copilot agentic across the developer workflow. GitHub Copilot Workspace is a developer environment, which leverages agents from start to finish so developers can go from spec to plan to code all in natural language. Copilot Autofix is an AI agent that helps developers at companies like Asurion and Auto Group fix vulnerabilities in their code over 3x faster than it would take them on their own. We’re also continuing to build on GitHub’s open platform ethos by making more models available via GitHub Copilot. And we are expanding the reach of GitHub to a new segment of developers introducing GitHub Spark, which enables anyone to build apps in natural language.

Microsoft 365 Copilot has a new Pages feature, which management thinks is the first new digital artefact for the AI age; Pages helps users brainstorm with AI and collaborate with other users; Microsoft 365 Copilot responses are now 2x faster and 3x better; daily users of Microsoft 365 have more than doubled sequentially; Microsoft 365 copilot saves Vodafone employees 3 hours per person per week, and will be rolled out to 68,000 employees; 70% of the Fortune 500 now use Microsoft 365 Copilot; Microsoft 365 copilot is being adopted at a faster rate than any other new Microsoft 365 feature; with Copilot Studio, organisations can build autonomous agents to connect with Microsoft 365 Copilot; more than 10,000 organisations have used Copilot Studio, up 2x sequentially; monthly active users of Copilot across Microsoft’s CRM and ERP portfolio grew 60% sequentially

We launched the next wave of Microsoft 365 Copilot innovation last month, bringing together web, work, and Pages as the new design system for knowledge work. Pages is the first new digital artifact for the AI age, and it’s designed to help you ideate with AI and collaborate with other people. We’ve also made Microsoft 365 Copilot responses 2x faster and improved response quality by nearly 3x. This innovation is driving accelerated usage, and the number of people using Microsoft 365 daily more than doubled quarter-over-quarter. We are also seeing increased adoption from customers in every industry as they use Microsoft 365 Copilot to drive real business value. Vodafone, for example, will roll out Microsoft 365 Copilot to 68,000 employees after a trial showed that, on average, they save 3 hours per person per week. And UBS will deploy 50,000 seats in our largest finserve deal to date. And we continue to see enterprise customers coming back to buy more seats. All up, nearly 70% of the Fortune 500 now use Microsoft 365 Copilot, and customers continue to adopt it at a faster rate than any other new Microsoft 365 suite…

…With Copilot Studio, organizations can build and connect Microsoft 365 Copilot to autonomous agents, which then delegate to Copilot when there is an exception. More than 100,000 organizations from Nsure, Standard Bank and Thomson Reuters to Virgin Money and Zurich Insurance have used Copilot Studio to date, up over 2x quarter-over-quarter…

…Monthly active users of Copilot across our CRM and ERP portfolio increased over 60% quarter-over-quarter. 

Azure is bringing AI to industry-specific workflows; DAX Copilot is used in over 500 healthcare organisations to document more than 1.3 million physician-patient encounters each month; DAX Copilot is growing revenue faster than GitHub Copilot did in its first year

We’re also bringing AI to industry-specific workflows. One year in, DAX Copilot is now documenting over 1.3 million physician-patient encounters each month at over 500 health care organizations like Baptist Medical Group, Baylor Scott & White, Greater Baltimore Medical Center, Novant Health and Overlake Medical Center. It is showing faster revenue growth than GitHub Copilot did in this first year. And new features extend DAX beyond notes, helping physicians automatically draft referrals, after-visit instructions and diagnostic evidence.

LinkedIn’s AI tools help hirers find qualified candidates faster, and hirers who use AI assistant messages see a 44% higher acceptance rate

LinkedIn’s first agent hiring assistant will help hirers find qualified candidates faster by tackling the most time-consuming task. Already hirers who use AI assistant messages see a 44% higher acceptance rate compared to those who don’t. And our hiring business continues to take share.

In September 2024, Microsoft introduced a new AI companion experience – powered by Copilot – that includes voice and vision capabilities, allowing users to browse and converse with Copilot simultaneously

With Copilot, we are seeing the first step towards creating a new AI companion for everyone with new Copilot experience we introduced earlier this month, includes a refreshed design and tone along with improved speed and fluency across the web and mobile. And it includes advanced capabilities like voice and vision that make it more delightful and useful and feel more natural. You can both browse and converse with Copilot simultaneously because Copilot sees what you see. 

Roughly half of Microsoft’s cloud and AI-related capex in 2024 Q3 (FY2025 Q1) are for long-lived assets that will support monetisation over the next 15 years and more, while the other half are for CPUs and GPUs; the capex spend for CPUs and GPUs are made based on demand signals; management will be looking at inference demand to govern the level of AI capex for training; management sees that growth in capex will eventually slow and revenue growth will increase, but how fast that happens will depend on the pace of adoption of AI; the capex that Microsoft has been committing is a sign of management’s commitment to grow together with OpenAI, and to grow Azure beyond OpenAI; Microsoft is currently not interested at all in selling GPUs for companies to train AI models and has turned such business away, and this gives management conviction about the company’s AI-related capex

Capital expenditures including finance leases were $20 billion, in line with expectations, and cash paid for PP&E was $14.9 billion. Roughly half of our cloud and AI-related spend continues to be for long-lived assets that will support monetization over the next 15 years and beyond. The remaining cloud and AI spend is primarily for servers, both CPUs and GPUs, to serve customers based on demand signals…

…The inference demand ultimately will govern how much we invest in training because that’s, I think, at the end of the day, you’re all subject to ultimately demand…

…I think in some ways, it’s helpful to go back to the cloud transition that we worked on over a decade ago, I think, in the early stages. And what you did see and you’ll see us do in the same time is you have to build to meet demand. Unlike the cloud transition, we’re doing it on a global basis in parallel as opposed to sequential given the nature of the demand. And then as long as we continue to see that demand grow, you’re right, the growth in CapEx will slow and the revenue growth will increase. And those 2 things, to your point, get closer and closer together over time. The pace of that entirely depends really on the pace of adoption…

…[Question] How does Microsoft manage the demands on CapEx from helping OpenAI with its scaling ambitions?

[Answer] I’m thrilled with their success and need for supply from Azure and infrastructure and really what it’s meant in terms of being able to also serve other customers for us. It’s important that we continue to invest capital to meet not only their demand signal and needs for compute but also from our broader customers. That’s partially why you’ve seen us committing the amount of capital we’ve seen over the past few quarters, is our commitment to both grow together and for us to continue to grow the Azure platform for customers beyond them…

…One of the things that may not be as evident is that we’re not actually selling raw GPUs for other people to train. In fact, that’s sort of a business we turn away because we have so much demand on inference that we are not taking what I would — in fact, there’s a huge adverse selection problem today where people — it’s just a bunch of tech companies still using VC money to buy a bunch of GPUs. We kind of really are not even participating in most of that because we are literally going to the real demand, which is in the enterprise space or our own products like GitHub Copilot or M365 Copilot. So I feel the quality of our revenue is also pretty superior in that context. And that’s what gives us even the conviction, to even Amy’s answers previously, about our capital spend, is if this was just all about sort of a bunch of people training large models and that was all we got, then that would be ultimately still waiting, to your point, for someone to actually have demand, which is real. And in our case, the good news here is we have a diversified portfolio. We’re seeing real demand across all of that portfolio.

Microsoft’s management continues to expect Azure’s growth to accelerate in FY2025 H2, driven by increase in AI capacity to meet growing demand

In H2, we still expect Azure growth to accelerate from H1 as our capital investments create an increase in available AI capacity to serve more of the growing demand.

Microsoft’s management thinks that the level of supply and demand for AI compute will match up in FY2025 H2

But I feel pretty good that going into the second half of even this fiscal year, that some of that supply/demand will match up…

…I do, as you heard, have confidence, as we get a good influx of supply across the second half of the year particularly on the AI side, that we’ll be better able to do some supply-demand matching and hence, while we’re talking about acceleration in the back half.

Microsoft’s management sees Microsoft’s partnership with OpenAI as having been super beneficial to both parties; Microsoft provides the infrastructure for OpenAI to innovate on models; Microsoft takes OpenAI’s models and innovates further, through post-training of the models, building smaller models, and building products on top of the models; management developed conviction on the OpenAI partnership after seeing products such as GitHub Copilot and DAX Copilot get built; management feels very good about Microsoft’s investment in OpenAI; Microsoft accounts for OpenAI’s financials under the equity method

The partnership for both sides, that’s OpenAI and Microsoft, has been super beneficial. After all, we were the — we effectively sponsored what is one of the most highest-valued private companies today when we invested in them and really took a bet on them and their innovation 4, 5 years ago. And that has led to great success for Microsoft. That’s led to great success for OpenAI. And we continue to build on it, right? So we serve them with world-class infrastructure on which they do their innovation in terms of models, on top of which we innovate on both the model layer with some of the post-training stuff we do as well as some of the small models we build and then, of course, all of the product innovation, right? One of the things that my own sort of conviction of OpenAI and what they were doing came about when I started seeing something like GitHub Copilot as a product get built or DAX Copilot get built or M365 Copilot get built…

… And the same also, I would say, we are investors. We feel very, very good about sort of our investment stake in OpenAI…

…  I would say, just a reminder, this is under the equity method, which means we just take our percentage of losses every quarter. And those losses, of course, are capped by the amount of investment we make in total, which we did talk about in the Q this quarter as being $13 billion. And so over time, that’s just the constraint, and it’s a bit of a mechanical entry. And so I don’t really think about managing that. That’s the investment and acceleration that OpenAI is making in themselves, and we take a percentage of that.

Microsoft’s management sees Copilot as the UI layer for humans to interact with AI; Copilot Studio is used to build AI agents to connect Copilot to other systems of the user’s choice; Copilot Studio can also be used to create autonomous AI agents but these AI agents are not fully autonomous because at some point, they will need to notify a human or require an input and that is where Copilot comes in again

The system we have built is Copilot, Copilot Studio, agents and autonomous agents. You should think of that as the spectrum of things, right? So ultimately, the way we think about how this all comes together is you need humans to be able to interface with AI. So the UI layer for AI is Copilot. You can then use Copilot Studio to extend Copilot. For example, you want to connect it to your CRM system, to your office system, to your HR system. You do that through Copilot Studio by building agents effectively.

You also build autonomous agents. So you can use even — that’s the announcement we made a couple of weeks ago, is you can even use Copilot Studio to build autonomous agents. Now these autonomous agents are working independently, but from time to time, they need to raise an exception, right? So autonomous agents are not fully autonomous because, at some point, they need to either notify someone or have someone input something. And when they need to do that, they need a UI layer, and that’s where, again, it’s Copilot.

So Copilot, Copilot agents built-in Copilot Studio, autonomous agents built in Copilot Studio, that’s the full system, we think, that comes together.

Netflix (NASDAQ: NFLX)

Within entertainment, Netflix’s management thinks the most important question for AI is whether it can help creators produce even better content; the ability of AI to reduce costs in content creation is of secondary importance

 Lots of hype, good and bad, about how AI is going to impact or transform the entertainment industry. I think that the history has been that entertainment and technology have worked hand-in-hand throughout the history of time. And it’s very important, I think, for creators to be very curious about what these new tools are and what they could do. But AI needs to pass a very important test. Actually, can it help make better shows and better films? That is the test and that’s what they got to figure out. But I’ve said this before and I will say it again. We benefit greatly from improving the quality of the movies and the shows much more so than we do from making them a little cheaper. So any tool that can go to enhance the quality, making them better is something that is going to actually help the industry a great deal.

Paycom Software (NYSE: PAYC)

Paycom’s management developed an AI agent internally for the company’s service team to help the team provide even better service; the AI agent improved Paycom’s immediate response rates by 25% without any additional human interaction; the AI agent was built in house; Paycom is using AI in other areas, such as in several existing and upcoming products

Internally, we developed and deployed an AI agent for our service team. This technology utilizes our own knowledge-based semantic search model and enables us to provide service to help our clients more quickly and consistently than ever before.The AI agent continually improves over time and is having an impact on helping our clients achieve even more value out of their relationship with Paycom. By utilizing our own AI agent, we were able to connect our clients to the right solution faster, improving our immediate response rates by 25% without any additional human interaction…

…[Question] Interesting to hear about using AI in the customer service organization. I’m curious if that’s technology that Paycom has built or if you’re using a third party.

[Answer] So that’s internal. We built it ourselves, and we’ve been using it. And so it gets better and better as we mentioned on the call. It’s sped up our process by 25% as far as being able to connect clients to the solution quicker, whether that be a configuration question, a tax question or what have you. And so that’s really been helpful to us, and it continues to do more and more from that perspective…

…[Question] A follow-up on the AI agent or the AI technology that you’re developing. Do you see an opportunity in the future to productize what you’re developing internally, maybe like in your — in future versions of your recruiting product or other products in your platform?

[Answer] I would say this isn’t the only area in which we’re using AI. We have it in several products that we both have released and will be releasing. And so there’s definitely opportunities to monetize AI. As far as this particular solution, it’s really helping us on the back end and helping our client as well. So I think we’re going to see results and benefits from that in other areas of efficiency across the board within our own organization.

Shopify (NASDAQ: SHOP)

Shopify recently enhanced Shopify Flow, a low-code workflow automation app, with a new admin API connector that provides an additional 304 new automation actions

Let’s start with Shopify Flow. A low-code workflow automation app that empowers merchants to build custom automations and help them run their businesses more efficiently. This includes a new automation trigger based on the merchant’s custom data and newly completed admin API connector that provides an additional 304 new actions to use in their automations. And as a result, Flow has become a much more powerful tool, enabling merchants to update products, process customer form submissions, edit orders and so much more.

The Shopify Inbox feature now uses AI to suggest personalised replies for merchants to respond to customer inquiries; half of merchants’ responses are now using the AI-suggested replies; fast customer response helps lift conversion rates for merchants; the replies feature may not seem like a big deal, but it actually helps free up a lot of time for merchants to focus on building products

Within Shopify Inbox, this product now uses AI to suggest replies based on each merchant’s unique store information making it super easy for merchants to respond quickly and accurately to customer inquiries. In fact, on average, merchants are using the Suggest Replies for about half of their responses, edited or not, showing just how effective this feature has become. Replying can quickly boost conversion rates, which means more sales for our merchants and in turn, for Shopify…

…I mentioned suggest replies in Shopify Inbox, which may not seem like a big deal, but it’s a huge deal because it means merchants can spend more of their time focused on the things that they need to be focused on like building our products.  

The Shop App has a new merchant-focused home feed that is powered by machine learning models to increase shopper engagement; the new home feed has led to an 18% increase in sessions where a buyer engaged with a recommendation; management thinks the combination of search with AI will make the search function on the Shop App a lot more relevant and personalised

This quarter, the Shop App launched a new merchant-focused home feed, showcasing the diversity and the richness of brands on Shop. The experience uses new machine learning models to help buyers keep up with the brands they love and discover new brands based on their preferences. These changes have already led to early success with an 18% increase in sessions where a buyer engaged with a recommendation…

…We also think Search and AI together makes the Shop search way more relevant, way more personalized. That is also very compelling.

Essentially every Shopify internal department is using AI to be more productive

Support engineering, sales, finance, just about every department internally is using AI in some way to get more efficient, more productive.

Shopify’s management thinks the integration of AI in search will change how consumers find merchants and products, but Shopify has helped merchants navigate many similar changes before, and Shopify will continue to help merchants navigate the AI-related changes

In terms of where consumers find merchants or find products, yes, AI and search is going to change. But to be clear, this entire flow and discovery process has been changing for many years. It’s the reason that you saw us integrate with places like YouTube or more recently, Roblox or TikTok or Instagram…

…You can rest assured that when consumers shift their buying preferences, their discovery preferences, their search preferences, and they’re looking for great products from great brands, Shopify will ensure that our merchants are able to do so. And that’s the reason even some of the more nuanced or some of the more — as you know, Shopify has an integration to Spotify. Why? Because some merchants that also have very large followings as a musician have massive followings on their artist profile, the fact that so you can now show Shopify products on your artist profile means for that particular segment of merchants, they can easily — they now have a new surface area in which to conduct business. And that’s the same thing when it comes to AI and search. 

Taiwan Semiconductor Manufacturing Company (NYSE: TSM)

TSMC’s management expects TMSC’s business in 2024 Q4 to be supported by strong AI-related demand; management sees very strong AI demand in 2024 H2, leading to higher capacity utilisation rate for TSMC’s leading-edge 3nm and 5nm process technologies; management now expects server AI processors to account for mid-teens percent of TSMC’s total revenue in 2024 (previous expectation was for low-teens percent)

Moving into fourth quarter. We expect our business to continue to be supported by strong demand for our leading-edge process technologies. We continue to observe extremely robust AI-related demand from our customers throughout the second half of 2024, leading to increasing overall capacity utilization rate for our leading-edge 3-nanometer and 5-nanometer process technologies…

…We now forecast the revenue contribution from server AI processors to more than triple this year and account for mid-teens percentage of our total revenue in 2024.

TSMC’s management defines server AI processors as GPUs, AI accelerators, and CPUs for training and inference

At TSMC, we defined server AI processor as GPUs, AI accelerators and CPUs performing training and inference functions and do not including — include networking, edge or on-device AI.

TSMC’s management thinks AI demand is real, based on TSMC’s own experience of using AI and machine learning in its operations; a 1% productivity gain for TSMC is equal to a tangible NT$1 billion return on investment (ROI); management thinks TSMC is not the only company that has benefitted from AI applications

Whether this AI demand is real or not, okay, and my judgment is real, we have talked to our customers all the time, including hyperscaler customers who are building their own chips, and almost every AI innovators is working with TSMC. And so we probably get the deepest and widest look of anyone in this industry. And why I say it’s real? Because we have our real experience. We have using the AI and machine learning in our fab and R&D operations. By using AI, we are able to create more value by driving greater productivity, efficiency, speed, qualities. And think about it, let me use, 1% productivity gain, that was almost equal to about TWD 1 billion to TSMC. And this is a tangible ROI benefit. And I believe we cannot be the only one company that have benefited from this AI application. So I believe a lot of companies right now are using AI and — for their own improving productivity, efficiency and everything.

TSMC’s management thinks AI demand is just at the beginning

[Question] Talk a little bit about what you think about the duration of this current semiconductor up-cycle? Do you think it will continue into the next couple of years? Or are we getting closer to the peak of the cycle?

[Answer] The demand is real and I believe it’s just the beginning of this demand, all right? So one of my key customers said, the demand right now is insane, that it’s just the beginning. It’s [ a form of scientific ] to be engineering, okay? And it will continue for many years.

When TSMC builds fabs to meet AI demand, management has a picture in mind of what the long-term demand picture looks like

[Question] Keen to understand how TSMC gets comfortable with customer demand for AI beyond 2025. And I ask this because it takes a couple of years before you can build a fab, so you need to be taking early — an early view on what does AI look like in 2026, 2027. So how are you specifically cooperating on long-term plans for capacity with these AI customers? And what commitments are these customers giving you?

[Answer]  let me say again that we did talk to a lot of our customers. Almost every AI innovator are working with us and that’s including the hyperscalers. So if you look at the long-term market — long-term structure and market demand profile, I think we have some picture in our mind and we make some judgment, of course, and we work with them on a rolling basis. So how we prepare our capacity, actually, just like Wendell said, we have a disciplined and [ a rollout ] system to plan the appropriate level of capacity. And that — to support our customers’ need, also to maximize our shareholders’ value. That’s what we’re always keeping our mind.

There’s more AI content that goes into the chips in PCs (personal computers) and smartphones; management expects the PC and smartphone business of TSMC to be healthy in the next few years because of AI-related applications 

The unit growth of PC and smartphone is still in the low single digit. But more importantly is the content. The content now we put more AI into that, they are cheap and so the silicon area increased faster than the unit growth. So again, I would like to say that for this PC and the smartphone business, not — is gradually increased and we expect it to be healthy in the next few years because of our AI-related applications.

Advanced packaging is currently a high single-digit percentage of TSMC’s revenue and management expects it to grow faster than TSMC’s overall business over the next 5 years; the margins of advanced packaging are improving, but it’s not at the corporate average level yet

Advanced packaging in the next several years, let’s say, 5 years, will be growing faster than the corporate average. This year, it accounts for about high single digit of our revenue. In terms of margins, yes, it is also improving. However, it’s still — it’s approaching corporate, but not there yet.

Demand for TSMC’s CoWoS (advanced packaging) continues to far exceed supply, even though TSMC has doubled CoWoS capacity compared to a year ago and will double it again

Let me share with you today’s situation is our customer’s demand far exceed our ability to supply. So even we work very hard and increase the capacity by about more than twice, more than 2x as of this year compared with last year and probably double again, but still not enough. And — but anyway, we are working very hard to meet the customers’ requirement.

Tencent (NASDAQ: TCEHY)

Tencent’s management is increasingly seeing tangible benefits from deploying AI across the company’s business; management wants to continue investing in AI; the most significant benefits are in content recommendation and targeting, which directly benefits Tencent’s business and advertising revenue; management also sees AI as a productivity tool, as Tencent’s Copilot is being used by Tencent’s software engineers frequently and is helping them generate efficiency gains; management is trying to incorporate AI into a lot of Tencent’s products, but they think it will take a few more quarters before real use cases show up

We are increasingly seeing a tangible benefit of deploying AI across our products and operations, including marketing services and cloud. And we’ll continue investing in AI technology, tools and solutions that assist users and partners…

…I think that the most significant one right now is actually around content recommendation and at targeting because the AI in — the AI engine in those two use cases are generating a significant amount of additional user time and at the same time, it’s generating a higher incremental targeting rate, response rate for our apps and both of them actually are direct benefits to the business and direct benefit to ad revenue. and both of the video accounts and our performance at revenue actually at scale…

… It’s actually a productivity tool that everybody is using on a frequent basis, for example, our Copilot is being used by our engineers across the board on a very frequent basis, and it’s actually generating efficiency gains for our business. and different businesses, a lot of our products are actually testing our Hunyuan and trying to incorporate AI into the — either the production process, right, so that they would gain efficiency or in the user experience use case so that it can actually make their user experience better. So I would say, right now, we are seeing more and more adoption among all our different products and services. It would take probably a few more quarters for us to see some real use cases at scale. 

Tencent’s management used the company’s foundation AI model, Tencent Hunyuan, to facilitate tagging and categorisation of content and advertising materials; Tencent also upgraded its machine learning platforms to deliver better advertising targeting; marketing services revenue from video accounts was up 60% year-on-year; Mini Programs marketing services revenue had robust growth; Tencent used large language models (LLMs) to improve the relevance of Weixin Search results, leading to higher commercial queries and click-through rates, and consequently, an increase in search revenue of more than 100%

Our Marketing Services revenue grew 17% year-on-year. Strength in games and e-commerce categories outweighed weakness in real estate and food and beverage. The Paris Olympics somewhat cushioned industry-wide weakness in brand ad revenue during the third quarter but this positive factor will be absent in the fourth quarter. We leveraged our foundation model, Tencent Hunyuan to facilitate tagging and categorization of content and ad materials. And we upgraded our machine learning platforms to deliver more accurate ad targeting.

By property, video accounts marketing services revenue increased over 60% year-on-year. As we systematically strengthen transaction capabilities in Weixin, advertisers increasingly utilize our marketing tools to boost their exposure and drive sales conversion. Mini Programs marketing services revenue grew robustly year-on-year as our Mini Games and Mini Dramas provided high-value rewarded video ad inventory and generated incremental closed-loop demand. And for Weixin Search, we utilized large language model capabilities to facilitate understanding of complex queries and content, enhancing the relevance of search results. Commercial queries increased and click-through rate improved, and our search revenue more than doubled year-on-year.

Tencent enjoyed swift year-on-year growth in GPU-focused cloud revenue and this revenue stream is now a teens percentage of Tencent’s infrastructure as a services revenue; Tencent has released Tencent Hunyuan Turbo, the new generation of its foundation AI model, which uses a heterogeneous mixture of experts architecture; compared to the previous generation, Hunyuan Turbo’s training and inference efficiency has doubled while its inference costs has halved; Hunyuan Turbo is ranked first for general capabilities among foundation AI models in China; Tencent has open-sourced Hunyuan models; management sees Tencent’s AI revenue being lesser than US cloud companies because China does not have a large enterprise, SaaS, and startup markets for AI services 

Our cloud revenue from GPUs primarily used for AI grew swiftly year-on-year and now represents a teens percentage of our infrastructure as a services revenue. We released Tencent Hunyuan Turbo, which utilizes a heterogeneous mixture of experts architecture, doubling our training and inference efficiency and halving inference cost versus its predecessor Hunyuan Pro. SuperCLUE ranked Hunyuan Turbo first for general capabilities among domestic peers. Last week, we made the Hunyuan large model and the Hunyuan 3D generation models available on an open-source basis. Our international cloud revenue increased significantly year-on-year. We leveraged domain expertise in areas such as games and live streaming and competitive pricing to win international customers…

…The IAS revenue is now in the teens generated by AI. But having said that, we think the amount of AI revenue is actually less than U.S. cloud companies. And the main reason is because, number one, China doesn’t really have a every big enterprise market. And if you look at the U.S., a lot of enterprises are actually sort of fitted in with AI and the — in testing out how AI can do for their business that they’re actually buying a lot of compute, which is not happening in China yet. There’s a very big SaaS ecosystem in the U.S., which everybody is actually trying to add AI to their functionality and thus charge the customers more. And that SaaS ecosystem is not really that vibrant in China. And thirdly, there are also fewer AI start-ups in China, which are actually buying a lot of compute. So as a result, the AI revenue in China on the cloud side is somewhat sort of at scale for us, but I think it will not be exploding like in the U.S. 

Tencent’s management does not want to embed commercial search results into the company’s AI chatbot, YongBao right now; the current focus for YongBao is on growing usage, not monetisation

[Question] Will you ramp up the Gen AI chatbot, would that eventually embed with the commercial sponsor answer as well?

[Answer] In terms of whether YongBao will embed commercial search results, the answer is no. for the current time, we’re focused on making YongBao be as appealing and attractive to users as it can be and we’re not focused on premature monetization.

Tencent’s management plans to invest in capex for AI, but the amount of investment will be small compared to the companies in the USA

If you look at CapEx, right, we believe we have a progressive CapEx plan, especially given that the development of a cloud business and the advent of AI, but at the same time, it’s measured compared to a lot of the U.S. companies. 

Tencent’s management sees the company’s advertising business being driven by 3 factors, namely consumer spending, Tencent’s ability to utilise AI to continue boosting click-through rates from currently low levels, and deployment of more inventory

In terms of the drivers for 2025, the overall macro environment would obviously be important accelerator or decelerator or neutral force for the aggregate advertising market. And that in turn will be a function primarily of consumer confidence. And consumer spending behavior. Now within that overall environment, our relative performance will be a function of, first of all, our advertising technology and our ability to utilize GPUs, utilize neural networks to continue boosting click-through rates from the current very low levels to higher levels that mechanically translates into more revenue. And then secondly, our deployment of specific inventories, in particular, video accounts, in particular, Weixin Search.

Tesla (NASDAQ: TSLA)

Tesla’s management released FSD v12.5 in 2024 Q3, which has increased data and training compute, and 5x increase in parameter count; Tesla also released Actually Smart Summon (your vehicle will autonomously drive to you in parking lots) and FSD for Cybertruck, which includes end-to-end neural nets for highway driving for the first time; version 13 of FSD is coming soon and it is expected to have a 5-6 fold improvement in miles between interventions compared to version 12.5; over the course of 2024, FSD’s improvement in miles between interventions has been at least 3 orders of magnitude; management expects FSD to become safer than human in 2025 Q2; Tesla vehicles on autopilot have 1 crash per 7 million miles, compared to 1 crash per 700,000 miles for the US average; Tesla has earned $236 million in revenue in 2024 Q3 from the release of FSD for Cybertruck and Actually Smart Summon

In Q3, we released the 12.5 series of FSD (Supervised)1 with improved safety and comfort thanks to increased data and training compute, a 5x increase in parameter count, and other architectural choices that we plan to continue scaling in Q4. We released Actually Smart Summon, which enables your vehicle to autonomously drive to you in parking lots, and FSD (Supervised) to Cybertruck customers, including end-to-end neural nets for highway driving for the first time…

…Version 13 of FSD is going out soon… We expect to see roughly a 5- or 6-fold improvement in miles between interventions compared to 12.5. And actually, looking at the year as whole, the improvement in miles between interventions, we think will be at least 3 orders of magnitude. So that’s a very dramatic improvement in the course of the year, and we expect that trend to continue next year.  The current total expectation, internal expectation for the Tesla FSD having longer miles between interventions [indecipherable] is the second quarter of next year, which means it may end up being in the third quarter but it’s next — it seems extremely likely to be next year…

…miles between critical interventions, mentioned by Elon already made 100x improvement with 12.5 from starting of this year and then with v13 release, we expect to be 1,000x from the beginning, from January of this year on production software. And this came in because of technology improvements going to end-to-end, having higher frame rate, partly also helped by hardware force, more capabilities, so on. And we hope that we continue to scale the neural network, the data, the training compute, et cetera. By Q2 next year, we should cross over the average, even in miles per critical intervention [indiscernible] in that case…

…Our internal estimate is Q2 of next year to be safer than human and then to continue with rapid improvements thereafter…

… So we published Q3 vehicle safety report, which shows 1 crash for every 7 million miles on autopilot that compares with the U.S. average of crash roughly every 700,000 miles. So it’s currently showing a 10x safety improvement relative to the U.S. average…

…We released FSD for Cybertruck and other features like actually small [indiscernible] like Elon talked about in North America, which contributed $326 million of revenues in the quarter. 

Tesla has deployed a 29,000 H100 cluster and expects to have a 50,000 H100 cluster by the end of October 2024, to support FSD and Optimus; Tesla is not training compute-constrained; Tesla’s AI has gotten so good that it now takes a long time to decide which version of the software is better because mistakes happen so infrequently and that is the big bottleneck to Tesla’s AI development; management is being very careful with AI-spending

We deployed and are training ahead of schedule on a 29k H100 cluster at Gigafactory Texas – where we expect to have 50k H100 capacity by the end of October…

…We continue to expand our AI training capacity to accommodate the needs of both FSD and Optimus. We are currently not training compute-constrained. [indiscernible] probably the big limiting factors of the FSD is actually getting so good that it takes us a while to actually find mistakes. And when you start getting to where it can take 10,000 miles to find a mistake, it takes a while to actually figure out which it is, is software A better than software B? It actually takes a while to figure it out because neither 1 of them makes the mistakes, would take a long time to make mistakes. So it’s actually the single biggest limiting factor is how long does it take us to figure out which version is better? Sort of a high-class problem…

… One thing which I’d like to elaborate is that we’re being really judicious on our AI compute spend to and saying how best we can utilize the existing infrastructure before making further investments…

…We still got to take which models are performing better. So the validation network to picking the models because as mentioned the miles between intervention is pretty large. We had to drive a lot of miles going close to. We do have simulation and other ways to get those metrics. Those 2 help, but in the end, that’s a big bottleneck. That’s why we’re not training-compete constrained alone. 

In the 10 October 2024 “We, Robot” event by Tesla, the company had showcased 50 autonomous vehicles, including 20 Cybercabs; the Cybercabs had no steering wheel, brake, or accelerator pedals, so they were truly autonomous

On October 10, we laid out a vision for an autonomous and future that I think is very compelling that the Tesla team did a phenomenal job there with actually giving people an option to experience the future, where you have humanoid robots working among the craft, not with a canned video and a presentation or anything but walking among crowd so he drinks and whatnot. And we had 50 autonomous vehicles. There were 20 Cybercabs but there were an additional 30 Model Ys, operating fully autonomously the entire night, carrying thousands of people with no incidents the entire night…

…Worth emphasizing that the Cybercab had no steering wheel or brake or accelerator panels, meaning there was no way for anyone to intervene manually a unit if they wanted to and the whole night went very smoothly.

Tesla is already offering autonomous ridehailing for Tesla employees in the Bay Area; the ridehailing service currently has a safety driver; Tesla has been testing autonomous ridehailing for some time; Elon Musk expects ridehailing to be rolled out to the public in California and Texas in 2025, and maybe other states in the USA; California has a lot of regulations around ridehailing, but there’s still a regulatory pathway; Tesla actually has passed Federal regulations for ridehailing, but it’s the state level where there are problems

We have for Tesla employees in the Bay Area, we already are offering ridehailing capabilities. So you can actually, with the development app, you can request a ride and it will take you anywhere in the Bay Area. We do have a safety driver for now but it’s not required to do that…

… We’ve been testing it for the good part of the year. And the building blocks that we needed in order to build this functionality and deliver it to production, we’ve been thinking about working on for years…

…So it’s not like we’re just starting to think about this stuff right now while we’re building out the early stages of our ridehailing network. We’ve been thinking about this for quite a long time, and we’re excited to get the functionality out there…

…We do expect to roll out ridehailing in California and Texas next year to the public. Now California is somewhere — there’s quite a long regulatory approval process. I think we should get approval next year but it’s contingent upon regulatory approval. Texas is a lot faster so it’s — we’ll definitely have available in Texas and probably have it available in California, subject to regulatory approval. And then — and maybe some other states actually next year as well, but at least California and Texas…

…[Question] Elon mentioned unsupervised FSD in California and Texas next year. Does that mean regulators have agreed to it in the entire state for existing hardware 3 and 4 vehicles?

[Answer] As I said earlier, California loves regulation… here’s a pathway. Obviously, Waymo operates in California so there’s just a lot of forms and a lot of approvals that are required. I mean, I’d be shocked if we don’t get approved next year, but it’s just not something we totally control. But I think we will get approval next year in California and Texas. And towards the Bay Area, branch out beyond California and Texas…

…I think it’s important to reiterate this like on our certifying a vehicle at the federal level in the U.S. is done by meeting FMVSS regulations. Our vehicles today that are produced there capable to meet all those regulations, the Cybercab regulations. And so the deployment of the vehicle to the road is no limitation, but its limitation is what you said at the state level where they control autonomous vehicle deployment. Some states are relatively easy, as you mentioned, for Texas. It’s other ones have always like California that may take a little longer. The other ones hadn’t set up anything yet. 

Tesla’s management acknowledges that there’s a chance that Tesla vehicles with Hardware Version 3 may not support unsupervised full self-dricing, and if so, Tesla will replace the hardware for those vehicle fleets for free into Hardware Version 4

By some measure, Hardware 4 has really several times the capability of Hardware 3. It’s easier to get things to work with then it takes a lot of effort to sort of squeeze that box analyst hat Hardware 3. And there is some chance that Hardware 3 is — does not achieve the safety level that allows for unsupervised FSD. There is some chance of that. And if that turns out to be the case, we will upgrade those group bought Hardware 3 FSD for free. And we have designed the system to be upgradeable so it’s really just to sort of switch out the computer thing, the camera, the cameras are capable. But we don’t actually know the answers of that. But if it does turn out, we’ll make sure we take care of those who are.

Tesla’s management thinks real-world AI in self-driving cars is different from LLMs (large language models) in that (1) real-world AI requires massive amounts of context that needs to be processed with a small amount of compute power and the way around this limitation is to do massive amounts of training so that the amount of inference that needs to be done is tiny, and (2) it’s difficult to sort out what data coming in from the video feed is important for the training

The nature of real world AI is different from LLM in that you have a massive amount of context. So like the — you’ve got a case of Tesla cameras that [indiscernible] if you include tunnel camera that — so you’ve got some context. And that is then distilled down into a small number of control outputs, whereas it’s like it’s very rare to have, in fact, I’m not sure any LLM out there can do gigabytes of context. And then you’ve got to then process that in the car with a very small amount of compute power. It’s all doable and it’s happening, but it is a different problem than what, say, a Gemini or OpenAI is doing.

And now part of the way you can make up for the fact that the inference computer is quite small, it is by spending a lot of effort on training. And just like a human the way you train on something, the less metal work takes when you try to — when you do it, like when the first time like a driving it absorbs your whole mind. But then as you train more and more on driving then the driving becomes a background task. It doesn’t — it only absorbs a small amount of your mental capacity because you have a lot of training. So we can make up for the fact that the inference computers — it’s tiny compared to a 10-kilowatt bank of GPUs because you’ve got a few hundred watts of inference compute. We can make up that with heavy training.

And then there’s also vast amounts to the actual petabytes of data coming in are tremendous. And then sorting out what training is important, of the vast amounts of video data coming in the feed, what is actually most important for training. That’s also quite difficult.

Tesla’s management thinks Elon Musk’s xAI AI-startup has been helpful to Tesla, but the 2 companies are focused on very different kinds of AI problems 

Well, I should say that xAI has been helpful to Tesla AI quite a few times in terms of things like scaling it, like training, just even like recently in the last week or so, improvements in training, where if you’re doing a big training run and it fails, being able to continue training and to recover from a training run, has been pretty helpful. But there are different problems. xAI actually is working on artificial general intelligence or artificial super intelligence. Tesla is autonomous cars and autonomous robots. There are different problems…

…Yes, Tesla is focused on real-world AI. And I was saying earlier, it is quite a bit different from LLM. But you have massive context in the form of video and some amount of audio, that’s going to be distilled like extremely efficient inference compute. I do think Tesla is the most efficient in the world in terms of inference compute because out of necessity, we have to be very good at efficient inference. We can’t put 10 kilowatts of GPUs in a car. We’ve got a couple of hundred watts. And it’s a pretty well designed Tesla AI chip, but it’s still a couple hundred watts. But there are different problems. I mean, the stuff at xAI. We’re running inference. I mean, it is running inference, answering questions on a 10-kilowatt rack. It’s like you can’t put that in a car. It’s a different problem.

Elon Musk created xAI because he thought there wasn’t a truth-seeking AI company being built

xAI is because I felt there wasn’t there wasn’t a truth-seeking digital super intelligence company out there, like that’s what it came down to. There needed to be a truth-seeking AI company that is very [indiscernible] about being truthful. I’m not saying xAI is perfect, but that is truth, but that is at least the explicit aspiration, even if something is politically incorrect, it would still be truhtful. I think this is very important for AI safety. So I think xAI has been helpful to Tesla and will continue to be helpful to Tesla, but they are very different problems.

There are no other car companies that has a world-class AI and chip-design team like Tesla

And like what other car company has a world-class chip design team? Like zero. What other car company has a world-class AI team like Tesla does? 0. Those were all startups that were created from scratch.

The Trade Desk (NASDAQ: TTD)

The incorporation of AI into Kokai, Trade Desk’s ad-buying platform, is encouraging adoption of Trade Desk by CFOs and CMOs

While there has been a lot of macro focus on the reduction in inflation rates, historic highs for stock market indices and growing indications of a soft landing, that’s not necessarily translating to consumer confidence, which is why CMOs are becoming much more closely aligned with their CFOs. CFOs want more evidence than ever that marketing is working. And for CFOs that doesn’t just mean traditional marketing KPIs. It means growing the top line business. All of our AI and data science injection into Kokai, our latest product release, is encouraging CMOs and CFOs to lean more and more on TTD to deliver real, measured growth…

…When CMOs faced pressure to achieve more with less, they turn to platforms like ours for flexibility, precision and measurable results.

Companies need an AI strategy, and Trade Desk’s AI product, Koa, is a great copilot for advertising traders; Trade Desk has plenty of opportunities in an AI-world because of the data assets it has, and management wants to improve all aspects of the company through AI

Every company needs an AI strategy. Our AI product, Koa, is a great copilot for traders. But this is only the beginning. There are endless possibilities for us as we have 1 of the best data assets on the Internet. The learnings that come from buying the global open Internet outside of walled gardens. To win in this new frontier, we’re looking across our entire suite of products, algorithms and features and asking how they all can be advanced by AI.

Visa (NYSE: V)

For Risk and Identity Solutions within value-added services, Visa wants to acquire Featurespace, an AI payments protection tech company that will enable Visa to enhance fraud prevention tools to clients and protect consumers in real time; Worldline, a Visa partner, will be using Decision Manager to provide businesses with AI-based e-commerce fraud detection abilities; Featurespace is a world leader in providing AI solutions to fight fraud

In Risk and Identity Solutions, we recently announced our intent to acquire Featurespace, a developer of real-time artificial intelligence payments protection technology. It will enable Visa to provide enhanced fraud prevention tools to our clients and protect consumers in real-time across various payment methods.  And Worldline, already a Visa partner and leading European acquirer, will soon be launching an optimized fraud management solution, utilizing Decision Manager to provide businesses with AI-based e-commerce fraud detection capabilities…

…Featurespace is a world leader in providing AI-driven solutions to combat that fraud, to reduce that fraud, to enable our clients and partners to continue to serve their customers in a safe way.

Visa’s management sees AI as being a driver of productivity across multiple functions in the company, and as a differentiator in its products and services

[Question] I just wanted to ask how you see AI playing into the business model. Do you see it more as driving VAS or incremental business model, uplift revenue or cost improvement? Or is it more of a competitive differentiator that will just keep you ahead of your competition?

[Answer] As it relates more broadly to especially generative AI at Visa, I see it really in 2 different buckets. The first is we are adopting it aggressively across our company to drive productivity. And we’ve seen some great results from everywhere to our engineering teams, to our accounting teams, to our sales teams, our client service teams. And we’re still in the early stages of, I think, the very significant impact this will have on the productivity of our business. I also see it as a real differentiator to the products and services that we’re putting in market. You’ve heard me talk about some of the new risk capabilities, risk management capabilities, for example, that we’ve deployed in the account-to-account space, which are all enabled with generative AI. You mentioned Featurespace. We’ve had some really good success in other parts of both our value-added services business and the broader consumer payments business as well. And we’ve got a product pipeline that is very heavily tilted towards some, we think, very exciting generative AI capabilities that hopefully you’ll hear more from us on soon.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. I have a vested interest in Alphabet, Amazon, Apple, ASML, Coupang, Datadog, Fiverr, Mastercard, Meta Platforms, Microsoft, Netflix, Shopify, TSMC, Tesla, and Visa. Holdings are subject to change at any time.