What We’re Reading (Week Ending 12 May 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 12 May 2024:

1. From Blueprint to Therapy: The Evolution and Challenges of Nucleic Acid Interventions – Biocompounding

Nucleic Acid Therapies (NATs) offer a targeted approach to rectify these underlying genetic issues. By employing strategies like antisense oligonucleotides, mRNA therapy, RNA interference, or CRISPR-based gene editing, NATs can directly modify or regulate the expression of genes responsible for the disease. These therapies can repair or silence defective genes, replace missing genes, or modulate gene expression, thereby addressing the root cause of the disease at the molecular level. This precision in targeting genetic defects makes NATs a promising and revolutionary approach in modern medicine, potentially offering cures or significant treatment improvements for numerous genetic and acquired diseases.

The modalities of NATs vary based on their mechanism of action, type of nucleic acid used, and therapeutic goals. Here’s an introduction to the different modalities of NATs:

  1. Antisense Oligonucleotides (ASOs): These are short, synthetic strands of DNA or RNA that are designed to bind to specific RNA molecules within a cell. By binding to their target RNA, ASOs can interfere with the process of protein production. They can inhibit the expression of a gene, modify RNA splicing, or promote the degradation of the RNA molecule. ASOs are used in conditions like Duchenne Muscular Dystrophy and Spinal Muscular Atrophy. Example: Sarepta Therapeutics
  2. RNA Interference (RNAi): This modality uses small interfering RNA (siRNA) or microRNA (miRNA) to silence specific genes. RNAi works by degrading the mRNA of a target gene, preventing it from being translated into a protein. This approach is particularly useful in diseases where inhibiting the expression of a certain gene can be therapeutic. RNAi has been explored for various applications including cancer therapy and viral infections. Currently FDA-approved siRNAs are used in conditions such as Hereditary transthyretin-mediated amyloidosis. Example: Alnylam Pharmaceuticals
  3. AAV Gene Therapy: Adeno-associated virus (AAV) vectors are commonly used in gene therapy. AAVs are small viruses that can deliver genetic material into cells without causing disease. In AAV gene therapy, the therapeutic gene is packaged into an AAV vector, which then delivers the gene into patient’s cells. This modality is useful for treating genetic disorders, such as Hemophilia A, by providing a functional copy of a defective or missing gene. Example: Spark Therapeutics
  4. mRNA Therapy: mRNA therapies involve the use of messenger RNA to produce therapeutic proteins inside the body. Unlike traditional gene therapy that alters the DNA within cells, mRNA therapy delivers mRNA that is translated into the desired protein, offering a temporary but effective treatment. This approach has gained significant attention, especially in the development of COVID-19 vaccines. Currently there are several attempts to develop cancer vaccines by the key players in this space. Example: Moderna, BioNTech, Pfizer
  5. CRISPR/Cas9 and Genome Editing: This revolutionary technology enables precise editing of the genome. CRISPR/Cas9 can be used to add, delete, or alter specific DNA sequences in the genome, offering the potential to correct genetic defects at their source. While still in the experimental stages for many applications, it holds promise for treating a range of genetic disorders. In Dec 2023, the first ever FDA-approved CRISPR-based gene therapy was used to treat sickle cell disease. Example: CRISPR Therapeutics, Vertex

2. China Is Still Rising – Nicholas Lardy

Those who doubt that China’s rise will continue point to the country’s weak household spending, its declining private investment, and its entrenched deflation. Sooner than overtake the United States, they argue, China would likely enter a long recession, perhaps even a lost decade.

But this dismissive view of the country underestimates the resilience of its economy. Yes, China faces several well documented headwinds, including a housing market slump, restrictions imposed by the United States on access to some advanced technologies, and a shrinking working-age population. But China overcame even greater challenges when it started on the path of economic reform in the late 1970s. While its growth has slowed in recent years, China is likely to expand at twice the rate of the United States in the years ahead.

Several misconceptions undergird the pessimism about China’s economic potential…

…A second misconception is that household income, spending, and consumer confidence in China is weak. The data do not support this view. Last year, real per capita income rose by 6 percent, more than double the growth rate in 2022, when the country was in lockdown, and per capita consumption climbed by nine percent. If consumer confidence were weak, households would curtail consumption, building up their savings instead. But Chinese households did just the opposite last year: consumption grew more than income, which is possible only if households reduced the share of their income going to savings…

…Another misconception concerns the potential for a collapse in property investment. These fears are not entirely misplaced; they are supported by data on housing starts, the number of new buildings on which construction has begun, which in 2023 was half what it was in 2021. But one has to look at the context. In that same two-year period, real estate investment fell by only 20 percent, as developers allocated a greater share of such outlays to completing housing projects they had started in earlier years. Completions expanded to 7.8 billion square feet in 2023, eclipsing housing starts for the first time. It helped that government policy encouraged banks to lend specifically to housing projects that were almost finished; a general easing of such constraints on bank loans to property developers would have compounded the property glut…

…By 2014, private investment composed almost 60 percent of all investment—up from virtually zero percent in 1978. As private investment is generally more productive than that of state companies, its expanding share of total investment was critical to China’s rapid growth over this period. This trend went into reverse after 2014 when Xi Jinping, having just assumed the top leadership position, aggressively redirected resources to the state sector. The slowdown was modest at first, but by 2023, private investment accounted for only 50 percent of total investment…

…But here again, the pessimism is not supported by the data. First, almost all the decline in the private share of total investment after 2014 resulted from a correction in the property market, which is dominated by private companies. When real estate is excluded, private investment rose by almost ten percent in 2023. Although some prominent Chinese entrepreneurs have left the country, more than 30 million private companies remain and continue to invest.

3. The Cloud Under The Sea – Josh Dzieza

In the family tree of professions, submarine cable work occupies a lonely branch somewhere between heavy construction and neurosurgery. It’s precision engineering on a shifting sea using heavy metal hooks and high-tension lines that, if they snap, can cut a person in half. In Hirai’s three decades with Kokusai Cable Ship Company (KCS), he had learned that every step must be followed, no matter how chaotic the situation. Above all else, he often said, “you must always be cool.”…

…The world’s emails, TikToks, classified memos, bank transfers, satellite surveillance, and FaceTime calls travel on cables that are about as thin as a garden hose. There are about 800,000 miles of these skinny tubes crisscrossing the Earth’s oceans, representing nearly 600 different systems, according to the industry tracking organization TeleGeography. The cables are buried near shore, but for the vast majority of their length, they just sit amid the gray ooze and alien creatures of the ocean floor, the hair-thin strands of glass at their center glowing with lasers encoding the world’s data.

If, hypothetically, all these cables were to simultaneously break, modern civilization would cease to function. The financial system would immediately freeze. Currency trading would stop; stock exchanges would close. Banks and governments would be unable to move funds between countries because the Swift and US interbank systems both rely on submarine cables to settle over $10 trillion in transactions each day. In large swaths of the world, people would discover their credit cards no longer worked and ATMs would dispense no cash. As US Federal Reserve staff director Steve Malphrus said at a 2009 cable security conference, “When communications networks go down, the financial services sector does not grind to a halt. It snaps to a halt.”…

…Fortunately, there is enough redundancy in the world’s cables to make it nearly impossible for a well-connected country to be cut off, but cable breaks do happen. On average, they happen every other day, about 200 times a year. The reason websites continue to load, bank transfers go through, and civilization persists is because of the thousand or so people living aboard 20-some ships stationed around the world, who race to fix each cable as soon as it breaks.

The industry responsible for this crucial work traces its origins back far beyond the internet, past even the telephone, to the early days of telegraphy. It’s invisible, underappreciated, analog. Few people set out to join the profession, mostly because few people know it exists…

…Once people are in, they tend to stay. For some, it’s the adventure — repairing cables in the churning currents of the Congo Canyon, enduring hull-denting North Atlantic storms. Others find a sense of purpose in maintaining the infrastructure on which society depends, even if most people’s response when they hear about their job is, But isn’t the internet all satellites by now? The sheer scale of the work can be thrilling, too. People will sometimes note that these are the largest construction projects humanity has ever built or sum up a decades-long resume by saying they’ve laid enough cable to circle the planet six times…

…The world is in the midst of a cable boom, with multiple new transoceanic lines announced every year. But there is growing concern that the industry responsible for maintaining these cables is running perilously lean. There are 77 cable ships in the world, according to data supplied by SubTel Forum, but most are focused on the more profitable work of laying new systems. Only 22 are designated for repair, and it’s an aging and eclectic fleet. Often, maintenance is their second act. Some, like Alcatel’s Ile de Molene, are converted tugs. Others, like Global Marine’s Wave Sentinel, were once ferries. Global Marine recently told Data Centre Dynamics that it’s trying to extend the life of its ships to 40 years, citing a lack of money. One out of 4 repair ships have already passed that milestone. The design life for bulk carriers and oil tankers, by contrast, is 20 years.

“We’re all happy to spend billions to build new cables, but we’re not really thinking about how we’re going to look after them,” said Mike Constable, the former CEO of Huawei Marine Networks, who gave a presentation on the state of the maintenance fleet at an industry event in Singapore last year. “If you talk to the ship operators, they say it’s not sustainable anymore.”

He pointed to a case last year when four of Vietnam’s five subsea cables went down, slowing the internet to a crawl. The cables hadn’t fallen victim to some catastrophic event. It was just the usual entropy of fishing, shipping, and technical failure. But with nearby ships already busy on other repairs, the cables didn’t get fixed for six months. (One promptly broke again.)

But perhaps a greater threat to the industry’s long-term survival is that the people, like the ships, are getting old. In a profession learned almost entirely on the job, people take longer to train than ships to build.

“One of the biggest problems we have in this industry is attracting new people to it,” said Constable. He recalled another panel he was on in Singapore meant to introduce university students to the industry. “The audience was probably about 10 university kids and 60 old gray people from the industry just filling out their day,” he said. When he speaks with students looking to get into tech, he tries to convince them that subsea cables are also part — a foundational part — of the tech industry…

…To the extent he is remembered, Cyrus Field is known to history as the person responsible for running a telegraph cable across the Atlantic Ocean, but he also conducted what at the time was considered an equally great technical feat: the first deep-sea cable repair.

Field, a 35-year-old self-made paper tycoon, had no experience in telegraphy — which helps explain why, in 1854, he embarked on such a quixotic mission…

…“When it was first proposed to drag the bottom of the Atlantic for a cable lost in waters two and a half miles deep, the project was so daring that it seemed to be almost a war of the Titans upon the gods,” wrote Cyrus’ brother Henry. “Yet never was anything undertaken less in the spirit of reckless desperation. The cable was recovered as a city is taken by siege — by slow approaches, and the sure and inevitable result of mathematical calculation.”

Field’s crew caught the cable on the first try and nearly had it aboard when the rope snapped and slipped back into the sea. After 28 more failed attempts, they caught it again. When they brought it aboard and found it still worked, the crew fired rockets in celebration. Field withdrew to his cabin, locked the door, and wept.

Cable repair today works more or less the same as in Field’s day. There have been some refinements: ships now hold steady using automated dynamic positioning systems rather than churning paddle wheels in opposite directions, and Field’s pronged anchor has spawned a medieval-looking arsenal of grapnels — long chains called “rennies,” diamond-shaped “flat fish,” spring-loaded six-blade “son of sammys,” three-ton detrenchers with seven-foot blades for digging through marine muck — but at its core, cable repair is still a matter of a ship dragging a big hook along the ocean floor. Newfangled technologies like remotely operated submersibles can be useful in shallow water, but beyond 8,000 feet or so, conditions are so punishing that simple is best…

…Debates about the future of cable repair have become a staple of industry events. They typically begin with a few key facts: the ships are aging; the people are aging; and it’s unclear where the money will come from to turn things around.

For much of the 20th century, cable maintenance wasn’t a distinct business; it was just something giant, vertically integrated telecom monopolies had to do in order to function. As they started laying coaxial cables in the 1950s, they decided to pool resources. Rather than each company having its own repair vessel mostly sitting idle, they divided the oceans into zones, each with a few designated repair ships.

When the telcos were split up at the turn of the century, their marine divisions were sold off. Cable & Wireless Marine became Global Marine. AT&T’s division is now the New Jersey-based SubCom. (Both are now owned by private equity companies; KCS remains a subsidiary of KDDI.) The zone system continued, now governed by contracts between cable owners and ship operators. Cable owners can sign up with a nonprofit cooperative, like the Atlantic Cable Maintenance & Repair Agreement, and pay an annual fee plus a day rate for repairs. In exchange, the zone’s three ships — a Global Marine vessel in Portland, UK, another in Curaçao, and an Orange Marine vessel in Brest, France — will stand ready to sail out within 24 hours of being notified of a fault.

This system has been able to cope with the day-to-day cadence of cable breaks, but margins are thin and contracts are short-term, making it difficult to convince investors to spend $100 million on a new vessel.

“The main issue for me in the industry has to do with hyperscalers coming in and saying we need to reduce costs every year,” said Wilkie, the chair of the ACMA, using the industry term for tech giants like Google and Meta. “We’d all like to have maintenance cheaper, but the cost of running a ship doesn’t actually change much from year to year. It goes up, actually. So there has been a severe lack of investment in new ships.”

At the same time, there are more cables to repair than ever, also partly a result of the tech giants entering the industry. Starting around 2016, tech companies that previously purchased bandwidth from telcos began pouring billions of dollars into cable systems of their own, seeking to ensure their cloud services were always available and content libraries synced. The result has been not just a boom in new cables but a change in the topology of the internet. “In the old days we connected population centers,” said Constable, the former Huawei Marine executive. “Now we connect data centers. Eighty percent of traffic crossing the Atlantic is probably machines talking to machines.”…

…In 2022, the industry organization SubOptic gathered six cable employees in their 20s and 30s for a panel on the future of the industry. Most of them had stumbled into their jobs inadvertently after college, and the consensus was that the industry needed to be much better about raising public awareness, especially among the young.

“I don’t know if anyone saw, but during the pandemic, submarine cables actually went viral on TikTok,” said one panelist, a young cable engineer from Vodafone. “People didn’t know they existed, and then suddenly, out of nowhere, they were viral. I think it’s engaging with youth and children through their own avenues — yes, you can have science museums and things like that, but they are online, they are on their iPads, they’re on their phones.”

“We’ve got some pretty senior decision-makers and influencers in the subsea cable industry here,” said one audience member. “Did any of us know that we went viral on TikTok?” he asked, to laughter.

“As this panel rightfully said upfront, it’s not that we have a brand problem,” said another audience member, “we just don’t have a brand at all.”

4. Looking for AI use-cases – Benedict Evans

I’ve been thinking about this problem a lot in the last 18 months, as I’ve experimented with ChatGPT, Gemini, Claude and all the other chatbots that have sprouted up: ‘this is amazing, but I don’t have that use-case’.

The one really big use-case that took off in 2023 was writing code, but I don’t write code. People use it for brainstorming, and making lists and sorting ideas, but again, I don’t do that. I don’t have homework anymore. I see people using it to get a generic first draft, and designers making concept roughs with MidJourney, but, again, these are not my use-cases. I have not, yet, found anything that matches with a use-case that I have. I don’t think I’m the only one, either, as is suggested by some of the survey data – a lot of people have tried this, especially since you don’t need to spend $12,000 on a new Apple II, and it’s very cool, but how much do we use it, and what for?…

…Suppose you want to analyse this month’s customer cancellations, or dispute a parking ticket, or file your taxes – you can ask an LLM, and it will work out what data you need, find the right websites, ask you the right questions, parse a photo of your mortgage statement, fill in the forms and give you the answers. We could move orders of magnitude more manual tasks into software, because you don’t need to write software to do each of those tasks one at a time. This, I think, is why Bill Gates said that this is the biggest thing since the GUI. That’s a lot more than a writing assistant.

It seems to me, though, that there are two kinds of problem with this thesis.

The narrow problem, and perhaps the ‘weak’ problem, is that these models aren’t quite good enough, yet. They will get stuck, quite a lot, in the scenarios I suggested above. Meanwhile, these are probabilistic rather than deterministic systems, so they’re much better for some kinds of task than others. They’re now very good at making things that look right, and for some use-cases this is what you want, but for others, ‘looks right’ is different to ‘right’…

…The deeper problem, I think, is that no matter how good the tech is, you have to think of the use-case. You have to see it. You have to notice something you spend a lot of time doing and realise that it could be automated with a tool like this…

…The cognitive dissonance of generative AI is that OpenAI or Anthropic say that we are very close to general-purpose autonomous agents that could handle many different complex multi-stage tasks, while at the same time there’s a ‘Cambrian Explosion’ of startups using OpenAI or Anthropic APIs to build single-purpose dedicated apps that aim at one problem and wrap it in hand-built UI, tooling and enterprise sales, much as a previous generation did with SQL. Back in 1982, my father had one (1) electric drill, but since then tool companies have turned that into a whole constellation of battery-powered electric hole-makers. One upon a time every startup had SQL inside, but that wasn’t the product, and now every startup will have LLMs inside.

I often compared the last wave of machine learning to automated interns. You want to listen to every call coming into the call centre and recognise which customers sound angry or suspicious: doing that didn’t need an expert, just a human (or indeed maybe even a dog), and now you could automate that entire class of problem. Spotting those problems and building that software takes time: machine learning’s breakthrough was over a decade ago now, and yet we are still inventing new use-cases for it – people are still creating companies based on realising that X or Y is a problem, realising that it can be turned into pattern recognition, and then going out and selling that problem.

You could propose the current wave of generative AI as giving us another set of interns, that can make things as well as recognise them, and, again, we need to work out what. Meanwhile, the AGI argument comes down to whether this could be far, far more than interns, and if we had that, then it wouldn’t be a tool anymore.

5. TIP622: Finding Certainty In An Uncertain World w/ Joseph Shaposhnik – Clay Finck and Joseph Shaposhnik

[00:29:29] Joseph Shaposhnik: I think of the credit bureaus and I think of a partner of theirs, which we’ll spend a minute talking about in a second, but. As you may know, the credit bureaus, there’s three of them in the United States. And they run an incredible oligopoly. If you want to secure a mortgage, get a car loan, rent a home, they’re involved in all of those decision making situations by the owners of those assets.

[00:29:55] Joseph Shaposhnik: As an example, if you go for a mortgage, all three credit bureaus will be pinned to get a score on you. All of them will be paid a couple of dollars for that score and all of that information that they’re pulling is contributory data. So there’s a relatively insignificant amount of incremental cost to generate that score and deliver it to the customer.

[00:30:21] Joseph Shaposhnik: You know, it’s a 95% incremental margin business. I mean, this is an incredible business. It’s basically an override on all economic activity in the United States and outside the United States where they play. And they’re just incredible businesses. But surprisingly not incredible stocks. You know, how could that be?

[00:30:40] Joseph Shaposhnik: It’s shocking to give you a sense organic growth if you look back, the last 5 years for the businesses have been approximately 7% a year. So, 3 or 4 times. Global GDP or U.S. GDP. They’ve outgrown the S&P the average S&P business over that period of time. They started with 30% EBITDA margins at the beginning of the 5 year period, so very profitable businesses.

[00:31:09] Joseph Shaposhnik: Yet over the last five years, two out of the three credit bureaus have underperformed the S&P, and over a 10 year period, they’ve been just in line performers with the S&P and so, I mean, they run an oligopoly. How could that possibly be? I used to be the credit bureau analyst at TCW, so I’m very familiar with these businesses, and they’re just incredible companies.

[00:31:33] Joseph Shaposhnik: And what happened is all three of these businesses spent more money on M&A than they generated in free cash flow over that five year period of time. They spent more money on M&A than all of the free cash flow they generated over the last five years. And they generate a lot of free cash flow. And let me say, let me just tell you, this is not on synergistic M&A.

[00:31:57] Joseph Shaposhnik: This was, I mean, they would call it synergistic, but it’s very difficult to synergize a near utility that they operate. And instead of just sticking to their knitting, they decided to acquire a lot of different data assets. that were incredibly expensive, generally from private equity, which doesn’t give assets away.

[00:32:18] Joseph Shaposhnik: And those returns are always, the returns on those businesses are always going to be lower than the returns on this incredible oligopoly that they run. And so, as interestingly as that, so of course, margins have been under pressure, returns have gone way down for these businesses because of all the acquisitions, these poor acquisitions at high multiples.

[00:32:42] Joseph Shaposhnik: And one of the most surprising things is we looked at the data on this, two out of the three businesses engaged in near zero share of purchases over that five year period of time. So you have this incredible business, you know, these three businesses that run an oligopoly, basically just an override on all economic activity.

[00:33:03] Joseph Shaposhnik: And they find all of these other businesses more attractive to allocate capital to than their own business, which is a 95% incremental margin business. Incredible. No wonder the stocks have not performed well, even though those businesses and those stocks should be like shooting fish in a barrel.

[00:33:20] Joseph Shaposhnik: So it’s incredible they bought back no, no stock, two out of the three businesses bought back no meaningful amount of stock. And not surprisingly, those businesses underperformed. In contrast to that, they have a partner, which is Fair Isaacs. And so Fair Isaacs, which is, the ticker is FICO provides the formula to the credit bureaus, which generates the score.

[00:33:44] Joseph Shaposhnik: The credit bureaus contribute the data. and the data with the formula creates a score that they can then sell to their end customers. So the bureaus pay FICO a fee for the formula, and they take the formula, and they generate a score, and they sell it to their customer. So you would think that FICO is basically in this ecosystem, has similar growth dynamics, has similar returns going into that 5 year period of time, similar EBITDA margins, tied to the same end markets, relatively similar company.

[00:34:17] Joseph Shaposhnik: Yet, over that five year period of time, FICO took all of its free cash flow, all of it, and used it to repurchase its shares. And so over the last five years, FICO has reduced share count by 20%, has engaged in no meaningful acquisitions to dilute its incredible franchise, and has generated a five bagger over the last five years.

[00:34:44] Joseph Shaposhnik: compared to the bureaus that have generated 15 to 100% return, total return over that 5 year period of time. So, a 5 bagger, which has outperformed the market by a ton, compared to an underperforming or an inline performance for the bureaus, I think just tells the tale of how important great capital allocation decision making is, how important it is to be aligned with a management team that understands how to generate value for shareholders.

[00:35:13] Joseph Shaposhnik: And I think for us and for everybody, serves as a warning when we think about investing with teams that are acquiring businesses in general and certainly acquiring businesses That are not as attractive as the core business. So capital allocation makes or breaks stories all the time, and incentives generally drive these decisions, but often times it just takes and an investor oriented CEO to see the big opportunity, which is usually in its core and not far afield.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google) and Meta Platforms. Holdings are subject to change at any time.

What We’re Reading (Week Ending 05 May 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 05 May 2024:

1. Karen Karniol-Tambour on investment anomalies at Sohn 2024 (transcript here) – Karen Karniol-Tambour and Jawad Mian

Jawad Mian (00:30): So 6 months ago equities were rallying in anticipation of lower interest rates but now we’ve seen year-to-date equities are rallying despite higher bond yields. So with a strong economy and inflation less of an issue, are you reverting to the typical inverse relationship between equities and bonds.

Karen Karniol-Tambour (00:49): The relationship between equities and bonds – it’s not an immutable fact of life. It’s not just a thing that occurs. It’s a function of the fundamental building blocks in stocks and bonds. When you look at stocks and bonds, they have a lot of things in common. They’re all future cash flows you’re discounting to today. So if you raise that, it’s bad for both and they both don’t do great when inflation is strong. The real inverse comes from their reaction to growth for the reason you’re saying. If growth is strong, then you can get equities rising and at the same time you can actually get the central bank tightening in response to that growth, which is bad for the bonds. And actually, the anomaly has been the years leading up to 2022, where inflation was just just a non-factor and the only dominant macro issue was growth. And so we’ve gotten really used to the idea that stocks and bonds have this inverse relationship. But that’s actually the anomaly. It’s not that normal to have a world where inflation just doesn’t matter. And finally, we live through this period where it’s like, “Wait a minute, inflation, its gravitational pull was at such a low level it was irrelevant – it’s becoming relevant again.” And we got this positive correlation where they both did badly, because you need to tighten in response to that inflation rearing its head.

Today – knock on wood – we look like we’re back to a world where inflation is not a non-issue, but it’s not a dominant issue, where we can have the kind of market action we’ve enjoyed so far in 2024, where we find out growth’s pretty damn resilient, growth’s doing great, companies can do well, earnings can do well, and at the same time the FED can ease less-than-expected or tighten relative to expectations at the same time. If they were tightening to stop very bad inflation, that would be a very different outcome. So the fundamental question as an investor is sort of where is the gravitational pull of inflation going to be? Is this going to be a major topic that then lead stocks and bonds sometimes to act the same way? Or is it going to go back to being kind of a non-issue?…

…Mian (02:53): A second anomaly. For the last 50 years, we’ve seen the US budget deficit average around 3% and it’s projected to be 6% over the next decade. So far we have seen markets being willing to finance these record deficits, in contrast to the UK for example. How come?

Karniol-Tambour (03:11): I think the best answer to this starts with actually the current account deficits, because obviously that’s part of who’s buying all the bonds we’re issuing. And it is a really weird anomaly because the United States is buying way more foreign goods than they’re buying ours. And typically if countries do that, their currency is weak because they have to convince someone to hold all the currency on the other side of that, so they have to attract all this financing. That the United States is running a massive current account deficit and yet the dollar is strong because what’s happening on the other end is people are just so enthusiastic about buying dollar financial assets. It’s so extreme that I think the United States has kind of a version of a Dutch disease.

So the classic Dutch disease is, you’re Saudi Arabia, you have oil. No one’s buying oil because you’re Saudi Arabia. No one’s thinking, “I really want Saudi oil.” They just need to fill up their car. So whatever the gas, is the gas is. But as Saudi Arabia, you get uncompetitive outside of it because money’s flooding in just for your oil, for nothing else. The United States has kind of become that on financial assets, which is people aren’t really thinking “I just want US financial assets.” It’s just that United States financial assets have done so well, they’re the dominant part of the index in stocks and in bonds. So anyone that needs to save any money around the world just ends up in US assets. As long as you care at all about market cap – which anyone reasonable would – and you’re going to the big market around the world, if you’re saving, you’re giving the United States money. And so we’re ending up with this flood of money that is a huge anomaly where we actually have a rising currency making everything else kind of uncompetitive, because people just want to buy stocks and bonds and no one else enjoys that. So we can run these huge deficits and sort of not worry about it.

2. Remembering Daniel Kahneman: A Mosaic of Memories and Lessons – Evan Nesterak and many others

To be continued …

By Richard Thaler, Professor of Behavioral Science and Economics, University of Chicago

My fondest memories of working with Danny come from 1984 to ’85 when I spent a year visiting him in Vancouver at The University of British Columbia. Danny had just begun a new project with Jack Knetsch on what people think is fair in market transactions and they invited me to join them. We had the then-rare ability to ask survey questions to a few hundred randomly selected Canadians each week. We would draft three versions of five questions, fax them to Ottawa Monday morning, get the results faxed back to us Thursday afternoon. Who needs Mturk! We then spent the weekend digesting the results and writing new questions.

We learned that raising the price of snow shovels the morning after a blizzard might make sense to an economist, but would make customers angry. Danny displayed two of his most prominent traits. He was always a skeptic, even (especially?) about his own ideas, so we stress-tested everything. And he was infinitely patient in that pursuit. Was our finding just true for snow shovels? What about water after a hurricane? Flu medicine? How about late-season discounts (which of course are fine). It was total immersion; meeting in person several times a week and talking constantly. We were in the zone.

Although we spent another year together in New York seven years later, we were unable to recreate that intensity. We had too many other balls in the air. But we continued our conversations and friendship until the end. Every conversation ended the same way: “To be continued.”…

...I’m more like a spiral than a circle

By Dan Lovallo, Professor of Strategy, Innovation and Decision Sciences, University of Sydney

Many people have heard that Danny changes his mind—a lot. This is certainly true. I have never written even a 5,000-word essay with him that didn’t take a year. Let me add another dimension to the discussion. During our last working dinner at a bistro in New York, and possibly out of mild frustration, I said, “Danny, you know you change your mind a lot.” It wasn’t a question. He continued chewing. I continued my line of non-question questioning: “And often you change it back to what it was at the beginning.”

Danny, having finished his bite and without missing a beat, looked up and in his characteristic lilt said, “Dan, that’s when I learn the most.” Then using his finger he drew a circle in space. “I don’t go around and around a problem. It might seem like it, but I am getting deeper and deeper.” The circle morphed into a three-dimensional spiral. “So, you’re missing all the learning,” he explained, as he displayed the invisible sculpture. “I’m more like a spiral than a circle.” Happy with this new idea, Danny grinned as only Danny could…

A case in character

By Angela Duckworth, Professor of Psychology, University of Pennsylvania

One evening, more than twenty years ago, I was the last one in the lab when the phone rang. “Hello?” I said, I hope not brusquely. I was a Ph.D. student at the time and eager to get back to my work. “Hello?” came the reply of an uncommonly polite older gentleman, whose accent I couldn’t quite place. “I’m so sorry to trouble you,” he continued. “I believe I’ve just now left my suitcase there.” Ah, this made sense. We’d hosted an academic conference that day. “It’s a terrible inconvenience, I know, but might you keep it somewhere until I can return to pick it up?” “Sure,” I said, cradling the receiver and grabbing a notepad. “How do you spell your name?” “Thank you so very much. It’s K-A-H-N-E-M-A-N.” I just about fainted. “Yes, Dr. Kahneman,” I said, coming to my senses, likely more deferentially than when I’d first picked up.

When I hung up, I thought to myself, Oh, it’s possible to be a world-famous genius—the most recently anointed Nobel laureate in economics, among other honors—and interact with anybody and everybody with utmost respect and dignity, no matter who they are. In the years that followed, I got to know Danny Kahneman much better, and when I did, that view was only confirmed. Confirmation bias? Halo effect? No and no. What then? Character. The world is mourning the loss of Danny Kahneman the genius, as we should, but I am missing Danny Kahneman the person…

Anxious and unsure

By Eric Johnson, Professor of Business, Columbia University

A few months before the publication of Thinking, Fast and Slow in 2011, the Center for Decision Sciences had scheduled Danny to present in our seminar series. We were excited because he had decided to present his first “book talk” with us. Expecting a healthy crowd, we scheduled the talk in Uris 301, the biggest classroom in Columbia Business School.

I arrived in the room a half hour early to find Danny, sitting alone in the large room, obsessing over his laptop. He confided that he had just changed two-thirds of the slides for the talk and was quite anxious and unsure about how to present the material. Of course, after the introduction, Danny presented in his usual charming, erudite style, communicating the distinction between System 1 and System 2 with clarity to an engaged audience. Afterwards, I asked him how he thought it went, and he said, “It was awful, but at least now I know how to make it better.” Needless to say, the book went on to become an international bestseller.

This was not false modesty. Having studied overconfidence throughout his career, Danny seemed immune to its effects. While surely maddening to some coauthors, this resulted in work that was more insightful and, most importantly to Danny and to us, correct. He was not always right, but always responsive to evidence, supportive or contradictory. For example, when some of the evidence cited in the book was questioned as a result of the replication crisis in psychology, Danny revised his opinion, writing in the comments of a critical blog: “I placed too much faith in underpowered studies.”

The best tribute to Danny, I believe, is adopting this idea, that science and particularly the social sciences, is not about seeming right, but instead, being truthful…

Practical problem solving

By Todd Rogers, Professor of Public Policy, Harvard University

I was part of a group helping some political candidates think about how to respond to untrue attacks by their political rivals. We focused on what cognitive and social psychology said about persuasive messaging. Danny suggested a different emphasis I hadn’t considered.

He directed us to a literature in cognitive psychology on cognitive associations. Once established, associations cannot simply be severed; attempting to directly refute them often reinforces them, and logical arguments alone can’t undo them. But these associations can be weakened when other competing associations are created.

For instance, if falsely accused of enjoying watching baseball, I’d be better off highlighting genuine interests—like my enjoyment of watching American football or reality TV—to dilute the false association with baseball. This anecdote is one small example of the many ways Danny’s profound intellect has influenced practical problem-solving. He’ll be missed and remembered.

Premortems

By Michael Mauboussin, Head of Consilient Research, Morgan Stanley

The opportunity to spend time with Danny and the chance to interview him were professional delights. One of my favorite lessons was about premortems, a technique developed by Gary Klein that Danny called one of his favorite debiasing techniques. In a premortem, a group assumes that they have made a decision (which they have yet to do), places themselves in the future (generally a year from now), and pretends that it worked out poorly. Each member independently writes down the reasons for the failure.

Klein suggested that one of the keys to premortems was the idea of prospective hindsight, that putting yourself into the future and thinking about the present opens up the mind to unconsidered yet relevant potential outcomes. I then learned that the findings of the research on prospective hindsight had failed to replicate—which made me question the value of the technique.

Danny explained that my concern was misplaced and that prospective hindsight was not central to the premortem. Rather, it was that the technique legitimizes dissent and allows organizations the opportunities to consider and close potential loopholes in their plans. That I had missed the real power of the premortem was a revelation and a relief, providing me with a cherished lesson…

Eradicating unhappiness

By George Loewenstein, Professor of Economics and Psychology, Carnegie Mellon University

For Danny, research was intensely personal. He got into intellectual disputes with a wide range of people, and these would hurt him viscerally, in part because it pained him that people he respected could come to different conclusions from those he held so strongly. He came up with, or at least embraced, the concept of “adversarial collaboration” in which researchers who disagreed on key issues would, however, agree upon a definitive test to determine where reality lay. A few of these were successful, but others (I would say most) ended with both parties unmoved, perhaps reflecting Robert Abelson’s insight that “beliefs are like possessions,” and, hence subject to the endowment effect.

I was spending time with Danny when he first got interested in hedonics—happiness—and that was a personal matter as well. His mother was declining mentally in France, and he agonized about whether to visit her; the issue was that she had anterograde amnesia, so he knew that she would forget his visit as soon as it ended. The criterion for quality of life, he had decided, should be the integral of happiness over time; so that—although she would miss out on the pleasure of remembering it—his visit would have value if she enjoyed it while it was happening.

Showing the flexibility of his thinking, and his all-too-rare willingness to learn from the data, his perspective changed as he studied happiness. He became more concerned about the story a life tells, including, notably, its peak and end; he concluded that eradicating unhappiness was a more important goal than fostering happiness, and began to draw a sharp distinction between happiness and life satisfaction, perhaps drawing, again, on his own experience. He always seemed to me to be extremely high in life satisfaction, but considerably less so in happiness.

3. Paradox of China’s stock market and economic growth – Glenn Luk

Joe Weisenthal of Bloomberg and the Odd Lots posed this question on Twitter/X:

“Given that the stock market hasn’t been especially rewarding to the volume-over-profits strategy undertaken by big Chinese manufacturers, what policy levers does Beijing have to sustain and encourage the existing approach?”

Many people may have noticed that despite the impressive growth of Chinese manufacturers in sectors like electric vehicles, the market capitalizations of these companies are dwarfed by Tesla. This seeming paradox lies at the heart of the the question posed by Joe.

In 2020, I shared an observation that China cares a lot more about GDP than market capitalization. I was making this observation in the context of Alibaba1 but would soon broaden the observation to encapsulate many more situations. In sharp contrast to Americans, Beijing just does not seem to care that much about equity market valuations but do seem to very much care about domestic growth and economic development…

…With respect to private sector market forces, Chinese policymakers tend to see its role as coordinators of an elaborate “game” that is meant to create an industry dynamic that drives desired market behaviors. The metaphor I sometimes use is as the Dungeon Master role in Dungeons & Dragons.

These “desired market behaviors” tend to overwhelmingly revolve around this multi-decade effort to maximize economic development and growth. Beijing has been very consistent about the goal to become “fully developed” by the middle of the 21st century.

To date, I would say that Chinese policymakers have been relatively successful using the approaches and principles described above to drive economic growth:

  • Priority on labor over capital / wage growth over capital income growth. Prioritizing labor is a key pillar of China’s demand-side support strategy. Growth in household income drives growth in domestic demand (whether in the form of household gross capital formation or expenditures).
  • Setting up rules to foster the create competitive industry dynamics and motivate economic actors to reinvest earnings back into growth.
  • Periodic crackdowns to disrupt what is perceived to be rent-seeking behavior, particularly from private sector players that have accumulated large amounts of equity capital (vs. small family businesses):
    • Anti-competitive behavior (e.g. Alibaba e-commerce dominance in the late 2010s)
    • Regulatory arbitrage (moral hazards inherent in Ant Financial’s risk-sharing arrangement with SOE banks)
  • Societal effects (for-profit education driving “standing on tiptoes” approach to childhood education)
  • Supply-side support to encourage dynamic, entrepreneurial participation from private sector players like in the clean energy transition to drive rapid industry through scale and scale-related production efficiencies. China has relied on supply-side strategies to support economic for decades despite repeated exhortations by outsiders to implement OECD-style income transfers.
  • Encouraging industry consolidation (vs. long drawn-out bankruptcies) once sectors have reached maturity although there are often conflicting motivations between Beijing and local governments.

A consistent theme is Beijing’s paranoia to rent-seeking behavior by capitalists (especially those who have accumulated large amounts of capital). It is sensitive to the potential stakeholder misalignment when capitalists — who are primarily aligned with one stakeholder class (fiduciary duty to equity owners).

It would prefer that rent-seeking behavior be handled by the party instead, whose objective (at least in theory) is to distribute these rents back to “The People” — although naturally in practice it never turns out this way; Yuen Yuen Ang has written multiple volumes about the prevalence of Chinese-style corruption and its corrosive economic effects.

So to bring it back to Joe’s question, the answer on whether Chinese policymakers can continue these policies going forward very much revolves around this question of rent-seeking: is it better to be done by the government or by private sector capitalists? What should be abundantly clear is that Beijing is definitive on this question: the party will maintain a monopoly on rent-seeking.

4. What Surging AI Demand Means for Electricity Markets – Tracy Alloway, Joe Weisenthal, and Brian Janous

Brian (09:58):

Yeah, and you’re right, I mean it’s not like we didn’t know that Microsoft had a partnership with OpenAI and that AI was going to consume energy. I think everyone though was a bit surprised at just how quickly what ChatGPT could do just captured the collective consciousness.

You probably remember when that was released. I mean it really sort surprised everyone and it became this thing where suddenly, even though we sort of knew what we were working on, it wasn’t until you put it out into the world that you realize maybe what you’ve created. That’s where we realized we are running up this curve of capability a lot faster than we thought. A number of applications that are getting built on this and the number of different ways that it’s being used and how it’s just become sort of common parlance. I mean, everyone knows what Chat GPT-3 is, and no one knew what it was the month before that.

So there was a bit, I think of a surprise in terms of just how quickly it was going to capture the collective consciousness and then obviously lead to everything that’s being created as a result. And so we just moved up that curve so quickly and I think that’s where the industry maybe got, certainly the utilities were behind because as you may have seen there, a lot of them are starting to restate their low-growth expectations.

And that was something that was not happening right before that. And so we’ve had massive changes just in the last two years of how utilities are starting to forecast what forecast. So if you take a look at a utility like Dominion in Virginia, so that’s the largest concentration of data centers in the United States. So they’re pretty good representative of what’s happening. If you go back to 2021, they were forecasting load growth over a period of 15 years of just a few percent.

I mean it was single-digit growth over that entire period. So not yearly growth, but over 15 years, single-digit growth. By 2023, they were forecasting to grow 2X over 15 years. Now keep in mind this is an electric utility. They do 10-year planning cycles. So because they have very long lead times for equipment for getting rights of away for transmission lines, they aren’t companies that easily respond to a 2X order of magnitude growth changed over a period of 15 years.

I mean, that is a massive change for electric utility, particularly given the fact that the growth rate over the last 15 to 20 years has been close to zero. So there’s been relatively no load growth in 15 to 20 years. Now suddenly you have utilities having to pivot to doubling the size of their system in that same horizon.

Tracy (13:10):

I want to ask a very basic question, but I think it will probably inform the rest of this conversation, but when we say that AI consumes a lot of energy, where is that consumption actually coming from? And Joe touched on this in the intro, but is it the sheer scale of users on these platforms? Is it, I imagine the training that you need in order to develop these models. and then does that energy usage differ in any way from more traditional technologies?

Brian (13:43):

Yeah, so whenever I think about the consumption of electricity for AI or really any other application, I think you have to start at sort of the core of what we’re talking about, which is really the human capacity for data, like whether it’s AI or cloud, humans have a massive capacity to consume data.

And if you think about where we are in this curve, I mean we’re on some form of S-curve of human data consumption, which then directly ties to data centers, devices, energy consumption ultimately, because what we’re doing is we’re turning energy into data. We take electrons, we convert them to light, we move them around to your TV screens and your phones and your laptops, etc. So that’s the uber trend that we’re riding up right now. And so we’re climbing this S-curve. I don’t know that anyone has a good sense of how steep or how long this curve will go.

If you go back to look at something like electricity, it was roughly about a hundred year. S-curve started in the beginning of last century. And it really started to flat line, as I mentioned before, towards the beginning of this century. Now we have this new trajectory that we’re entering, this new S-curve that we’re entering that’s going to change that narrative. But that S-curve for electricity took about a hundred years.

No one knows where we are on that data curve today. So when you inject something like AI, you create a whole new opportunity for humans to consume data, to do new things with data that we couldn’t do before. And so you accelerate us up this curve. So we were sitting somewhere along this curve, AI comes along and now we’re just moving up even further. And of course that means more energy consumption because the energy intensity of running an AI query versus a traditional search is much higher.

Now, what you can do with AI obviously is also much greater than what you can do with a traditional search. So there is a positive return on that invested energy. Oftentimes when this conversation comes up, there’s a lot of consternation and panic over ‘Well, what are we going to do? We’re going to run out of energy.’

The nice thing about electricity is we can always make more. We’re never going to run out of electricity. Not to say that there’s not times where the grid is under constraint and you have risks of brownouts and blackouts. That’s the reality. But we can invest more in transmission lines, we can invest more in power plants and we can create enough electricity to match that demand.

Joe (16:26):

Just to sort of clarify a point and adding on to Tracy’s question, you mentioned that doing an AI query is more energy intensive than, say, if I had just done a Google search or if I had done a Bing search or something like that. What is it about the process of delivering these capabilities that makes it more computationally intensive or energy intensive than the previous generation of data usage or data querying online?

Brian (16:57):

There’s two aspects to it, and I think we sort of alluded to it earlier, but the first is the training. So the first is the building of the large language model. That itself is very energy intensive. These are extraordinarily large machines, collections of machines that use very dense chips to create these language models that ultimately then get queried when you do an inference.

So then you go to ChatGPT and you ask it to give you a menu for a dinner party you want to have this weekend, it’s then referencing that large language model and creating this response. And of course that process is more computationally intensive because it’s doing a lot more things than a traditional search does. A traditional search just matched the words you put into a database of knowledge that it had put together, but these large language models are much more complex and then therefore the things you’re asking it to do is more complex.

So it will almost by definition be a more energy intensive process. Now, that’s not to say that it can’t get more efficient and it will, and Nvidia just last week was releasing some data on some of its next generation chips that are going to be significantly more efficient than the prior generation.

But one of the things that we need to be careful of is to think that because something becomes more efficient, then therefore we’re going to use less of the input resource. In this case, electricity. That’s not how it works, because going back to the concept of human capacity for consuming data, all we do is we find more things to compute. And this is, you’ve probably heard of Jevons paradox, and this is the idea that, well, if we make more efficient steam engines, he was an economist in the 1800s and he said ‘Well, if make more efficient steam engines, then we’ll use less coal.’

And he is like ‘No, that’s not what’s going to happen. We’re going to use more coal because we’re going to mechanize more things.’ And that’s exactly what we do with data just because we’ve had Moore’s Law for years, and so chips has become incredibly more efficient than they were decades ago, but we didn’t use less energy. We used much more energy because we could put chips in everything.

So that’s the trend line that we’re on. It’s still climbing that curve of consumption. And so no amount of efficiency is going to take us at this point, at least because I don’t believe we’re anywhere close to the bend in that S-curve. No amount of efficiency is going to take us off of continuing to consume more electricity, at least in the near term…

…Brian (22:35):

Well, this is where it gets a little concerning is that you have these tech companies that have these really ambitious commitments to being carbon neutral, carbon negative, having a hundred percent zero carbon energy a hundred percent of the time, and you have to give them credit for the work they’ve done.

I mean, that industry has done amazing work over the last decade to build absolutely just gigawatts upon gigawatts of new renewable energy projects in the United States all over the world. They’ve been some of the biggest drivers in the corporate focus on decarbonization. And so you really have to give that industry credit for all it’s done and all the big tech companies have done some amazing work there.

The challenge though that we have is the environment that they did that in was that no growth environment we were talking about. They were all growing, but they were starting from a relatively small denominator 10 or 15 years ago. And so there was a lot of overhang in the utility system at that time because the utilities had overbuilt ahead of that sort of flatlining. So there was excess capacity on the system.

They were growing inside of a system that wasn’t itself growing on a net basis. So everything they did, every new wind project you brought on, every new solar project you bought on, those were all incrementally reducing the amount of carbon in the system. It was all net positive.

Now we get into this new world where their growth rates are exceeding what the utilities had ever imagined in terms of the absolute impact on the system. The utilities’ response is ‘The only thing we can do in the time horizon that we have is basically build more gas plants or keep online gas plants or coal plants that we were planning on shuttering.’

And so now that the commitments that they have to zero carbon energy to be carbon negative, etc., are coming into contrast with the response that the utilities are laying out in their what’s called integrated resource plans or IRPs.

And we’ve seen this recently just last week in Georgia. We’ve seen it in Duke and North Carolina, Dominion and Virginia. Every single one of those utilities is saying ‘With all the demand that we’re seeing coming into our system, we have to put more fossil fuel resources on the grid. It’s the only way that we can manage it in a time horizon we have.’ Now, there’s a lot of debate about whether that is true, but it is what’s happening…

…Brian (30:29):

That’s right. And that’s the big challenge that good planners have today is what loads do you say yes to and what are the long-term implications of that? And we’ve seen this play out over the rest of the globe where you’ve had these concentrations of data centers. This is a story that we saw in Dublin, we’ve seen it in Singapore, we’ve seen it in Amsterdam.

And these governments start to get really worried of ‘Wait a minute, we have too many data centers as a percentage of overall energy consumption.’ And what inevitably happens is a move towards putting either moratoriums on data center build out or putting very tight restrictions on what they can do and the scale at which they can do it. And so we haven’t yet seen that to any material degree in the United States, but I do think that’s a real risk and it’s a risk that the data center industry faces.

I think somewhat uniquely in that if you’re the governor of a state and you have a choice to give power to a say new EV car factory that’s going to produce 1,500, 2,000 jobs versus a data center that’s going to produce significantly less than that, you’re going to give it to the factory. The data centers are actually the ones that are going to face likely the most constraints as governments, utilities, regulators start wrestling with this trade-off of ‘Ooh, we’re going to have to say no to somebody.’…

…Tracy (36:36):

What are the levers specifically on the tech company or the data center side? Because again, so much of the focus of this conversation is on what can the utilities do, what can we do in terms of enhancing the grid managing supply more efficiently? But are there novel or interesting things that the data centers themselves can do here in terms of managing their own energy usage?

Brian (37:02):

Yes. There’s a few things. I mean, one is data centers have substantial ability to be more flexible in terms of the power that they’re taking from the grid at any given time. As I mentioned before, every data center or nearly every data center has some form of backup generation. They have some form of energy storage built into this.

So the way a data center is designed, it’s designed like a power plant with an energy storage plant that just happens to be sitting next to a room full of servers. And so when you break it down into those components, you say, okay, well how can we better optimize this power plant to be more of a grid resource? How can we optimize the storage plant to be more of a grid resource? And then in terms of even the servers themselves, how can we optimize the way the software actually operates and is architected to be more of a grid resource?

And that sort of thinking is what is being forced on the industry. Frankly, we’ve always had this capability. I mean, we were doing, I mean we did a project like 2016 with a utility where we put in flexible gas generators behind our meter because the utility was going to have to build a new power plant if we didn’t have a way to be more flexible.

So we’ve always known that we can do this, but the industry has never been pressurized to really think innovatively about how can we utilize all these assets that we have inside of the data center plant itself to be more part of the grid. So I think the most important thing is really thinking about how data centers become more flexible. There’s a whole ‘nother line of thinking, which is this idea of, well, utilities aren’t going to move fast enough, so data centers just need to build all their own power plants.

And this is where you start hearing about nuclear and SMRs and infusion, which is interesting, except it doesn’t solve the problem this decade. It doesn’t solve the problem that we’re facing right now because none of that stuff is actually ready for prime time. We don’t have an SMR that we can build today predictably on time, on budget.

So we are dependent on the tools that we have today, which are things like batteries, grid enhancing technologies, flexible load, reconductoring transmission lines to get more power over existing rights of ways. So there’s a number of things we can do with technologies we have today that are going to be very meaningful this decade and we should keep investing in things that are going to be really meaningful next decade. I’m very bullish on what we can do with new forms of nuclear technology. They’re just not relevant in the time horizon. The problem we’re talking about [now].

Joe (39:52):

At some point, we’re going to do an Odd Lots episode specifically on the promise of small modular reactors and why we still don’t have them despite the seeming benefits. But do you have a sort of succinct answer for why this sort of seeming solution of manufacturing them faster, etc., has not translated into anything in production?

Brian (40:14)

Well, quite simply, we just forgot how to do it. We used to be able to build nuclear in this country. We did it in the seventies, we did it in the eighties, but every person that was involved in any one of those projects is either not alive or certainly not still a project manager at a company that would be building nuclear plants, right?

I think we underestimate human capacity to forget things. Just because we’ve done something in the past doesn’t mean that we necessarily can do it. Again, we have to relearn these things, and as a country, we do not have a supply chain. We don’t have a labor force. We don’t have people that manage construction projects that know how to do any of these things.

And so when you look at what South Korea is doing, you look at what China’s doing, they’re building nuclear plants with regularity. They’re doing it at a very attractive cost. They’re doing it on a predictable time horizon, but they have actually built all of those resources that we just simply don’t have in this country that we need and we need to rebuild that capability. It just doesn’t exist today…

…Brian (41:50):

Absolutely. And so if you go back to the era that we’ve been in of relative no load growth, if you’re a utility regulator and utility comes and asks you for a billion dollars for new investment and you’re used to saying ‘no,’ you’re used to saying ‘Well, wait a minute. Why do you need this? What is this for? How is this going to help manage again, reliability, cost, predictability, etc.?’

Now you’re in this whole new world and going back to this concept of we easily forget things — no one who’s a regulator today or the head of utility today has ever lived through an environment where we’ve had this massive expansion of the demand for electricity. So everyone now, including the regulators are having to relearn, okay, how do we enable utility investment in a growth environment? It’s not something they’ve ever done before. And so they’re having to figure out, okay, how do we create the bandwidth for utilities to make these investments?

Because one of the fundamental challenges that utilities have is that they struggle to invest if there’s no customer sitting there asking for the request, so they can’t sort of invest. I mean, if I’m Nvidia and I’m thinking about the world five years from now and think ‘Wow, how many chips do I want to sell in 2030?’ I can go out and build a new factory. I can go out and invest capital and I can go do all, I mean, I don’t need to have an order from a Microsoft or an Amazon or a Meta to go do that. I can build speculatively.

Utilities can’t really do that. They’re basically waiting for the customer to come ask for it. But when you have all this demand show up at the same time, well, what happens? The lead time start to extend. And so instead of saying ‘Yeah, I’ll give you that power in a year or two years,’ it’s now like, ‘Well, I’ll give it to you in five to seven years.’ And so that’s an unsustainable way to run the electric utility grid. So we do need regulators to adapt and evolve to this new era of growth.

5. Reflections from the heart of Japan’s ancient cedar forest – Thomas Chua

Yakushima was particularly memorable, an island near Kagoshima famous for its wildlife and ancient cedar forests. These majestic cedars, some of the oldest trees in the world, grow steadily through centuries, unaffected by the transient storms and seasonal fluctuations.

This is Sennensugi, which means a thousand-year-old cedar tree even though it’s still young. Yakushima’s oldest tree (and the oldest tree in Japan) is Jōmon Sugi, which is estimated to be between 2,170 and 7,200 years old.

This resonates deeply with my investment strategy. Just as these enduring cedars are not swayed by the fleeting changes in their environment, I focus on “Steady Compounders”—companies with significant economic moats and consistent intrinsic value growth.

When friends learn about my extensive travels, they often ask, “What about your investments? Don’t you need to monitor them constantly?” What they usually mean about ”monitoring” isn’t analyzing quarterly business results, but rather obsessively tracking stock prices and consuming every tidbit of news to stay perpetually informed.

However, I liken such constant vigilance to setting up a camera in a forest to watch the trees grow, this approach isn’t just tedious—it’s unnecessary and potentially harmful, often prompting rash decisions.

Everyone invests to grow wealth, but understanding why you invest is crucial. For me, it serves to enrich my curiosity and intellect, rewards my eagerness to learn, and more importantly, grants me the freedom to live life on my terms and cherish moments with my loved ones.

Therefore, I don’t pursue obscure, unproven companies which require intensive monitoring. Instead, I look for Steady Compounders — firms with a significant economic moat that are growing their intrinsic value steadily.

Like the steady growth of Yakushima’s cedars, these firms don’t need constant oversight; they thrive over long periods through economic cycles, much as the cedars endure through seasonal changes. Investing in such companies gives me the freedom to explore the world, knowing my investments are growing steadily, mirroring the quiet, powerful ascent of those ancient trees.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google), Amazon, Meta Platforms, Microsoft, and Tesla. Holdings are subject to change at any time.

What We’re Reading (Week Ending 28 April 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 28 April 2024:

1. 10 Questions with Chris Beselin – Michael Fritzell and Chris Beselin

Today, I’ll be interviewing Chris Beselin, who runs a Vietnam-focused activist fund and two tech businesses from his base in Ho Chi Minh City…

3. Why do you Vietnam has been so successful as an economy – why has it developed faster than almost any other nation on earth?

There are a range of factors, of course, but just to outline a few:

It’s a balanced economy and growth model – it’s not your typical emerging market, where the economy is overly dependent on one or a handful of commodities.

Rather, the Vietnamese growth model has multiple core engines: its one of the most trade-focused economies in the world (measured as (export+import)/GDP) with free trade agreements signed with countries representing 60% of global GDP, it has a young and well-educated population where English proficiency is on par with e.g. India and South Korea, it has a sizeable and confident middle class that is rapidly growing and it has a stable government that has been focused on pro-market deregulations for the past 35 years.

And in contrast to what many people think from the outset, Vietnamese exports are primarily engineering-driven (as opposed to lower value-add textiles and similar). Around 45% of the exports are electronics, smartphones, laptops and machinery components. In this sense, my conviction is that Vietnam is much more the next South Korea or Japan than the next China.

To me, this all boils down to the fact that the number one asset of the country is its young, savvy and hungry engineering population (ca. 100,000 engineers are educated per year in Vietnam, of which around 50% are within software). The attractiveness of the Vietnamese engineering talent pulls foreign capital to invest in onshore engineering-centered manufacturing, which in turn has vast ripple effects on the employment of thousands of additional factory workers around the engineers…

5. What misconceptions do you think foreigners typically have about the country?

I think there are many. Just to name a few:

The first one is perhaps “Vietnam is almost like China, but smaller and less developed”. I went through a bit of the difference in the fabric of the economies and demographics previously, but then there is also the very important difference in politics. Geopolitically, Vietnam is not and will never be or perceive itself to be a global superpower like China – it doesn’t have any geopolitical ambitions outside its own borders like China has.

Vietnam is primarily interested in developing its economy through trade and FDI, this in turn means that Vietnam in practice benefits from being geopolitically neutral between East and West and by trading/being friends with “everyone”. So far the country has managed this balance very astutely for decades.

Another common misconception (particularly for Westerners growing up during the Vietnam War) is that “Vietnam is just getting back on its feet after the recent war”. Obviously, this perspective is wildly outdated, but it’s still surprisingly common among foreign visitors. To put it into perspective, perhaps a suitable analogy is if you would have been saying/thinking similar things about France or the UK in the mid 90s… (also then ca. 50 years from the end of the Second World War, just like Vietnam is today 50 years away from its war ending in 1975).

2. Private Credit Offers No Extra Gains After Fees, New Study Finds – Justina Lee

A trio of academics has a bold take on the booming $1.7 trillion private credit market: after accounting for additional risks and fees, the asset class delivers virtually no extra return to investors…

…“It’s not a panacea for investors where they can earn 15% risk-free,” said Michael Weisbach, a finance professor at Ohio State University who co-wrote the research with Isil Erel and Thomas Flanagan. “Once you adjust for the risk, they basically are getting the amount they deserve, but no more.”

Behind the research is complex math to try to untangle the alpha part of a return that’s down to skill, and the beta part that might just come from stumbling into a bull market. While comparing stock pickers to a market benchmark like the S&P 500 is standard by now, it’s not obvious what the right yardstick is for private-credit funds, which make idiosyncratic and opaque loans to a wide array of companies…

..The three economists dissected MSCI data on 532 funds’ cash flows, covering their incoming capital and distributions to investors. They compare the industry’s performance to stock and credit portfolios with similar characteristics, whose fluctuations end up explaining the majority of private-credit returns. The study makes the case that these private credit funds also carry some equity risk, since around 20% of their investments contain equity-like features such as warrants.

After accounting for those risks, they find that there’s still alpha left on the table — which only vanishes once fees paid to these managers are deducted…

…As private markets boom, some quants — most notably Cliff Asness of AQR Capital Management — have suggested that investors are being misguided by returns that mask volatility and may be less impressive than they appear.

True at Adams Street Partners, who co-wrote one of the first papers on private-credit performance, cautions that until the industry faces its first downturn it may be hard to determine real alpha. But he says the NBER study is a good step toward digging beneath the surface of private-credit returns.

3. Americans are still not worried enough about the risk of world war – Noah Smith

When Americans think of World War 2, we usually think of the roughly four years of the war that we participated in, from late 1941 through 1945. Those years were indeed the most climactic and destructive of the war, by far, but the war actually began earlier. In fact, although the official start date is September 1, 1939, it’s easy to make an argument that the war began long before that…

…But throughout the 1930s, there were a number of conflicts that led into World War 2, and eventually merged with that overall conflict, like tributaries emptying into a great river. Let’s do a quick timeline of these.

In 1931-32, Japan seized Manchuria from China, an act that led inexorably to a wider war between the two powers. The Manchurian war and occupation also set Japan on a path toward militarism, bringing to power the regime that would ultimately prosecute WW2 itself.

In 1935-36, fascist Italy invaded and conquered Ethiopia. The League of Nations halfheartedly tried to stop the war and failed, leading to the League being discredited and the post-WW1 order being greatly weakened. That emboldened the fascist powers. Ethiopian resistance to Italian rule would eventually become a minor theater of WW2.

From 1935 through 1939, Japan and the Soviet Union fought an undeclared border war, ultimately culminating in major battles in 1939, which the USSR won. That led to Japan seeking an alliance with Nazi Germany, and eventually led to the Soviets’ entry into the war against Japan at the very end of WW2. (The realization that Japan couldn’t defeat the Soviets and conquer Siberian oil fields also prompted Japan to try to take Southeast Asian oil instead, when it needed oil to prosecute its war against China; this led to Pearl Harbor and the Pacific War.)

From 1936 through 1939, Nazi Germany, fascist Italy, and the Soviet Union fought each other in a proxy war: the Spanish Civil War. Units from all three powers officially or unofficially engaged in the fighting. When the Nationalists won, it emboldened the fascist powers even further.

In 1937, Japan invaded the remainder of China, in what’s called the Second Sino-Japanese War. This became a major theater of World War 2, accounting for almost as many deaths as the Nazi invasion of the USSR. It also prompted Japan to go to war with Britain and the U.S., in order to seize the oil fields of Indonesia to support the invasion of China. (The fact that we don’t count this as the start of WW2 seems like pure eurocentrism to me.)

In 1939, before the Soviet Union joined World War 2, it invaded Finland in what’s known as the Winter War, seizing some territory at great cost. This war would continue all the way through WW2 itself.

So there were no fewer than six wars in the 1930s that ended up feeding into World War 2 itself!..

…So if you were living at any point in 1931 through 1940, you would already be witnessing conflicts that would eventually turn into the bloodiest, most cataclysmic war that humanity has yet known — but you might not realize it. You would be standing in the foothills of the Second World War, but unless you were able to make far-sighted predictions, you wouldn’t know what horrors lurked in the near future.

In case the parallel isn’t blindingly obvious, we might be standing in the foothills of World War 3 right now. If WW3 happens, future bloggers might list the wars in Ukraine and Gaza in a timeline like the one I just gave.

Or we might not be in the foothills of WW3. I think there’s still a good chance that we can avert a wider, more cataclysmic war, and instead have a protracted standoff — Cold War 2 — instead. But I’m not going to lie — the outlook seems to be deteriorating. One big reason is that China appears to be ramping up its support for Russia…

…So it makes sense to view the Ukraine War as a European proxy conflict against Russia. But what’s more ominous is that it also makes an increasing amount of sense to view it as a Chinese proxy conflict against Europe.

A little over a year ago, there were reports that China was sending lethal aid to Russia. Believing these reports, I concluded — perhaps prematurely — that China had gone “all in” on Russia’s military effort. Some of the reports were later walked back, but the fact is, it’s very hard to know how much China is doing to provide Russia with drones and such under the table. But now, a year later, there are multiple credible reports that China is ramping up aid in various forms.

For example, the U.S. is now claiming that China is providing Russia with both material aid and geospatial intelligence (i.e. telling Russia where Ukrainian units are so Russia can hit them):

The US is warning allies that China has stepped up its support for Russia, including by providing geospatial intelligence, to help Moscow in its war against Ukraine.

Amid signs of continued military integration between the two nations, China has provided Russia with satellite imagery for military purposes, as well as microelectronics and machine tools for tanks, according to people familiar with the matter.

China’s support also includes optics, propellants to be used in missiles and increased space cooperation, one of the people said.

President Joe Biden raised concerns with Xi Jinping during their call this week about China’s support for the Russian defense industrial base, including machine tools, optics, nitrocellulose, microelectronics, and turbojet engines, White House National Security Council spokesperson Adrienne Watson said.

This is very similar to the aid that Europe and the U.S. are providing Ukraine. We also provide geospatial intelligence, as well as materiel and production assistance. If reports are correct — and this time, they come from the U.S. government as well as from major news organizations — then China is now playing the same role for Russia that Europe and the U.S. have been playing for Ukraine.

In other words, the Ukraine War now looks like a proxy war between China and Europe…

…Of course, World War 3 will actually begin if and when the U.S. and China go to war. Almost everyone thinks this would happen if and when China attacks Taiwan, but in fact there are several other flashpoints that are just as scary and which many people seem to be overlooking.

First, there’s the South China Sea, where China has been pressing the Philippines to surrender its maritime territory with various “gray zone” bullying tactics…

…The U.S. is a formal treaty ally of the Philippines, and has vowed to honor its commitments and defend its ally against potential threats.

And then there’s the ever-present background threat of North Korea, which some experts believe is seriously considering re-starting the Korean War. That would trigger a U.S. defense of South Korea, which in turn might bring in China, as it did in the 1950s.

It’s also worth mentioning the China-India border. China has recently reiterated its claim to the Indian state of Arunachal Pradesh, calling it “South Tibet” and declaring that the area was part of China since ancient times. India has vigorously rejected this notion, of course. An India-China border war might not start World War 3, but the U.S. would definitely try to help India out against China, as we did in 2022…

…America hasn’t mustered the urgency necessary to resuscitate its desiccated defense-industrial base. China is engaging in a massive military buildup while the U.S. is lagging behind. This is from a report from the Center for Strategic and International Studies:

The U.S. defense industrial base…lacks the capacity, responsiveness, flexibility, and surge capability to meet the U.S. military’s production needs as China ramps up defense industrial production. Unless there are urgent changes, the United States risks weakening deterrence and undermining its warfighting capabilities against China and other competitors. A significant part of the problem is that the U.S. defense ecosystem remains on a peacetime footing, despite a protracted war in Ukraine, an active war in the Middle East, and growing tensions in the Indo-Pacific in such areas as the Taiwan Strait and Korean Peninsula.

The United States faces several acute challenges.

First, the Chinese defense industrial base is increasingly on a wartime footing and, in some areas, outpacing the U.S. defense industrial base…Chinese defense companies…are producing a growing quantity and quality of land, maritime, air, space, and other capabilities. China…is heavily investing in munitions and acquiring high-end weapons systems and equipment five to six times faster than the United States…China is now the world’s largest shipbuilder and has a shipbuilding capacity that is roughly 230 times larger than the United States. One of China’s large shipyards, such as Jiangnan Shipyard, has more capacity than all U.S. shipyards combined, according to U.S. Navy estimates.

Second, the U.S. defense industrial base continues to face a range of production challenges, including a lack of urgency in revitalizing the defense industrial ecosystem…[T]here is still a shortfall of munitions and other weapons systems for a protracted war in such areas as the Indo-Pacific. Supply chain challenges also remain serious, and today’s workforce is inadequate to meet the demands of the defense industrial base.

“The Chinese defense industrial base is increasingly on a wartime footing.” If that isn’t a clear enough warning, I don’t know what would be. You have now been warned!

4. Memory Shortage and ASML – Nomad Semi

Although we are down to 3 major DRAM manufacturers, memory has always been a commodity since the 1970s. The wild swing in gross margin for SK Hynix in a year proves the point. Supply is the underlying driver of the memory cycle. Memory prices should always trend down over time with cost as memory manufacturers migrate to the next process node that allows for higher bit per wafer. There is always a duration mismatch in demand and supply due to the inelasticity of supply. When there is supernormal profit, capex will go up and supply will follow in 2 to 3 years. Supply will exceed demand and DRAM prices should fall to cost theoretically. This post will instead focus on the current upcycle and how it could actually be stronger than the 2016 cycle.

How we get here today is a result of the proliferation of AI and the worst downcycle since the GFC. To be fair, the 2022 downcycle was driven by a broad-based inventory correction across all the verticals rather than very aggressive capex hike from the memory producer. The downcycle led to negative gross margin for SK Hynix, Micron and Kioxia. Negative gross margin led to going concern risk, which led to Hynix and Micron cutting their capex to the lowest level in the last 7 years. This is despite the fact that we have moved from 1x nm to 1b nm which will require higher capex per wafer.

HBM (High Bandwidth Memory) has become very important in AI training, and you can’t run away from talking about HBM if you are looking at DRAM…

…HBM affects current and future DRAM supply in 2 different ways. The 1st is cannibalization of capex from DRAM and NAND where Fabricated Knowledge gave a very good analogy. The 2nd is as Micron mentioned in the last call that “HBM3E consumes approximately three times the wafer supply as DDR5 to produce a given number of bits in the same technology node”. In fact, this ratio will only get worse in 2026 as HBM4 can consume up to 5x the wafer supply as DDR5. The way it works is a HBM die size is double that of DDR5 (which already suffers from single digit die size penalty vs DDR4). HBM die size will only get bigger as more TSV is needed. Yield rate of HBM3e is below 70% and will only get harder as more dies are stacked beyond 8-hi. Logic base die of the HBM module is currently produced in-house by Micron and Hynix although this could be outsourced to TSMC for HBM4. In summary, not only is HBM consuming more of current wafer supply, but it is also cannibalizing the capex for future DRAM and NAND capacity expansion

In past upcycles, capex will often go up as memory producers gain confidence. Nobody wants to lose market share as current capex = future capacity → future market share. However, SK Hynix and Micron will be unable to expand their DRAM wafer capacity meaningfully in 2024 and 2025.

SK Hynix has limited cleanroom space available for DRAM expansion (~45k wpm) at M16 and this will be fully utilized by 2025. The company will have to wait till 2027 before Yong-In fab can be completed. Even when the balance sheet situation for SK Hynix improves in 2025, it will be limited by its cleanroom space.

For Micron, the situation is slightly better. Taichung fab also has limited space available for capacity expansion, but this will likely be earmarked for HBM production. Micron will have to wait until the new Boise fab is ready in 2026. Both Micron and Hynix will be limited in capacity expansion in 2025 against their will.

5. Artificial intelligence is taking over drug development – The Economist

The most striking evidence that artificial intelligence can provide profound scientific breakthroughs came with the unveiling of a program called AlphaFold by Google DeepMind. In 2016 researchers at the company had scored a big success with AlphaGo, an AI system which, having essentially taught itself the rules of Go, went on to beat the most highly rated human players of the game, sometimes by using tactics no one had ever foreseen. This emboldened the company to build a system that would work out a far more complex set of rules: those through which the sequence of amino acids which defines a particular protein leads to the shape that sequence folds into when that protein is actually made. AlphaFold found those rules and applied them with astonishing success.

The achievement was both remarkable and useful. Remarkable because a lot of clever humans had been trying hard to create computer models of the processes which fold chains of amino acids into proteins for decades. AlphaFold bested their best efforts almost as thoroughly as the system that inspired it trounces human Go players. Useful because the shape of a protein is of immense practical importance: it determines what the protein does and what other molecules can do to it. All the basic processes of life depend on what specific proteins do. Finding molecules that do desirable things to proteins (sometimes blocking their action, sometimes encouraging it) is the aim of the vast majority of the world’s drug development programmes.

Because of the importance of proteins’ three-dimensional structure there is an entire sub-discipline largely devoted to it: structural biology. It makes use of all sorts of technology to look at proteins through nuclear-magnetic-resonance techniques or by getting them to crystallise (which can be very hard) and blasting them with x-rays. Before AlphaFold over half a century of structural biology had produced a couple of hundred thousand reliable protein structures through these means. AlphaFold and its rivals (most notably a program made by Meta) have now provided detailed predictions of the shapes of more than 600m.

As a way of leaving scientists gobsmacked it is a hard act to follow. But if AlphaFold’s products have wowed the world, the basics of how it made them are fairly typical of the sort of things deep learning and generative AI can offer biology. Trained on two different types of data (amino-acid sequences and three-dimensional descriptions of the shapes they fold into) AlphaFold found patterns that allowed it to use the first sort of data to predict the second. The predictions are not all perfect. Chris Gibson, the boss of Recursion Pharmaceuticals, an AI-intensive drug-discovery startup based in Utah, says that his company treats AlphaFold’s outputs as hypotheses to be tested and validated experimentally. Not all of them pan out. But Dr Gibson also says the model is quickly getting better…

…A lot of pharma firms have made significant investments in the development of foundation models in recent years. Alongside this has been a rise in AI-centred startups such as Recursion, Genesis Therapeutics, based in Silicon Valley, Insilico, based in Hong Kong and New York and Relay Therapeutics, in Cambridge, Massachusetts. Daphne Koller, the boss of Insitro, an AI-heavy biotech in South San Francisco, says one sign of the times is that she no longer needs to explain large language models and self-supervised learning. And Nvidia—which makes the graphics-processing units that are essential for powering foundation models—has shown a keen interest. In the past year, it has invested or made partnership deals with at least six different AI-focused biotech firms including Schrodinger, another New York based firm, Genesis, Recursion and Genentech, an independent subsidiary of Roche, a big Swiss pharmaceutical company.

The drug-discovery models many of the companies are working with can learn from a wide variety of biological data including gene sequences, pictures of cells and tissues, the structures of relevant proteins, biomarkers in the blood, the proteins being made in specific cells and clinical data on the course of disease and effect of treatments in patients. Once trained, the AIs can be fine tuned with labelled data to enhance their capabilities.

The use of patient data is particularly interesting. For fairly obvious reasons it is often not possible to discover the exact workings of a disease in humans through experiment. So drug development typically relies a lot on animal models, even though they can be misleading. AIs that are trained on, and better attuned to, human biology may help avoid some of the blind alleys that stymie drug development.

Insitro, for example, trains its models on pathology slides, gene sequences, MRI data and blood proteins. One of its models is able to connect changes in what cells look like under the microscope with underlying mutations in the genome and with clinical outcomes across various different diseases. The company hopes to use these and similar techniques to find ways to identify sub-groups of cancer patients that will do particularly well on specific courses of treatment.

Sometimes finding out what aspect of the data an AI is responding to is useful in and of itself. In 2019 Owkin, a Paris based “AI biotech”, published details of a deep neural network trained to predict survival in patients with malignant mesothelioma, a cancer of the tissue surrounding the lung, on the basis of tissue samples mounted on slides. It found that the cells most germane to the AI’s predictions were not the cancer cells themselves but non-cancerous cells nearby. The Owkin team brought extra cellular and molecular data into the picture and discovered a new drug target. In August last year a team of scientists from Indiana University Bloomington trained a model on data about how cancer cells respond to drugs (including genetic information) and the chemical structures of drugs, allowing it to predict how effective a drug would be in treating a specific cancer.

Many of the companies using AI need such great volumes of high quality data they are generating it themselves as part of their drug development programmes rather than waiting for it to be published elsewhere. One variation on this theme comes from a new computational sciences unit at Genentech which uses a “lab in the loop” approach to train their AI. The system’s predictions are tested at a large scale by means of experiments run with automated lab systems. The results of those experiments are then used to retrain the AI and enhance its accuracy. Recursion, which is using a similar strategy, says it can use automated laboratory robotics to conduct 2.2m experiments each week…

…The world has seen a number of ground breaking new drugs and treatments in the past decade: the drugs targeting GLP-1 that are transforming the treatment of diabetes and obesity; the CAR-T therapies enlisting the immune system against cancer; the first clinical applications of genome editing. But the long haul of drug development, from discerning the biological processes that matter to identifying druggable targets to developing candidate molecules to putting them through preclinical tests and then clinical trials, remains generally slow and frustrating work. Approximately 86% of all drug candidates developed between 2000 and 2015 failed to meet their primary endpoints in clinical trials. Some argue that drug development has picked off most of biology’s low-hanging fruit, leaving diseases which are intractable and drug targets that are “undruggable”.

The next few years will demonstrate conclusively if AI is able to materially shift that picture. If it offers merely incremental improvements that could still be a real boon. If it allows biology to be deciphered in a whole new way, as the most boosterish suggest, it could make the whole process far more successful and efficient—and drug the undruggable very rapidly indeed. The analysts at BCG see signs of a fast-approaching AI-enabled wave of new drugs. Dr Pande warns that drug regulators will need to up their game to meet the challenge. It would be a good problem for the world to have. 


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google DeepMind), ASML, and TSMC. Holdings are subject to change at any time.

What We’re Reading (Week Ending 21 April 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 21 April 2024:

1. The Anguish of Central Banking – Arthur F. Burns

Why, in particular, have central bankers, whose main business one might suppose is to fight inflation, been so ineffective in dealing with this worldwide problem?

To me, as a former central banker, the last of these questions is especially intriguing. One of the time-honored functions of a central bank is to protect the integrity of its nation’s currency, both domestically and internationally. In monetary policy central bankers have a potent means for fostering stability of the general price level. By training, if not also by temperament, they are inclined to lay great stress on price stability, and their abhorrence of inflation is continually reinforced by contacts with one another and with like-minded members of the private financial community. And yet, despite their antipathy to inflation and the powerful weapons they could wield against it, central bankers have failed so utterly in this mission in recent years. In this paradox lies the anguish of central banking…

…Analyses of the inflation that the United States has experienced over the past fifteen years frequently proceed in three stages. First are considered the factors that launched inflation in the mid-1960s, particularly the governmental fine tuning inspired by the New Economics and the loose financing of the war in Vietnam. Next are considered the factors that led to subsequent strengthening of inflationary forces, including further policy errors, the devaluations of the dollar in 1971 and 1973, the worldwide economic boom of 1972-73, the crop failures and resulting surge in world food prices in 1973-74, the extraordinary increases in oil prices that became effective in 1974, and the sharp deceleration of productivity growth from the late 1960s onward. Finally, attention is turned to the process whereby protracted experience with inflation has led to widespread expectations that it will continue in the future, so that inflation has acquired a momentum of its own.

I have no quarrel with analyses of this type. They are distinctly helpful in explaining the American inflation and, with changes here and there, that in other nations also. At the same time, I believe that such analyses overlook a more fundamental factor: the persistent inflationary bias that has emerged from the philosophic and political currents that have been transforming economic life in the United States and elsewhere since the 1930s. The essence of the unique inflation of our times and the reason central bankers have been ineffective in dealing with it can be understood only in terms of those currents of thought and the political environment they have created…

…the period from World War II to the mid-1960s was marked not only by a dampening of the business cycle but also by persistent increases in the prosperity of American families…

…This experience of economic progress strengthened the public’s expectations of progress. What had once been a quiet personal feeling that the long future would be better than the past, particularly for one’s children, was transformed during the postwar years into an articulate and widespread expectation of steady improvement in living standards—indeed, into a feeling of entitlement to annual increases in real income.

But the rapid rise in national affluence did not create a mood of contentment. On the contrary, the 1960s were years of social turmoil in the United States, as they were in other industrial democracies…

…In the innocence of the day, many Americans came to believe that all of the new or newly discovered ills of society should be addressed promptly by the federal government. And in the innocence of the day, the administration in office attempted to respond to the growing demands for social and economic reform while waging war in Vietnam on a rising scale. Under the rubric of the New Economics, a more activist policy was adopted for the purpose of increasing the rate of economic growth and reducing the level of unemployment…

…The interplay of governmental action and private demands had an internal dynamic that led to their concurrent escalation. When the government undertook in the mid-1960s to address such “unfinished tasks” as reducing frictional unemployment, eliminating poverty, widening the benefits of prosperity, and improving the quality of life, it awakened new ranges of expectation and demand. Once it was established that the key function of government was to solve problems and relieve hardships—not only for society at large but also for troubled industries, regions, occupations, or social groups—a great and growing body of problems and hardships became candidates for governmental solution…

…Many results of this interaction of government and citizen activism proved wholesome. Their cumulative effect, however, was to impart a strong inflationary bias to the American economy. The proliferation of government programs led to progressively higher tax burdens on both individuals and corporations. Even so, the willingness of government to levy taxes fell distinctly short of its propensity to spend. Since 1950, the federal budget has been in balance in only five years. Since 1970, a deficit has occurred in every year. Not only that, but the deficits have been mounting in size. Budget deficits have thus become a chronic condition of federal finance; they have been incurred when business conditions were poor and also when business was booming. But when the government runs a budget deficit, it pumps more money into the pocketbooks of people than it withdraws from their pocketbooks; the demand for goods and services therefore tends to increase all around. That is the way the inflation that has been raging since the mid-1960s first got started and later kept being nourished.

The pursuit of costly social reforms often went hand in hand with the pursuit of full employment. In fact, much of the expanding range of government spending was prompted by the commitment to full employment. Inflation came to be widely viewed as a temporary phenomenon—or, provided it remained mild, as an acceptable condition. “Maximum” or “full” employment, after all, had become the nation’s major economic goal— not stability of the price level. That inflation ultimately brings on recession and otherwise nullifies many of the benefits sought through social legislation was largely ignored…

…And so I finally come to the role of central bankers in the inflationary process. The worldwide philosophic and political trends on which I have been dwelling inevitably affected their attitudes and actions. In most countries, the central bank is an instrumentality of the executive branch of government—carrying out monetary policy according to the wishes of the head of government or the ministry of finance. Some industrial democracies, to be sure, have substantially independent central banks, and that is certainly the case in the United States. Viewed in the abstract, the Federal Reserve System had the power to abort the inflation at its incipient stage fifteen years ago or at any later point, and it has the power to end it today. At any time within that period, it could have restricted the money supply and created sufficient strains in financial and industrial markets to terminate inflation with little delay. It did not do so because the Federal Reserve was itself caught up in the philosophic and political currents that were transforming American life and culture…

…Facing these political realities, the Federal Reserve was still willing to step hard on the monetary brake at times—as in 1966, 1969, and 1974—but its restrictive stance was not maintained long enough to end inflation. By and large, monetary policy came to be governed by the principle of undernourishing the inflationary process while still accommodating a good part of the pressures in the marketplace. The central banks of other industrial countries, functioning as they did in a basically similar political environment, appear to have behaved in much the same fashion.

In describing as I just have the anguish of central banking in a modern democracy, I do not mean to suggest that central bankers are free from responsibility for the inflation that is our common inheritance. After all, every central bank has some room for discretion, and the range is considerable in the more independent central banks. As the Federal Reserve, for example, kept testing and probing the limits of its freedom to undernourish the inflation, it repeatedly evoked violent criticism from both the Executive Branch and the Congress and therefore had to devote much of its energy to warding off legislation that could destroy any hope of ending inflation. This testing process necessarily involved political judgments, and the Federal Reserve may at times have overestimated the risks attaching to additional monetary restraint…

…Monetary theory is a controversial area. It does not provide central bankers with decision rules that are at once firm and dependable. To be sure, every central banker has learned from the world’s experience that an expanding economy requires expanding supplies of money and credit, that excessive creation of money will over the longer run cause or validate inflation, and that declining interest rates will tend to stimulate economic expansion while rising interest rates will tend to restrict it; but this knowledge stops short of mathematical precision…

…It is clear, therefore, that central bankers can make errors—or encounter surprises—at practically every stage of the process of making monetary policy. In some respects, their capacity to err has become larger in our age of inflation. They are accustomed, as are students of finance generally, to think of high and rising market interest rates as a restraining force on economic expansion. That rule of experience, however, tends to break down once expectations of inflation become widespread in a country. At such a time, lenders expect to be paid back in cheaper currency, and they are therefore apt to demand higher interest rates. Since borrowers have similar expectations, they are willing to comply. An “inflation premium” thus gets built into nominal interest rates. In principle, no matter how high the nominal interest rate may be, as long as it stays below or only slightly above the inflation rate, it very likely will have perverse effects on the economy; that is, it will run up costs of doing business but do little or nothing to restrain overall spending. In practice, since inflationary expectations, and therefore the real interest rates implied by any given nominal rate, vary among individuals, central bankers cannot be sure of the magnitude of the inflation premium that is built into nominal rates. In many countries, however, these rates have at times in recent years been so clearly below the ongoing inflation rate that one can hardly escape the impression that, however high or outrageous the nominal rates may appear to observers accustomed to judging them by a historical yardstick, they have utterly failed to accomplish the restraint that central bankers sought to achieve. In other words, inflation has often taken the sting out of interest rates— especially, as in the United States, where interest payments can be deducted for income tax purposes…

…There is a profound difference between the effects of mistaken judgments by a central bank in our age of inflation and the effects of such judgments a generation or two ago. In earlier times, when a central bank permitted excessive creation of money and credit in times of prosperity, the price level would indeed tend to rise. But the resulting inflation was confined to the expansion phase of the business cycle; it did not persist or gather force beyond that phase. Therefore, people generally took it for granted that the advance of prices would be followed by a decline once a business recession got under way. That is no longer the case.

Nowadays, businessmen, farmers, bankers, trade union leaders, factory workers, and housewives generally proceed on the expectation that inflation will continue in the future, whether economic activity is booming or receding. Once such a psychology has become dominant in a country, the influence of a central bank error that intensified inflation may stretch out over years, even after a business recession has set in. For in our modern environment, any rise in the general price level tends to develop a momentum of its own. It stimulates higher wage demands, which are accommodated by employers who feel they can recover the additional costs through higher prices; it results in labor agreements in key industries that call for substantial wage increases in later years without regard to the state of business then; and through the use of indexing formulas, it leads to automatic increases in other wages as well as in social security payments, various other pensions, and welfare benefits, in rents on many properties, and in the prices of many commodities acquired under long-term contracts…

…If my analysis of central banking in the modern environment is anywhere near the mark, two conclusions immediately follow. First, central banks have indeed been participants in the inflationary process in which the industrial countries have been enmeshed, but their role has been subsidiary. Second, while the making of monetary policy requires continuing scrutiny and can stand considerable improvement, we would look in vain to technical reforms as a way of eliminating the inflationary bias of industrial countries. What is unique about our inflation is its stubborn persistence, not the behavior of central bankers. This persistence reflects the fundamental forces on which I dwelt earlier in this address—namely, the philosophic and political currents of thought that have impinged on economic life since the Great Depression and particularly since the mid-1960s…

…The precise therapy that can serve a nation best is not easy to identify, and what will work well in one country may work poorly in another. In the case of the American inflation, which has become a major threat to the well-being of much of the world as well as of the American people, it would seem wise to me at this juncture of history for the government to adopt a basic program consisting of four parts. The first of these would be a legislative revision of the federal budgetary process that would make it more difficult to run budget deficits and that would serve as the initial step toward a constitutional amendment directed to the same end. The second part would be a commitment to a comprehensive plan for dismantling regulations that have been impeding the competitive process and for modifying others that have been running up costs and prices unnecessarily. The third part would be a binding endorsement of restrictive monetary policies until the rate of inflation has become substantially lower. And the fourth part would consist of legislation scheduling reductions of business taxes in each of the next five years—the reduction to be quite small in the first two years but to become substantial in later years. This sort of tax legislation would release powerful forces to improve the nation’s productivity and thereby exert downward pressure on prices; and it would also help in the more immediate future to ease the difficult adjustments forced on many businesses and their employees by the adoption of the first three parts of the suggested program.

2. Two Things I’m Not Worried About – Ben Carlson

Here are two things a lot of other people are worried about but not me:

Stock market concentration. Here’s a chart from Goldman Sachs that shows by one measure, the U.S. stock market is as concentrated as it has ever been:

To which my reply is: So what?

Yes, the top 10 stocks make up more than one-third of the S&P 500. All this tells me is that the biggest and best companies are doing really well. Is that a bad thing?

Stock markets around the globe are far more concentrated than the U.S. stock market. Emerging markets rose to their highest level since June 2022 yesterday. Out of an index that covers 20+ countries, a single stock (Taiwan Semiconductor) accounted for 70% of the move.

Stock market returns over the long run have always been dominated but a small minority of the biggest, best-performing companies…

… Bloomberg is out with a new report that sounds the alarm on U.S. government debt levels:

With uncertainty about so many of the variables, Bloomberg Economics has run a million simulations to assess the fragility of the debt outlook. In 88% of the simulations, the results show the debt-to-GDP ratio is on an unsustainable path — defined as an increase over the next decade.

In the end, it may take a crisis — perhaps a disorderly rout in the Treasuries market triggered by sovereign US credit-rating downgrades, or a panic over the depletion of the Medicare or Social Security trust funds — to force action. That’s playing with fire.

I’ll believe it when I see it.

People have been sounding the alarm on government debt in this country for decades. There has been no panic. No financial crisis. No debt default…

… Interest expense relative to the size of the economy has shot higher in recent years from the combination of more debt and higher rates:

But we’re still well below the highs from the 1980s and 1990s. And when you look at the absolute numbers here, going from 1.5% of GDP to 3% of GDP isn’t exactly the end of the world…

…Debt-to-GDP is now as high as it was in World War II:

That seems scary until you realize in Japan, debt-to-GDP is closer to 300%. I’m not saying we should test our limits but there is no pre-set line in the sand on these things.

3. The inequity method of accounting – Sujeet Indap

The fundamental bargain of M&A seems pretty simple. At the closing of a deal, the buyer pays the seller, and gets a business in return.

It hasn’t been so straightforward for the family who agreed in 2022 to sell its California supermarket Save Mart to the private equity firm Kingswood Capital Management, which valued the grocery chain at $245mn.

Three months after the papers were signed, Kingswood demanded that Save Mart’s prior owners, the Piccinini family, fork back over $109mn after already surrendering the company. In effect, Kingswood wanted to receive a net $77mn payment to take over Save Mart.

And thanks to some ballsy lawyering and nebulous bookkeeping, it seems the PE firm might actually succeed, its gambit upheld by a controversial arbitration ruling in September 2023…

…When Kingswood signed the deal for Save Mart, it was really acquiring two separate businesses. One was the Save Mart grocery chain, comprised of 200 stores and more than $4bn in annual revenue. Save Mart separately held a majority stake in Superstore Industries (SSI), a successful food wholesaler/distributor that had two other owners…

…The two sides agreed that Save Mart’s equity stake in SSI, the joint venture, would be valued at $90mn, a significant step up from the ~$22.5mn value that Save Mart had assigned the investment on its books.

The increase reflected SSI’s valuable land portfolio, according to one person familiar with the transaction. And it enables Kingswood to lower SSI’s tax basis should it ever want to sell SSI, according to a person involved in the transaction.

Those seem reasonable enough. Still, the accounting of SSI’s value is what laid the foundation for this dispute.

For context, a company’s investments can be recorded on its balance sheet in three ways: cost method, equity method, and full consolidation.

Save Mart selected the equity method for its SSI stake.

To explain a bit further: Let’s imagine a company with $100 of asset value and $60 in liabilities, which leaves it with an equity value of $40. Say this company has a 50-per-cent owner, meaning it owns $20 in equity. The owner’s balance sheet would list that $20 as a single line item, called “equity in unconsolidated affiliates”. That account would grow with the subsidiary’s proportional net income, and decrease with any net losses or dividends.

Save Mart’s stake in SSI was listed as a single line on its balance sheet — worth $22.5mn…

…In March 2022, Kingswood and Save Mart closed their deal with the PE firm sending payments based on the family’s proposed accounting. That then set off a final round of post-closing negotiations, where Kingswood got 90 days to argue with the Piccinini’s maths…

…But Kingswood dropped in one massive adjustment with the boilerplate.

It added back $109mn of gross SSI debt, and asserted that the figure counted as official “Indebtedness”. And it argued it should be paid back for all that additional debt.

The PE firm pointed to the language in the deal contract, and said the definition of “Indebtedness” included any Save Mart “group” debt…

…Arbitrator Joseph Slights III, a lawyer in private practice who was formerly a Delaware Vice Chancellor, did not ultimately buy any of what the Piccinnis were selling.

He wrote in the arbitration decision:  “Delaware law is more contractarian than most, and Delaware courts will enforce the letter of the parties’ contract without regard for whether they have struck a good deal or bad deal . . . the result is not absurd or commercially unreasonable.”…

…The Piccinnis, understandably, believe writing a cheque for $109mn is indeed “absurd” and “commercially unreasonable”. They have accused Kingswood of “bad faith” and “gamesmanship” in their court papers.

They will now appeal to the Delaware Supreme Court, pointing to a 2017 decision that said in a post-closing adjustment dispute, the legal system should aim to uphold the broader spirit of the contract instead of narrow contract definitions…

…Kingswood had believed, all along prior to signing and closing, that the gross SSI debt belonged on Save Mart’s main balance sheet. But they decided to keep quiet about that until after the deal closed.

One implication is that they were happy to close on the Piccininis’ terms, and winning on the SSI debt issue would be a bonus, given that there was no guarantee of winning the arbitration.

The firm’s equity check on the $240mn transaction was just $60mn (see the sources and uses table above). If Kingswood is eventually paid the $109mn, it will receive nearly two times their equity contribution by weaponising accounting and legal technicalities.

4. Don’t Be Afraid – Michael Batnick

All-time highs are interesting in the emotions they elicit. Some people might be euphoric as their accounts reach dollar amounts never seen before. Others might fear this is as good as it’s going to get and worry about a trap-door scenario.

Your emotional state might also depend on your asset allocation. If you’re sitting on a large cash pile, it’s understandable that you might be hesitant to go “all in” at a record price. It might not “feel” right.

The good news is the data doesn’t support those feelings. On average since 1970, the S&P 500 has done better 1, 3, and 5 years after making an all-time high than picking a random day.

5. An Interview with Google Cloud CEO Thomas Kurian About Google’s Enterprise AI Strategy – Ben Thompson and Thomas Kurian

You did mention that, “People are moving out of proof-of-concept into actually doing products”. Is that actually happening? What are the actual use cases that companies are actually rolling out broadly as opposed to doing experiments on what might be possible?

TK: Broad-brush, Ben, we can break it into four major categories. One category is streamlining internal processes within the organization, streamlining internal processes. In finance, you want to automate accounts receivable, collections, and cashflow prediction. In human resources, you want to automate your human help desk as well as improve the efficiency with which you can do benefits matching, for example. In procurement and supply chain, you want for example, look at all my suppliers, their contracts with me and tell me which ones have indemnification and warranty protection, so I can drive more volume to those that give me indemnification and warranties and less to those that don’t, for example. These are all practical cases we have customers live in deployment with.

Second is transforming the customer experience. Transforming the customer experiences, how you market, how you merchandise, how you do commerce, how you do sales and service. An example is what Mercedes-Benz CEO Ola Källenius talked about how they’re building a completely new experience for the way that they market and sell and service their vehicles.

Third is that some people are integrating it into their products, and when I say re-imagining their products, re-imagining their core products using AI. We had two examples of companies who are in the devices space. One is Samsung and the other one is Oppo, and they’re re-imagining the actual device itself using AI with all the multimodality that we provide.

There are quite a few companies now re-thinking that if a model can change the way that I see it, that I can process multimodal information. For example, in media we have people saying, “If your model can read as much information as it can, can it take a long movie and shrink it into highlights? Can I take a sports recording of the NCAA basketball final and say, ‘find me all the highlights by this particular player’?” and not have to have a human being sit there and splice the video, but have it do it and I can create the highlights reel really quickly. So there are lots of people re-imagining the product offerings that they have.

And finally, there are some people saying, “With the cost efficiency of this, I can change how I enter a brand new market because, for example, I can do personalized offers in a market where I may not have a physical presence, but I can do much higher conversion rate for customers with online marketing and advertising because now I can do highly tailored campaigns because the cost of creating the content is much lower.” So broad-brush, streamline the core processes and back office, transform the customer experience and it doesn’t mean call centers or chatbots, it can be actually transferring the product itself, transforming the nature of the product you build and enter new markets.

Is it fair to say then when you talk about, “Moving from proof-of-concept to actual production”, or maybe that’s not the words you used, but people are saying, “Okay, we’re going to build this” because this stuff’s not showing up yet, in the real world. Is it the case that, “We see that this could be valuable, now we’re in”, and that’s why you’re emphasizing the platform choice now because they’ve committed to AI broadly, and now it’s like, “Where are we going to build it”?

TK: We have people experimenting, but we also have people actually live deployment and directing traffic. Orange, the telecom company, was talking about how many customers they’re handling online, Discover Financial was talking about how their agents are actually using AI search and AI tools to discover information from policy and procedure documents live. So there are people actually literally running true traffic through these systems and actually using them to handle real customer workload.

Are you seeing the case in a lot of in customers, or maybe you’re hearing from potential customers, that AI is rolling out, if that’s the right word, in an employee arbitrage situation? Where there’s individual employees that are taking on themselves to use these tools and they are personally benefiting from the increased productivity — maybe they’re doing less work or maybe they’re getting more done — and the companies want to capture that more systematically. Is that a theme that you’re seeing?

TK: We’re seeing three flavors. Flavor one is a company has, we’re going to try eight or nine, what they call customer journeys or use cases, we’re going to pick the three that we see as the maximum return, meaning value and value does not mean cost savings always. It could be, for example, we have one who is handling 1 million calls a day through our customer service system. Now a million calls a day, if you think about it, Ben, an average person can do about 250 calls a day, that’s a certain volume in an eight-hour day. If you handled a million, that is a lot of people, so the reality is that several of them were not being answered and people never called because the wait time was so long. So in that case, it was not about cost savings, it’s the fact that they’re getting able to reach many more customers than they could do before. So that’s one. One part is people saying, “I have a bunch of scenarios, I’m going to pick the three”, and in many cases, they’re actually augmenting something they’re doing or doing something they couldn’t do before, that’s scenario one.

Scenario two was I have, for example, there’s a large insurance company that’s working with us. Today, when they do claims and risk calculation, it takes a long time to handle the claims and the risk, particularly the risk calculation, because there’s thousands of pages of documents, there’s a lot of spreadsheets going back and forth. They put it into Gemini and it was able to run the calculations much, much more quickly. So second is I’m picking a very high value use case for my organization, which is the core function, and I’m going to implement it because I can get a real competitive advantage. In their case, it’s the fact that they can both get more accurate scoring on the risk and they can also do a much more accurate job, faster job in responding.

And the third scenario is what you said. “Hey, we’ve got a bunch of people, we’re going to give it to a certain number of developers”. For example, our coding tool, “They are going to test it, they say it helps me generate much better unit tests, it helps me write better quality code”. Wayfair’s CTO was talking about what their experience is, and then they say, “Let’s go broadly”, so all three patterns are being seen…

Do you see AI, though, in all this talk about, “You need to choose a platform? Sure, our platform’s going to be open, you can use it anywhere” — but do you see this as a wedge to be like, “Okay, this is a reboot broadly for the industry as far as cloud goes, and sure, your data may be in AWS, or in Azure, or whatever it might be, but if you have a platform going forward, you should start with us”? Then maybe we’ll look up in ten, fifteen years, and all the center of gravity shifted to wherever the platforms are?

TK: For sure. I mean, it’s a change in the way that people make purchase decisions, right? Ten years ago, you were worried about commodity computing, and you were like, “Who’s going to give me the lowest cost for compute, and the lowest cost for storage, and the lowest cost for networking?”. Now the basis of competition has changed and we have a very strong position, given our capability both at the top, meaning offering a platform, offering models, et cetera, and building products that have long integrated models.

Just as an example, Ben, integrating a model into a product is not as easy as people think; Gmail has been doing that since 2015. On any daily basis, there are over 500 million operations a day that we run and to do it well, when a partner talked about the fact that 75% of people who generate an image for slides actually end up presenting it, it’s because we have paid a lot of attention over the years on how to integrate it.

So we play at the top of the stack, and we have the infrastructure and scale to do it really well from a cost, performance, and global scale that changes the nature of the competition. So we definitely see this, as you said, as a reset moment for how customers thinking of choosing their cloud decision.

If you’re talking about a lot of choices about models, and customers were over-indexed on choosing the correct model, that implies that models are maybe a commodity, and that we’ve seen with GPT-4 prices are down something like 90% since release. Is that a trend you anticipate continuing, and is it something that you want to push and actually accelerate?

TK: Models — whether they’re a commodity or not, time will tell, these are very early innings. All we’re pointing out is every month, there’s a new model from a new player, and the existing models get better on many different dimensions. It’s like trying to pick a phone based on a camera, and the camera’s changing every two weeks, right? Is that the basis on which you want to make your selection?

Well, but if you make that basis, then you might be locked into the operating system.

TK: That’s right, and so that’s why we say you should choose an open platform, and you should be able to use a collection of different models, because it’s changing, and don’t lock into a particular operating system at a time when the applications on top of it are changing, to use your analogy.

Why is your platform open as compared to others? Microsoft has announced you can use other models, not just OpenAI models. Amazon is sort of, to the extent you can ascertain a strategy, it’s like, “Look, we’re not committing to anything, you could do whatever you want.” Why do you feel comfortable saying, “No, we’re the open one,” and they’re not?

TK: Well, first of all, the completeness of our platform; Vertex has a lot more services than you can get with the other platforms. Secondly, in order to improve a platform, you have to have your own model, because there’s a bunch of things you do when you engineer services with that model.

I’ll give you a really basic example. You use a model, you decide to ground the answers. Grounding improves quality, but can also introduce latency. How do you make sure that when you’re grounding, you’re not serially post-processing a model’s answer to add latency? Unless you have your own model, you wouldn’t even get to that. So because we have our own model, we’re able to engineer these things, but we make them available as services with other models, so you can use enterprise grounding as a very specific example. There are lots of customers using it with Mistral and with Llama and with Anthropic.

Second thing, we are not just offering models, but we’re actually helping the third party go to customers with us. I met a lot of customers today jointly with [CEO] Dario [Amodei] from Anthropic, and it’s a commitment to make sure we’re not just giving you our infrastructure, we’re not just training, integrating a model into Vertex, we’re not just making it a first-class model, but we’re actually bringing it to clients together.

I think that’s what we mean by open. One of the other players has no models of their own, so naturally they’re offering a bunch of models, and the other player has outsourced their model development to a third party…

How important is that million context window in the story you are telling? My perception is, there’s a lot of stuff you could do if you build a lot of infrastructure around it, whether it be RAG or other implementations, but it feels like with Gemini 1.5 there are jack-of-all-trades possibilities that seem to open up to a much greater extent, and there’s a bit where, you had that compliance bit, the statements of work and they had to compare it to the 100-page compliance document. I got some comments like, “Maybe companies shouldn’t have 100-page compliance notebooks or whatever it might be”, but the reality is, that’s the case, the world has that. My perception of the keynote is, that was the killer feature, that seemed to undergird everything. Was that the correct perception?

TK: Yeah, there are two reasons. Just to be perfectly clear, Ben, the long context window allows you to do three things that are important. First of all, when you look at high definition video, for example, and other modalities, and just imagine you’re dumping a high definition video in and you want to create out of the NCAA final, which just happened, the highlight reel but you don’t want to specify every attribute about what you want spliced into the highlight reel. The model has to digest it and because it has to process it, it’s a fairly dense representation of the video because there are objects, there are people moving, there are actions, like I’m throwing a pass. They could be, I have my name on the back of my t-shirt, there could be a score like, “When did they change from 24 to 26 points? Did they score three pointers?”, so there are many, many, many dimensions. So reasoning becomes a lot better when you can take a lot more context, that’s one, and it’s particularly true of modality.

The second is, today people don’t use models to maintain state or memory, meaning they ask it a question, the next time they think, “Hey, it may not remember”, so when you’re able to maintain a longer context, you can maintain more state, and therefore you can do richer and richer things rather than just talk back-and-forth with a very simplistic interface. You see what I mean?

The third thing is, there are certainly complex scenarios, it’s the unfortunate reality, there’s lots of policies and procedure books that are even longer than what we showed, and so there are scenarios like that that we have to be able to deal with. But in the longer term, the real breakthrough is the following. Context length, if you can decouple the capabilities of the model and the latency to serve a model from the context length, then you can fundamentally change how quickly you can scale a model.

Is this ultimately, from your perspective, a question of infrastructure, and that just leans into Google’s biggest advantage?

TK: It’s a question of global infrastructure, but also optimizations at every layer in the infrastructure, which we can co-engineer with DeepMind…

Sundar Pichai mentioned in his video greeting, he emphasized the number of AI startups, and particularly AI unicorns using Google Cloud. To go back to the reboot idea, do you view the AI Era as a restart in terms of capturing the next generation of companies? I mean, obviously, AWS had a huge advantage here as far as general cloud computing, the entire mobile app ecosystem was by and large built on AWS. In the enterprise era, you have to deal with what’s there, what they’ve already dealt with, you have to have the integrations. Do you see yourself as having this as a big focus, “We’re going to own this era of startups”?

TK: Yes. And by the way, every one of those startups is being pursued by the other two, and the fact that 90% of the unicorns and 60% of all AI-funded startups, up in each case by ten points in eight months, and they are the most discerning ones. I mean, just to be frank, the unicorns, for them, it is the really biggest cost of goods sold in their P&L.

So what’s the driver there?

TK: The efficiency of our infrastructure.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google), Amazon (parent of AWS), Microsoft, and TSMC. Holdings are subject to change at any time.

What We’re Reading (Week Ending 14 April 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 14 April 2024:

1. Perplexity is ready to take on Google – Alex Heath and Aravind Srinivas

What’s it like on the frontlines of the AI talent war right now?

I made mistakes in chasing the wrong people. Recently there was a really senior backend engineer who ended up joining X.AI. He was talking to us, too.

I was talking to Patrick Collison for advice on this, and he said, “Why are you even in this race? Why are you trying to compete with these people? Go after people who want to actually build the stuff that you’re building and don’t chase AI clout.”

There are a lot of good engineers who are applying to us and Anthropic and OpenAI and X.AI and Character.ai. These are the top five choices of AI startups. And people normally just go to the highest bidder. Whoever has the highest valuation will be able to win this race all the time because, on paper, you’re always going to be able to offer the same amount of shares but the dollar value is going to be much higher…

...Have you taken any kind of lesson away from the Gemini diversity scandal? I saw you recently integrated photo generation into Perplexity.

Factfulness and accuracy is what we care about. Google has many other cultural things that they care about, and that’s why they made their products that way. They should only prioritize one aspect, which is giving an accurate answer. They don’t do that for whatever reasons. They have all these other people in the room trying to make decisions.

If I learned one thing, it’s that it’s better to be neutral. Don’t try to have any values you inject into the product. If your product is an answer engine, where people can ask questions and get answers, it better respond in a scholarly way. There’s always a nerd in your classroom who’s just always right, but you don’t hate them for having a certain political value, because they are just going to give you facts. That’s what we want to be. And Google’s trying to be something different. That’s why they got into trouble.

What are you hearing generally about the state of Google from people there right now?

The researchers are still pretty excited about what they’re doing. But the product team messes up their releases. The Gemini product team was fine-tuning all these models to put in the product. There’s a lot of bureaucracy, basically.

I know Sergey Brin being there is making things faster and easier for them. You might have seen the video that was circulating of him being at some hackathon. He brushed it [the Gemini diversity scandal] off as just some kind of a small bug, right?

It’s not a small bug. It’s actually poor execution. The image generation thing is actually very easy to catch in testing. They should have caught it in testing. When you consider Google as the place for de facto information and correctness, when they make mistakes it changes the way you perceive the company…

How much of your tech is in-house versus fine-tuning all these models that you work with? What’s your tech secret sauce?

In the beginning, we were just daisy-chaining GPT-3.5 and Bing. Now, we post-train all these open-source models ourselves. We also still use OpenAI’s model.

We are never going to do the full pre-training ourselves. It’s actually a fool’s errand at this point because it takes so much money to even get one good model by pre-training yourself. There are only four or five companies that are capable of doing that today. And when somebody puts out these open-source models, there’s no reason for you to go and recreate the whole thing.

There is a new term that has emerged in this field called post-training. It’s actually like fine-tuning but done at a much larger scale. We are able to do that and serve our models ourselves in the product. Our models are slightly better than GPT-3.5 Turbo but nowhere near GPT-4. Other than Anthropic and Gemini, nobody has actually gotten to that level yet.

How are you doing to solve AI hallucination in your product? Can you?

The reason why we even have sources at the top of the answer is because we want to make sure that users have the power to go verify the answer. We precisely tell you which link to go to versus showing ten blue links and you not being sure which to read.

The other way is constantly improving the authority of which sources we use to cite the answer and then getting rid of the bad ones. When you don’t have sufficient information, it’s better to say you don’t know rather than saying something you made up.

2. Book Summary: Our Investing Strategy, who does the market smile upon – Made In Japan

He goes by the name of Tatsuro Kiyohara, who was the CIO of Tower Investment Management, which ran the flagship K-1 fund that compounded 20% annually during his 25-year run (that’s 9300%). Compare this to the TOPIX which did an annualized return of roughly 3%.

But its not just the numbers that he posted that were inspiring, the journey to get there was a tumultuous one that would be almost impossible for us to replicate. He is built differently. Who else is willing to pour in almost their entire net worth when the fund is down -72% in an attempt to save his fund not just for his sake, but for the clients that decided to stick with him amid all the redemptions?..

…During his stint in New York, his clients included the familiar hedge funds we’d all heard of. One of which was Julian Robertson’s Tiger Management. Meeting Tiger was perhaps the first instance he decided that he wanted to do something with hedge funds, he would spend his days talking stock at their office. Tiger appreciated him too – one time he realized that Tiger was short a stock, Kawasaki Steel, and also realized Nomura was attempting to ‘promote’ the stock (what you call ‘pump’ these days), he almost had to fight with Tiger to convince exiting and that the stock was going up regardless of fundamentals which they finally obliged. This stock 3xed not long after. Not a surprise that he was invited to Tiger’s annual party to be awarded best salesman of the year…

…The game is to figure out when you’re in the minority. This doesn’t always mean that because everyone is bullish, that you being bullish isn’t a variant perception. If for example the market expects a company to grow by 10% a year for the next 5 years, and you believe it will be more like 30%, you are still in the minority.

It is more difficult to thus have a good investment idea in large caps because its harder to figure out what’s priced in.

From an opportunity cost standpoint for an individual investor, the return on time is so much better researching cheap micro & small-caps with the potential to grow. If you had an hour with the CEO of a large company or a small one the likelihood you will get more valuable insights (your alpha) is much higher in the latter. In general, you also won’t lose much if you’re wrong here, these companies aren’t priced for growth anyway. For him personally, he tries to avoid investing in stocks that have already gone up as much as possible…

…One of his first investments with the fund that falls into this archetype was Nitori, which is a household name today but when he first invested nobody touched it. It was a Hokkaido-based company, and the furniture market in Japan was shrinking. What he realized though was that the market was very fragmented, and he saw Nitori as the one to take market share over time with an exceptional founder. Which proved to be correct. His investment in NItori 10xed all the while the market halved. The lesson here was that even with a shrinking market if you find the right company, you can generate strong returns. No doubt there are some diamonds in the rough…

…In the end, 90% of investing in Small/Micro-caps is about Management
Heres what he looks for:

  1. Operators with a growth mindset
  2. Talented employees that are aligned with the CEO’s Vision and Mission
  3. Doesn’t get crushed by competitors
  4. A widening core competence (read = Moat) as it grows
  5. Not in an industry where you’re pulling forward demand (There is a finite pile of demand that once absorbed will be gone, he points out the Japanese M&A industry as one)
  6. The management is impeccable with their word that is, they do what they say
  7. Companies that have positive feedback loops…

…With one company, every time he requested a meeting with Investor Relations, the CEO showed up every time without fail which he found strange. Eventually, though this business got caught in an accounting scandal and went under. Maybe if the CEO shows up too readily you need to be careful. Another business had zero interest in doing IR, showing up in their factory uniform, and wasn’t too friendly. One day however they show up in suits! This business also didn’t too well…

…There are mavericks among them. Founder of Zensho Holdings (the operator of Beef bowl chain Sukiya). This was an unpopular stock that IPOd. The first thing he saw when visiting the founder’s office was a bench press. He was ‘ripped like Popeye’. At the meeting, all he talked about was how superior a food a ‘beef bowl’ (Gyuudon) was. If Japanese people had eaten enough beef bowls and benchpressed enough Japan wouldn’t have lost the war (LOL).

One time one of Kiyohara-sans employees told this founder he’s also been going to the gym, he immediately challenged said employee and tested him with a ‘lariat’ a type of wrestling tackle. The key is to find a CEO who knows his wrestling moves.

The IR was also interesting. In its mid-term plan, they even included the P/E multiple the company should be trading at in 5 years…

…As he once reflected on his portfolio, he realized that the business and its management was like looking at himself in the mirror. If assessing management correctly is the key to investing, also understand that there is a self-selection bias. If the Founder and CEO is that much more brilliant than you – you won’t even realize how brilliant he is. That said you’d never invest in a ‘dumb’ CEO, so ultimately you end up selecting people ‘on your level’. This it appears, to be the reality for investing in microcaps…

…He was adamant about that whether large or small – he had no interest in buying an expensive business. If the P/E was too high, that was a pass.

Over the years he tried building various growth models and realized this had almost no benefit to making money so just stopped. Was a waste of time.

He screens for such companies by finding a high net cash ratio which was just net cash over market cap. (So basically net-nets).

He also liked to invert the problem, by looking at the current P/E of the stock you can figure out the kind of earnings growth it implied. No rocket science here – he tries to figure out the Terminal Multiple with a Perpetual Growth model. For example, if the risk-free rate is 2% and the P/E is 10x that implies a terminal growth of -8.2% all else equal if earnings growth was -3.1% instead then the P/E should be 20x. (Yes this is all negative growth)

The 40-year average of the risk-free rate is about 1.7% in Japan so that sounds fair to him. Also, he says, forget about the concept of equity risk premium – this is just a banker’s term to underwrite uncertainty. If you’re uncertain just model that into your earnings projections…

…He doesn’t look at P/B where Investors may be calculating the liquidation price which can be inaccurate.

The point he thinks many miss – if a company is loss-making would the hypothetical buyer buy the assets at face value? And say that the business is a decent profitable business – no one’s going to be looking at the P/B they’ll be fixated on the P/E…

…Risks of small/micro caps:

  • Illiquidity discount
  • Many businesses are small suppliers to a much larger company
  • It operates in an industry with low entry barriers
  • Limited talent within the organization
  • Succession issues and nepotism – the son of the owner can be a dumb ass
  • Because no one notices them – the likelihood of a fraud/scandal is higher
  • When the owner retires, this person may pay out a massive retirement bonus
  • Because it’s an owner-operator and harder to take over, no one keeps them in check and may screw up
  • When there is an accounting fraud, the damage will be large
  • They don’t have the resources to expand overseas
  • They have an incentive to keep their valuation low to minimize the inheritance tax.

3. Three reasons why oil prices are remarkably stable – The Economist

Shouldn’t oil prices be surging? War has returned to the Middle East. Tankers in the Red Sea—through which around 12% of seaborne crude is normally shipped—are under attack by Houthi militants. And opec, a cartel of oil exporters, is restricting production. Antony Blinken, America’s secretary of state, has invoked the spectre of 1973, when the Yom Kippur war led to an Arab oil embargo that quadrupled prices in just three months. But oil markets have remained calm, trading mostly in the range of $75 and $85 per barrel for much of last year…

…Oil production is now less concentrated in the Middle East than it has been for much of the past 50 years. The region has gone from drilling 37% of the world’s oil in 1974 to 29% today. Production is also less concentrated among members of OPEC… That is partly because of the shale boom of the 2010s, which turned America into a net energy exporter for the first time since at least 1949…

…Another reason for calm is opec members’ ample spare production capacity (ie, the amount of oil that can be produced from idle facilities at short notice)…

…America’s Energy Information Administration (eia) estimates that opec’s core members have around 4.5m barrels per day of spare capacity—greater than the total daily production of Iraq…

…The world still has a big appetite for oil: according to the eia demand hit a record in 2023 and will be higher still in 2024, thanks in part to growth in India. But that is unlikely to push prices much higher. Global growth is not at the levels seen in the early 2000s. China, long the world’s biggest importer of oil, is experiencing anaemic economic growth. Structural changes to its economy also make it less thirsty for the stuff: next year, for example, half of all new cars sold in the country are expected to be electric.

4. How We’ll Reach a 1 Trillion Transistor GPU – Mark Liu and H.S. Philip Wong

All those marvelous AI applications have been due to three factors: innovations in efficient machine-learning algorithms, the availability of massive amounts of data on which to train neural networks, and progress in energy-efficient computing through the advancement of semiconductor technology. This last contribution to the generative AI revolution has received less than its fair share of credit, despite its ubiquity.

Over the last three decades, the major milestones in AI were all enabled by the leading-edge semiconductor technology of the time and would have been impossible without it. Deep Blue was implemented with a mix of 0.6- and 0.35-micrometer-node chip-manufacturing technology. The deep neural network that won the ImageNet competition, kicking off the current era of machine learning, was implemented with 40-nanometer technology. AlphaGo conquered the game of Go using 28-nm technology, and the initial version of ChatGPT was trained on computers built with 5-nm technology. The most recent incarnation of ChatGPT is powered by servers using even more advanced 4-nm technology. Each layer of the computer systems involved, from software and algorithms down to the architecture, circuit design, and device technology, acts as a multiplier for the performance of AI. But it’s fair to say that the foundational transistor-device technology is what has enabled the advancement of the layers above.

If the AI revolution is to continue at its current pace, it’s going to need even more from the semiconductor industry. Within a decade, it will need a 1-trillion-transistor GPU—that is, a GPU with 10 times as many devices as is typical today…

…Since the invention of the integrated circuit, semiconductor technology has been about scaling down in feature size so that we can cram more transistors into a thumbnail-size chip. Today, integration has risen one level higher; we are going beyond 2D scaling into 3D system integration. We are now putting together many chips into a tightly integrated, massively interconnected system. This is a paradigm shift in semiconductor-technology integration.

In the era of AI, the capability of a system is directly proportional to the number of transistors integrated into that system. One of the main limitations is that lithographic chipmaking tools have been designed to make ICs of no more than about 800 square millimeters, what’s called the reticle limit. But we can now extend the size of the integrated system beyond lithography’s reticle limit. By attaching several chips onto a larger interposer—a piece of silicon into which interconnects are built—we can integrate a system that contains a much larger number of devices than what is possible on a single chip…

…HBMs are an example of the other key semiconductor technology that is increasingly important for AI: the ability to integrate systems by stacking chips atop one another, what we at TSMC call system-on-integrated-chips (SoIC). An HBM consists of a stack of vertically interconnected chips of DRAM atop a control logic IC. It uses vertical interconnects called through-silicon-vias (TSVs) to get signals through each chip and solder bumps to form the connections between the memory chips. Today, high-performance GPUs use HBM extensively…

…With a high-performance computing system composed of a large number of dies running large AI models, high-speed wired communication may quickly limit the computation speed. Today, optical interconnects are already being used to connect server racks in data centers. We will soon need optical interfaces based on silicon photonics that are packaged together with GPUs and CPUs. This will allow the scaling up of energy- and area-efficient bandwidths for direct, optical GPU-to-GPU communication, such that hundreds of servers can behave as a single giant GPU with a unified memory. Because of the demand from AI applications, silicon photonics will become one of the semiconductor industry’s most important enabling technologies…

…We can see the trend already in server GPUs if we look at the steady improvement in a metric called energy-efficient performance. EEP is a combined measure of the energy efficiency and speed of a system. Over the past 15 years, the semiconductor industry has increased energy-efficient performance about threefold every two years. We believe this trend will continue at historical rates. It will be driven by innovations from many sources, including new materials, device and integration technology, extreme ultraviolet (EUV) lithography, circuit design, system architecture design, and the co-optimization of all these technology elements, among other things.

Largely thanks to advances in semiconductor technology, a measure called energy-efficient performance is on track to triple every two years (EEP units are 1/femtojoule-picoseconds).

In particular, the EEP increase will be enabled by the advanced packaging technologies we’ve been discussing here. Additionally, concepts such as system-technology co-optimization (STCO), where the different functional parts of a GPU are separated onto their own chiplets and built using the best performing and most economical technologies for each, will become increasingly critical.

5. The illusion of moral decline – Adam Mastroianni

In psychology, anything worth studying is probably caused by multiple things. There may be lots of reasons why people think morality is declining when it really isn’t.

  • Maybe people say that morality is declining because they think it makes them look good. But in Part I, we found that people are willing to say that some things have gotten better (less racism, for instance). And people still make the same claims when we pay them for accuracy.
  • Maybe because people are nice to you when you’re a kid, and then they’re less nice to you when you’re an adult, you end up thinking that people got less nice over time. But people say that morality has declined since they turned 20, and that it’s declined in the past four years, and all that is true for old people, too.
  • Maybe everybody has just heard stories about how great the past is—like, they watch Leave It to Beaver and they go “wow, people used to be so nice back then.” But again, people think morality has declined even in the recent past. Also, who watches Leave It to Beaver?
  • We know from recent research that people denigrate the youth of today because they have positively biased memories of their own younger selves. That could explain why people blame moral decline on interpersonal replacement, but it doesn’t explain why people also blame it on personal change.

Any of these could be part of the illusion of moral decline. But they are, at best, incomplete.

We offer an additional explanation in the paper, which is that two well-known psychological phenomena can combine to produce an illusion of moral decline. One is biased exposure: people pay disproportionate attention to negative information, and media companies make money by giving it to us. The other is biased memory: the negativity of negative information fades faster than the positivity of positive information. (This is called the Fading Affect Bias; for more, see Underrated ideas in psychology).

Biased exposure means that things always look outrageous: murder and arson and fraud, oh my! Biased memory means the outrages of yesterday don’t seem so outrageous today. When things always look bad today but brighter yesterday, congratulations pal, you got yourself an illusion of moral decline.

We call this mechanism BEAM (Biased Exposure and Memory), and it fits with some of our more surprising results. BEAM predicts that both older and younger people should perceive moral decline, and they do. It predicts that people should perceive more decline over longer intervals, and they do. Both biased attention and biased memory have been observed cross-culturally, so it also makes sense that you would find the perception of moral decline all over the world.

But the real benefit of BEAM is that it can predict cases where people would perceive less decline, no decline, or even improvement. If you reverse biased exposure—that is, if people mainly hear about good things that other people are doing—you might get an illusion of moral improvement. We figured this could happen in people’s personal worlds: most people probably like most of the people they interact with on a daily basis, so they may mistakenly think those people have actually become kinder over time.

They do. In another study, we asked people to answer those same questions about interpersonal replacement and personal change that we asked in a previous study, first about people in general, and then about people that they interact with on a daily basis. When we asked participants about people in general, they said (a) people overall are less moral than they were in 2005, (b) the same people are less moral today than in 2005 (personal change) and (c) young people today are less moral than older people were in 2005 (interpersonal replacement). Just as they did before, participants told us that morality declined overall, and that both personal change and interpersonal replacement were to blame.

But we saw something new when we asked participants about people they know personally. First, they said individuals they’ve known for the past 15 years are more moral today. They said the young folks they know today aren’t as moral as the old folks they knew 15 years ago, but this difference was smaller than it was for people in general. So when you ask people about a group where they probably don’t have biased exposure—or at least not biased negative exposure—they report less moral decline, or even moral improvement.

The second thing that BEAM predicts is that if you turn off biased memory, the illusion of moral decline might go away. We figured this could happen if you asked people about times before they were born—you can’t have memories if you weren’t alive. We reran one of our previous studies, simply asking participants to rate people in general today, the year in which they turned 20, the year in which they were born, 20 years before that, and 40 years before that.

People said, basically, “moral decline began when I arrived on Earth”:

Neither of these studies mean that BEAM is definitely the culprit behind the illusion of moral decline, nor that it’s the only culprit. But BEAM can explain some weird phenomena that other accounts can’t, and it can predict some data that other accounts wouldn’t, so it seems worth keeping around for now.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google) and TSMC. Holdings are subject to change at any time.

What We’re Reading (Week Ending 07 April 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 07 April 2024:

1. China’s capitalist experiment – Michael Fritzell

I just read a great new book by analyst Anne Stevenson-Yang. It’s called Wild Ride and is available for pre-order on Amazon.

The book tells the story of China’s economic miracle from the late 1970s until today - how Deng Xiaoping’s reforms unleashed a wave of entrepreneurship and led to China’s economy becoming one of the largest in the world.

However, it also discusses some of the system’s fragilities and how the country now seems to be turning inwards again…

…China under Mao Zedong was a closed-off, repressive society. Meat was a once-in-a-week luxury. Cooking was done outside. And personal freedoms were more or less non-existent…

…After Mao died in 1976, a power struggle ensued. Ultimately, Mao’s former ally, Deng Xiaoping, emerged victorious from this struggle. One of his first tasks was to open up the economy to the outside world. For this, he would need hard currency.

Practical considerations took priority in those early years. When Deng Xiaoping travelled to the United States in 1979, he ordered an inventory of all hard currency in China’s banks. He came up with only US$38,000 - hardly enough to pay for his delegation.

This was a low point for the Chinese economy. Deng recognized that China needed exports. Japan, Korea, and Taiwan became wealthy by promoting the export of manufactured goods. So Deng adopted a twin strategy of promoting exports in special economic zones while shielding ordinary Chinese from foreign cultural influences…

..Deng’s special economic zones were newly incorporated entities acting as quasi-governments. What made them different was that their managers were rewarded by meeting targets focused on the scale of capital investment and gross tax revenues…

…Initially, foreign influence was kept at bay. Foreign nationals were required to live in special compounds, use separate medical facilities, and even use special currencies. Romantic relationships between foreigners and Chinese were forbidden as well.

The special economic zones in the Southern parts of the Guangdong province, such as Shenzhen, were particularly successful. One of the reasons was that they were near port facilities. But perhaps even more importantly, they had access to financial powerhouse Hong Kong, with its banks and talented entrepreneurs. While, of course, having access to hundreds of millions of workers from inland provinces.

In fact, Shenzhen became a model for the China that was about to develop. It was the first city to abolish the food coupon system, thus allowing residents to buy food with their own money. And residents were soon allowed to lease their own land…

…Another important part of Deng’s reforms was allowing farmers to grow whatever they pleased after meeting some quota. They could then sell any surplus in newly established markets. This unleashed immense rural income growth of 12% per year throughout the 1980s.

A similar system was later introduced to state-owned enterprises as well. They were now allowed to retain profits, either for reinvestment or pay them out as bonuses to employees. Managers suddenly realized they had incentives to increase revenues and profits, and some became wealthy…

…But beneath the surface, discontent was growing. Students were devouring books brought in from overseas. They were clamoring not only for economic gains but also for political reforms. By 1987, Beijing students regularly held marches from the university districts to Tian’anmen Square to protect against political restrictions…

…The crackdown on the student demonstrations in Beijing in June 1989 led to a significant political shift. For two years after the massacre, the country closed off, and dissidents were hunted down and jailed. Anyone who participated in the protests was either disappeared, jailed, demoted or unable to attend university or get a good job.

After the student protests, the Communist Party shifted its strategy to maintaining control. It upped its propaganda efforts, conveying that if the party were to collapse, China would end up in total anarchy.

In the aftermath of Tian’anmen, a communication system was established that improved the party’s control over the provinces. Tax collection and audits were tightened, and a criminal detection and surveillance system was developed…

…One of Deng’s buzzwords during this era was “to get rich is glorious” (致富光荣). You no longer had to be ashamed of pursuing wealth; it was promoted from the top down.

The Communist Party bet that as long as people felt their livelihoods improved, they would not rock the boat. The restive students who protested at Tian’anmen Square would now focus on economic opportunity rather than spiritual dissatisfaction.

After his come-back in the early 1990s, Deng picked out young talent Zhu Rongji to push for further reforms. In a long list of achievements, Zhu Rongji managed to:

  • Cut the government bureaucracy in half
  • Privatize housing
  • Sell off 2/3 of the companies in the state sector
  • Unify the dual currencies used prior to 1994
  • Introduce a nationwide tax system
  • Take control of the appointment of all provincial-level governors…

…After the reforms of the 1990s, China’s economic growth really took off. Exporters in China’s coastal regions benefitted from the country’s admission into the WTO, and Chinese returnees started businesses left and right…

…It was also during the 2000s that the property boom really kicked into high gear. In the late 1990s, Zhu Rongji instituted reforms that allowed state-owned enterprises to sell worker housing back to tenants for a pittance. As prices rose throughout the 2000s, tenants now held significant household equity, which they could then leverage to buy new, even fancier, commodity housing.

A change in the tax structure also incentivized local governments to promote construction. In the mid-1990s, the central government established its own offices to collect taxes directly. In other words, local governments had less ability to raise taxes themselves, instead relying on remittances from the central government. Local governments thus became cash-poor.

To fund their spending programs, they instead set up local government financing vehicles (LGFVs), which used land as collateral for borrowing. And since they were government entities, they were seen as quasi-sovereign borrowers enjoying full access to loans from state banks. Over time, the number of LGFVs grew to over 10,000. They operate urban infrastructure, subway systems, water and gas utilities, etc. Some of them are profitable, but many of them are not…

…The privatization of China’s housing market, which provided collateral for new loans, created one of the biggest credit booms the world has ever seen. Later on, in just five years, more credit was created than the entire value of the US banking system…

…After the Great Financial Crisis of 2008, the Communist Party leadership unleashed a CNY 4 trillion stimulus program that brought forward demand for infrastructure and spending targets.

At this point, it was already becoming clear that the capital stock for infrastructure was starting to exceed those of most other developing or even developed economies. By 2012, China had 8x the length of highways per unit of GDP as that of Japan. At the time, more than 70% of China’s airports were failing to cover their own costs, even though such costs tend to be modest…

…Meanwhile, with the state pushing for big stimulus packages, the government increasingly directed economic resources. Concepts such as “advance of the state, retreat of the private sector” (国进民退) became more common, reflecting a shift in the economy away from private sector entrepreneurship…

…And indeed, with the emergence of Xi Jinping, the state has started to reassert control. State companies are now receiving most of the loans from China’s banks. State media is now talking of “national rejuvenation”, trying to unite the country around nationalist sentiment and acceptance of a “moderately prosperous lifestyle” (小康社会). This is a clear break from the era of Deng Xiaoping’s reforms when getting rich was perhaps the greatest virtue in life…

…Further, she believes that a Russia-Iran-China bloc is currently being formed and that China’s financial system could serve as a bedrock for trade within the bloc:

“If, however, China were someday to shrink its network of trading partners to other dictatorships like Russia and North Korea, its dedicated financial system could become the principal one used for trade among those nations.”

In other words, Anne believes that China is withdrawing from its informal pact with Western nations about open trade, with the experiment in Western-style capitalism that commenced in 1979 over. The Chinese economy is now morphing into a different system, one where the state reigns supreme and will become an influential partner in a new trading bloc formed by China’s current geopolitical allies.

2. 20 Lessons From 20 Years of Managing Money – Ben Carlson

1. Experiences shape your perception of risk. Your ability and need to take risk should be based on your stage in life, time horizon, financial circumstances and goals.

But your desire to take risk often trumps all that, depending on your life experiences. If you worked at Enron or Lehman Brothers or AIG or invested with Madoff, your appetite for risk will be forever altered.

And that’s OK as long as you plan accordingly.

2. Intelligence doesn’t guarantee investment success. Warren Buffett once wrote, “Investing is not a game where the guy with the 160 IQ beats the guy with the 130 IQ. Once you have ordinary intelligence, what you need is the temperament to control the urges that get other people into trouble in investing.”

I’ve met so many highly educated individuals who are terrible investors. They can’t control their emotions because their academic pedigree makes them overconfident in their abilities.

Emotional intelligence is the true sign of investment smarts.

3. No one lives life in the long-term. Long-term returns are the only ones that matter but you have to survive a series of short-terms to get there.

The good strategy you can stick with in those short-terms is preferable to the perfect strategy you can’t stick with…

9. The biggest risks are always the same…yet different. The next risk is rarely the same as the last risk because every market environment is different.

On the other hand, the biggest mistakes investors make are often the same — timing the market, recency bias, being fearful when others are fearful and greedy when others are greedy and investing in the latest fads.

It’s always a different market but human nature is the constant…

16. Experience is not the same as expertise. Just because you’ve been doing something for a long time doesn’t mean you’re an expert.

I know plenty of experienced investors who are constantly fighting the last war to their own detriment.

How many people who “called” the 2008 crash completely missed the ensuing bull market? All of them?

How many investment legends turn into permabears the older they get becasue they fail to recognize how markets have changed over time?

Loads of investment professionals who have been in the business for many years make the same mistakes over and over again…

18. There is a big difference between rich and wealthy. Lots of rich people are miserable. These people are not wealthy, regardless of how much money they have.

There are plenty of people who wouldn’t be considered rich based on the size of their net worth who are wealthy beyond imagination because of their family, friends and general contentment with what they have.

19. Optimism should be your default. It saddens me to see an increasing number of cynical and pessimistic people every year.

I understand the world can be an unforgiving place and things will never be perfect but investing is a game where the optimists win.

3. 8 Google Employees Invented Modern AI. Here’s the Inside Story – Steven Levy

EIGHT NAMES ARE listed as authors on “Attention Is All You Need,” a scientific paper written in the spring of 2017. They were all Google researchers, though by then one had left the company…

…Recurrent neural networks struggled to parse longer chunks of text. Take a passage like Joe is a baseball player, and after a good breakfast he went to the park and got two hits. To make sense of “two hits,” a language model has to remember the part about baseball. In human terms, it has to be paying attention. The accepted fix was something called “long short-term memory” (LSTM), an innovation that allowed language models to process bigger and more complex sequences of text. But the computer still handled those sequences strictly sequentially—word by tedious word—and missed out on context clues that might appear later in a passage. “The methods we were applying were basically Band-Aids,” Uszkoreit says. “We could not get the right stuff to really work at scale.”

Around 2014, he began to concoct a different approach that he referred to as self-attention. This kind of network can translate a word by referencing any other part of a passage. Those other parts can clarify a word’s intent and help the system produce a good translation. “It actually considers everything and gives you an efficient way of looking at many inputs at the same time and then taking something out in a pretty selective way,” he says. Though AI scientists are careful not to confuse the metaphor of neural networks with the way the biological brain actually works, Uszkoreit does seem to believe that self-attention is somewhat similar to the way humans process language.

Uszkoreit thought a self-attention model could potentially be faster and more effective than recurrent neural nets. The way it handles information was also perfectly suited to the powerful parallel processing chips that were being produced en masse to support the machine learning boom. Instead of using a linear approach (look at every word in sequence), it takes a more parallel one (look at a bunch of them together). If done properly, Uszkoreit suspected, you could use self-attention exclusively to get better results…

…The transformer crew set about building a self-attention model to translate text from one language to another. They measured its performance using a benchmark called BLEU, which compares a machine’s output to the work of a human translator. From the start, their new model did well. “We had gone from no proof of concept to having something that was at least on par with the best alternative approaches to LSTMs by that time,” Uszkoreit says. But compared to long short-term memory, “it wasn’t better.”

They had reached a plateau—until one day in 2017, when Noam Shazeer heard about their project, by accident. Shazeer was a veteran Googler—he’d joined the company in 2000—and an in-house legend, starting with his work on the company’s early ad system. Shazeer had been working on deep learning for five years and recently had become interested in large language models. But these models were nowhere close to producing the fluid conversations that he believed were possible.

As Shazeer recalls it, he was walking down a corridor in Building 1965 and passing Kaiser’s workspace. He found himself listening to a spirited conversation. “I remember Ashish was talking about the idea of using self-attention, and Niki was very excited about it. I’m like, wow, that sounds like a great idea. This looks like a fun, smart group of people doing something promising.” Shazeer found the existing recurrent neural networks “irritating” and thought: “Let’s go replace them!”

Shazeer’s joining the group was critical. “These theoretical or intuitive mechanisms, like self-attention, always require very careful implementation, often by a small number of experienced ‘magicians,’ to even show any signs of life,” says Uszkoreit. Shazeer began to work his sorcery right away. He decided to write his own version of the transformer team’s code. “I took the basic idea and made the thing up myself,” he says. Occasionally he asked Kaiser questions, but mostly, he says, he “just acted on it for a while and came back and said, ‘Look, it works.’” Using what team members would later describe with words like “magic” and “alchemy” and “bells and whistles,” he had taken the system to a new level.

“That kicked off a sprint,” says Gomez. They were motivated, and they also wanted to hit an upcoming deadline—May 19, the filing date for papers to be presented at the biggest AI event of the year, the Neural Information Processing Systems conference in December. As what passes for winter in Silicon Valley shifted to spring, the pace of the experiments picked up. They tested two models of transformers: one that was produced with 12 hours of training and a more powerful version called Big that was trained over three and a half days. They set them to work on English-to-German translation.

The basic model outperformed all competitors—and Big earned a BLEU score that decisively shattered previous records while also being more computationally efficient. “We had done it in less time than anyone out there,” Parmar says. “And that was only the beginning, because the number kept improving.”…

…TRANSFORMERS DID NOT instantly take over the world, or even Google. Kaiser recalls that around the time of the paper’s publication, Shazeer proposed to Google executives that the company abandon the entire search index and train a huge network with transformers—basically to transform how Google organizes information. At that point, even Kaiser considered the idea ridiculous. Now the conventional wisdom is that it’s a matter of time.

A startup called OpenAI was much faster to pounce. Soon after the paper was published, OpenAI’s chief researcher, Ilya Sutskever—who had known the transformer team during his time at Google—suggested that one of its scientists, Alec Radford, work on the idea. The results were the first GPT products. As OpenAI CEO Sam Altman told me last year, “When the transformer paper came out, I don’t think anyone at Google realized what it meant.”

The picture internally is more complicated. “It was pretty evident to us that transformers could do really magical things,” says Uszkoreit. “Now, you may ask the question, why wasn’t there ChatGPT by Google back in 2018? Realistically, we could have had GPT-3 or even 3.5 probably in 2019, maybe 2020. The big question isn’t, did they see it? The question is, why didn’t we do anything with the fact that we had seen it? The answer is tricky.”

Many tech critics point to Google’s transition from an innovation-centered playground to a bottom-line-focused bureaucracy. As Gomez told the Financial Times, “They weren’t modernizing. They weren’t adopting this tech.” But that would have taken a lot of daring for a giant company whose technology led the industry and reaped huge profits for decades. Google did begin to integrate transformers into products in 2018, starting with its translation tool. Also that year, it introduced a new transformer-based language model called BERT, which it started to apply to search the year after.

But these under-the-hood changes seem timid compared to OpenAI’s quantum leap and Microsoft’s bold integration of transformer-based systems into its product line. When I asked CEO Sundar Pichai last year why his company wasn’t first to launch a large language model like ChatGPT, he argued that in this case Google found it advantageous to let others lead. “It’s not fully clear to me that it might have worked out as well. The fact is, we can do more after people had seen how it works,” he said…

…Does Google miss these escapees? Of course, in addition to others who have migrated from the company to new AI startups. (Pichai reminded me, when I asked him about the transformer departures, that industry darling OpenAI also has seen defections: “The AI area is very, very dynamic,” he said.) But Google can boast that it created an environment that supported the pursuit of unconventional ideas. “In a lot of ways Google has been way ahead—they invested in the right minds and created the environment where we could explore and push the envelope,” Parmar says. “It’s not crazy that it took time to adopt it. Google had so much more at stake.”

Without that environment: no transformer. Not only were the authors all Google employees, they also worked out of the same offices. Hallway encounters and overheard lunch conversations led to big moments. The group is also culturally diverse. Six of the eight authors were born outside the United States; the other two are children of two green-card-carrying Germans who were temporarily in California and a first-generation American whose family had fled persecution, respectively.

4. In Depth: Local Governments Struggle to Tackle Mountain of Hidden Debt – Cheng Siwei, Wang Juanjuan, Zhang Yuzhe, Ding Feng and Zhang Yukun

The central government has been trying to address the problem of LGFV debt for years, mainly through piecemeal measures that had limited success. But in July, the Politburo vowed to formulate and implement a comprehensive strategy to resolve local government hidden debts.

These off-the-books liabilities, which include LGFV bonds with implicit official backing, have accumulated over the years to around 30 trillion to 70 trillion yuan according to some estimates, and become a threat to the country’s fiscal and financial stability and sustainability.

One of the main instruments being used to repay hidden debt in this round of debt resolution is special refinancing bonds — on-balance-sheet local government bonds whose proceeds are used to repay outstanding hidden debt. Issuance has stepped up significantly since early October after the Ministry of Finance launched a special refinancing bond swap program.

From October to December, almost all provincial-level regions on the Chinese mainland issued these special refinancing bonds, raising nearly 1.4 trillion yuan to repay hidden borrowings, according to calculations by analysts at Tianfeng Securities Co. Ltd. The regions include heavily indebted Guizhou province, which topped the list with issuance of 226.4 billion yuan.

Many regions have announced plans to issue more such bonds in February and March, with planned issuances totaling more than 100 billion yuan, the Tianfeng analysts wrote in a January report.

The campaign to resolve hidden debt has tightened rules for new debt issuance and cut some localities off from their previous financing channels, depriving them of resources to pay interest on hidden debt. The proceeds of special refinancing bonds cannot be used to make interest payments.

“The core issue now is that we can’t make our interest payments,” a source who works for an economic development zone in West China told Caixin, noting that without new financing, the fiscal revenue of the region can only sustain government agencies’ day-to-day operations and preferential policies for attracting businesses. He said his local government has stopped making all other payments, including those to project developers, to ensure it can meet interest payments on outstanding LGFV debt…

…The renewed push to bring hidden debt onto the books and restructure or swap LGFV debt, however, has reinforced the belief that the central government won’t allow LGFVs to default on their bonds, reviving investor sentiment. That’s led to a surge in demand for LGFV bonds over the past few months, even as the central government has repeatedly highlighted the need to stem any renewed buildup in hidden debt…

…Although LGFV bonds are back in hot demand, tightened oversight has made it more difficult for some vehicles, especially those with heavy debt burdens, to continue issuing new debt. This has curbed growth in hidden debt to some extent, but it has added to default risks of some LGFV bonds as there is less money available to make the interest repayments.

The central government ordered provincial officials to compile a list of LGFVs owned by local authorities in their jurisdictions…

…Obtaining new bank loans has become much harder for LGFVs on the list, as banks heed the central government’s instruction to prevent new LGFV debt.

Regarding existing LGFV debt, the State Council in September issued guidance that banks, among the most important creditors of LGFVs, should become involved in debt resolution in 12 provincial-level regions with high government leverage, which include Liaoning, Heilongjiang, and Jilin, the three rustbelt provinces in Northeast China. The guidance set out that banks should focus on restructuring or swapping existing loans, high-interest non-standard debt, and other types of borrowing.

5. Conviction and Quality – Josh Tarasoff

Conviction is no doubt the foundation of long-term business ownership. How is it formed? What is it like to have it? Why does it falter? In my experience there are two distinct kinds of conviction. Explicit conviction, as I call it, comes from having figured something out. It entails a useful prediction, like “our ETA is 5pm” or “majoring in economics will lead to better career prospects than majoring in philosophy.” There is an underlying logic to it, which can be explained and used to persuade. Implicit conviction, on the other hand, is exemplified by the trust one might have in a family member, a dear friend, a close colleague, to do the right thing, to get the job done, to come through. It is felt as opposed to believed. This kind of conviction doesn’t make predictions so much as align with what is good. It doesn’t theorize about goodness but rather knows it when it sees it…

…In the context of investing, one might develop the thesis that a particular company can capture X% market share, generate Y dollars in annual revenue, achieve Z% operating margins, and therefore has an intrinsic value within a certain range. One might have high confidence because of the presence of competitive advantages and management with a very good track record. One would have a range of expected returns from owning the shares over time. All of this would fall into the explicit category.

Sooner or later, the investment would encounter a confounding surprise. Perhaps execution turns choppy, a new competitive vector emerges out of nowhere, an exogenous crisis turns the world upside down, etc. Old projections are now in doubt, previous plans and strategies are being reworked, everything is less fun. These things are actually happening all the time— something explicit conviction has a way of tuning out! Only genuine and well-placed implicit conviction, a qualitative knowing that the company will do what it needs to and ought to do, is equipped to ably traverse this kind of terrain. Unlike analysis-based explicit conviction, implicit conviction comes from something deeper than the cause and effect we perceive in the unfolding of events—it is both analytical and, crucially, intuitive (about which more later)…

…While in everyday life implicit conviction arises naturally, in the context of investing I can’t help but feel it is somewhat alien. In part, this is because few companies are truly deserving. Even so, I suspect that implicit conviction is proffered by investors even less than it ought to be. It isn’t difficult to see why the investment industry is inhospitable to implicit conviction, and why its partner rules the roost. Implicit conviction forms of its own accord and cannot be planned. It defies quantification, eliciting the charge of being too “fuzzy” to matter. Nor can it be fully captured in words. Implicit conviction is impossible to transmit from analyst to portfolio manager or from portfolio manager to client, which is highly inconvenient for the business of managing money. It is primarily personal. It is quiet. By contrast, the appeal of the explicit is clear. Explicit conviction furnishes the comfort of knowability and modeled outcomes. It projects the legitimacy of diligence and precision. It is thought to be reliably manufactured via “repeatable process.” It is clever and self assured….

…Nonetheless, because literal communication necessitates choosing a word, I will use “Quality” (capitalized to distinguish it from the ordinary sense of the term) to indicate the deeper-something on which implicit conviction is based. Using “Quality” in this way is consistent with my prior writing and pays homage to the work of Robert Pirsig, which was a formative influence.

Analysis plays an important but limited role in detecting Quality. For example, the following is a selection of (neither necessary nor sufficient) indicators that I have found to be suggestive of Quality in companies:

  • “Wow” customer experiences
  • Mission to solve an important problem
  • Domain mastery (the best at what they do)
  • First-principles-based thinking and invention
  • Unlimited ambition combined with no-nonsense realism
  • Overcapitalized balance sheet
  • Founder mentality (life’s work).

While carefully looking for indicators like these is helpful, I think it would be a misstep to attempt to systematize the search, constructing a Grand Unified Theory of Quality and attendant comprehensive processes for finding and evaluating it. Quality emerges from the complexity of the system in action; it is in the how rather than the what. Thus, when Quality is broken down into parts and analyzed, its essence is lost. This explains why analysis alone has trouble discerning the authentic from the artificial. Moreover, Quality frozen in a theory or process cannot be recognized in sufficiently new contexts, such as in a company that is novel to one’s experience or in the same company as it evolves (they always do!).

So where does that leave us? With intuition. Well-honed intuition does what analysis cannot by perceiving Quality directly, as opposed to through an intellectual process. What I suspect is happening in the direct perception of Quality is subconscious pattern recognition, based upon a dynamic, holistic experience of the thing in question. Of course, the ability to intuitively recognize patterns in a specific domain must be earned through experience and feedback; indeed, I have found that the value of my own intuition has grown (starting at zero) over many years. Interestingly, I also find that experiencing Quality in any one domain (e.g., music or meditation, to use examples that are dear to me) can be helpful for recognizing it in other domains (including business) because Quality’s nature is universal, even as its manifestations are necessarily particular.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google) and Microsoft. Holdings are subject to change at any time.

What We’re Reading (Week Ending 31 March 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 31 March 2024:

1. Gold-Medalist Coders Build an AI That Can Do Their Job for Them – Ashlee Vance

Take the case of Cognition AI Inc.

You almost certainly have not heard of this startup, in part because it’s been trying to keep itself secret and in part because it didn’t even officially exist as a corporation until two months ago. And yet this very, very young company, whose 10-person staff has been splitting time between Airbnbs in Silicon Valley and home offices in New York, has raised $21 million from Peter Thiel’s venture capital firm Founders Fund and other brand-name investors, including former Twitter executive Elad Gil. They’re betting on Cognition AI’s team and its main invention, which is called Devin.

Devin is a software development assistant in the vein of Copilot, which was built by GitHub, Microsoft and OpenAI, but, like, a next-level software development assistant. Instead of just offering coding suggestions and autocompleting some tasks, Devin can take on and finish an entire software project on its own. To put it to work, you give it a job—“Create a website that maps all the Italian restaurants in Sydney,” say—and the software performs a search to find the restaurants, gets their addresses and contact information, then builds and publishes a site displaying the information. As it works, Devin shows all the tasks it’s performing and finds and fixes bugs on its own as it tests the code being written.

The founders of Cognition AI are Scott Wu, its chief executive officer; Steven Hao, the chief technology officer; and Walden Yan, the chief product officer…

…Wu, 27, is the brother of Neal Wu, who also works at Cognition AI. These two men are world-renowned for their coding prowess: The Wu brothers have been competing in, and often winning, international coding competitions since they were teenagers…

…Sport-coding—yes, it’s a real thing—requires people to solve puzzles and program with speed and accuracy. Along the way, it trains contestants to approach problems in novel ways. Cognition AI is full of sport-coders. Its staff has won a total of 10 gold medals at the top international competition, and Scott Wu says this background gives his startup an edge in the AI wars…

…One of the big claims Cognition AI is making with Devin is that the company has hit on a breakthrough in a computer’s ability to reason. Reasoning in AI-speak means that a system can go beyond predicting the next word in a sentence or the next snippet in a line of code, toward something more akin to thinking and rationalizing its way around problems. The argument in AI Land is that reasoning is the next big thing that will advance the industry, and lots of startups are making various boasts about their ability to do this type of work.

Devin does appear to be well ahead of the other coding assistants in many respects. You can give it jobs to do with natural language commands, and it will set off and accomplish them. As Devin works, it tells you about its plan and then displays the commands and code it’s using. If something doesn’t look quite right, you can give the AI a prompt to go fix the issue, and Devin will incorporate the feedback midstream. Most current AI systems have trouble staying coherent and on task during these types of long jobs, but Devin keeps going through hundreds and even thousands of tasks without going off track.

In my tests with the software, Devin could build a website from scratch in 5 to 10 minutes, and it managed to re-create a web-based version of Pong in about the same amount of time. I had to prompt it a couple of times to improve the physics of the ball movement in the game and to make some cosmetic changes on its websites, all of which Devin accomplished just fine and with a polite attitude…

…Exactly how Cognition AI made this breakthrough, and in so short a time, is something of a mystery, at least to outsiders. Wu declines to say much about the technology’s underpinnings other than that his team found unique ways to combine large language models (LLMs) such as OpenAI’s GPT-4 with reinforcement learning techniques. “It’s obviously something that people in this space have thought about for a long time,” he says. “It’s very dependent on the models and the approach and getting things to align just right.”

2. Geopolitics in the C-Suite – Jami Miscik, Peter Orszag, and Theodore Bunzel

But even though national security and foreign policy occasionally intruded on corporate America during that time, until very recently, few executives concerned themselves with geopolitics. In the post–Cold War world, with globalization on the march, the idea that national interests might be at odds with open markets and expanding trade came to seem alien to American executives.

But the changes that have roiled the geopolitical landscape in recent years have left an impression in C-suites around the United States. In a recent poll of 500 institutional investors, geopolitics ranked as the top risk to the global economy and markets in 2024…

…As governments lean on economic restrictions and industrial policies to achieve geopolitical ends, corporations have increasingly become both the objects and instruments of foreign policy…

…The centrality of economic competition to today’s foreign policy problems represents a qualitative break from the past. During the Cold War, for example, the United States and the Soviet Union hardly interacted economically: trade between them peaked at a paltry $4.5 billion in 1979; in recent years, the United States and China have generally traded that much every week or two, adjusting for inflation. In the post–Cold War era, U.S. foreign policy was focused on opening markets and reducing international economic barriers rather than erecting them. Era-defining crises such as the 9/11 attacks did little to change the relationship between U.S. policymakers and American corporations; if anything, the “war on terror” further solidified the idea that foreign policy was primarily concerned with security and military issues, not economics.

But in the background, global economic integration was transforming the playing field. In 1980, trade accounted for just 37 percent of global GDP. Today, that figure is 74 percent, and economies have become intertwined to a degree never seen in the twentieth century. Globalization is not new, of course; it has been a centuries-long process. What is new, however, is the emergence of great-power rivalry in a highly interconnected world. Military power still matters, but economic and technological competition have become the main battlefield of global politics. Under the so-called Washington consensus that dominated policymaking for decades, the question of where a semiconductor manufacturer would build its next factory or whether German auto companies would decide to throttle their investments in China would have seemed relatively unimportant to policymakers. Now, such questions are at the center of almost every major foreign policy debate.

Greater economic integration has also created a complex web of links between geopolitical rivals that policymakers now seek to leverage for strategic ends. This is especially true when it comes to financial and technological networks, where Washington holds a privileged position…

…But as great-power tensions have increased, so has the number of sectors caught in the fray of what Farrell and Newman call “weaponized interdependence.” Consider, for example, the way that G-7 countries have taken advantage of Russian dependence on shipping insurers based in the West, an industry that most foreign policymakers had probably never thought about before Russia’s 2022 invasion of Ukraine. To try to cap the price of Russian oil exports, the G-7 prevented these companies from insuring Russian crude oil cargoes unless they had been sold at a maximum of $60 per barrel.

Western powers are not the only ones playing this game. In 2010, after a Chinese fishing trawler and Japanese Coast Guard patrol boats collided in disputed waters, setting off a diplomatic row between Beijing and Tokyo, China banned exports to Japan of the rare-earth minerals that are critical components of batteries and electronics, thus raising costs and creating shortages for Japanese manufacturers of everything from hybrid cars to wind turbines…

…More recently, a number of American consulting firms have been caught in the middle of the complex U.S.-Saudi relationship, with Congress demanding details about their contracts with Saudi Arabia that Riyadh has forbidden them to provide.

All these dynamics are being turbocharged by an intensifying competition between the United States and China, the two countries with the largest and most globally intertwined economies. Both aim to dominate the twenty-first-century economy, which means gaining the upper hand in computing technologies, biotechnology, and clean energy. And the foreign policies of both countries are now driven by a shared desire to shape their economies in ways that reduce their vulnerability and increase their leverage. China calls this “self-reliance.” Washington calls it “de-risking.” For the United States, what it looks like in practice is expanded export controls on advanced semiconductors and manufacturing equipment, enhanced government screening of investments by U.S. companies in foreign markets, and major subsidies for industries such as electric vehicles and microchips, primarily through the Inflation Reduction Act and the CHIPS Act. In this brave new world, the secretary of commerce is as important to foreign policy as the secretaries of state and defense.

Washington is hardly alone in taking such steps. State-sponsored drives for greater self-reliance have taken hold in nearly every major economy, particularly after the supply-chain disruptions of the COVID-19 pandemic. The number of countries introducing or expanding investment screening, for example, jumped from three between 1995 and 2005 to 54 between 2020 and 2022. Meanwhile, a wave of industrial policies has increased trade barriers in an attempt to induce companies to reshore their supply chains. At the same time, the understanding of what matters to national security has also expanded, as countries seek to advance or protect everything from software and microchips to pharmaceuticals and foodstuffs.

Many of the complications of this new era are rooted in the difference between the way the public and private sectors view time horizons. Policymakers set bright lines with immediate operational implications—for example, suddenly forbidding companies from exporting or importing certain goods from certain countries. But companies need to make long-term investment decisions. Should a company set up another plant in China if there is market demand and doing so is currently allowed by law? Should a pharmaceutical company set up advanced R & D centers in mainland China or purchase a Chinese biotech firm, given the long-run trajectory of relations between Beijing and the West? Should a consumer electronics firm purchase Chinese-made chips if they are the most cost-efficient option? Answering these questions requires executives to forecast the outcomes of highly volatile political debates and policymaking choices over which they have little control. And yet whatever decisions they make have a significant effect on whether, for example, the United States can effectively “de-risk” its economic relationship with China.

The example of semiconductors is instructive. Washington is seeking to reshore semiconductor manufacturing, but the success of its flagship industrial policy, the CHIPS Act, depends only in part on how the Commerce Department distributes the legislation’s $39 billion in subsidies over the next five years. A much more important factor is whether the Taiwanese chip manufacturer TSMC will risk setting up facilities in the United States despite high costs and a relative scarcity of human capital, and whether Apple decides to buy slightly more expensive chips made by U.S. fabricators instead of less expensive ones produced in Asia. And the CHIPS Act is only one input in those decisions.

3. Get Smart: Chasing Nvidia? Don’t succumb to FOMO – Chin Hui Leong

Are you feeling left out because you missed Nvidia’s (NASDAQ: NVDA) massive stock rise?

Well, we have good news and bad news.

Let’s start with the bad news: that tightening in your chest you are feeling right now is the fear of missing out — or better known by its initials “FOMO”. 

And ladies and gentlemen, FOMO is real.

It’s that sneaky emotion which spurs you to buy a stock based on a feeling rather than proper research…

…But hang on, amid the hype — there’s good news too.

If you recognise that you are feeling FOMO, then congratulations — you have just taken the first step in recognising what you have to deal with: your runaway emotions.

The next step is to keep your emotions in check.

On the other side of FOMO, is its cousin FOJI — or the fear of joining in.

Like FOMO, FOJI is also a strong emotion.

That’s especially true for some investors who are bearing scars from 2022 when US stocks took a beating amid a punishing bear market.

These scars can emit another fear — FOJI — which is paralysing for investors.

The fear of looking stupid if you buy today only to watch the stock fall the very next day…

…Whether it is FOMO or FOJI, you won’t invest well if feelings dictate your actions.

Recognising the presence of both emotions is key…

…Beyond FOMO and FOJI, there’s JOMO or the joy of missing out.

Don’t feel down if you decide to give Nvidia a pass.

As Peter Lynch once said — you can’t kiss all the frogs to find out which will turn into a prince.

Unconvinced?

In Lynch book’s “One Up on Wall Street”, he wrote down the names of 65 stocks which returned at least 10 times their original price (he calls them 10-baggers).

Except that the fund that he ran owned NONE of them.

Before you start rolling your eyes, consider this point: Peter Lynch achieved a stunning 29% per year in annualized returns over 13 years, outpacing the benchmark S&P 500 index (INDEXSP: .INX) by more than two times…

…By sharing the list of missed winners, Lynch had a salient point to make: you do not need to be in every 10-bagger to deliver enviable returns.

4. How the richest woman in the world—mocked as a ‘miser’ in the press—helped bail out New York City during the panic of 1907 – Will Daniel

Hetty Green is remembered as the “world’s greatest miser” and the “Witch of Wall Street,” but these days, Green would likely be seen as an eccentric investing icon. After all, while she became famous for her frugal nature and gruff exterior, Green pioneered value investing strategies that have made billionaires out of many of today’s leading investors. And when the chips were down, when people really needed help, the whaling heiress turned independent investor, business tycoon, and world’s wealthiest woman often used her fortune to save the day…

…Over a three-week period after the panic began on Oct. 22, 1907, the New York Stock Exchange plummeted nearly 50% from its 1906 peak. And a year later, in 1908, Gross National Product (GNP), a measure akin to today’s Gross Domestic Product (GDP), cratered 12%. The problems for the banking system were so severe during the knickerbocker crisis that they spurred the establishment of the Federal Reserve System…

…As the situation deteriorated, John Pierpont Morgan, the American financier who founded what is now JPMorgan Chase, was eventually forced to call together a group of Wall Street’s best and brightest at the Morgan Library to help decide how to prop up the ailing economy and stock market. Hetty Green was the only woman who was invited to attend that meeting during the height of the panic…

…“I saw this situation coming,” she said, noting that there were undeniable signs of stress. “Some of the solidest men of the Street came to me and wanted to unload all sorts of things, from palatial residences to automobiles.”

Green said that she then gave The New York Central Railroad company a “big loan” after they came knocking, and that made her “sit up and do some thinking.” She decided to begin gathering as much cash as possible, understanding that a panic could be on the way…

…Green described how men came to New York from all over the country to ask for loans during the panic of 1907. But despite being labeled a “miser” throughout her life, she didn’t take advantage of the situation.

“Those to whom I loaned money got it at 6%. I might just as easily have secured 40%,” she explained…

…Usury, or charging excessive interest for a loan, was against Green’s moral code, which was born of her Quaker roots…

…Green would go on to lend the government of New York City $1.1 million at the peak of the 1907 panic, which is equivalent to roughly $33 million in today’s dollars…

…“On more than one occasion, when New York was running low on money, she would lend money to the city,” explained Charles Slack, the author of Green’s biography, Hetty: The Genius and Madness of America’s First Female Tycoon. “And she always did so at reasonable rates. She didn’t gouge or hold the city over a barrel.”

5. Transcript for Yann Lecun: Meta AI, Open Source, Limits of LLMs, AGI & the Future of AI | Lex Fridman Podcast #416 – Lex Fridman and Yann Lecun

Lex Fridman (00:50:40) I would love to sort of linger on your skepticism around auto regressive LLMs. So one way I would like to test that skepticism is everything you say makes a lot of sense, but if I apply everything you said today and in general to I don’t know, 10 years ago, maybe a little bit less, no, let’s say three years ago, I wouldn’t be able to predict the success of LLMs. So does it make sense to you that autoregressive LLMs are able to be so damn good?

Yann LeCun (00:51:20) Yes.

Lex Fridman (00:51:21) Can you explain your intuition? Because if I were to take your wisdom and intuition at face value, I would say there’s no way autoregressive LLMs, one token at a time, would be able to do the kind of things they’re doing.

Yann LeCun (00:51:36) No, there’s one thing that autoregressive LLMs or that LLMs in general, not just the autoregressive one, but including the bird style bidirectional ones, are exploiting and its self supervised running, and I’ve been a very, very strong advocate of self supervised running for many years. So those things are a incredibly impressive demonstration that self supervised running actually works. The idea that started, it didn’t start with BERT, but it was really kind of good demonstration with this.

(00:52:09) So the idea that you take a piece of text, you corrupt it, and then you train some gigantic neural net to reconstruct the parts that are missing. That has produced an enormous amount of benefits. It allowed us to create systems that understand language, systems that can translate hundreds of languages in any direction, systems that are multilingual, so it’s a single system that can be trained to understand hundreds of languages and translate in any direction, and produce summaries and then answer questions and produce text.

(00:52:51) And then there’s a special case of it, which is the auto regressive trick where you constrain the system to not elaborate a representation of the text from looking at the entire text, but only predicting a word from the words that are come before. And you do this by constraining the architecture of the network, and that’s what you can build an auto aggressive LLM from.

(00:53:15) So there was a surprise many years ago with what’s called decoder only LLM. So since systems of this type that are just trying to produce words from the previous one and the fact that when you scale them up, they tend to really understand more about language. When you train them on lots of data, you make them really big. That was a surprise and that surprise occurred quite a while back, with work from Google, Meta, OpenAI, et cetera, going back to the GPT kind of work, general pre-trained transformers.

Lex Fridman (00:53:56) You mean like GPT2? There’s a certain place where you start to realize scaling might actually keep giving us an emergent benefit.

Yann LeCun (00:54:06) Yeah, I mean there were work from various places, but if you want to place it in the GPT timeline, that would be around GPT2, yeah.

Lex Fridman (00:54:19) Well, because you said it so charismatic and you said so many words, but self supervised learning, yes. But again, the same intuition you’re applying to saying that auto aggressive LLMs cannot have a deep understanding of the world. If we just apply that, same intuition, does it make sense to you that they’re able to form enough of a representation in the world to be damn convincing, essentially passing the original touring test with flying colors?

Yann LeCun (00:54:50) Well, we’re fooled by their fluency, right? We just assume that if a system is fluent in manipulating language, then it has all the characteristics of human intelligence, but that impression is false. We’re really fooled by it.

Lex Fridman (00:55:06) What do you think Alan Turing would say, without understanding anything, just hanging out with it?

Yann LeCun (00:55:11) Alan Turing would decide that a Turing test is a really bad test, okay? This is what the AI community has decided many years ago that the Turing test was a really bad test of intelligence.

Lex Fridman (00:55:22) What would Hans Marvek say about the larger language models?

Yann LeCun (00:55:26) Hans Marvek would say that Marvek Paradox still applies. Okay, we can pass-

Lex Fridman (00:55:32) You don’t think he would be really impressed?

Yann LeCun (00:55:34) No, of course everybody would be impressed. But it’s not a question of being impressed or not, it’s the question of knowing what the limit of those systems can do. Again, they are impressive. They can do a lot of useful things. There’s a whole industry that is being built around them. They’re going to make progress, but there is a lot of things they cannot do, and we have to realize what they cannot do and then figure out how we get there. And I’m seeing this from basically 10 years of research on the idea of self supervised running, actually that’s going back more than 10 years, but the idea of self supervised running. So basically capturing the internal structure of a piece of a set of inputs without training the system for any particular task, to learning representations.

(00:56:26) The conference I co-founded 14 years ago is called International Conference on Learning Representations. That’s the entire issue that deep learning is dealing with, and it’s been my obsession for almost 40 years now. So learning representation is really the thing. For the longest time, we could only do this with supervised learning, and then we started working on what we used to call unsupervised learning and revived the idea of unsupervised running in the early 2000s with your [inaudible 00:56:58] and Jeff Hinton. Then discovered that supervised running actually works pretty well if you can collect enough data. And so the whole idea of unsupervised, self supervised running kind of took a backseat for a bit, and then I tried to revive it in a big way starting in 2014, basically when we started FAIR and really pushing for finding new methods to do self supervised running both for text and for images and for video and audio.

(00:57:29) And some of that work has been incredibly successful. I mean, the reason why we have multilingual translation system, things to do, content moderation on Meta, for example, on Facebook, that are multilingual, that understand whether a piece of text is hate speech not or something, is due to that progress using self supervised running for NLP, combining this with transformer architectures and blah, blah, blah.

(00:57:53) But that’s the big success of self supervised running. We had similar success in speech recognition, a system called WAVE2VEC, which is also a joint embedding architecture, by the way, trained with contrastive running. And that system also can produce speech recognition systems that are multilingual with mostly unlabeled data and only need a few minutes of labeled data to actually do speech recognition, that’s amazing. We have systems now based on those combination of ideas that can do real time translation of hundreds of languages into each other, speech to speech.

Lex Fridman (00:58:28) Speech to speech, even including, which is fascinating, languages that don’t have written forms.

Yann LeCun (00:58:34) That’s right.

Lex Fridman (00:58:34) Just spoken only.

Yann LeCun (00:58:35) That’s right. We don’t go through text, it goes directly from speech to speech using an internal representation of speech units that are discrete, but it’s called Textless NLP. We used to call it this way. But yeah, so I mean incredible success there. And then for 10 years, we tried to apply this idea to learning representations of images by training a system to predict videos, learning intuitive physics by training a system to predict what’s going to happen in the video.

(00:59:02) And tried and tried and failed and failed, with generative models, with models that predict pixels. We could not get them to learn good representations of images. We could not get them to learn good representations of videos. And we tried many times, we published lots of papers on it, where they kind of sort of work, but not really great. They started working, we abandoned this idea of predicting every pixel and basically just doing the joint embedding and predicting and representation space, that works. So there’s ample evidence that we’re not going to be able to learn good representations of the real world using generative model. So I’m telling people, everybody’s talking about generative AI. If you’re really interested in human level AI, abandon the idea of generative AI…

…Yann LeCun (01:35:29) I actually made that comment on just about every social network I can, and I’ve made that point multiple times in various forums. Here’s my point of view on this, people can complain that AI systems are biased and they generally are biased by the distribution of the training data that they’ve been trained on that reflects biases in society, and that is potentially offensive to some people or potentially not. And some techniques to de-bias then become offensive to some people because of historical incorrectness and things like that.

(01:36:23) And so you can ask two questions, the first question is, is it possible to produce an AI system that is not biased? And the answer is, absolutely not. And it’s not because of technological challenges, although they are technological challenges to that, it’s because bias is in the eye of the beholder. Different people may have different ideas about what constitutes bias for a lot of things, there are facts that are indisputable, but there are a lot of opinions or things that can be expressed in different ways. And so you cannot have an unbiased system, that’s just an impossibility.

(01:37:08) And so what’s the answer to this? And the answer is the same answer that we found in liberal democracy about the press, the press needs to be free and diverse. We have free speech for a good reason, is because we don’t want all of our information to come from a unique source because that’s opposite to the whole idea of democracy and progressive ideas and even science. In science, people have to argue for different opinions and science makes progress when people disagree and they come up with an answer and consensus forms, and it’s true in all democracies around the world.

(01:37:58) There is a future which is already happening where every single one of our interaction with the digital world will be mediated by AI systems, AI assistance. We’re going to have smart glasses, you can already buy them from Meta, the Ray-Ban Meta where you can talk to them and they are connected with an LLM and you can get answers on any question you have. Or you can be looking at a monument and there is a camera in the glasses you can ask it like, what can you tell me about this building or this monument? You can be looking at a menu in a foreign language, and I think we will translate it for you, or we can do real time translation if we speak different languages. So a lot of our interactions with the digital world are going to be mediated by those systems in the near future.

(01:38:53) Increasingly, the search engines that we’re going to use are not going to be search engines, they’re going to be dialogue systems that we just ask a question and it will answer and then point you to perhaps appropriate reference for it. But here is the thing, we cannot afford those systems to come from a handful of companies on the west coast of the US because those systems will constitute the repository of all human knowledge, and we cannot have that be controlled by a small number of people. It has to be diverse for the same reason the press has to be diverse, so how do we get a diverse set of AI assistance? It’s very expensive and difficult to train a base model, a base LLM at the moment, in the future it might be something different, but at the moment, that’s an LLM. So only a few companies can do this properly.

(01:39:50) And if some of those top systems are open source, anybody can use them, anybody can fine tune them. If we put in place some systems that allows any group of people, whether they are individual citizens, groups of citizens, government organizations, NGOs, companies, whatever, to take those open source AI systems and fine tune them for their own purpose on their own data, then we’re going to have a very large diversity of different AI systems that are specialized for all of those things.

(01:40:35) I tell you, I talked to the French government quite a bit, and the French government will not accept that the digital diet of all their citizens be controlled by three companies on the west coast of the US. That’s just not acceptable, it’s a danger to democracy regardless of how well-intentioned those companies are, and it’s also a danger to local culture, to values, to language. I was talking with the founder of Infosys in India, he’s funding a project to fine tune Llama 2, the open source model produced by Meta, so that Llama 2 two speaks all 22 official languages in India, it is very important for people in India. I was talking to a former colleague of mine, Moustapha Cisse, who used to be a scientist at Fair and then moved back to Africa, created a research lab for Google in Africa and now has a new startup Co-Kera.

(01:41:37) And what he’s trying to do, is basically have LLM that speak the local languages in Senegal so that people can have access to medical information because they don’t have access to doctors, it’s a very small number of doctors per capita in Senegal. You can’t have any of this unless you have open source platforms, so with open source platforms, you can have AI systems that are not only diverse in terms of political opinions or things of that-

Yann LeCun (01:42:00) … AI systems that are not only diverse in terms of political opinions or things of that type, but in terms of language, culture, value systems, political opinions, technical abilities in various domains, and you can have an industry, an ecosystem of companies that fine tune those open source systems for vertical applications in industry. I don’t know, a publisher has thousands of books and they want to build a system that allows a customer to just ask a question about the content of any of their books, you need to train on their proprietary data. You have a company, we have one within Meta, it’s called Metamate, and it’s basically an LLM that can answer any question about internal stuff about the company, very useful.

(01:42:53) A lot of companies want this. A lot of companies want this not just for their employees, but also for their customers, to take care of their customers. So the only way you’re going to have an AI industry, the only way you’re going to have AI systems that are not uniquely biased is if you have open source platforms on top of which any group can build specialized systems. So the direction of inevitable direction of history is that the vast majority of AI systems will be built on top of open source platforms…

…Lex Fridman (02:04:21) You often say that a GI is not coming soon, meaning not this year, not the next few years, potentially farther away. What’s your basic intuition behind that?

Yann LeCun (02:04:35) So first of all, it’s not going to be an event. The idea somehow, which is popularized by science fiction and Hollywood, that somehow somebody is going to discover the secret to AGI or human-level AI or AMI, whatever you want to call it, and then turn on a machine and then we have AGI, that’s just not going to happen. It’s not going to be an event. It’s going to be gradual progress. Are we going to have systems that can learn from video how the world works and learn good representations? Yeah. Before we get them to the scale and performance that we observe in humans it’s going to take quite a while. It’s not going to happen in one day. Are we going to get systems that can have large amount of associated memory so they can remember stuff? Yeah, but same, it’s not going to happen tomorrow. There is some basic techniques that need to be developed. We have a lot of them, but to get this to work together with a full system is another story.

(02:05:37) Are we going to have systems that can reason and plan perhaps along the lines of objective-driven AI architectures that I described before? Yeah, but before we get this to work properly, it’s going to take a while. Before we get all those things to work together, and then on top of this, have systems that can learn hierarchical planning, hierarchical representations, systems that can be configured for a lot of different situation at hand, the way the human brain can, all of this is going to take at least a decade and probably much more because there are a lot of problems that we’re not seeing right now that we have not encountered, so we don’t know if there is an easy solution within this framework. So it’s not just around the corner. I’ve been hearing people for the last 12, 15 years claiming that AGI is just around the corner and being systematically wrong. I knew they were wrong when they were saying it. I called their bullshit…

…Lex Fridman (02:08:48) So you push back against what are called AI doomers a lot. Can you explain their perspective and why you think they’re wrong?

Yann LeCun (02:08:59) Okay, so AI doomers imagine all kinds of catastrophe scenarios of how AI could escape or control and basically kill us all, and that relies on a whole bunch of assumptions that are mostly false. So the first assumption is that the emergence of super intelligence is going to be an event, that at some point we’re going to figure out the secret and we’ll turn on a machine that is super intelligent, and because we’d never done it before, it’s going to take over the world and kill us all. That is false. It’s not going to be an event. We’re going to have systems that are as smart as a cat, have all the characteristics of human-level intelligence, but their level of intelligence would be like a cat or a parrot maybe or something. Then we’re going to work our way up to make those things more intelligent. As we make them more intelligent, we’re also going to put some guardrails in them and learn how to put some guardrails so they behave properly.

(02:10:03) It’s not going to be one effort, that it’s going to be lots of different people doing this, and some of them are going to succeed at making intelligent systems that are controllable and safe and have the right guardrails. If some other goes rogue, then we can use the good ones to go against the rogue ones. So it’s going to be my smart AI police against your rogue AI. So it’s not going to be like we’re going to be exposed to a single rogue AI that’s going to kill us all. That’s just not happening. Now, there is another fallacy, which is the fact that because the system is intelligent, it necessarily wants to take over. There is several arguments that make people scared of this, which I think are completely false as well.

(02:10:48) So one of them is in nature, it seems to be that the more intelligent species otherwise end up dominating the other and even distinguishing the others sometimes by design, sometimes just by mistake. So there is thinking by which you say, “Well, if AI systems are more intelligent than us, surely they’re going to eliminate us, if not by design, simply because they don’t care about us,” and that’s just preposterous for a number of reasons. First reason is they’re not going to be a species. They’re not going to be a species that competes with us. They’re not going to have the desire to dominate because the desire to dominate is something that has to be hardwired into an intelligent system. It is hardwired in humans. It is hardwired in baboons, in chimpanzees, in wolves, not in orangutans. The species in which this desire to dominate or submit or attain status in other ways is specific to social species. Non-social species like orangutans don’t have it, and they are as smart as we are, almost, right?

Lex Fridman (02:12:09) To you, there’s not significant incentive for humans to encode that into the AI systems, and to the degree they do, there’ll be other AIs that punish them for it, I’ll compete them over it.

Yann LeCun (02:12:23) Well, there’s all kinds of incentive to make AI systems submissive to humans.

Lex Fridman (02:12:26) Right.

Yann LeCun (02:12:27) Right? This is the way we’re going to build them. So then people say, “Oh, but look at LLMs. LLMs are not controllable,” and they’re right. LLMs are not controllable. But objectively-driven AI, so systems that derive their answers by optimization of an objective means they have to optimize this objective, and that objective can include guardrails. One guardrail is, obey humans. Another guardrail is, don’t obey humans if it’s hurting other humans within limits.

Lex Fridman (02:12:57) Right. I’ve heard that before somewhere, I don’t remember

Yann LeCun (02:12:59) Yes, maybe in a book.

Lex Fridman (02:13:01) Yeah, but speaking of that book, could there be unintended consequences also from all of this?

Yann LeCun (02:13:09) No, of course. So this is not a simple problem. Designing those guardrails so that the system behaves properly is not going to be a simple issue for which there is a silver bullet for which you have a mathematical proof that the system can be safe. It’s going to be a very progressive, iterative design system where we put those guardrails in such a way that the system behave properly. Sometimes they’re going to do something that was unexpected because the guardrail wasn’t right and we’re dd correct them so that they do it right. The idea somehow that we can’t get it slightly wrong because if we get it slightly wrong, we’ll die is ridiculous. We are just going to go progressively. It is just going to be, the analogy I’ve used many times is turbojet design. How did we figure out how to make turbojet so unbelievably reliable?

(02:14:07) Those are incredibly complex pieces of hardware that run at really high temperatures for 20 hours at a time sometimes, and we can fly halfway around the world on a two-engine jetliner at near the speed of sound. Like how incredible is this? It’s just unbelievable. Did we do this because we invented a general principle of how to make turbojets safe? No, it took decades to fine tune the design of those systems so that they were safe. Is there a separate group within General Electric or Snecma or whatever that is specialized in turbojet safety? No. The design is all about safety, because a better turbojet is also a safer turbojet, so a more reliable one. It’s the same for AI. Do you need specific provisions to make AI safe? No, you need to make better AI systems, and they will be safe because they are designed to be more useful and more controllable…

…Lex Fridman (02:28:45) Well, it’ll be at the very least, absurdly comedic. Okay. So since we talked about the physical reality, I’d love to ask your vision of the future with robots in this physical reality. So many of the kinds of intelligence that you’ve been speaking about would empower robots to be more effective collaborators with us humans. So since Tesla’s Optimus team has been showing us some progress on humanoid robots, I think it really reinvigorated the whole industry that I think Boston Dynamics has been leading for a very, very long time. So now there’s all kinds of companies Figure AI, obviously Boston Dynamics.

Yann LeCun (02:29:30) Unitree.

Lex Fridman (02:29:30) Unitree, but there’s a lot of them.

Yann LeCun (02:29:33) There’s a few of them.

Lex Fridman (02:29:33) It’s great. It’s great. I love it. So do you think there’ll be millions of humanoid robots walking around soon?

Yann LeCun (02:29:44) Not soon, but it’s going to happen. The next decade I think is going to be really interesting in robots, the emergence of the robotics industry has been in the waiting for 10, 20 years without really emerging other than for pre-program behavior and stuff like that. And the main issue is, again, the Moravec paradox, how do we get those systems to understand how the world works and plan actions? And so we can do it for really specialized tasks. And the way Boston Dynamics goes about it is basically with a lot of handcrafted dynamical models and careful planning in advance, which is very classical robotics with a lot of innovation, a little bit of perception, but it’s still not, they can’t build a domestic robot.

(02:30:41) We’re still some distance away from completely autonomous level five driving, and we’re certainly very far away from having level five autonomous driving by a system that can train itself by driving 20 hours like any 17-year-old. So until we have, again, world models, systems that can train themselves to understand how the world works, we’re not going to have significant progress in robotics. So a lot of the people working on robotic hardware at the moment are betting or banking on the fact that AI is going to make sufficient progress towards that…

…Yann LeCun (02:38:29) I love that question. We can make humanity smarter with AI. AI basically will amplify human intelligence. It’s as if every one of us will have a staff of smart AI assistants. They might be smarter than us. They’ll do our bidding, perhaps execute a task in ways that are much better than we could do ourselves, because they’d be smarter than us. And so it’s like everyone would be the boss of a staff of super smart virtual people. So we shouldn’t feel threatened by this any more than we should feel threatened by being the manager of a group of people, some of whom are more intelligent than us. I certainly have a lot of experience with this, of having people working with me who are smarter than me.

(02:39:35) That’s actually a wonderful thing. So having machines that are smarter than us, that assist us in all of our tasks, our daily lives, whether it’s professional or personal, I think would be an absolutely wonderful thing. Because intelligence is the commodity that is most in demand. That’s really what I mean. All the mistakes that humanity makes is because of lack of intelligence really, or lack of knowledge, which is related. So making people smarter, we just can only be better. For the same reason that public education is a good thing and books are a good thing, and the internet is also a good thing, intrinsically and even social networks are a good thing if you run them properly.

(02:40:21) It’s difficult, but you can. Because it helps the communication of information and knowledge and the transmission of knowledge. So AI is going to make humanity smarter. And the analogy I’ve been using is the fact that perhaps an equivalent event in the history of humanity to what might be provided by generalization of AI assistant is the invention of the printing press. It made everybody smarter, the fact that people could have access to books. Books were a lot cheaper than they were before, and so a lot more people had an incentive to learn to read, which wasn’t the case before.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Apple, Alphabet (parent of Google), Microsoft, Meta Platforms, and Tesla. Holdings are subject to change at any time.

What We’re Reading (Week Ending 24 March 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 24 March 2024:

1. The future of ‘communist capitalism’ in China – Martin Wolf

What is the economic future of China? This question raises many specific issues, notably China’s persistent macroeconomic imbalances, the threat of population decline and worsening relations with important parts of the outside world, above all, an increasingly hostile US. But underneath all of these lies a deeper one: is “communist capitalism”, that seemingly self-contradicting invention of Deng Xiaoping, inexorably fading away under Xi Jinping? Will China’s regime ossify and, in the end, collapse, as the Soviet Union did?…

…Much light on this issue is shed by China’s World View, a recently published book by David Daokui Li, a distinguished Harvard-trained professor of economics, who teaches at Tsinghua University. People interested in China, be they hawks or doves, should read Li’s valuable book carefully.

Perhaps its most startling observation is that “from 980 until 1840, the beginning of China’s modern history”, income per head declined. Ancient China was in a Malthusian trap. This picture is even worse than the one shown in the work of the late Angus Maddison. Even after 1840, this grim reality did not get much brighter. Only after Deng Xiaoping’s “reform and opening up” did it change.

By freeing the private economy, relying on market forces and opening up to the world economy, Deng created the conditions for an extraordinary transformation. Yet, by repressing demands for democracy in Tiananmen Square in 1989, he also reinforced communist party control. He invented a new political economy: today’s China is the result.

Is it also sustainable? Li’s book answers a clear “yes” to this question. In essence, he argues that China’s political system should be viewed not as Soviet, but as a modernised form of the traditional Chinese imperial state. This state is paternal. It is responsible for the people, but not accountable to them, except in one fundamental way: if it loses mass support, it will be overthrown. Its job is to provide stability and prosperity. But, in doing so, it does not try to run everything from the centre. That would be crazy in so vast a country: it decentralises to local levels. The communist party should, he argues, be seen fundamentally as the national party of China.

From this perspective, the Xi regime does not represent an abandonment of the goals of the Deng era, but rather an attempt to remedy some of the problems created by its reliance on “go-go” capitalism, namely, pervasive corruption, soaring inequality and environmental damage…

…When considering the prospects for China, one should not focus mainly on the list of obvious problems — falling property prices, excessive debt, excess savings, an ageing population and western hostility. All these can be dealt with by a country with China’s human resources and growth potential, even if with difficulty.

The bigger issue is whether, in the centralising, cautious and conservative era of Xi, Deng’s move from stagnation to explosive growth is doomed to reverse back into stagnation. If people come to believe that the dynamism of the recent past has been lost for good, then there is a risk of a downward spiral of disappointed hopes. But the force of 1.4bn people wanting a better life is extremely powerful. Will anything be allowed to halt it? The answer, I suspect, is still “no”.

2. Is diversification a blessing or curse? – Chin Hui Leong

DIVERSIFICATION is good or bad for you, depending on whom you ask. Warren Buffett, the legendary investor and businessman, once said that if you know what you’re doing, it makes little sense to diversify.

But Peter Lynch, a star mutual fund manager of the 1980s, had a different approach. He believed that the more stocks you own, the better your chances of finding a winner. Lynch was famous for holding up to 1,400 stocks in his portfolio.

Here’s the surprise: They both achieved remarkable success, despite their opposing positions. What does this mean for you as an investor? Should you diversify, or not?…

…As we delve deeper into diversification, we should not lose sight of its goal to reduce risk. This is where buying businesses from unrelated industries or geographies can go wrong. In fact, investors who diversify into areas where they lack expertise are taking more risk, not less. It makes little sense to do so, says Lynch. How well you know your stocks matters more than how many sectors or regions you spread your money across.

I agree with Lynch. Diversify only if you want to boost your chances of finding more winning stocks in your portfolio.

Here is a point you shouldn’t miss: you should always be looking to learn more about new businesses and industries. As you become more knowledgeable, you can grow your portfolio with more stocks you know well, but without exceeding your limits.

Remaining humble is key. Knowing the limits of your knowledge in any new area is how you keep yourself in check. As author Carl Richards once said, risk is what’s left when you think you’ve thought of everything…

…Here’s a simple rule of thumb to help you. If you’ve been following a new company for a year, invest no more than 1 per cent of your portfolio into the stock. If it’s five years, then up to 5 per cent. You can adjust the percentage to fit your risk appetite.

The point of this strategy is to have a reference point where you can match your risk level with your knowledge level…

…Finally, investing over time helps to spread your risk over years. Don’t worry about starting small in a stock. A winning stock is only known in hindsight. Here’s the point most people miss: if a stock is destined to be a winner, the stock price rise will happen over years, if not decades…

…Here’s the final conundrum: the mark of a successful portfolio is a concentrated portfolio. How can that be? Let’s say you invested $1,000 each into 10 stocks. Each stock will make up a tenth of this $10,000 portfolio.

After five years, the first one skyrockets, increasing by 10 times and is worth $10,000, while the last one goes to zero. The other eight stocks stay the same at $1,000. Do the math and you’ll end up with $18,000 in total. The big difference is, the winning stock will comprise more than 55 per cent of the five-year old portfolio…

…As you diversify to find more winners, the best of them will naturally rise to the top – thereby concentrating your portfolio in the right set of winning stocks. That’s more than any investor can wish for.

3. China has little choice but stimulus – Ethan Wu

The near universal reaction in the west to China’s refreshed 5 per cent gross domestic product growth target: good luck with that…

…The old growth drivers — property, infrastructure and manufacturing — all face major constraints. Property’s structural decline is well known; home prices and sales keep falling. Meanwhile, infrastructure is running into the limit of high debt levels. Chinese officials were dispatched last year to prod local governments to delever. It began with easy cost cuts: withholding wages from civil servants, delaying payments to vendors, slashing city services. But more recently, the deleveraging drive has been hitting infrastructure projects already under way, as Reuters reported in January:

Increasing its efforts to manage $13 trillion in municipal debt, the State Council in recent weeks issued a directive to local governments and state banks to delay or halt construction on projects with less than half the planned investment completed in 12 regions across the country, the sources said…

…Lastly, manufacturing. Since about 2020, the credit that once flowed to the property sector has been redirected to manufacturing, especially in politically favoured sectors such as solar and electric vehicles. The year-over-year growth rate of loans to Chinese industry has risen steadily, though the level is now declining..

…This pivot back to manufacturing is “radical”, says Adam Wolfe of Absolute Strategy Research, and it has generated important victories for China. Most notably, BYD is now the world’s biggest EV maker, and China the biggest auto exporter. But it has also created an enormous oversupply of manufactured goods, which, when combined with limp demand at home, is crushing industrial margins and fuelling deflation…

…China’s manufacturing trade surplus is already huge, perhaps 2 per cent of world GDP. As Gavekal’s Yanmei Xie wrote in the FT last month, western countries sensibly fear China dumping cheap goods into export markets. A cheap renminbi heightens the threat; trade retaliation is widely anticipated. If that is right, export-led growth probably can’t be China’s escape valve.

This glum picture suggests that China may soon be forced into stimulus. Assuming the GDP target is at least somewhat binding, no sector of the Chinese economy stands ready to get growth to 5 per cent. A pick-up in consumption could do it, but we’ve heard no convincing story for why anxious consumers would suddenly become gripped by animal spirits…

…The unclear stimulus outlook has left the bulk of investors nervous, but equity outflows have at least stopped. The stock market has rallied 14 per cent since early February, but only because of ample support from the state. Value trade or value trap?

What keeps us sceptical is the fact that Chinese stocks are not loads cheaper than global stocks. After the rally, the CSI 300 trades at 13x forward earnings, versus 14x for the MSCI all-country world ex-US index. To us the risks in China stocks are much clearer than the reward.

4. Exxon Barges in on Hess Deal – Matt Levine

I, on the other hand, used to be a convertible bond investment banker, so I have somewhat more than the usual familiarity with them. I could tell you, for instance, that it is common in the US for a convertible to be done as a Rule 144A offering, meaning that the bonds are sold to large “qualified institutional buyers” (QIBs) in a private placement and then can’t be resold to retail investors. Doing a 144A deal is generally faster and cheaper than doing a public deal that is registered with the US Securities and Exchange Commission, and retail investors don’t really buy convertibles anyway.

But eventually the institutional buyers of a 144A deal will want to be able to convert their bonds into regular, publicly traded stock, so there needs to be some mechanism for turning “144A” convertibles into “registered” ones. I am old enough that, when I started as a converts banker, the way to do this was to file a registration statement with the SEC, but the modern approach is pretty much that you wait six months or a year and the convertible becomes freely tradeable as a legal matter.

As a practical matter, though, the way this works is that the bonds, when they are originally issued, have a “restrictive legend” on them saying that they can be sold only to institutional buyers, and after a year the company sends a notice to its transfer agent saying “you can take that legend off the bonds now.” And when the bonds have the legend, they can’t be freely traded; once the legend is off, they can be. Here I am pretending, as one does, that “the bonds” are pieces of paper with a legend stamped on them, but of course they are actually entries in an electronic database; what really happens is that the original bonds have a “restricted CUSIP” (the identification number that every security has), telling transfer agents and depositaries and brokers and everyone else that they can only be sold to QIBs, and then after a year the company gets them a new “unrestricted CUSIP” and they trade freely. This is not hard — it’s a phone call or an email, maybe a legal opinion — but the company has to do it…

…So for instance here is the indenture for Avid Bioservices Inc.’s 1.25% exchangeable senior notes due 2026, a convertible bond it issued in 2021.4 Section 4.06(e) of the indenture, the 94-page contract governing the bonds, says:

If, and for so long as, the restrictive legend on the Notes specified in ‎‎Section 2.05(c) has not been removed, the Notes are assigned a restricted CUSIP or the Notes are not otherwise freely tradable … as of the 370th day after the last date of original issuance of the Notes, the Company shall pay Additional Interest on the Notes at a rate equal to 0.50% per annum of the principal amount of Notes outstanding until the restrictive legend on the Notes has been removed. …

…Avid forgot, for two years, to take the restrictive legend off of its convertible. This was very understandable: Its obligation to remove the restricted legend was boring and technical and buried in Section 4.06(e) of a bond indenture that surely nobody read. It could only remove the legend a year after it issued the bonds, after everyone had stopped paying attention. And, as Avid points out, it “did not receive any notices and was not otherwise made aware” of this provision in, sure, a contract that it signed, but a very long and boring contract. (And, to be fair, the holders forgot too!) And because it completely forgot about its obligation to remove the legend, Avid also forgot to pay the 0.5% penalty interest rate for two years. And because it forgot to pay the extra interest, it created a non-curable default on the bonds: The holders can demand all of their money back, with interest, immediately, with no chance for Avid to fix the problem by removing the legend and paying the overdue interest…

…This is a bad oopsie by Avid, which probably should have put a reminder in its calendar to unrestrict the CUSIP. But it’s a clever trade by whoever this holder was: The old bonds are far out-of-the-money (that is, they’re not going to convert into stock), and Bloomberg tells me that they were trading in the high 70s as recently as a month ago (the high 80s more recently). If you had noticed Avid’s extremely technical oopsie, you could have bought the bonds at, say, 80 cents on the dollar, sent them a letter saying “we gotcha hahahaha,” and made a quick 20 points, plus interest. The holder owns “at least 25%” of the bonds (the amount required to accelerate), and there are $143.75 million of bonds outstanding; 20 points on 25% of $143.75 million is $7.2 million. Plus interest.

5. Sora, Groq, and Virtual Reality – Ben Thompson

Groq was founded in 2016 by Jonathan Ross, who created Google’s first Tensor Processing Unit; Ross’s thesis was that chips should take their cue from software-defined networking: instead of specialized hardware for routing data, a software-defined network uses commodity hardware with a software layer to handle the complexity of routing. Indeed, Groq’s paper explaining their technology is entitled “A Software-defined Tensor Streaming Multiprocessor for Large-scale Machine Learning.”

To that end Groq started with the compiler, the software that translates code into machine language that can be understood by chips; the goal was to be able to reduce machine-learning algorithms into a format that could be executed on dramatically simpler processors that could operate at very high speed, without expensive memory calls and prediction misses that make modern processors relatively slow.

The end result is that Groq’s chips are purely deterministic: instead of the high-bandwidth memory (HBM) used for modern GPUs or Dynamic Random Access Memory (DRAM) used in computers, both of which need to be refreshed regularly to function (which introduces latency and uncertainty about the location of data at a specific moment in time), Groq uses SRAM — Static Random Access Memory. SRAM stores data in what is called a bistable latching circuitry; this, unlike the transistor/capacitor architecture undergirding DRAM (and by extension, HBM), stores data in a stable state, which means that Groq always knows exactly where every piece of data is at any particular moment in time. This allows the Groq compiler to, in an ideal situation, pre-define every memory call, enabling extremely rapid computation with a relatively simple architecture.

It turns out that running inference on transformer-based models is an extremely ideal situation, because the computing itself is extremely deterministic. An LLM like GPT-4 processes text through a series of layers which have a predetermined set of operations, which is perfectly suited to Groq’s compiler. Meanwhile, token-based generation is a purely serial operation: every single token generated depends on knowing the previous token; there is zero parallelism for any one specific answer, which means the speed of token calculation is at an absolute premium…

…One of the arguments I have made as to why OpenAI CEO Sam Altman may be exploring hardware is that the closer an AI comes to being human, the more grating and ultimately gating are the little inconveniences that get in the way of actually interacting with said AI. It is one thing to have to walk to your desk to use a PC, or even reach into your pocket for a smartphone: you are, at all times, clearly interacting with a device. Having to open an app or wait for text in the context of a human-like AI is far more painful: it breaks the illusion in a much more profound, and ultimately disappointing, way. Groq suggests a path to keeping the illusion intact.

It is striking that Groq is a deterministic system running deterministic software that, in the end, produces probabilistic output. I explained deterministic versus probabilistic computing in ChatGPT Gets a Computer:

Computers are deterministic: if circuit X is open, then the proposition represented by X is true; 1 plus 1 is always 2; clicking “back” on your browser will exit this page. There are, of course, a huge number of abstractions and massive amounts of logic between an individual transistor and any action we might take with a computer — and an effectively infinite number of places for bugs — but the appropriate mental model for a computer is that they do exactly what they are told (indeed, a bug is not the computer making a mistake, but rather a manifestation of the programmer telling the computer to do the wrong thing).

I’ve already mentioned Bing Chat and ChatGPT; on March 14 Anthropic released another AI assistant named Claude: while the announcement doesn’t say so explicitly, I assume the name is in honor of the aforementioned Claude Shannon. This is certainly a noble sentiment — Shannon’s contributions to information theory broadly extend far beyond what Dixon laid out above — but it also feels misplaced: while technically speaking everything an AI assistant is doing is ultimately composed of 1s and 0s, the manner in which they operate is emergent from their training, not proscribed, which leads to the experience feeling fundamentally different from logical computers — something nearly human — which takes us back to hallucinations; Sydney was interesting, but what about homework?

The idea behind ChatGPT Gets a Computer is that large language models seem to operate somewhat similarly to the human brain, which is incredible and also imprecise, and just as we need a computer to do exact computations, so does ChatGPT. A regular computer, though, is actually the opposite of Groq: you get deterministic answers from hardware that is, thanks to the design of modern processors and memory, more probabilistic than you might think, running software that assumes the processor will handle endless memory calls and branch prediction.

In the end, though, we are back where we started: a computer would know where the bow and stern are on a ship, while a transformer-based model like Sora made a bad guess. The former calculates reality; the latter a virtual reality.

Imagine, though, Sora running on Groq (which is absolutely doable): could we have generated videos in real-time? Even if we could not, we are certainly much closer than you might have expected. And where, you might ask, would we consume those videos? How about on a head-mounted display like the Apple Vision Pro or Meta Quest? Virtual reality (my new definition) for virtual reality (the old definition).


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Apple and Meta Platforms. Holdings are subject to change at any time.

What We’re Reading (Week Ending 17 March 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 17 March 2024:

1. The Ultra-Pure, Super-Secret Sand That Makes Your Phone Possible – Vince Beiser

Spruce Pine is not a wealthy place. Its downtown consists of a somnambulant train station across the street from a couple of blocks of two‑story brick buildings, including a long‑closed movie theater and several empty storefronts.

The wooded mountains surrounding it, though, are rich in all kinds of desirable rocks, some valued for their industrial uses, some for their pure prettiness. But it’s the mineral in Glover’s bag—snowy white grains, soft as powdered sugar—that is by far the most important these days. It’s quartz, but not just any quartz. Spruce Pine, it turns out, is the source of the purest natural quartz—a species of pristine sand—ever found on Earth. This ultra‑elite deposit of silicon dioxide particles plays a key role in manufacturing the silicon used to make computer chips. In fact, there’s an excellent chance the chip that makes your laptop or cell phone work was made using sand from this obscure Appalachian backwater. “It’s a billion‑dollar industry here,” Glover says with a hooting laugh. “Can’t tell by driving through here. You’d never know it.”

In the 21st century, sand has become more important than ever, and in more ways than ever. This is the digital age, in which the jobs we work at, the entertainment we divert ourselves with, and the ways we communicate with one another are increasingly defined by the internet and the computers, tablets, and cell phones that connect us to it. None of this would be possible were it not for sand.

Most of the world’s sand grains are composed of quartz, which is a form of silicon dioxide, also known as silica. High‑purity silicon dioxide particles are the essential raw materials from which we make computer chips, fiber‑optic cables, and other high‑tech hardware—the physical components on which the virtual world runs. The quantity of quartz used for these products is minuscule compared to the mountains of it used for concrete or land reclamation. But its impact is immeasurable…

…In the mid‑1950s, thousands of miles from North Carolina, a group of engineers in California began working on an invention that would become the foundation of the computer industry. William Shockley, a pathbreaking engineer at Bell Labs who had helped invent the transistor, had left to set up his own company in Mountain View, California, a sleepy town about an hour south of San Francisco, near where he had grown up. Stanford University was nearby, and General Electric and IBM had facilities in the area, as well as a new company called Hewlett‑Packard. But the area known at the time as the Santa Clara Valley was still mostly filled with apricot, pear, and plum orchards. It would soon become much better known by a new nickname: Silicon Valley.

At the time, the transistor market was heating up fast. Texas Instruments, Motorola, and other companies were all competing to come up with smaller, more efficient transistors to use in, among other products, computers. The first American computer, dubbed ENIAC, was developed by the army during World War II; it was 100 feet long and 10 feet high, and it ran on 18,000 vacuum tubes.

Transistors, which are tiny electronic switches that control the flow of electricity, offered a way to replace those tubes and make these new machines even more powerful while shrinking their tumid footprint. Semiconductors—a small class of elements, including germanium and silicon, which conduct electricity at certain temperatures while blocking it at others—looked like promising materials for making those transistors.

At Shockley’s startup, a flock of young PhDs began each morning by firing up kilns to thousands of degrees and melting down germanium and silicon. Tom Wolfe once described the scene in Esquire magazine: “They wore white lab coats, goggles, and work gloves. When they opened the kiln doors weird streaks of orange and white light went across their faces . . . they lowered a small mechanical column into the goo so that crystals formed on the bottom of the column, and they pulled the crystal out and tried to get a grip on it with tweezers, and put it under microscopes and cut it with diamond cutters, among other things, into minute slices, wafers, chips; there were no names in electronics for these tiny forms.”

Shockley became convinced that silicon was the more promising material and shifted his focus accordingly. “Since he already had the first and most famous semiconductor research and manufacturing company, everyone who had been working with germanium stopped and switched to silicon,” writes Joel Shurkin in his biography of Shockley, Broken Genius. “Indeed, without his decision, we would speak of Germanium Valley.”

Shockley was a genius, but by all accounts he was also a lousy boss. Within a couple of years, several of his most talented engineers had jumped ship to start their own company, which they dubbed Fairchild Semiconductor. One of them was Robert Noyce, a laid‑back but brilliant engineer, only in his mid‑20s but already famous for his expertise with transistors.

The breakthrough came in 1959, when Noyce and his colleagues figured out a way to cram several transistors onto a single fingernail‑sized sliver of high‑purity silicon. At almost the same time, Texas Instruments developed a similar gadget made from germanium. Noyce’s, though, was more efficient, and it soon dominated the market. NASA selected Fairchild’s microchip for use in the space program, and sales soon shot from almost nothing to $130 million a year. In 1968, Noyce left to found his own company. He called it Intel, and it soon dominated the nascent industry of programmable computer chips.

Intel’s first commercial chip, released in 1971, contained 2,250 transistors. Today’s computer chips are often packed with transistors numbering in the billions. Those tiny electronic squares and rectangles are the brains that run our computers, the Internet, and the entire digital world. Google, Amazon, Apple, Microsoft, the computer systems that underpin the work of everything from the Pentagon to your local bank—all of this and much more is based on sand, remade as silicon chips.

Making those chips is a fiendishly complicated process. They require essentially pure silicon. The slightest impurity can throw their tiny systems out of whack.

Finding silicon is easy. It’s one of the most abundant elements on Earth. It shows up practically everywhere bound together with oxygen to form SiO2, aka quartz. The problem is that it never occurs naturally in pure, elemental form. Separating out the silicon takes considerable doing.

Step one is to take high‑purity silica sand, the kind used for glass. (Lump quartz is also sometimes used.) That quartz is then blasted in a powerful electric furnace, creating a chemical reaction that separates out much of the oxygen. That leaves you with what is called silicon metal, which is about 99 percent pure silicon. But that’s not nearly good enough for high‑tech uses. Silicon for solar panels has to be 99.999999 percent pure—six 9s after the decimal. Computer chips are even more demanding. Their silicon needs to be 99.99999999999 percent pure—eleven 9s. “We are talking of one lonely atom of something that is not silicon among billions of silicon companions,” writes geologist Michael Welland in Sand: The Never-Ending Story.

Getting there requires treating the silicon metal with a series of complex chemical processes. The first round of these converts the silicon metal into two compounds. One is silicon tetrachloride, which is the primary ingredient used to make the glass cores of optical fibers. The other is trichlorosilane, which is treated further to become polysilicon, an extremely pure form of silicon that will go on to become the key ingredient in solar cells and computer chips.

Each of these steps might be carried out by more than one company, and the price of the material rises sharply at each step. That first‑step, 99 percent pure silicon metal goes for about $1 a pound; polysilicon can cost 10 times as much.

The next step is to melt down the polysilicon. But you can’t just throw this exquisitely refined material in a cook pot. If the molten silicon comes into contact with even the tiniest amount of the wrong substance, it causes a ruinous chemical reaction. You need crucibles made from the one substance that has both the strength to withstand the heat required to melt polysilicon, and a molecular composition that won’t infect it. That substance is pure quartz.

THIS IS WHERE Spruce Pine quartz comes in. It’s the world’s primary source of the raw material needed to make the fused‑quartz crucibles in which computer‑chip‑grade polysilicon is melted. A fire in 2008 at one of the main quartz facilities in Spruce Pine for a time all but shut off the supply of high‑purity quartz to the world market, sending shivers through the industry.

Today one company dominates production of Spruce Pine quartz. Unimin, an outfit founded in 1970, has gradually bought up Spruce Pine area mines and bought out competitors, until today the company’s North Carolina quartz operations supply most of the world’s high‑ and ultra‑high‑purity quartz. (Unimin itself is now a division of a Belgian mining conglomerate, Sibelco.)

In recent years, another company, the imaginatively titled Quartz Corp, has managed to grab a small share of the Spruce Pine market. There are a very few other places around the world producing high‑purity quartz, and many other places where companies are looking hard for more. But Unimin controls the bulk of the trade.

The quartz for the crucibles, like the silicon they will produce, needs to be almost absolutely pure, purged as thoroughly as possible of other elements. Spruce Pine quartz is highly pure to begin with, and purer still after being put through several rounds of froth flotation. But some of the grains may still have what Glover calls interstitial crystalline contamination—molecules of other minerals attached to the quartz molecules.

That’s frustratingly common. “I’ve evaluated thousands of quartz samples from all over the world,” says John Schlanz, chief minerals processing engineer at the Minerals Research Laboratory in Asheville, about an hour from Spruce Pine. “Near all of them have contaminate locked in the quartz grains that you can’t get out.”

Some Spruce Pine quartz is flawed in this way. Those grains are used for high‑end beach sand and golf course bunkers—most famously the salt‑white traps of Augusta National Golf Club, site of the iconic Masters Tournament. A golf course in the oil‑drunk United Arab Emirates imported 4,000 tons of this sand in 2008 to make sure its sand traps were world‑class, too.

The very best Spruce Pine quartz, however, has an open crystalline structure, which means that hydrofluoric acid can be injected right into the crystal molecules to dissolve any lingering traces of feldspar or iron, taking the purity up another notch. Technicians take it one step further by reacting the quartz with chlorine or hydrochloric acid at high temperatures, then putting it through one or two more trade‑secret steps of physical and chemical processing.

The result is what Unimin markets as Iota quartz, the industry standard of purity. The basic Iota quartz is 99.998 percent pure SiO2. It is used to make things like halogen lamps and photovoltaic cells, but it’s not good enough to make those crucibles in which polysilicon is melted. For that you need Iota 6, or the tip‑top of the line, Iota 8, which clocks in at 99.9992 percent purity—meaning for every one billion molecules of SiO , there are only 80 molecules of impurities. Iota 8 sells for up to $10,000 a ton. Regular construction sand, at the other end of the sand scale, can be had for a few dollars per ton…

…Unimin sells this ultra‑high‑purity quartz sand to companies like General Electric, which melts it, spins it, and fuses it into what looks like a salad bowl made of milky glass: the crucible. “It’s safe to say the vast majority of those crucibles are made from Spruce Pine quartz,” Schlanz says.

The polysilicon is placed in those quartz crucibles, melted down, and set spinning. Then a silicon seed crystal about the size of a pencil is lowered into it, spinning in the opposite direction. The seed crystal is slowly withdrawn, pulling behind it what is now a single giant silicon crystal. These dark, shiny crystals, weighing about 220 pounds, are called ingots.

The ingots are sliced into thin wafers. Some are sold to solar cell manufacturers. Ingots of the highest purity are polished to mirror smoothness and sold to a chipmaker like Intel. It’s a thriving multi-billion dollar industry in 2012.

The chipmaker imprints patterns of transistors on the wafer using a process called photolithography. Copper is implanted to link those billions of transistors to form integrated circuits. Even a minute particle of dust can ruin the chip’s intricate circuitry, so all of this happens in what’s called a clean room, where purifiers keep the air thousands of times cleaner than a hospital operating room. Technicians dress in an all‑covering white uniform affectionately known as a bunny suit. To ensure the wafers don’t get contaminated during manufacture, many of the tools used to move and manipulate them are, like the crucibles, made from high‑purity quartz.

The wafers are then cut into tiny, unbelievably thin quadrangular chips—computer chips, the brains inside your mobile phone or laptop. The whole process requires hundreds of precise, carefully controlled steps. The chip that results is easily one of the most complicated man‑made objects on Earth, yet made with the most common stuff on Earth: humble sand.

The total amount of high‑purity quartz produced worldwide each year is estimated at 30,000 tons—less than the amount of construction sand produced in the United States every hour. (And even construction sand is in high demand; there’s a thriving black market in the stuff.) Only Unimin knows exactly how much Spruce Pine quartz is produced, because it doesn’t publish any production figures. It is an organization famously big on secrecy. “Spruce Pine used to be mom‑and‑ pop operations,” Schlanz says. “When I first worked up there, you could just walk into any of the operations. You could just go across the street and borrow a piece of equipment.”

NOWADAYS UNIMIN WON’T even allow staff of the Minerals Research Laboratory inside the mines or processing facilities. Contractors brought in to do repair work have to sign confidentiality agreements. Whenever possible, vice‑president Richard Zielke recently declared in court papers, the company splits up the work among different contractors so that no individual can learn too much.

Unimin buys equipment and parts from multiple vendors for the same reason. Glover has heard of contractors being blindfolded inside the processing plants until they arrive at the specific area where their jobs are and of an employee who was fired on the spot for bringing someone in without authorization. He says the company doesn’t even allow its employees to socialize with those of their competitors.

It was hard to check out Glover’s stories, because Unimin wouldn’t talk to me. Unlike most big corporations, its website lists no contact for a press spokesperson or public relations representative. Several emails to their general inquiries address went unanswered. When I called the company’s headquarters in Connecticut, the woman who answered the phone seemed mystified by the concept of a journalist wanting to ask questions.

She put me on hold for a few minutes, then came back to tell me the company has no PR department, but that if I faxed (faxed!) her my questions, someone might get back to me. Eventually I got in touch with a Unimin executive who asked me to send her my questions by email. I did so. The response: “Unfortunately, we are not in a position to provide answers at this point in time.”

2. It was never about LLM performance – Justin

The LLM community is obsessed with benchmarking model performance. Mistral released their new “flagship” model this week, and immediately focused the discussion on how it performs on “commonly used benchmarks” relative to other models:

The entire blog post (I’d recommend reading it) is just a read through of how this model performs relative to other models on benchmarks, from math and coding to multilingual capabilities…

…This tendency to fixate on benchmarks is understandable – right now, it’s basically the only semi-objective way to measure how these models stack up against each other. It’s something vendors in other spaces, like data streaming, do too. But it is dangerous because it misses the point of where this whole AI thing is going, and is a textbook product marketing anti-pattern.

In a trend that we’ve seen hundreds of times in developer tooling, the underlying LLM is not going to matter within a few years. Large Language Model performance is already highly commoditized, and will continue to head in that direction. All that will matter is the experience that you build on top of these models, and what that enables for your customers.

Let’s take a look at the ChatGPT interface. Here’s a common prompt I’ve been using for testing, asking the model to summarize the contents of an external link into a tweet thread. Unrelated aside, the responses to this prompt are virtually identical across every major LLM.

Which parts of this interface are the underlying model – GPT-4 in this case – and which are an experience built by OpenAI on top of the underlying model?

The text response, minus any formatting, is what the model generated. But the:

  • Ability of the model to access and scrape content from a web page
  • Context of the prompt, including setting the system as a helpful assistant
  • Formatting the response, like changing the numbers to gray UI for typing the prompt
  • Filepicker for attaching media to the prompt
  • Prompt history
  • Model switcher / picker (this one is meta)
  • Ability to persist and share the model responses

…and more not show here

are all not GPT-4, they’re features built by OpenAI on top of GPT-4 to create an experience that is helpful and worth paying for. Some of these are harder to build than others – OpenAI’s secret sauce obviously isn’t the little arrow that scrolls down to the bottom of the response. ChatGPT would be nothing without GPT-4 – but the reverse may also be true!

The retort to this line of reasoning is that these chat interfaces are primarily for non-technical users, while the real money for these model providers comes from developer use cases, building LLMs into user-facing applications. I’ve worked closely with one of the major model compute providers, so this is not foreign to me. But experience matters to developers too!

OpenAI has dedicated significant resources to building a seamless developer experience beyond “docs for the model.” Here’s their playground for prompting GPT models – you can adjust parameters like temperature and penalties, plus change the system prompt to be any other style…

…For a closed source model provider like OpenAI, the difference between what is model and what is experience is academic – you’re paying for both. They are one thing. But where this really matters is in open source. Does the convergence of open source performance to closed source performance really matter if the experience of using that open source is bad?…

…The open source discussion has been too anchored on reaching performance parity with OpenAI models. This is a small piece of the puzzle. For developers looking to build applications with these open source models, and especially the pro-sumer chat use case, users need to consider the holistic experience that model providers offer. Integrating LLMs into your app is almost never going to be the “drop in” experience you see on marketing sites – and my concern is that the “open source is approaching parity with OpenAI!” narrative is not actually true in a meaningful way.

Folks working in AI can look to previous examples of this phenomenon in developer tools for guidance: A couple of years ago, I wrote about how underlying performance of production relational databases is becoming commoditized, and vendors are focusing much more on developer experience. It’s going to happen here too, the question is just when.

3. Aravind Srinivas – Building An Answer Engine – Patrick O’Shaughnessy and Aravind Srinivas

Patrick: [00:07:28] It’s really cool to think about the sequencing to get there. We’ve had search engines. Like you said, it’s a hack to get the answers. You’re building what I think of today as an answer engine. I type something in, it’s just giving the answer directly with great citation and all this other stuff we’ll talk about. And the vision you’re articulating is this question engine can anticipate the things that I want to learn about and give them to me beforehand.

And I’d love to build up towards that. So maybe starting with the answer engine, explain to us how it works. Maybe you could do this via the time line of how you’ve built the product or something. But what are the components? What is happening behind the scenes when I type something into Perplexity either a question or a search query or whatever? Walk us through in some detail the actual goings on behind the scenes in terms of how the product works itself?

Aravind: [00:08:13] Yes. So when you type in a question into Perplexity, the first thing that happens is, it first reformulates the question, it tries to understand the question better, expands the question in terms of adding more suffixes or prefixes to it, to make it more well formatted. It speaks to the question engine part. And then after that, it goes and pulls so many links from the web that are relevant to this reformulated question.

There are so many paragraphs in each of those links. It takes only the relevant paragraphs from each of those links. And then an AI model, we typically call it large language model. It’s basically a model that’s been trained to predict the next word on the Internet and fine-tuned for being good at summarization and chats.

That AI model looks at all these chunks of knowledge the bits of study that surface from important or relevant links and takes only those parts that are relevant to answering your query and gives you a very concise four or five sentence answer, but also with references. Every sentence has a reference to which webpage or which chunk of knowledge it took from which webpage and puts it at the top in terms of sources.

That gets you to a nicely formatted rendered answer, sometimes in markdown bullets, or sometimes just generic paragraphs, sometimes it has images in it. But a great answer with references or citation so that if you want to dig deeper, you can go and visit the link. If you don’t want and just read the answer and ask a follow-up, you can engage in a conversation, both modes of usage are encouraged and allowed. So this is what happens on Perplexity today.

Patrick: [00:09:51] What percent of users end up clicking beneath the summarized answer into a source webpage?

Aravind: [00:10:01] At least 10%.

Patrick: [00:10:02] So 90% of the time, they’re just satisfied with what you give them?

Aravind: [00:10:06] It depends on how you look at it. If you wanted to be 100% of the time, people always click on a link, that’s the traditional Google. And you want to be 100% of the time where people never click on links, that’s ChatGPT. We think the sweet spot is somewhere in the middle. People should click on link sometimes to go do their work there. Let’s say, you’re just booking a ticket, you might actually want to go away Expedia or something.

Let’s say you’re deciding where to go first. You don’t need to go away and read all these SEO blogs and get confused on what you want to do. You first make your decision independently with this research body that’s helping you decide. And once you finished your research and you have decided, then that’s when you actually have to go out and do your actual action of booking your ticket. That way, I believe there is a nice sweet spot of one product providing you both the navigational search experience as well as the answer engine experience together. And that’s what we strive to be doing…

Patrick: [00:13:54] Can you explain from an insider’s perspective and someone building an application on top of these incredible new technologies, what do you think the future might look like or even what you think the ideal future would be for how many different LLM providers there are, how specialized they get to scale the primary answer, so there’s only going to be a few of them. How do you think about all this and where you think it might go?

Aravind: [00:14:16] It really depends on who you’re building for. If you’re building for consumers, you do want to build a scalable infrastructure because you do want to ask many consumers to use your product. If you’re building for the enterprise, you still want a scalable infrastructure.

Now it really depends, are you building for the people within that company who are using your product. Let’s say, you’re building an internal search engine, you only need to scale to the size of the largest organization, which is like maybe 100,000 people. And not all of them will be using your thing at one moment. You’re decentralizing it, you’re going to keep different servers for different companies and you can elastically decide what’s the level of throughput you need to offer.

But then if you’re solving another enterprise’s problem, where that enterprise is serving consumers and you’re helping them do that, you need to build scalable infrastructure indirectly at least. For example, OpenAI. Their APIs are used by us, other people to serve a lot of consumers. So unless they solve that problem themselves, they’re unable to help other people solve their problem. Same thing with AWS.

So that’s one advantage you have of actually having a first-party product that your infrastructure is helping you serve. And by doing that, by forcing yourself to solve that hard problem, whatever you build can be used by others as well. Amazon build AWS first for Amazon. And because Amazon.com requires very robust infrastructure, that can be used by so many other people and so many other companies emerged by building on top of AWS.

Same thing happened with OpenAI. They needed robust infrastructure to serve the GPT-3 developer API and ChatGPT as a product. But once they got it all right, then they can now support other companies that are building on top of them. So it really depends on what’s your end goal and who you’re trying to serve and what’s the scale of our ambition…

Patrick: [00:19:02] And when I think about the history of the product, which I was a pretty early user of, the first thing that pops to my mind is that it solves the hallucination problem, which has become less of a problem. But early on, everyone just didn’t know how to trust these things and you solved that. You gave citations, you can click through the underlying webpages, et cetera.

I’d love you to walk through what you view the major time line product milestones have been of Perplexity dating back to its start. The one I just gave could be one example. There was this possibility, but there was a problem and you solved it, at least that was my perception as a user. What have been the major milestones as you think back on the product and how it’s gotten better?

Aravind: [00:19:41] I would say the first major thing we did is really making the product a lot faster. When we first launched, the latency for every query was seven seconds, then we actually had to speed up the demo video to put it on Twitter so that it doesn’t look embarrassing.

And one of our early friendly investors, Daniel Gross who co-invests a lot with Nat Friedman, he was one of our first testers before we even released the product. And he said, you guys should call it a submit button for a query. It’s almost like you’re submitting a job and waiting on the cluster to get back. It’s that slow.

And now we are widely regarded as the fastest chatbot out there. Some people even come and ask me, why are you only as fast as ChatGPT? Why are you not faster? And little did they realize that ChatGPT doesn’t even use the web by default. It only uses it on the browsing mode on Bing.

So for us to be as fast as ChatGPT already tells you that in spite of doing more work to go pull up links from the web, read the chunks, pick the relevant ones and use that to give you the answer with sources and a lot more work on the rendering, despite doing all the additional work, if you’re managing an end-to-end latency as good as ChatGPT that shows we have like even a superior back end to them.

So I’m most proud about the speed at which we can do things today compared to when we launched, the accuracy has been constantly going up, primarily few things. One is we keep expanding our index and like keep improving the quality of the index. From the beginning, we knew all the mistakes that previous Google competitors did, which is obsessed about the size of your index and focus less on the quality.

So we decided from the beginning we would not obsess about the size. Size doesn’t matter and index actually, what matters is the quality of your index. What kind of domains are important for AI chatbots and question-answering and knowledge workers. That is what we care about. So that decision ended up being right.

The other thing that has helped us improve the accuracy was training these models to be focused on hallucinations. When you don’t have enough information in the search snippets, try to just say I don’t know, instead of making up things. LLMs are conditioned to always be helpful, will always try to serve the user’s query despite what it has access to, may not be even sufficient to answer the query. So that part took some reprogramming, rewiring. You’ve got to go and change the ways. You can’t just solve this with prompt engineering. So we have spent a lot of work on that.

The other thing I’m really proud about is getting our own inference infrastructure. So when you have to move outside the OpenAI models to serve your product, everybody thinks, “Oh, you just train a model to be as good as GPT and you’re’ done.” But reality is OpenAI’s mode is not just in the fact that they have trained the best models, but also that they have the most cost-efficient, scalable infrastructure for serving this on a large-scale consumer product like ChatGPT. That is itself a separate layer of mode. You can build that mode, you can build.

And so we are very proud of our inference team, how fast, high throughput, low latency infrastructure we built for serving our own LLMs. We took advantage of the open source revolution, Llama and Mistral and took all these models, trained them to be very good at being great answer bots and served them ourselves on GPU so that we get better margins on our product. So all these three layers, both in terms of speed through actual product back-end orchestration, accuracy of the AI models and serving our own AI models, we’ve done a lot of work on all these things…

Patrick: [00:28:50] Can you expand on index. You’ve referenced that a few times for those that haven’t built one or haven’t thought about this. Just explain that whole concept and the decisions that you’ve made. You already mentioned quality versus size. But just explain what it means to build an index, why it’s so important, et cetera?

Aravind: [00:29:07] Yes. So what does an index mean, it’s basically a copy of the web. The web has so many links and you want a cache, you want a copy of all those links in the database, so a URL and the contents in that URL. Now the challenge here is new links are being created every day on the web and also existing links keep getting updated on the web as well. New sites keep getting updated. So you’ve got to periodically refresh them. The URL needs to be updated in the cache with a different version of it.

Similarly, you got to keep adding new URLs to your index, which means you’ve got to build a crawler. And then how you store a URL, the contents in that URL also matters. Not every page is native HTML anymore. The web has upgraded a lot, rendering JavaScript a lot, and every domain has custom-based rendered the JavaScript. So you’ve got to build parsers. So you’ve got to build a crawler, indexer, parser and that together makes up for a great index.

Now the next step comes to retrieval, which is now that you have those index, every time you hit a query, which links do you use? And which paragraphs in those links do you use? Now that is the ranking problem. How do you figure out what is relevance and ranking? And once you retrieve those chunks, like the top few chunks relevant to a query that the user is asking, that’s when the AI model comes in. So this is a retrieve part. Now the generic part. That’s why it’s called retrieve and generic.

So once you retrieve the relevant chunks from the huge index that you have, the AI model will come and read those chunks and then give you the answer. Doing this ensures that you don’t have to keep training the AI model to be up to date. What you want the AI model to do is to be intelligent, to be a good reasoning model.

Think about this as when you were a student, I’m sure you would have written an open book exam, open notes exam in school or high school or college. What are those exams test you for? They don’t test you for rote learning. So it doesn’t give an advantage to the person who has the best memory power. It gives advantage to person who has read the concepts, can immediately query the right part of the notes, but the questions required you to think on the fly as well.

That’s what we want to design systems. It’s very different philosophy from OpenAI, where OpenAI wants this one model that’s so intelligent, so smart, you can just ask it anything. It’s going to tell you. We rather want to build a small efficient model that’s smart, capable, can reason on facts that it’s given on the fly. And this ambiguate different individuals with different names or saved as not sufficient information, not get confused about dates.

When you’re asking something about the future, say that was not yet happened. These sort of corner cases handle all of those with good reasoning capabilities yet have access to all of the world’s knowledge at an instant through a great index. And if you can do both of these together end-to-end orchestrated with great latency and user experience, you’re creating something extremely valuable. So that’s what we want to build…

Patrick: [00:37:26] Do you think that the transformer architecture is here to stay and will remain the dominant tool or architecture for a long time?

Aravind: [00:37:33] This is a question that everybody asks in the last six years or seven years since the first transformer came. Honestly, nothing has changed. The only thing that has changed is the transformer became a mixture of experts model, where there are multiple models and not just a single model. But the core self-attention model architecture has not changed. And people say there are shortcomings, the quadratic attention, complexities there. But any solution to that incurs costs somewhere else too.

Most of the people are not aware that majority of the computation in a large transformer like GPT-3 or 4 is not even spent on the attention layer. It’s actually spent on the matrix multiplies. So if you’re trying to focus more on the quadratic part, you’re incurring costs and the matrix multiples, and that’s actually the bottleneck in the larger scaling.

So honestly, it’s very hard to make an innovation on the transformer that can have a material impact at the level of GPT-4 complex cost of training those models. So I would bet more on innovations, auxiliary layers, like retrievable augmented generation. Why do you want to train a really large model when you don’t have to memorize all the facts from the Internet, when you literally have to just be a good reasoning model?

Nobody is going to value Patrick for knowing all facts. They’re going to value you for being an intelligent person, fluid intelligence. If I give you something very new that nobody else has an experience in, are you well positioned to learn that skill fast and start doing it really well. When you hire a new employee, what do you care about? Do you care about how much they know about something? Or do you care about whether you can give them any task and they would still get up to speed and do it, which employee would you value more?

So that’s the sort of intelligence that we should bake into these models, and that requires you to think more on the data. What are these models training on? Can we make them train on something else and just memorizing all the words on the Internet? Can we make reasoning emerge in these models through a different way? And that might not need innovation on the transformer, that may need innovation more on what data you’re throwing at these models.

Similarly, another layer of innovation that’s waiting to happen is the architecture like sparse versus dense models. Clearly, mixture of experts is working, GPT-4 is a mixture of experts, Mixtral is a mixture of experts, Gemini 1.5 is a mixture of experts. So even there, it’s not one model for coding, one model for reasoning and math, one model for history that depending on your input, it’s getting routed to the right model. It’s not that spares.

Every individual tokened is routed to a different model, but it’s happening every layer. So you’re still spending a lot of compute. How can we create something that’s actually 100 humans in one company? So the company itself has aggregated so much smarter. We’ve not created the equivalent at a model layer, more experimentation on the sparsity and more experimentation on how we can make reasoning emerge in a different way is likely to have a lot more impact than thinking about what is the next transformer.

4. Training great LLMs entirely from ground up in the wilderness as a startup – Yi Tay

People always assume it’s simply a question/debate of accelerator choice (TPUs vs GPUs etc) and all GPU clusters are created equal. For us, this soon proved to be false. As we sampled across different service providers, we find that the variance of hardware quality differs vastly even for the same hardware, i.e., GPUs (H100s). Note that here, hardware refers to overall cluster quality and not necessarily the chips or accelerators per se. Just like a lottery. Basically:

Not all hardware is created equal. The variance of cluster quality across hardware providers is so high that it is literally a lottery pertaining to how much pain one would have to go through to train good models. In short, a hardware lottery in the era of LLMs.

More specifically, we’ve leased a few clusters from several compute providers, each with a range of hundreds to thousands of chips. We’ve seen clusters that range from passable (just annoying problems that are solvable with some minor SWE hours) to totally unusable clusters that fail every few hours due to a myriad of reasons. Specifically, some clusters have nodes that fail every N hour with issues ranging from cabling issues (where N is unreasonably small), GPU hardware errors etc. Even more surprisingly, every cluster across the same provider could also be vastly different in terms of how robust it was…

…Did I mention you’ll also get a different Model Flop Utilisation (MFU) for different clusters!? This was a non negligible amount of compute wasted if one is unlucky enough to find a provider with badly cabled nodes or some other issues. Systems with very sub-optimal file systems would have the MFU of training runs tank the moment a team mate starts transferring large amounts of data across clusters.

Every service provider also had different levels of support. These range from being polite to nonchalant, “chatgpt-style” canned responses to blaming the user for every single thing that goes wrong.

Overall, every single cluster we tried feels like they have their own vibe, struggles and failure modes. It was also almost as though every single cluster needed their own hot-fixes for their own set of issues – some more tolerable than others. That said, we’ve learned that fail safes are important, and finding fast hot fixes for any clusters could be key…

…We’re training our models on GPUs for the most part at Reka. Personally, I’ve used TPUs all my life when it comes to large language model training at Google pre-Reka life. CUDA and nccl were the most alien thing to me ever. (I only learned it’s pronounced “Nickel” from one of my coworkers who used to work at Nvidia lol)

I was completely taken aback by the failure rate of GPUs as opposed to my experiences on TPUs at Google. In fact, I don’t actually recall TPUs failing much even for large runs, though I was not sure if I was protected from knowing this just by the sheer robustness of the outrageously good infra and having a dedicated hardware team. In fact, the UL2 20B model (at Google) was trained by leaving the job running accidentally for a month. It never failed. If this were in GPU land, it would have failed within the first few days for sure.

That said, I think this could be more about the competency of the hardware team that manages your accelerators rather than the underlying chip. The presence of having good hardware support (from your compute provider) is important. And so much hinges on them being actually competent, reinforcing the notion of the “hardware lottery”…

…It is no secret that my favourite codebase of all time is T5X and Mesh Tensorflow (named tensors ftw) but these options quickly became not viable as 1) they don’t get as much support outside Google, 2) they are kind of deprecated and 3) they are not friendly to folks on our team that are not xooglers.

We ended up going for something vanilla, seemingly stable and more popular (i.e., pytorch) that is more accessible to most people on the team (except me lol). In my first few months, I was tripping all over pip, git, docker and all these wild life stuff. Then again, I am not 100% sure about how stable or user friendly it would be to use a google codebase externally (it would have been pretty nasty I guess).

To be very frank, I would have to say the quality of codebases externally significantly lag behind those I’ve been used to at Google. Primarily because codebase within Google tends to be written by ML rockstars themselves (e.g, Noam Shazeer, Barret Zoph, Adam Roberts, Hyung Won Chung et al.) and just feel better (e.g., superior vibes) compared to those I’ve tried externally. In particular, I found myself super annoyed with the code quality when dabbling with stuff built by other companies (some way worse than others 🤗).

5. How The Interstate Highway System Changed American Industry – Lawrence Hamtil

Signed into law in 1956 by then President Dwight Eisenhower, the Federal Highway Act created the Interstate Highway System, which would become the largest and costliest public works project in history.  Measuring almost 48,000 miles in total distance, the Interstate Highway System was completed only in 1992, more than three decades after work began, and for a total cost in today’s dollars of more than $500 billion…

…Among the beneficiaries of this huge outlay were the quarry owners and aggregate miners, who provided the gravel and rock on which the interstates were laid, the heavy machinery manufacturers who provided the graders, tractors, and steamrollers that turned those rocks into roads, and the oil and gas producers and refiners who made the gasoline and diesel that fueled the project…

…As families began to set out exploring the country on the new interstate system, restauranteurs such as Ray Kroc and Howard Johnson recognized the need to provide traveling families with predictable, familiar service.  The idea of the chain restaurant was born as interstate exit ramps guided hungry motorists to McDonald’s and Howard Johnson’s.  Families would also need places to say on longer journeys, so hotels followed restaurants in the chain model as franchises like Holiday Inn became a staple of interstate exits; early ads for the hotel underlined the value of the familiar by stating, “The best surprise is no surprise.”

The logistical flexibility provided by the interstate system also gave rise to a whole new model of retailing:  big box stores began to set up in small towns offering rich variety and low prices to consumers previously left unserved by larger retailers.  Walmart’s 1975 annual report detailed just such a model…

…Whereas not quite a century before the railroads had aided in the rise of Sears, Roebuck, and Co. as the first retailer with national reach, the interstate in the 1960s and 1970s would provide the backbone of Walmart’s logistical operations, with large distribution centers situated at critical points throughout the interstate network to facilitate inventory replenishment, as Professor Jesse LeCavalier has noted on his blog. 


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google), Amazon, Apple, and Microsoft. Holdings are subject to change at any time.

What We’re Reading (Week Ending 03 March 2024)

The best articles we’ve read in recent times on a wide range of topics, including investing, business, and the world in general.

We’ve constantly been sharing a list of our recent reads in our weekly emails for The Good Investors.

Do subscribe for our weekly updates through the orange box in the blog (it’s on the side if you’re using a computer, and all the way at the bottom if you’re using mobile) – it’s free!

But since our readership-audience for The Good Investors is wider than our subscriber base, we think sharing the reading list regularly on the blog itself can benefit even more people. The articles we share touch on a wide range of topics, including investing, business, and the world in general. 

Here are the articles for the week ending 03 March 2024:

1. The Future of Ecommerce That Wasn’t: An In-depth Look into What Went Wrong with Wish – Speedwell Research

A good explanation is a story that is hard to vary. If we did a postmortem of WebVan (grocery delivery) or Pets.com (online specialty store for Pets), what would we say went wrong? If we did the postmortem in 2006, most likely we would have said it was a silly and unrealistic idea. But if we were to do a postmortem now, with the existence of Instacart (grocery delivery) and Chewy (online specialty store for pets), how would our understanding change?

This is not a trivial exercise. It is far too easy to be dismissive about a failing business and think it was the entrepreneur’s ill-thought-out idea or just incompetence, but this does not hold scrutiny.

Look at Apple. For how many years was Steve Jobs and his insistence on not licensing the Mac operating system seen as the impetus for their failure? And the same thing happened again when the iPhone was released: analysts thought their unwillingness to license the phone’s iOS would ultimately lead to their demise. Now though, Apple’s success is attributed to their close integration and their proprietary software is a key selling point, which wouldn’t be possible if they licensed it.

If you took over Lego in 2004 when it was nearing bankruptcy, what would you diagnose as the problem? Would you have thought that with digital entertainment kids just don’t want to play with toy blocks anymore? Or would you have thought the focus on “noncore” activities like theme parks, clothing, and video game development were the issues? Perhaps the product was good but was simply too expensive? You know that today there are vastly more digital entertainment options than there were in 2004, they still have theme parks and video games, and their products are still expensive, so what was it?

If you were appointed CEO of Crocs in 2008 when their stock dropped 98% and was on the verge of entering bankruptcy, tell us that you wouldn’t be tempted to lay the blame on the aesthetics of the shoes. It is the most ridiculed shoe design with “ugly” virtually synonymous with Crocs and yet they now sell over 150 million of them a year. Again, what some people would identify as the problem of the business turned out to be a virtue…

…So, if we are saying Wish was unsuccessful because of their focus on cheap items with slow shipping, we shouldn’t be able to point to another company that did something similar and was successful…

…We will do one final analysis to estimate churn before concluding. However, we want to note that this analysis is unnecessary to make the point we are about to. If you simply saw that they lost users despite spending >80% of revenues on marketing, or almost ~$1.5bn, is there any explanation you would accept that could convince you the business was healthy? Imagine your Chief Marketing Officer just told you they spent $1.5bn to lose 2mn buyers and grow revenues 2%. How would anyone possibly see that as a good thing? And yet, with a little bump in numbers from Covid in 2020, it was overlooked by investors in favor of the hope of buying the next Amazon at IPO.

In the S-1 they disclose the following LTV chart. They calculate “LTV” as cumulative gross profits over a period of time attributable to new buyers acquired in a given cohort, divided by total new buyers, which is most certainly not what an LTV really is.

For example, let’s say a cohort generated $15 of gross profit in year one and then another $10 of gross profit in year two. They would add those two numbers up and say the “LTV” of the customer in year 2 is $25.  Therefore, if you wanted to calculate how much each cohort generated in gross profits in a given year, you just have to take the difference between each year. In this example, this cohort generated $10 in gross profits in the second year versus $15 in the first year, suggesting ample churn. What you would want to see is each year’s incremental figure stay steady or increase.

The chart above shows cumulative gross profit by cohort. If it was a perfectly linear line, then that would mean in each period the cohort bought the same amount of goods as the previous period.

We will focus just on the 2017 cohort for simplicity. We annotated it to show how we estimated incremental gross profits. The average buyer from the 2017 cohort earns $15 in gross profits in year 1, which drops to $10 in year 2, and then to $6 in year 3. We can already see that by year 3, each cohort is generating about 1/3rd what it did in year 1, which suggests heavy churn. Remember that Wish’s payback period is about 2 years, which means it isn’t until year 3 they make that small incremental gross profit. And remember, this is just to pay back the initial marketing investment, not other S&M they spent on promotions to reengage that buyer.

Here, we can see that that the difference between the gross profit for total buyers divided by the gross profit per average active buyer gives us a churn estimate. At the end of year 1, 100% of buyers are active (by definition) and by year three that drops to 19%. That comes out to about 44% annual churn over two years. It is also noteworthy that the churn is much worse in the first year. A full 67% of buyers do not return after buying once.

Now, remember that their average payback period is under 2 years. That is rather problematic in the context of almost no one being left after 2 years! They have a thin amount of remaining users that not only need to cover all of the reengagement marketing, but also all of their G&A and R&D cost. And that’s before they can even make a profit!

This is a fundamentally broken business. Users do not stay long enough, they have to pay to get users to return, and users are not profitable…

…Earlier we said that an explanation is a story that cannot easily vary. Well, we have trouble figuring out exactly what the story is that cannot vary. There is nothing in principle wrong with an ecommerce offering catered to the consumers in the low-end of household earnings. Some would note that the low average order value would make it hard to make enough contribution profit per order, but that is essentially what Pinduoduo did in China, what Shopee is doing in Southeast Asia and Brazil, and what Temu is doing in the US. While we don’t know exactly if all of those initiatives will end up being profitable, it is hard to claim it is the idea itself that is rotten.

Clearly, Wish had a problem with both their high cost to acquire users and their ability to retain them. We know that Pinduoduo had a better customer acquisition engine piggy backing off of Tencent’s Weixin with preferred placement and the Community Group Buy model was a novel way to spur consumer sharing, free of charge. Shein had TikTok and went viral early on with “Shein hauls”, where influencers would post everything they purchased. They would later lean into influencer marketing on TikTok to much success. Amazon has Amazon Prime which helps retain users, and their optimal customer service helps keep customers satisfied at potential churn events.  Wish was lacking something in the customer acquisition and retention area, but exactly what isn’t obvious.

Perhaps it was a mix of everything that individually created customer churn events from slow shipping to “unreliable shipping”, fraud, fake listings, sub-par customer service, inadequate item selection, poor item quality, inaccurate recommendations, or perhaps even internal issues. But again, other companies have survived similar or worse issues. And the longer the list, the more it speaks to our lack of confidence in any one variable. As an investor from the outside, it isn’t apparent what the key problem was, at least not to us.

What is crystal clear though is that there were issues since at least 2019, and some red flags prior. An investor only needed the company’s IPO prospectus to see these problems brewing, and could have avoided even worrying about any potential “narrative fallacy” by just focusing on the financials.

2. Bill Ackman: Investing, Financial Battles, Harvard, DEI, X & Free Speech | Lex Fridman Podcast #413 (partial transcript here) – Lex Fridman and Bill Ackman

Bill Ackman (57:12): So this was at the time of the Financial Crisis, circa November 2008. Real estate’s always been a kind of sector that I’ve been interested in. I began my career in the real estate business working for my dad, actually arranging mortgages for real estate developers. So I have kind of deep deep ties and interest in the business. General Growth was the second largest shopping mall company in the country – Simon Properties many people have heard of – General Growth was number two. They own some of the best malls in the country…

…General Growth the company, the CFO in particular,  was very aggressive in the way that he borrowed money. He borrowed money from a kind of Wall Street – not long-term mortgages – but generally relatively short-term mortgages. He was pretty aggressive. As the value went up, he would borrow more and more against the assets and that helped the short-term results of the business. The problem was during the Financial Crisis, the market for what’s called CMBS – commercial mortgage backed securities – basically shut. And the company, because its debt was relatively short-term, had a lot of big maturities coming up that they had no ability to refinance. The market said, “oh my God, the lenders are going to foreclose and the shareholders are going to get wiped. The company’s going to go bankrupt, they’re going to get wiped out.” The stock went from $63 a share to 34 cents. There was a family, the Bucksbaum family owned I think about 25% of the company and they had a $5 billion stock that was worth $25 million or something by the time we bought a stake in the business.

What interested me was, I thought the assets were worth substantially more than the liabilities. The company had $27 billion of debt and had $100 million value of the equity, down from like $20 billion. And sort of an interesting place to start with a stock down 99%. But the fundamental drivers – the mall business – are occupancy, how occupied are the malls? Occupancy was up year on-year between ‘07 and ‘08. Interestingly, net operating income, which is kind of a measure of cash flow from the malls – that was up year-on-year. So the underlying fundamentals were doing fine. The only problem they had is they had billions of dollars of debt that they had to repay – they couldn’t repay

If you examine the bankruptcy code, it’s precisely designed for a situation like this where it’s this resting place you can go to restructure your business. Now the problem was that every other company that had gone bankrupt, the shareholders got wiped out. And so the market, seeing every previous example the shareholders get wiped out, the assumption is the stock is going to go to zero. That’s not what the bankruptcy code says. What the bankruptcy code says is that the value gets apportioned based on value, and if you could prove to a judge that the assets’ worth more than the liabilities, then the shareholders actually get to keep their investment in the company. And that was the bet we made.

So we stepped into the market. We bought 25% of the company in the open market. We had to pay up. It started at 34 cents – I think there were 300 million shares – so it was at a $100 million value. By the time we were done, we paid an average of – we paid $60 million for 25% of the business, so about $240 million for the equity of the company. And then we had to get on the board to convince the directors the right thing to do. The board was in complete panic, didn’t know what to do, spending a ton of money on advisers…

…And the key moment, if you’re looking for fun moments, is there’s a woman named Maddie Bucksbaum who was from the Bucksbaum family. Her cousin John was chairman of the board, CEO of the company. And I said – as she calls me after we disclose our stake in the company, she’s like “Billy Ackman, I’m really glad to see you here.” I met her – I don’t think it was a date – but I kind of met her in a social context when I was 25 or something. And she said, “I’m really glad to see you here and is there anything I can do to help you, call me.” I said, “Sure.” We kept trying to get on the board of the company, they wouldn’t invite us on.  Couldn’t really run a proxy contest, not with a company going bankrupt, and their advisers actually were Goldman Sachs and they’re like, “You don’t want the fox in the hen house” and they were listening to their advisors. So I called Maddie up and I said, “Maddie, I need to get on the board of the company to help.” And she says, “I will call my cousin and I’ll get it done.” She calls back a few hours later, “You’ll be going on to the board.” I don’t know what she said, but she was convincing.

Next thing you know, I’m invited to the board of the company and the board is talking about the old equity of General Growth. Old equity is what you talk about when the shareholders are getting wiped out. I said, “No, no, no. This board represents the current equity of the company. I’m a major shareholder, John’s a major shareholder, there’s plenty of asset value here. This company should be able to be restructured for the benefit of shareholders.” And we led a restructuring for the benefit of shareholders and it took let’s say eight months and the company emerged from Chapter 11. We made an incremental investment into the company and the shareholders kept the vast majority of their investment. All the creditors got their face amount of their investment – par plus accrued interest. And it was a great outcome. All the employees kept their jobs, the malls stayed open, there was no liquidation. The bankruptcy system worked the way it should. I was in court all the time and the first meeting with the judge, the judge is like “Look, this would never have happened were it not for a financial crisis.” And once the judge said that, I knew we were going to be fine because the company had really not done anything fundamentally wrong – maybe a little too aggressive in how they borrowed money.

Stock went from 34 cents to $31 a share…

…Lex Fridman (1:05:44): How hard is it to learn some of the legal aspects of this? You mentioned bankruptcy code – I imagine it’s very sort of dense language and dense ideas and the loopholes and all that kind of stuff. If you’re just stepping in and you’ve never done distressed investing, how hard is it to figure out?

Bill Ackman (1:06:05): It’s not that hard. I literally read a book on distressed investing. Ben Branch or something, on distressed investing.

Lex Fridman (1:06:12): So you were able to pick up the intuition from that, just all the basic skills involved, the basic facts to know, all that kind of stuff.

Bill Ackman (1:06:20): Most of the world’s knowledge has already been written somewhere. You just got to read the right books.

3. Why is Google killing cookies? – Eric Benjamin Seufert

What is Google’s underlying motivation in deprecating third-party cookies in Chrome? Suspicion is warranted. Google’s mission statement for its Privacy Sandbox initiative is to “Protect [user] privacy online,” across its Chrome browser and its Android operating system (Google intends to deprecate its GAID Android identifier at some point). Cookies, unquestionably, present severe data leakage risks to consumers: they allow anonymous services to observe the web activities of users with little preventative recourse. But as I point out in this piece, “privacy” is an abstract social concept, and firms – but especially multi-trillion dollar market leaders – don’t make dramatic, sweeping policy changes absent commercial benefit. Believing that a company would utterly reform the mechanics of digital advertising solely in service of increased user privacy is as absurd as believing that two firms would engage in a merger as an expression of friendship. To not assume a commercial motive in cookie deprecation is naive.

Apple’s App Tracking Transparency (ATT) privacy policy is an apt example of this. Apple launched an international PR campaign championing the privacy safeguards of the iPhone following its introduction of ATT in April 2021. Yet as I point out in this piece, Apple collects and utilizes consumer data in the ways that ATT was ostensibly designed to prevent. Apple positions its use of install and purchase data collected via consumer engagement in apps that it doesn’t own as “ads personalization” and not “tracking.” Apple claims first-party privileges over this consumer data because Apple exerts (and is stridently maintaining a firm grip on) control over iOS payments, giving it exclusive, proprietary access to that data. And in a court filing from December 2023, Apple had the following to say about the logical contortions of its privacy policies (all emphasis from the document):

The Allow Apps to Request to Track setting governs whether apps can ask to track users across apps or websites owned by other companies, as Apple’s descriptions of the setting consistently make clear … Plaintiffs also include a screen shot of the Tracking disclosure, which explains that Apple “requires app developers to ask for permission before they track your activity across Apps or websites they don’t own.” … Given Apple’s extensive privacy disclosures, no reasonable user would expect that their actions in Apple’s apps would be private from Apple.

This isn’t to say that Google and Apple don’t employ well-meaning, intelligent, and highly effective people whose efforts are centered on promoting their conceptions of digital privacy. But digital privacy initiatives from publicly traded, multi-trillion-dollar corporations must be viewed in a broader commercial context…

…So given that Google must have a commercial motivation in deprecating cookies, what is it? The most obvious is simply margin expansion: Google’s network business, which serves ads on third-party websites and apps, will almost certainly suffer if the Privacy Sandbox is less effective for targeting and measurement than cookies (and early indicators suggest it is). If the economics of buying third-party open web inventory through Google’s tools degrades, some of that demand may simply be routed to Google’s owned-and-operated channels. And these channels feature much higher margin for Google than its Network business: Bernstein estimated in December 2022 that Google’s margin on Network revenue is 10%, while it’s 15% for YouTube and 55% for Search. As I argue in this piece, because advertising budgets are deployed against absolute performance, Google will likely lose some degree of top-line revenue if its Network business unit declines. But Google doesn’t need to shift all of the revenue from Network to these channels to maintain its current bottom line given the margin differentials: $1BN in Network revenue produces the same margin as $181MM in Search revenue. 

4. Twitter thread on life and investing lessons from climbing Mt Kinabalu – Eugene Ng

1 | Hiking is a marathon, not a sprint. It is about finishing, whether or not you finish first or last it doesn’t matter, as long as you finish. There is no gold medals for the fastest, and only rescue for those who don’t. It only matters as long as you finish, and that you remain safe when you finish. Safety first, go slow.

It is the same with investing. Never be permanently wiped out, avoid all unlimited downside trades, and then you can focus on making asymmetric bets with unlimited upside…

3 | I was in awe out the scale of the human labour require for the entire operations. We saw numerous porters carrying up 20-40+ kg of fresh water, food, furniture, equipment for the lodge where we stayed at. There were also a number of porters who carried up luggage for some climbers as well. Without them, the support, the ecosystem, none of these would have been possible for us to experience the climb.

It is ever easier than ever to get data, but we cannot be lazy. We need to  learn to appreciate the ecosystem and what we have now with the internet, versus 50 years old with libraries and faxes. Use them to your advantage. Easily available does not mean everyone will actually read them. Do not confuse it.

4 | We had a fantastic mountain guide who was one of the oldest and fittest at 52 years and he has  been doing it over 32 years old since he was 20 years old. He still does this three times a week and will retire next year. He was leading us in our walk through easier path with such a controlled and comfortable pace, like he is meditating. Without him, it would have been so much more difficult. Having the right leader to help guide you really matters.

Having the right mentors, the right people around you matters. They can have the right expertise, experience to share that can help you in your journey to become better and to avoid the pitfalls…

6 | When ascending up and descending back down, it is not an individual effort but a collective team effort. The company matters. Without the right company to support you mentally on every step of the way makes so much of a difference, everyone has a role to play.

Like in investing, there are going to be ups and downs. The right people/investors to stand by you matters, and not run away at the first sight of trouble. Choosing carefully the right team to the best of your ability matters.

7 | Sometimes a member of your team is not going to be feeling very well or can be injured, it is being prepared and bringing along extra supplies, medical or food, and continuously supporting them with what you have physically and mentally. Remember that if they can’t finish, you can’t finish.

Know that the businesses that we invest in are not going to do well all the time, it is not going to be a straight line. There are going to be ups and downs, and they will zig and zag from time to time. We need to have the patience to stand by them through difficult times, and the good times, and not sell them.

8 | Run your own race at your own pace, sometimes you will overtake and sometimes others will overtake. Don’t be stressed by someone behind you trying to push you go faster. You set your own pace. If they want to overtake, just stand to the side and let them overtake, if not just chill. Separately, if you want to overtake someone slower, then just overtake on the side.

Find your own investment strategy that suits you best, that energises you.  The real race is against yourself, not against others. There will always be someone who will do better than you in any given year, so chill. It is not about being the top 10% in a year, but the top 10% after 10 or 20 years and more…

11 | Always remember to never get complacent and choose speed, or get distracted. Do a misstep over a loose rock, and you may just end up spraining your ankle (like me), and end up not finishing the climb. But thankfully, it was serious and painful, but it was still okay enough for me to complete the last 8km. It was insanely painfully with every descend as my right ankle landed on every step.

Never think highly of yourself. Stay humble, have humility. The moment you lose that, you stop listening, you stop absorbing, you stop learning, and then with a mis-step you might just result in eventual failure. Never do that.

12 | At the end, despite how much preparation, it is really willpower at the end that gets everyone through to the end. It can be so powerful, the human mind and the will power. Despite how tough it is, we were just highly focused on taking step at a time mindfully and carefully, that’s all that mattered. I sprained my right ankle horribly with 8km left, it was really painful, but I kept persisting, and my teammates were patient with me and walked slower. “Stay hard” by David Goggins was our slogan to keep us going.

Investing too is a slog, managers get paid to endure all the emotional and psychological elements with all the ups and downs. It is knowing when to keep pursuing and staying the course even especially when the going gets tough…

14 | Memories over medals. We did not finish first, but we finished in the end, and that’s what matters. To Team Endurance!

If you beat the index after 10 or 20 years, you will be in the top quartile. You want to keep staying and playing the game, and keeping doing okay and eventually you will do very well.

5. Things I Don’t Know About AI – Elad Gil

In most markets, the more time passes the clearer things become. In generative AI (“AI”), it has been the opposite. The more time passes, the less I think I actually understand.

For each level of the AI stack, I have open questions…

…There are in some sense two types of LLMs – frontier models – at the cutting edge of performance (think GPT-4 vs other models until recently), and everything else. In 2021 I wrote that I thought the frontier models market would collapse over time into an oligopoly market due to the scale of capital needed. In parallel, non-frontier models would more commodity / pricing driven and have a stronger opensource presence (note this was pre-Llama and pre-Mistral launches).

Things seem to be evolving towards the above:

Frontier LLMs are likely to be an oligopoly market. Current contenders include closed source models like OpenAI, Google, Anthropic, and perhaps Grok/X.ai, and Llama (Meta) and Mistral on the open source side. This list may of course change in the coming year or two. Frontier models keep getting more and more expensive to train, while commodity models drop in price each year as performance goes up (for example, it is probably ~5X cheaper to train GPT-3.5 equivalent now than 2 years ago)

As model scale has gotten larger, funding increasingly has been primarily coming from the cloud providers / big tech. For example, Microsoft invested $10B+ in OpenAI, while Anthropic raised $7B between Amazon and Google. NVIDIA is also a big investor in foundation model companies of many types. The venture funding for these companies in contrast is a tiny drop in the ocean in comparison. As frontier model training booms in cost, the emerging funders are largely concentrated amongst big tech companies (typically with strong incentives to fund the area for their own revenue – ie cloud providers or NVIDIA), or nation states wanting to back local champions (see eg UAE and Falcon). This is impacting the market and driving selection of potential winners early.

It is important to note that the scale of investments being made by these cloud providers is dwarfed by actual cloud revenue. For example, Azure from Microsoft generates $25B in revenue a quarter. The ~$10B OpenAI investment by Microsoft is roughly 6 weeks of Azure revenue. AI is having a big impact on Azure revenue revently. Indeed Azure grew 6 percentage points in Q2 2024 from AI – which would put it at an annualized increase of $5-6B (or 50% of its investment in OpenAI! Per year!). Obviously revenue is not net income but this is striking nonetheless, and suggests the big clouds have an economic reason to fund more large scale models over time.

In parallel, Meta has done outstanding work with Llama models and recently announced $20B compute budget, in part to fund massive model training. I posited 18 months ago that an open source sponsor for AI models should emerge, but assumed it would be Amazon or NVIDIA with a lower chance of it being Meta. (Zuckerberg & Yann Lecunn have been visionary here)…

...Are cloud providers king-making a handful of players at the frontier and locking in the oligopoly market via the sheer scale of compute/capital they provide? When do cloud providers stop funding new LLM foundation companies versus continuing to fund existing? Cloud providers are easily the biggest funders of foundation models, not venture capitalists. Given they are constrained in M&A due to FTC actions, and the revenue that comes from cloud usage, it is rational for them to do so. This may lead / has led to some distortion of market dynamics. How does this impact the long term economics and market structure for LLMs? Does this mean we will see the end of new frontier LLM companies soon due to a lack of enough capital and talent for new entrants? Or do they keep funding large models hoping some will convert on their clouds to revenue?…

What happens in China? One could anticipate Chinese LLMs to be backed by Tencent, Alibaba, Xiaomi, ByteDance and others investing in big ways into local LLMs companies. China’s government has long used regulatory and literal firewalls to prevent competition from non-Chinese companies and to build local, government supported and censored champions. One interesting thing to note is the trend of Chinese OSS models. Qwen from Alibaba for example has moved higher on the broader LMSYS leaderboards…

How much of AI cloud adoption is due to constrained GPU / GPU arb? In the absence of GPU on the main cloud providers companies are scrambling to find sufficient GPU for their needs, accelerating adoption of new startups with their own GPU clouds. One potential strategy NVIDIA could be doing is preferentially allocating GPU to these new providers to decrease bargaining power of hyperscalers and to fragment the market, as well as to accelerate the industry via startups. When does the GPU bottleneck end and how does that impact new AI cloud providers? It seems like an end to GPU shortages on the main clouds would be negative for companies whose only business is GPU cloud, while those with more tools and services should have an easier transition if this were to happen…

…ChatGPT launched ~15 months ago. If it takes 9-12 months to decide to quit your job, a few months to do it, and a few months to brainstorm an initial idea with a cofounder, we should start to see a wave of app builders showing up now / shortly.


Disclaimer: The Good Investors is the personal investing blog of two simple guys who are passionate about educating Singaporeans about stock market investing. By using this Site, you specifically agree that none of the information provided constitutes financial, investment, or other professional advice. It is only intended to provide education. Speak with a professional before making important decisions about your money, your professional life, or even your personal life. We currently have a vested interest in Alphabet (parent of Google), Amazon, Apple, Meta Platforms, Microsoft, and Tencent. Holdings are subject to change at any time.