In this issue:
- How Many Trillion-Dollar Companies Should There Be?—Better hiring decisions are clearly a massive opportunity, and there's a known tool for aggregating the kind of knowledge that's useful for exactly these decisions. So why is it so hard to pick up the trillion-dollar bill on the sidewalk? And what does this reveal about the kinds of companies that do produce that much value?
- Choice—Retailers have reduced the number of newly-introduced products from 5% in 2019 to 2% in 2023, thanks to a combination of one-time pandemic disruptions and long-term logistics trends.
- Durables—Stanley has built a $750m business off durable goods, but has dodged many of the problems that make this such a challenging category.
- Lobbying—AI lobbyists have to navigate DC's scope insensitivity problem, where minor contributions to endlessly-debatable issues get more attention than meaningful changes in the state of the world.
- The China Model—A bribery case illustrates China's accidental state capacity model.
- Training Data—One perk of the metaverse: it can be a proprietary source of training data for a category that, unlike text, images, and video, does not have a large and well-labeled corpus.
How Many Trillion-Dollar Companies Should There Be?
A few months ago I was at a conference where Robin Hanson spoke about prediction markets.[1] He argued that given how much of companies' outcomes are driven by who they choose to hire, and how non-rigorous the process of selecting employees and revisiting those selections is, there's a literal trillion-dollar opportunity in getting it right. Prediction markets are Hanson's preferred method here—employees could anonymously bet on whether a given new hire was likely to get promoted or fired in the next two years, which, through the pricing of these bets, would inform management about collective perceptions, while simultaneously enabling the more perceptive employees to have a greater impact on the market, thanks to the actual profit and loss of those bets.
We'll revisit this plan in a moment, but for now we can focus on the high level assessment, i.e. whether or not this could theoretically be a trillion-dollar opportunity. The addressable market is sizable: total wages and salaries in the US were $10.5tr as of 2022, and let's suppose we could make the labor market 5% more efficient with better data. That's $525bn in value creation, and it's not implausible for the company creating the value to capture 20% of it, or about $105bn. Software companies at that scale tend to have net profit margins in the 20-35% range, so we don't need a heroic earnings multiple to slap a $1 trillion valuation on this hypothetical company.
And yet, this hypothetical trillion-dollar company doesn’t exist; the inefficiency persists. Instead we live in a world where an employee’s single most important contribution could be making one specific hire—a contribution likely to go unrecognized (e.g. the people who interviewed Satya Nadella for his first Microsoft job in the early 90s probably didn’t receive a retroactive seven-figure finder's fee).
Meanwhile, basically every major company agrees that their most valuable resource is their employees. That's true in the fuzzy sense that companies are made out of people, and it's even true in the ploddingly literal sense that corporate America's biggest expense is labor, and any given company's non-labor expenses (materials, software, etc.) is some other company's revenue which is being used to pay its people. And there's room to improve employee-to-employer matching. The natural frictions in hiring and firing indicate that we’re far from the theoretical optimum, and the phenomenon of late bloomers also shows that there’s significant upside to matching the right person with exactly the right tasks.
But the matching business itself is just not that big. LinkedIn produces ~$14bn in revenue and is plausibly worth ~$100bn, Japan's Recruit is worth $66bn, recruiting firm Robert Half is worth $9bn, Workday is valued at $70bn. The multi-trillion dollar problem of who to hire and how to judge them has piecemeal solutions at various points, but hasn't produced a single dominant winner just yet.
If we take another angle, we can look at what the world's trillion-dollar companies actually do:
- Smartphones, computers, and software, with the last as a disproportionate contributor to both margins and growth.
- Business software and cloud computing
- Extract and refine oil
- Search and cloud computing
- Online retail and cloud computing
- GPUs and a software stack that makes it hard to exit their ecosystem
Aramco, the oil company, is arguably an outlier that shouldn't really be counted.[2] What the other companies have in common is that they have some kind of continuously improving returns to scale: Apple can do tighter hardware/software interaction than any other major company because they're operating at scale in both businesses; Microsoft has incredible distribution into big companies, meaning that they can rapidly expand in new areas like AI; Google can run more search ranking experiments than competitors and has more comprehensive data on what people click on when they aren’t searching ($, Diff); Amazon's loop is endlessly complex but tends to mean that they can offer faster shipping on more products at lower prices than anyone with a weaker infrastructure footprint, not to mention their cloud business's ability to continuously slice up different mixes of computing resources optimally; and while Nvidia is in a cyclical industry ($, The Diff), it's also been able to extend its lead as a standards-setter: if the biggest customers build their AI product roadmap partly around Nvidia's hardware roadmap, it's hard for anyone to interrupt the process.
So what these companies have in common is a combination of 1) relatively higher fixed than marginal costs, such that each incremental dollar of revenue from a given business line is more profitable than the last, and 2) some kind of scaling effect where each dollar of revenue makes the next dollar easier to earn in at least some respect.
Does hiring have this? Some parts of it clearly do: LinkedIn is a classic network effects business, and the company realized early that a big growth driver was ranking well for people's names.[3] A comprehensive directory of professional backgrounds ends up being useful for some kinds of metadata analysis that will be harder for a smaller-scale site—looking for which startup hired the most alumni from some larger company, for example, is straightforward when two thirds of the relevant professionals have an up-to-date profile, but meaningless when the share is, say, 5%.
But it's harder for LinkedIn to make inroads into the predict-employee-future-performance business, since 1) in many companies, that doesn't exist as a discrete function, with firing decisions made on a more ad hoc basis, and 2) to the extent that it does, LinkedIn ends up competing with its own customers, the HR departments. Any software that enhances the work of some professional has to embody an assumption about which parts of the job are commoditized and which parts add value. Anything that speeds up the commoditized portion is probably a good feature to add, at least some day, but if it touches on the more qualitative part of the user's work, they'll worry that adopting it means losing their job.
Another reason this particular category is hard is that transparency's value is so situational. At one level, everyone wants full transparency in the hiring process: jobseekers want to know if they're being interviewed out of genuine interest, as a box-checking exercise ("We know who we're going to hire, but we need to interview at least three people"), as an attempt to keep relationships warm, or as an exercise in competitive intelligence. Employers want to know if they're talking to someone who is actively looking for a job, desperately looking for one, trying to get a backup offer so they have more leverage getting the job they really want, trying to see if all that Leetcode practice is actually paying off, or gathering competitive intelligence themselves. And, critically, each instance of information asymmetry benefits the person with the extra information!
This makes it hard to have some kind of meaningful data loop in hiring: the more information you'd theoretically get, the more incentive the other side has to obscure it. The information where you can get a good data loop includes:
- Internal, company-specific metrics that are not going to be shared elsewhere. This is not quite as good as a loop where the data can be used externally, but it does have some value, since it makes it harder for customers to leave, and makes that a function of both their size and their tenure. It's not the network effect of which trillion-dollar valuations are made, but it does nice things to an LTV calculation if the revenue is going up at the same time that the churn rate is going down.
- Aggregation of public-facing information—undergraduate degrees with the GPA sanded off, "best places to work" awards that omit any mention of how much lobbying and system-gaming were required to achieve them, etc. This, too, is valuable, but necessarily limited, because for strategic reasons neither side shares the most actionable information.[4]
If a product is going to affect the hiring process, it will only get an opt-in if both sides feel that they're getting some advantage. For internal work software, the opt-in is a simple take-it-or-leave-it—if your employer starts using some kind of fancy ML-powered employee evaluation tool, you might be able to stop them by complaining, but you’ll probably have to just put up with it or quit. There are cases where both sides opt in to the same system, but in that particular instance, medical residency matching, the company that manages the process is implementing a version of a published algorithm, which competitors could also implement. This doesn't make it impossible for them to turn a profit (stay tuned for an upcoming Diff piece on open source economics), but it does place some limits.
The tricky thing about solving hiring with prediction-market-based businesses is that it's hard to find a model where a) everyone opts in to a value-creating system, and b) whoever creates that system actually captures a big share of that value. There's a near possible world where more companies are running lightweight internal prediction markets, but where these markets can't offer anything truly proprietary without creating the perception that they might be unfair—at which point employees will be reluctant to participate.
In the end, this process should give us more appreciation for how unique the handful of trillion-dollar companies out there really is. There are many more technologies that have created that much value—not just the big general-purpose ones, like electricity, but specific applications like air conditioning—than there are technologies that can create a trillion dollars of market value for one specific company. But for the few companies that do capture the enormous wealth they create, the kernel of the business is often a single insight (like "there should be a nice-looking computer available for home use" or "college kids love socializing and they all have computers now"), the actual wealth creation comes from countless instances of favorable compounding, particularly from creating incentives for users to slowly leak information that can be used to make the product much better for them and slightly better for everyone else. If there's one generalizable difference between trillion-dollar ideas and ideas that can create a trillion-dollar company, it's this: when the thing you figured out first eventually becomes obvious to everyone else, have you set up a system that ensures that you're still accumulating new and valuable secrets faster than anyone else?
Disclosure: Long Microsoft, Meta, Amazon.
For those of you who are familiar with Hanson's work, this does not narrow it down much. The talk was this one at Manifest. ↩︎
The oil industry's structure, including the size and profitability of companies, is partly a real-world economics question, but given the ludicrously profitable nature of extracting an energy-dense material that can be subdivided almost infinitely and cheaply transported worldwide, and the fact that the business requires fixed assets in specific locations controlled by nation-states, the split between oil companies and host countries is a matter of a) negotiation in advance, and b) countries' and companies' willingness to honor the terms of their agreement after they see how much money is at stake. Aramco is an example of this: foreign oil companies signed lopsided agreements with Saudi Arabia in the early days of the global oil industry, and those agreements got gradually revised up to the point that Saudi Arabia fully nationalized the oil assets. And then, a few decades later, started selling them again. ↩︎
This was a sort of targeted outreach. The people who are googling the names of professionals and clicking on the result that looks like a résumé are disproportionately likely to be either looking for a job or looking for an employee. So they're the exact demographic that should be signing up for LinkedIn in the first place. ↩︎
You can think of the big online ad companies as being in the business of finding and monetizing secrets. Ads work best when they match people's desires, but people can't even be honest with themselves about those desires. Clicks will, in the aggregate, reveal truths that even extensive introspection keeps buried. ↩︎
Diff Jobs
Companies in the Diff network are actively looking for talent. See a sampling of current open roles below:
- A systematic hedge fund is looking for researchers and portfolio managers who have experience using alternative data (NYC).
- A company building ML-powered tools to accelerate developer productivity is looking for software engineers. (Washington DC area)
- A company building the new pension of the 21st century and building universal basic capital is looking for a frontend engineer. (NYC)
- A private credit fund denominated in Bitcoin needs a credit analyst that can negotiate derivatives pricing. Experience with low-risk crypto lending preferred (i.e. to large miners, prop-trading firms in safe jurisdictions). (Remote)
- A startup building a new financial market within a multi-trillion dollar asset class is looking for generalists with banking and legal experience. (US, Remote)
Even if you don't see an exact match for your skills and interests right now, we're happy to talk early so we can let you know if a good opportunity comes up.
If you’re at a company that's looking for talent, we should talk! Diff Jobs works with companies across fintech, hard tech, consumer software, enterprise software, and other areas—any company where finding unusually effective people is a top priority.
Elsewhere
Choice
Newly-introduced products were 5% of retailers' selections in 2019 and were 2% in 2023 ($, WSJ), as many retailers responded to supply chain issues during the pandemic by cutting less popular product lines, and found that it didn't pay to reintroduce them. There are several forces at work:
- Expectations for fast shipping and in-stock products put a premium on a smaller set of faster-selling items. This is not just because there are fewer things to keep in stock, but because the turnover for low-demand products is both lower and less predictable. Meanwhile, for people willing to tolerate longer shipping times, there are plenty of goods that are available but not in stock by way of discount retailers with long shipping times.
- More expensive labor also has a disproportionate effect on lower-turnover products.
- The pandemic was a one-time data dividend for retailers who moved more of their sales to apps ($, The Diff), which may have collapsed the uncertainty around whether or not some products were worth keeping. Grocery stores know that a gallon of milk or a dozen generic eggs will move off the store shelves at a predictable pace, but might have collected enough information to know that the third variety of vegan burger patty wasn't adding much sales compared to the first two.
Durables
CNBC profiles the sudden rise of the Stanley Quencher water bottle. This is another instance of the complexities of selling durable goods in a mature market: if demand is growing, it's hard to tell when the market will be saturated. In Stanley's case, they have some collectors who buy dozens of Quenchers, and the company's president, recruited from Crocs, knows how to deliberately limit availability to both a) keep prices high and b) control working capital (no, it's not deficient demand planning—it's a limited-edition drop!). These businesses can still be viable, but this is a category where it's far harder to run a growing company than a stagnant one. Replacement demand is relatively straightforward to model, but new-deployment demand means figuring out what customers want, and how non-customers might become future customers. It's tough to get right, but rewarding when it works (and for what it's worth, Google Trends shows demand for the Quencher up year-over-year).
Lobbying
AI companies and nonprofits are lobbying heavily, but often have a hard time casting their agenda in terms of salient political issues instead of genuinely important existential risk: they're talking about AI's potential to cause unprecedented growth, mass unemployment, and possibly the end of the world, rather than trying to frame it around current issues. (Of course, there's a selection effect at work in politics: the ideal issue is one with relevance to everyone, no clear solution, and a close to but not quite 50/50 split in popular opinion. This makes it infinitely newsworthy and means that learning the issue has a long half-life.) It's a bit like the story about how, after the first use of the atomic bomb, a farmer wrote to Oak Ridge to ask if the new weapons could be used to clear tree stumps: no matter how powerful a given technology is, once it's a creature of popular opinion it gets framed in terms of popular issues, even if it potentially renders some of them obsolete.
The China Model
A consistent claim in evolution, science, and economics is that if some approach works, it may end up being repeatedly independently discovered from a variety of starting points. A fun example of this is that China has, indirectly, implemented the neoliberal dream of a Georgist taxation system that funds a competent and well-paid bureaucracy. It’s hard to do that directly in a Marxist-Leninist state, but they do achieve it indirectly by making real estate-related bribes the simplest way for underpaid civil servants to catch up economically. This does lead to the occasional arrest for multi-million-dollar bribery scandals, but that’s a small price to pay for a system that directly incentivizes political leaders to maximize GDP growth.
Training Data
Late last year, Meta’s CTO gave an interview with a focus on AI. He notes that when he was recruited to the company then known as Facebook, in 2006, he was recruited specifically for his AI experience—big tech has been interested in AI for a long time, and what’s new is that this interest is paying off and people outside the companies are noticing. He makes one notable point on the Metaverse: there’s already a large corpus for text, audio, video, and images, but there isn’t one for static 3D objects or for moving ones. This is another sense in which Meta’s metaverse ambitions are forward-looking; just as Meta wants to set rules for VR given how costly it was not to set the rules for mobile apps, they also want to maximize the number of generative AI categories where they have a growing data advantage. They have this in text and images, but competitors can at least partly catch up. If the main place where one category of media exists is within a Meta-controlled walled garden, there can be a generative AI category where they have a monopoly.