How the machines actually work
The term collaborative filtering dates to Goldberg, Nichols, Oki and Terry's 1992 Communications of the ACM paper describing Xerox PARC's "Tapestry." The Netflix Prize (2006–2009) made matrix factorization — decomposing the sparse user–item rating matrix into low-dimensional latent factors — the industry default. Modern systems are hybrids: collaborative signals, content embeddings, sequence models, and real-time contextual features operating together in milliseconds.
Each major platform's architecture reflects what it is paid to maximize. YouTube's published system (Covington, Adams and Sargin, RecSys 2016) optimizes expected watch time per impression, not click-through rate — a distinction with large downstream consequences. YouTube's Chief Product Officer confirmed at CES 2018 that more than 70% of watch time comes from recommendations. Netflix (Gomez-Uribe and Hunt, ACM TMIS 2015) openly states its optimization target is "improved member retention" proxied through hours streamed, and values the system at "more than $1 billion per year" in retained subscribers. TikTok's leaked internal "Algo 101" document exposed a simplified scoring formula weighting likes, comments, playtime, and completions — with "retention" and "time spent" as the company's ultimate north-star metrics. Meta's optimization target since January 2018 has been "Meaningful Social Interactions" — a weighted composite that the Facebook Files revealed weighted angry-emoji reactions 5× a Like.
These systems share ML infrastructure with but are architecturally distinct from the ad-tech pipeline. Content recommenders are single-tenant rankers running in 50–300 milliseconds. Real-time bidding is a multi-party auction clearing inside a 100-millisecond budget, governed by the IAB's OpenRTB specification, in which supply-side platforms broadcast impression opportunities with behavioral metadata to dozens or hundreds of demand-side platforms. Large platforms run both stacks simultaneously — recommender engagement generates the inventory that the ad auction monetizes.
The central critique, articulated by Jonathan Stray at Berkeley's Center for Human-Compatible AI and by Stuart Russell in Human Compatible (2019), is a Goodhart's Law problem: engagement is a proxy for value, and once a proxy becomes a target, a sufficiently capable optimizer will achieve it by modifying the user — making them angrier, more addicted, more predictable — rather than by genuinely serving them. Haugen's disclosures are empirical confirmation: Meta's own documents describe MSI rewarding outrage despite a publicly stated goal of "time well spent."