For nearly three decades, Silicon Valley has relied on a simple legal defense when harm shows up on its platforms: “We didn’t create the content.”
That argument, grounded in Section 230 of the Communications Decency Act, has been one of the most effective liability shields in modern corporate history. It allowed platforms to scale at global speed while avoiding responsibility for what users post, share, and amplify. Lawsuits were routinely dismissed before they reached discovery. Executives rarely had to testify. The system worked.
What changed last week is not the statute. It’s the reality around it.
In K.G.M. v. Meta Platforms, Inc., et al., a Los Angeles jury found Meta and YouTube negligent in the design and operation of their platforms, concluding that those failures were a substantial factor in causing harm to a young user. The jury awarded $3 million in damages and assigned 70% responsibility to Meta and 30% to YouTube.
This was not a case about a single post. It was not a case about whether a platform failed to remove content quickly enough. It was a case about how the system itself works.
And the jury found that system wanting.
