Opinion editor's note: Star Tribune Opinion publishes a mix of national and local commentaries online and in print each day. To contribute, click here.
•••
The lawsuit filed by the New York Times against OpenAI and Microsoft for copyright infringement pits one of the great establishment media institutions against the purveyor of a transformative new technology. Symbolically, the case promises a clash of the titans: labor-intensive human newsgathering against push-button information produced by artificial intelligence.
But legally, the case represents something different: a classic instance of the lag between established law and emerging technology.
Copyright law, a set of rules that date back to the printing press, was not designed to cover large language models like ChatGPT. It will have to be consciously evolved by the courts — or amended by Congress — to fit our current circumstances.
The key legal issue in the case will be the doctrine known as "fair use." Codified in the Copyright Act of 1976, fair use tells you when it's acceptable to use text copyrighted by someone else. The fair use test has four factors. Educational and nonprofit uses are more likely to be found to be fair use. Creative work gets more copyright protection than technical writing or news. The amount of the work that has been copied matters, as does the centrality to the copied work of the material that's been copied.
And perhaps most important for the Times' lawsuit, courts also consider whether the copying will harm the present or future market for the work copied.
Once you know the law, you can guess roughly how the legal arguments in the case are going to go. The Times will point to examples where a user asks a question of ChatGPT or Bing and it replies with something substantially like a New York Times article. The newspaper will observe that ChatGPT is part of a business and charges fees for access to its latest versions, and that Bing is a core part of Microsoft's business. The Times will emphasize the creative aspects of journalism. Above all, it will argue that if you can ask a large language model (LLM) powered search engine for the day's news, and get content drawn directly from the New York Times, that will substantially harm and maybe even kill the Times' business model.