Before they gobbled up headlines everywhere, large language models ingested truly staggering amounts of data to train their models. That training data didn’t emerge from the ether: Some of it came from other people’s creativity and work.
https://themarkup.org/hello-world/2023/04/22/copyright-showdown-ais-next-frontier