
Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.
Sign Up For Latest News
Foundation models are often trained on large volumes of copyrighted material. In the United States, AI researchers have long relied on fair use doctrine to avoid copyright issues with training data. However, our U.S. case law analysis in this brief highlights that fair use is not guaranteed for foundation models and that the risk of copyright infringement is real, though the exact extent remains uncertain. We argue that the United States needs a two-pronged approach to addressing these copyright issues—a mix of legal and technical mitigations that will allow us to harness the positive impact of foundation models while reducing intellectual property harms to creators.
Large Language Models Can Be Strong Differentially Private Learners