PhD Proposal: Watermarking for Transparent Large Language Model Development and Deployment

Talk
John Kirchenbauer
Time: 
11.12.2025 13:00 to 14:30
Location: 

IRB-5105 or https://umd.zoom.us/j/97837256573 (Passcode: wm4attr)

Over the last five years, the rapid increase in the capability and ubiquity of generative models has revealed a clear and pressing need to incorporate transparency enhancing technologies into these systems. In the first part of this talk, I will introduce our work on output watermarking for large language models and showcase how this technology can enable robust content provenance in realistic deployment scenarios. Next, I will discuss our work on training data attribution and memorization in large language models while bringing to light the ways in which web-scale pretraining presents unique and fundamental challenges in this space. Then, motivated by prescient issues at the intersection of intellectual property law and generative model training, I will re-contextualize output watermarks as general-purpose tools for partitioning the output space of a generative model into sets that are trivially distinguishable with the watermark secret, but indistinguishable without it. Finally, I will use this framing to motivate a proposal for making the problems of training data attribution and membership inference more tractable via proactive, selective watermarking by content owners and creators.