How to Protect Copyright Data in Optimization of Large Language Models?

被引:0
|
作者
Chu, Timothy [1 ]
Song, Zhao [2 ]
Yang, Chiwun [3 ]
机构
[1] Google, Mountain View, CA 94043 USA
[2] Adobe Res, San Jose, CA USA
[3] Sun Yat Sen Univ, Guangzhou, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) and generative AI have played a transformative role in computer research and applications. Controversy has arisen as to whether these models output copyrighted data, which can occur if the data the models are trained on is copyrighted. LLMs are built on the transformer neural network architecture, which in turn relies on a mathematical computation called Attention that uses the softmax function. In this paper, we observe that large language model training and optimization can be seen as a softmax regression problem. We then establish a method of efficiently performing softmax regression, in a way that prevents the regression function from generating copyright data. This establishes a theoretical method of training large language models in a way that avoids generating copyright data.
引用
收藏
页码:17871 / 17879
页数:9
相关论文
共 50 条
  • [11] Research on Fine-Tuning Optimization Strategies for Large Language Models in Tabular Data Processing
    Zhao, Xiaoyong
    Leng, Xingxin
    Wang, Lei
    Wang, Ningning
    BIOMIMETICS, 2024, 9 (11)
  • [12] Extracting Training Data from Large Language Models
    Carlini, Nicholas
    Tramer, Florian
    Wallace, Eric
    Jagielski, Matthew
    Herbert-Voss, Ariel
    Lee, Katherine
    Roberts, Adam
    Brown, Tom
    Song, Dawn
    Erlingsson, Ulfar
    Oprea, Alina
    Raffel, Colin
    PROCEEDINGS OF THE 30TH USENIX SECURITY SYMPOSIUM, 2021, : 2633 - 2650
  • [13] Leveraging Large Language Models for Sensor Data Retrieval
    Berenguer, Alberto
    Morejon, Adriana
    Tomas, David
    Mazon, Jose-Norberto
    APPLIED SCIENCES-BASEL, 2024, 14 (06):
  • [14] Empowering Large Language Models for Textual Data Augmentation
    Li, Yichuan
    Ding, Kaize
    Wang, Jianling
    Lee, Kyumin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 12734 - 12751
  • [15] Making Large Language Models Better Data Creators
    Lee, Dong-Ho
    Pujara, Jay
    Sewak, Mohit
    White, Ryen W.
    Jauhar, Sujay Kumar
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 15349 - 15360
  • [16] Automating Qualitative Data Analysis with Large Language Models
    Parfenova, Angelina
    Denzler, Alexander
    Pfeffer, Juergen
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 4: STUDENT RESEARCH WORKSHOP, 2024, : 101 - 109
  • [17] Leveraging large language models for data analysis automation
    Jansen, Jacqueline A.
    Manukyan, Artur
    Al Khoury, Nour
    Akalin, Altuna
    PLOS ONE, 2025, 20 (02):
  • [18] Are You Copying My Model? Protecting the Copyright of Large Language Models for EaaS via BackdoorWatermark
    Peng, Wenjun
    Yi, Jingwei
    Wu, Fangzhao
    Wu, Shangxi
    Zhu, Bin
    Lyu, Lingjuan
    Jiao, Binxing
    Xu, Tong
    Sun, Guangzhong
    Xie, Xing
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 7653 - 7668
  • [19] How to Use Large Language Models for Empirical Legal Research
    Choi, Jonathan H.
    JOURNAL OF INSTITUTIONAL AND THEORETICAL ECONOMICS-ZEITSCHRIFT FUR DIE GESAMTE STAATSWISSENSCHAFT, 2024, 180 (02): : 214 - 233
  • [20] Will Large Language Models Really Change How Work Is Done?
    Cappelli, Peter
    Tambe, Prasanna
    Yakubovich, Valery
    MIT SLOAN MANAGEMENT REVIEW, 2024, 65 (03)