Opening a conversation on responsible environmental data science in the age of large language models

被引:0
|
作者
Oliver, Ruth Y. [1 ]
Chapman, Melissa [2 ]
Emery, Nathan [3 ]
Gillespie, Lauren [4 ]
Gownaris, Natasha [5 ]
Leiker, Sophia [1 ]
Nisi, Anna C. [6 ]
Ayers, David [7 ]
Breckheimer, Ian [8 ]
Blondin, Hannah [9 ]
Hoffman, Ava [10 ]
Pagniello, Camille M. L. S. [11 ]
Raisle, Megan [12 ]
Zimmerman, Naupaka [12 ]
机构
[1] Univ Calif Santa Barbara, Bren Sch Environm Sci & Management, Santa Barbara, CA USA
[2] Univ Calif Santa Barbara, Natl Ctr Ecol Anal & Synth, Santa Barbara, CA USA
[3] Univ Calif Santa Barbara, Ctr Innovat Teaching Res & Learning, Santa Barbara, CA USA
[4] Stanford Univ, Dept Comp Sci, Palo Alto, CA USA
[5] Gettysburg Coll, Dept Environm Studies, Gettysburg, PA USA
[6] Univ Washington, Dept Biol, Ctr Ecosyst Sentinels, Seattle, WA USA
[7] Univ Calif Davis, Wildlife Fish & Conservat Biol Dept, Davis, CA USA
[8] Rocky Mt Biol Labs, Crested Butte, CO USA
[9] Univ Miami, Cooperat Inst Marine & Atmospher Studies CIMAS, Miami, FL USA
[10] Fred Hutchinson Canc Ctr, Data Sci Lab, Seattle, WA USA
[11] Univ Hawaii Manoa, Hawaii Inst Marine Biol, Kaneohe, HI USA
[12] Univ San Francisco, Dept Biol, San Francisco, CA USA
来源
关键词
bias; ChatGPT; data ethics; generative AI; pedagogy;
D O I
10.1017/eds.2024.12
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
The general public and scientific community alike are abuzz over the release of ChatGPT and GPT-4. Among many concerns being raised about the emergence and widespread use of tools based on large language models (LLMs) is the potential for them to propagate biases and inequities. We hope to open a conversation within the environmental data science community to encourage the circumspect and responsible use of LLMs. Here, we pose a series of questions aimed at fostering discussion and initiating a larger dialogue. To improve literacy on these tools, we provide background information on the LLMs that underpin tools like ChatGPT. We identify key areas in research and teaching in environmental data science where these tools may be applied, and discuss limitations to their use and points of concern. We also discuss ethical considerations surrounding the use of LLMs to ensure that as environmental data scientists, researchers, and instructors, we can make well-considered and informed choices about engagement with these tools. Our goal is to spark forward-looking discussion and research on how as a community we can responsibly integrate generative AI technologies into our work. Impact Statement With the recent release of ChatGPT and similar tools based on large language models, there is considerable enthusiasm and substantial concern over how these tools should be used. We pose a series of questions aimed at unpacking important considerations in the responsible use of large language models within environmental data science.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Promoting interactions between cognitive science and large language models
    Qu, Youzhi
    Du, Penghui
    Che, Wenxin
    Wei, Chen
    Zhang, Chi
    Ouyang, Wanli
    Bian, Yatao
    Xu, Feiyang
    Hu, Bin
    Du, Kai
    Wu, Haiyan
    Liu, Jia
    Liu, Quanying
    Innovation, 2024, 5 (02):
  • [32] Can Large Language Models Transform Computational Social Science?
    Ziems, Caleb
    Held, William
    Shaikh, Omar
    Chen, Jiaao
    Zhang, Zhehao
    Yang, Diyi
    COMPUTATIONAL LINGUISTICS, 2023, 50 (01) : 237 - 291
  • [33] Performance of Large Language Models in a Computer Science Degree Program
    Krueger, Tim
    Gref, Michael
    ARTIFICIAL INTELLIGENCE-ECAI 2023 INTERNATIONAL WORKSHOPS, PT 2, XAI3, TACTIFUL, XI-ML, SEDAMI, RAAIT, AI4S, HYDRA, AI4AI, 2023, 2024, 1948 : 409 - 424
  • [34] The use of ChatGPT and other large language models in surgical science
    Janssen, Boris, V
    Kazemier, Geert
    Besselink, Marc G.
    BJS OPEN, 2023, 7 (02):
  • [35] Promoting interactions between cognitive science and large language models
    Qu, Youzhi
    Du, Penghui
    Che, Wenxin
    Wei, Chen
    Zhang, Chi
    Ouyang, Wanli
    Bian, Yatao
    Xu, Feiyang
    Hu, Bin
    Du, Kai
    Wu, Haiyan
    Liu, Jia
    Liu, Quanying
    INNOVATION-THE EUROPEAN JOURNAL OF SOCIAL SCIENCE RESEARCH, 2024, 5 (02)
  • [36] MaScQA: investigating materials science knowledge of large language models
    Zaki, Mohd
    Jayadeva
    Mausam
    Krishnan, N. M. Anoop
    DIGITAL DISCOVERY, 2024, 3 (02): : 313 - 327
  • [37] Large language models in food science: Innovations, applications, and future
    Ma, Peihua
    Tsai, Shawn
    He, Yiyang
    Jia, Xiaoxue
    Zhen, Dongyang
    Yu, Ning
    Wang, Qin
    Ahuja, Jaspreet K. C.
    Wei, Cheng -, I
    TRENDS IN FOOD SCIENCE & TECHNOLOGY, 2024, 148
  • [38] Reconciling the contrasting narratives on the environmental impact of large language models
    Ren, Shaolei
    Tomlinson, Bill
    Black, Rebecca W.
    Torrance, Andrew W.
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [39] Extracting Training Data from Large Language Models
    Carlini, Nicholas
    Tramer, Florian
    Wallace, Eric
    Jagielski, Matthew
    Herbert-Voss, Ariel
    Lee, Katherine
    Roberts, Adam
    Brown, Tom
    Song, Dawn
    Erlingsson, Ulfar
    Oprea, Alina
    Raffel, Colin
    PROCEEDINGS OF THE 30TH USENIX SECURITY SYMPOSIUM, 2021, : 2633 - 2650
  • [40] How Large Language Models Will Disrupt Data Management
    Fernandez, Raul Castro
    Elmore, Aaron J.
    Franklin, Michael J.
    Krishnan, Sanjay
    Tan, Chenhao
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2023, 16 (11): : 3302 - 3309