BB-GeoGPT: A framework for learning a large language model for geographic information science

被引:12
|
作者
Zhang, Yifan [1 ]
Wang, Zhiyun [1 ]
He, Zhengting [1 ]
Li, Jingxuan [1 ]
Mai, Gengchen [2 ,3 ]
Lin, Jianfeng [4 ]
Wei, Cheng [1 ]
Yu, Wenhao [1 ,5 ]
机构
[1] China Univ Geosci, Sch Geog & Informat Engn, Wuhan 430078, Peoples R China
[2] Univ Texas Austin, Dept Geog & Environm, SEAI Lab, Austin, TX 78712 USA
[3] Univ Georgia, Dept Geog, SEAI Lab, Athens, GA 30602 USA
[4] Meituan, Beijing 100102, Peoples R China
[5] China Univ Geosci, Natl Engn Res Ctr Geog Informat Syst, Wuhan 430078, Peoples R China
基金
中国国家自然科学基金;
关键词
Large language model; GIS knowledge corpus; Domain adaptation; Self-instruct instructions; DISAMBIGUATION;
D O I
10.1016/j.ipm.2024.103808
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large language models (LLMs) exhibit impressive capabilities across diverse tasks in natural language processing. Nevertheless, challenges arise such as large model parameter size and limited model accessibility through APIs such as ChatGPT and GPT-4, which prohibits the model deployment on mobile devices and domain adaptation or fine-tuning. Moreover, while LLMs excel in general domains, their performance in specialized fields such as GIS may not always align with the expectations of domain experts. This is primarily attributed to the diverse disciplinary origins of the training data, which often lack comprehensive coverage and treatment of knowledge specific to individual disciplines (e.g., GIS). Therefore, there is a crucial need to train and adapt LLMs specifically designed for different professional fields. In this paper, our focus is on the GIS domain, where we introduce BB(BaBy)-GeoGPT, a large language model with GIS-specific knowledge. To achieve this goal, we curated a comprehensive set of resources, comprising model pretraining data (BB-GeoPT, 26,907 documents), supervised fine-tuning data (BB-GeoSFT, 35,876 instructions), and evaluation data (BB-GeoEval, 600 objective questions and 150 subjective questions). BB-GeoGPT is developed by first adapting an open-source generaldomain LLM, the LLaMA-2-7B model, to our pretraining data. Subsequently, we use instruction tuning to further fine-tune the model on our BB-GeoSFT. Through extensive experiments on the evaluation dataset, BB-GeoGPT demonstrates improvements ranging from 10.55% to 47.57% for objective questions and from 7.87% to 27.73% for subjective questions, when compared to general LLMs of similar size in terms of accuracy. Moreover, our data collection strategy and the amassed data can serve as a foundation for advancing LLM research in the GIS domain, fostering further development.
引用
收藏
页数:19
相关论文
共 50 条
  • [21] Teaching science with the "universal language" of music: alignment with the Universal Design for Learning framework
    Crowther, Gregory J.
    Adjapong, Edmund
    Jenkins, Lekelia D.
    ADVANCES IN PHYSIOLOGY EDUCATION, 2023, 47 (03) : 491 - 498
  • [22] MapReader: a framework for learning a visual language model for map analysis
    Zhang, Yifan
    Zhang, Wenbo
    Zeng, Ziyi
    Jiang, Keying
    Li, Jingxuan
    Min, Wen
    Luo, Wei
    Guan, Qingfeng
    Lin, Jianfeng
    Yu, Wenhao
    INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE, 2025,
  • [23] UNIREX: A Unified Learning Framework for Language Model Rationale Extraction
    Chan, Aaron
    Sanjabi, Maziar
    Mathias, Lambert
    Tan, Liang
    Nie, Shaoliang
    Peng, Xiaochang
    Ren, Xiang
    Firooz, Hamed
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [24] UNIREX: A Unified Learning Framework for Language Model Rationale Extraction
    Chan, Aaron
    Sanjabi, Maziar
    Mathias, Lambert
    Tan, Liang
    Nie, Shaoliang
    Peng, Xiaochang
    Ren, Xiang
    Firooz, Hamed
    PROCEEDINGS OF WORKSHOP ON CHALLENGES & PERSPECTIVES IN CREATING LARGE LANGUAGE MODELS (BIGSCIENCE EPISODE #5), 2022, : 51 - 67
  • [25] Automated Building Information Modeling Compliance Check through a Large Language Model Combined with Deep Learning and Ontology
    Chen, Nanjiang
    Lin, Xuhui
    Jiang, Hai
    An, Yi
    BUILDINGS, 2024, 14 (07)
  • [26] Incorporating passage feature within language model framework for information retrieval
    Dang, Ke
    Zhao, Tiejun
    Qi, Haoliang
    Zheng, Dequan
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, 2007, 4394 : 476 - +
  • [27] A Retrieval-Augmented Framework for Tabular Interpretation with Large Language Model
    Yan, Mengyi
    Rene, Weilong
    Wang, Yaoshu
    Li, Jianxin
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2024, PT 2, 2025, 14851 : 341 - 356
  • [28] MatChat: A large language model and application service platform for materials science
    Chen, Zi-Yi
    Xie, Fan-Kai
    Wan, Meng
    Yuan, Yang
    Liu, Miao
    Wang, Zong-Guo
    Meng, Sheng
    Wang, Yan-Gang
    CHINESE PHYSICS B, 2023, 32 (11)
  • [29] MatChat: A large language model and application service platform for materials science
    陈子逸
    谢帆恺
    万萌
    袁扬
    刘淼
    王宗国
    孟胜
    王彦棡
    Chinese Physics B, 2023, (11) : 208 - 213
  • [30] BookGPT: A General Framework for Book Recommendation Empowered by Large Language Model
    Li, Zhiyu
    Chen, Yanfang
    Zhang, Xuan
    Liang, Xun
    ELECTRONICS, 2023, 12 (22)