BB-GeoGPT: A framework for learning a large language model for geographic information science

被引:12
|
作者
Zhang, Yifan [1 ]
Wang, Zhiyun [1 ]
He, Zhengting [1 ]
Li, Jingxuan [1 ]
Mai, Gengchen [2 ,3 ]
Lin, Jianfeng [4 ]
Wei, Cheng [1 ]
Yu, Wenhao [1 ,5 ]
机构
[1] China Univ Geosci, Sch Geog & Informat Engn, Wuhan 430078, Peoples R China
[2] Univ Texas Austin, Dept Geog & Environm, SEAI Lab, Austin, TX 78712 USA
[3] Univ Georgia, Dept Geog, SEAI Lab, Athens, GA 30602 USA
[4] Meituan, Beijing 100102, Peoples R China
[5] China Univ Geosci, Natl Engn Res Ctr Geog Informat Syst, Wuhan 430078, Peoples R China
基金
中国国家自然科学基金;
关键词
Large language model; GIS knowledge corpus; Domain adaptation; Self-instruct instructions; DISAMBIGUATION;
D O I
10.1016/j.ipm.2024.103808
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large language models (LLMs) exhibit impressive capabilities across diverse tasks in natural language processing. Nevertheless, challenges arise such as large model parameter size and limited model accessibility through APIs such as ChatGPT and GPT-4, which prohibits the model deployment on mobile devices and domain adaptation or fine-tuning. Moreover, while LLMs excel in general domains, their performance in specialized fields such as GIS may not always align with the expectations of domain experts. This is primarily attributed to the diverse disciplinary origins of the training data, which often lack comprehensive coverage and treatment of knowledge specific to individual disciplines (e.g., GIS). Therefore, there is a crucial need to train and adapt LLMs specifically designed for different professional fields. In this paper, our focus is on the GIS domain, where we introduce BB(BaBy)-GeoGPT, a large language model with GIS-specific knowledge. To achieve this goal, we curated a comprehensive set of resources, comprising model pretraining data (BB-GeoPT, 26,907 documents), supervised fine-tuning data (BB-GeoSFT, 35,876 instructions), and evaluation data (BB-GeoEval, 600 objective questions and 150 subjective questions). BB-GeoGPT is developed by first adapting an open-source generaldomain LLM, the LLaMA-2-7B model, to our pretraining data. Subsequently, we use instruction tuning to further fine-tune the model on our BB-GeoSFT. Through extensive experiments on the evaluation dataset, BB-GeoGPT demonstrates improvements ranging from 10.55% to 47.57% for objective questions and from 7.87% to 27.73% for subjective questions, when compared to general LLMs of similar size in terms of accuracy. Moreover, our data collection strategy and the amassed data can serve as a foundation for advancing LLM research in the GIS domain, fostering further development.
引用
收藏
页数:19
相关论文
共 50 条
  • [41] Quantitative monitoring of pollution in Mitrovica Lake using empirical model and geographic information science & systems
    Maliqi, Edon
    Jusufi, Kaltrina
    INTERNATIONAL JOURNAL OF ENVIRONMENTAL ANALYTICAL CHEMISTRY, 2024,
  • [42] A General Embedding Framework for Heterogeneous Information Learning in Large-Scale Networks
    Huang, Xiao
    Li, Jundong
    Zou, Na
    Hu, Xia
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2018, 12 (06)
  • [43] MapGPT: an autonomous framework for mapping by integrating large language model and cartographic tools
    Zhang, Yifan
    He, Zhengting
    Li, Jingxuan
    Lin, Jianfeng
    Guan, Qingfeng
    Yu, Wenhao
    CARTOGRAPHY AND GEOGRAPHIC INFORMATION SCIENCE, 2024, 51 (06) : 717 - 743
  • [44] A novel large-language-model-driven framework for named entity recognition
    Wang, Zhenhua
    Chen, Huiru
    Xu, Guang
    Ren, Ming
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (03)
  • [45] StructGPT: A General Framework for Large Language Model to Reason over Structured Data
    Jiang, Jinhao
    Zhou, Kun
    Dong, Zican
    Ye, Keming
    Zhao, Wayne Xin
    Wen, Ji-Rong
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9237 - 9251
  • [46] TALLRec: An Effective and Efficient Tuning Framework to Align Large Language Model with Recommendation
    Bao, Keqin
    Zhang, Jizhi
    Zhang, Yang
    Wang, Wenjie
    Feng, Fuli
    He, Xiangnan
    PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, : 1007 - 1014
  • [47] Large Language Model Evaluation Criteria Framework in Healthcare: Fuzzy MCDM Approach
    Hamzeh Mohammad Alabool
    SN Computer Science, 6 (1)
  • [48] LUNA: A Model-Based Universal Analysis Framework for Large Language Models
    Song, Da
    Xie, Xuan
    Song, Jiayang
    Zhu, Derui
    Huang, Yuheng
    Felix, Juefei-Xu
    Ma, Lei
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2024, 50 (07) : 1921 - 1948
  • [49] METAL: Metamorphic Testing Framework for Analyzing Large-Language Model Qualities
    Hyun, Sangwon
    Guo, Mingyu
    Babar, M. Ali
    2024 IEEE CONFERENCE ON SOFTWARE TESTING, VERIFICATION AND VALIDATION, ICST 2024, 2024, : 117 - 128
  • [50] A framework for collaborating a Large Language Model tool in brainstorming for triggering creative thoughts
    Chang, Hung-Fu
    Li, Tong
    THINKING SKILLS AND CREATIVITY, 2025, 56