BB-GeoGPT: A framework for learning a large language model for geographic information science

被引:12
|
作者
Zhang, Yifan [1 ]
Wang, Zhiyun [1 ]
He, Zhengting [1 ]
Li, Jingxuan [1 ]
Mai, Gengchen [2 ,3 ]
Lin, Jianfeng [4 ]
Wei, Cheng [1 ]
Yu, Wenhao [1 ,5 ]
机构
[1] China Univ Geosci, Sch Geog & Informat Engn, Wuhan 430078, Peoples R China
[2] Univ Texas Austin, Dept Geog & Environm, SEAI Lab, Austin, TX 78712 USA
[3] Univ Georgia, Dept Geog, SEAI Lab, Athens, GA 30602 USA
[4] Meituan, Beijing 100102, Peoples R China
[5] China Univ Geosci, Natl Engn Res Ctr Geog Informat Syst, Wuhan 430078, Peoples R China
基金
中国国家自然科学基金;
关键词
Large language model; GIS knowledge corpus; Domain adaptation; Self-instruct instructions; DISAMBIGUATION;
D O I
10.1016/j.ipm.2024.103808
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large language models (LLMs) exhibit impressive capabilities across diverse tasks in natural language processing. Nevertheless, challenges arise such as large model parameter size and limited model accessibility through APIs such as ChatGPT and GPT-4, which prohibits the model deployment on mobile devices and domain adaptation or fine-tuning. Moreover, while LLMs excel in general domains, their performance in specialized fields such as GIS may not always align with the expectations of domain experts. This is primarily attributed to the diverse disciplinary origins of the training data, which often lack comprehensive coverage and treatment of knowledge specific to individual disciplines (e.g., GIS). Therefore, there is a crucial need to train and adapt LLMs specifically designed for different professional fields. In this paper, our focus is on the GIS domain, where we introduce BB(BaBy)-GeoGPT, a large language model with GIS-specific knowledge. To achieve this goal, we curated a comprehensive set of resources, comprising model pretraining data (BB-GeoPT, 26,907 documents), supervised fine-tuning data (BB-GeoSFT, 35,876 instructions), and evaluation data (BB-GeoEval, 600 objective questions and 150 subjective questions). BB-GeoGPT is developed by first adapting an open-source generaldomain LLM, the LLaMA-2-7B model, to our pretraining data. Subsequently, we use instruction tuning to further fine-tune the model on our BB-GeoSFT. Through extensive experiments on the evaluation dataset, BB-GeoGPT demonstrates improvements ranging from 10.55% to 47.57% for objective questions and from 7.87% to 27.73% for subjective questions, when compared to general LLMs of similar size in terms of accuracy. Moreover, our data collection strategy and the amassed data can serve as a foundation for advancing LLM research in the GIS domain, fostering further development.
引用
收藏
页数:19
相关论文
共 50 条
  • [31] A Framework for Agricultural Intelligent Analysis Based on a Visual Language Large Model
    Yu, Piaofang
    Lin, Bo
    APPLIED SCIENCES-BASEL, 2024, 14 (18):
  • [32] GeoLLM: A specialized large language model framework for intelligent geotechnical design
    Xu, Hao-Ruo
    Zhang, Ning
    Yin, Zhen-Yu
    Njock, Pierre Guy Atangana
    COMPUTERS AND GEOTECHNICS, 2025, 177
  • [33] Accelerate Large Language Model Inference on Edge TPU with OpenVX framework
    Wu, You-En
    Wu, Hsin-I
    Chin, Kuo-Cheng
    Yang, Yi-Chun
    Tsay, Ren-Song
    2024 IEEE 6TH INTERNATIONAL CONFERENCE ON AI CIRCUITS AND SYSTEMS, AICAS 2024, 2024, : 502 - 506
  • [34] A Generative Adaptive Context Learning Framework for Large Language Models in Cheapfake Detection
    Pham, Long-Khanh
    Vo-Hoang, Hoa-Vien
    Tran, Anh-Duy
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 1288 - 1293
  • [35] Pareto Optimal Learning for Estimating Large Language Model Errors
    Zhao, Theodore
    Wei, Mu
    Preston, Joseph S.
    Poon, Hoifung
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 10513 - 10529
  • [36] A wholistic model for teaching computer and information science: learning on a WhIM
    Lorence, Daniel
    Jones, Arthur
    INTERNATIONAL JOURNAL OF INNOVATION AND LEARNING, 2007, 4 (04) : 376 - 390
  • [37] A large language model and denoising diffusion framework for targeted design of microstructures with commands in natural language
    Kartashov, Nikita
    Vlassis, Nikolaos N.
    COMPUTER METHODS IN APPLIED MECHANICS AND ENGINEERING, 2025, 437
  • [38] A novel large language model enhanced joint learning framework for fine-grained sentiment analysis on drug reviews
    Zou, Haochen
    Wang, Yongli
    NEUROCOMPUTING, 2025, 626
  • [39] GIS-KG: building a large-scale hierarchical knowledge graph for geographic information science
    Du, Jiaxin
    Wang, Shaohua
    Ye, Xinyue
    Sinton, Diana S.
    Kemp, Karen
    INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE, 2022, 36 (05) : 873 - 897
  • [40] FinBERT: A Large Language Model for Extracting Information from Financial Text
    Huang, Allen H.
    Wang, Hui
    Yang, Yi
    CONTEMPORARY ACCOUNTING RESEARCH, 2023, 40 (02) : 806 - 841