Content area

Abstract

Geospatial code generation is emerging as a key direction in the integration of artificial intelligence and geoscientific analysis. However, there remains a lack of standardized tools for automatic evaluation in this domain. To address this gap, we propose AutoGEEval, the first multimodal, unit-level automated evaluation framework for geospatial code generation tasks on the Google Earth Engine (GEE) platform powered by large language models (LLMs). Built upon the GEE Python API, AutoGEEval establishes a benchmark suite (AutoGEEval-Bench) comprising 1325 test cases that span 26 GEE data types. The framework integrates both question generation and answer verification components to enable an end-to-end automated evaluation pipeline—from function invocation to execution validation. AutoGEEval supports multidimensional quantitative analysis of model outputs in terms of accuracy, resource consumption, execution efficiency, and error types. We evaluate 18 state-of-the-art LLMs—including general-purpose, reasoning-augmented, code-centric, and geoscience-specialized models—revealing their performance characteristics and potential optimization pathways in GEE code generation. This work provides a unified protocol and foundational resource for the development and assessment of geospatial code generation models, advancing the frontier of automated natural language to domain-specific code translation.

Details

1009240
Business indexing term
Company / organization
Title
AutoGEEval: A Multimodal and Automated Evaluation Framework for Geospatial Code Generation on GEE with Large Language Models
Author
Wu Huayi 1   VIAFID ORCID Logo  ; Shen Zhangxiao 2 ; Hou Shuyang 2 ; Liang Jianyuan 2 ; Jiao Haoyue 3 ; Yaxian, Qing 2   VIAFID ORCID Logo  ; Zhang Xiaopu 2   VIAFID ORCID Logo  ; Xu, Li 2 ; Gui Zhipeng 4   VIAFID ORCID Logo  ; Guan Xuefeng 2   VIAFID ORCID Logo  ; Longgang, Xiang 2   VIAFID ORCID Logo 

 State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China; [email protected] (H.W.); [email protected] (Z.S.); [email protected] (J.L.); [email protected] (Y.Q.); [email protected] (X.Z.); [email protected] (X.L.); [email protected] (X.G.); [email protected] (L.X.), Collaborative Innovation Center of Geospatial Technology, Wuhan University, Wuhan 430079, China 
 State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China; [email protected] (H.W.); [email protected] (Z.S.); [email protected] (J.L.); [email protected] (Y.Q.); [email protected] (X.Z.); [email protected] (X.L.); [email protected] (X.G.); [email protected] (L.X.) 
 School of Resource and Environmental Sciences, Wuhan University, Wuhan 430079, China; [email protected] 
 School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China; [email protected] 
Volume
14
Issue
7
First page
256
Number of pages
31
Publication year
2025
Publication date
2025
Publisher
MDPI AG
Place of publication
Basel
Country of publication
Switzerland
Publication subject
e-ISSN
22209964
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
Publication history
 
 
Online publication date
2025-06-30
Milestone dates
2025-05-15 (Received); 2025-06-29 (Accepted)
Publication history
 
 
   First posting date
30 Jun 2025
ProQuest document ID
3233222639
Document URL
https://www.proquest.com/scholarly-journals/autogeeval-multimodal-automated-evaluation/docview/3233222639/se-2?accountid=208611
Copyright
© 2025 by the authors. Published by MDPI on behalf of the International Society for Photogrammetry and Remote Sensing. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2025-07-25
Database
2 databases
  • ProQuest One Academic
  • ProQuest One Academic