Select Page

Huan Ning, Zhenlong Li | April 13, 2025

In a recent vision paper, GIScience in the Era of Artificial Intelligence: A Research Agenda Towards Autonomous GIS, we proposed three scales of autonomous GIS agents: local, centralized, and infrastructure. Some GIS agents have been developed on local and centralized scales, such as LLM-Geo and Google Geospatial Reasoning. What do the infrastructure agents look like? In the vision paper, we discussed that “A key challenge [to develop the infrastructure-scale GIS agents] lies in establishing secure standards, protocols and policies to let generative AI manage computational resources.” The release of Google’s A2A protocol (early April 2025) to support communications of agents shined a light on the infrastructure scale agents, allowing AI agents can collaborate with each other to accomplish complex tasks using the resources across different types of infrastructure, such as data and computational resources.

Besides protocols among AI agents like A2A, Anthropic AI’s Model Context Protocol (MCP) focuses on AI-ready resources (e.g., data and tools). Released in late 2024, it aims to expose resources for large language models (LLMs). Our understanding is that MCP is a wrapper of existing resources (e.g., data, services) to provide information or materials for LLMs, so that they can provide better answers for users, rather than output plausible text according to LLM’s internal knowledge. In GIScience, geospatial data sharing attempts led by Open Geospatial Consortium (OGC) have been started since 1994, benefiting the geospatial community. Online geospatial resources that meet OGC standards can be shared across programs via API (application programming interface). In addition, the strong needs for geospatial data sharing, integration, and discovery led to the development of National Spatial Data Infrastructure (NSDI) and the large NSF-funded initiatives such as EarthCube and I-GUIDE.  In the era of artificial intelligence (AI), these geospatial resources may need to be wrapped up for AI applications, especially for autonomous GIS agents. Our paper, Autonomous GIS: the next-generation AI-powered GIS has mentioned the need to prepare data for AI agents, and we implemented a data retrieval agent LLM-Find and it’s QGIS implementation, demonstrating a practical and plug-and-play manner for AI applications to access geospatial data. MCP is probably a more general way to share data for AI, especially authorization, but whether it can be adapted for geospatial data need further investigation.

Due to the nature of LLMs, many of these solutions or practices, like A2A, MCP, LLM-Find, and GIS Copilot, used a straightforward method – feeding descriptions of all available resources (e.g., agents, tools, and data) into an LLM and expect the LLM chooses the needed resources. This strategy faces the scalability problem: what if there are hundreds or thousands of resources? For example, there are about 30,000 Census variables and 200,000 OGC services (by PolarHub). Obviously, it is challenging to put the super-long descriptions of all variables and services into the LLM. Our team is testing the RAG (retrieval augmented generation) in GIS Copilot to help the Copilot to select appropriate tools to process geospatial data for users. We think fine-tuning multiple LLMs may be an alternative. Since humans have created countless public and private services for geospatial data infrastructure, autonomous GIS agents at the infrastructure scale need to discover, test, document, and recommend these resources, but how? We think this is a fascinating and useful research topic.