Applying AI methods to digitised map collections around the world has emerged as a major area of interest, supporting library and archive goals of improving collection discovery and reuse (McDonough, 2024; McDonough and Vitale, 2024). The timely intersection of web mapping, imaging, and computer vision technologies with maturing IIIF infrastructure and digital humanities expertise created the foundation for projects like Allmaps (https://allmaps.org/), Machines Reading Maps (MRM, https://machines-reading-maps.github.io/), and MapReader (https://github.com/maps-as-data/MapReader, Hosseini et al, 2022). In particular, MRM’s efforts to automatically transcribe text on maps were a key demonstration of how AI experiments in libraries are also useful to scholarly communities in other disciplines: text on maps is a major new way of exploring the history of cartography, landscapes, and societies (McDonough et al, 2024; Nelson, 2024). Allmaps and MapReader are now linking up the data, software, and underlying infrastructures of these projects to further support open map research and collection discovery.
Using the mapKurator text spotting pipeline (now PALETTE, Lin and Chiang, 2024), MRM created data for all the text printed on 56,628 georeferenced maps in the David Rumsey Historical Map Collection, radically transforming how students, researchers, and the general public can search the collection (Vitale, 2023; Larsen et al, 2024). More recently, the open-source software MapReader has implemented text spotting as well as support for IIIF input (https://mapreader.readthedocs.io/en/latest/using-mapreader/step-by-step-guide/1-download.html#downloading-maps-from-iiif-servers), aiming to create accessible, and humanities-oriented computer vision research tools (Wood et al, 2024; Coleman, 2024). For Fantastic Futures 2025, we will share recent work to lower barriers to working with the output of text spotting, best practices in data interoperability, and examples of research enabled by these efforts.
Text spotting outputs represent visual, textual, and spatial data simultaneously, but with new kinds of errors that curators and scholars are just now beginning to understand: viewing these outputs on a map is an essential part of devising evaluation metrics, and formulating research questions based on imperfect data. We are therefore developing a new IIIF-friendly approach to formatting text spotting output and simplifying it for browser-based visualisation. This turns the inferred outputs from AI methods into accessible, explorable data useful in both library discovery interfaces and in research applications.
Text spotting output from mapKurator for the Rumsey collection was delivered as 56,628 geojson files (1 per map) collectively containing 111,072,386 instances of text (zipped 51.65 GB). (3 versions of the data exist: versions 1 and 2 are openly available via the Stanford Digital Repository (Machines Reading Maps Collection, 2023); version 3 will be released alongside a forthcoming publication. Work here is based on version 3.) Each filename refers to a UID in the Rumsey metadata, but no metadata was contained in these output files. This data was added alongside the Rumsey catalog to enable “Text on Maps”-based searching of the Rumsey collection. However, because of the size of this data, and the absence of metadata, the data was very difficult to evaluate or analyze.
If the Rumsey site facilitates initial exploration and map-by-map viewing, we now need more robust tools for viewing,reading, filtering, and exporting across maps based on metadata and other data-driven criteria. To work towards this goal, we used Allmaps tools to transform each instance of text in the original geojson mapKurator output into IIIF Georeference Annotations (https://iiif.io/api/extension/georef/). Next, we combined annotations from all sheets into a single composite JSON file which also includes selected metadata (e.g. title, creator, publication date). We then transformed the JSON data into vector tiles. With Allmaps, it is now possible to plot and explore this data on a web map, but also to plot the text annotations in their correct pixel locations on top of the original scanned map image: pixel coordinates can be transformed into geospatial coordinates, and vice versa. Examples of this interoperability are viewable at https://observablehq.com/d/f5b83904983ce90b and https://observablehq.com/@allmaps/rumsey (van Wissen et al, 2025). A significant advantage of this approach is that if the georeferencing metadata is updated in the future, the text spotting data can also be easily updated. All new data will be published shortly on Zenodo and code is available at https://github.com/allmaps/text-on-maps-viz.
One aim of this work is that the results of this data curation work will serve as a model for future text spotting output formatting: going forward we plan to implement IIIF georeference annotations as a standard output option for the text spotting task in MapReader. We are embedding open standards and IIIF interoperability across the emerging ‘Open Maps’ landscape (Schoonman and Baptist, 2024), in particular in Allmaps and MapReader functionality. Simplifying the ability to move between JSON/GeoJSON and Web Annotations/IIIF Georeference Annotations is powerful, reducing file sizes and enabling work with both pixel or geospatial data.
In this presentation, we will discuss these workflows and visualisation interfaces through examples of research in progress that is enabled by this work. After using the Rumsey data as the initial use case for the approach to structuring text spotting output as IIIF annotations, we are developing approaches to visualising change over time in both the Rumsey data and in two editions of large-scale Ordnance Survey (OS) sheets from the National Library of Scotland. While the Rumsey collection’s global scope over 5 centuries presents certain visualisation and research opportunities, the OS is more circumspect both geographically (to Britain) and temporally (to the nineteenth and early twentieth centuries). With each collection, the team is learning how to interact with text spotting output given both cartographic and AI biases. Through such research, the broader community can become more adept at interrogating big data derived from digital image collections as well as better prepared to develop approaches to presenting this data to the library and archive publics via search interfaces. We hope that from these initial case studies, and the tools and lessons emerging from our work, colleagues will have models for creating, sharing, and exploring text on maps data for their own collections.
References
Coleman, C. N. “Art, Cartography, and Machines: AI Tool Design for Historians.” Imago Mundi 76, no. 2 (2024): 307–313. https://doi.org/10.1080/03085694.2024.2453339
Ducatteeuw, V., Danniau, F. & Verbruggen, C. “Mapping Ghent’s cultural heritage: a place-based approach with web GIS.” International Journal of Digital Humanities (2025). https://doi.org/10.1007/s42803-025-00099-4
Hosseini, K., Wilson, D. C. S., Beelen, K., and McDonough, K. “MapReader: A Computer Vision Pipeline for the Semantic Exploration of Maps at Scale.” In Proceedings of the 6th ACM SIGSPATIAL International Workshop on Geospatial Humanities. GeoHumanities ’22. New York, NY, USA: Association for Computing Machinery: 8–19 (2022). https://doi.org/10.1145/3557919.3565812
Larsen, K. L., Thornberry, E., & Vitale, V. “Teaching and Learning with Text on Maps: A David Rumsey Collection Case Study.” Imago Mundi 76, no. 2 (2024): 290–296. https://doi.org/10.1080/03085694.2024.2453335
Lin, Y. and Chiang, Y.-Y. “Hyper-Local Deformable Transformers for Text Spotting on Historical Maps.” In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD '24). Association for Computing Machinery, New York, NY, USA (2024): 5387–5397. https://doi.org/10.1145/3637528.3671589
Machines Reading Maps Collection, Stanford Digital Repository (2023) https://searchworks.stanford.edu/view/rc349kh8402
McDonough, K. “Maps as Data.” In Computational Humanities, ed. by J. M. Johnson, D. Mimno, & L. Tilton. University of Minnesota Press, 2024.
McDonough, K., Beelen, K., Wilson, D. C. S., & Wood, R. “Reading Maps at a Distance: Texts on Maps as New Historical Data.” Imago Mundi 76, no. 2 (2024): 296–307. https://doi.org/10.1080/03085694.2024.2453336
McDonough, K., and Vitale, V. “Introduction.” Imago Mundi 76 (2): 279–84. https://doi.org/10.1080/03085694.2025.2453329
Nelson, R. K. “Mapping Environmental Inequalities: Using MapReader to Uncover Mid-Century Industrial Burdens.” Imago Mundi 76, no. 2 (2024): 284–290. https://doi.org/10.1080/03085694.2024.2453333
Schoonman, J. and Baptist, V. Open Maps Meeting, November 5-6 2024, TU Delft, https://openmapsmeeting.nl/.
van Wissen, L., Kuruppath, M. and Petram, L. “Unlocking the Research Potential of Early Modern Dutch Maps.” European Journal of Geography 16, no. 1 (2025): s12-s17. https://doi.org/10.48088/ejg.si.spat.hum.l.wis.12.17
Vitale, V. “Searching Maps by Words: How Machine Learning Changes the Way We Explore Map Collections.” Journal of Cultural Analytics 8, no. 1 (Apr. 2023). https://doi.org/10.22148/001c.74293.
Wood, R., Hosseini, K., Westerling, K., Smith, A., Beelen, K., Wilson. D. C. S., & McDonough, K. “MapReader: Open software for the visual analysis of maps.” Journal of Open Source Software 9, no. 101 (2024): 6434. https://doi.org/10.21105/joss.06434