11-19, 14:00–14:25 (Pacific/Auckland), WG607
This talk presents applications of GeoAI using street-level imagery and spatial analytics: detecting and mapping traffic signs, prioritising urban heat mitigation for active travel, assessing liveable streets through social-spatial configurations, and comparing perceived versus measured built environments. Together, they demonstrate scalable, data-driven approaches to inform urban planning.
Seeing Through the Crowd’s Eyes: AI-Powered Urban Insights from Street Imagery
Introduction
Cities are dynamic systems shaped by the interplay between built form and human experience. Traditional approaches to urban informatics often rely on top-down spatial data, such as remote sensing and census statistics, which are limited in capturing how people perceive and interact with urban space. The emergence of human-centred GeoAI offers an alternative approach by integrating street-level imagery with spatial analytics to understand cities from the perspective of their inhabitants.
Street view imagery (SVI) from platforms such as Google Street View and Mapillary provides immersive, ground-level visual data that mirrors the pedestrian experience of the city. Combined with advances in computer vision, machine learning, and geographic information systems (GIS), these data enable fine-grained assessments of urban conditions, ranging from infrastructure quality and visual aesthetics to safety and comfort. This paper presents four implementation case studies that apply object detection, semantic segmentation, and perceptual modelling to address pressing issues in urban infrastructure, climate resilience, street livability, and urban form perception.
Background: Human-Centred GeoAI and Street-Level Imagery
Human-centred GeoAI refers to the integration of artificial intelligence techniques with spatial data to model, interpret, and support human experiences in the built environment. It draws from computer vision, geospatial science, and human geography to create tools and insights that are sensitive to both objective spatial metrics and subjective perceptions.
Street-level imagery is particularly well-suited for human-centred GeoAI. Unlike satellite or aerial imagery, SVI captures the visual field as experienced by pedestrians, enabling assessments of visual enclosure, greenery, safety cues, signage, and cleanliness. Combined with deep learning techniques, these images can be analysed to detect objects, classify urban scenes, or infer human perceptions at scale.
Methods
Across all four case studies, we employed deep learning techniques, including object detection and semantic segmentation, applied to SVI. Outputs were integrated with spatial data in GIS to produce interpretable urban metrics and policy-relevant insights.
Object Detection: In the first study, we used the DNN object detection framework to identify Stop and Give Way signs from Google Street View images. These bounding boxes were geo-located using photogrammetric triangulation and integrated into GIS for validation and spatial pattern analysis.
Semantic Segmentation for Environmental Features: For climate-sensitive urban assessment, we developed a convolutional neural network model to segment tree canopy, buildings, and sky from SVI. The percentage of each class was calculated for each street segment and analysed in conjunction with land surface temperature and social vulnerability indices.
Space Syntax and Visual Composition: In the third study, segmented streetscape compositions (sky, tree, and building percentages) were mapped and compared with space syntax metrics such as integration and connectivity. These spatial-social variables were analysed to determine the alignment between physical form and social interaction potential.
Semantic-Based Human-Labelled Perceptions: Human perception was measured using Mapillary SVI and the MIT Places Pulse 2.0 dataset through semantic segmentation performed with deep residual networks (ResNet50), pre-trained on the ADE20K dataset. The Places Pulse dataset includes over 100,000 SVIs across 56 cities, including Melbourne, rated by participants on six perceptual indicators: Beautiful, Wealthy, Livable, Safe, Boring, and Depressing.
To infer perceptual scores from SVIs, we trained six Radial Basis Function (RBF) kernel Support Vector Machine (SVM) models on the Places Pulse dataset. Semantic segmentation translated the visual content of each image into categorical features, which were then used as input for the SVM models. Five-fold cross-validation ensured robustness across varying parameter settings. These models were subsequently applied to the Mapillary SVI dataset in Melbourne, generating spatial surfaces of perceived urban quality.
Case Studies
Traffic Sign Detection and Spatial Registration
Local governments require accurate records of street signage for safety and regulatory compliance. Our object detection pipeline achieved 95.63% detection accuracy and 97.82% classification accuracy for Stop and Give Way signs in selected areas of Melbourne. The derived locations were mapped in GIS and compared against council asset layers. This open-source workflow enables cost-effective monitoring and maintenance of road infrastructure, and is transferable to other signage types and cities.
Heat-Resilient Streetscape Planning for Active Travel
We identified microclimatic disparities in Bendigo's active travel corridors using street-level segmentation of tree cover and sky openness. Combining these results with satellite LST and census-derived vulnerability indices revealed areas where thermally uncomfortable walking environments overlapped with high pedestrian exposure and low greening. This informed a prioritised plan for tree planting and shading infrastructure. The method provides a scalable, equity-focused tool for urban heat adaptation.
Assessing Streetscape and Social Integration Using Space Syntax
Focusing on Greater Bendigo, we examined whether visually well-designed streetscapes aligned with high social interaction potential. Integration and connectivity scores from space syntax were correlated with visual segmentation outputs. Results showed that areas with high space syntax scores did not always align with aesthetically rich environments. This suggests the need for integrated design and network planning to promote livability. Demographic overlays showed older populations gravitating to highly connected areas, highlighting the social implications of spatial configuration.
Perceived vs. Measured Urban Form
In metropolitan Melbourne, we compared the Places Pulse-based perceptual scores derived from ResNet + SVM models with objective spatial metrics for the 5D dimensions of urban design: Density, Diversity, Design, Distance to Transit, and Destination Accessibility. Our analysis confirmed that areas with moderate density, high walkability, and green space were perceived as livable and safe. However, in very high-density zones, perceptions shifted negatively, with terms like "Depressing" and "Boring" appearing more frequently. These mismatches point to the need for planners to consider not just what is built, but how it is experienced.
Discussion
These case studies collectively demonstrate the potential of human-centred GeoAI to bridge the gap between technical urban analytics and lived urban experience. By leveraging crowd-sourced imagery and advanced computer vision models, planners and policymakers can generate fine-scale, scalable insights into infrastructure, climate vulnerability, social inclusion, and perceptual quality.
The combination of image analysis and spatial reasoning opens new avenues for participatory planning and targeted intervention. For example, the ability to spatially map perceptions of safety or beauty enables a deeper understanding of place attachment and mental well-being. Similarly, detecting infrastructure assets and environmental features at scale supports more equitable service delivery.
Conclusion and Future Work
Human-centred GeoAI grounded in street-level imagery offers a promising path for more inclusive, responsive, and perceptive urban analytics. Future work will extend these methods to other cities across Australia and globally, refine perception models with more culturally specific training data, and develop interactive tools for real-time urban planning and public engagement.
By embedding human perception into spatial modelling and leveraging scalable, open-source tools, we move closer to cities that are not only functionally efficient but also emotionally resonant and equitable for all.
Associate Professor Qian (Chayn) Sun is a geospatial scientist specializing in urban informatics, GeoAI, and environmental health. She leads the GISail research group at RMIT University, focusing on developing advanced geospatial solutions to improve urban resilience and sustainability. Her research integrates spatial analytics, remote sensing, and machine learning to address complex challenges in urban planning, climate change mitigation, and public health. Chayn has led several key projects with AURIN, including the Integrated Heat Vulnerability Assessment Toolkit, and the development of a national database on Culturally and Linguistically Diverse (CALD) communities to analyze health and environmental inequalities, has published extensively on social vulnerability and heat-health risk in Australian cities. She is also dedicated to mentoring postgraduate students and fostering interdisciplinary collaboration.