FOSS4G 2023
Opening session with institutional greetings.
In this keynote, we will explore the significance of seeding in the context of open-source software. Using QField as an example, we will explore the steps needed to turn a student's project into the leading fieldwork app that helps hundreds of thousands of people with their work and can help address many of the Sustainable Development Goals.
We will discuss the challenges faced during the initial stages of development and what steps played a crucial role in overcoming them. We will also highlight the importance of community and industry involvement and how these helped QField reach global success and over 800K downloads.
Through this keynote, attendees will gain insights into the role of seeding and commitment in developing and growing open-source software, highlighting its impact on innovation, collaboration, and sustainability.
Join us for an insightful discussion on planting seeds and the potential to drive positive change through open-source software.
DigiAgriApp is a client-server application to manage different kinds of data related to farming fields. It is able to store information about crops (specie, farming forms/system...), any kind of sensor data (included sensors and device hardware, weather, soils...), irrigation information (system type, openings...), field operations (pruning, mowing, treatments...), remote sensing data (taken from different devices as mobiles, drone, satellites) and production quantities.
The DigiAgriApp server is composed of a PostgreSQL/PostGIS database and a REST API service to interface with it. The server is developed using Django and the Django REST framework extension with other minor extensions are used to create the REST API. This service plays the key interface between the database and the client. We choose a nested way to create the API, of which the main element is the farm; this way the user can see only the farms related to him and from there he can look to other nested elements, first of all the farm’s fields and later other elements like sensor and remote data or other sub-fields like rows and plants. The REST API is using JavaScript Object Notation as input and output format to simplify and standardize the communication with it.
To obtain data from the sensors the server is also composed of a growing number of services to work with data providers, of which currently only a few are implemented. The Message Queue Telemetry Transport provider is a demon listening continuously to a broker (backend system to coordinate different clients) and several topics to obtain data as soon as they are provided; the second provided that is already implemented is related to remote sensing data and uses the SpatioTemporal Asset Catalogs specification to obtain the data. STAC is a common language to describe geospatial information, so it can more easily be worked with, indexed and discovered.
The client side instead is developed using Flutter, an open-source UI software development kit based on dart, a programming language designed for client development. Flutter is able to create cross-platform applications and it was chosen precisely because of its ability to realize cross platform applications.
All the code is released as Free and Open Source software with a GNU General Public License Version 3 license; it is available in the DigiAgriApp repository on GitLab and the client application will be published also in the main stores for mobile apps.
From detecting vegetation hazard to measuring catenary geometry, it all begins with three trains equipped with LiDaR mapping systems, roaming the french railway network.
Let us see how geospatial opensource softwares enabled us to build a cost-effective and comprehensive solution, starting from basic raw data processing up to setting up a geographic information system, full of relevant data that brings up railway maintenance to another level.
Over the recent years, Cloud Optimized Raster format have gain popularity not only because they ease access but also because the enable fast visualisation of the data. During this talk I'll go over the principles of dynamic tiling and talk about the different cloud optimized raster format. I'll also present the latest news about TiTiler.
Welcome to GeoServer, a popular web service for publishing your geospatial data using industry standards for vector, raster and mapping.
If the previous sentence made no sense to you, or if you are new to foss4g, or even just new to GeoServer, attend this talk to get pointed in the right direction!
This presentation provides a gentle introduction to FOSS4G and we will do our best to say the quiet part out loud:
- Demo: We have learned from experience, and will introduce GeoServer using a demo.
- Usage: Concepts using both a demo, and diagrams to connect to your data and publish as a spatial service.
- Checklist: Preflight check-lists capturing common oversights when deploying GeoServer for the first time.
- Value: What role GeoServer plays in your organization and what value the application provides.
- Community: How the project is managed, and a discussion of the upcoming activities.
Attend this presentation to get a running start on using GeoServer in your organization!
Attend this presentation to get a running start on using GeoServer in your organization!
Spatial data interoperability has been on the spot among the Open Geospatial Consortium members for almost 30 years, but the current moment is notable for several reasons. An enormous amount of data is growing exponentially due to the novel sensors that bring observations from previously inaccessible areas in such resolution. We can observe and explore the global ocean with modern computational resources and AI models. Federated Data Spaces initiatives emerge with the paradigm of multi-source data integration harmoniously supporting heterogeneous models.
Speakers will present recent advancements in the data mesh methods based on two environments endorsing open source implementations used for the integrations.
First is the Federated Marine SDI (FMSDI) Pilot, which focuses on advancing the implementation of open data standards, architecture, and prototypes for use with the creation, management, integration, dissemination, and onward use of marine and terrestrial data services for the Arctic. Use cases developed in the recent phase of the FMSDI pilot further demonstrated the capabilities and use of OGC, IHO and other community standards in response to a grounding event and the evacuation of a cruise ship or research vessel in the Arctic.
The approach collated with Iliad - Digital Twin of the Ocean and its interoperability patterns model. Based on the specific requirements for data transfer, access and computation, it looks to generalise core architectural patterns with standard implementations. These patterns address the core issues of data publishing, aggregation and extensive analyses close to the data. Together, they enable a viable overall digital twin ecosystem. Data mesh of observations with data lakes and assembly are essential building blocks that allow the flow and synchronisation of data between different data owners. A open, common information model, defined on the domain-specific and well-known generic ontologies, Analysis Ready Data, and Essential Variables concepts, allows for the traceability of provenance and various expressions. It is a critical prerequisite to achieving data interoperability and explainable AI. Application packaging of processing chains allows for seamless compute-to-data, remote computation, or even mobile control when data is too big to flow. The computation is executed in a controlled environment, and the results harmonised for further use or available as decision-ready information.
Presenters will describe these patterns and illustrate them with OGC and partners' open implementations (like OGC-NA, EDR, geoXACML, HubOcean sync API) from the projects.
QGIS turned 20 years old last year. The first lines of code were written in mid-February 2002 and when the programme was first compiled and run, it could do precisely one thing:
Connect to a PostGIS database and draw a vector layer.
Nowadays, QGIS is the go-to GIS solution for millions of users, and to make sure that QGIS's future is as bright as its past, we did a lot of work on communication, strategy and outreach.
In this talk, I’ll overview all the work done, the current status and the future of QGIS and its community.
Oskari is used world wide to provide web based map applications that are built on top of existing spatial data infrastructures. Oskari offers building blocks for creating and customizing your own geoportals and allows embedding maps to other sites that can be controlled with a simple API. In addition to showing data from spatial services, Oskari offers hooks for things like using your own search backend and fetching/presenting statistical data.
This presentation will go through the improvements to existing functionalities and new features introduced in Oskari during the last year including:
- Theme support
- UI rewrite progress
- Cloud compatibility improvements
You can try some of the functionalities Oskari offers out-of-the-box on our sample application: https://demo.oskari.org.
Link: https://oskari.org
The SpatioTemporal Asset Catalog (STAC) specifications are a flexible language for describing geospatial information across domains and for a variety of use cases. This talk will present the current state of the specifications, which includes the core STAC specification and the API specification built on top of OGC APIs. While the core specification has been stable for roughly two years and doesn't need a lot of updates, the API specification got numerous updates and is finally close to a stable release. This presentation digs into additions to STAC extensions and the latest community developments. We survey the updates to the open-source STAC ecosystem, which includes software written in Python, Node.js, and more. Finally, let's also look into the near future.
Editors of OpenStreetMap can use my software to search for a place or region, generating a list of candidate matches from Wikidata, which can then be checked and saved to OpenStreetMap.
Linking the two projects isn't without controversy. They use different licenses which raises questions about what information from one project can be copied to the other.
In the presentation I will give details of a new version of the editing tool.
I will talk about the benefits of linking, the process of finding matches, the community response - including the controversy - and how people can get involved.
This talk is going to reveal the secret of building and running development or
user environments as you always wanted. Each of your projects can run in
isolated, fully self contained environment, using the latest, or really old, or
heavily customized geospatial packages regardless of Linux distro or Mac version you
use. You can have as many environments as you want, and the environment will change as you
change between your projects, branches or commits.
No, we are not going to run containers, Flatpaks of Snaps for that. We are going
to enjoy the most advanced package manager Nix, the
largest collection of software in the world called Nix packages
(nixpkgs), unique tooling they provide and
Geonix Devenv projects built on top of that.
More than 250 years ago, Giovanni Battista Nolli, an Italian architect, engineer and cartographer, was concerned with how and where space is or is not publicly accessible. In his map 'La nuova topografia di Roma Comasco', he mapped publicly accessible interior and exterior spaces of Rome with an impressively high level of detail as a figure-ground map. Since Nolli’s time, both the character and diversity of public spaces as well as cartographic technology have changed. In my Master thesis, I aim to adapt Nolli's underlying idea for today’s circumstances on the basis of open data, and seek to develop methods for processing volunteered geographical information from OpenStreetMap (OSM) to identify, categorize, and map public spaces based on thematic and geometric information.
First, it has to be clarified what is considered public space and what is not. Among the defining aspects that appear most frequently in the examined definitions are accessibility, use and activities, democratic function, control/operation, ownership, and publicness. Given the data available via OSM as well as in terms of feasibility, I focus on the aspect of public accessibility and exclude indoor spaces. Most of the information needed for this is either available in OSM via tags (e.g. access, opening_hours), can be derived based on tags (e.g. from amenity, leisure), or can be inferred based on geometries (e.g. access to areas depending on surrounding barriers). Data processing is implemented as a Python script based on existing OSM and geospatial Python packages.
In case studies, the developed methods are applied to different parts of the city of Vienna, Austria. The identified public spaces are mapped, and in this step the quality and completeness of the available data in the study area are also evaluated. The resulting maps can provide information about the extent, type and potential of public space in the cases studied.
Upon completion of the project, the data processing methods will be published as open source in order to allow further collaborative development as well as adaptations for similar research and visualizations. With OSM as the database used, international applications are evident and could result in interesting comparisons and findings, though one most keep in mind, that the underlying definitions of public space are not equally applicable in non-Western cities and the extent and quality of the data can lead to limitations.
Based on the implementation of the Global Statistical and Geospatial Framework (GSGF) proposed by the UN and implemented in Latin America and the Caribbean by the Economic Commission for Latin America and the Caribbean (ECLAC), a set of specific technological components were developed, such as a geoportal, a statistical manager and an API with the possibility of consuming information from different applications. At the same time, components already existing in the community were implemented such as Kobo Toolbox, GeoNode, Airflow, MapLibre, Nominatim and Metabase for the integration of information from the collection in the territory to the publication of the data. The project was initially carried out with a group of countries: Argentina, Paraguay, Honduras, Guatemala, Dominican Republic, Costa Rica and Ecuador.
A presentation and demonstration of data cube functionality implemented based on OGC API Standards and draft Candidate Standards.
Including:
- OGC API - Tiles,
- OGC API - Maps,
- OGC API - Coverages,
- OGC API - Discrete Global Grid Systems,
- OGC API - Processes - Part 1: Core, and Part 3: Workflows and Chaining ("Nested Processes", "Collection Input", "Collection Output"),
- OGC Common Query Language (CQL2)
with a focus on providing efficient access to analysis-ready sentinel-2 data and additional processing close to the data, in the context of wildfire risk assessment.
In the last decade, 5 complementary assets have intersected, creating a series of new capabilities for our community. Modern geospatial did not exist even five years ago, and openness - the combination of open standards, open data all glued together with open source code is a key contributing factor.
This talk will present the case for openness being a competitive advantage for a modern, innovative technology company. We will discuss why we have been right all along, and why we will end up being even righter in the future. If you want a solid dose of confirmation bias, this is the talk for you!
This presentation will delve into the intricacies of packaging geospatial software for Debian Linux and its derivatives, including Ubuntu, OSGeoLive, and others.
It will begin by contrasting the differences between packaging for an operating system and application-level package managers. The presentation will then provide an introduction to the Debian GIS Team and their established practices for packaging, including resources for finding information. The focus will then shift to the crucial steps involved in preparing the software for distribution, such as creating metadata and dependencies, building the package, testing its functionality, and ultimately making it available to end-users for easy installation and use.
Protomaps is a simple, self-hostable system for tiled vector datasets. In the year since last FOSS4G, we've rolled out a new compressed specification (V3), added support for tile generation tools, and open sourced key integrations with content delivery networks. This talk will give an overview of:
- Why you might want to, or not want to, deploy Protomaps for your application
- PMTiles write support in the popular Tippecanoe and Planetiler tools
- The new open source integrations of Protomaps with AWS Lambda and Cloudflare
- Overview of real-world deployments for users in web GIS, journalism and the public sector
STAC is a well-known and acknowledged spatiotemporal metadata standard within the community. There are many applications with open-source data; however, there are few adoptions by premium satellite imagery providers. At UP42, we adopted STAC as the core metadata system within our applications and provided STAC API for users to manage their data easily. The ongoing adoption challenges with multiple data providers taught many takeaways that we would like to share with the community.
- UP42: a short introduction
- Data management challenges at UP42
- Solution with STAC & standard archive format
- STAC implementation: lessons learned
- Current state and way forward
This talk gives an overview of the current state of the GRASS GIS project for both users and developers. Latest version of GRASS includes even more tools parallelized using OpenMP to speed up massive data processing. The graphical user interface is changing as the single-window layout matured and is becoming the number one choice and a default setting. This adds to a quicker startup without a need for a welcome screen and streamlined data management. The code quality of C and C++ code improved significantly in the last year, the code compiles with strict compiler settings and we are heading towards pedantic compliance. Last but not least, this summer GRASS GIS celebrates its 40th birthday!
Suomi.fi-maps offers to the public administration and government agencies a centralized service for utilizing maps and location data. In the Suomi.fi-maps service, a user may compile their own map views from the map layers available in the service, as well as from their own objects and materials provided by service interfaces of their own organization.
Oskari platform is used to implement the Suomi.fi-maps system. Suomi.fi-maps is used to enable all the Finnish residents to use maps and the location data to find about the services they are interested in.
In addition to other open data the open materials of the National Land Survey may also be used: various terrain and background maps, property boundaries and aerial photographs. User may connect their own interfaces to the Suomi.fi-maps service or add their own objects to the map to be published.
This presentation describes with examples, how the Oskari platform and its features are used used to implement the Suomi.fi-maps service and lessons learned.
geoserverx is a modern Python package that provides an efficient and scalable way to interact with Geoserver REST APIs. It leverages the asynchronous capabilities of Python to offer a high-performance and reliable solution for managing Geoserver data and services.
With geoserverx, users can easily access and modify data in Geoserver, such as uploading and deleting shapefiles, publishing layers, creating workspaces, styles, etc. . The package supports asynchronous requests along with synchronous method to the Geoserver REST API, which enables users to perform multiple tasks simultaneously, improving performance and reducing wait times.
Apart from being implemented in Python Projects, geoserverx also provides CLI support for all of it's operations. Which makes it useful for people who want to avoid Python all-together.
In this talk we discover for the very first time about how geoserverx work and underlying code ideology. Along with that we'll also spread some light on upcoming modules to be integrated in geoserverx
Working in large open source projects, with several people contributing to the code, can be challenging, especially trying to keep everyone on the same page, and generating code that has enough similarities to allow shared maintenance.
The advent of platforms like GitHub also made it easier for one time contributors to donate small and large bits of code to the platform, generating in the process a fair amout of “review stress” in the project maintainers.
The presentation covers how pull request checks, formatting and static analysis tools have been used to streamline basic checks in the code:
- Testing the code on a variety of operating systems, Java versions and integrations with data sources before the code can be contributed to the project
- Enforcing common formatting
- Adding basic checks with CheckStyle
- Locating obvious errors, leftover code, basic optimization issues using the Java compiler linting, ErrorProne, PMD and SpotBugs
- Improving readability of the code as well as enforcing best practices and common approaches with the same tools.
- Effects on the dynamics of code reviews
The presentation will cover all those aspects, with examples from the author’s experience with the GeoTools, GeoWebCache and GeoServer projects.
Field data is often collected from various sources and in various formats using different gadgets and applications, making it prone to errors and inconsistencies. In this presentation, we discuss the challenges of cleaning field data and present best practices for ensuring data quality upload to OSM. We identify common sources of errors in field data, such as incorrect coordinates and incorrect attribute values, tags and discuss methods for detecting and correcting these errors. We will also discuss the importance of establishing data quality standards and procedures for ensuring that field data is accurate, complete, and consistent. Our findings provide valuable insights for OpenStreetMap Users and practitioners working with GIS field data. All tools used here are Open Source.
The Open Science Persistent Demonstrator (OSPD) is a long-term inter-agency initiative aiming to enable and communicate reproducible Earth Science across global communities of users and amplify inter-agency Earth Observation mission data, tools, and infrastructures. This talk will highlight the status and roadmap of the initiative (kicked off in 2023) and will provide an outlook on the first pilot activities of the demonstrator, as well as opportunities for participation for the FOSS4G community.
In the scope of this activity, ESA, NASA and OGC work together on the development of a long-term Open Science framework (e.g., a permanent open science demonstrator) in which participating organisations provide data, tools, and infrastructure in a coordinated approach, building on existing investments where appropriate.
In the frame of this activity, the OGC supports the Open-Source and Open Science Community by developing a persistent demonstrator that makes Open Science more tangible to a bigger audience, helps in exploring new forms of communication of scientific results to stakeholders, and helps develop the necessary standards to ensure the highest levels of interoperability across participating organizations. At the same time, it makes Earth Observation results available to other disciplines and communities, creates attention beyond the Earth Observation community, and directly impacts decision makers and political agendas.
The goal here is to demonstrate interoperable, collaborative research that allows reuse of existing components. These other resources are either offered as part of emerging Open Science Environments or in the form of either directly accessible “cloud-native” data/functions or by means of Web APIs. To reach this goal, it is essential to empower communities of practice to share FAIR (Findable, Accessible, Interoperable, Reusable) descriptions of their resources and capabilities. To allow this system to scale, it is crucial to avoid infinite combinations of community and application specific metadata, functions, data and products.
One focus is the facilitation of direct participation of the scientific community as the primary users of this framework, and of the open-source for geospatial community as essential contributors to the activity. To handle modelling complexity, OGC, NASA and ESA will define manageable processes and best practices for communities conducting geoscience research in multiple domains using heterogeneous data and tools on a distributed infrastructure. These agreements will include, but not limited to, standards, vocabularies, and ontologies for data and workflows and develop community-wide open source science mechanisms, modeling considerations and design patterns.
The Ontology discipline made its way into the Computer Science domain in the
1990s, filling a gap in the architecture aspect of a still infant engineering
domain. Its most visible impact happened around the industry consortium Object
Management Group (OMG), leading first to the Unified Modelling Language (UML)
and later to the Model Driven Architecture (MDA). MDA became the base
infrastructure of data architectures and exchange mechanisms specified by
institutions such as the Open Geo-spatial Consortium (OGC) or the European
Commission (through the INPIRE directive).
However, a parallel path has been treaded by the World Wide Web Consortium
(W3C). First with the specification of the Resource Description Framework (RDF),
a new paradigm for data encoding leveraged on the WWW, and later with the Web
Ontology Language (OWL), a pragmatic approach to ontology encoding, building on
RDF. This infrastructure developed by the W3C became known as the Semantic Web,
and also as Linked Data, for the innovative paradigm through which it connects
disparate data sources and data domains.
The OGC would eventually approach the semantic web, specifying GeoSPARQL in
2013, an ontology and query language for linked geo-spatial data. However,
technologies supporting this new standard were slow in materialising.
More recently, the specification by the OGC of a new set of data standards based
on the OpenAPI technology set out a clear path for the convergence of
geo-spatial data with the Semantic Web. New software is emerging, opening
an entirely new world to geo-spatial data provision, a clear step forwards in
practically, usability and semantics.
This address starts by reviewing the core concepts of the Semantic Web and
then reviews state-of-the-art software for the management, publication
and exploration of linked geo-spatial data. This addressed is targeted at SDI
professionals and data scientists wishing to upgrade the semantics of the data
they create and use.
A typical GeoServer deployment involves exposing it as a front service to publish a number of layers directly to the internet, where a single instance, or even a couple, and an on-premise deployment model is enough.
Within larger companies though, more often than not GeoServer is a critical component of a more significant infrastructure, used to host tens of thousands of layers to accommodate organization requirements across various departments and workflows that involve several other systems, and complex cloud deployments.
These scenarios are where GeoServer Cloud shine, enabling devOps teams to set up clusters of GeoServer pods that are scalable, have improved resiliency, security, and resource utilization; and increased observability and integration with telemetry systems for monitoring, debugging, and tracing.
In this talk, we'll explore in depth how GeoServer Cloud achieves these goals, from technology and design choices to detailed overviews of technical improvements that were required, supported by success stories of current CampToCamp customers that got GeoServer Cloud in production.
MapStore is an open source product developed for creating, saving and sharing in a simple and intuitive way maps, dashboards, charts and geostories directly online in your browser. MapStore is cross-browser and mobile ready, it allows users to:
- Search and load geospatial content served using widely used protocols (WMS, WFS, WMTS, TMS, CSW) and formats (GML, Shapefile, GeoJSON, KML/KMZ etc..)
- Manage maps (create, modify, share, delete, search), charts, dashboard and stories directly online
- Manage users, groups and their permissions over the various resources MapStore can manage
- Edit data online via WFS-T with advanced filtering capabilities
- Deeply customize the look&feel to follow strict corporate guidelines
- Manage different application contexts through an advanced wizard to have customized WebGIS MapStore viewers for different use cases (custom plugins set, map and theme)
You can use MapStore as a product to deploy simple geoportals by using the standard functionalities it provides but you can also use MapStore as a framework to develop sophisticated WebGIS portals by reusing and extending its core building blocks.
MapStore is built on top of React and Redux and its core does not explicitly depend on any mapping engine but it can support both OpenLayers, Leaflet and Cesium; additional mapping engines could be also supported (for example MapLibre GL) to avoid any tight dependency on a single engine.
The presentation will give the audience an extensive overview of the MapStore functionalities for the creation of mapping portals, covering both previous work as well work for the future releases. Eventually, a range of MapStore case studies will be presented to demonstrate what our clients (like City of Genova, City of Florence, Halliburton, Austrocontrol and more) and partners are achieving with it.
Oskari (https://www.oskari.org, https://github.com/oskariorg) provides a super-easy-to-use tool for creating mobile friendly maps that can be embedded onto websites or used as is. When embedding the maps on existing websites one can utilise the RPC API to further leverage the capabilities of Oskari. The API allows for integrating with existing services and external data sources so that the end result will be a seamless spatially enabled service running on any modern web browser.
While creating maps with Oskari requires no expertise in programming, utilising the RPC API requires basic knowledge of JavaScript. This talk will present the possibilities of Oskari RPC API among with some examples of live services created using it.
SMASH , the digital field mapping application for android and IOS that superseded the well known app geopaparazzi has been around for some years now. The last two years were a positive development storm after a quite calm year and brought many fixes as well as enhancements. Examples are better postgis and geopackage support, but also some hidden gems like geocaching.
The big news is on the serverside though. A new survey server has been developed in tight cooperation with a local government agency to best create effective surveying workflows and tools for survey teams. To attract a wider developer community to contribute to the project, the django framework was chosen for the server backend.
This presentation will give an overview of everything happened lately in the SMASH field mapping world.
We will give a status report on the GDAL software, focusing on recent developments and achievements in the 3.6 and 3.7 GDAL versions released during the last year, but also on the general health of the project.
The discussed topics will be as various as the scope of GDAL is, covering the new single CMake build system, the full open source write vector support for the Esri FileGeodatabase format, a Arrow-based columnar oriented read API for vector layers implement in the Arrow, (Geo)Parquet, GeoPackage and FlatGeoBuf drivers, new vector layer API for table relationsihp management, new raster drivers for the JPEG-XL, KTX2, BASISU, NSIDCbin formats, multi-threaded read capabilities in the GeoTIFF driver, multiple performance improvements in the GeoPackage driver, advanced API to read raster compressed data, a new vector driver for the General Transit Feed Specification (GTFS), support for the new Seek Optimized ZIP (SOZip) specification, etc.
The SpatioTemporal Asset Catalog (STAC) ecosystem for/in JavaScript has evolved in the last year. This talk will update you on the current state of the ecosystem and gives an outlook on what is missing. This talk will cover libraries such as stac-js, stac-layer, stac-browser, stac-node-validator, and more. We'll dive into what the libraries do, how they relate to each other and give you some hints how you get started.
Earth-Search is a publicly available SpatioTemporal Asset Catalog (STAC) API providing an index for some of the public datasets available through the AWS Registry of Open Data (RODA) and has been shown to be a valuable resource for accessing the Sentinel-2 archive as Cloud-Optimized GeoTIFFs. A new version of Earth-Search is an update and enhancement of the Sentinel-2 metadata as well as new Collections of data available on AWS, including Landsat Collection 2, NAIP, and Sentinel-1.
This talk will include a summary of the STAC catalog, what STAC extensions are used and how the data is best accessed based on file formats. We will also dive into the datasets that are available through the API and will present the architecture for indexing including a discussion of data latency. We will provide resources and tutorials for how to get started with public geospatial datasets on AWS.
GeoServer is a web service for publishing your geospatial data using industry standards for vector, raster and mapping. It powers a number of open source projects like GeoNode and geOrchestra and it is widely used throughout the world by organizations to manage and disseminate data at scale.
What can you do with GeoServer? This visual guide introduces some of the best features of GeoServer, to help you publish geospatial data and make it look great!
GeoServer has grown into an amazing, capable and diverse program - attend this presentation for:
- A whirl-wind tour of GeoServer and everything it can do today.
- A visual guide to some of the best features of GeoServer.
- Our favorite tricks we are proud of!
New to GeoServer - attend this talk and prioritize what you want to look into first. Expert users - attend this talk and see what tricks and optimizations you have been missing out on.
OpenLayers makes it easy to put a dynamic map in any web page. In this talk we'll discover how Openlayers works, what is the methodology behind adding new data to it. We'll understand how to spin up new openlayers map under 5 minutes using React.
Openlayers allows us to put data on map, interact with it depending upon whether you use it on computer or mobile, export data with map, etc. We'll explore examples few examples from https://openlayers.org/en/latest/examples/ .
Apart from this we'll also take a look at https://viglino.github.io/ol-ext/ this is extension to the openlayers using which we can add more types of interactions, visualisations, controls, etc.
When publishing (raster and vector) data in the form of a web mapping application, the first step is always to prepare a cache of the data. Currently, tiled images seem to be the industry standard - and the internal format of the tiles is either PBF (for vector data) or PNG/JPEG/WebP or similar raster data formats supported by current web browsers and desktop mapping applications (e.g. QGIS).
Most of the tools out there are going to store the raster tiles in a file-system structure, using directories for the Z and X tile coordinates and file names for the Y coordinate. This is limiting for practical purposes as on some filesystems you can exceed the maximum number of files easily. While for the vector data, the OpenMapTiles project seems to be well established, along with Tippecanoe and Planetiler, for the raster data tiles, the field of tiling possibilities is wide open.
The tiling process can be very demanding on hardware resources and time-consuming. Having the possibility to parallel process the data or even use a cluster of machines for faster tiling could be crucial for some applications.
In this talk, we will give an overview of the current possibilities for tiling, focused (but not exclusively) on the raster data tiles. Gdal2tiles, QGIS tile generating tools, mapproxy-seed, mapcache_seed, and others. Each of the tools has its place in the geospatial data provider ecosystem, and so does MapTiler-Engine. With MapTiler-Engine, users can process large amounts of geospatial data and store them in various output tile formats. It supports many input data formats and adds modifications such as output color, resolution, and more. It also supports different tile matrix sets. MapTiler-Engine has a graphical user interface for easy usage, but it also has a command line interface, so you can make it part of a larger toolchain.
We will discuss the algorithms inside geowarp, a high-performance and very low-level JavaScript library for reprojection, resampling and cropping of data from GeoTIFFs and other rasters. This talk will be at the abstract algorithmic level and is suitable for everyone. Here are some of the various algorithms that we will discuss:
- proj-turbo: fit an unknown reprojection function to a simple affine transformation
- fast-min/fast-max: calculating the range of your raster data leveraging the theoretical limits of the data types
- near-vectorize: automatically determining the optimal resampling algorithm based on relative pixel size
- dufour-peyton-intersection: calculate the pixels of an arbitrary raster inside an arbitrary polygon
- various resampling techniques including nearest, bilinear, vectorization, and box-based statistical methods
Generalization is a crucial topic in the map production process, describing the derivation of a map of a smaller scale from another one. It combines maintaining essential features and removing less important ones to offer a readable map. Often, this complex topic is reduced to a selection of attributes, creating label geometries, and simplifying line and area geometries.
The presentation shares the knowledge of the cartographer's toolkit by introducing the whole set of available generalization operators and showing less-known approaches for creating better maps. The entire collection of operators consists of simplification, smoothing, aggregation, amalgamation, collapse, merging, refinement, exaggeration, enhancement, and displacement, which can be implemented by algorithms.
The goal is to go behind the standards of creating centroids for labelling and using a Douglas-Peucker Algorithm for line simplification. A showcase of polygon simplification and creating label geometries are shown, demonstrating how to implement the operators using PostGIS with OpenStreetMap data. Several existing and working solutions for simplifying geometries and labels are presented to showcase possibilities.
The need to integrate geospatial data into products and services has resulted in a proliferation of Free and Open Source web APIs which often do not adopt any standards, thus requiring more development time and a lack of interoperability between solutions. For instance a bounding box has been written in multiple ways, depending on whether developers use the coordinates of the four corners, only upper left and lower right, latitude or longitude first, or some other variation.
The good news is that the Open Geospatial Consortium, a neutral, consensus-based organization, has been developing open standards for geospatial information. These standards are developed as building blocks, which means they could be easily incorporated into existing applications in order to enable a piece of geospatial functionality. The location building blocks are freely available to anyone to download and use.
In this presentation, we describe the conceptual model for the existing building blocks, which uses semantic annotations to define the different components. We also describe a practical example of how a building block could be integrated into an application and provide some resources for developers who want to build applications with the location building blocks.
Welcome to GeoNetwork and FOSS4G! GeoNetwork is a leading open-source web catalog for keeping track of the spatial information.
This is an orientation session, so if you are new to foss4g we can help explain how everything fits together, and how the pieces of the puzzle form a whole. If you are migrating from ESRI environment this is a critical talk to attend as open source technology is often presented in isolation.
Jody is an experienced open source developer, digging how this technology works. Jonna is part of the QGIS community looking how to successfully use GeoNetwork.
This presentation shares our findings and experience with you, and touches on what makes GeoNetwork succeed:
- We look at what GeoNetwork is for, the business challenge it is faced with, and the amazing technical approach taken by the technology.
- We will demo the the publishing workflow to see what is required, and look at how harvesting can jump start your catalog contents
- We peek under the hood at how the editor works, and discover the central super-power of GeoNetwork
- Look at examples of how GeoNetwork has been extended by organizations to see what is possible with this technology
GeoNetwork is an established technology - recognized as an OSGeo project and member of the foss4g community for over a decade. We would love to welcome you to the conference and share what this project has to offer.
eoAPI is an open source project which aim to create a full Earth Observation API, combining STAC metadata API (stac-fastapi), a Raster dynamic tile service (TiTiler) and a Vector Tiles service (TiPg).
Using eoAPI AWS CDK template you're almost two command lines away of setting your own Earth Observation services.
MIERUNE is a geospatial tech company in Japan. We set FOSS4G as a foundation of us and continuously join the communities as an user, a developer or a contributor. Thesedays we have been committing our new service - MIERUNE BASE. MIERUNE BASE is focussing on easily serving and sharing datasets on a simple architecture based on serverless and FOSS4G. In this talk, we will introduce the architecture or techniques of MIERUNE BASE.
The quality of geospatial data is generally measured by its logical consistency, completeness,
positioning quality, semantic quality, temporal quality and genealogy [1]. In fact, concerning the
situation of geospatial data in Madagascar in the past, since 1992, the old orthophotos had been
attached to the national reference system which is the international 1924 with Laborde as a
projection. The first old orthophotos were achieved during the environmental program in 90s. In
other hand, the remain old orthophotos were produced with the mission as national securing land
tenure. However, the geometric accuracy and details of all the old orthophotos are different as well
as they do not cover the national territory. If they cover a large area for about 60 000 km2, some
users have noticed discrepancies of a few meters or even more than a dozen meters on certain
points, even though the field of application is land. In December 2019, a ministerial order was
developed to define the technical specifications of photogrammetric work in the country. In this
specification, according to Chapter 4, Section 14, the accuracy of the orthophoto / orthoimage is
estimated by the planimetric root mean square deviation (emqXY) calculated from the differences
between the ground coordinates and measured orthoimage coordinates of certain clearly identifiable
topographic features. For the orthophoto / orthoimages in urban areas, the emqXY must be better
than 1 m CE90 which is the circular error at the 90th percentile. For the rest of the territory other
than the urban area, it must be better than 3 m CE90 [2].
Therefore, not only is it crucial to be able to measure this quality, but also to control, to improve,
and finally to guarantee it [3]. The basic map in Madagascar is the topographic map at the scale of 1
: 100 000. However, the average age of these maps is 60 years. Consequently, the contained
information no longer meets the needs of most users. On the other hand, orthoimages produced later
seem to be much more accurate. To evaluate the accuracy of the 1 : 100 000 topographic map, we
first identified an orthophoto that could be used as a reference. Furthermore, we considered the
orthobase elaborated in 2014 from the SPOT5 image and the control result of the CASEF
(Agricultural Growth and Land Security) project orthoimage. The 2014 orthobase was produced
within the framework of our cooperation with the La Reunion (France) region, while the CASEF
orthoimage was developed for the purpose of land tenure security in Madagascar.
In order to conduct this study, we tried to answer the following series of questions : 1) what is the
most accurate orthoimage to serve as a reference; 2) what is the average value of the deviations of
the objects on the 1 : 100 000 topographic maps as well as those of these derived products (SCAN
100 and BD 100) compared to those of the reference orthophotos. 3) Finally, is there a set of
parameters to reposition the SCAN 100 / BD 100 on this orthoimage?
To achieve this study, several steps were taken including literature reviews, collection of a few
samples and observations of results from previous work. We also made researches on the reference
data from which the BD 100, the topographic maps at 1 : 100 000 and SCAN 100 will be evaluated.
From this comparison, we could see that the attachment to the national reference system of the
CASEF orthoimage is more accurate than that of the orthobase. In addition to that, coordinate
pointing of identifiable geographic objects on both datasets were made with statistical evaluation of
the differences. Related to tools that we are adopting, since that our budget has been limited in
terms of software license, so that we are using open source geospatial software to make our
organization better with QGIS during the evaluation process.
After evaluating four (04) sheets on the 1 : 100 000 map of Mahajanga, Antalaha, Manjakandriana
and Toamasina, we quantified the root mean square errors at 109.3 m, 108.6 m, 128.4 m and 51.9 m
respectively. The deviations are disparate, therefore there is no single set of parameters to reposition
the 1,100,000 topographic map. We concluded that the BD 100 should be left as it is, and that a newset of geographic databases should be developed at different scales, in particular the new version of
the BD 100.
Catasto-Open is an open-source set of tools for the Italian Cadastre that manages geospatial data in a user-friendly and efficient manner. The tool is designed to store, retrieve and manipulate cadastral data, including property boundaries, ownership information, and other relevant details. By leveraging GeoServer and MapStore technologies, it allows for the integration with existing GIS systems, making it a versatile and valuable resource for managing geospatial data in an OGC-compliant pipeline. The tool is accessible to a wide range of users, including government agencies, private companies, and individual property owners, also Catasto-Open can be easily customizable to meet the specific needs of different users.
Vector tiles are changing the way we create maps. Client-side rendering offers endless possibilities to the cartographer and has introduced new map design tools and techniques. Let’s explore an innovative approach to modern cartography based on simplicity and a comprehensive vector tiles schema.
Take a tour of vector tiles cartography basics and learn about the latest trends through a number of examples illustrated with the MapTiler maps. Get an overview of best practices and learn about simple open-source recipes, towards advanced combinations of fills, patterns, fonts, and symbols. Selected layer parameters and style expressions will be discussed in a visual way and explained with basic syntax that you can take away.
The OGC APIs are a fresh take at doing geo-spatial APIs, based on WEB API concepts and modern formats, including:
- Small core with basic functionality, extra functionality provided by extensions
- OpenAPI/RESTful based
- JSON first, while still allowing to provide data in other formats
- No mandate to publish schemas for data
- Improved support for data tiles (e.g., vector tiles)
- Specialized APIs in addition to general ones (e.g., DAPA vs OGC API - Processes)
- Full blown services, building blocks, and ease of extensibility
This presentation will provide an introduction to various OGC APIs and extensions, such as Features, Styles, Maps and Tiles, STAC and CQL2 filtering.
Some have reached a final release, some are in draft: we will discuss their trajectory towards official status, as well as how good the GeoServer implementation is tracking them, and show examples based on the GeoServer HTML representation of the various resources.
There are no Free (as in Beer) and Open Source Cloud Datastores. Let's have an opinionated look at some of the better alternatives to store and modify, private and public data for spatial applications.
Having build FOSS cloud interfaces 4 Geo since forever I decided to look at the current state of data stores.
We have pretty much figured out how to do serverless in the cloud. Data at rest though is a completely different beast. The going gets tough the closer you work to the metal. There is an overwhelming multitude of formats, models and standards to chose from. Should we consider relational, document, and/or [column orientated] data files?
With too many to discuss we put the spotlight on some exciting new players such as bit.io and geoparquet.
A recent Panorama (BBC) report asked; Is the cloud damaging the planet? Is it?
Is there anything we can do? We want to share some best practices in regards to building data store interfaces as well as running these services at scale, and in production.
Development of Tippecanoe, a widely-used open-source C++ tool for creating vector map tilesets, has moved to Felt, where it is a key component of the zero-configuration data ingestion pipeline that processes Felt’s public data library layers as well as uploads from external users.
Version 2 of Tippecanoe improves its automatic choice of zoom levels, and makes visual improvements to coordinate rounding, small polygons, and the distribution of points in low zoom levels. It now runs faster and uses less memory and disk space. There are new options to generate label points for polygons, to order features by attributes, and to use Visvalingam line simplification. Tippecanoe now accepts FlatGeobuf input as well as GeoJSON and CSV, and can generate output in PMTiles as well as MBTiles.
GeoMapFish is an open source WebGIS platform developed in close collaboration with a large user group. It targets a variety of uses in public administrations and private groups, including data publication, geomarketing and facility management. OpenLayers and an OGC architecture allow the use of different cartographic engines: MapServer, QGIS Server, GeoServer. Recently new features have been added such as vector tiles integration, from raw data to visualization. In order to get rid of the AngularJS dependency, a roadmap has been established for a migration to a web components architecture. K8S support is evolving with the implementation of the necessary tools for Azure environments. A highly integrated platform, a large number of features, fine grained security and a mature reporting engine are characteristics of the GeoMapFish solution. In this talk, we will present the key usages, the state of the migration process to web components and latest functional developments, including backend - frontend decoupling allowing to plug in multiple front-end WebGIS clients.
Get your preferred OSM dataset (ie. country) running in a local Geoserver instance with only 2 commands and avoid any dependence on an external provider.
Simple, fast, clean solution. Lowering the barrier to entry to geospatial technology use and development.
Docker-compose setup which assembles the necessary components to implement a Geoserver instance that publishes the OpenStreetMap (OSM) layers locally on a single host/machine (Postgis is required to store the OSM layers).
Instructions for this project are based on this repository OSM-Styles, but make a much simpler execution plan.
The steps and scripts are intended to run in the context of Linux, Mac and Windows environments.
Open source tools have played a significant role in enriching OpenStreetMap (OSM) with community mapping in Tanga, Tanzania. These tools have enabled local communities to actively participate in mapping their own areas, which has led to a more accurate and detailed representation of the community on OSM. The use of open source tools in community mapping has also allowed for increased collaboration and sharing of data between community members, as well as with other organizations and researchers.
One such open source tool that has been used in community mapping in Tanga is QGIS. This tool has been used to create detailed maps of the community, including roads, buildings, and other infrastructure. The use of QGIS has also allowed for data analysis, which has helped community members identify areas in need of improvement and target resources more effectively.
Another open source tool that has been used in community mapping in Tanga is OpenDataKit (ODK). ODK has been used to collect data in the field, such as information on the availability of healthcare facilities and services. This data has been used to create detailed maps of the community, which has helped community members identify areas in need of improvement and target resources more effectively.
The use of open source tools in community mapping in Tanga has also led to increased collaboration and sharing of data between community members, as well as with other organizations and researchers. For example, community members have been able to share their data with organizations working on healthcare and education projects, which has helped these organizations target resources more effectively.
Overall, the use of open source tools in community mapping in Tanga has been a significant factor in the success of OSM in the area. These tools have enabled local communities to actively participate in mapping their own areas, which has led to a more accurate and detailed representation of the community on OSM. The use of open source tools in community mapping has also allowed for increased collaboration and sharing of data between community members, as well as with other organizations and researchers, which has helped to improve the community and target resources more effectively.
MapMint is a comprehensive task manager for publishing web mapping applications. It is a robust open-source geospatial platform allowing users to organize, edit, process, and publish spatial data online. MapMint includes a complete administration tool for MapServer and simple user interfaces to create mapfiles visually. MapMint is based on the extensive use of OGC standards and automates WMS, WFS, WMT-S, and
WPS. Most MapMint core functions are run through WPS requests called general or geospatial web services: vector and raster operations, mapfiles creation, spatial analysis, queries, and more. MapMint server-side is built on top of ZOO-Project, MapServer, and GDAL, and its numerous WPS services are written in C, Python, and JavaScript. MapMint client-side is based on OpenLayers and Jquery provide user-friendly tools to create, publish and view maps. MapMint architecture and main features will be introduced in this presentation. Its modules (dashboard, distiller, manager, and publisher) will be described with an emphasis on the OGC standards and OSGeo software they use. Some short but relevant case studies and examples will finally illustrate key MapMint functionalities.
Kontur is geospatial company providing disaster management solutions, such as Disaster Ninja - an emergency mapping tool for Humanitarian OpenStreetMap Team. Disaster Ninja is one of the applications based on Kontur Platform.
The platform modules include front-end & back-end supporting multiple apps, ui-kit, ETL for geospatial datasets, Kubernetes setup.
Last year we’ve started to open source our products to meet the global goal - increasing the availability and quality of open data and providing tools to work with them.
I want to present the following:
- Open source modules our platform has to offer
- How to use those modules together
- What use cases were solved and can be solved
Redmine Geo-Task-Tracker (GTT) Plugin provides geospatial support for Redmine. Redmine is a well-known OSS issue management system. GTT Plugin enables to attach geospatial information(Point, Polyline and Polygon) to each issues. It is effective in management many issues based on geospatial infromation(ex. Road and park management). This talk introduces features and some use cases.
The GeoNetwork-opensource project is a catalog application facilitating the discovery of resources within any local, regional, national or global "Spatial Data Infrastructure" (SDI). GeoNetwork is an established technology - recognized as an OSGeo Project and a member of the foss4g community for over a decade.
The GeoNetwork team would love to share what we have been up to in 2023!
The GeoNetwork team is excited to talk about the different projects that have contributed with the new features added to the software during the last twelve months. Our rich ecosystem of schema plugins continues to improve; with national teams pouring fixes, improvements and new features into the core application.
We will also talk about the UI revamp through the geonetwork-ui framework, and the new perspectives it could bring to your catalogs. Progress of our main branches (4.2.x), and release schedule.
Attend this presentation for the latest from the GeoNetwork community and this vibrant technology platform.
The delivery of national census programs to aid nations in coming up with better strategies for serving the population’s needs and better plans for sustainability. While on the other hand, several developing nations around the world have not been able to deliver highly accurate census data and results to aid in these efforts. This leads to the implementation of policies that are not inclusive among other limitations introduced along the way.
By leveraging on open data platforms such as OpenStreetMap, open-source geo applications can be built to aid developing nations in accurate and location-driven data-capturing processes. Having digital location strategies and innovation as the major component for census data collection can potentially lead to vast growth in digital economies across developing nations and unleash endless possibilities and potential innovations which are inclusive and fit for purpose. This also provides a platform and chance to have more contributions towards OSM at the national level while delivering accurate and much-needed data.
UNVT Portable is a package for RaspberryPi that allows users to access a map hosting server via a web browser within a local network, primarily for offline use during disasters. It is designed to aid disaster response by combining aerial drone imagery with OpenStreetMap and open data tile datasets.
"UNVT Portable" is a map server that allows you to freely use web maps from devices such as smartphones even in an offline environment. It is mainly designed to work in an offline environment in the event of a major disaster, and various open data tiles are prepared in advance, such as drone aerial images taken after a disaster, OpenStreetMap, and satellite images released for free by JAXA(Japan Aerospace Exploration Agency), etc. Combine sets to create the maps you need in times of disaster. We envision a use case for municipalities, etc. to understand the situation after a disaster and to respond to disasters. It is built using open source software such as Apache and MapLibre and Raspberry Pi, and is completely open source. Unlike tools such as Google Maps, which are difficult to use for secondary purposes, it is being developed as open source so that it can be released in a form that can be easily used by anyone, including local governments, international organisations and private companies.
The Web Map Service (WMS) is the most popular standard of sharing data remotely. It is commonly used as a basemaps, a way of presenting governmental spatial data, and as a data source when creating vector datasets. Creating a WMS requires original data to be read and then rendered. This process can be slow, especially if the source data is heavy and not optimized. This is the case, for example, with Sentinel 1 global satellite data, which is a collection of daily revisions with a total volume of 250 GB per one day. Here we demonstrate an efficient way to share such a very large data set as WMS using Mapserver scaled with Kubernetes.
Mapserver is used as engine of our WMS, because of it speed and ease of automation. In order to optimise the performance of the service and therefore the user experience, it is recommended to store the data in the right format, with the right file structure also being aware of limitations of storage, bucket or disk read speed. GDAL provides a set of options that can be executed in a single command to overwrite the original data with new, cloud optimized. It is usually good practice to store selected zoom levels as a cache, but for time series data that is enriched daily, the cache is not overwritten as new data arrives, but is incremented.
Despite its popularity and advantages, WMS as a standard of serving data has its limitations. The potentially large disk read time is multiplied by the number of users sending requests. Tests using JMeter (100 users sending 100 GetMap requests in a loop) have shown that on a relatively strong processor (32CPU), the greatly increased traffic acts as a distributed denial-of-service (DDoS) - the server stops responding.
This problem is solved using Kubernetes (K8s) which allows metric-based automatic horizontal scaling of containerised applications, in this case – Mapserver. Prometheus as a K8s cluster monitoring tool allows custom metrics to be defined e.g., number of http requests per time interval. Prometheus makes it possible to distribute the traffic between newly created pods so that all requests can be answered.
The aim of the talk is to stimulate discussion, confront the idea with experts and demonstrate good practice in creating a publicly accessible WMS, with a focus on optimising speed under heavy source data conditions, supported by a working example and statistics.
Surface runoff is one of the processes with direct impact on water erosion. Surface runoff has two basic components: a) sheet runoff and b) rill runoff. Observation of these phenomena at various scales and then using mathematical models to describe their observations plays a key role for soil protection. One of the models developed to compute these phenomena is SMODERP, used for example in the flexible and adaptive approach to land management and landscape planning called Model of Living Landscape project. Innovative application of the SMODERP model (https://github.com/storm-fsv-cvut/smoderp2d) named SMODERP Line is presented in this contribution. SMODERP Line is accessible through various interfaces including OGC Web Processing Service (WPS) which can be easily integrated into user-defined processing pipelines or web applications. Usage of SMODERP2D Line will be demonstrated in the QGIS environment through a new OWSLib-based QGIS WPS Client Plugin (https://github.com/OpenGeoLabs/qgis-wps-plugin).
This contribution was supported by grant RAGO - Living landscape (SFZP 085320/2022) and Using remote sensing to assess negative impacts of rainstorms (TAČR - SS01020366).
The National Land Survey of Finland (NLS) is a government agency that maintains finnish property register and uses various administrative information systems that handle crucial data. To develop, manage, and maintain these systems, NLS follows a Business Technology Standard model and aims to publish its own production applications as open source software and use open source applications in development when possible.
During the development of new information systems, NLS follows an agreed and approved management model and uses only components and software that meet development guidelines. Examples of such components are QGIS and PostgreSQL. However, if NLS needs to adopt and evaluate components that are not yet included in the development guidelines, it must evaluate associated open source projects, record and process considerations, and accept them in accordance with the change management process.
To evaluate the maturity of open source projects, NLS has developed a tool that continuously evolves to reflect the needs of the organization. The tool is a checklist of criteria that can be used to assess the maturity of a project and compare it to similar products. The presentation explains the items in the tool and their significance as part of the metrics.
The tool that NLS has developed could be valuable for individuals and companies in similar positions when evaluating open source projects for their needs. The experiences gained by NLS can also help improve weak points that open source software producers may not have considered in their own projects.
IDeAMapSudan is a 2.5-year project finishing in March 2023. The project aims to develop a community-led geospatial database for mapping deprived urban areas (e.g., informal settlements) that will support the decision-making process for displacement and socio-economic reconstruction in Khartoum, Sudan. To that end, nine trainers from different governmental and non-governmental organizations were selected to be trained by a team of international experts from the Faculty ITC of the University of Twente, The Netherlands; the Universite Libre de Bruxelles, Belgium; and from the African Population and Health Research Center Kenya. These nine trainers were taught the essential competencies in using Free, and Open Source Geospatial Software to produce, compile, curate and distribute spatial data. Once the training of the nine trainers was completed, a series of community workshops were organized so that the trainers could train local community actors in tasks related to spatial data curation in close relation to their communities. The datasets produced from this process were then used to create a deprivation model and additional open data sets that can be used to help local communities and actors to take actions to mitigate several types of deprivations:
Unplanned urbanization - e.g. small, high-density, disorganized buildings
Social risk - e.g. no social safety net, crime
Environmental risk - e.g. flood zone, slopes
Lack of facilities - e.g. schools, health facilities
Lack of infrastructure - e.g. roads, bus service
Contamination - e.g. open sewer, trash piles
Land use/rights - e.g. non-residential zoning
This talk will describe three significant aspects of the project: the curriculum of competencies and the software tools used to teach these competencies; the phases and challenges of assembling a team and infusing it with a sense of community and participation; and the importance of disseminating results and evaluate the social impact open source software and open data can have.
MapComonents is an open-source framework extending React for mapping applications. It can be used to develop browser-based applications that do not require any backend, as well as web clients that use an arbitrary number of backend services. MapComponents uses MapLibre for rendering, raster, and vector tiles.
It provides working defaults wherever possible enabling the usage with minimal parameters. At the same time, it exposes the entire MapLibre API allowing very granular control of the result where it is needed. Solutions for more complex and common requirements such as PDF export, a feature editor, layer tree, WMS loader, measure tools, or bookmarks are provided as ready-to-use and highly configurable drop-in components. Exotic requirements include the swipe tool, the magnifying glass that partially shows two synchronized MapLibre instances or components that render 3D meshes or deck.gl.
Layers on the map are covered by several components and example codes in our lab repository. It can be combined with a backend for managing a more extensive data set. In addition, it also works as a progressive web app offline with most functions. Creating dashboards and complex user interfaces that combine maps and diagrams MapComponents is more straightforward than traditional approaches, given the declarative nature of React and its vast ecosystem of existing components.
The presentation will show and explain an actual example and its function. MapComponents framework is available under the MIT license and developed by WhereGroup GmbH.
MapTiler SDK is a TypeScript layer that adds new capabilities on top of MapLibre GL, both in terms of UI and core features. It also comes with an interface to MapTiler Cloud REST API.
The features we have added on top of MapLibre are of two kinds: many convenient helpers to make the developers' life easier, and plenty of built-in defaults that are specially made to use MapTiler data without having to specify annoying URLs or {ZXY} patterns, yet keeping it 100% backward compatible with MapLibre. In addition, all our services from MapTiler Cloud API, such as geocoding, IP geolocation, coordinate transforms, or static maps generation, are now easily accessible with well-documented TypeScript functions. All this with an open-source license.
In the talk, we are going to present the library, showing practical examples and outputs. We believe, that the SDK is going to make the life of the web mapper easier not only by providing a close integration of MapTiler services but also by the new components and library itself.
The demo will feature some nice weather visualization we’ve been working on lately!
Landfill sites are for storing waste in a secure and secluded manner but they can cause a lot of damage to the environment by generating greenhouse gases and contaminating soils by releasing heavy metals and toxins. Monitoring the area of landfill sites from space is a challenging problem because of the huge amount of unstructured data and unavailability of standard datasets and procedures. By combining open-source tools with geospatial data, we present a global dataset that monitors the changes in the landfill area. We have achieved this by developing a deep learning based segmentation model that uses multispectral satellite data and segments the landfill areas from them. In order to develop the model, we have labelled landfill sites from optical imagery from all over the world. Our current segmentation model has 31 million parameters and has achieved an accuracy of 77.6% on the test set. Currently, the dataset contains temporal data from 2021 of the major landfill sites from more than 7 countries and it is growing daily as new data is coming in. In future, we aim to enhance this dataset by adding more variables other than the area, for instance height of the landfill and will also explore other higher resolution data for validating our results further.
Processing satellite data is a challenging task that requires high computational power and storage, leading to substantial time and cost investments. This can often pose a significant obstacle for organizations, especially those with limited resources. In addition, the processing of large amounts of data can be complex, requiring a flexible infrastructure to effectively manage it. These challenges can make processing satellite data an overwhelming task, requiring significant financial and technical resources. Some organizations have lost all their funding in the AWS bills.
This is where serverless computing options such as AWS Lambda offer an efficient and cost-effective way to process satellite data. With Lambda, organizations can avoid the need to maintain a server full-time, which can significantly reduce costs. Additionally, AWS Lambda functions can be invoked through event-based triggers, such as file creation or update in S3, or scheduled using cron jobs. This approach allows organizations to process data on an as-needed basis, further reducing costs.
Let’s talk about a generic framework to leverage AWS Lambda functions, Step Functions for orchestration, and S3 storage for data storage. With AWS Lambda, organizations can process large amounts of data in parallel, with each Lambda function offering up to 10GB of memory and 6vCPU. Lambda supports up to 1000 concurrent executions, allowing organizations to efficiently process huge amounts of data. With AWS Step Functions, organizations can easily orchestrate multiple Lambda functions to create complex workflows for satellite products. Finally, S3 storage provides a scalable and secure way to store data at a lower cost, with S3 multi-tier storage options like Glacier allowing for analysis over a large period of time.
We’ll cover the basics of AWS Lambda, Step functions, and S3 using gdal functions to process a diverse set of satellite products like sentinel 2 and Landsat where we can clip the rasters, create COGS, calculate values like NDVI, EVI, etc
One decade ago, we saw the launch of the first earth observation cubesats by Planet Labs. In the years since we have seen hundreds of satellites launched, and dozens of startup companies launching taskable satellites. While this has led to incredible opportunities to leverage multiple sensors and sensor modalities, the massive increase of data has also created challenges in data management, discovery, and usage. The community driven SpatioTemporal Asset Catalog (STAC) specification was an important step forward in exposing data to users in a standard way that enables cloud-native workflows and has been successful across government and industry.
The process of actually tasking satellites, however, is still very much non-standard; each data provider exposes a unique API, if at all. Some data aggregators have created a single tasking API that proxies and translates to multiple data provider APIs, but this is still non-standard, and proprietary.
Element 84 has been leading an effort to create a community standard API around how users order future data and how providers respond to those requests. Working with government groups, commercial satellite operators, and data integrators, we have hosted working sprints to develop a specification and open-source tooling demonstrating the power of a tasking API specification.
This talk will cover the current status of the community tasking API specification, future plans, and a demonstration of how to use the API to order data.
GeoServer is a web service for publishing your geospatial data using industry standards for vector, raster and mapping. Choose additional extensions to process data (either in batch or on the fly) and catalog records.
GeoServer is widely used by organizations throughout the world to manage, disseminate and analyze data at scale. GeoServer web services power a number of open source projects like GeoNode and geOrchestra.
This presentation provides an update on our community as well as reviews of the new and noteworthy features for the latest releases. In particular, we will showcase new features landed in 2.22 and 2.23, as well as a preview of what we have in store for 2.24 (to be released in September 2023).
Attend this talk for a cheerful update on what is happening with this popular OSGeo project, whether you are an expert user, a developer, or simply curious what GeoServer can do for you.
Valhalla proved, since its inception in 2015, to be a valuable part of the OSM software universe, occupying an important niche in the routing section. It's arguably one of the most feature-rich open-source routing engines, serving many different use cases and integrations/deployments.
However it's a fairly complex system which is hard to comprehensibly document and new users or developers are often overwhelmed. So, I'd like to introduce its general architecture, capabilities and showcase "new" features (the last talk was given in 2016 on FOSS4G NA), as well as the accompanying open-source software, like various libraries, clients and docker image(s).
Chemical incidents, such as accidents at heavy chemical plants or large-scale toxic gas leaks, are difficult to assess accurately because of the large spatial extent of the damage and the rapidly changing scope/level/target of the damage over time. These characteristics also make it hard to conduct experiments to recreate or simulate large-scale chemical incidents in real world. In the case of large-scale chemical accidents or release, post-incident damage assessment is as important as prevention, but spatial ambiguity makes it difficult to assess the extent of damage to victims, and there is little way to identify fake victims from real ones.
In this 5 year-long study, we aim to combine the results of a chemical diffusion model and the location data of mobile service subscribers on the incident spot over time. For this, FOSS4G based 3D geospatial web service using GeoServer, Postgresql/PostGIS, Cesium, etc. will be developed to assess the level of chemical exposure of each victim and calculate the level of damage based on it.
In 2022, the first year of the study, we developed a prototype that combines the time-dependent output of the chemical diffusion model with the time-dependent location data of individuals and successfully visualized it in a Web 3D globe. In the coming year, we'll further develop this system into an integrated risk assessment platform for chemical accidents by combining chemical exposure assessment model and damage calculation model.
These days we have an incredible amount of (open-source) geo-spatial data, remote sensing data and insights, plus the tools to share them with the world! But when building a web map application or dashboard we often end up with too cluttered visualizations, confusing jargon, scary technology or struggle in communicating with the geo-data illiterate. GIS technology can be hard to understand.
How do we design and build a map application showing a huge amount of geo-data accompanied by the elaborate functionality to discover it?
As GIS experts we think from a technological perspective, adding more and more buttons, layers, panels, pop-ups, legends, draw tools, scale-bars. But these GIS terms makes an application confusing, scary and technically hard to understand for the user..
On the other hand, UX and IX designers think about usability, smooth experiences and helping users to easily navigate, see, use and interpret an application. But they lack the understanding of specific map related design requirements and map related interactivity. Here, the map is taken for granted and is often not well designed..
I often find myself mediating between the GIS and cartographic professionals, web-developers, UX and IX designers and data-designers. I believe there is still a lot we can improve with each other!
So let’s bridge the gap and join the conversation about Interactive Cartography! In this talk I will give some clear useful examples. What is Interactive Cartography and what can we learn in this? Be amazed with some simple examples which can quickly improve your web map application!
Here comes a developer story about contributing to GeoDjango.
An unfortunate combination of a valid, but unconventional spatial reference on the one hand, and "smart" logic for a mixed-geometry dataset: Geometries supposed to be located in Austria ended up in the Near East.
Investigation showed that GeoDjango's behaviour for returning the SRID of the dataset was not according to its documentation (see Django ticket #34302). While fixing the issue, additionally, an incorrect type cast from None
to string was discovered.
In this talk you will also learn:
1. How to set up the GeoDjango test suite with a PostGIS docker container
2. How the Django code review process looks like
Ground-based weather sensor networks are essential in monitoring local weather patterns and climate. Integration of such data into GIS environments is critical to supporting manifold applications including urban planning, public health studies, and weather forecasting.
These networks use scattered geolocalized sensors to measure multiple atmospheric variables (e.g. air temperature, wind speed, precipitations). Often, data is distributed online by network managers which can be either local/national authorities, private companies, or volunteers. Due to the diversity of data providers, both formats and access patterns of meteorological sensor data are heterogeneous and the preprocessing tasks (e.g. temporal aggregations, spatial filtering) are generally time-consuming.
Given the above and to increase end-users exploitation of such sensor data, we present the development of an experimental QGIS plugin facilitating access and preprocessing of openly available data from ground-based sensor networks and enabling their direct use in QGIS. The plugin is designed to implement REST APIs connections and HTTP requests to download data. A user interface allows for selecting time intervals and types of observation to be downloaded. Once data is retrieved, the plugin provides options for filtering, outliers removal, time aggregation with summary statistics as well as observation mapping into a standard GIS layer. These functionalities are only partially available in similar existing QGIS plugins. The plugin leverages FOSS Python libraries for data handling including Pandas. The Dask parallel computing library is also exploited to speed up I/O operations on raw data.
The current version of the plugin is developed to retrieve and process weather sensor data provided by the Environmental Protection Agency of Lombardy Region (ARPA Lombardia), Northern Italy. The data retrieval is based on the Sodapy Python library, a Python client for the Socrata Open Data API. The plugin's work-in-progress source code is available at (https://github.com/gisgeolab/ARPA_Weather_plugin) released under MIT license. The plugin is being developed within the LCZ-ODC project (agreement n. 2022-30-HH.0) funded by Italian Space Agency (ASI), which aims to identify Local Climate Zones within the Metropolitan City of Milan.
Ongoing work includes the extension of the plugin functionalities to incorporate additional data providers, starting from other Italian regional ARPAs. The goal of this project is to provide a reproducible framework to access and handle weather data into QGIS, thus extending the capability of the software to support a wider range of practitioners and applications.
OpenStreetMap is an open source data which any one can access it free. This data is contributed by the local communities or individuals voluntarily. For them to gather together, we use mapathons to bring them for the mapping. A lot of data is added during these mapathons to help vulnerable people around the Globe.
Various disciplines such as traffic simulations, driving simulations and applications in autonomous driving require highly detailed road network datasets. OpenDRIVE evolved as an open industry standard for modelling of lane-level road networks (HD maps). Acquiring such datasets is very expensive tough because it has to be done through mobile mapping in most cases. We want to introduce to the FOSS4G community two recently and openly published road network datasets from Brunswick (https://doi.org/10.5281/zenodo.7071846) and Wolfsburg (https://doi.org/10.5281/zenodo.7072631). Investment in both datasets has been funded by German authorities and covered more than 100.000 Euro. We will also give a short appetiser on how to use this data with free and open GIS tools.
The public administration of the Swiss canton Aargau chose to use OSS for the publication of all open WMS, using GeoServer-Cloud and PostgreSQL. Meanwhile, the decentral offices, which gather geographical data and style this data are used to using proprietary software for this purpose. The strategy chosen was to provide a soft transition to OSS, by providing automated conversion processes based on a new FOSS project and by improving existing OSS with regards to styling conversions towards SLD.
Securing a modern API in an effective way is critical to prevent unauthorized access and ensure the privacy and integrity of data. In general, there are three common mechanisms that can be used for API security: API keys, OAuth2/OpenID Connect, and JSON Web Tokens (JWT). Each of these mechanisms provides a different level of security and flexibility, depending on the requirements of the API. Modern OGC APIs are agnostic and rely completely on the adoption of OpenAPI security schemes so the implementers can use the mechanism that perfectly fits with their requirements.
fastgeoapi is a new open-source tool designed to be an authentication and authorization layer on top of a vanilla pygeoapi that offers out-of-the-box a secured infrastructure easily pluggable and configurable through the a standard OpenID Connect protocol.
This talk aims to describe the recipe to configure and protect a vanilla pygeoapi with Keycloak and Open Policy Agent in order to publish secured OGC APIs in a standard manner.
This talk will describe the usage of Jason-3 Altimeter data, which records the topographic height of the surface of the earth every ~10 days, to help measure the changes in water level of reservoirs across the globe. The use of NASA Common Metadata Repository (CMR) API to download and subset is described along with navigating the maze of various Jason-3 Level-2 Products depending on the use-case.
This talk introduces to this open dataset and various other altimetry missions, to allow for multi-mission monitoring of reservoirs of the world. It further uses Free and Open Source Software (CMR Specification, Xarray) to pre-process the data for use.
In the field of disaster management, communication and collaboration are key components for successful response and recovery efforts.
Next-Generation Incident Command System (NICS) is an open-source web-based command and control environment designed for small to extreme-scale incidents. It facilitates collaboration across state and local/municipal levels of preparedness, planning, response and recovery for all-risk/all-hazard events.
Although the active development stage of NICS was in 2000s - early 2010s, MIT, its initial developer, is no longer supporting the project.
Kontur Platform, another open-source product for disaster management, has the potential to replace NICS modules in accordance with modern technology standards.
In this talk I'll present:
- Research into the current NICS state of technology stack, code, features and installation process
- Live demo of NICS
- Our ability to maintain the current version of NICS
- Vision for NICS 2.0 based on Kontur Platform.
With these efforts, NICS can continue to provide a critical tool for disaster incident management in its current and upgraded states
This presentation will introduce the attendees to those which are GeoNode's current capabilities and to some practical use cases of particular interest in order to also highlight the possibility of customization and integration. We will provide a summary of new features added to GeoNode in the last release together with a glimpse of what we have planned for next year and beyond, straight from the core developers.
We will discuss the state of GeoRasterLayer, a JavaScript library that renders GeoTIFFs directly on a LeafletJS map without a server. This will include an introduction of new features, including the following:
- shifting warping off the main thread to a pool of web workers
- improved support for extent calculations by increasing vertex density of polygon representations of bounding boxes
- high-resolution support by using geowarp
We will also look to the future and discuss the following:
- support for raster types other than GeoTIFF/COG
- geozarr support
- similar integrations into other web mapping libraries
Audience feedback and ideas will be most welcome!
Do you get regular data-drops from suppliers, and struggle with viewing changes between releases and keeping everything synchronised? In this talk we'll demonstrate how the great new geodata versioning tool Kart is being used to solve this problem with the official New Zealand cadastre datasets.
—
We’re drowning in data, but the geospatial world lags badly behind in versioning tools compared to our software counterparts. Kart (https://kartproject.org) is solving this with a practical open tool for versioning datasets, enabling you to work more efficiently and collaborate better.
Kart allows you to quickly and easily manage history, branches, data schemas, and synchronisation for large & small datasets between different working copy formats, operating systems, and software ecosystems.
Modern version control unlocks efficient collaboration, both within teams and across organisations meaning everyone stays on the same page, you can review and trace changes easily: ultimately using your time more efficiently.
Motivation:
Spatial Data Infrastructures (SDI) developed for the exchange of environmental
has heretofore been greatly shaped by the standards issued by the Open
Geospatial Consortium (OGC). Based on the Simple Object Access Protocol (SOAP),
services like WMS, WFS, WCS, CSW became digital staples for researchers and
administrative bodies alike.
In 2017 the Spatial Data on the Web Working Group (SDWWG) questioned the overall
approach of the OGC, based on the ageing SOAP technology
[@SDWWG2017]. The main issues identified by the SDWWG can be summarised as:
- Spatial resources are not identified with URIs.
- Modern API frameworks, e.g. OpenAPI, are not being used.
- Spatial data are still shared in silos, without links to other resources.
- Content indexing by search engines is not facilitated.
- Catalogue services only provide access to metadata, not the data.
- Data difficult to understand by non-domain-experts.
To address these issues the SDWWG proposed a five point strategy inspired on the
Five Star Scheme [@BernersLee2006]:
- Linkable: use stable and discoverable global identifiers.
- Parseable: use standardised data meta-models such as CSV, XML, RDF, or JSON.
- Understandable: use well-known, well-documented, vocabularies/schemas.
- Linked: link to other resources whenever possible.
- Usable: label data resources with a licence.
The work of the SDWWG triggered a transformational shift at the OGC towards
specifications based on the OpenAPI. But while convenience of use has been the
focus, semantics has been largely unheeded. A Linked Data agenda has not
been pursued.
However, the OpenAPI opens the door to an informal coupling of OGC services with
the Semantic Web, considering the possibility of adopting JSON-LD as
syntax to OGC API responses. The introduction of a semantic layer to digital
environmental data shared through state-of-the-art OGC APIs is becoming a
reality, with great benefits to researchers using or sharing data.
This communication lays down a simple SDI set up to serve semantic environmental
data through a SensorThings API created with the glrc
software. A use case is
presented with soil data services compliant with the GloSIS web ontology.
SensorThings API:
SensorThings API is an OGC standard specifying a unified framework to
interconnect Internet of Things resources over the Web [@liang2016ogc].
SensorThings API aims to address both the semantic, as well as syntactic,
interoperability. It follows ReST principles [@fielding2002principled],
promotes data encoding with JSON, the OASIS OData protocol
[@chappell2011introducing] and URL conventions.
The SensorThings API is underpinned on a domain model aligned with the ISO/OGC
standard Observations & Measurements (O&M) [@Cox2011], targeted at the
interchange of observation data of natural phenomena. O&M puts forth the
concept of Observation
has an action performed on a Feature of Interest
with the goal of measuring a certain Property
through a specific Procedure
.
SensorThings API mirrors these concepts with Observation
, Thing
,
ObservedProperty
and Sensor
. This character makes of SensorThings API a
vehicle for the interoperability of heterogeneous sources of environmental
data.
glrc
:
grlc
(pronounced "garlic") is a lightweight server that translates SPARQL
queries into Linked Data web APIs [@merono2016grlc] compliant with the OpenAPI
specification. Its purpose is to enable universal access to Linked
Data sources through modern web-based mechanisms, dispensing the use of the
SPARQL query language. While losing the flexibility and federative capacities
of SPARQL, web APIs present developers with an approachable interface that can
be used for the automatic generation of source code.
A glrc
API is constructed from a SPARQL query to which a meta-data section is
prepended. This section is declared with a simplified YAML syntax, within a
SPARQL comment block, so the query remains valid SPARQL. The meta-data provide
basic information for the API set up and most importantly, the SPARQL end-point
on which to apply the query. The listing shows an example.
#+ endpoint: http://dbpedia.org/sparql
PREFIX dbo: <http://dbpedia.org/ontology/>
PREFIX dbr: <http://dbpedia.org/resource/>
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
SELECT ?band_label {
?band rdf:type dbo:Band ;
dbo:genre dbr:Hard_Rock ;
rdfs:label ?band_label .
} ORDER BY ?band_label
A special SPARQL variable formulation is used to map into API parameters. By
adding an underscore (_
) between the question mark and the variable name,
glrc
is instructed to create a new API parameter. A prefix separated again
with an underscore informs glrc
of the parameter type. The ?band_label
variable in [Listing @lst:1] can be expanded to ?_band_label_iri
to create a
new API parameter of the type IRI.
Use case: GloSIS:
The Global Soil Partnership (GSP) is a network of stakeholders in the soil
domain established by members of the United Nations Food and Agriculture
Organisation (FAO). Its broad goals are to raise awareness to the importance of
soils and to promote good practices in land management towards a sustainable
agriculture.
Acknowledging difficulties in exchanging harmonised soil data as an important
obstacle to its goals, the GSP launched in 2019 an international consultancy to
assess the state-of-the-art and propose a path towards a Global Soil Information
System (GloSIS) based on a unified exchange. A domain model resulted, based
on the ISO 28258 standard for soil quality [@SchleidtReznik2020], augmented with
code-lists compiled from the FAO Guidelines for Soil Description [@Jahn2006].
This domain model was then transformed to a Web Ontology, relying on the Sensor,
Observation, Sample, and Actuator ontology (SOSA) [@Janowicz2019], and other
Semantic Web standards such as GeoSPARQL, QUTD and SKOS. The GloSIS web ontology
has been successfully demonstrated as a vehicle to exchange soil information as
Linked Data [@GloSIS].
A prototype API for the GloSIS ontology, formulated in compliance with the
SensorThings API specification, will be presented in this communication. It
demonstrates how the same set of SPARQL queries can be used to query through a
ReST API any end-point available over the internet, sharing linked soil data in
accordance with the GloSIS ontology. Thus providing a clear step towards the
federated and harmonised system envisioned by the GSP.
Following the work we did with TiTiler (a python module which is designed to create Raster services), we decided to develop the same kind of project but for Vector. Using Postgres/PostGIS as datasource and FastAPI/Pydantic for the web framework, TiPG is a lightweight application which user can include into their own FastAPI application and easily customized.
During this session I'll go over the design principle of the TiPg python module and also show some of its great features.
Recently, sveltekit is becoming a more popular framework for developing web application. It has been released as v1.0.0 last December. However, there are still not many use cases of developing maplibre applications in sveltekit compared to other frameworks like react. The author is involved in developing maplibre application with sveltekit in United Nations Development Programme (geohub), and also developing sveltekit based Web-GIS applications for water asset management at Eastern African countries (watergis). Hence, several useful maplibre boilerplate and components were developed in sveltekit during those projects' work. watergis/sveltekit-maplibre-boilerplate is a template which can start developing maplibre application in sveltekit with minimuum source code. Furthermore, watergis/svelte-maplibre-components consists of various useful maplibre components to add more functionality easily to your web application (all components are documented here). For instance, this component library provides you features of exporting maps, adding legends, styling maps, sharing maps, measuring distance and integrating with Valhalla api, etc. In this talk, these maplibre boilerplate and components will be briefly introduced.
Maybe you've heard of Kart, the great new geodata versioning tool from the team at Koordinates? But did you know that Kart also has a QGIS plugin so you can do real data versioning without needing to leave QGIS?
In just 5 minutes we'll demonstrate how to import data into a new Kart repository, make and review some changes, merge a branch, and push everything to a remote server. All from QGIS!
—
We’re drowning in data, but the geospatial world lags badly behind in versioning tools compared to our software counterparts. Kart (https://kartproject.org) is solving this with a practical open tool for versioning datasets, enabling you to work more efficiently and collaborate better.
Kart allows you to quickly and easily manage history, branches, data schemas, and synchronisation for large & small datasets between different working copy formats, operating systems, and software ecosystems.
Modern version control unlocks efficient collaboration, both within teams and across organisations meaning everyone stays on the same page, you can review and trace changes easily: ultimately using your time more efficiently.
Static type hints according to PEP 484 (and its extensions) have been a part of Python since version 3.5, which came out in 2015. Research from 2021 shows that 3 out of 4 Python developers already use optional type hinting at least sometimes in their projects. Time is ripe for static type hints to enter the FOSS4G Python world!
A GitHub issue on fiona
's issue tracker to add static type hints to the library recently gained some traction. Currently, it is envisioned to create type stubs for fiona
1.9 and possibly move the type hints into core fiona
with the future 2.0 version.
This talk will give an overview on the current status of the effort to add type hints to fiona
. Furthermore it will briefly discuss considerations and the reasoning behind design decisions taken up until then. Contributions to the effort are very much welcome – just take part in the discussion on GitHub.
Geospatial information from satellites is increasingly being used by decision-makers and scientists alike. However, there are two fundamental issues with this kind of data and related handling technologies. Firstly, data processing typically requires long time and a-priori expert knowledge compared to traditional data sources. Second, integrating satellite data into processing pipelines can be expensive in terms of software and application development efforts. The OpenDataCube (ODC) was created to help users solve these issues. Although ODC offers an alternative to being used as a data management application, its deployment is typically challenging for inexperienced users. Therefore, the primary purpose of this work is to provide potential ODC users with a ready-to-use, portable instance of this software.
The software is produced and published in a Docker container. In comparison to the traditional installation and configuration of the ODC, the tool proposed here provides an environment where the ODC database is already set up. It helps to avoid occasional conflicts that are common in SQL and Python installations. Even though other ODC implementations are available as a Docker container, the proposed solution has some advantages. Specifically, Python geospatial libraries are integrated in the container to support data manipulation. While available ODC instances are designed to process satellite images only (mainly Sentinel and Landsat data), the tool contains scripts to automatically adapt and ingest non-satellite data (e.g. raw ground-sensor network data, land cover/soil maps, etc.) by creating also metadata files when they are missing. The proposed solution makes available processing pipelines to re-grid, georeference and import datasets into the ODC. Both scripts and pipelines can be used through Jupyter notebook interfaces, which allow users also to perform exploratory analyses on the ingested data.
The source code is available at (https://github.com/gisgeolab/LCZ-ODC) and is released under a MIT license. The software is being developed within the LCZ-ODC project (agreement n. 2022-30-HH.0) funded by the Italian Space Agency (ASI) and aimed to identify Local Climate Zones within the Metropolitan City of Milan. Given the nature of the datacube development, this tool promotes Open Geospatial Consortium (OGC) compliant data sharing. Ongoing work focuses on the development and integration of additional pre-processing scripts with the aim of supporting the ingestion of additional types of data as well as providing new ready-to-use embedded processing functionalities.
Because environmental impact assessment(EIA) process is a combination of detailed fields that require a lot of expertise (e.g., noise, air pollution, odor, water pollution, ecological environment, living environment, etc.), despite its long history, the process is still complex and slow, and it is not easy to break away from the document/drawing-centered work process. Since the nature of the environment involves many geographic/spatial context, if it can be assisted with a spatio-temporal system, it can be expected to show very high efficiency compared to the current process.
To verify the feasibility of such a system, we adopted a FOSS4G-based approach and developed a pilot system in this study. Specifically, we used GeoServer and Postgresql/PostGIS for handling and providing data spatially, and Cesium for 3D geospatial based visualization. We focused on the design and implementation of APIs to assemble the sub-processes of EIA, as well as the visualization and UI of the pilot system.
This system demonstrates how the noise propagate during and after the construction in an interactive way. We expect the system will increase the non-expert stakeholder's understanding of noise propagation visually.
Through this presentation, we will discuss our findings implemented in a EIA process centered on the noise, from the first step of applying for approval from the civil/construction operator to the last step of deriving the final evaluation opinion by the noise expert in charge, and provide clues to the future of Digital EIA.
In the future, we believe that the expansion to other EIA media and the smooth implementation of current legal and administrative tasks will make it a system that can be used in the field.
The SpatioTemporal Asset Catalog (STAC) is a standard for describing geospatial data and enabling interoperability. As the STAC ecosystem has grown, the need for robust validation tools to ensure metadata integrity has also increased. STAC-Validator, an open-source validation tool developed by Sparkgeo, was expanded upon by STAC-Check. Originally funded by the Radiant Earth Foundation, STAC-Check provides additional linting and validation capabilities beyond STAC-Validator with a focus on adherence to STAC Best Practices.
In this talk, we will discuss the features, usage, and benefits of STAC-Check and STAC-Validator. Attendees will learn how these tools can identify common metadata issues and ensure compliance with STAC Best Practices. We will also explore how the tools can be integrated into existing workflows for automated metadata validation. By attending, participants will gain a better understanding of metadata validation in the STAC ecosystem and the collaborative efforts of the community to develop and maintain open-source validation tools.
How do you run an aid station in case of a disaster? Scenarios are planned for each city, but there are limitations in applying them to actual aid station operations. In our presentation, we will present a case study on the development and simulation of a aid station management tool using digital twin technology and share various visualization techniques in a 3D city model environment.
The study site is Ulju-gun, a county of about 220,000 people in southern South Korea, with two nuclear power plants operating within a few kilometers of each other. Moving people to shelters to protect them in the event of a disaster such as a radioactive leak is very essential and crucial part of disaster management.
The aid station management tool presented in this presentation leverages ground-truth 3D modeling data of the shelter buildings that will be operational during a disaster to provide facility placement and editing capabilities. This allows relief tents to be automatically placed or edited based on the scenario. It also provides the ability to monitor the overall changes that may occur at the shelter through a dashboard, including real-time victim status, food, beverage, and medical support, supply status, shelter information, and disaster situation information.
The Cesium platform is used to service the data and the Three.js library is used to handle the viewing and placement of 3D model data in glTF format. Other open source implementations include React, Turf.js, Apache ECharts, and GeoServer.
We believe that the findings mentioned in this study provide a good example of how 3D city model-based shelter operations and visualization techniques can be applied to disaster preparedness systems to support effective decision-making and resource allocation.
Access to high-quality data on existing bicycle infrastructure is a requirement for evidence-based bicycle network planning, which can support a green transition of human mobility. However, this requirement is rarely met: Data from governmental agencies or crowdsourced projects like OpenStreetMap often suffer from unknown, heterogeneous, or low quality. Currently available tools for road network data quality assessment often fail to account for network topology, spatial heterogeneity, and bicycle-specific data characteristics.
To fill these gaps, we introduce BikeDNA, an open-source tool for reproducible quality assessment tailored to bicycle infrastructure data. BikeDNA performs either a standalone analysis of one data set or a comparative analysis between OpenStreetMap and a reference data set, including feature matching. Data quality metrics are considered both globally for the entire study area and locally on grid cell, thus exposing spatial variation in data quality with a focus on network structure and connectivity. Interactive maps and HTML/PDF reports are generated to facilitate the visual exploration and communication of results.
BikeDNA is based on open-source python libraries and Jupyter notebooks, requires minimal programming knowledge, and supports data quality assessments for a wide range of applications - from urban planning to OpenStreetMap data improvement or transportation network research. In this talk we will introduce how to use BikeDNA to evaluate and improve local data sets on bicycle infrastructure, examine what BikeDNA can teach us on the current state of data for active mobility, and discuss the importance of local quality assessments to support increased uptake of open and crowd-sourced data.
This talk is about a prototype that enables collaborative mapping without the need of any internet connectivity, only a local network is required. It runs fully in the browser, hence is cross-platform, it basically runs on any smartphone. The users form a peer-to-peer network in order to exchange their data.
It can be used in situations where there either is no internet infrastructure, it's spotty or it was destroyed. In the disaster response case, only a local network, without any server infrastructure, would be needed.
In the talk you'll learn about content-addressing, WebRTC and peer-to-peer networks and of course experience a live demonstration of the prototype.
The tech-stack is [Svelte] for the application, OpenLayers for displaying the map, IPFS for the storage, libp2p for the networking. The project is licensed under the Apache/MIT licenses.
FieldMaps.io is a personal initiative originally created to develop offline interactive reference maps for humanitarian actors. However, in short time, it transitioned to helping develop common operational datasets that form the foundation for humanitarian response planning. Over the past 2 years, enormous effort has gone into releasing a high-resolution composite dataset able to be updated daily from multiple sources. This talk will cover 3 aspects of the project.
Algorithm
Edge-matching resolves gaps and overlaps between hundreds of separate national data sources, requiring an algorithm that can perform at global scale. The resulting methodology uses something akin to a euclidean allocation raster applied to vector space, free of the compromises other approaches like generalization and snapping make. If you've ever been challenged by topology or data cleaning, you might find some insights into solving your own problems with the ideas contained here.
Pipeline
The edge-matching algorithm involves multiple complex and computationally intensive steps. Although Geopandas and GDAL usually come to mind when building multi-step geoprocessing scripts, PostGIS ended up being the fastest and best scaling tool for transforming gigabytes of vector data. I'll challenge your assumptions of how it can be used to create pipelines on both desktops and in the cloud, and make a case for why you should include it in your next project.
Sources
A composite dataset is only as good as the foundations it builds upon, and great care was taken in selecting which sources were used in this project. For international boundaries, I'll go into detail about how I used only public domain sources to create an ISO 3166 compliant dataset. At the subnational level, I'll highlight two projects that each curate updated administrative boundaries: one by the United Nations, another by an academic institution.
Whether you're a remote sensing specialist in search of the best topologically valid boundaries to run zonal statistics with, a Python developer frustrated by your pipelines constantly running into memory limits, or just want to run this tool on your own boundaries, I hope you come away from this talk with a valuable concept you can apply to your own work.
Data: https://fieldmaps.io/data
Tool: https://github.com/fieldmaps/edge-extender
In Cameroon, the planning and monitoring of a measles vaccination campaign is implemented in an open source software called Iaso built on a Python based backend combining Django and Postgres/Postgis ; the frontend is React based. Iaso aims to provide a number of core functionalities to support ongoing geospatial data management: a mobile application, a web dashboard, a mapping function to merge various data sources, a user-friendly API for data science and scripting, and a seamless bi-directional integration with DHIS2 (standard health information system in low- and middle-income countries).
Iaso is articulated around three essential components : a central georegistry interface, a mobile data collection tool and a micro planning interface. Those tools are integrated seamlessly with each other to provide a powerful platform to manage, update, merge and validate multiple data sources and structured information collected. Geospatial data from GPS collection to the management of multiple reference lists of organization units (Health, Administrative or School pyramid) are Iaso's foundation. Those features allow interconnecting collected data to existing hierarchical features coupled with planification and collection of survey campaigns in the field through the mobile application and the web platform.
Iaso exposes a full API providing various endpoints allowing data scientists to integrate data analysis pipeline through external analytic platform. As a geospatial data management platform, it provides versioning of every dataset and is designed to keep a full history of all the changes on the data of interest from the forms to the geometry or metadata of the organization units. It also features seamless integration with QGIS and other desktop applications through a templated Geopackage format.
In this presentation, the tool is explained and described from the planning of the vaccination campaign in Cameroon to the near real-time monitoring of the campaign (eg. stock and team planning management).
Source : https://github.com/BLSQ/iaso
In the field of surveying, expensive equipment is a barrier for many organizations and professionals, especially those with limited financial resources. However, recent advances in low-cost differential GPS technology using U-blox ZED-F9P GNSS Receiver, have shown that it is possible to achieve accuracy that is on par with more expensive equipment. This innovation is making it possible for young and local professionals to access reliable and affordable surveying tools. With my experience setting up and using open-source GNSS receivers in Tanzania, Uganda, and Liberia, we were able to conduct surveying despite challenging conditions while collecting high-accuracy geospatial data. This development has the potential to transform how we conduct surveys in areas with constrained resources. Moreover, the technology has far-reaching implications for the future of surveying. With the increased availability of low-cost GNSS receivers, we can expect to see more young professionals entering the field, bringing fresh ideas and perspectives. Additionally, democratizing access to surveying equipment could make it possible for more organizations, regardless of their financial resources, to conduct high-quality surveys. This technology has the potential to revolutionize the field of surveying.
The Joint Research Centre (JRC) of the European Commission is committed to providing independent, evidence-based science and knowledge that supports EU policies. To facilitate this, the JRC has developed the Big Data Analytics Platform (BDAP), a data platform that allows data scientists to easily access, analyze, view, and reuse scientific data to generate and communicate evidence-based insights and foresight.
BDAP hosts spatiotemporal data at petabyte scale from various domains, including elevation, meteorological, administrative, and satellite Earth Observation data. Its architecture leverages almost entirely on Free and Open Source software and tools. The platform offers a cluster environment with both CPU and GPU machines, allowing for large-scale data processing. Additionally, users can visualize and interactively analyze their data through Jupyter Notebooks and Voilà dashboards.
Recently, BDAP implemented the Spatio Temporal Asset Catalog (STAC) specification to describe its data. The catalog hosts different types of data, which share the basic STAC fields. Thanks to the STAC modularity each data type can be described with its own STAC extensions.
BDAP reuses and benefits from various STAC Free and Open Source software and tools. In particular, from the STAC ecosystem it implements the STAC Browser for displaying and searching data, it provides STAC compliant APIs through STAC FAST-API backed by an elasticsearch instance, and uses PySTAC as a Python library for working with STAC metadata. This implementation helps BDAP in its FAIRification process improving users' search, access, and reuse of data.
In this presentation, the design and implementation of the STAC compliant set of software tools will be described. Some real use cases will be presented, with an example on the creation of analysis ready data cubes from Sentinel-2 Earth Observation satellite imagery.
STAC Browser is a full-fledged web interface for browsing and searching static STAC catalogs and STAC APIs. It has been rewritten from scratch with a lot of new functionality. This talk will introduce STAC Browser, showcase new functionality and uncover some unexpected gems such as the broad range of customization possibilities. Lastly, the presentation will guide you through a set of best practices for your static STAC catalog or STAC API so that you get the most out of STAC Browser with regards to functionality and user experience.
This project was a pilot of a larger upcoming project, where the aim is to produce a national interoperable data model for every valid zoning and city plan in Finland. The project is part of the development of the Finnish Environment Institute’s Built Environment Information System and the harmonization of national land use planning information.
The aim of this presentation is to present the overall workflow of the project and the transition from proprietary data towards an open source national database with common spatial and descriptive information. Currently the data used in municipal decision making processes in Finland consists of proprietary data that is lacking spatial information or is outdated.
The transformation of the zoning and city plans from two different data providers created a lot of topological errors and unmatched geometries. QGIS was a key tool for fixing these errors - the digitizing and geometry repair tools were used in solving these issues.
This pilot project was implemented in Southern Savonia, Finland. In the region, zoning has been executed for approximately 80 % of the whole land area. The focus of the project was to investigate the compatibility of the base data and how to automate the processes of merging, fixing, updating and comparing the data. The data was in vector format and was provided by the National Land Survey of Finland and municipalities of Southern Savonia.
The automation processes were built with a python script and the quality control was made with manual digitization. The official documentation of the zoning and city plans were included in the borderline vector data. The final product was uploaded to a GitHub repository. The project also managed to produce a timeline for the upcoming nationwide project and the distribution between automated and manual workload in similar projects.
The methods and the results of the project could be duplicated in other countries or lead the way towards more open national or regional land use planning.
Mapillary is an open platform for street-level imagery and map data that began in 2013. Since then around 1.8 billion images have been contributed from around the world. Imagery has been contributed from horseback in Kyrgyzstan, boats in the canals of Amsterdam, and bicycles on the streets and trails of Sydney. As Mapillary approaches 2 billion images, we’d like to summarize the latest features, acknowledge some of the amazing contributions, and hint at some of the updates that are coming.
Some of the things that we have been working on include:
Desktop Uploader improvements including support for videos and popular cameras.
Improvements to Mapillary Tools, command line scripts for working with and uploading geotagged imagery and video.
Mobile app updates including multi-tasking, redesigns, multi-language support, and upload improvements.
Camera Grant programs in the US and Europe, providing 360º cameras for people interested to map pedestrian infrastructure.
Integrations with Rapid Editor, an AI powered OpenStreetMap editor which we will demo in more detail at a workshop.
Updated Help Pages to make capturing, uploading, and using street-level imagery far easier.
After walking through the latest Mapillary improvements, we will take a look at case studies of organizations contributing and using imagery. We’ll zoom in on an NGO, a government agency, and a commercial entity, each of which are using Mapillary in different ways.
We’ll finish our talk with an exploration of upcoming Mapillary features and projects. We encourage questions and suggestions in the Q&A and hope for a productive conversation at the end as we walk together towards 2 billion images.
An overview of the Core Models and Encodings for Styling and Symbology - Part 1: Core ("SymCore") 2.0 draft candidate Standard.
In comparison to the current OGC Symbology Conceptual Model: Core Part ("SymCore") version 1.0, the new draft candidate Standard aims to better reflect its classification as an OGC Implementation Standard by including the requirements classes needed to enable the implementation of interoperable encodings, renderers (e.g., OGC API - Maps / OGC API - Tiles) and systems parsing and/or generating style definitions (e.g., OGC API - Styles, visual style editors, style transcoders).
It does so by featuring:
- A modular logical and conceptual model for styling capabilities,
- A minimal Core requirements class including clear extension mechanisms, through the definition of abstract Selectors, Symbolizers, and Expressions,
- a basic Vector Styling requirements class,
- a basic Coverage Styling requirements class,
- requirements classes providing additional styling functionality,
- a JSON encoding of the conceptual and logical model facilitating machine readability,
- a CSS-inspired encoding of the conceptual and logical model facilating hand-editing.
The latest version of the draft is available in HTML (https://opengeospatial.github.io/ogcna-auto-review/18-067r4.html) or PDF (https://opengeospatial.github.io/ogcna-auto-review/18-067r4.pdf).
The official GitHub repository is at: https://github.com/opengeospatial/styles-and-symbology
pgRouting is evolving rapidly, many changes have been taking place. Lets catch on.
The focus of this talk will be on the topology functions that were created on 2013, Its been 10 years, and its their time to go: * Why "I" don't want to use them any more * New specialized functionality has been created that substitute the work that the topology functions are doing in a very rustic way. * A quick guide on how not to use the "soon to be deprecated topology functions"
In the submitted paper, the topicality of the building stock in municipalities in Baden-Württemberg, part of the Federal Republic of Germany, is examined. Three municipalities were selected and included in the study according to the spatial type concept of the Federal Office for Building and Regional Planning (BBSR 2023): rural town 2,000-5,000 inhabitants, small town 5,000-20,000 inhabitants, medium-sized town, 20,000-100,000 inhabitants. The analysis concept is explained and the quantitative and qualitative results of the project, which is currently in its final phase, are presented. The aim is to use these results to derive and communicate recommendations for action for the municipalities, but also for the public surveying administration, in order to contribute to timely and effective action by municipal decision-makers and citizens through faster provision of geospatial data.
“Cloud-Native Geospatial” is a new paradigm for performing efficient data access and compute the cloud in an interoperable way in order to achieve scalable and repeatable analysis of geospatial data. The last few years have seen major developments in open standards and open software that make this possible, supporting full end to end interoperable workflows on remote sensing data, starting from data discovery to publishing of derived products.
This talk will provide an overview of what Cloud-Native geospatial is and why it is important for building scalable architectures. It will cover the current state of the Spatio Temporal Asset Catalog (STAC) specifications, and the landscape of cloud-optimized file formats, for raster, vector, and point-cloud data formats (COG, GeoZarr, GeoParquet, COPC).
On February 6, 2023 a sequence of major earthquakes with magnitudes 7.8 and 7.5 have struck Southern Turkiye and Northern Syria, causing massive damage and very high number of casualties in both countries. The sequence of earthquakes were followed with hundreds of aftershocks within the month following the earthquakes, as well as triggering other major earthquakes, such as the 6.4 magnitude earthquake that had struck Antakya on February 20. Humanitarian OpenStreetMap Team (HOT), with Yer Çizenler (YÇ), HOT’s local partner within the Turkish OSM community, have activated to map the missing road and building base data with the help of regional and global OpenStreetMap communities.
More than 7 thousand contributors from these communities, together, have contributed to the addition of more than 1.4 million buildings, 70,000 km of roads into OpenStreetMap for the use of field volunteers and organizations worldwide.
In this talk, the audience will be informed about the coordinated efforts within this mapping activation, the impact of the data created with some example use cases within the response activities. The audience will be informed about various open data sources that were used to enhance the existing OSM data, and their licensing and compatibility considerations during the mapping process. The presenters will also describe the validation, data quality assurance and monitoring methods, approaches and tools utilized for ensuring the OSM data is reliable, current and is able to meet community standards within both short and long terms.
Cloud computing is revolutionizing the way companies develop, deploy and operate software and GeoSpatial software is no exception. With cost savings to simplified management, flexibility, lower downtime and scalability of dynamic environments more and more companies are migrating their on premise systems to the cloud but cloud based setups have their own set of hurdles and challenges.
The migration of the series itself can be challenging. Monitoring, debugging and scaling of applications are very much different than what you are used to.
In this presentation we will share with you the lessons we have learned at GeoSolutions and share some common patterns for the migration of on premise GeoServer clusters to the cloud. We'll share with you tips on how to:
- best practices to migrate your existing GeoServer cluster to the cloud
- insights on your geoserver cluster using centralized logging and Monitor plugin
- avoid common bottlenecks to best set up a distributed scalable GeoServer cluster
- work containers and container orchestrators like Kubernetes
The Knowledge Laboratory, in short k.LAB, is a software stack that embraces the FAIR principles: findable, accessible, interoperable and reusable. Its objective is to support linked knowledge across the borders of the domains of single modelers and scientists. k.LAB’s fascinating novelty is the use of semantics to create a natural language to describe the models and the qualities that want to be observed.
Modelers can develop their models and publish them to the network. Publishing makes them findable and accessible within the network. Since everything in the network is observable, when running a model, k.LAB looks for the best knowledge unit able to resolve the particular request. Interoperability is build and reusability is a natural consequence.
The k.LAB software stack is free and open source and relies on various projects of the Osgeo community as Geoserver, Openlayers and the Hortonmachine. It has been in development for almost 2 two decades and got a particular visibility boost in 2021, when the Statistics Division of the UN Department of Economic and Social Affairs and the UN Environment Program, in collaboration with the Artificial Intelligence for Environment & Sustainability at the Basque Centre for Climate Change, launched the Artificial Intelligence powered application for rapid natural capital accounting: the ARIES for SEEA Explorer.
Lately a python client that allows interaction with k.LAB has been released. This opens up to new ways to observe the world from within common GIS tools as for example QGIS.
An overview of the state of the art of the project will be given.
In this presentation, we showcase a unique approach to analyzing Capital Bikeshare trips in Washington D.C. using Open-Source Geospatial (FOSS4G) tools and technologies. Our project involved loading trip data into a PostGIS database, utilizing the Valhalla routing engine and OpenStreetMap data to find the optimal routes between each pair of stations, and then constructing a topogeometry table to represent these routes. Using this topogeometry table, we are able to estimate the number of Capital Bikeshare trips that occur on each road in Washington D.C.
The use of FOSS4G tools and technologies allowed us to perform this analysis in a cost-effective and efficient manner, while also providing high-quality results. The results of our analysis have important implications for urban planning and mobility research, as they can be used to understand the patterns and impacts of bike-share usage in cities.
Our presentation will provide an overview of the methodology used in our project, as well as a discussion of the results and their implications. We will also share our experiences using FOSS4G tools and technologies and provide insights on how these tools can be used in similar projects. This presentation is of interest to geospatial professionals, urban planners, and anyone interested in using FOSS4G tools for data analysis and mobility research.
SOZip (Seek-Optimized ZIP) is a new open specification on top of the ZIP archive format to compress one or several files organized and annotated such that a SOZip-aware reader can perform very fast random access (seek) within a compressed file.
SOZip makes it possible to access large compressed files directly from a .zip file without prior decompression. It is not a new file format, but a profile of the existing ZIP format, done in a fully backward compatible way. ZIP readers that are non-SOZip aware can read a SOZip-enabled file normally and ignore the extended features that support efficient seek capability.
We will present how SOZip works under the hood and discuss about SOZip implementations, in particular in GDAL, which make it possible for its downstream users, in particular QGIS, to read seamlessly and efficiently large compressed files in GeoPackage, FlatGeoBuf, or shapefile formats.
The surge of Street View Imagery (SVI) as an essential data source for urban analytics, especially in streetscape audit studies, has been catalyzed by the proliferation of imagery platforms, advances in computer vision, machine learning, and the availability of computing resources. However, it has been noted that the mainstream data provided by government agencies and private companies (e.g., Google Street View, Baidu Total View) have limitations in spatial resolution, update frequency (e.g., GSV updates data every few years), and data application (e.g., barriers to free data download, historical data retrieval, and free data use). These restrictions present issues when achieving more accurate and dynamic streetscape monitoring.
The emergence of the Web 2.0 era has fostered the potential for individuals to contribute and access information through multiple resources, which has also facilitated the collection of massive Volunteered Street View Imagery (VSVI). The VSVI data have the potential to provide more open, comprehensive, and
diverse geographic information, which is however conditional on a set of criteria such as data completeness and quality. To better understand the value of this novel type of data in streetscape monitoring studies, this study aims to analyze the generation process of VSVI and examine relevant characteristics in Tokyo using the typical VSVI data of Mapillary.
The generation process of Mapillary from 2014 (the inception of Mapillary) to 2022 is analyzed from the perspective of road expansion, data amount accumulation, and hotspot change of VSVI data; the examination of characteristics includes the assessment of spatial distribution (road coverage and spatial
density), contribution time distribution (revisit time, update frequency, and seasonal diversity), and image quality (image type and shooting perspective). These analyses use GSV imagery data as a benchmark.
Every year, there's a new Postgres major release that improves on performance in certain areas and could provide new hooks for extensions like PostGIS to take advantage from them. If not planned well, upgrading your production databases can become a pain. Sooner than you think you'll be running on EOL (End-of-Life) versions because the upgrade has been postponed too many times. Don't!
Did you know Postgres upgrades can be greatly automatized these days with downtimes of only a few seconds? This talk will show you how and will also present some essential features from recent Postgres and PostGIS versions to get you excited for the new upgrade.
Geospatial analysis welcomes an audience to interact with complex interactions and dynamic shifts in ecosystem balance. Location intelligence collected as data layers mirror a symphony or chapters in a book. We will explore the potential risks of vulnerable cities by exploring the environment, economics, built infrastructure, and how they intersect. We build the story or music over time while exploring the tensions we create. Let’s examine the edges of eco-geomorphic frameworks and listen for a narrative.
We will explore how Re:Earth as a digital public good could support a "Peaceful Profitable Society" and create new employment opportunities.
Re:Earth is an open source platform built around a geographic information system that digitally represents geospace and enables analysis and visualization of cities and regions. The use of such digital public goods offers opportunities to develop new ways of working and improve their own lives, especially for the socially vulnerable.
In particular, we will explore the potential for vulnerable populations, such as refugees and single mothers, to use Re:Earth to pave the way for self-empowerment. We will also delve into how digital public goods such as Re:Earth can impact society as a whole, especially how they can be a tool for the vulnerable to improve their own lives and contribute to the realization of a "society where peace is profitable".
This speech will provide insight into how such digital public goods can impact individual lives and society as a whole, and how they can help shape a "society where peace is profitable".
Geochicas is a initiative born in State of the Map Sao Paolo and adopted by FOSS4G communities over the past years. We would like to share with you what had happened in the last couple of years and what we foresee in the future of the initiative. How Geochicas is part of a larger ecosystem of siblings organizations working towards having a more balanced presence of women and minority groups in the Geospatial communities.
Gleo is a nascent javascript WebGL mapping library. It aims to find a niche alongside Leaflet, OpenLayers, MapLibre and Deck.gl.
This library was presented at FOSS4G 2022, with an emphasis on its architectural foundations: geometry/reprojection/antimeridian handling, and object-oriented abstractions for WebGL data structures.
This session provides a tour of the features developed during the last year. These include, among others:
- Work done as part of the OSGeo-OGC codesprints (OGC API clients, experimental symbols)
- Animated symbols (render loop)
- Symbol class decorators (ability to add more functionality to a cartographic symbol class during runtime)
- Flexibility of scalar field manipulation (symbols that render as a magnitude instead of a colour, then the field renders as e.g. a heatmap)
These functionalities are a fresh approach to cartographic rendering and will provide a glimpse of the potential of Object-Oriented WebGL manipulation for cartographic rendering.
If you have ever had the experience of having to write code to draw on web maps, you'll know how painful the process can be - especially when situations get more complex.
Terra Draw is an open source JavaScript library that provides a new way to add drawing functionality to a host of web mapping libraries, including Leaflet, OpenLayers, Google Maps, MapboxGL JS and MapLibreGL JS.
The library provides a selection of built in modes that 'just work' across different mapping libraries. These features include elementary drawing tools like point, line and polygon, as well as supporting more advanced concepts like snapping, rotation and scaling.
Terra Draw is also designed to be extendable so that you can write your own custom modes and adapters (thin wrappers for each mapping library). The architecture of the library means that any mode work can work with any adapter and vice versa creating a strong multiplier affect as new modes and adapters are written. This decoupling has the added benefit that drawing libraries can be swapped out without breaking your app!
The talk will examine the history of the library, how to get started, and also an opportunity to hear more about the future of Terra Draw.
In our allocated 15 minutes, we would like to take you on a trip following the winding roads of building a community, the Romanian geospatial community: geo-spatial.org. We want to share our story, beyond our geodata and knowledge portal, to the very core of the values and principles that have guided us through difficult times and made our overcame challenges even brighter.
In our more than a decade of existence, we’ve organised over 25 national FOSS workshop, a regional FOSS4G in 2013 and a global FOSS4G in 2019, we’ve initiated collaborative geo-related projects and managed to infuse the geospatial component in various non-spatial organisations, such as the ones in education or investigative journalism.
The Long Island Zoning Atlas is an interactive web map that displays zoning data, public services, and demographic data for municipalities all across Long Island excluding New York City. The app focuses on statistics that help affordable housing advocates plan housing projects. This year we rebuilt the Long Island Zoning Atlas using our new FOSS stack. The project presented a problem very common to GIS projects: transforming data from many different sources, in this cases towns. We were given the data in many different formats and needed to transform it all into clean, usable data which is organized to our needs and renders quickly and efficiently on the web.
Published in 2020, the European strategy for data sets the vision for Europe to become a leader in a data-driven society by establishing so-called common European data spaces in all strategic societal sectors. Data spaces are envisioned as sovereign, trustworthy and interoperable data sharing environments where data can fairly flow within and across actors, in full respect of European Union (EU) values to the benefit of European economy and society. The development of data spaces is accompanied by a set of horizontal legislative measures, including, among others, an Implementing Act on high-value datasets under the Open Data Directive that lays down a list of datasets (many of which being geospatial) that EU Member States public sector organisations are required to make available for free, under open access licenses, in machine-readable formats and via Application Programming Interfaces (APIs).
The talk will describe the activities around open source geospatial software and open geospatial data that the European Commission’s Joint Research Centre (JRC) has performed to support the development of the common European Green Deal data space, focused on environmental data sharing and instrumental to address climate changes and environmental challenges in line with the top priority of Von der Leyen’s Commission 2019-2024.
A key enabler to bring public data into this data space is the infrastructure setup for the EU INSPIRE Directive, which is technically coordinated, maintained and operated by the JRC. The INSPIRE Directive itself, together with the Directive on public access to environmental information, are currently subject of an impact assessment that might lead to a revision of the legal framework (GreenData4All initiative). This is accompanied by an overall modernisation of the technical infrastructure, increasingly based on open source software both at the Commission side (GeoNetwork for the INSPIRE Geoportal, ETF for the INSPIRE Reference Validator and Re3gistry for the INSPIRE Registry) and at the Member States side, where FOSS4G tools are the primary choice for both serving and consuming data. Thanks to a number of INSPIRE Good Practices promoted by the community, new standards and approaches for data encoding and sharing (e.g. based on OGC APIs) are bringing additional value to the INSPIRE stack. The same set of approaches ensures the full alignment and complementarity between INSPIRE and the Implementing Act on high-value datasets, thus positioning open source geospatial software as a true enabler for the Green Deal data space.
By using a mobile device, such as a smartphone or tablet, to collect geographic data, mobile mapping involves applying that data to OpenStreetMap (OSM). Because it enables real-time data collection and updates, this technology has significantly increased the efficiency and accuracy of mapping. Navigation, digital twin generation, and crowdsourced mapping are some uses for mobile mapping on OSM. By allowing community members to contribute to the map, crowdsourced mapping improves the completeness and accuracy of the map. A digital twin of a physical location can be created using mobile mapping data and utilized for planning and decision-making. OSM data gathered by mobile mapping can be used by navigation applications to provide more precise and current routing information. Overall, mobile mapping has significantly enhanced the capabilities of OSM and will continue to play a major role in mapping and location-based services.
In this workshop, we shall present about the different mobile applications that we use in editing, updating and mapping on OSM.
Applications:
OrganicMaps
OSMAnd
Vespucci
ODK
Mapilary
And many others!
QGIS releases three new versions per year and each spring a new long-term release (LTR) is designated. Each version comes with a long list of new features. This rapid development pace can be difficult to keep up with, and many new features go unnoticed. This presentation will give a visual overview of some of the most important new features released over the last calendar year.
In March of 2023 a new Long-term release was published (3.28), and shortly before FOSS4G, the latest stable version of QGIS (3.32) will be released. I will start by comparing the new LTR (3.28) to the previous (3.22). Here I will also summarize by category the new features found in the latest LTR (GUI, processing, symbology, data providers etc.).
I will then turn my attention to the important new features found in the latest releases (3.30 & 3.32). Each highlighted feature will not simply be described but will be demonstrated with real data. The version number for each feature will also be provided. If you want to learn about the current capabilities of QGIS, this talk is for you!
Potential topics include: Annotation layers * GUI enhancements * New Expressions * Point cloud support * Print layout enhancements * New renderers and symbology improvements * Mesh support * 3D * Editing
Do you want to broaden your horizons by learning about geospatial support for the United Nations operations? Or are you interested in developing highly efficient and portable geospatial apps which make use of PMTiles, COPC, COG, Raspberry Pi, and a cool Web3 technology named IPFS (Inter-planetary File System)? We are doing both in the Domain Working Group 7 (DWG 7) on Smart Maps of the UN Open GIS Initiative.
In this participatory and voluntary DWG established in Firenze in August 2023, participants bring in their objectives and combine efforts within the Partnership for Technology in Peacekeeping to bring greater involvement to peacekeeping through innovative approaches and technologies that have the potential to empower UN global operations. In addition to our core objective to support the use of UN Vector Tile Toolkit in the UN Global Service Centre, DWG 7 is supporting domestic and campus-level service operations, and supporting 3D geospatial data such as point clouds and 3D city models. We are combining efforts to define and implement the concept of Smart Maps.
We are happy to share with you our new effort named Model UN Development and Operations (MUNDO) that simulates geospatial support for the United Nations operations by making use of existing open geospatial data and our Smart Maps technologies. MUNDO project is not only useful for demonstrating the technology for the UN staff, but also useful for learning about the situation and the UN’s effort. We are also happy to share with you our new concept of WebMaps3, which introduces Web3 technology for web maps. By combining IPFS and cloud optimized formats like PMTiles, COPC, and COG, we were successful in hosting a vector tiles service from a newly released nation-wide cadastre dataset on a Raspberry Pi, within 10 days after the release, by producing a 14GB PMTiles file.
deck.gl is a popular open source data visualization library that uses the power of WebGL to render huge amounts of data performantly in the browser. A collection of versatile layers allows the user to create many different types of visualizations, with excellent support for geospatial data in particular.
The core layers can be extended by the means of deck.gl extensions to create interactive experiences which are not possible in other data visualization frameworks.
This talk will give an overview of deck.gl, including some of the core layers and will then focus on three of the latest extensions:
- The MaskExtension implements realtime masking of data by an arbitrary spatial boundary. An example use case is clipping a set of roads and places of interest to the boundary of a city.
- The CollideExtension avoids collisions between features on screen. This can be used to selectively show large cities in preference to small ones on a map when they would otherwise overlap.
- The TerrainExtension offsets the 3D component of features by referencing a separate 3D layer. For example, a set of pins on a map can be placed at the correct height relative to a 3D terrain layer.
OpenLayers is a powerful web-mapping library, and it has been around for quite a while. Far from being stuck in a past state where it offered most features anyone could expect, the community of contributors and maintainers are continuously pushing it forward, rethinking orientations and taking in new trends. Be it cloud-native formats, emerging standards or drastic performance improvements, more and more innovations are becoming parts of OpenLayers feature set.
This talk will give you an overview of the past few years of development, and show in how many incredibly useful ways OpenLayers can be used nowadays. We will also discover the exciting developments that are shaping up for the future, and how all this is being made possible.
30DayMapChallenge is a daily map making challenge which is held since 2019 every year in november on social network. This challenge has become year after year popular for the mapmakers community, and more than 8000 maps have been posted in 2022 session.
Last year was my first participation, it was a great opportunity to try to make unusual maps, complete sleeping projects, and be updated with geospatial technologies.
In this talk will be presented how this challenge has been completed and especially which open tools has been used to make the 30 maps.
The European Ground Motion Service (EGMS) is part of the Copernicus Land Monitoring Service (CLMS) lead by the EEA (European Environment Agency). EGMS is based on the full resolution InSAR processing (20x5m) of the European Space Agency (ESA) Sentinel-1 (S1). This massive geospatial timeseries dataset is composed by ~10.000 million timeseries distributed over 31 European countries. The baseline covers 2015-2020 and updates are being published on a yearly basis. It is publicly accessible at https://egms.land.copernicus.eu/ with a 3D viewer and download service.
This open dataset consists of three product levels (Basic, Calibrated and Ortho). The Basic and Calibrated are offered at full resolution 20x5m (Line of Sight) whereas the Ortho product offers horizontal (East-West) and vertical (Up-Down) anchored to the reference geodetic model resampled at 100x100m.
Sixense is coordinating a consortium responsible for the independent validation of this continental scale geospatial dataset. The validation goal is to assess that the EGMS products are consistent with user requirements and product specifications, covering the expected range of applications. To evaluate the fitness of the EGMS ground motion data service seven reproducible validation activities (VA) have been developed gathering validation data from different sources across 12 European countries:
• VA1 – Point density check performed by Sixense.
• VA2 – Comparison with other ground motion services carried out by NGI (Norwegian Geotechnical Institute).
• VA3 – Comparison with inventories of phenomena/events performed by BRGM (French Geological Survey).
• VA4 – Consistency check with ancillary geo-information carried out by NGI.
• VA5 – Comparison with GNSS data performed by TNO (Dutch Geological Survey).
• VA6 – Comparison with insitu monitoring data performed by GBA (Austrian Geological Survey).
• VA7 – Evaluation XYZ and displacements with Corner Reflectors performed by TNO.
The validation environment developed and maintained by Terrasigna includes all the necessary elements to perform all the validation tasks from data collection and description to execution of the different methodologies. The objective of this portable Kubernetes/Terraform cloud-based system is to guarantee reproducibility of all the validation activities:
• A MinIO web-based validation data upload tool where scientists can upload their validation data and EGMS subsets.
• A validation data catalogue based on GeoNode (based on OGC CSW) where all validation sites data is properly described and georeferenced to ensure reproducibility.
• JupyterHub notebook environment where scientists can develop their validation scripts (Python/R). These notebooks produce graphs and figures to be included in the yearly validation reports.
We developed a free graph-based geo-intelligence engine that serves fast, scalable, and reliable data analysis. The engine's value lies in its flexibility and applicability to any relational dataset, as well as its integration of open-source technologies and libraries. We chose to build our geo-intelligence engine on a graph infrastructure to enable faster, index-free queries and better support for interconnected data.
To showcase the capabilities of our engine, we have developed a geo-financial software that provides users with a powerful tool for analyzing financial scores of companies based on geo-location. Businesses can quickly and easily analyze data to gain valuable insights into competitors, potential partnerships, and market trends. Our software presents the results of the analysis in a user-friendly and visually appealing format, making it accessible even to non-technical users.
Our geo-financial analysis software is based on user-specified location and range. The user interacts with an Angular frontend, which incorporates the Leaflet library for map interaction and an OpenStreetMap basemap. The backend is based on Golang, which handles authentication and message queueing interaction with a Python analysis tool. The data retrieved for Python processing comes from a Neo4j graph database, which is accessed through Cypher queries and networking algorithms. All of the software components are located in separate containers, promoting flexible and independent scalability achieved with Docker Compose and orchestrated by Kubernetes.
In this presentation, we will discuss our graph-based geo-intelligence engine, which is the backbone of our application. We will showcase the geo-financial analysis application itself, providing a demo and demonstrating how it can be used for business geo-intelligence analysis. Throughout the presentation, we will continuously discuss the open-source technologies that are at the core of our work and focus on the value that each of them has brought to our achievements.
Using generic or standard content management system (CMS) like Wordpress or Strapi for managing geospatial data isn't an optimal solution. Since object geometry isn't just one of many data fields, it requires special handling for setting the data (e.g., on the map), storing data, transforming data for various needs (geometry output format, CRS etc.) and using them for spatial analysis.
When talking about a geospatial CMS, one would think that using GeoServer should be a must. How else would you vizualize a non-trivial amount of data on the map, right? Although Geoserver might be a good answer, that's not the only one. We, at our company, have developed our custom geospatial CMS using the OpenLayers mapping library on the frontend and PostgreSQL (with PostGIS, of course) on the backend, using PHP Laravel and GeoJSON as middle man between the data store and the frontend.
CMS platforms frequently have one specific feature. Different objects may have various attributes. Using the EAV (entity-attribute-value) model is one of the methods that is frequently utilized, although this choice usually comes with a number of issues, such as querying and storing the data. We used the possibility to swap out the EAV model for a straightforward json field in our CMS.
This talk will present what choices we had to make to build solution in such way and what some of our challenges were.
UNVT Portable is a package for RaspberryPi that allows users to access a map hosting server via a web browser within a local network, primarily for offline use during disasters. It is designed to aid disaster response by combining aerial drone imagery with OpenStreetMap and open data tile datasets.
Since we have introduced QGIS 3D in 2017, it has gone through major improvements. In addition to new features, several new data formats have been also integrated to QGIS.
This presentation will cover the latest improvements made as result of the recent crowdfunding efforts to introduce point cloud processing, enhance 3D maps for elevation data.
OL-Cesium is a popular Open source Javascript library that you can leverage to add 3D to a new or existing OpenLayers application. You code the logics in a single place and it gets applied to both OpenLayers 2D map and Cesium 3D globe. The library handles the synchronization of the view, layers, styling, for you. This behaviour is customizable.
Since its creation, 9 years ago, the library has attracted a large community of users. It has evolved to follow OpenLayers, Cesium and the global javascript ecosystem.
This talk is about the strengths of the library, its state and the plans for the future.
FlowmapBlue is a free tool for rendering origin-destination data (counts of movements between pairs of locations).
It is used to visualize urban mobility, commuting behavior, bus, subway and air travels, bicycle sharing, human and bird migration, refugee flows, marine traffic, freight transportation, trade, supply chains, scientific collaborations, epidemiological and historical data and data on many other topics.
Flowmap.gl is the open-source flow map drawing library which FlowmapBlue is based on. Flowmap.gl is now a part of the vis.gl framework which belongs to the OpenJS foundation.
Geographic networks often contain large numbers of connections and displaying them on a map presents significant challenges both in terms of the rendering performance and the map readability. Instead of always showing the data in full detail, Flowmap.gl applies automated clustering to render useful and readable summaries which constantly adapt to the map viewport. In this talk, we'll discuss the motivation behind this approach and describe the design decisions and the inner workings of the tool.
Managing and building an open-source developers community can be time-consuming, labor-intensive, and most of the time, it is a hit and trial for what to do and what not to do. This can bring an unavoidable churn in your community, with developers leaving the community because of not finding enough value resulting in grief and frustration. In this session, we’ll discuss how to build successful open source communities and programs that resulted in a highly engaged community of developers that are passionate about open source, contributing PRs, and sharing valuable and authentic feedback that made our open source project successful.
vis.gl is a suite of composable, interoperable open source geospatial visualization frameworks (GPU powered) centered around deck.gl. During the last 4 years vis.gl has played an essential role in the development of geospatial applications during the last 4 years.
With close to 100K daily downloads from npm, it’s widely used today in many areas and industries: from academics teams, to enterprise companies like Uber, Foursquare, CARTO, Google or Amazon.
The open governance of vis.gl has guaranteed the evolution and maintenance of the framework, the project joined the OpenJS foundation in 2022 with the main goal of re-enforcing the open evolution of the project.
During this talk we’ll do a quick and high level introduction of the most important frameworks that belong to this suite (deck.gl, kepler.gl, loaders.gl, etc.), we’ll do an update of the most important features and milestones achieved in the last year, and we’ll share the strategy and direction for the next year.
Traffic signs are a key feature for navigating and managing traffic safely, affecting all of us on a daily basis. However, traffic sign datasets are lacking on open government data portals as well as OpenStreetMap (OSM).
Mapillary’s computer vision capabilities can extract more than 1,500 classes of traffic signs globally from street-level imagery. Generated traffic signs are available on iD Editor, Rapid and JOSM Mapillary plugin to enrich OpenStreetMap data.
Our team wanted to know how the accuracy of traffic signs detected by Mapillary compared with the reality on the ground (the ground truth). To answer this question we collected more than thousands ground truth data in San Francisco and used this information to produce the recall, precision, and positional accuracy of our machined generated traffic sign data. This provided some interesting insights in OpenStreetMap and the level of completeness and gaps of that dataset.
In this talk, we will cover Mapillary’s traffic sign extraction capabilities, Mapillary generated traffic sign data against ground truth data and OSM’s traffic sign coverage in San Francisco’s downtown. We will be also addressing how data quality can be improved using various data collection techniques and the role of post-processing with Structure from Motion and control points annotations.
Gisquick (https://gisquick.org/) is an open-source platform for publishing GIS projects on the web. A GIS project is defined by a QGIS project file including data sources (files, databases, even virtual layers) and symbology defined in the QGIS desktop application using the styling tool.
With the help of the Gisquick plugin for QGIS, it is possible to upload the data to the Gisquick server and host the map.
Gisquick is a fully featured hosting platform, where the project administrator can fine-tune web publishing attributes, set predefined scales, bounds, or visibility. Also group permissions on the project level as well as layer level (query, edit, export) may be defined. Vector data - geometry and attributes - can be edited directly on the web.
Interface between the frontend and backend is based on open standards (OGC WMS and WFS). The mapping application has standard components from the GIS point of view: decent layer switcher, attribute table, zoomable map, printing tool (based on QGIS templates), and customizable feature-detail form.
All this can be tested on our demo platform https://demo.gisquick.org/ - but you can also make your own deployment via Docker images. Gisquick is open-source software published under the GNU GPL.
In the presentation, we are going to present various features of Gisquick and show practical examples and discuss technologies used for its development.
In this talk we introduce a European initiative with global effects that aims to support the uptake of Earth Observation (EO) data products and services by increasing European capability to generate timely, accurate, disaggregated, people-centred, accessible and user-friendly environmental information based on EO data. The initiative - Open Earth Monitor Cyberinfrastructure - is following a well defined workflow:
(1) Identify gaps and needs analysis : finding out what are the bottlenecks of data platforms together with stakeholders;
(2) Use open source EO computing engine : integrating EO with in-situ data to obtain improved geospatial data services and products;
(3) Build better data portals: harmonise, bridge and improve existing open source platforms;
Make data platforms FAIR: improve accessibility of data with open source licences and capacity building;
(4) Serve concrete goals: all Open Earth Monitor activities are centred around pre-defined use cases with various stakeholders.
We do not plan to reinvent the wheel, therefore all our efforts will focus on improving existing open source solutions and other initiatives, such as: OpenEO.org, Geopedia.world, GlobalEarthMonitor.eu, EarthSystemDataLab.net, OpenLandMap.org, EcoDataCube.eu., LifeWatch.eu, XCUB and EuroDataCube.com. Our developments will materialise in a series of monitoring tools at European as well as global level in various fields: forestry, natural hazards, biodiversity, crop monitoring etc.
In the context of Open Earth Monitor, Cyberinfrastructure is defined as the coordinated aggregate of software, hardware, human expertise and other technologies required to support current and future discoveries in science and engineering, enabling relevant integration of often disparate resources to provide an useful and usable framework for research, discovery and decision-making characterised by broad access and "end-to-end" coordination.
Open Earth Monitor Cyberinfrastructure has received funding from the European Union's Horizon Europe research and innovation programme under grant agreement No. 101059548. (HORIZON-CL6-2021-GOVERNANCE-01).
Knowledge graphs help contextualize and enrich data by modeling and querying relationships between entities using a graph database and have been successfully used alongside geospatial data and map tooling for use cases such as logistics and supply chain analysis, fraud detection, investigations, suitability analysis, real estate, and data journalism. In this presentation, we examine how the open-source Neo4j graph database can be used with QGIS and Python for making sense of geospatial data using graph algorithms and graph data visualization alongside maps while combining data from OpenStreetMap, cadastral data, and public data portals to find insights that address the use cases mentioned above. We introduce an open-source QGIS plugin for Neo4j called Quantum Graph that enables data loading, graph analytics, and visualization from within QGIS: https://github.com/johnymontana/quantum-graph
MapServer, a founding OSGeo projects, has been powering mapping systems since the mid 1990s. This talk gives an overview of the many features of MapServer that have been developed over the past 25 years, with a focus on advanced functionality that is not well-known as they deserve.
Features will be shown using sample Mapfiles - the configuration files used by MapServer. Examples will include advanced symbology, special layer types such as graticules, charts, and contours, displaying data from S3 buckets, and more!
Mergin Maps has become a popular open-source GIS platform for collecting, managing and sharing geospatial data. In the past year, we have introduced several new features and improvements to the platform. Our goal is to provide a flexible and powerful GIS solution that is accessible to users of all levels, from seasoned professionals to those just getting started. In this talk, we will highlight the latest developments and demonstrate how they can benefit users in various fields.
One of the significant updates is the introduction of workspaces, which allows users to organize their projects, data, and users in a hierarchical structure. This new feature streamlines the management of multiple projects and simplifies the process of adding and removing users.
Another update is the implementation of tracking, which enables users to collect and visualize location data. This feature is particularly useful for tracking vehicles, equipment, and personnel in the field, and can be customized to include various attributes.
Finally, we will discuss the Mergin Maps roadmap for the future, including plans for new features, enhanced integrations and community-driven development. We believe these changes will make Mergin Maps more accessible and user-friendly for everyone, regardless of their level of experience.
Whether you are a seasoned GIS professional or new to the world of geospatial data, this talk will provide valuable insights into the latest developments in Mergin Maps and its potential for your work.
Road Surface Inspector is a system developed by IT34 with the purpose of speeding up the process of road damage registration by using deep learning. The time consuming process of inspection and registration of road damage is reduced significantly by using our Road Scanner Inspector app that can be placed in the windshield of any vehicle. The app records a video and gps coordinates, which are later processed in order to find different types of damage - potholes, cracks, damaged markings using deep learning.
The system can also detect other types of assets such as traffic signs, traffic lights, manholes and others that can be used in fx digitalization tasks.
The results of the image analysis are presented on a webgis portal as heatmaps presenting the condition of the road in the areas that were inspected using the app. The heatmaps are further used by the decision makers in order to prioritize the road maintenance work.
While using the app, Gps logs are built in realtime based on the positions sent by the phone while driving. These are further used for street inspection documentation.
Open source components.
Postgres + Postgis for storing the data and for geometry based analysis
PyTorch and Yolo7 for deep learning
OpenLayers for visualizing the images/detection results as rasters in webgis
Geoserver for publishing data as WMS/WFS
QGis as an external visualization tool for the data
Privacy aware Content Managment System (CMS) operators don't let their viewers accept cookies from an external map provider. But creating a map used to require specialized GIS knowledge and hosting a map server is not everyone's cup of tea.
This talk explains how non-experts can serve a map based on OpenStreetMap vector tiles from a CMS. A MapLibre GL JS based Wordpress plugin displaying a self-hosted PMTiles dataset is shown as an example.
Are you interested in open geospatial tech for humanitarian purposes? Have you ever wondered who the people behind the geospatial technologies are? The collective brains? In this talk, we will tap into the power of the tech collective at Humanitarian OpenStreetMap Team, share our experience, excite you about joining the collective and get some hands-on input from YOU!
Meet two members of the Humanitarian OpenStreetMap Team (HOT) - Petya & Synne. We are a global team that operates with four regional Open Mapping Hubs: https://www.hotosm.org/hubs/. In developing and improving open geospatial tech for humanitarian purposes, our vision is to creatively meet the needs of the communities through collective, community-centered efforts. Our mission? To amplify community-led innovation for impact through diversity, creativity & passion!
Some of the stories we will share will be about our experiences and lessons learnt on collective projects and products (https://github.com/hotosm/) ranging from the HOT Tasking Manager collective , collaborating with Kathmandu Living Labs (KLL) in Nepal, to development of a Field Mapping Tasking Manager (FMTM). We will also share some of the boldest regional activities, including OpenStreetMap (OSM) Hackfest in Asia Pacific and the Ideas Lab in Eastern and Southern Africa.
You will also find out how YOU can get involved by contributing to open geospatial tech. Expect a short participatory exercise [the collective brains/ power of collective intelligence] during this session!
United Nations Development Programme (UNDP) is a United Nations agency tasked with helping countries eliminate poverty and achieve sustainable economic growth and human development.
Recent advances in technology and information management have resulted in large quantities of data being available to support improved data driven decision making across the organization. In this context, UNDP has developed a corporate data strategy to accelerate its transformation into a data-driven organisation. Geo-spatial data is included in this strategy and plays an important role in the organization. However, the large scale adoption and integration of geo-spatial data was obstructed in the past by issues related to data accessibility (silos located in various country offices), interoperability as well as sub-optimal hard and soft infrastructure or know-how.
All this issues have been addressed recently, when UNDP SGD integration started developing a geospatial hub - GeoHub - to provide geospatial data visualisation and analytical tools to UNDP staff and policymakers.
UNDP GeoHub is a repository of a wide array of data sets of the most recent time span available at your fingertips! It is a centralized ecosystem of geospatial data and services to support development policymakers. It allows users to search and visualise datasets, compute dynamic statistics and download the data. In addition, GeoHub provides a feature to share their maps with the community easily. With our repository, you can also upload to share your valuable data to share with the community! It connects geospatial knowledge and know-how across the organization to enhance evidence-based decision-making with relevant data-led insights.
Geohub ecosystem consists of sveltekit & maplibre based frontend web applications and various FOSS4G software in the backend side. PostgreSQL/PostGIS, titiler, pg_tileserv and martin are deployed in Azure Kubernetes (AKS) to provide advanced visualisation and analysis for users. All source code is published in Github with an open-source license.
One of the primary motivations for the Open Mapping Hub Asia-Pacific to increase the quantity and quality of OpenStreetMap (OSM) data in the region is the region's high exposure to multiple types of hazards.
Apart from assisting response efforts following a disaster event by providing access to critical geospatial information, the hub aims to ensure that OSM data is already available in high-risk areas, even before a disaster occurs, to be used in critical anticipatory action such as developing early warning systems and mitigation plans. It is critical to have a systematic method for determining the OSM mapping requirements in these disaster hotspots.
Although some tools separately assess the Completeness of OSM Data and the Disaster Risk Level of a location, a new tool that combines these assessments is required to highlight the areas that should be prioritized for mapping in OSM.
The Open Mapping Hub Asia-Pacific created a data-driven method for determining which areas in OSM disaster mapping should be prioritized. The resulting method is deployed as a QGIS plug-in and distributed to OSM communities for offline assessments to identify disaster-prone areas that have not yet been mapped in OSM.
Europe is a world leader in Earth Observation (EO) and climate change studies. An outstanding example is Copernicus, the most ambitious EO programme worldwide, which in addition to being an independent system is also a strong component of the Group on Earth Observation (GEO), an intergovernmental partnership aiming to improve the availability, access and use of open EO to support policy and decision making in a wide range of sectors.
Since 2005, the Global Earth Observation System of System (GEOSS) has been a key initiative by GEO to integrate platforms and connect existing infrastructures using common standards for sharing and using digital resources. Europe is delivering a regional contribution to GEO, named EuroGEO, by covering the last mile of the EO value chain. However, this regional node lacks the effective interoperability needed to implement a European ecosystem to fully support the policy cycle.
To fill this gap, the development of a sustainable EuroGEOSS ecosystem connecting many European assets including data, sensor networks, analytical methods and models, computing infrastructures, products and services that support European objectives (i.e. a EuroGEOSS ecosystem), is of a vital importance in the evolution of the initiative.
The purpose of this talk is to present the rationale and the development status of a EuroGEOSS prototype, that the European Commission’s Joint Research Centre is conceptualizing.
Starting with the analysis of use cases with the highest European policy priority, five of them were identified as the prominent ones to be replicated. Along with the replication of use cases, a monitoring framework of issues and gaps identified in the life cycle will be populated meanwhile.
The EuroGEOSS prototype architecture will implement the following patterns: a) Portal and Single Sign On; b) Meta catalogue of the services (data, models, infrastructures, etc.); c) High flexibility and modularity level; d) Adoption of the Machine Learning operation (MLOps) methodology.
The EuroGEOSS ecosystem is not conceived as another platform. It will rather be a virtual platform leveraging on: a) open sources and open interoperability standards (normative and de facto); b) interconnection of novel technologies; c) inclusion of relevant European communities such as those around EuroGEO and INSPIRE; d) Scalable interoperable infrastructures: CREODIAS, OpenEO, etc.
The development of a EuroGEOSS prototype will last until the end of 2024, documenting the status of gaps, challenges in the available data and infrastructure, as well as assisting a future scenario and business model and a possible operationalization.
When it comes to styling of geodata many tools have their own solution: SLD, QGIS-Styles, OpenLayers-Styles, Leaflet, …
But what to do if you need to share the same style across different formats?
GeoStyler brings the solution. With its standalone parsers, nearly any (layer based) style can be converted from one format to another - from SLD to OpenLayers, QGIS, Mapfile, and vice versa.
On top of this, GeoStyler offers a library of React UI elements to easily create styles in your own WebGIS.
This talk will give an overview of possible use cases for GeoStyler, its latest developments such as the new layout and the support for expressions, as well as past and upcoming community events.
Welcome to the Open Source Geospatial Foundation, proud hosts of FOSS4G, and advocate for free and open source geospatial software everywhere. This is a call out to open source software developers; please join OSGeo and help us help you!
Join OSGeo today:
- Even just listing your project on the osgeo.org website is a great first step. Help us promote your technology so users can discover and enjoy your software.
- The OSGeo “community program” gives project teams a chance to join the foundation with an emphasis on supporting innovation and new projects. The foundation provides some direct support, assistance along with endorsement and recognition from our board.
- For established projects please join our “incubation program” to be recognized for excellence and as a full OSGeo committee.
Unlike other foundations OSGeo does not require that you give up or transfer any Intellectual Property; we simply ask that you be spatial, open-source, and open to participation.
This presentation gives clear instructions on how to join OSGeo, and representatives from recent successful projects will be on hand to answer your questions.
The mobile application QField is based on QGIS and allows fieldwork to be carried out efficiently based on QGIS projects, offline or online. Developments in recent months have added additional functions to the application that are useful for fieldwork. Examples are used to present the most important new features. Discover the most recent features like 3D-layers handling, printing of reports and atlases, elevation profiling of terrain and layers, multi-column support in feature form, azimuth values in the measuring tool, locked screen mode, the QR-code reader, stakeout functionalities, the official release of the iOS version and many more.
Since April 2022 I've been manipulating projected digital maps in collaboration with improvising musicians, dancers, and spoken word artists across Europe and North America. Constraining my project to use only web mapping technologies, "A Synesthete's Atlas" is a curious mutation of expanded cinema, applying strategies from experimental film & animation, color theory, the Light and Space movement, and concrete poetry to geography.
I'll present Carto-OSC, an assemblage of open source libraries, data, and protocols, plus 1000+ lines of JavaScript that integrates it all into a touch-surface interface. I'll discuss my motivations and use of the OSC protocol to control the manipulations, offer aesthetic observations, and present video excerpts of previous performances.
PostGIS supports geometries with a Z dimension and geometries with measure values, but there are not a lot of examples of both of these being used together. One use case is the analysis of airplane tracks which requires both - that is to say every vertex has an altitude and a timestamp. This talk will show how to convert ADS-B data transmitted from aircraft into these LINESTRINGZM geometries, which can then be analyzed as trajectories using native PostGIS functions.
With this dataset and this tool, we can do things such as determine anomalous changes in an aircraft's velocity or altitude and find the exact point in time at which two aircraft came closest to one another.
I will cover how to use Python and PostgreSQL's PL/Python language extension to import the data and QGIS to render the data, but the analysis will be be done in SQL.
How do you create a near-real-time source of 3D geospatial data from around the world?
The French Institute of Cartography and start-up Extra are collaborating to develop a decentralized protocol for this purpose. The Circum protocol will merge LiDAR datasets from various providers, sell this data source to consumers, and redistribute the value back to the original providers.
Circum uses blockchain technology and 3D surface reconstruction algorithms to carry out its mission. Learn about the protocol’s key mechanisms with the team at this conference.
3geonames.org is a free api for fast reverse geocoding, using a new technique of locality-preserving hashing of 2d/3d spatial points to 1d integers via a combination of Hilbert curve and bit interlacing. This talk expands on the use-case and the performance/accuracy advantages of this technique.
The Hydro Network-Linked Data Index (NLDI) is a system that can index data to a hydrographic network and offers a RESTful web service to discover indexed information upstream and downstream of arbitrary points along the stream network. This allows users to search for and retrieve geospatial representations of stream flowlines, catchments, and relevant water monitoring locations contributed by the water data community - without downloading the national dataset or establishing links themselves.
This is done by data providers publishing open information about the locations of their data within the context of the U.S. stream network. Data linked to the NLDI includes various federal, state and local water infrastructure features and water quantity and quality monitoring locations. The NLDI is being developed as an open source project and welcomes contributions to both its code and indexed data, with the main implementation currently being maintained by the U.S. Geological Survey.
The community of practice surrounding the NLDI extends to R and python developers working on clients that allow scientists to quickly retrieve data relevant for specific hydrologic analyses. As the NLDI community grows, a similar concept could be applied at a global scale, facilitating the development of downstream tools and applications.
While the NLDI is limited to the US, global work would be possible by leveraging global stream network datasets such as MERIT-Hydro. A proof-of-concept global River Runner allowing discovery of the flowpath downstream of arbitrary points anywhere on Earth has already been implemented using MERIT-Hydro and OGC-API Processes in pygeoapi. This session includes demonstrations of the NLDI and the global River Runner.
In a context of digital transition and the increasing availability of urban data, Rennes Métropole wishes to better equip its decisions and public policies on the basis of data and cooperation.
Ultimately, the goal is to :
- Promote cooperation and the contribution of the actors of the territory, in particular the citizens
- "Enlighten" public decisions and policies, in particular the democratic, ecological and energy transition projects carried out by Rennes Métropole.
Issues of transparency, public service efficiency and cost control are also sought.
The metropolitan cooperation platform that is currently developed will consist of one or more tools based on the digital twin intended to equip public decisions and policies on the basis of data and cooperation.
The platform is developed partly on VC Map which is an Open-Source JavaScript framework and API for building dynamic and interactive maps on the web. It can display 2D data, oblique imagery and massive 3D data including terrain data, vector data, mesh models, and point clouds making it easy for users to explore and interact with the data in an integrated and high-performance map application. VC Map is built upon open, proven, and reliable GIS and web technologies such as OpenLayers and Cesium for the visualization of 2D and 3D geo-data.
A particular effort was made on the design in order to offer users, mainly citizens of Rennes Metropole, a pleasant user experience that allows an exploration of the development projects of the metropole in 2D and 3D.
We will present the cooperation platform through three use cases of interest for Rennes Metropole :
Solar Cadaster : Simulation of the photovoltaic production potential of the roofs and comparison with the energy consumption of the residents, the costs and the capacity of the network.
Linear transport systems : Mediation (including visualization) and consultation with citizens and communities for the implementation of a linear transport infrastructure
Exposure to electromagnetic waves : Visualization of exposure levels to electromagnetic waves (simulations and real and real measured values) as well as objects (radioelectric relays and sensors) on the territory of the City of Rennes.
Use case for the implementation of a platform that supports data that contributes to the publication and management of Digital Twins, based on the use of MapLibre as a web viewer and at the same time consuming information from different geospatial sources, including Mesh, Raster, DEM; and near real time data sources such as OneBusWay or OpenTripPlanner based on GTFS formats, for the comparison and analysis of information.
GIS instructors at an American technical college have created a five-course certificate in GeoAI. The first cohort of undergraduate students has completed the degree requirements two years later. This presentation will discuss the formation for the degree, the courses, and the resulting graduates. The presentation will discuss the learning outcomes for the degree and individual AI and machine learning GIS courses.
In this talk, I'll share some practical tips and tricks for managing an enterprise GIS workflow with QGIS and PostGIS. I'll showcase some real-world examples to highlight the benefits of using a centralized spatial database to manage GIS data, and I'll walk through the steps to set up a QGIS project for creating, updating, and deleting data directly from QGIS.
My goal is to help organizations that are planning to set up a PostGIS-powered QGIS workflow and are looking for innovative ways to maximise the benefits of the joint powerhouse of QGIS and PostGIS.
As we dive deeper, I'll explore some of the key technical aspects of using QGIS and PostGIS for enterprise GIS. I'll share some tips for configuring and integrating the tools, and showcase how to set up an easily accessible end-user workflow for creating and editing data in QGIS using QGIS forms.
Throughout the talk, I'll also share some stories from different projects to illustrate how these tips and tricks have been successfully applied in practice. I will do my best to ensure that you’ll leave the talk with an understanding of the benefits of using QGIS and PostGIS as part of their Enterprise GIS workflows.
Whether you're a GIS professional, team leader / project manager or anyone seeking to optimize their GIS data management, this talk will provide valuable insights and practical advice for optimizing your GIS data management. Join me as we explore the power of open source tools for enterprise GIS!
Modeling of traffic, especially public, has its own specific features in the settlements. For today, there are some commercial solutions initiated and maintained by city administration and transportation data providers. Open source and easy to use timetable tool could be practical either for specialists in the city development and for daily commuting around the city. In the Open Source QGIS protocol repositories, transport solutions with public transport schedule calculations (bus, tram, trolleybus) are still rare, some examples could be noted, for instance, "Site Schedule Optimization" (https://plugins.qgis.org/plugins/site_schedule_optimization/)
Developed plugin called “Public transport schedule” and aimed for use for the settlements of different size and transport networks of varying complexity. Such a solution could be suitable for small cities that do not have means or technical resources to implement commercial products. Basic knowledge of QGIS is required to practice the plugin.
Initial data for the plugin are a linear layer of the route and a point set of stations (stops) of public transport. Plugin was developed using QGIS processing elements, Shapely and Numpy libraries. The plugin uses city data from the Yandex and is designed for the conditions of a large metropolis of Moscow, interface languages are Russian and English. A special case of public transport routes is considered: throughout the route, the movement of vehicles should be right-hand and two-way. The procedure accepts for input two vector layers - point (stations) and linear (route), as well as the data specified in the plugin window. Stations can be located along the line in any order, at no more than 50 meters from the line. The line layer of the route should consist of a single geometry object of type Line (LineString), for the reason that using the MultiLine (MultiLineString) geometry format, the algorithm will not be able clearly determine the side of the station location relative to the line. Plugin’s workflow consists of successive stages: (1) data input, (2) additional tools connection, (3) validation of user input data, (4) geometry processing of vector layers, (5) choosing the start station (user input), (6) attributes manipulating, (7) scheduling, (8) data output.
The first stage of user input includes data collecting, namely: point layer of stations; linear layer of the route; parking time at stations (second); turnaround time at end stations (second); total number of vehicles, serving the route; maximum speed of movement (km per hour). At the second stage additional calculation tools were used: traffic activity schedule on weekdays, number of buses serving the Moscow transport system at a certain moment of time). Some needed data are schedule of traffic activity on the weekend in Moscow (https://yandex.ru/company/researches/2020/moscow/trolltrambus/) and a model of acceleration/deceleration of the vehicle (https://infopedia.su/12x37b7.html/). At the third stage the data entered by the user have to be validated. Geometry processing of vector layers at the fourth stage consists in performing the following operations: working with map projection; calculation of projections of station points on the line (bus stop places); finding terminal stations; adjustment of the line length according to the position of the end stations. At the fifth stage, user chooses the start station (one of the two terminal ones): depending on the start station selected by the user, the direction of the linear layer of the route changes or remains unchanged. The attributes manipulating of the sixth stage consists in attributing projection points of stations and vehicle positions while moving along the side of movement (straight or backward in relation to the starting station). To build a model of vehicle movement, plugin creates all points of a possible position inside the processed line of a single vehicle during its operation, then attributes these points in order and according to the time of their passage.
Subsequent scheduling includes the following actions: calculating the time for changing the number of buses on the line, estimating the time to leave the line, or completing the run of a single vehicle; building a universal model of track along the line. In order to complete the work on the line, a single bus should arrive at the start station. Physical characteristics of train movement are taken from the JSON file (https://infopedia.su/12x37b7.html/). When a certain distance is reached, equal to the braking distance of the bus to the next stop, the bus starts to slow down; extrapolation of the universal model of movement (finding the moments of the beginning and end of the running of individual buses). Scheduling is drawn up on the data from Yandex analysis (https://yandex.ru/company/researches/2020/moscow/trolltrambus/) of traffic activity on weekdays in Moscow. Plugin predicts integer number of "cycles" a single vehicle can pass during its plying along the line. The final stage is data output: two tables of timetable for the vehicles arrival at the stations on opposite sides of the movement.
Plugin’s output is a schedule table for the time the bus moves along the route with an accuracy of up to a second. In the paper we have shown the first results of plugin’s development which is then planned to be presented in a diploma at the GIS Department of the Federal State Budget Educational Institution of Higher Education «MIREA — Russian Technological University» (RTU MIREA) and put into the QGIS plugin repository for the open use.
Our talk presents an initiative that works to develop an open, interactive, user intuitive platform for a constantly updated, comprehensive and detailed overview of the dynamic environment of the open source digital infrastructure for geospatial data storage, processing and visualisation systems. OSS4gEO is designed as a repository that functions as an extended metadata catalogue, curated by the community and a tool for metrics computation, visualisation, ecosystem statistical analysis and reporting.
The initial development of the Open Source for Geospatial Software Resources platform builds on previous extensive work started in 2016 that has materialised into a pioneering overview of open source solutions for geospatial, voluntarily updated by the team. Starting in 2023, OSS4gEO has become a part of a wider ESA EO Open Innovation initiative to actively support and contribute to the EO and geospatial open source community and it is intended as a seed action to better understand, represent and harvest the geospatial open source ecosystem.
There are 3 main objectives that OSS4gEO aims to achieves:
(1) It aims to offer an informed and as complete as possible overview of the open source for geospatial and EO ecosystem, together with various capabilities of filtering and visualisations, within the platform as well as technical solutions to programmatically access and extract data from the database (APIs) to use in any purpose, including commercial;
(2) It aims to provide guidance through the complexity of the geospatial ecosystem so that one can choose the best solutions, while understanding their sustainability, technical and legal interoperability and all the dependencies levels;
(3) It aims to serve as a community building, a promoting and maintaining platform for new and innovative open source solutions for EO and geospatial, developed within various projects, research centres, small or large companies, universities or through individual initiatives.
Our talk will outline the OSS4gEO initiative as a community-led, bottom-up initiative, highlight current and future developments and co-development activities and introduce the wider ESA EO Open Innovation context.
GeoServer deployments in the cloud and kubernetes are becoming the norm, while the amount of data published is also growing, both in terms of layers and size of data. As a result, the need for scaling up is becoming more and more common.
This presentation covers GeoServer clustering approaches, comparing the available options and their suitability to different environments. We will cover: * Managing the GeoServer configuration, stable configuration with planned upgrades versus dynamic runtime changes. * Deployment options (monolithic, separate tiling, microservice oriented) * Dynamic configuration clustering with JMS, external database storage, and distributed memory.
Attend this presentation to get an update on GeoServer cloud and clustering options, and pick the option that is the best match for your specific use case.
PDAL is Point Data Abstraction Library. It is a C/C++ open source library and applications for translating and processing point cloud data. It is not limited to LiDAR data, although the focus and impetus for many of the tools in the library have their origins in LiDAR. PDAL allows you to compose operations on point clouds into pipelines of stages. These pipelines can be written in a declarative JSON syntax or constructed using the available API. This talk will focus on the current state of the PDAL Pointcloud processing library and related projects such as COPC and Entwine, for pointcloud processing. Coverage of the most common filters, readers and writers along with some general introduction on the library, coverage of processing models, language bindings and command line based batch processing. First part will be covering new features for current users. Some discussion of installation method including Docker, binaries from package repositories, and Conda packaging. For more info see https://pdal.io
GTFS is stands for General Transit Feed Specification, which is developed by Google and used for describing schedules of public transpotation. A bunch of dataset is distributed in the world and GTFS includes geospatial information - stops and routes. To utilize such intresting data, we have developed GTFS-GO - QGIS plugin to process GTFS. You can translate GTFS to GIS data and visualize them by GTFS-GO. The plugin can be used for analyzing public transportaion by aggregating traffic frequencies on each stop or route. In this talk, you can see how GTFS is visualized or analyzed by using GTFS-GO on QGIS.
Climate change’s impact on public transportation tends to focus on improving transit infrastructure to reduce stoppages. While this is important, it does not take into account the effect it has on communities, often already underserved, that rely on the transit system. As part of The Opportunity Project’s Building Climate Change Resilience Through Public Transit sprint, our team at Data Clinic set out to develop an open source, user-friendly, and scalable tool to communicate intersectional risks faced by transit infrastructure and community access at the local level. This solution was inspired by both the event, and user research with key stakeholders in transit agencies, academia, and community organizations.
In this presentation, we will demonstrate TREC: Transit Resiliency for Essential Commuting, and expose key decisions that resulted in a geospatial solution designed for wide audiences, and geographic and data scalability. TREC’s transit stop-level insights can become crucial tools for transit planners and community organizations to prioritize and advocate for infrastructure improvements that take community effects into account.
Focused initially on two locations- one small (Hampton Roads, Virginia) and one large (New York City) transit system, each station is treated as a destination providing access to essential services during localized climate change events. In this MVP, we employ flooding as our climate scenario, the event most cited as recurring and disruptive by our stakeholders.
Using OpenStreetMap to calculate walksheds around each station obtained from GTFS data, we categorize importance in accessing essential services such as hospitals and jobs around a transit stop. Layered onto this, we bin current flood risk for each station using the prevalence of buildings with moderate- to extreme high-risk of flooding according to open data, and provide polygons representing projected flood risk in 2050.
While we built the TREC UI to maximize accessibility of this contextualized data to multiple stakeholders, we also seek to optimize usability of the repo to allow tech-mature transit planners to adopt the tool internally and incorporate their proprietary fine-grained data. Further, we are committed to expanding the functionality of TREC according to user feedback.
The threat of climate change disrupting daily life on a recurring basis, beyond large-scale disasters, continues to grow. With the help of this tool, we hope to democratize relevant data, inspire the open publication of localized geospatial data related to climate change, and enable human-centered decisionmaking through a multidimensional lens.
The goal of this presentation is to give an overview of the different options available for deploying a GeoServer configuration to different environments. In addition to the common data_dir folder deployment option, we will explore the possibilities offered by existing extensions and by the REST API, including different client libraries around it. We will also discuss the advantages that can be brought by Terraform for this use case.
pgRouting, a PostGIS extension containing algorithms and tools for working with graph data, has become a highly flexible member of the FOSS routing engine family. In this talk, I want to demonstrate just how flexible it can be by showing how routable networks (called 'topologies' in pgRouting) can be made editable.
I will take the audience from theoretical conception of editable topologies (how can edits, insertions and deletions be handled in PostGIS?) through its implementation. Finally, I will end with a demonstration of a fully editable topology in a web mapping application based on a real world example using OpenStreetMap data.
The Web has an increasing number of web applications being developed to freely provide their information and is a hub for open data publishing. For this to happen as a self-sustained ecosystem, data must be findable, accessible, interoperable, and reusable to both humans and machines across the wider web. This session delves into Web Best Practices for publishing data using open source and standards-based solutions.
The geoconnex.us project is about providing technical infrastructure and guidance to create an open, community-contribution model for a knowledge graph linking hydrologic features in the United States as an implementation of Internet of Water principles. This knowledge graph can be leveraged to create a wide array of information products to answer innumerable water-related questions.
Implementation has two parts: persistently identified real world objects and organizational monitoring locations that collect data about them. Both must be published to the Web using persistent URIs and communicated with common linked data semantics in order for a knowledge graph to be constructed.
The Internet of Water Coalition supports the first part with a Permanent Identifier Service and reference hydrologic reference features (e.g. watersheds, monitoring locations, dams, bridges, etc.) within the US.
In support of the second part, geoconnex.us takes advantage of pygeoapi using the OGC API - Features standard to publish structured metadata resources about individual hydrologic objects and the data about them. pygeoapi supports extending this standard by incorporating domain-specific structured data into the HTML format at the feature level, and allowing for external HTTP URI identification. In addition, pygeoapi’s flexible plugin architecture enables for custom integration and processes. This means that individual features from various sources can have structured, standardized metadata harvested by search engines and assembled into a useful knowledge graph.
This spatial feature-based linked data architecture enables data interoperability between independent organizations who hold information about the same real world thing without centralizing data infrastructure - answering important questions like, “Who is collecting water data about my local stream and its tributaries?” or “What data do we have about water upstream and downstream of East Palestine, Pennsylvania?”
At the National Land Survey of Finland (NLS) we are developing multiple QGIS plugins, and we needed a way to share the common code and break the components to smaller independent plugins while still providing a good developer experience.
One of the main issues when sharing library code between different QGIS plugins is the runtime environment uncertainty. Since Python import machinery is not easily configurable to support multiple versions of dependencies (like nested node_modules in nodejs-world), the runtime is limited by default to a single version of a library, and later access to the same module is cached. This limits the version available to all plugins in a single QGIS session to the code that is first run, which makes sharing code difficult, especially when breaking API changes are necessary to the dependency library code.
At NLS we developed tooling to work around these limitations, which improves the developer experience and allows sharing of common QGIS plugin code easily via standard Python libraries. Tool provides a streamlined developer workflow and necessities like typing and IDE helpers, and a way to package a plugin that depends on other standard Python libraries.
Development environment for a QGIS plugin can be initialized simply by using a virtual environment, installing the dependencies and launching QGIS with the plugin and its dependencies fully setup. This works with bootstrap code passed on the command line, which will provide QGIS access to the virtual environment, setups the plugin from the environment with access to any library dependencies. Tool also provides a debugger session and could also provide for example hot reload signals for the plugin when code is changed. This provides a quicker and easier feedback cycle for the developer and simplifies the workflows when developing QGIS plugins.
Runtime dependencies are reorganized at build-time to be imported for a sub-package of the plugin, so only the exact packaged version of a dependency is used at runtime. This works by rewriting external library dependency import statements in the source code. Tool also generates the metadata.txt file in a way that is compatible with standard Python packaging tools, for example setuptools. This allows easily sharing the same code both as Python library and as a QGIS plugin.
MobiDataLab is the EU-funded lab for prototyping new mobility data sharing solutions.
Our aim is to foster data sharing in the transport sector, providing mobility organising
authorities with recommendations on how to improve the value of their data,
contributing to the development of open tools in the cloud, and organising hackathons
aiming to find innovative solutions to concrete mobility problems.
Started in 2022, the project investigated mobility data and services and did grown an
open knowledge base about mobility data as one of the four main pillars of the project.
With the realization of tools and the combination of data and services in the Transport Cloud,
which is the second pillar of the project, a representative set of technical
"mobility data sharing enablers" has been grown.
In the second half of the project, these assets are being provided to the public.
The Virtual and Living Labs will host environments for mobility data stakeholders
to explore the state of the art for data, services and their interaction to solve
mobility data challenges. All aligned with the FAIR statement - making data and services
findable, accessible, interoperable and reusable.
The challenges are mainly based on a broad set of use-cases, defined by the core project group,
the reference group and external stakeholders. These challenges are the core of the Livind and
Virtual Labs, where participants building solutions for the given challenges and exploring new
opportunities with the shared mobility data and services.
With the feedback of the labs, our partners, the reference group and external stakeholders
- mobility data providers from public and private sector, municipalities,
governmental institutions, start-up communities and stakeholders from research and industry,
the project will make challenges transparent and remove barriers for data sharing.
Since the project started in February 2021, we will present our achievements
provide an outlook on the last mile of the project, where we are bringing the
tools on the road.
Further information on the project is available via https://mobidatalab.eu and https://github.com/mobidatalab .
MobiDataLab is funded by the EU under the H2020 Research and Innovation Programme (grant agreement No 101006879).
In Norway we now get more up-to-date maps for land resource map (AR5), because the domain experts on agriculture in the municipalities in Norway have got access to a easy to use client. This system includes a simple web browser client and a database built on Postgis Topology.
In this talk we will focus on, what is it with Postgis Topology that makes it easier to build user friendly and secure tools for updating of land resource maps like AR5. We will also say a couple of words about advantages related to traceability and data security, when using Postgis Topology.
In another project, where we do a lot ST_Intersection and ST_Diff on many big Simple Feature layers that covers all of Norway, we have been struggling with Topology exceptions, wrong results and performance for years. Last two years we also tested JTS OverlayNG, but we still had problems. This year we are switching to Postgis Topology and tests so far are very promising. We also take a glance on this project here in this talk.
A Postgis Topology database modell has normalised the data related to borders and surfaces as opposed to Simple Feature where this is not the case. Simple Feature database modell may be compared to not using foreign keys between students and classes in a database model, but just using a standard spreadsheet model where each student name are duplicated in each class they attend.
URL’s that relate this talk
https://gitlab.com/nibioopensource/pgtopo_update_gui
https://gitlab.com/nibioopensource/pgtopo_update_rest
https://gitlab.com/nibioopensource/pgtopo_update_sql
https://gitlab.com/nibioopensource/resolve-overlap-and-gap
The LH Urban Digital Twin Platform is a comprehensive solution for new town planning and development that utilizes open source digital twin technology. The platform combines real-world data with spatial information context to offer a three-dimensional sharing/collaboration integration support system.
Developers will appreciate the platform's flexibility and scalability, which are based on a microservice architecture that connects multiple modules independently and loosely. The platform utilizes open standards WMS, WFS, WCS, WPS OGC Web Service standard features through GeoServer and GeoWebCache, a tile cache server that accelerates map delivery built into GeoServer. Additionally, the platform supports visualization of data in various formats using mago3D, F4DConverter, and Smart Tiling.
The platform offers a range of services, including automatic apartment building placement, construction site safety management, 3D urban landscape simulation, environmental planning simulation, and underground facility visualization simulation. The platform also features real-time monitoring and visualization of IoT-based data, which is of particular interest to developers interested in smart city development.
Firstly, the presentation will show how open source based digital twin visualize the complex 3D city models in a web browser. Secondly it will showcase the platform's features and data, including actual system's functions and service UI/UX through a video. Attendees will gain insights into how the platform can be used to support rational decision making during complex urban planning, design, development, and operation stages.
This presentation is of interest to developers working in the field of urban planning, design, and development, as well as those interested in open source digital twin technology.
LH Corp is one of the largest public companies in Korea providing land and housing for public purpose. They are owned and controlled by the Korean government. They’ve played a large role in new town development and housing welfare.
QFieldCloud enables the synchronisation and consolidation of field data collected by teams using QField. From small individual projects to large data collection campaigns, the platform allows you to manage the collaboration of multiple people on the same project, assign different roles and rights to different users, work online and offline, and keep track of changes made. In 2022, QFieldCloud was testable as a beta version. Already during the beta phase, over 40,000 registered users synchronised their projects via the platform. Beginning of 2023, the official version was released. A brief overview of how QFieldCloud works and how the platform is built is given.
The use of free open source software is catching on and (at least) in Finland governmental institutions are making the big switch to open source software from other solutions. This opens up the need and possibility for training.
Training needs may differ from no previous training or knowledge to advanced GIS professionals so customising the training and exercises are important. Some might need to start with basic GIS and spatial information in general and continue to hands-on learning and multiple different exercises to help them learn the use of different tools and workflows in QGIS.
For more advanced users, training and helping with different programs for example GeoServer and QField or deepening the knowledge of different workflows such as visualisation or Python in QGIS are more in order.
FOSS4G has also been catching on and spreading in schools and universities. These new professionals that have used FOSS4G from the very beginning of their studies can be more efficient and skillful using these different programs. They may also demand more from the software and think of new ways to modify and perfect their workflows and produce new innovations. They can be a new and very important resource for developing different areas of FOSS4G.
Training new and more experienced professionals in FOSS4G is a very important step for implementing new tools and workflows into different industries and businesses. Training also works both ways, through discussion and hands-on exercises some new and interesting needs may emerge and those could be possible to develop further into new tools or plugins.
The more institutions, businesses and other users are interested in switching to FOSS4G, the more new opportunities and needs for different tools and working methods arise. This in turn helps to develop the software further.
Apache Superset is one of the most used no-code platforms for business intelligence. It allows for the exploration and visualization of data, from simple line charts to highly detailed geospatial charts, without the need for programming skills. These charts can be published on interactive dashboards to provide users with meaningful and up-to-date information. Currently, a plug-in for visualizing cartodiagrams is in development which is based on the OSGeo projects OpenLayers and GeoStyler. This plug-in gives users the ability to use any visualization of Superset within a geospatial context, so that e.g. simple pie charts or even complex location based timeseries can be displayed on a map. Thereby, Superset becomes a powerful tool for visualizing geospatial data.
This talk gives a brief overview of Superset and possible use cases while focussing on geospatial data.
Use cases should drive product development, not the other way around. Maps and the products we use to consume them have the biggest impact on the world when these principles are adhered to. How many government portals have you visited where a carefully curated map is presented that hardly anyone sees let alone uses? Presenting the data to the user in an intuitive way that helps them make a decision or take action is essential.
Large paper maps of the 1700s were well suited to a captain’s desk as their ships traversed the oceans. Road atlases of the 20th century helped to spur family adventures and weekend getaways as highway networks were constructed around the world. The small computers in our pockets today allow us to see when the next train will arrive and which one will get us home sooner. These examples took the technology of the day and used it to make products with significant impact on society. The mobile internet in particular changed mapping in one of the most notable ways since humans started abstracting 3D space on 2D surfaces.
We’re on the cusp of another great shift in the way maps are used with many exciting use cases awaiting discovery. The technology powering this potential is Augmented Reality (AR). This talk will explore some of the use cases that AR is supporting and where it might be useful in future. We’ll look at how AR can be accessed and how the medium of access affects its utility. With these use cases in mind, we’ll assess how open tools and map data enable AR. Some of the data and tools we’ll look at include:
Geometries of pedestrian ways
Associated attributes: Incline, safety, lighting, access, surface type, accessibility features
Building entrances
3D building data
VPS for localisation
Routing algorithms
The talk will conclude with a summary of Meta’s approach to map building and how open source geospatial technology powers the maps we build for today and the years ahead.
The National Land Survey (NLS) of Finland decided in the fall of 2020 to develop a national topographic data production system based on open source technologies and especially on QGIS client. Since then, many significant steps have been taken to implement the MVP of the application for the operators of the NLS at the start of 2025.
The latest significant expansion of our product has been the development of a comprehensive and user-friendly way to handle data quality management for the operators. Our aim was to develop it in a way that changes for the quality rules could be easily made and maintained and that would be as informative as possible. The basic idea behind quality management is clear: our customers want high-quality data, and we want the operators to have clear and easy-to-understand checks for their workflow that do not limit their productivity. For this, we have developed a tool, simply named Quality management tool.
The reason we couldn't use the basic QGIS tools was that they were not easily modifiable or extensive enough for our use cases and quality demands for our data. We have been able to use some of them, such as geometry checks, but for the most part, the quality tools had to be manually selected and configured, which would take the operator's time.
The key concept of quality management is that the operator gets real time feedback about the quality, so the errors can be fixed as part of the basic workflow and there is no need for separate phases for quality control. Additionally, we would not limit the user from saving their work to their local database, regardless of the errors they may have, so that the workflow would not be interrupted.
At this moment we have published the graphical user interface for visualizing the quality check results (can be found here: https://github.com/nlsfi/quality-result-gui) but on this talk I would show how it can work on a larger scale. For this purpose, I would present the tool with use case videos. I would also like to talk about the architecture of the tool and how we are going to develop the tool even further. I hope that some listeners can apply this tool for their workflows and benefit from this example.
Introducing Felt, a new map sharing and collaboration product.
We connect closely with the current ecosystem of open source mapping tools and make it easier to work together with colleagues inside and outside mapping. In this talk, we will show:
- How current users of programs like QGIS bring Felt into their workflows
- Where Felt lets them expand into new areas like community feedback
- How we’ve used and expanded core OSS libraries like MapLibre, GDAL, Pelias, and Tippecanoe
- Why we’re pushing forward emerging formats and standards like PMTiles
Session attendees will gain an important new tool for their stack, a product made for extending the reach of existing open source mapping tools and improving collaborative map-making beyond analysis.
The Lazio Region Authority (Italy) has been using for several years a system based on the G3W-SUITE and QGIS application which has allowed it, not only to publish public web services, but to prepare web cartographic management systems dedicated to internal staff for the management of territorial aspects of own competence:
* management of damages caused by wildlife and related reimbursement procedures
* environmental impact assessment practices
* wolf genetics
* signaling the presence of wild boar in urban areas
* nests and strandings of sea turtles
* road accidents with wildlife
The close integration between the suite and QGIS has allowed to create web cartographic management systems characterized by:
* numerous geometry editing features
* customization of the structure of the editing and attribute consultation forms
* simplification of attributes compilation thanks to the ability to inherit from QGIS: editing widgets,
* mandatory and uniqueness constraints, default values, conditional forms and drill down cascade based on expressions
* possibility of defining geographical constraints in visualization and editing in order to divide the
territory based on areas of competence associated with individual users
* possibility to differentiate the information content accessible on the basis of different users and roles
* descriptive analysis of the data through integration with the graphs created with the DataPlotly plugin
Thanks to the contribution and funding from the Lazio Region dedicated to the development and integration with the QGIS functions related to data editing, G3W-SUITE is configured as a valid tool for the preparation of advanced geographic data management systems on the web.
As an example, we report a series of use cases:
* Environmental Protection Agency of the Piemonte Region: post-event damage and usability census, management and cartographic representation of post-earthquake inspection requests
* Gran Paradiso National Park: park route signage management
* Piemonte Region: preparation of Civil Protection Plans
* Environmental Protection Agency of the Lombardy Region: Hydrological Information System
The Uber H3 library is a powerful geospatial indexing system that offers a versatile and efficient way to index and query geospatial data. It provides a hierarchical indexing scheme that allows for fast and accurate calculations of geospatial distances, as well as easy partitioning of data into regions. In this proposal, we suggest using the Uber H3 indexing library in Postgres for geospatial data processing.
Postgres is an open-source relational database management system that provides robust support for geospatial data processing through the PostGIS extension. PostGIS enables the storage, indexing, and querying of geospatial data in Postgres, and it offers a range of geospatial functions to manipulate and analyze geospatial data.
However, the performance of PostGIS can be limited when dealing with large datasets or complex queries. This is where the Uber H3 library can be of great use. By integrating Uber H3 indexing with Postgres, we can improve the performance of PostGIS, especially for operations that involve partitioning of data and distance calculations.
We propose to demonstrate the use of Uber H3 indexing library in Postgres for geospatial data processing through a series of examples and benchmarks. The proposed presentation will showcase the benefits of using Uber H3 indexing for geospatial data processing in Postgres, such as improved query performance and better partitioning of data. We will also discuss the potential use cases and applications of this integration, such as location-based services, transportation, and urban planning.
The proposed presentation will be of interest to developers, data scientists, and geospatial analysts who work with geospatial data in Postgres. It will provide a practical guide to integrating Uber H3 indexing with Postgres, and offer insights into the performance gains and applications of this integration.
Pointview is a product developed by IT34 for working with Lidar and Photogrammetry data. It gives the user the possibility to upload, process, visualize and work with data.
Lidar data formats such as LAS, LAZ, E57 can be uploaded, processed and visualized in the browser.
Photogrammetry: Images from drones or video from phones can be uploaded, processed into a 3d point cloud and visualized in a browser.
In addition, data can be captured using our SmartSurvey app that captures video which is used for building a 3d pointcloud, together with an ortophoto and dem. The data is later available for visualization in Pointview or in QGIS though a WFS service.
Moreover, the system offers a complete management system where the user can create projects for organizing the data, can share the data with other users and manage the access.
The system uses various data processing workflows for data processing based on open source components such as:
PostgresSql + Postgis for storing the data and for geometry based analysis.
OpenLayers for visualizing the images and ground control points results as rasters
Geoserver for publishing data as WMS/WFS,
QGis for visualizing data,
PDAL for lidar data processing,
GDAL for raster data processing,
CloudCompare for lidar data processing,
Potree for Data Visualization
The Massive Open Online Course - Earth Observation Open Data Science (MOOC EOODS) teaches the concepts of data cubes, cloud platforms, and open science in the context of Earth Observation (EO).
It aims at Earth Science students, researchers, and Data Scientists who want to increase their technical capabilities onto the newest standards in EO cloud computing. The course is designed as a MOOC that explains the concepts of cloud native EO and open science by applying them to a typical EO workflow from data discovery, data processing up to sharing the results in an open and FAIR way.
The EO College platform hosts the course and hands-on exercises are carried out directly on European EO cloud platforms, such as Euro Data Cube or openEO Platform, using open science tools like the Open Science Data Catalogue and STAC to embed the relevance of the learned concepts into real-world applications. The MOOC is an open learning experience relying on a mixture of animated lecture content and hands-on exercises created together with community renowned experts.
After finishing the course, the participants will understand the concepts of cloud native EO, be capable of independently using cloud platforms to approach EO related research questions and be confident in how to share research by adhering to the concepts of open science.
The MOOC is valuable for the EO community and open science as there is currently no learning resource available where the concepts of cloud native computing and open science in EO are taught jointly to bridge the gap towards the recent cloud native advancements in EO. The course is open to everybody, thus serving as teaching material for a wide range of purposes including universities and industry, maximizing the outreach to potential participants.
Our talk will give an overview of the MOOC at the current status. Furthermore, we encourage review, feedback on its content and discussion.
Deprecation of a used framework is a common risk for software projects. Migrations are very time-consuming and costly, without showcasing any new functional features. This can make them an unpopular task, that tends to be postponed until there is no other choice, be it for a customer or the community of an open source project.
During the last decade for instance, AngularJS has been one of the most popular web frameworks around. This was not any different in FOSS4G projects, where it had been adopted in geoportals and other frontend components. With the end of the decade, active development of AngularJS came to an end and since summer 2021 no more security updates are provided. This has become a major challenge for many web ecosystems - including FOSS4G ones - where AngularJS is still very present, but will have to be replaced in the long run.
This talk will present various open source projects and how they differently approach this challenge. It will reflect on lessons learned so far and aspires to provide inspiration for other projects in a similar situation.
Geomapfish is a WebGIS framework that allows to build geoportals. It is a community driven project. Its frontend is based on the ngeo javascript library, which has been built on top of AngularJS and OpenLayers. Due to its wide functionality, the project’s goal is to prevent a one shot migration. It has been decided for a continuous migration based on (Lit Element) web components, that allow to integrate migrated functionalities step by step.
Geoportal.lu is the national geoportal of Luxembourg. It is based on the Geomapfish framework, but has a very customized frontend. The requirement here is similar. Instead of migrating all at once, the different parts should be continuously integrated. After following the Geomapfish migration strategy based on web components at first, the project is finally migrated to another javascript framework (vue), without giving up on the continuous migration.
Geonetwork is a well-known FOSS4G catalog application. On top of its powerful backend, sits a frontend that is also based on AngularJS. Once again, its functionality is so vast, that a complete rewrite would be enormous. Thus came up the idea of geonetwork-ui: A new project that could live alongside Geonetwork without the goal to become isofunctional, but to complement it. A project providing libraries specialized in proposing user interfaces by leveraging Geonetwork’s backend capabilities.
This presentation will cover the support GeoServer provides to publish rich data models (complex features with nested properties and multiple-cardinality relationships), through OGC services and OGC API - Features, focusing on the recent Smart Data Loader and Features Templating extensions, covering in detail ongoing and planned work on GeoServer.
As far as the INSPIRE scenario is concerned, GeoServer has extensive support for implementing view and download services thanks to its core capabilities but also to a number of free and open-source extensions; undoubtedly the most well-known (and dreaded) extension is App-Schema, which can be used to publish complex data models and implement sophisticated download services for vector data.
We will also provide an overview of how those extensions are serving as a foundation for new approaches to publishing rich data models: publishing them directly from MongoDB, embracing the NoSQL nature of it, and supporting new output formats like JSON-LD which allows us to embed well-known semantics in our data.
Real-world use cases from the organizations that have selected GeoServer and GeoSolutions to support their use cases will be introduced to provide the attendees with references and lessons learned that could put them on the right path when adopting GeoServer.
To address the urgent need for affordable, high quality, non-sewered sanitation solutions for urban residents, our organization distributes and manages urine-diverting, container-based toilets with accompanying safe emptying, transportation, and treatment of waste through a subscription model in urban informal settlements. Our goal is to reach one million residents over the next five years at a cost of less than $5 per person per year. To achieve this goal, we need to invest in scalable and efficient systems, particularly for waste collection logistics.
Currently, our waste collection routing and scheduling is performed manually and unsystematically. Managers decide which toilets to collect from each day based on filling rate data and assign a cluster of toilets to each waste collector. Collectors then have the autonomy to choose their own route for collecting those toilets based on their knowledge of the location and movement paths in the area. The process is time consuming, provides opportunities for expensive inefficiencies in routes, and is not scalable. To improve and automate the process, we leveraged route planning algorithms by modifying and piloting an existing route optimization tool. In the first phase of the project, we collected data for the tool such as GPS coordinates of toilets, locations of intermediate holding points where collectors can consolidate collections, considerations that collectors currently make when planning their routes, and challenges they face. We were also able to use existing street maps from OSM that included both formal roads and informal footpaths, which is critical data for the tool to perform well. The routes generated by the updated tool were estimated to reduce collection time by 48% and distance traveled by 62%. In the current phase of the project we are syncing the tool with our Customer Relationship Management (CRM) system to reduce the need for manual data entry and with an open-source mobile application( (OSMAnd) that will provide turn-by-turn navigation instructions based on the routes produced from the tool. We will then conduct a pilot in the Mathare informal settlement in Nairobi to measure actual time saved by using the tool for collections.
EOReader is a remote-sensing opensource python library reading optical
and SAR constellations, loading and stacking bands, clouds, DEM and spectral indices in a sensor-agnostic way.
Optical | SAR |
---|---|
Sentinel-2 and Sentinel-2 Theia Sentinel-3 OLCI and SLSTR Landsat 1 to 9Harmonized Landsat-Sentinel PlanetScope , SkySat and RapidEye Pleiades and Pleiades-Neo SPOT-6/7 SPOT-4/5 Vision-1 Maxar SuperView-1 GEOSAT-2 |
Sentinel-1 COSMO-Skymed TerraSAR-X , TanDEM-X and PAZ SAR RADARSAT-2 and RADARSAT-Constellation ICEYE SAOCOM Capella |
It also implements sensor-agnostic features, such as load
and stack
many bands:
- satellite bands (optical or SAR)
- spectral indices
- clouds
- DEM
Context
As one of the Copernicus Emergency Management Service Rapid Mapping and Risk and Recovery Mapping operators,
SERTIT needs to deliver geoinformation (such as flood or fire delineation, landslides mapping, etc.) based on multiple EO constellations.
In rapid mapping, it is important to have access to various sensor types, resolutions, and satellites. Indeed, SAR sensors are able to detect through clouds and during nighttime while optical sensors benefit from of multi spectral bands to better analyze and classify the crisis information.
This is why SERTIT decided to decouple the sensor handling from the extraction algorithms: the latter should be able to ingest semantic bands without worrying about how to load the specific sensor band or in what unit it is.
The assumption was made that all the spectral bands from optical sensors could be mapped bands between each other, in addition to the natural mapping between SAR bands.
Examples
- Why EOReader?
- Basic tutorial
- Optical data
- SAR data
- VHR data
- Water detection on multiple products
- STAC
Simplified method of comparing covers from different years of MapBiomas 7 (Brazilian Land Cover Classification), using standard QGIS tools:
- Raster Calculator;
- Pixels to points;
- Sample Raster Values;
Code and data being organized in https://github.com/leandromet/geo_postgis/tree/master/2023_carbon_biodiversity.
For all processes, database hosting, data analysis and storage it was used a notebook computer with i7-10750H processor (6 cores, 12 threads) with 16GB RAM, 1TB SSD running Ubuntu 22.10 with QGIS 3.22, PostgeSQL 14.7, PostGIS 3.2.3 and PGAdmin 4 v6.21. There was no tuning, variable change or performance adjust from the default installation of all the tools out of Ubuntu repositories via apt-get.
To create images that represent land use change and a database of pixels for combinatorial analysis time of classes in the same places over the years, for the production of maps and representation of changes in Sankey Diagram-type graphs using javascript flow diagram (https://sankeymatic.com/).
Mapbiomas consists of 30 meter resolution raster files for every year between 1985 and 2020, covering the whole Brazilian territory (4,000x4,000 kilometers). It has 25 classes of land use represented by digital numbers in the raster file, like Natural Forest, Pasture, Urban and Silviculture.
We started by using Raster Calculator to aggregate Natural classes from blocks of years with threshold masks as simple as:
( "extent_year1985@1" = 3 or "extent_year1985@1" = 4 or "extent_year1985@1" = 5 or "extent_year1985@1" = 11 )
OR
( "extent_year1986@1" = 3 or "extent_year1986@1" = 4 or "extent_year1986@1" = 5 or "extent_year1986@1" = 11 )
And then multiplying results from distant periods with applied weights to visually enhance areas that lost or recovered certain land use classes. Once we had the mapped regions it is pretty hard to count all locations for many combinations with different years, like looking at same pixels that went from forest to pasture in 1988 but went back to forest in 1995 and then rock in 2005. The flux of uses has too many options for them to be organized by comparing the evolution by year and then by decades to try and see what happened in a larger region.
Since a raster file is in reality a table, in this case 36 tables with 155,239 columns and 158,459 lines, why not use it in a structured database for handling all the 885 billion pixels, as point geometries? In fact we need only 24 billion points, each with 36 attributes of 1 byte per year, and maybe we can store it in a tablespace equivalent to the original 36GB raster files.
With the objective of doing simple queries in a database that accessed all of the pixels from every year, in a way that anybody could do and use in a comprehensible and transparent way, it was created a spatial point table. With 250 million lines from a 500x700 kilometer rectangle on the Atlantic Forest, a table with 38 columns that can be queried and the land use classes grouped, spatially filtered and exported in a few seconds.
It took about two weeks to organize and develop a methodology based on those concepts, and we reproduced the Rondônia State in the Amazon biome to verify it all went fine, with a similarly sized rectangle and just over 300 million point registries. It takes about 12 hours of processing for all pixels from a 36 band TIFF file directly to a database table. Both datasets could be filtered using PostGis functions like ST_Intersects with benefits from spatial indexes (30 minutes to build it) and even with a final size of 60GB database from 8GB raster data, the derivation of new information became much faster than reprocessing images every time you change something in the analysis algorithm.
Once we had the complete table we can filter it with an administrative division from Brazil like:
select count(*) from ibge.proc_microrregioes_2021_4326_rupestre, mapbiomas.uso_solo_mapbio7
where cd_micro = '32008' and st_intersects( proc_microrregioes_2021_4326_rupestre.geom , uso_solo_mapbio7.geom)
---- result 3,931,037 pixels from 250 million
And also create aggregated data from many possible combinations of classes and years like:
select leg1.description as classe1985, leg1.legend as class_1985, mb_01_1985, mb_06_1990, mb_11_1995, mb_16_2000,
mb_21_2005, mb_26_2010, mb_31_2015, mb_36_2020, count(*) from
mapbiomas.campos_rupestres_pci_2023 , mapbiomas.uso_solo_mapbio7 , mapbiomas.mapbiomas_legend leg1
where
leg1.mapbiomas=mb_01_1985 and st_intersects( campos_rupestres_pci_2023.geom , uso_solo_mapbio7.geom)
group by leg1.legend, mb_01_1985, mb_06_1990, mb_11_1995, mb_16_2000,
mb_21_2005, mb_26_2010, mb_31_2015, mb_36_2020
---- SELECT 8,624 Query returned successfully in 2 secs 520 msec.
By doing tests like these we could determine that we find over 8 thousand different combinations of land use evolution looking at a 5 year interval in a 12x12 kilometer square, and over 150 thousand for the 500x700 km data. Most of the combinations have less than 1 to 10 hectares but it shows how impossible it would be to look at all possibilities detectable in the area mapping and coloring approach.
With the grouped information we did diagrams of the Sankey type to see the full period and step by step change in regions with similar pixels evolution. We could differentiate protected areas from private properties, the main differences from the Atlantic Forest where tenure is consolidated from Amazon conflict areas.
We have places that lost less than 5% of forest cover in Indigenous land, others in public settlements with over 50% of forest becoming pasture in the Amazon and in contrast the Atlantic Forest lost coverage until the early 2000´s and recovered, but the original areas are distinct from the recovered ones. For a sample close to our institution for example, we saw a 200k hectares decrease from 8 million original forests but when looking at where are all pixels we find that over a million hectare was lost and the same amount regenerated in other abandoned areas. That might indicate a loss of biodiversity and other environmental services and quality due to a lack of management is a broader point of view.
Introduction to basic but important concepts about Coordinate Reference Systems (what is doable in 20 min ;)
- Geographic Coordinate (Reference) Systems
- Different Datums/Ellipsoids
- Projections (Mercator, UTM, LCC, ...)
- EPSG catalog
- WKT (well known text) description
- Reference to PROJ.org library
The purpose is to explain basic concepts to have a good basis to understand later more complex problems. The presentation will have a lot of links to go deeper into any area of interest.
This talk is about the current state of MilMap and its ongoing development. MilMap is a military geo-portal system widely and successfully used in every sectors of Korean military. The system is now undergoing major change from geo-portal to military digital twin system.
MilMap is developed on top of numerous open source projects such as PostGIS, GeoServer, GeoWebCache, Cesium, OpenLayers, mago3D, OpenGXT. The system provides several functionalities like POI search, geospatial data search, layer control, satellite image search and download, spatial terrain analysis, coordinates reading, and map notes, to the military officers through the intranet. Although the system provides geospatial analytics functions through OGC WPS(Web Processing Service), the current system is basically a web based 3D GIS for data viewing and printing. Thanks to MilMap, military officers can now access the huge amount of geospatial data(maps, imagery, 3D, POI, and others) in their browser without installing additional software.
MilMap is now undergoing major development to be a more customized, automated, and analytical system. The future MilMap will support user data uploading for intelligence sharing, more bespoke battle field analysis and others. In the long run, MilMap is expected to be a cloud based military digital twin system for geospatial intelligence sharing and battle field analysis & simulation.
This is the story of 2 twin projects (namely AIR-BREAK and USAGE) undertaken by Deda Next on dynamic sensor-based data, from self-built air quality stations to the implementation of OGC standard compliant client solution.
In the first half of 2022, within AIR-BREAK project (https://www.uia-initiative.eu/en/uia-cities/ferrara), we involved 10 local high schools to self-build 40 low-cost stations (ca. 200€ each, with off-the-shelf sensors and electronic equipment) for measuring air quality (PM10, PM2.5, CO2) and climate (temperature, humidity). After completing the assembling, in late 2022 stations were installed at high schools, private households, private companies and local associations. Measurements are collected every 20 seconds and pushed to RMAP server (Rete Monitoraggio Ambientale Partecipativo = Partecipatory Environmental Monitoring Network - https://rmap.cc/).
Hourly average values are then ingested with Apache NiFi into OGC’s SensorThings API (aka STA) compliant server of the Municipality of Ferrara (https://iot.comune.fe.it/FROST-Server/v1.1/) based on the open source FROST solution by Fraunhofer Institute (https://github.com/FraunhoferIOSB/FROST-Server).
STA provides an open, geospatial-enabled and unified way to interconnect Internet of Things (IoT) devices, data and applications over the Web (https://www.ogc.org/standard/sensorthings/). STA is an open standard, it builds on web protocols and on OGC’s SWE standards and has an easy-to-use REST-like interface, providing a uniform way to expose the full potential of the IoT (https://github.com/opengeospatial/sensorthings/).
In second half of 2022, within USAGE project (https://www.usage-project.eu/), we released the v1 of a QGIS plugin for STA protocol.
The plugin enables QGIS to access dynamic data from heterogeneous domains and different sensor/IoT platforms, using the same standard data model and API. Among others, dynamic data collected by the Municipality of Ferrara will be CC-BY licensed and made accessible from municipal open data portal (https://dati.comune.fe.it/).
During the talk, a live demo will be showcased, accessing public endpoints exposing measurements (timeseries) about air quality (from EEA), water (BRGM), bicycle counters, traffic sensors, etc.
Japan fascinates the world with its rich culture, materialized with a full of cultural sites in its territory as example. To protect it, the Law for the protection of cultural properties established a “cultural heritage” designation system, where designated places should be preserved.
With the collaboration of the Nara National Research Institute for Cultural Properties, Japan cultural heritages has been mapped as a WebGIS tool where more than 100,000 places can be visualized.
In this talk will be presented tool functionalities and technically its OpenSource based architecture.
OSGeoLive is a self-contained bootable DVD, USB thumb drive or Virtual Machine based on Lubuntu, that allows you to try a wide variety of open source geospatial software without installing anything. It is composed entirely of free software, allowing it to be freely distributed, duplicated and passed around. It provides pre-configured applications for a range of geospatial use cases, including storage, publishing, viewing, analysis and manipulation of data. It also contains sample datasets and documentation. OSGeoLive is an OSGeo project used in several workshops at FOSS4Gs around
the world.
The OSGeoLive project has consistently and sustainably been attracting contributions from ~ 50 projects for over a decade. Why has it been successful? What has attracted hundreds of diverse people to contribute to this project? How are technology changes affecting OSGeoLive, and by extension, the greater OSGeo ecosystem? Where is OSGeoLive heading and what are the challenges and opportunities for the future? How is the project steering committee operating? In this presentation we will cover current roadmap, opportunities and challenges, and why people are using OSGeoLive.
In recent years, 3D city models have gained popularity for supporting urban planning, citizen engagement, and research. As technology and infrastructure have improved, many cities and countries now use 3D models to address urban issues, encourage participation, and inform decision-making.
The Japanese government, including the Ministry of Land, Infrastructure, Transport and Tourism's Project PLATEAU, have promoted open 3D city models and 3D point cloud data. Over 100 cities are currently developing and releasing open digital twin data in CityGML format as of February 2023. Binyu et al. published the results of these efforts, which are also highlighted in the 3D City Index benchmarking report. The report shows that seven out of 40 cities (18%) compared were Japanese cities.
This report discusses the current state of open digital twin data in Japan, which is compatible with the open database license ODbL. The data can be imported into popular tools such as OpenStreetMap, and converters have been developed for this purpose. Since 2022, import work has been conducted on an experimental basis in collaboration with national and international communities. Sharing the results and challenges of this work is expected to promote the use of 3D city model data globally.
The routing machine is about the route track a user can take from one point to the other with directions after reaching each point. For paid services such as Google maps, this already exists, and Google has applied a centralized model of usage. In this talk, we will talk about the type of libraries and already existing implementations that are almost deprecated but we can keep alive, since for the open source community, the ability to customize and change, they are essential. There are no active Open Source or community versions of the routing machine for maps. We need to change that. We can do that by improving a couple of things that already exist. Having more wrappers for different types of implementations, say Vue, or React, and finally Svelte. The routes should be updated and the selection of the type of route, car, bike, or walking should reflect the data received from maps. And define a safer business model. Open Source is more active and strong than paid and centralized services. We need to make sure that what we are offering and implementing as services to our clients can reflect a similar dedication the first have.
This is a plug-in created using pyQGIS, and an example of using it as basic data for decision-making on noise measurement station selection policy will be presented.
As data for use in decision-making by public institutions, we introduce cases in which basic public data are utilized and processed to ultimately be used as core data for decision-making.
It will be time to talk about how text-based data held and provided by public institutions is being used for their spatial expression and policy making, and why the opening of public data will play a more important role in the future.
Natural Earth is a public domain map dataset available at 1:10m, 1:50m, and 1:110 million scales. Featuring tightly integrated vector and raster data, with Natural Earth one can build a variety of visually pleasing, well-crafted maps with cartography or GIS software.
GeoServer GeoCSS is a CSS inspired language allowing you to build maps without consuming fingertips in the process, while providing all the same abilities as SLD.
In this presentation we’ll show how we have built a world political map and a world geographic map based on Natural Earth, using CSS, and shared the results on GitHub. We’ll share with you how simple, compact styles can be used to prepare a multiscale map, including: * Leveraging CSS cascading. * Building styles that respond to scales in ways that go beyond simple scale dependencies. * Various types of labeling tricks (conflict resolution and label priority, controlling label density, label placement, typography, labels in various scripts, label shields and more). * Quickly controlling colors with LessCSS inspired functions. * Building symbology using GeoServer large set of well known marks.
Join this presentation for a relaxing introduction to simple and informative maps.
Discrete Global Grid Systems (DGGS) are gaining popularity as a new method of geospatial data representation. This presentation will provide an overview of the concept of DGGS and its advantages over traditional geospatial data representation methods.
We will explore the similarities and differences between these different DGGS frameworks, including their cell shapes, grid resolutions, and ability to handle different types of geospatial data. We will also discuss the benefits of using DGGS in geospatial data applications, such as remote sensing, climate modeling, and environmental monitoring.
Overall, this presentation will provide a comprehensive overview of the concept of DGGS and its potential applications in geospatial data analysis and visualization. Attendees will gain a deeper understanding of the advantages and challenges associated with different DGGS frameworks and will gain insights into the ongoing research efforts in this field.
The open source based environmental impact assessment(EIA) decision support verification tool(verification tool) is a web-based tool for verifying the EIA algorithm based on the EIA review decision support algorithm using data for each Environmental Impact.
This verification tool was developed using open source projects such as PostGIS, GeoServer, and Openlayers. However, the flowchart library used a commercial software called GoJS.
This verification tool is intended to verify the adequacy of the implementation of the EIA algorithm developed by experts in each Environmental Impact.
It is possible to support comprehensive decision-making, including opinion gathering, by operating the review decision-making algorithm based on data by Environmental Impact and environmental impact analysis results.
The spatial analysis required to verify the algorithm was developed using OpenGXT of the OGC WPS service. It includes a way to visualize the result processed through this spatial analysis function.
The world is big. So is the OpenStreetMap dataset -- and it's growing bigger each day. This is great, of course, since the #1 collection of free geospatial data is becoming ever more detailed. But working with this massive data volume requires a high-end workstation, guru-level database skills, and patience (or at least two out of three).
Or, perhaps, it requires a new approach?
GeoDesk is a database engine specifically designed to work with OSM. Instead of importing OSM data into a traditional database, users convert it into a GOL (a "Geographic Object Library"). Not only is this process twenty times faster, the resulting GOL file consumes one-tenth of the storage required by an SQL-based database.
Features can be retrieved based on their characteristics and spatial relationships, using a simple and intuitive query language that executes 50 times faster than SQL. The ability to export regions in bulk makes it easy to archive and share geodata.
The GeoDesk software consists of two parts: a Java-based toolkit that enables developers to incorporate the database engine into their own applications, and a stand-alone utility for creating, managing and querying GOLs.
The GeoDesk utility converts OSM data into a variety of popular formats. It also allows users without programming knowledge to generate basic reports, from counting the pubs in Dublin to finding the longest rivers in Colorado.
GeoDesk is 100% FOSS, and its hardware requirements are modest -- a decade-old dual-core laptop will work just fine. This makes free and open geospatial analysis accessible to a broader audience of developers.
Learn more at www.geodesk.com, or follow @GeoDeskTeam.
GeoServer is the start of so many great open source success stories.
This talk introduces the core GeoServer application and explores the ecosystem that has developed around this beloved OSGeo application. Our presentation draws on the GeoServer ecosystem for use-cases and examples of how the application has been used successfully by a wide range of organizations.
Each use-case highlights a capability of GeoServer providing an overview of the technology drawn from practical examples.
- Andrea Amie is on hand to share success stories highlighting GeoServer use in managing vulnerable ecosystems, agriculture information management, and marine data management.
- Jody Garnett will look at how GeoServer technology powers cloud services
- Gabriel will look at am amazing remixes for Cloud Native GeoServer
- GeoServer technology powering the OSGeo community, including GeoNode, geOrchestra
- A showcase of examples collected from our user list
Attend this talk to learn what GeoServer is good for out-of-the-box, and for inspiration on what is possible using GeoServer and the FOSS4G community.
National Land Survey of Finland (NLS) has built multiple feature services based on the OGC API Features standard since 2019. These services provide cadastral and topographic data, buildings, geographic names, and addresses both as open and contract-based APIs.
The engine behind these services is Hakunapi – a high performance server implementation to easily build “off-the-shelf” Simple Features and customized Complex Features services with geospatial data backed by a PostGIS database. Currently the OGC API Features (Part 1, 2 and 3) standard is supported. The codebase is based on Java, and it utilizes also other geospatial libraries such as JTS Topology Suite and GeoTools.
Hakunapi is now Free Open-Source Software available at GitHub, and with the version 1.0 to be released during the spring 2023. On the last few years NLS has internally used the library for services providing both Simple Features (like traditional topographic database) and Complex Features (cadastral registry and geographic names with some hierarchical feature structures too).
This talk presents key features and benefits of using Hakunapi for implementing feature services based on the OGC API Features standard. Also experiences and best practices by NLS on developing these services and our roadmap towards modern OGC API services is discussed.
Demo: https://beta-paikkatieto.maanmittauslaitos.fi/inspire-addresses/features/v1/
Code: https://github.com/nlsfi/hakunapi
In January 2022, OSGeo and OGC signed a new and updated version of the Memorandum of Understanding (MoU) that aims to maximize the achievement of the mission and goals of the two organizations: promoting the use of Open Standards and Open source software within the geospatial developer community. Identifying open source technologies that could be used as Reference Implementations for OGC Standards and validating OGC compliance tests are examples of activities that can take place within the scope of the agreement.
More than one year after the agreement was signed and almost one year after it was introduced to the OSGeo community in a keynote at FOSS4G 2022, this presentation will summarize all activities accomplished and future plans, including the establishment of the OSGeo Standards Committee within OSGeo and the organisation of the 3rd joint code sprint, in Switzerland, together with the Apache Software Foundation.
The presentation will also reiterate the benefits of the new agreement, which allows OSGeo charter members to represent the priorities of OSGeo in the development of OGC Standards and supporting documents and services.
The ever-increasing threat from disaster is an urgent call for a proactive discourse on pragmatic elimination and reduction of the challenges and stresses caused by disasters. This study, therefore, leverages on the research gap as it applies to the application of crowdsourced rapid response mapping in a developing country of Nigeria, where, critical geospatial data is grossly unavailable to respond to vulnerable resilient communities. The study deployed two research techniques namely: participatory crowdsourced mapping and gamification. The HOT tasking manager data analytics was used to analyze the level of participation and contribution of volunteer mappers over time while QGIS was used to produce maps unveiling building footprints generated in OSM, before and after Mapathon. The study delineated 8 LGAs for a mapping task of 2015 grids and 639 grids for Mapathon battle season-1 and 2 respectively.Season-1 was the months of flood(Rainy Season) while Season-2 was the flood receding months (dry season) Results unveiled analysis of flood response mapping season-1, had a total of 571,659 edits comprising 481,912 buildings and 22,244km of roads contributed by initial 7,601 participants, but completed by 1,644 volunteers, mapping 4,946 grids within a timeline of 38months at the rate of three hours 38minutes per task. 70% of volunteer mappers engaged were beginner mappers Maps showing before and after Mapathon in OSM were produced for ONELGA, Numan Sarbon Birnin and Ilorin West LGAs respectively. However, analysis of flood response mapping season-2, unveiled a total of 357,168 edits comprising 325,023 buildings and 7,438km of roads were contributed by the initial 4,006 participants, but completed by 801 volunteer mappers using 2,238 grids within a timeline of 14months at the rate of two hours 33minutes per task. Maps showing before and after in OSM were produced for Afikpo North, Warri South, Logo and Jamare LGA respectively. The Study contributed to a measurable target of SDGs 1 to 7, 11, 13, 15 and 17. The study generated massive critical open geospatial data needed for effective disaster response and SDGs, and paving way for effective geoinformation e-governance in Nigeria. Lastly, the study promotes the relevance of citizen-generated data for national geospatial data infrastructure development and participatory crowdsourced mapping using OpenStreetMap at local levels. The study has also bridged a critical scientific research gap and inquiry in OSM GIScience.
The interest on urban pedestrian networks is growing, with impacts centered at UN SDGs numbered 3, 11, 10 and 13: the improvement of accessibility helps in reducing inequalities and the fostering of non-motorized locomotion improves well-being and sustainability in urban scenarios. The idea behind OpenSidewalkMap is to leverage the multi-purpose OpenStreetMap data for the pedestrian network data. The structure of the project is decentralized, with localities deployed as nodes on a world web-map. At each node there’s a modular structure within a webpage, containing apps that have a different role, in order to create what is intended to be a full-fledged inventory, whose functionality can be expanded as new modules can be added. Currently there are four modules: “Webmap” containing an interactive cartographic representation of the data; “Optimized Routing” that uses the data to create an optimized routing, currently only for a wheelchair profile based on an empiric equation; “Dashboard” featuring statistical charts to look at the bigger picture of the data, mainly focused on value percentages, thus giving attribute completeness, also giving a look at the data aging and number of revisions; “Data quality tool” looking at most common possible errors on data, giving direct link to editors, being at this point focused on finding invalid values, with geometrical and topological error detection planned to be included; there are 4 planned modules: “data watching” to monitor changes on data, to track and combat possible vandalism against data since OSM data is universally editable; “Tiles” giving raster and maybe vector tiles; “API” giving features on request; “Surveying And Validation” to list projects in different platforms/editors to expand and validate available data. This way the inventory will include continuously the full cycle of data: creation and collection; storage, maintenance and management; application and analysis. The project is aimed to have zero-maintenance costs, as long as everything is hosted using current freely available Microsoft github infrastructure, with all code and data being maintained inside github repositories, webpages deployed with github pages, updated using github actions. In case of shutdown of any of these services, the software can still be deployed in another server infrastructure with a similar workflow. There is lots of room for improvement, with only the node for the city of Curitiba being available as of february 2023. The homepage of the code is available at: https://kauevestena.github.io/opensidewalkmap/ .
Project PLATEAU is an initiative led by the Ministry of Land, Infrastructure, Transport and Tourism of Japan (MLIT), to develop and utilize 3D city models compliant with CityGML standards. MLIT aims to establish rules of creation of 3D city models as part of general operations in each local government, and also to release them as open data to promote utilization for urban planning and business creation.
Raster data is a type of digital image data that is stored and processed as a grid of cells, each of which represents a specific area or location in the image. This grid is known as a raster or pixel grid, and each cell contains a value that represents a characteristic of the corresponding area or location in the image, such as color, elevation, temperature, or other attributes. Depending upon the resolution of the data these raster file sizes can vary from a few MBs to few GBs. Hence reading data from a large set of raster dataset which has time dimension associated with it is challenging.
PostgreSQL can be used to store time series raster datasets, which are raster datasets that have a time dimension associated with them. This can be useful for storing and analyzing raster data that changes over time, such as satellite images, climate data, or land cover change data.
To store time series raster datasets in PostgreSQL, we will use the postgis_raster extension, which provides support for storing and manipulating raster data in the database, and the TimescaleDB extension to add time series functionality to PostgreSQL, allowing us to store and query raster data with a time dimension.
Using the TimeScaleDb extension we will partition the raster table by converting it to hypertable which is what TimescaleDB uses to optimally store and process time series data. This can help us to optimize query time.
For aggregated values from raster data over time and space, we will use the Continuous aggregate feature of TimescaleDB which is a form of materialized view to pre-compute and store raster data over time.
Moreover, TimescaleDB allows compression of data which can be very helpful in cases where the data is huge which is usually the case with raster datasets in postgres saving us space in the Database and optimizing some queries.
The proposed presentation will be of interest to developers, data scientists, and geospatial analysts who work with Raster datasets. It will provide a practical guide to querying the raster datasets in PostgreSQL with TimescaleDB and postgis_raster extension.
The View Server (VS) is MIT licensed, Docker based, cloud-native, scalable software stack providing external services for searching, viewing, and downloading Earth Observation (EO) data. Services implementations are following OGC Web services standards STAC, OpenSearch, WMS, WMTS, WCS.
Having EOxServer and MapCache as core components, enables EO Data publication in a modular and configurable way. The process starts with data harvesting, preprocessing and metadata ingestion and ends with serving pre-cached and on demand rendered images through an attached Web client based on OpenLayers and EOxC libraries or on individual service endpoints.
EOxServer allows dynamic generation of visual images from multi-spectral data. In this way, specific bands or channels of the original images can be selected as the grey or red, green, and blue output colour channels. It also supports flexible rendering based on previously extracted image statistics, pansharpening on the fly, filtering the long time periods of products intersecting with the query in CQL syntax utilizing metadata parameters and more.
VS provides both S3, OpenStack Swift, HTTP and local files support when considering data storage and can be deployed in Docker Swarm environment via docker-compose templates or in Kubernetes environment as a set of Helm charts.
The software stack was and is used by EOX in a quite a number of operational deployments for ESA, like the VirES projects, Copernicus Space Component Data Access system (CSCDA), or more recently Earth Observation Exploitation Platform Common Architecture.
Links:
https://eox.at/2021/09/eoxserver-1-0/
https://eoxserver.org/
https://github.com/EOxServer/eoxserver/
https://gitlab.eox.at/vs/vs-deployment
Powerful earthquakes hit southern Turkey and Syria on 6 February 2023. These earthquakes in Turkey and Syria caused thousands of casualties and destroyed cities. Geospatial infrastructure is critical to respond to these earthquakes during rescue operations, humanitarian effort as well as planning recovery activities.
Yercizenler coordinated mapping activation with the collaboration of Humanitarian OpenStreeMap team to improve open geodata infrastructure in the earthquake affected region and supporting humanitarian response in the scope of mapping.
Türkiye Earthquakes Mapping Response aims to complete open map data infrastructure before and after the event in affected areas. This response is structured with following workstreams; Remote Mapping, Post-disaster Field Data Collection, Global Community Activation and Geo-data Integration.
In this talk; we will talk about how open data and community activation helped save lives after earthquakes, what challenges we faced and what we have learnt during the Türkiye Earthquakes mapping Response effort.
“Whether it is to know where children are, what access they have to facilities (education, health, transportation), what environment they live in (water, air), where risks exist (hazards, diseases), where events happen or where services and resources are available; most of the operational data used by UNICEF is geospatial” (UNICEF Geospatial Roadmap, 2019). At UNICEF we realize that we need to leverage geospatial information to enhance decision-making and optimize resource allocation and drive effective interventions. Geo-enabling UNICEF’s data, systems and processes aims at transforming data into easily accessible, readily available, and actionable geospatial information that can address key questions, such as: “How many children have been affected by a flood?”, “Where children have limited access to schools and limited access to health services?”. This information is critical to support decision-making to ultimately drive better results for children.
UNICEF has recently adopted a hybrid corporate geospatial architecture, which aims at bringing together the advantages of both commercial and open-source GIS world. This presentation aims at discussing how UNICEF is leveraging modern open-source geospatial solutions to address some of the key data-management challenges.
Specifically, two open-source geospatial projects developed by UNICEF will be showcased and discussed: GeoRepo and GeoSight. GeoRepo is a web-based system that will help us store, manage and share a commonly agreed, versioned, official set of administrative boundaries and other core geospatial datasets. It will help us ensure that geospatial data is used consistently across all internal systems and will also strengthen our interoperability with external systems. GeoSight, on the other hand, is a web geospatial data platform developed by UNICEF to bridge the gap between web mapping systems and the Business Intelligence / data analytical platforms. GeoSight is specifically designed to simplify the process of harmonizing data from multiple data sources. It also allows users to easily create online maps for visualizing multiple indicators at subnational levels (e.g. at the province or district level). Both platforms are built using Django and React and use modern open-source geospatial standards and libraries, such as MapLibre and vector tiles.
*t:*
Are you tired of managing your database infrastructure and paying exorbitant fees for cloud-based solutions? We have the solution for you! Join us as we demonstrate how to build a self-managed database on a Kubernetes cluster using open-source technologies, including the PostgreSQL operator by CrunchyData, TimescaleDB, PostGIS, and Uber H3.
Working with geospatial data can be a daunting task and if you add a temporal dimension on top, it becomes a bigger puzzle. While frameworks like STAC provide a way to work with such use cases, let’s talk about database solutions combining open-source technologies
- Postgres (the power of SQL)
- Postgis (our friendly neighborhood geospatial handler)
- timescaleDB (the new cool thing which gives command over the very time)
- uber-h3 (the secret weapon to improve speed)
And combine all this with postgres-operator by crunchyData and you have one-stop solution for all your database needs.
We will talk about how to set up a Kubernetes cluster on any cloud provider, such as AWS EKS or Linode, and show how to deploy a self-managed database using the PostgreSQL operator. We will highlight the benefits of using TimescaleDB and PostGIS as extensions to add temporal and spatial capabilities, respectively. We will also showcase Uber H3 for faster spatial analysis, allowing you to easily process large amounts of spatial data.
To ensure high availability and disaster recovery, we’ll see how to use replicas (separate read and write and automatic switchover) and backups with pgBackRest. Additionally, we will see how to monitor your database infrastructure using Prometheus and Grafana which are inbuilt into postgres-operator with a very minimum configuration that can is super flexible and all open source.
This solution provides organizations with a scalable and cost-effective database infrastructure that is resilient, easy to deploy, and easy to manage.
In the end, let’s compare the cost of running a managed solution such as RDS (x2) and flexibility (None).
GeoNode is a Web Spatial Content Management System that uses the Django Python web framework. MapStore is an open source WebGIS product and highly customizable framework that has been used as the default user interface to visualize catalog, map viewer and geospatial applications in GeoNode.
This presentation provides an overview of the integration of the MapStore framework inside the GeoNode ecosystem and the main differences with the MapStore product, along with guidelines and references to resources for its customization and the development of custom functionality.
This talk describes the creation of a water quantification dataset for the entire world. Tracking changes of water-bodies over time helps in timely action to combat drought and floods. The tools used to build this dataset are all free and open source (postgis, gdal, geopandas, scipy) and are built on top of data from OpenStreetMap.
The dataset is updated everyday with new measurements of lake water extent across the globe. The solution to detect and track water bodies involved fetching satellite data using STAC API, pre-processing it to remove cloud cover and invalid pixels, identifying water bodies using band ratio, converting to vector and applying post-processing filters to avoid false-positive detection to finally serve it through an API. This solution has allowed us to track and quantify changes in a lake's water extent over time with high accuracy.
Local governments can use OpenStreetMap data in a variety of ways to help them reach their full potential in areas including planning, infrastructure management, community development, and more. Participation, crisis management, economic growth, and numerous other things.
OpenStreetMap Uganda is utilizing this data to create a GIS-based digital property tax database that will allow town agents to search and explore details about various properties across four cities in Uganda using free and open-source software. Geographic Information Systems (GIS) can use OpenStreetMap data to give robust mapping and analytical capabilities.
This enables local governments to offer simple access to useful information and resources for planning, decision-making, promoting openness, and increasing community engagement.
Motivation
New, evolving technologies allow to host data and program code (smart contracts) on distributed blockchains. Beside other aspects, like validating geospatial data and their transactions, this technology might also be interesting for building distributed services for the ‘classical’ spatial data infrastructures.
Prototype
During the last months a prototype was developed to test the capabilities of smart contracts to distribute spatial data using the OGC API – Features specification and gain some experiences in its design, typical workflows, limitations etc.
The prototype is designed as smart contract on the ‘Internet Computer (IC)’ blockchain (see https://internetcomputer.org/). This allows to store program code and the spatial data in one container on the blockchain and execute it on demand.
To simplify the test, a fixed workflow is implemented.
1) Data providers upload a spatial dataset (currently glider GNSS tracks in the IGC format) on a simple webpage running within the container
2) The spatial data is persisted on the IC blockchain
3) Users access the data via OGC API – Features with their browser (html representation) or with their GIS
Presentation
In the presentation, I would like to share some experiences on developing geospatial interfaces in a blockchain environment and show the current state of the prototype. Especially the coding with the programming language ‘Motoko’, the exposed interfaces, and the distribution on the blockchain with its costs will be addressed.
I would further like to discuss use cases of the approach, e.g. a simplified data distribution for smaller data providers, and the potential extensions on this approach, like introducing a user management, adding metadata or integrating dynamic data sources.
Links
- Entrypoint (OGC API Features): https://mtlom-hiaaa-aaaah-abtkq-cai.raw.ic0.app/
- Github page (mainly Motoko code): https://github.com/janschu/igc_tools
Related:
- Internetcomputer (IC): https://internetcomputer.org/
- IGC - International Gliding Commission – GNSS Flight Recorders Spec: https://www.fai.org/sites/default/files/igc_fr_specification_2020-11-25_with_al6.pdf
- OGC API – Features: https://ogcapi.ogc.org/features/
Virtual Constellations-as-a-Service and Virtual Image Catalogs
Sharing remote sensing assets among multiple tenants is crucial to unlock the value of new space earth imaging constellations. In these schemes, a tenant has access to a so-called virtual constellation consisting of dedicated access to a number of assets as well as automated mechanisms to procure additional imagery from other assets. Access to this virtual constellation is mediated through a virtual catalog client-side that looks to the user as if it comes from its own dedicated assets and is fully interoperable with open-source standards for cloud optimized pipelines, such as STAC and COG.
Satellogic Inc., a leader in sub-meter resolution Earth Observation data collection recently reached a three-year agreement with the Government of Albania to develop a Dedicated Satellite Constellation. This unique program derives from Satellogic's Constellation-as-a-Service model and will provide Albania with responsive satellite imagery capabilities across its sovereign territory. Two satellites, ALBANIA1 and ALBANIA2 were launched in January 2023, to provide imagery for national map generation to support emergency response, land use management as well as environmental monitoring of sustainability goals.
To support this government effort we have developed a secure, encrypted end-to-end data platform, continuously updated archival imagery in dedicated client-side cloud along with support for open source standards such as STAC and COG. We also discuss future directions in terms of the resulting ability to build integrations with external image processing platforms and open source data exploitation projects.
In recent years, the software industry has witnessed a remarkable trend away from traditional standalone applications and towards online multiplayer platforms that offer users a more integrated and collaborative experience.
As this trend continues, it is becoming increasingly important for open source tools to stay competitive by providing seamless access to data and connectivity.
In this talk I will introduce mapstack, outline our mission to bring all of the world’s open location data together in one place, and share my thoughts on how such an unprecedented open resource will benefit the wider FOSS4G ecosystem.
ZOO-Project is a WPS (Web Processing Service) platform which is implemented as an Open Source project and following the OGC standards, it was released under an MIT/X-11 style license and is currently in incubation at OSGeo. It provides a WPS compliant developer-friendly framework to easily create and chain WPS Web services. This presentation gives a brief overview of the platform and summarizes new capabilities and enhancement available in the new version. A brief summary of the Open Source project history with its direct link with FOSS4G will be presented. The new release comes up with a brand new ZOO-Kernel Fast Process Manager and, with the approved standard OGC API - Processes part 1: core. The new functionalities and concepts available in the latest release will be presented and described, also highlight their interests for applications developers and users. Apart from that, various use of OSGeo software, such as GDAL, GEOS, PostGIS, pgRouting, GRASS, OTB, SAGA-GIS, as WPS services through the ZOO-Project will be presented. Then, the ongoing developments and future innovations will be explored.
Diagonal is a steward-owned data science consultancy working with projects in the built environment. We build interactive tools to help people understand the tradeoffs inherent in their plans to evolve cities. Our tools are powered by B6 - an in-memory geospatial analysis engine we built to work with large data sets describing the built environment. We typically use it work work with OpenStreetMap and open government data. To enable others to repeat our analyses, we recently released B6 as open source. In this talk, we'll give an overview of B6, including how it's implemented, and how we use it in our commercial work.
Albania is one of the most vulnerable countries in terms of the trend of climate change in the Western Balkans. Changing weather patterns have already been observed over the last 15 years with increasing temperatures, decreasing precipitation, and more frequent extreme events like floods and droughts. Among the most affected cities is Tirana, where a time series analysis was done using FOSS data and tools. Our aim was to provide accurate map representations of local climate zones (LCZs) to track the changes of the last decade based on an open online platform running on Google Earth Engine. This is called LCZ generator and aims to use free data sources from the Copernicus Hub (Demuzere et al. 2021). The satellite data based analysis was done by using 5-15 training areas for each LCZ types. It provided a 100 by 100 m ground resolution supervised classification for the entire municipality of Tirana. The analysis shows that the quick urbanization process resulted in a decreasing proportion of green areas, and unpaved surfaces in the municipality of Tirana, which consequently increased the vulnerability of the city to extreme weather events.
A large-scale map was also compiled using a free and open source Geographic Information System (QGIS), which seems to be the most effective in identifying the varying urban climate zones on the city planning level, since it shows the city's structures and even highlights the role of a building or small park (Cenameri, 2021).
We're living in the world of APIs. CRUD operations are base of lot of operations. Many smart frameworks such as Django, Flask, Laravel provides out of the box solutions to filter the data, which covers almost all needs to separate data based on column values.
When it comes to Geospatial data, we expect to filter data based on their location property instead of metadata. This is where things get complicated, if you are using framework that doesn't have package, library built to handle such use cases, you are likely to be dependent on either database or any external package to handle it.
Fortunately Geodjango[https://docs.djangoproject.com/en/4.0/ref/contrib/gis/] (Django's extension) allows us to create databases which understands geometry and can process it[https://docs.djangoproject.com/en/4.0/ref/contrib/gis/geoquerysets/#gis-queryset-api-reference]. It also provides support to write APIs using Rest Framework extension [https://pypi.org/project/djangorestframework-gis/] which takes this to next level allowing user to output the data in various formats, creating paginations inside geojson, create TMSTileFilters, etc.
In this talk we'll scratch the surface of this python package and see how to build basic CRUD APIs to push, pull GIS data along with filtering it to the PostgreSQL database
At the Norwegian Water and Energy Directorate (NVE), the OSGeo Community project actinia was introduced together with the Open Source Apache Airflow software as a platform for delivering operational Copernicus services at national scale.
In the presentation, we will illustrate how Airflow and actinia work together and present current and future applications operationalized on the platform.
Those applications cover currently:
- Avalanches
- Flooding
- snow cover
- lake ice
More services related to NVE`s area of responsibility are being investigated, like landslides, slush flows, glacier lake outburst floods, or specific land cover changes...
Born in 2016 thanks to the funding of the National Operational Program for Metropolitan Cities (PON METRO 2014-2020) the current Spatial Data Infrastructure (SDI) of the city of Genova is a hybrid infrastructure, where open source components and technologies are merged together with proprietary ones (such as the Oracle Database) in a well designed platform with respect of all national guidelines (promoted by AgID - Agenzia per l’Italia Digitale) and international standards.
To support the Geoportal initiative, the city of Genova has collaborated with GeoSolutions as a company closely involved in the most important Open Source projects worldwide in the geospatial field with the aim to provide the necessary support for all the SDI stack in terms of deployment, development but also the staff training to make it autonomous as much as possible in the maintenance of the overall system.
The city of Genova Geoportal as well as the wider Geospatial Infrastructure are both reachable online.
A simple and at the same time robust WebGIS based on the Open Source MapStore software is provided with the inclusion of both advanced GSI functionalities and also most common geospatial tools like:
- Geospatial data search via OCG Web Services and Nominatim
- 2D and 3D visualization of geospatial data using a map agnostic engine supporting OpenLayers, Leaflet and Cesium for the 3D
- Editing and Styling of geospatial layers
- Download functions of geospatial data working on top of OGC services
- And many more
The aim is to provide ready-to-use tools for all users (both citizens and employed analysts worked in the PA) by leveraging the maturity of the Open Source Software as well as the simplicity of integration with the pre-existing COTS software in order to maximize the reuse of the existing infrastructure and minimize the need for customizations and a possible use of commercial support even for educational purposes.
Many cross-cutting projects usually gravitate around the SDI in the Public Administration and its own Geoportal. To date, more than 300 geospatial layers are available in the Geoportal which allows them to be viewed and consulted within preconfigured MapStore maps, dashboards and geostories and/or used through geospatial services (such as WMS, WMTS, WFS, WCS and CSW) developed according to international standards (OGC - Open Geospatial Consortium) and exposed through GeoServer and GeoNetwork with also a fine grained security tier represented by GeoFence to manage authorizations on geospatial data.
Open data and geospatial technology have the potential to revolutionize decision-making processes across a variety of sectors, including urban planning, disaster response, environmental management, and more. However, the use of open data in the geospatial domain poses its own set of challenges, including data quality, reliability, and standardization concerns. Managing, maintaining, and updating large datasets can also be resource-intensive, posing a challenge for organizations and communities that rely on open data.
This talk will explore the opportunities and challenges of using open data in the context of geospatial technology. I will begin by discussing the potential benefits of open data, including increased transparency, improved collaboration, and the ability to make more informed decisions. I will then delve into the key challenges of using open data in geospatial contexts, including issues related to data quality and reliability, standardization, and the sheer volume of data. We will explore strategies for managing and maintaining large datasets, such as crowdsourcing and automated data processing, and discuss best practices for ensuring data quality and reliability.
This talk is relevant to anyone interested in the intersection of open data and geospatial technology, including data scientists, GIS professionals, policymakers, and community leaders. Attendees will come away with a deeper understanding of the opportunities and challenges of using open data in geospatial contexts and gain practical insights on how to leverage this data to drive social and economic impact. By the end of the talk, attendees will be equipped with the knowledge and tools they need to make the most of open data in the geospatial domain.
Never before have we had such a rich collection of satellite imagery available to both companies and the general public. Between missions such as Landsat 8 and Sentinels and the explosion of cubesats, as well as the free availability of worldwide data from the European Copernicus program and from Drones, a veritable flood of data is made available for everyday usage.
Managing, locating and displaying such a large volume of satellite images can be challenging. Join this presentation to learn how GeoServer can help with with that job, with real world examples, including:
- Indexing and locating images using The OpenSearch for EO and STAC protocols
- Managing large volumes of satellite images, in an efficient and cost effective way, using Cloud Optimized GeoTIFFs.
- Visualize mosaics of images, creating composite with the right set of views (filtering), in the desired stacking order (color on top, most recent on top, less cloudy on top, your choice)
- Perform both small and large extractions of imagery using the WCS and WPS protocols
- Generate and view time based animations of the above mosaics, in a period of interest
- Perform band algebra operations using Jiffle
Attend this talk to get a good update on the latest GeoServer capabilities in the Earth Observation field.
pycsw is an OGC CSW server implementation written in Python and is an official OSGeo Project. pycsw implements clause 10 HTTP protocol binding - Catalogue Services for the Web, CSW of the OpenGIS Catalogue Service Implementation Specification, version 3.0.0 and 2.0.2. pycsw allows for the publishing and discovery of geospatial metadata, providing a standards-based metadata and catalogue component of spatial data infrastructures. The project is certified OGC Compliant, and is an OGC Reference Implementation.
The project currently powers numerous high profile catalogues such as IOOS, NGDS, NOAA, US Department of State, US Department of Interior, geodata.gov.gr, Met Norway and WMO WOUDC. This session starts with a status report of the project, followed by an open question answer session to give a chance to users to interact with members of the pycsw project team. This session will cover how the project PSC operates, the current project roadmap, and recent enhancements focused on ESA's EOEPCA, Open Science Data Catalogue and OGC API - Records.
WiGISKe, with the support of CPCS Transcom Ltd An Industry leader in Transportation Analytics, conducted a study to better understand the experiences of public transit users in Kenya, with a particular focus on the safety concerns of women and girls. Public transit trips have the potential to be less safe for women and girls, as they often must walk through, or wait in, unsafe areas in order to access public transit. Additionally, many women must travel through the city very early in the morning or late at night, which can make them more vulnerable to harm.
The study utilized community-driven remote data collection methods, such as surveys and focus groups, to document security pain points and strategies adopted in Nairobi to ensure inclusive development. The findings highlighted several key safety concerns raised by women and girls, such as harassment, assault, and theft. However, the study also documented successful strategies that have been implemented to address these concerns, such as increased lighting and security patrols at transit stops.
The implications of these findings are significant. Policymakers and public transit providers must take action to ensure that women and girls have access to safe and reliable public transit, which is essential for their daily lives and economic opportunities. This may require changes in policy, infrastructure, and funding, as well as increased awareness and education about the importance of safe public transit.
Overall, the study provides valuable insights into the experiences of public transit users in Kenya and the need for safer and more inclusive public transit systems. By highlighting both the challenges and successes in addressing safety concerns, the study can serve as a valuable resource for policymakers, public transit providers, and other stakeholders seeking to improve public transit in Kenya and beyond
The Open Metadata platform allows the integration of data and metadata for the management of governance within an organization to integrate different sources, control its publication, its access, standardize the processing and even to be able to analyze the lineage. What we are going to share is the adaptation of one of the data sources to the OGC - CSW service to be able to consume the cataloged metadata transparently in the system.
In this talk we discover various aspects of styling in Geoserver. It allows us to style using various languages such as SLD, YSLD, CSS, etc. along with ability to create style using no-code way ( Geostyler ) .
In this talk we explore how to style Point, LineString, Polygon, Rasters. We'll cover following tasks
1. Attribute based Styling
2. Zoom based Styling
3. Variable Styling
4. Labelling and it's optimisation
5. Legends
n the West African Sahel, farmers and herders are critically vulnerable to climate shocks and need access to climate information to secure their livelihoods. Herders use data on pasture and water availability to move their livestock and farmers need weather predictions for planting. While satellite imagery has made much of this information readily accessible to the spatial community, few channels exist to transmit this information to farmers and herders. As a result, climate data has become more powerful than ever before, yet mostly inaccessible to those who depend on this information for their livelihoods.
This talk will share the lessons of the GARBAL programme, an initiative that seeks to bridge this gap. GARBAL is a call center that uses Copernicus Earth Observation imagery and field data to provide farmers & herders with information on pasture, water and markets in Mali, Niger and Burkina Faso. GARBAL was first developed in 2015 and this talk will provide lessons from several years of practice.
The GARBAL interface uses an open-source stack including PostGIS and Mapserver to create a user-friendly interface for call center agents, who then use that interface to answer questions from callers on pasture conditions, market prices and weather forecasts (among others).
The talk will share lessons from the technical and programmatic aspects of the project. The technical side will go over the architecture of the data treatment, demo the interface, talk about successes and failures and show how you can play with the data yourself. The programmatic side focuses more on how the user needs evolved over the years, techniques for translating GIS data into information useful to farmers and herders, operating in areas of active conflict and how EO data fits into existing centuries-old traditional data collection systems in the Sahel.
The amount of data we have to process and publish keeps growing every day, fortunately, the infrastructure, technologies, and methodologies to handle such streams of data keep improving and maturing. GeoServer is an Open Source web service for publishing your geospatial data using industry standards for vector, raster, and mapping. It powers a number of open-source projects like GeoNode and geOrchestra and it is widely used throughout the world by organizations to manage and disseminate data at scale. We integrated GeoServer with some well-known big data technologies like Kafka and Databricks, and deployed the systems in Azure cloud, to handle use cases that required near-realtime displaying of the latest AIS received data on a map as well background batch processing of historical Maritime AIS data.
This presentation will describe the architecture put in place, and the challenges that GeoSolutions had to overcome to publish big data through GeoServer OGC services (WMS, WFS, and WPS), finding the correct balance that maximized ingestion performance and visualization performance. We had to integrate with a streaming processing platform that took care of most of the processing and storing of the data in an Azure data lake that allows GeoServer to efficiently query for the latest available features, respecting all the authorization policies that were put in place. A few custom GeoServer extensions were implemented to handle the authorization complexity, the advanced styling needs, and big data integration needs.
The Copernicus Ground Motion Service (EGMS) is a European Union (EU) initiative under the Copernicus program, which aims to provide near-real-time information about ground deformation caused by natural or man-made hazards. The service uses a variety of data sources, including satellite radar imagery, to monitor and analyze ground motion in areas prone to landslides, sinkholes, earthquakes, and other hazards. Given the sensitive nature of the service, EGMS product validation is a key activity in assuring the user community (especially the decision makers) of the quality of the ground motion and deformation information provided.
The main goals of the EGMS validation system are as follows: to provide a reproducible environment on top of modern cloud infrastructures (with a particular focus on the European geo clouds), to enable the development of scientific tools that validate EGMS characteristics, to facilitate the reproducibility of the validation tasks, and to account for key performance indicators (which will allow shareholders to monitor the quality of the primary EGMS product).
To achieve the first goal of providing a reproducible environment, we have focused on providing Terraform modules that facilitate the deployment of our software stack on any supported cloud platform. The software stack is built on top of the Kubernetes container orchestration system, which runs on top of a managed cloud environment. Kubernetes provides uniform services regardless of the underlying cloud platform.
For the goals of developing the validation tools and the execution of those tools we decided on using an unified approach based on the JupyterHub solution. JupyterHub is used for providing an unified development environment based on R and Python EO software tools (based on modified Pangeo Docker images). Also Jupyter is used for executing the validation tools outside of JupyterHub by leveraging an internal python service that uses papermill to execute the notebook and then “nbconvert” to generate a html webpage containing the required visualizations and documentation in human readable form.
The validation system is complemented by an bespoke web dashboard aimed for providing reports and information related to the status of the various key performance indicators.
Overall the whole validation system was developed by solely using FOSS4G components: GeoPandas, RasterIO, GeoNode, GeoServer and JupyterHub.
(or what happens when GeoServer and PostGIS meet Active Directory)
This talk will present a case study of how Astun implemented a single sign on (SSO) system for a large
commercial client. The client stored their spatial data in a PostGIS database and provided both direct access
to the database via QGis and from QGis via WMS using GeoServer to carry out the styling and rendering of the
data. Staff are divided into 4 teams and then are subdivided by end client in to small groups. Some of the
data in the system is restricted to just the group working on a specific problem for a specific client, other
data is shared with the whole team, and some is available to the whole company.
The client brief was to move their on site system to "the cloud", and to allow staff to connect to the data
from anywhere in the world with only one user account and password for access to PostGIS and GeoServer data.
Initially, the project planned to leverage the existing corporate Azure Active Directory system to provide the
necessary authentication and authorizations. However, early experiments showed that the time between
requesting a new group and it appearing on the server was (sometimes) longer than the lifetime of the new
group.
Astun provided an open source solution, using Keycloak to handle the user and administrator facing frontends,
with user data being stored in an OpenLDAP server. It was then possible to make use of the LDAP service to
perform authentication and authorization of users to both PostGIS and GeoServer, making sure that data
restrictions applying in one were duplicated in the other.
The talk will cover details of the process and look at some of the issues that were encountered during the
project.
pygeoapi is an OGC API Reference Implementation. Implemented in Python, pygeoapi supports numerous OGC APIs via a core agnostic API, different web frameworks (Flask, Starlette, Django) and a fully integrated OpenAPI capability. Lightweight, easy to deploy and cloud-ready, pygeoapi's architecture facilitates publishing datasets and processes from multiple sources. The project also provides an extensible plugin framework, enabling developers to implement custom data adapters, filters and processes to meet their specific requirements and workflows. pygeoapi also supports the STAC specification in support of static data publishing.
pygeoapi has a significant install base around the world, with numerous projects in academia, government and industry deployments. The project is also an OGC API Reference Implementation, lowering the barrier to publishing geospatial data for all users.
This presentation will provide an update on the current status, latest developments in the project, including new core features and plugins. In addition, the presentation will highlight key projects using pygeoapi for geospatial data discovery, access and visualization.
We show how Mergin Maps can be used in various real-world situations to use the power of QGIS ecosystem to speed up and effectively capture data in the field and reliably collaborate with your team. We will not dive into technical details, but focus more on general understanding of what can be done nowadays in the field of professional geo-data capturing.
Do you need to capture the location of plants or animals with your personal phone? Or distribute this task to a group of volunteers without need to train them? Or your company has a network of pipes or fiber cables, you use QGIS in the office for analysis and you want to use the same map as your colleagues on site? Are you fed up with using for such tasks a camera and MS Excel or even pen and paper? This talk can show you how others solve these challenges with Mergin Maps.
Mergin Maps is a free and open-source platform powered by QGIS rendering engine to capture and share geo-data with ease. It has been developed by Lutra Consulting since 2017 and it has served thousands of companies and individuals in full production for more than 2 years. It comes with Android, iOS apps that do not need any training to be used by the general public. Also a powerful server to store, version and collaborate on your QGIS projects.
MapLibre GL JS, Leaflet, Esri Leaflet, OpenLayers, and Cesium JS are all great mapping libraries. However, it can be difficult to decide which one to use for different applications. In this talk, I compare the strengths and weaknesses of each library based on different criteria. The criteria include the following:
- Library footprint and modularity.
- Load times for vector tile and image tile layers.
- Rendering performance of GeoJSON data.
- Styling and rendering features.
- Viewport performance and screen size responsiveness.
This presentation is the follow up of the datahub paradigm presented last year: The confluence of geo data and open data. This time we will look at the metadata edition and maintenance aspect.
Writing metadata to describe a dataset is an essential part of managing a catalog. Each record in a catalog has been written, or at the very least enriched, by actual humans. GeoNetwork is a very widely used open-source metadata catalog; as such, it offers powerful tools in this regard: custom edition forms, batch editing, templates, custom XSL processing, advanced edition in XML, etc.
Despite all these features, authoring metadata is often felt as a difficult process, involving complex actions, convoluted validity rules and an intricate knowledge of metadata schemas like ISO19139.
Our vision for this new metadata editor can be summed up in three phrases:
- Make metadata accessible to everyone
- Forget about metadata schemas
- Build your own editor
This editor is made to feed content into your Datahub. Whether you want to describe open data, geo data or anything else, the editor will make it simple for you! Come and discover the concepts behind the scenes.
At the end of 2022 the Swiss geodata catalogue, geocat.ch, was migrated to GeoNetwork version 4. A more modern user interface as well as a more powerful search based on Elasticsearch makes it easier to search the more than 14000 geometadata contained in geocat.ch.
This new version of geocat.ch has been the subject of a usability study focusing on geodata search. Some developments based on the results of this study have been proposed to the GeoNetwork developer community. To discuss these proposals with the other users of GeoNetwork, a GeoNetwork user community should be founded and could be helpful in the further developments of GeoNetwork. In addition, the usability study showed that the search for geodata is very dependent on the quality of the information entered into the catalog.
The geometadata in geocat.ch come from different organizations (direct entry or harvesting), have different spatial extents, are multilingual and some have different data models. The harmonized entries of the most important information are essential and form the basis for efficient searches. The Swiss geometadata standard (GM03), which is currently under review with the aim of simplifying and updating the Swiss geometadata model, always based on international standards.
geOrchestra is a complete spatial data infrastructure (SDI) and combines a number of widely used open source components. These include GeoNetwork as a metadata catalogue, GeoServer, GeoWebCache, GeoFence, and Jasig CAS. During this talk we will present the project and its latest developments.
geOrchestra is an open source, modular, interoperable and secure spatial data infrastructure designed by people for people.
The technical architecture is based on modularity and interoperability. The extensive use of the Spring Framework allows the integration of additional components. Compliance with OGC standards is central, because only then can the various components and any external IDS work together.
geOrchestra is supported by an underlying server infrastructure, which can be configured in an automated way if necessary. We support deployment on Kubernetes as well as Ansible. geOrchestra has proven to be an innovative IDS in a highly orchestrated environment. Its modular architecture allows it to deploy individual components as microservices. Individual components such as GeoServer-cloud or GeoNetwork Microservices can therefore be scaled as needed.
Nevertheless, an SDI must be user-friendly and adopt a user-centric approach. This is the latest development that the geOrchestra community has started to follow. New modules such as the Datafeeder simplifies the data registry and the Datahub portal makes it very easy for a user to find the right dataset.
Current developments related to geOrchestra include a rewrite of the GeoNetwork metadata catalogue to provide a complete new user interface for editing metadata.
The substantial reduction of disaster risk and life losses, a major goal of the Sendai Framework by the United Nations Office for Disaster Risk Reduction (UNISDR), requires a clear understanding of the dynamics of the built environment and how it affects, in case of natural disasters, the life of communities, represented by local governments and individuals. The framework states that communities participating in risk assessments should increase their understanding of efficient risk mitigation measures.
Earthquakes are threatening many regions in the world with constantly increasing risk due to rapid urbanization and industrialization. Earthquakes do not kill people, buildings do. Thus, the main threat of earthquakes comes from building damage and collapse. To improve resilience and preparedness, we need to estimate the risk, the possible damage of buildings and the related human and financial losses. This requires not only the position, size and class of buildings, but also the reconstruction value and the number of people inside the building at any time. For this, exposure models are used that translate the physical earthquake hazard to building damage, human and financial losses. Exposure models usually describe the built environment of administrative regions as groups (aggregates) of different building classes and their frequency.
We present our open, dynamic, and global approach to describe, model, and classify every building on Earth with the greatest level of detail. Our model is based on the building data from OpenStreetMap and engineering information from open exposure models, combining these two sources to a building-by-building description of the exposed assets. We retain the aggregated descriptions where the building coverage in OpenStreetMap is incomplete and describe every building separately where building data is available. Due to the near-real-time computations of our model, it directly profits from the growth of OpenStreetMap and with about 5 million buildings added each month (or approx. 2 per second), the areas of incomplete coverage are constantly shrinking, making way for our building-specific exposure model.
Here, we introduce shortly the earthquake phenomenon, how it affects the built environment, why a high level of detail is necessary for useful assessments of the impact and the consequences of earthquakes, how OpenStreetMap and other open data helps us to achieve this goal and how communities can benefit for the model for their own risk assessments.
Finland is reputed to be the Land of a Thousand Lakes, but a more precise estimate is that Finland has 57000 lakes which are larger than one hectare. The precise shorelines of all the lakes have been available as open data since 2012 but the situation with the bathymetric data is not as good. Depth contours are available for about 80% of the total lake area, but oldest soundings are from the end of the 19th century. Bathymetric data of the lakes has not been considered particularly important and the old measurements have not been systematically updated and verified. Therefore, the most common acquisition method in the existing bathymetric data is still manual measurement with a plumb line through the ice. Because the depth points are frequently 75-100 meters apart, such data are only usable for creating rather approximate depth contours.
However, since mid 1980s the Finnish Environment Agency, the Finnish Transport and Communications Agency Traficom, and their predecessors, have been mapping lake bathymetry with sonar sounding. In recent years these agencies have published their depth point data as open data under the CC-BY 4.0 license. These new datasets are essentially XYZ point clouds. Thanks to open source GIS programs anybody can take these datasets and create digital elevation models (DEM) of the lake bottoms, colored hillshade visualizations, 3D-models, and even traditional depth contours.
This presentation will dig into the nature of the data that is collected with sonar soundings and how it affects the selection of the interpolation method. A complete open source workflow that is using GDAL and Generic Mapping Tools (GMT) will be presented. The workflow begins from raw point measurements and lake shoreline vectors, and yields a DEM, hillshade visualization with a color table, and depth contours. Results for more than 1800 Finnish lakes will be available online, but the main outcome is the workflow itself. Because only command line tools which can be scripted and parameterized are used, it is simple to tune the process so that the output will suit different needs.
In the FOSS4G 2021 programme, the word 'notebook' appeared ten times and the word 'jupyter' ten times too in the abstracts of four workshops and four presentations.
In 2022, 'jupyter' and 'notebook' appear in two workshops and two presentations abstracts.
More discreetly, at least three workshops and one scientific paper used notebooks without mentioning them.
As we can see, notebooks are becoming increasingly common in data science and the geospatial world.
But what is a notebook? What is it useful for? What are its limitations?
Are there other platforms than Jupyter?
Can we do anything other than Python? What about geospatial? Are these tools FOSS?
These are some of the questions that this presentation will try to answer.
(TL;DR: yes!)
If you have never heard of Quarto, Observable or Org-mode, this presentation is for you.
In this talk we give an example of how open source tooling enables companies to fast-track software development, while simultaneously benefitting the FOSS4G community. Our use case is the development of the user interface for hydrodynamic simulation software, including editing and analysis, called the 3Di Modeller Interface.
Traditionally hydrodynamic simulation software companies develop their own user interfaces, usually closely resembling GIS packages, (re-)implementing features like background maps, layer management, geoprocessing tools, and styling options. In our approach we turned it around. Instead of developing our own GIS-like software, we used QGIS to leverage development. Specifically for larger governmental agencies (where a certain well-known proprietary GIS suite is often the only GIS that employees are allowed to use), we packaged our implementation in an installer, enabling modellers to use QGIS for hydrodynamic analysis within their organisations.
This approach has several advantages for users and for the FOSS4G community. For users, hydrodynamic modelling tools seamlessly integrate with the ever expanding GIS capabilities that QGIS has to offer; and users can built their own custom tooling, combining our own open libraries for hydrodynamic modelling with FOSS4G libraries like PyQGIS, Shapely, NetworkX, GDAL or QGIS.
For the FOSS4G community, this approach increases the user base, including users that are into developing their own plugins, it increases sustainable memberships, and creates job opportunities for FOSS4G developers.
The 3Di Modeller Interface is developed by Nelen & Schuurmans, a Dutch water and IT company, in collaboration with Lutra Consulting, a European FOSS4G company. Its development relies on several open source projects: QGIS, Shapely, GDAL, GeoAlchemy2, and NetworkX, amongst others. When we started in software development, we used open source mainly because it was free of cost. During the development, the board of directors became convinced that contributing to several open source projects (financially and/or developing) is the way forward.
The extractive sector in Malawi has been marked as one of the development enablers to achieve the 2063 Agenda established by African nations. As the mining sector continues to develop, open-source software such as QGIS has been a vital and cost-effective tool in monitoring mining activities for the purpose of tracking the effects of mining on the environment and human populations and encouraging accountability from stakeholders in relation to the Malawi government regulations. Open-source software and data have also been vital in resolving compensation issues in communities that exist in mining areas and allow for geoscientists to give needed to advice to affected stakeholders.
The Model of Living Landscape (MLL) is a set of empirical based tools for land management and landscape planning. It recognizes the complexity of the interactions between humans and the natural environment, and it aims to create a sustainable and resilient landscape that supports the well-being of both people and nature. One of the core MLL components is a process-based model for rainfall-runoff and erosion computation called SMODERP. The model operates on the principle of cell-by-cell mass balance, calculated at each time step. SMODERP (https://github.com/storm-fsv-cvut/smoderp2d) is open-source software implemented in Python language to ensure compatibility with most GIS software solutions. The current implementation supports Esri ArcGIS, GRASS GIS and QGIS. In this contribution, a new QGIS SMODERP plugin linking the hydrologic model outputs to MLL will be presented. The plugin performs the input data preparation on the background using GRASS GIS data provider, computation is done by SMODERP Python package, and results visualised with predefined map symbology in QGIS map canvas.
This contribution was supported by grant RAGO - Living landscape (SFZP 085320/2022) and Using remote sensing to assess negative impacts of rainstorms (TAČR - SS01020366).
As well trained and experienced members of the free software community from Turkey, we were caught off guard, when the earthquakes happened on February 6, 2023. We started mapping campaigns with HOT, we aggregate different data sources on a GeoServer installation, we did several visualizations on QGIS, but we always felt like something was missing.
If we had a guideline of disaster response for free software communities, we would feel better at the beginning.
This session's aim is, to prepare a dynamic guideline of disaster response actions for geospatial communities, focused on free software and open data.
In preparation for a new Alpine Club map by the Institute of Cartography of the TU Dresden around Mt. Ushba in Georgia in the Great Caucasus, the decision was made to use OpenStreetMap as the primary data source for the map. As a result, the fieldwork in place contributed to OpenStreetMap to use gained information for map production by using OpenStreetMap. In the past, data import and organized mapping had already happened, leaving gaps only fillable by fieldwork.
Mapping campaigns took place in 2021 and 2022. In preparation, it was necessary to identify missing or uncertain information. The catalogue of objects which should be mapped was derived from existing Alpine Club maps and the feature tags of OpenStreetMap. Several trails currently missing in OpenStreetMap were identified by collecting and comparing openly available GPS tracks, hiking guides, and old maps. The comprehensive information collection summarized the knowledge of all the sources. It became central for planning the office work on the data and organizing the extensive on-site mapping.
Based on the collected information, the routes were planned in advance and during the fieldwork assigned to the mapping teams. On tour, new data was collected, which could not be obtained from aerial images such as small paths, hiking routes, guideposts, and POIs.
The collection of geographical names worked similar to the collection of missing paths. After reviewing and selecting various sources, an updated set of names has been compiled. Old maps play an important role because they sometimes contain names that need to be added or allow updates for more recent documents. Combined with background literature on the region, uncertainties in assigning geographical features can frequently be solved. Asking locals helped in finding the ideal spelling. The result is a much more consistent toponym base both in the OpenStreetMap database and in the derived produced map.
The presentation will share the knowledge on preparing and organizing the fieldwork for such a project. Significant aspects are how to identify missing ways and to collect geographic names.
The Open Geospatial Consortium API family of standards (OGC API) are being developed to make it easy for anyone to provide geospatial data to the web, and are the next generation of geospatial web API standards designed with resource-oriented architecture, RESTful principles and OpenAPI. In addition, OGC APIs are being built for cloud capability and agility.
The OGC API - Processes standard supports the wrapping of computational tasks into executable processes that can be offered by a server through a Web API and be invoked by a client application. The standard specifies a processing interface to communicate over a RESTful protocol using JavaScript Object Notation (JSON) encodings. Typically, these processes execute well-defined algorithms that ingest or process vector and/or coverage data to produce new datasets or analyses.
pygeoapi is an open source Python server implementation of the OGC API suite of standards. The project emerged as one of the most effective reference implementations that provides the capability for organizations to deploy OGC API endpoints using OpenAPI, GeoJSON, and HTML. pygeoapi is built on an extensible plugin framework in support of clean, adaptive data integration and easy customization.
Prefect is an open source data workflow orchestration platform developed in Python. It provides robust orchestration of workflows and offers a large set of features that range from monitoring to supporting cloud storage, to periodic execution, etc. It is a robust and very capable workflow engine, which is a perfect fit for managing execution of OGC API – Processes requests in pygeoapi.
This presentation will provide an overview of the prefect process manager plugin for pygeoapi and will demonstrate:
- How to use pygeoapi for handling OGC API - Processes use cases
- How the pygeoapi prefect plugin is a good match for managing the execution of processes and what are its main strengths as a geospatial data processing platform
R is well-known for its unsurpassed provision of well documented statistical functions and packages in the default installation. Less well-known is its excellent support for spatial data through packages such as sf, terra, and stars. A thriving ecosystem of diverse and often topic-specific packages build on these foundations, making R a powerful command-line GIS (Geographic Information System) for reproducible research. However, dedicated GIS software (e.g. QGIS) offers specific processing algorithms that are either not available in R, or may achieve a higher level of performance than their equivalents in R. This presentation describes how it is now possible to combine the strengths of R and QGIS through R packages that interface processing algorithms provided by QGIS. These packages (qgisprocess, qgis) allow users to create data processing pipelines that combine R and QGIS algorithms almost seamlessly. We discuss the current state of these R packages and demonstrate the usage of their most important functions by example. Finally, we shed light on future development directions and seek feedback from the community.
This presentation focuses on the use of MapStore to navigate urban scenarios using its 3D tools and capabilities. Latest versions of MapStore include improvements and tools related to the exploitation of 3D data such as Map Views, Styling, 3D Measurements and more. Support for 3D Tiles and glTF models through the Cesium mapping library has also been greatly enhanced to provide support for more powerful integration.
Attendees will be presented with a selection of use cases around the following topics: visualization of new projects for urban planning, relations between different levels of a city and descriptions of events inside a city. At the end of the presentation attendees will be able to use the presented workflows to replicate them on different urban scenarios using the 3D tools of the MapStore WebGIS application.
In the OGC world, you have a catalog to look for metadata/datasets, and the OGC API Features to fetch the data, paginate, filter and so on.
The use cases have evolved since then and data consumers expect more complete abilities from their data catalogs. Nowadays we want to analyze, understand and reuse our datasets and providing such tools is a great way to encourage data owners to share and open their warehouse. A data API could then offer:
Full text search on data points
Data fetching, paging, sorting and filtering
Data analytics, aggregation, computation
Data joining
And those operations should perform in an optimized and scalable manner.
You might have heard about columnar formats, and columnar vector formats such as Arrow, Parquet… After an introduction of the context and the expectation of a well shaped data API, we’ll present different approaches and types of flow architectures
- Warehouse formats
- Static files (parquet)
- Index
- Databases (PostGIS, Cytus)
- Api models and implementation
- OGC API Features limitation
- Duck DB
- Pure SQL
And compare the different stack in terms of efficiency depending on various use cases.
The final goal is to provide an API which serves search, analytics and dataviz purposes.
Orfeo ToolBox is now a mature software with more than 100 applications dedicated to remote sensing and data extraction.
It is used both in academic works, in operational processing chains.
OTB now needs to be more modular ("core", "machine learning", "SAR", "feature extraction") and also easier to use through Python.
We will present the recent developments and our roadmap.
Initiated in 2002, the OSGeo project deegree has developed to an important and mature building block for Spatial Data Infrastructures (SDI) over the last 20 years. The project provides 9 official Reference Implementations of OGC Standards such as GML, WFS, WMS, and OGC API - Features.
In this talk, we will focus on the recent improvements available in deegree webservices v3.5 and the updated roadmap for the next version which lists support of Java 17. We will also show how the OGC Standards OGC API - Features Core and CRS have been implemented and can be used with existing configurations.
Finally, we will present the future directions of the project and what developments are currently planned.
Deck.gl is a framework for visualization, animation and 3D editing of large volumes of data (up to millions of points), in the browser, with optimal performance thanks to WebGL technology and the computing power of the GPU.
Deck.gl is prepared to work seamlessly with WebGL based map libraries such as MapLibre GL JS, Mapbox GL JS or Google Maps. It extends their capabilities with a large number of formats, data types and layer visualizations, such as point clouds (tessellated or not), real 3D vector data, 3D models, on-the-fly clustering, trip animations, GPU filtering, etc. The deck.gl code is not only free, but designed with extensibility in mind, making it very easily customizable.
In this presentation we will show 4 use cases developed for companies and administrations with specific needs. We chose deck.gl (over Mapbox/MapLibre alone) to provide rich interactivity and the ability to visually analyze large amounts of data.
We will expose the challenges we faced and how deck.gl was used:
1. Information system for precision irrigation: in a region of 25,000 plots, we show animated time series of evapotranspiration data, vegetative vigor, or water needs during an annual cycle.
2. Biodiversity world map: instant loading of a dataset of 200,000 points with GPU filtering, providing interactivity and refresh rates far beyond the ones offered by Mapbox or MapLibre.
3. Precision topographic measurements on terrain surface models: visualization of point clouds, terrains, textures, contour lines and other vector cartography in 3D, multi-profiles, and in-browser 3D editing.
4. Urban data control panel: from a dataset of 40,000 georeferenced records, we apply spatiotemporal and categorical filtering, 3D dynamic aggregation and symbolization, and computation of indicators and graphs in real time.
BBOX is a new OGC API Open Source implementation, with support for established OGC services driven by MapServer or QGIS Server. BBOX is implemented in Rust, with a built-in high-performance web server.
- Supported OGC API Services:
- OGC API - Maps, with support for OGC WMS 1.3
- OGC API - Tiles, with support for WMTS and XYZ endpoints
- OGC API - Features
- OGC API - Processes, with multiple processing engine backends
- Enterprise ready:
- Authentication / Authorization
- Instrumentation + Monitoring
- First class Docker / K8s support
- Simple usage:
- bbox serve –map alaska.qgz
The National Land Survey of Finland (NLS) is rebuilding its topographic data management system using open source components. The new system will be based on QGIS and PostgreSQL. The goals of the renewal are:
- Utilization of new technologies and standards
- Advancement in the transition from producing map data to producing spatial data
- Enhancement of the quality and timeliness of data
- Enhancement of the production through automation and better tools
The current system has been in use for over 20 years and has been developed throughout its lifespan. NLS is planning to replace the current production system after the first phase of development in 2025.
In this talk, I will talk about the status of the development, elaborate the main objectives of the first phase and introduce the published OS components so far. In the first two years of the development the focus was on concurrent data management by 100 operators and on the integration of the stereo mapping tools (proprietary). In addition, we have designed and implemented OS quality assurance tools to ensure the logical consistency of the features concerning the attributes, the geometries and the topology. These tools also include a topological rule set for topographic data management in PostgreSQL.
We have also published some plugins for the operators to improve the digitizing workflow. To facilitate the development work, we have contributed some development tools for QGIS plugin developers. The OS publications of the service and client components of the concurrent data management tools are not yet on the roadmap although our final goal.
The current process of maintaining topographic data includes some field work too. QField is the chosen OS tool for that purpose. Now, we are defining the additional functionalities needed to make the field work efficient enough and to smooth out the data transfer between the main system and the mobile application.
Afterwards, we have yet to make significant progress in the integration of TDMS with the systems that produce and provide products. In relation to our products, we need to find a way to easily maintain base topographic data and its enriched cartographic derivates and place names, as part of the production process.
G3W-SUITE is a modular, client-server application (based on QGIS-Server) for managing and publishing interactive QGIS cartographic projects of various kinds in a totally independent, simple and fast way.
Accessing administration, consultation of projects, editing functions and use of different modules are based on a hierarchic system of user profiling, open to editing and modulation.
The suite is made up of two main components: G3W-ADMIN (based on Django and Python) as the web administration interface and G3W-CLIENT (based on OpenLayer and Vue) as the cartographic client that communicate through a series of API REST.
The application, released on GitHub with Mozilla Public Licence 2.0, is compatible with QGIS LTR versions and it is based on strong integration with the QGIS API.
This presentation will provide a brief history of the application and insights into key project developments over the past year, including:
* new editing functions and greater integration with QGIS tools and widgets in order to simplify the preparation of web cartographic management systems
* QGIS embedded project management
* WMS-T and MESH data management and integration of TimeSeries functions
* on/off management for the individual symbology categories as in QGIS
* integration of the QGIS Processing API to allow the integration of QGIS analysis modules and perform online geographic analysis
* structured management for log consultation on three levels: G3W-SUITE, QGIS-SERVER and DJANGO
The talk, accompanied by examples of application of the features, is dedicated to both developers and users of various levels who want to manage their cartographic infrastructure based on QGIS
Keeping (OGC) Geospatial Web Services up-and-running is best accommodated by continuous monitoring: not only downtime needs to be guarded,
but also whether the services are functioning correctly and do not suffer from performance and/or other Quality of Service (QoS) issues.
GeoHealthCheck (GHC) is an Open Source Python application for monitoring uptime and availability of OGC Web Services.
In this talk we will explain GHC basics, how it works, how you can use and even extend GHC (plugins).
There is an abundance of standard (HTTP) monitoring tools that may guard for general status and uptime of web services.
But OGC web services often have their own error, "Exception", reporting not caught by generic HTTP uptime
checkers. For example, an OGC Web Mapping Service (WMS) may provide an Exception as a valid XML response or
in a error message written "in-image", or an error may render a blank image.
A generic uptime checker may assume the service is functioning as from those requests and an HTTP status "200" is returned.
Other OGC services may have specific QoS issues not directly obvious. A successful and valid "OWS GetCapabilities" response may not
guarantee that individual services are functioning correctly. For example an OGC Web Feature Service (WFS) based on a dynamic database may
return zero Features on a GetFeature response caused by issues in an underlying database. Even standard HTTP checkers supporting "keywords"
may not detect all failure cases. Many OGC services will have multiple "layers" or feature types,
how to check them all?
What is needed is a form of semantic checking and reporting specific to OGC services!
GeoHealthCheck (GHC) is an Open Source (MIT) web-based framework through which OGC-based web services can be monitored. GHC is written in
Python (with Flask) under the umbrella of the GeoPython GitHub Organization. It is currently an OSGeo Community Project.
GHC consists of a web-UI through which OGC service endpoint URLs and their checks can be managed,
and monitoring-results can be inspected, plus a monitoring engine that executes scheduled "health-checks" on OGC service endpoints.
A database stores results, allowing for various forms of reporting.
GHC is extensible: a plugin-system is available for "Probes" to support an expanding number of
cases for OGC specific requests and -checks. Work is in progress to provide a GHC API for various integrations.
Info, sources, demo: https://geohealthcheck.org
GeoSolutions has been involved in a number of projects, ranging from local administrations to global institutions, involving GeoNode deployments, customizations and enhancements. A gallery of projects and use cases will showcase the versatility and effectiveness of GeoNode, both as a standalone application and as a service component, for building secured geodata catalogs and web mapping services, dashboards and geostories. In particular the recent advancements in data ingestion and harvesting workflows will be presented, along with the many ways to expose its secured services to third party clients. Examples of GeoNode’s builtin capabilities for extending and customizing its frontend application will be showcased.
We’re drowning in data, but the geospatial world lags badly behind in versioning tools compared to our software counterparts. Kart (https://kartproject.org) is solving this with a practical open tool for versioning datasets, enabling you to work more efficiently and collaborate better.
We will introduce you to Kart and demonstrate some of the key features, including our QGIS plugin. And we'll highlight what’s coming next on our roadmap.
Since 2022 we have added support for Raster and Point Cloud datasets, and we'll be showing how we build on Kart's versioning and spatial filtering techniques to efficiently navigate, access, and use large and small datasets. For rasters and point-cloud datasets, we'll show how you can get the benefits of Kart without having to duplicate data that is already hosted in S3 in a useful format.
Kart allows you to quickly and easily manage history, branches, data schemas, and synchronisation for large & small datasets between different working copy formats, operating systems, and software ecosystems.
Modern version control unlocks efficient collaboration, both within teams and across organisations meaning everyone stays on the same page, you can review and trace changes easily: ultimately using your time more efficiently.
ST_LUCAS is an open-source system designed for providing harmonized space-time aggregated LUCAS data. LUCAS (Land Use and Coverage Area frame Survey) is an activity managed by Eurostat that performs in-situ surveys (points in 2x2km grid) over Europe every three years since 2006. For each LUCAS point, the land cover and land use classes are examined, five photos taken, and various agro-environmental attributes collected. Eurostat is providing data in plain CSV files. LUCAS nomenclature is changing each survey year, some attributes were removed, added or renamed.
ST_LUCAS was created with the goal to provide harmonized (each LUCAS survey is translated into common nomenclature) and space-time aggregated (for each LUCAS point, a single location and set of harmonized attributes for each survey year are provided) data. The ST_LUCAS system offers analysis-ready data through the Python API and QGIS plugin (“ST_LUCAS Download Manager”), which minimizes obstacles to use the data by the wider audience. Users may easily access land cover/use information about 1 350 847 points covering 28 EU countries measured from 2006 till 2018 by Eurostat. LUCAS points are retrieved from the ST_LUCAS system based on specified spatial, temporal, attribute, and thematic filters. The Python API and QGIS plugin also allow retrieving photos (one facing photo and four landscape photos in the cardinal compass directions) for each LUCAS point. Additionally, two analytical functions are available: user-defined LUCAS land cover classes aggregation and the possibility to translate LUCAS nomenclature into other nomenclatures.
See ST_LUCAS website https://geoforall.fsv.cvut.cz/st_lucas/ for detailed information.
Vectortile ecosystem have made big changes in Web Mapping, especially in terms of Client-side map rendering. Thesedays, costs of producing and streaming tiles have been dramatically reduced by some techniques - tippecanoe, PMTiles… and so on. However we have the problem important but unsolved yet: Dynamic tiles. Techniques which are matured and widely used are for Static tiles. Static tiles are not good at streaming data frequently updated but we sometimes need to dynamically serve such data. In this talk, I’ll survey techniques for Dynamic tiles which already exist and propose the solution for this.
OpenStreetMap (OSM) is the largest crowd-sourced mapping effort to date, with an infrastructure network that is considered near-complete. The mapping activities started as any crowd-sourced information platform: the community expanded OSM anywhere there was a collective interest. Initial efforts were found around universities or hometowns of mappers. Events, such as natural disasters can also trigger a major update. The recent earthquakes in Turkey and Syria lead to a massive contribution by the Humanitarian OSM Team (HOT) of more than 1.7 million buildings in the region in less than a month after the event1. This type of activities result in a map that is of non-uniform completeness, with some areas having all building footprints in, while other areas remain incomplete or even untouched. Currently, with 550 million footprints, OSM identifies between a quarter and half of the total building footprints in the world, if we estimate that there are around 1-2 billion buildings in the world.
A global view on the local completeness of buildings in OSM did not yet exist. Unlike other efforts, that only look at a subset of OSM building data (Biljecki & Ang 2020; Orden et al., 2020; Zhou et al., 2020), we have used the Global Human Settlement Layer (GHSL) to estimate completeness of the entire dataset. The remote sensing dataset is distributed onto a grid of approximately 100x100 meter tiles. In each tile of the grid, the built area of GHSL is compared to the total area of OSM building footprints. The computed ratio is measured against a completeness threshold that is calibrated using areas that were manually assessed.
Using information derived from remote sensing datasets can be problematic: GHSL does not only measure building footprints: it includes any human-built structures, including infrastructure and industrial areas. Next to that, due to sub-optimal input data or failing algorithms, the dataset is not of the same quality as the crowd-sourced data in OSM in areas that are complete. Even with these limitations, a comprehensive global completeness assessment is created. The assessment should not be used as ground truth, but rather as reflection on the OSM building dataset as is and as a guideline for priorities for the future. Statistics on regional completeness can be created and the quality of GHSL could be assessed on countries that are considered to be complete, such as France or the Netherlands.
Climate Change is affecting our daily lives. Already for many years, we are interested in how this will influence agriculture and livelihoods on the African continent. In this talk we will show a tracking methodology with open data and opensource software. The main data source is satellite imagery from METEOSAT (MSG) as well as rainfall estimates by NOAA to show trends in the last 15 years. We will share links to free data and scripts and make a list of all software used in a step-by-step guide.
We are currently living in an era for Earth Observations that maybe 20 years ago we could not image. Petabytes and Petabytes of data are being created, having so much data it is a good problem to have but the next question is how we can make sure that the Data created it is really being used to solve the challenges we are facing on Earth. The Copernicus Program has given us the opportunity of having Open Data from a variety of diverse sensors but at the same time more and more companies are part of the New Space era in the one commercial companies are launching Optical and SAR satellites that are complementing the Open Data sources.
In my daily job doing Partnerships in the industry, I have the chance to work together with most of the New Space companies trying to find the best way to promote how we all can take advantage of all the data we have available from the Open Data sources and the Commercial sources, this can be optical data working together with SAR and how it can be a game changer in many future projects in Earth Observation.
My presentation will go around how in the last few years there are more options to be able to build products helping to solve earth's challenges by taking advantage of the resources we have in the New Space Industry.
In the US, less than 20% of OpenStreetMap (OSM) buildings have a height tag (less than 10% globally). Providing buildings with height tags helps several use cases including 3D map visualization. At Meta, we have begun using open mapping data to estimate building heights and providing them back to the community. At the end of 2022, we used data from city GIS departments to estimate millions of heights and release them to the public through the Daylight Map Distribution (https://daylightmap.org/2022/12/02/building-heights.html). In 2023, we are using publicly available USGS/3DEP aerial lidar and releasing to the public through the Overture Maps Foundation – processing millions of square kilometers. This talk will cover the challenges, algorithm, QA process, and accuracy metrics from this effort. It is our hope that over the course of the year, we can estimate and publish heights for the majority of the buildings in the US and begin work on non-US open data sources as well.
QWC2 (QGIS Web Client 2) is the official web application of QGIS, that allows you to publish your projects with the same rendering, thanks to QGIS Server. The environment is composed of a modern responsive front-end written in JavaScript on top of ReactJS and OpenLayers, and several server-side Python/Flask micro-services to enhance the basic functionalities of QWC2 and QGIS Server.
QWC2 is modular and extensible, and provides both an off-the-shelf web application and a development framework: you can start simple and easy with the demo application, and then customize your application at will, based on your needs and development capabilities.
This talk aims at introducing this application and to show how easy it is to publish your own QGIS projects on the web. An overview of the QWC2 architecture will also be given. It will also be an opportunity to discover the last new features that have been developed in the past year and ideas for future improvements.
Radiant Earth is building a new data sharing utility called Source Cooperative that aims to make it trivially easy for data providers to publish data on the Internet. Source Cooperative is the next generation of Radiant MLHub which Radiant Earth built to share Earth observation training datasets. In this talk, we will share lessons learned about sharing data from working with NASA, Planet, Sinergise, AWS, Microsoft, and others. We will also share how we’re applying those lessons to create Source Cooperative.
Based on the implementation of a set of forms in Kobo Toolbox, an information flow for the Fire Management Commission of the Argentine Republic was created to be able to integrate from the field the fire reports (on line / off line) in a simple way and their different stages of evolution. The automation of the ingestion to a Geonode, as a geospatial data manager allows the integration with weather forecast data, near real time information, fire incidences, hot spot detection and predictive fire indexes.
The integration is done with the Airflow tool, which guarantees integration and monitoring of information flows, simplifying the process during incidents.
The presentation will provide a comprehensive introduction to GeoServer's own authentication and authorization subsystems. The authentication part will cover the various supported authentication protocols (e.g. basic/digest authentication, CAS, OAuth2) and identity providers (such as local config files, database tables, and LDAP servers). It will also cover the recent improvements implemented with the OpenID integrations and the refreshed Keycloak integration.
It will explain how to combine various authentication mechanisms in a single comprehensive authentication tool, as well as provide examples of custom authentication plugins for GeoServer, integrating it in a home-grown security architecture. We’ll then move on to authorization, describing the GeoServer pluggable authorization mechanism, and comparing it with an external proxy-based solution. We will explain the default service and data security system, reviewing its benefits and limitations.
Finally, we’ll explore the advanced authorization provider, GeoFence. The different levels of integration with GeoServer will be presented, from the simple and seamless direct integration to the more sophisticated external setup. Finally, we’ll explore GeoFence’s powerful authorization rules using:
- The current user and its roles.
- The OGC services, workspace, layer, and layer group.
- CQL read and write filters.
- Attribute selection.
- Cropping raster and vector data to areas of interest.
OpenMapTiles is an open-source set of tools for processing OpenStreetMap data into zoomable and web-compatible vector tiles to use as high-detailed base maps. These vector tiles are ready to use in MapLibre, Mapbox GL, Leaflet, OpenLayers, and QGIS as well as in mobile applications.
Dockerized OpenMapTiles tools and OpenMapTiles schema are being continuously upgraded by the community (simplification, performance, robustness). The presentation will demonstrate the latest changes in OpenMapTiles. The last release of OpenMapTiles greatly enhanced cartography and map styling possibilities, especially the enrichment of Points of Interest and improvement of land use or land cover layer. The new version of Natural Earth brought updated data to upper zoom levels and included a new OSM OpenMapTiles style, showing all features in well know colors for vector tiles. OpenMapTiles is also used for generating vector tiles from government open data secured by Swisstopo.
Cloud Optimized GeoTIFF (COG) is a file format for storing geospatial raster data, such as satellite imagery and digital elevation models, in a way that allows for efficient access and processing in cloud-based systems. COG is a specialized type of GeoTIFF that is optimized for cloud-based workflows, such as those that use Amazon Web Services (AWS) or Google Cloud Platform (GCP) for storage and processing.
COG achieves its optimized performance by storing the image data in a way that allows it to be easily accessed and processed in the cloud, without the need to download the entire image. This is achieved through a combination of techniques, including tiling, compression, and indexing. COG files are designed to be "lazy loaded", meaning that only the portion of the image that is needed for a specific operation is loaded into memory, reducing the amount of data that needs to be transferred and processed.
In this presentation we will use Python with GDAL and rasterstats for point and polygon query in COG.
For point query in COG we will use GDAL and load only the specific portion of image for which we need to read the value making point query in COG extremely fast. To do this we will first calculate the index of input location in COG using a formula and COG metadata.
For polygon query we will use rasterstats and a geosjson to read aggregated value for that geojson from COG.
The proposed presentation will be of interest to developers, data scientists, and geospatial analysts who work with Cloud Optimized GeoTIFF (COG). It will provide a practical guide to querying the COG using Python.
Various applications with the need of highly detailed road network models emerged within the last decade. Apart from traffic simulations in context of urban planning, especially the automotive industry plays an important role in geodata consumption for development, testing and validation of autonomous driving functions. In this domain, human-centred driving simulation applications with their realistic 3D virtual environments pose the highest demands on real-world data and lane-level road network models. It is not uncommon for such road network data to not only be mathematically continuously modelled, but also to contain all the necessary topological links and semantic information from traffic-regulating infrastructure – such as signs and traffic lights. Schwab and Kolbe [1] give a compact overview of the requirements of such fields of application and describe different domain-specific road data formats, which are commonly used for such tasks. Of these peculiar road description formats, OpenDRIVE [2] evolved as an open industry standard. In 2017 we proposed a driver for conversion of OpenDRIVE’s continuous road geometry elements into standardized GIS geometries according to OGC Simple Features Access [3] via the free and open-source Geospatial Data Abstraction Library (GDAL) [4]. By then, this was the first open source conversion tool from OpenDRIVE into more GIS-friendly encodings. Since then, other OpenDRIVE conversion tools have popped up, such as [5], [6], [7], [8]. But none of those allows such a comfortable integration into common GIS tools like our proposed GDAL extension by, for example, simply dragging and dropping an OpenDRIVE dataset into QGIS. We now present a refurbished version of our OpenDRIVE GDAL driver which is based on the novel C++ library libOpenDRIVE. It integrates well in GDAL’s new CMake building process and offers a more convenient starting point for developers and researchers who want to bring OpenDRIVE data easily into context with other geodata such as with aerial images, OpenStreetMap or cadastral data. Apart from OpenDRIVE, other specialized road network description formats are crucial to the automotive engineering and research domain. Where Road2Simulation [9] and laneLet2 [10] already come along in GIS-friendly encodings, RoadXML and NDS Open Lane Model [11] could also profit from such a GDAL-based conversion approach. By bringing the domains of automotive engineering and GIS closer together we hope to stimulate interdisciplinary knowledge transfer and the creation of an interconnected research community.
The CARTO Analytics Toolbox (AT) is a collection of spatial functions that add spatial capabilities to Data Warehouses. At the moment, BigQuery, Snowflake, Redshift and PostgreSQL versions are available.
This talk will show some of the main functions of the AT, and discuss some examples of spatial data analysis performed in different DWs. Special emphasis will be put on the functionality related to spatial indexes, particularly H3 and Quadbin.
The Analytic Toolbox functions are also the building blocks for other tools both from CARTO and outside of CARTO, which will be briefly introduced as well.
Imagine a future where entire communities can harness the power of the sun to fuel their homes and businesses, reducing their dependence on traditional energy sources and helping to build a more sustainable world. At FOSS4G, I am excited to share with you a groundbreaking project that is making this vision a reality in Kosovo, using the latest geospatial technology.
Through the USAID funded Kosovo Energy Security of Supply (KESS) activity, DT global is working to promote sustainable energy solutions in Kosovo. A partnership between DT Global and DevGlobal, are leveraging the power of drones, GIS software, and open-source machine learning models to revolutionize the way we evaluate the solar potential of individual structures. By accurately delineating the boundaries of rooftops using drone imagery, we can then apply cutting-edge photogrammetry analytics to determine the optimal placement of solar panels.
But we're not stopping there. By training the Ramp* open buildings model to successfully identify and delineate rooftops in Kosovo, using data obtained from the Kosovo Cadastral Agency's 2023 high-resolution aerial survey campaign, we are laying the groundwork for a national-level approach to mapping building footprints that can be utilized for a range of applications beyond evaluating rooftop solar potential.
*Ramp is an open-source machine learning model and toolset for extracting building footprints from high-resolution satellite imagery at scale.
FOSS4G 2023 conference closing session.
Announcement of the 2023 Sol Katz Award recipient.
OSGeo Foundation Annual General Meeting