This notebook demonstrates how the APEx Product Catalogue can be integrated with an openEO-based platform.
Important
Please note that openEO is an open standard, and not all platforms may support every use case. Additionally, the generated STAC metadata may vary based on your project’s requirements. Therefore, we strongly recommend verifying your specific use cases with your chosen provider.
Supported Use Cases
Scripted upload of results generated by openEO(supported by all openEO instances).
Direct publishing of results to object storage (requires support for the openEO ‘workspace’ API).
Direct publishing of results to a STAC catalogue (requires support for the openEO ‘workspace’ API, including merging into an existing catalogue).
In this notebook, we will focus on demonstrating use case 1 in detail, with a brief overview of how to implement use cases 2 and 3.
Requirement already satisfied: openeo in /opt/conda/lib/python3.11/site-packages (0.39.1)
Requirement already satisfied: owslib in /opt/conda/lib/python3.11/site-packages (0.32.1)
Requirement already satisfied: requests>=2.26.0 in /opt/conda/lib/python3.11/site-packages (from openeo) (2.32.3)
Requirement already satisfied: shapely>=1.6.4 in /opt/conda/lib/python3.11/site-packages (from openeo) (2.0.7)
Requirement already satisfied: numpy>=1.17.0 in /opt/conda/lib/python3.11/site-packages (from openeo) (2.2.3)
Requirement already satisfied: xarray!=2025.01.2,>=0.12.3 in /opt/conda/lib/python3.11/site-packages (from openeo) (2025.1.1)
Requirement already satisfied: pandas>0.20.0 in /opt/conda/lib/python3.11/site-packages (from openeo) (2.2.3)
Requirement already satisfied: pystac<1.12,>=1.5.0 in /opt/conda/lib/python3.11/site-packages (from openeo) (1.11.0)
Requirement already satisfied: deprecated>=1.2.12 in /opt/conda/lib/python3.11/site-packages (from openeo) (1.2.18)
Requirement already satisfied: lxml in /opt/conda/lib/python3.11/site-packages (from owslib) (5.3.1)
Requirement already satisfied: python-dateutil in /opt/conda/lib/python3.11/site-packages (from owslib) (2.9.0)
Requirement already satisfied: pyyaml in /opt/conda/lib/python3.11/site-packages (from owslib) (6.0.2)
Requirement already satisfied: wrapt<2,>=1.10 in /opt/conda/lib/python3.11/site-packages (from deprecated>=1.2.12->openeo) (1.17.2)
Requirement already satisfied: pytz>=2020.1 in /opt/conda/lib/python3.11/site-packages (from pandas>0.20.0->openeo) (2024.2)
Requirement already satisfied: tzdata>=2022.7 in /opt/conda/lib/python3.11/site-packages (from pandas>0.20.0->openeo) (2025.1)
Requirement already satisfied: six>=1.5 in /opt/conda/lib/python3.11/site-packages (from python-dateutil->owslib) (1.16.0)
Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/lib/python3.11/site-packages (from requests>=2.26.0->openeo) (3.4.0)
Requirement already satisfied: idna<4,>=2.5 in /opt/conda/lib/python3.11/site-packages (from requests>=2.26.0->openeo) (3.10)
Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/lib/python3.11/site-packages (from requests>=2.26.0->openeo) (2.2.3)
Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/lib/python3.11/site-packages (from requests>=2.26.0->openeo) (2024.8.30)
Requirement already satisfied: packaging>=23.2 in /opt/conda/lib/python3.11/site-packages (from xarray!=2025.01.2,>=0.12.3->openeo) (24.1)
Building the openEO Processing Graph
Before diving into the different use cases, we will first build a simple openEO workflow. In this example, we will calculate the NDVI for a single Sentinel-2 observation over Gran Canaria. This section assumes the reader has basic knowledge of openEO and will help us focus on the integration with the APEx Product Catalogue.
con = openeo.connect("openeo.dataspace.copernicus.eu").authenticate_oidc()
This use case is the most widely supported, as it involves retrieving the STAC metadata of openEO job results, downloading it, and then ingesting it into an APEx Product Catalogue.
Key drawbacks of this method include:
The need to first download data from the openEO backend and then re-upload it, which can be time-consuming and bandwidth-intensive. This process may also be interrupted by network instability. Such cases could be avoided when in use cases 2 and 3.
Additional project-specific code is required to automate this task.
We start this example by downloading the result through an openEO batch job.
job = result.execute_batch("gran_canaria.tiff",title="APEx Demo - Gran Canaria - NDVI")
In this next step we are going to download the job metadata and create a STAC collection using the pystac library.
stac_metadata_dict = job.get_results().get_metadata()stac_metadata_dict["id"] ="gran_canaria_ndvi"# #remove collection assets, we will rely on item links# del stac_metadata_dict["assets"]collection = pystac.Collection.from_dict(stac_metadata_dict)collection.license ="CC-BY-4.0"collection
type"Collection"
id"gran_canaria_ndvi"
stac_version"1.0.0"
description"Results for batch job j-250313110405469ea3095a5b2b06ce85"
Converting an online openEO Collection to a local Collection
An important consideration until now is that we are currently using openEO API URLs in the STAC metadata. It’s important to note that some of these URLs may require authentication. Additionally, openEO only stores batch job results temporarily. Therefore, if long-term access to the results is needed, the files should be stored in a separate, accessible storage solution, such as an S3 bucket. Use cases 2 and 3 provide a solution for this.
The code below provides an example of how to update the STAC metadata to point to a URL outside of openEO. For this example, we use a placeholder value (/tmp).
Currently, there is an open issue aimed at improving this behavior in openEO.
# Remove collection and canoncial links from both the collection and their itemscollection.remove_links(rel="collection")collection.remove_links(rel="canonical")items =list(collection.get_stac_objects(rel=pystac.RelType.ITEM))for i in items: i.remove_links(rel="collection") i.remove_links(rel="canonical")# Normalize all the hrefs to use the new path /tmpcollection.set_self_href("./collection.json")collection.normalize_hrefs('/tmp/', skip_unresolved=True)collection.save(catalog_type=pystac.CatalogType.SELF_CONTAINED)collection
type"Collection"
id"gran_canaria_ndvi"
stac_version"1.0.0"
description"Results for batch job j-250313110405469ea3095a5b2b06ce85"
The collection metadata has now been cleaned up and can be added to the STAC API. Here we create the full collection, but it’s also possible to only add the generated item to a new collection.
At this point, it is also recommended to improve the metadata quality by adding additional metadata properties.
Important note
At this point, the actual Geotiff is still located on the openEO backend, and can be accessed by any tool via the signed url. This url will however expire, so for more permanent catalogues, the data file needs to be moved to a better location. To do this, simply download the tiff file and upload it to an online location of your preference.
CATALOGUE ="https://catalogue.demo.apex.esa.int"
Authentication for APEx Product Catalogue
To interact with an APEx Product Catalogue, you need a valid authentication token. This token ensures that you can write metadata to an APEx-initialized STAC catalogue, provided you have been granted sufficient access rights.
To obtain an authentication token, follow the dedicated guide on generating your OIDC token. This step is essential for securely accessing and modifying catalogue records.
# Helper class to support the authentication with owslib and requests librariesclass BearerAuth(requests.auth.AuthBase):def__init__(self, token):self.token = tokendef__call__(self, r): r.headers["authorization"] ="Bearer "+self.tokenreturn rtoken =input("Please provide your OIDC token here")auth = BearerAuth(token)
r = Records(CATALOGUE,auth=Authentication(auth_delegate=auth))print(json.dumps(r.collections(), indent=2))
Use Case 2 and 3: Using the export_workspace Process
In these use cases, openEO exports the collection directly to external storage. This is enabled through the Workspaces API, which provides greater control over your data—allowing you to define where it is saved and from where it is loaded.
There are multiple options for creating a workspace. You can either provision one through a given backend or register your own storage location, such as an S3 bucket, to directly store openEO results.
Since these use cases rely on the availability of the export_workspace process, we first check whether it is supported.