banner



Open Science Demonstrator 2025: Call for Participation: Call for Participation (CFP)

1. Executive Summary

Open science aims to make scientific research more transparent, reproducible, and reusable. This document outlines an Open Science Demonstrator, a framework for designing and sharing scientific workflows (research processes) and data products. Others can easily reproduce results, adapt the workflow to new problems, or use the output data. By standardizing how we describe data, methods, and results, we enable researchers to reuse and recombine workflows instead of starting from scratch. This not only accelerates innovation but also builds trust: demonstrating reproducibility in research “shows integrity, inspires trust and respect, and encourages reuse”. Key benefits of this approach include (see the 2024 Open Science Pilot for further detail):

  • Reproducibility: Anyone can rerun the workflow and obtain the same results, which strengthens confidence in scientific findings.

  • Efficiency through Reuse: Researchers can apply existing workflows to new datasets or policy questions (for example, applying a climate analysis from one region to another) without re-developing methods, saving time and resources.

  • Interoperability: Using international standards (such as those from the Open Geospatial Consortium, OGC) ensures that data and tools from different organizations can work together. This makes it easier for governments and institutions to collaborate and share data.

  • Transparency: Important details about data sources, processing steps, and assumptions are explicitly documented. This clarity aids policymakers in understanding how results were obtained, fostering evidence-based decision-making.

In summary, the Open Science Demonstrator provides a template for standardizing scientific workflows. It allows a workflow to be packaged with all necessary information (data inputs, methods, parameters, and outputs) in a way that others can readily execute or modify. Policymakers can support such initiatives to increase the impact of research, ensure better return on investment in science (by avoiding duplicate efforts), and promote trustworthy, open scientific practices on issues of public importance. The initiative is co-sponsored by NASA, ESA, and the European Commission through the FOCAL project (project #101137787).

2. Introduction and Background

Modern scientific research often involves complex data workflows – sequences of steps where data is collected, processed, analyzed, and turned into results. Conceptually, a scientific workflow is similar to a data pipeline that transforms inputs (raw data) into outputs (insights, visualizations, etc.). These workflows can be implemented in various forms (computer scripts, notebooks, specialized software), but regardless of implementation, the goal is to test a hypothesis or fulfill a specific analytical task.

However, sharing and reusing such workflows is challenging. Traditionally, researchers publish papers with method descriptions that humans can read and attempt to follow. This narrative approach falls short when we want machine-readable, executable reproducibility. A written description in a paper may omit parameters or environmental details, making it hard for others to rerun the analysis exactly or adapt it to new data. The lack of formal structure also limits transparency, unless one invests heavily in curating data and methods with persistent references.

Several key issues need to be addressed to make scientific workflows truly reusable and interoperable in an open science context:

  • Access to Data Sources: Workflows depend on data from various sources (e.g., satellite images, sensor readings, survey results). These data sources must remain accessible over time via stable infrastructure. In other words, we need persistent, curated data services so that years later the workflow can still retrieve the same or updated data. Each data source often has its own formats and terminologies; understanding these is necessary to configure the workflow’s data input steps.

  • Publishing Results for Reuse: Beyond getting a result, we must consider how to publish the outputs in a reusable way. This allows one workflow’s outputs to become inputs for another workflow, enabling composed analyses to meet broader goals. For example, the results of a climate model might feed into a policy decision tool – if both adhere to common description standards, this chaining is much easier.

  • Descriptive Elements of Workflows: To reuse a workflow, one must describe what it does in a standardized manner – including the data it requires, the steps it takes (at least in outline), the parameters it uses, and the outputs it produces. The collection of all these descriptions forms a sort of “metadata” for the workflow. Reusability will depend on how well these descriptions are standardized. Common descriptive elements include things like data schemas, process parameters, assumptions, and provenance (lineage information about how each result was obtained). The more we can standardize these descriptions across different workflows, the easier it is to plug one workflow’s component into another or to swap out data sources.

Reusable Workflow Architecture: A foundational concept in making workflows reusable is separating the planning of the workflow from its execution. In a reference architecture for reusable workflows, we maintain an Execution Plan (which describes the sequence of steps, what data to fetch, and what parameters to use) separately from the actual Workflow Execution Engine that runs the steps. In practice, this means you could prepare a plan (perhaps as a configuration file or script) that is general, and then execute it on different computing platforms or with different input parameters as needed. The workflow engine reads the plan and carries out the analysis. By adjusting the plan (for example, pointing to a different data source or changing a threshold parameter), you can reuse the same core workflow in a new context without modifying the underlying code of the analysis steps.

f1
Figure 1. The conceptual diagram illustrates this architecture. The Workflow Execution Plan (red) contains references to data sources and parameters, while the Workflow Execution Engine (right side) reads this plan to perform the execution (yellow)

By adopting such an architecture, we make it possible to adjust workflows to new reuse contexts (for instance, running the same analysis in a different region or with updated data from a later date) simply by tweaking the execution plan. The execution engine can be a “black box” that does not need to change; only the plan and inputs change. This approach requires that the plan be written in a standard, transparent description of the workflow’s requirements (data needed, process steps, etc.). There are clearly many details to describe, and not all can be standardized at once. The strategy is to incrementally standardize the most important elements that enable reuse.

2.1. Challenges in Workflow Reuse and Complexity

Scientific workflows can be very complex, encompassing data collection methods, numerous processing steps, computational models, and visualization or decision-making outputs. Historically, capturing all this complexity for someone else to repeat has been done via scientific papers or technical reports. A paper is essentially a semi-structured description written for human understanding. While papers are indispensable for peer review and general knowledge sharing, they are not sufficient for directly reproducing or adapting the science in an automated way. Important specifics (like exact data retrieval queries, software versions, or intermediate data transformations) might be glossed over or buried in prose, making them easy to miss.

The consequence is that even if a workflow is published, another researcher often has to invest significant effort to recreate it. This is inefficient and prone to error. It also limits transparency – if someone wanted to audit or verify each step, they might not have access to the same environment or might misinterpret a description. True reusable workflows need more rigor in how details are captured and shared.

At the same time, we must recognize variability and avoid one-size-fits-all solutions. Different scientific domains (and even different projects in the same domain) have unique aspects. It’s impractical to impose a single standard that covers every detail of every workflow from the start. Instead, we need a flexible approach:

  • Reuse Patterns and Commonalities: Many workflows share certain patterns. For example, two environmental models might both access meteorological data and produce a time-series output, even if their internal computations differ. It is immediately obvious that similar types of data can and should be handled in similar ways across workflows. If two analyses both use satellite imagery, ideally they should fetch that imagery using a common standard API and format. Leveraging such commonalities means creating standardized formats and access methods (e.g., Web APIs) for those data types. It also means reusing descriptive patterns – for instance, a template for describing a dataset (its source, date, variables, quality, etc.) could be applied wherever that dataset appears in any workflow. By identifying these patterns, we can develop supporting software and libraries that help scientists create, manage, and present these descriptions with less effort.

  • Extensibility for new Elements: While standardizing common elements, the framework must allow introducing new descriptions for novel aspects of a workflow. Each workflow or domain might have unique requirements, so the system should enable defining new terms or schema components while maintaining interoperability. This calls for a mechanism to manage and reuse common elements easily, yet flexibly incorporate new ones when needed. In practice, this might be achieved through a modular approach to schema design (building new pieces from existing building blocks, see below).

  • Incremental Evolution: Not everything can be standardized overnight, especially across all domains of science. A pragmatic plan is to start with a minimal set of standards that deliver immediate reuse value, then gradually expand. For example, one might begin by standardizing how datasets and results are cited and described (since those are fundamental to reuse), then later tackle more complex aspects like how to describe an algorithm’s internal parameters. Standardization will evolve step by step, guided by real-world use cases and feedback.

A strategy to manage this complexity is to use existing examples and new technologies for assistance. Studying current scientific workflows (especially ones that have been successfully shared or repeated) can highlight which descriptive elements are most critical. If many workflows struggle with, say, describing provenance (what data came from where, and how it was processed), then provenance might be a priority area for standardization. Furthermore, the vast legacy of scientific papers and documentation could be mined with AI tools to automatically extract common workflow patterns or terminology. For instance, machine reading of methods sections across thousands of papers might reveal a handful of recurring steps or data types in a field, suggesting where standard descriptions would be most beneficial. Such AI-assisted analysis can complement expert opinion in building our reusable workflow framework.

2.2. Workflow Reuse Scenarios (Use Cases)

To clarify what we mean by “reusing” a workflow, consider the different ways a scientific workflow might be repurposed or repeated. We identify at least five reuse scenarios for science workflows:

  • Reproducibility: The ability to run the same workflow again and obtain the same results. This involves using the original inputs and conditions to validate the outputs and each step of the workflow. (This is the foundation of verification – ensuring the workflow is correctly documented to yield identical outcomes when repeated.)

  • Iteration: Running the workflow to get equivalent results in a new temporal context or with updated data. For example, a workflow that analyzed last year’s data is run again this year to see how results have changed. The steps are the same, but the input dataset is newer or extended.

  • Application (Reuse in a New Context): Reusing the executable workflow in a related context – for instance, applying a workflow developed for one region or domain to a different region or dataset. An example might be taking a flood prediction model built for one river basin and using it in another basin, or using a disease spread analysis developed for one pathogen to analyze another. This typically requires the workflow’s description of data inputs to be general enough that you can swap in a different (but similar) data source, and the processing steps will still work. This scenario is particularly important for interoperability, as discussed below.

  • Transcription: Rewriting or porting the workflow’s execution plan to run on a different infrastructure. For example, converting a workflow expressed as a Python script into a format that runs on a big data platform or rewriting a set of local analysis steps to use a cloud computing service. The core scientific steps remain, but the implementation is transcribed into a new form (perhaps to take advantage of faster hardware or integrate with a specific platform).

  • Pattern Reuse: Adopting and adapting part of a workflow (or the general pattern of a workflow) to perform a related function. Here, one might not reuse the entire workflow, but instead extract a sub-workflow or approach. For example, recognizing that a data cleaning procedure in one workflow could be applied to other datasets, or using the general structure of a successful workflow as a template for designing a new one.

Each of these reuse cases places slightly different requirements on how we describe and modularize workflows. The first (reproducibility) demands complete documentation of everything needed to run the workflow as-is. Iteration and application both require that inputs and parameters can be changed while maintaining the integrity of the process – implying standardized data interfaces and clearly defined parameters. Transcription focuses on portability, meaning the abstract workflow definition should be decoupled from any one software environment. Pattern reuse benefits from workflows being broken into well-defined components that can be mixed and matched.

Of these scenarios, the Application reuse case (using the workflow in a new context) is often the most critical for interoperability and broad impact. In application reuse, you assume the processing steps are black boxes – you don’t need to understand or modify their inner workings – as long as you have a transparent, standardized description of the data they expect and produce, and how to access that data. This scenario is often referred to as "processors" in Earth Observation sciences, e.g., an interferometric processor to process radar signals.

If every step in the workflow clearly states “I need data of type X with these characteristics, accessible via Y protocol”, then someone else can provide a different dataset of type X (from their own region, for instance) in the specified format/protocol and the step will still run. The actual code (the black box) can remain unchanged. This greatly simplifies reuse: instead of rewriting algorithms, one ensures the inputs/outputs and their access methods are well-described and standardized. Thus, much of our standardization effort focuses on data and metadata descriptions rather than the internal logic of processing steps. Interoperability is achieved by agreeing on those interface descriptions.

3. Implementation Plan: Building the Open Science Demonstrator

To implement this vision, we propose an incremental plan that starts with a simple workflow and gradually introduces standardization components. The implementation leverages existing standards and tools as building blocks and proceeds through a series of steps, each adding more capability for description and reuse. In work package three, OGC Building Blocks provide blueprints for structuring data services: OGC API - Features building blocks for feature/tabular data and metadata catalogs (alternatively, OGC API - Records), OGC API - Coverages/Tiles building blocks for gridded raster data, possibly OGC API - EDR (or similar) for convenient access to observation data, and OGC API - Processes to trigger various processes such as workflow deployment and execution.

In work package 4, we will dive more deeply into finer granular building blocks and use the OGC Building Blocks framework as a foundation for enhancing interoperability of specific aspects of reusable workflow and data. OGC Building Blocks can represent entire standards or parts thereof, thus we can develop building blocks for example for for vocabulary binding or provenance information expression. The OGC Building Blocks initiative provides modular specifications we can reuse for this purpose.

Using Building Blocks: The OGC Building Blocks (BB) framework offers a collection of reusable standard components and patterns that align with modern web standards. For sure, many other building block frameworks exist and qualify for experiments in this pilot, for example EOEPCA or EarthCODE Building Blocks.

Many elements we need for describing workflows have already been developed or prototyped within OGC and complementary initiatives. By using these, we avoid reinventing the wheel and ensure compatibility with broader community practices. Specifically, OGC Building Blocks provide standardized content and mechanisms for:

  • OGC API request and response schemas: Standard definitions of how to request data or services and what the response format is. (For example, OGC API – Features defines a RESTful way to query spatial data and get results in JSON.) Using these schemas means our workflow’s data access steps can be made uniform and web-friendly.

  • Data element descriptions (“semantic annotations”) for JSON: Guidelines to annotate JSON data structures with semantics, i.e., linking data fields to well-defined concepts. This often involves JSON-LD or similar techniques to ensure that anyone (or any machine) reading the data knows exactly what each field means (e.g., a field “temp” is identified as “air temperature” with units Celsius, via a reference to a vocabulary). These semantic annotations make data self-describing and unambiguous.

  • Vocabulary usage specifications: Rules or profiles for using controlled vocabularies in data. For instance, specifying that a field should use terms from a particular ontology or code list. This ensures consistency – e.g., everyone uses “precipitation” vs “rainfall” in a dataset according to a standard vocabulary entry. The Building Blocks include ways to formally constrain values to approved terms.

  • Schema composition from components: Methods to compose more complex data schemas out of simpler, reusable pieces. This is key to our modular approach – instead of crafting a huge bespoke schema for a workflow, we assemble it from generic components (like a “Dataset Reference” component, an “API Call” component, a “Provenance Record” component, etc.). The OGC BB framework supports design patterns for such composition, making our specifications more maintainable and interoperable.

  • Provenance schema patterns: Implementation patterns for recording provenance, meaning how a particular data item was produced – which source data and process led to it. The Building Blocks likely incorporate or align with standards like W3C PROV for capturing this. Reusing these patterns means our workflow provenance records will be in a format that others understand and can integrate with (e.g., linking to broader provenance traces or citations).

  • Data cataloguing standards: Standard ways to catalog and describe datasets and resources (for example, metadata standards such as DCAT or OGC API Records). This helps in publishing the inputs and outputs of workflows in a searchable catalog or repository so that others can find and cite them. By using standard catalog structures, the workflow’s data products become part of the broader ecosystem of data that tools and portals can index.

With this strong baseline in place, we outline a step-by-step implementation plan. Each step adds a layer of capability or standardization. The process is iterative – one can loop through some steps as needed to refine the approach. The steps are:

1. Identify Candidate Workflows: Select one or a few example scientific workflows that will serve as the test cases for the demonstrator. These should be workflows that are executable in a certain platform (e.g., in a Jupyter Notebook environment, a specific modeling software, or a workflow orchestration system) and representative of real scientific use cases. Choosing well-understood workflows provides a controlled way to develop and test the standards. For instance, we might pick a published analysis from environmental science that we have access to, which involves multiple data sources and steps.

2. Capture Initial Conditions of the Workflows: For each candidate workflow, gather all relevant information about its current form. This includes:

  • Workflow plans or code: Obtain the workflow’s implementation, such as the Jupyter notebooks, script files, Application Packages, Apache Airflow DAGs, or even FME workbenches (FME – Feature Manipulation Engine – is a data integration tool). Essentially, we need the executable recipe of the workflow in whatever format it currently exists.

  • Data sources: Identify all data inputs the workflow uses. What datasets, databases, or API endpoints are being accessed? For each data source, note its location (URL, file path, etc.) and nature (e.g., a CSV file of observations, a web service providing climate model data, etc.).

  • Data access queries and responses: Where the workflow fetches data from an API or database, record example queries (e.g., an API request URL or database query) and the format of the response data. This helps us understand how data retrieval currently works and what the data looks like coming in.

  • Data access method documentation: Collect any documentation about how to use those data sources (user guides for APIs, metadata descriptions, etc.). This will be useful when we later design standardized access methods – we’ll know the capabilities and constraints of existing services.

  • Execution environment details: It can also be useful to note the computing environment (software libraries, versions, etc.) though this is more relevant for reproducibility than for description standards. (Nevertheless, if environment matters, we can consider describing it too, perhaps as part of provenance or a workflow profile.)

At the end of this step, we should have a clear picture of the workflow’s “as-is” state: what it does, what it needs to run, and how it interacts with external resources.

3. Analyze the Workflows to Identify Key Descriptive Elements: With the collected information, perform an analysis to determine which descriptive elements are needed to formally describe the workflow. Ask: what information would someone (or a machine) need to fully understand or replay this workflow without the original context? For example, you will likely find the need to describe:

  • The inputs (their format, schema, source, temporal/spatial extent, etc.).

  • The parameters and configuration of each step (like algorithm settings).

  • The outputs (their schema, meaning of each output, units).

  • The sequence of steps or dependencies (though we might treat steps as black boxes, we still need to record the order or which outputs feed into which next step).

  • Any assumptions or prerequisites (e.g., “this step assumes the input image is already georeferenced” – that might be a detail to capture).

  • Provenance details required (e.g., tracking which input file version was used).

From this analysis, extract a list of elements that seem general and important. We are essentially listing what a workflow description standard should cover, based on real needs observed in the workflow. At this stage, we might also notice extraneous or overly specific details that we decide to abstract away – for instance, if a script contains a hard-coded file path, we realize the general description needs a variable for file location, not that specific path.

4. (Optional) Use AI to Enrich Understanding: We can explore using artificial intelligence or machine-learning tools to help parse the workflow and related literature. For example, an AI might read the code or notebook and automatically suggest what each step is doing, or it might parse the research paper associated with the workflow to find described inputs/outputs. AI could also help map the workflow’s steps to known concepts (like recognizing that a certain code cell is doing “outlier removal” on data). This could surface descriptive elements that our manual analysis missed. Additionally, AI text mining on scientific literature might identify common patterns (e.g., many workflows in climate science mention “netCDF file of temperature data” – indicating a common input type that warrants a standard descriptor).

While optional, this step can accelerate the identification of reusable components by leveraging a broader knowledge base than just our own experience.

5. Select High-Value Reusable Aspects: From the list of descriptive elements, prioritize a small set that have high re-use value. These are the aspects of the workflow description that will be useful across multiple workflows and domains. For instance, describing a dataset’s spatial resolution or the provenance of a result are broadly useful concepts. In contrast, a very specific parameter name might be unique to one model and not as reusable. We focus first on aspects like:

  • Data input descriptions (since many workflows use data inputs, a standard way to describe them is high value).

  • Key output descriptions (since outputs often feed other workflows or need publication).

  • Provenance tracking information.

  • Common configuration parameters that appear in multiple workflows (e.g., start/end dates for analysis, geographic region of interest). By selecting a handful of these, we set the scope for initial standardization – keeping it manageable.

6. Define Context-Specific Profiles: For the chosen aspects, we may need to profile them to the specific workflow or domain. A “profile” in this sense means a specialization or restriction of a general standard to fit a particular context. For example, if one chosen aspect is “dataset description”, the general standard might allow describing any kind of dataset; our profile for an Earth observation workflow might require certain fields like coordinate reference system, spatial resolution, and temporal coverage to always be included.

We create a context-specific profile that tailors the general descriptors to our workflow’s domain. This ensures the descriptions are both standardized and relevant to the domain at hand (e.g., environmental science). Profiles help maintain interoperability (by conforming to a base standard) while adding necessary detail for a given context.

7. Develop Building Block Schemas (Sub-schemas) for the Key Aspects: Now, using the OGC Building Blocks approach, create or assemble schema components that formalize the descriptions for each key aspect. Essentially, we’re building the pieces of a standard:

  • For data inputs: Create a schema (or reuse an existing one) that can describe a data input to a workflow step. This schema might include fields for the data source (with maybe a URI or service endpoint), data format, subset or query parameters (if the input is a subset of a larger data set), and so on. We might base this on existing standards like STAC (SpatioTemporal Asset Catalog) or OGC API Records, but simplified for our needs.

  • For workflow outputs: Define a schema for output data or results. This could include the data format, the meaning of the output (e.g., “Flood risk map for region X”), units or measurement scales, etc., along with metadata linking it to the workflow that produced it.

  • For provenance records: If not already covered by a building block, define how a provenance entry should look for one step of the workflow (possibly leveraging W3C PROV or OGC’s Patterns). This might include fields like step name, timestamp, input references, output references, and any runtime details (like software version or execution duration).

  • For parameters/configuration: Create a schema fragment that can capture a parameter name, value, and description (and possibly link it to a formal definition if available).

We should not do this from scratch if avoidable. Reusing existing building blocks is ideal. For instance, OGC has ongoing work on “Application Schemas” and JSON schemas with semantic annotations; those can be imported. The goal is to end up with a set of schema building blocks – think of them as LEGO pieces of our workflow description standard.

8. Document Examples and Test the Building Blocks: Take the schemas created in step 7 and apply them to our actual workflow as a test. For example, write a sample description of one input dataset from our workflow using the new schema: fill in its fields with the real info (source URL, format, etc.) and see if the schema captures everything needed. Do the same for an output or a step’s provenance. By creating concrete examples, we can validate whether the building blocks are sufficient or if they need tweaks. This also yields documentation examples that will be useful for others to understand how to use the standard.

Testing might involve using schema validation tools to ensure our example instances conform to the schemas. It could also involve asking someone not involved in development to see if they understand the example – a check on clarity.

9. Integrate the Descriptions into an Application Schema Package: Once the building block schemas are validated in isolation, integrate them into the overall workflow description schema (sometimes called an application schema or target description package). In other words, assemble a complete specification for describing our workflow by combining the relevant blocks. This could be a single JSON Schema or a set of linked schemas that together describe the workflow’s metadata.

For instance, the top-level might be a “Workflow Description” JSON that has sections for “Inputs” (each described by the input schema), “Outputs” (each described by the output schema), “Steps” (each with a reference to inputs/outputs and perhaps a link to a process description if needed), and “Provenance” (which could be a record of how the outputs were generated, linking to each step and input). The integration ensures that all pieces fit together and references among them are consistent (e.g., an output’s provenance can point to the step that created it, which in turn points to the input it used).

At this stage, our candidate workflow has a formal descriptive package that ideally can be published or exchanged. We can think of it as metadata that accompanies the workflow.

10. Apply Vocabulary and Terminology Standards (Qualify with Controlled Vocabularies): To further standardize and avoid ambiguity, we constrain certain fields to use controlled vocabularies. This step involves reviewing our schemas and determining where free text or arbitrary values should be replaced by references to standard terms. For example:

  • If there’s a field for “data type” or “phenomenon”, we might want it to use a controlled vocabulary of phenomena (e.g., a term for “Temperature” from a climate ontology).

  • Units of measure should ideally be standard (perhaps using UCUM codes or the OGC/ISO list of units).

  • Fields like spatial reference system could use EPSG codes or OGC definitions.

  • Keywords or categories might come from a standardized list relevant to the domain.

We modify the schema to qualify these fields: essentially adding constraints or annotations that say “this value must come from X vocabulary”. This improves interoperability because different workflows using the same vocabulary will use identical terms for the same concept. It also allows tools to automatically understand relationships (for instance, two outputs labeled with the same ontology term can be recognized as the same kind of thing).

11. Identify and Source the Vocabularies: For each vocabulary we decided to use in step 10, ensure we have a source for it. Many standard vocabularies exist (e.g., ISO, OGC, or domain-specific ones). We should:

  • Find existing vocabularies for each need (such as climate variable vocabularies for phenomena, library of unit definitions, etc.).

  • If an appropriate vocabulary doesn’t exist, we may define a small controlled list ourselves for the demonstrator (with the intention to align with a broader effort later).

  • Ensure that these vocabularies are accessible in a stable, machine-readable form. Ideally, each term has a URI (a resolvable identifier on the web) that returns information about the term.

For example, OGC’s RAINBOW platform (the OGC Definitions Server) can host such terms. RAINBOW is designed as a repository of concepts and definitions, providing each term a persistent URI and linking to its description in various formats (HTML, JSON-LD, etc.) . We may publish our demonstrator’s custom terms (if any) on a service like RAINBOW, or use RAINBOW’s existing content for standard terms. Using a service like this ensures the terms we reference in our workflow descriptions are FAIR (Findable, Accessible, Interoperable, Reusable) and governed (so that, for example, “Air Temperature” has a consistent definition everywhere).

12. Publish Vocabularies in Machine-Readable Form: Work with the chosen infrastructure (such as RAINBOW or another vocabulary service) to make sure our controlled vocabularies are online and accessible. This may involve uploading definitions or ensuring the terms we need are already present and stable. The outcome should be that any tool or person can resolve a term’s identifier to get its definition. For instance, if our workflow description says the output’s phenomenon is http://www.opengis.net/def/phenomenon/OGC/Atmosphere/AirTemperature (as a made-up example URI), navigating to that URI should provide information that this term means “Air Temperature” in degrees Celsius, etc. This step is about operationalizing the standard terminology aspect.

13. Develop Validation Rules and Tools: To maintain the integrity of our workflow descriptions, we create validation constraints. These could be additional schematron rules, ShACL shapes, or custom scripts that check the consistency and completeness of the descriptions beyond basic schema structure. For example, a rule might enforce that every input dataset listed in the workflow description must have a corresponding entry in the provenance for some output (meaning every input is actually used somewhere), or that the vocabulary terms are used correctly (matching the allowed list). Another rule could check that if an input is from a web service, a URL is provided and responds to a test query.

Validation ensures that when someone writes a new workflow description or modifies one, they haven’t broken the standard. It’s also a step where we test the robustness of our approach: run the validators on the demonstrator’s description to catch any errors or missing pieces.

14. Ensure Data Access Services are Available and Citable: For the demonstrator to be effective, the data sources and services it relies on must be up and reachable by others. This step involves making sure that any web APIs or data servers used in the workflow are properly documented, persistent, and citable. Two scenarios are common:

  • Existing Services: If the workflow uses, say, an open API provided by some agency, we ensure we have a stable reference to it (like a DOI or a well-known URL), and we might create a standardized description for it (using, for example, OGC API Records or another service metadata schema). We incorporate that description into our documentation so others know exactly what service to use, and perhaps provide a link to its documentation.

  • New Services for Demonstration: In some cases, we might stand up new instances of data services specifically for the demonstrator. For example, if originally the workflow read a local file, we might instead deploy an OGC API service that serves that data in a standard way (to demonstrate interoperability improvements). If we do this, we need to host it in a reliable way (even if just for the project’s duration) and provide a persistent identifier.

The phrase “citable” implies that even ephemeral services should have a reference that can be included in scientific publications or documentation. This could mean obtaining DOIs for datasets or using stable naming for APIs. The outcome of this step is that anyone attempting to rerun the workflow can access the required data sources easily, and they can reference those sources in their own work.

15. Execute an Enhanced Version of the Workflow: Now we put everything together by creating a new version of our candidate workflow that follows the open science demonstrator principles. In practice, we modify the original workflow (from step 1) to incorporate standard interfaces and capture metadata during execution. Concretely, the updated workflow should:

  • Use standardized data access: For at least one (preferably all) data retrieval steps, instead of using a proprietary method or a local file, use an OGC API or other standard service. For example, if the original workflow read data from a file, we might replace that with an HTTP call to an OGC API – Features or OGC API – Records service that delivers the same data. This demonstrates the workflow’s ability to fetch data in an interoperable way.

  • Capture configuration parameters: Ensure that any configuration (like input region, time range, thresholds) is not just hard-coded but captured in a config file or in the provenance. This means someone else can see what parameters were used, and potentially change them for a new run. We might refactor the code to read parameters from a JSON or YAML config that aligns with our description schema, rather than being embedded in the code.

  • Optionally use OGC API Processes: If feasible, we incorporate a step where the workflow calls an OGC API – Processes service (the OGC standard for web processing services) to execute a step. For instance, if one step was a heavy computation, we might offload it to a cloud service that implements OGC API Processes. This is not mandatory (hence optional), but including it would show how even the execution of a step can be standardized (not just data access).

  • Provenance logging for each step: Modify the workflow so that as it runs, it produces a provenance record for each step. This could be done by instrumenting the code: after each major step, output a log or file that details what was done, which inputs were used (with identifiers), and what outputs were produced. If using a workflow platform, it might have this capability built-in. The important part is that the information captured matches the provenance schema we devised. For example, if step 3 of the workflow smooths a dataset, the provenance might record: “Step 3 (Smoothing): input = dataURL, parameter = 5-day window, output = temp_smooth.csv (local file reference)”.

  • Conform output data to the specification: If the original output was, say, a chart or a CSV, ensure the final outputs are packaged or formatted according to our output schema. This might involve converting or adding metadata. For instance, instead of just producing an image, we might produce an image with a sidecar JSON that describes the image (what it represents, coordinate system, etc.), aligning with our standard.

  • Produce a complete provenance trace: At the end of the run, compile the step-by-step provenance into a trace document. This trace should list all execution steps and link each step’s inputs, parameters, and outputs. This could be a single JSON document or RDF graph that a user (or machine) can follow to see exactly how the final result was obtained. Note that intermediate outputs might not be saved as permanent files, but they can still be described in the provenance (e.g., “Step 2 output [not saved to disk] was passed as input into Step 3”). The provenance trace essentially encodes the workflow execution after the fact.

This enhanced workflow is now a demonstrator of how to do science in a more open, standard, and reusable way. We should test it by running it in our environment to ensure it still produces the expected scientific result, but now with the bonus that it also produces standardized descriptions and can be adjusted more easily.

16. Validate the Final Outputs and Descriptions: As a final step, we use the validation tools from step 13 and the standards we developed to validate everything:

  • Output data validation: Check that the outputs of the workflow conform to the output schema (for example, if we said output is GeoJSON, is the file valid GeoJSON and does it have the required properties described in our schema?).

  • Metadata validation: Validate the workflow’s description package against the schemas. Ensure the provenance trace is complete and consistent (e.g., every input mentioned in provenance was declared in the inputs section; all required vocabularies are used correctly).

  • Interoperability check: Optionally, we could attempt to use the workflow description to run the workflow in a different environment (if our description is formal enough, maybe another tool can consume it). While this might be beyond the scope of the demonstrator, it’s a good litmus test for how useful the standard is – can someone else take what we’ve produced (the description + the standardized code) and run or adapt it independently?

Any issues found are noted and addressed, leading to improvements in the building blocks or the workflow implementation as needed. After this, we will have a validated example of an open-science-compliant workflow.

4. Key Areas for Standardization

Throughout the above plan, certain elements emerged repeatedly as crucial to standardize for any scientific workflow. These candidate aspects for standardization are broadly applicable to many domains and workflows:

  • Terminology: Consistent use of terms and definitions. This means establishing shared vocabularies for things like phenomena (e.g., temperature, precipitation), processes (e.g., “classification”, “regression”), and even conceptual entities (what constitutes a “dataset” vs a “data product”). Standard terminology ensures that when one scientist says “humidity” and another says “relative humidity,” a machine or user can tell if those are the same or different concepts. Adopting community-agreed vocabularies or ontologies and referencing them (as described with RAINBOW) is key here.

  • Provenance: A standardized way to record how results were obtained. Provenance includes information on the origin of data, the sequence of processing steps, the agents (people or software) involved, and timings. Having a common provenance model (such as the W3C PROV standard or other simplified patterns) allows different workflows’ provenance traces to be understood together. This is critical for trust and reproducibility – one should be able to trace any result back to its raw inputs through a chain of documented steps.

  • Data Access Service Descriptions: A uniform method to describe data services (like APIs or servers) that provide the data for workflows. This could involve standard metadata records for services, including their endpoints, available parameters, and output formats. If every workflow documented its data sources in a standard way, one could easily plug new data sources into workflows or evaluate if two workflows use equivalent data. Standards like OGC API - Records or ISO 19119 (service metadata) could be leveraged.

  • API Access Parameters (Template and As-Used): When workflows call APIs or web services, we should standardize how we represent the template for such calls and the actual calls made. The “template” means a generalized form of the API request (e.g., a URL with placeholders for date or location), and “as used” means the specific instance with those placeholders filled (e.g., the actual URL used on January 1, 2024, for region XYZ). By standardizing this, one can clearly communicate how to call a service for new parameters, and also record exactly what was done in a specific run. This might be achieved with a notation or schema for API calls, possibly building on OpenAPI specifications or the OGC API standards. Essentially, anyone reading the workflow description can see both the pattern of data access and the concrete details for the run that was executed.

  • Data Quality Measures: Workflows often involve data quality considerations (filtering out poor data, calculating uncertainties, etc.). Standardizing how we describe data quality (both of inputs and outputs) can greatly aid reuse. For instance, if an output dataset comes with a quality score or error margin, that should be described in a consistent way so that others can interpret it or even decide if the data is usable for their purposes. Common quality descriptors (e.g., lineage confidence, accuracy, completeness) and metrics should be used. If we integrate quality information into the workflow’s output description in a standard form, it makes the results more comparable and trustworthy.

  • Data Dimensions: Scientific data often has multiple dimensions – spatial, temporal, thematic, etc. Standardizing how we describe these dimensions (for example, the geographic coverage of a dataset, the time period it spans, the resolution or sampling interval, and any other relevant axes like model ensemble member or statistical measure) is important. These are the facets that allow one to align or merge data from different sources. If every dataset description includes, say, spatialExtent, temporalExtent, resolution, coordinateReferenceSystem, etc., then users can quickly assess the compatibility of datasets. Standard representations for these (perhaps using ISO 19115 for spatial/temporal metadata or other community standards) should be adopted.

By focusing on these areas – terminology, provenance, service description, API usage, quality, and data dimensions – we cover the most critical information needed to make workflows self-describing and reusable. These aspects apply “to all cases” in the sense that any scientific workflow, whether in geosciences, life sciences, social sciences, etc., will need to address most of these in order to be fully shareable. They form a checklist for developing standards and tools in the open science demonstrator and beyond.

The Open Science Demonstrator architecture and implementation plan detailed above provide a roadmap for making complex scientific workflows easier to understand, reproduce, and adapt. By separating the workflow plan from execution, leveraging community standards (like OGC Building Blocks), and incrementally building up a library of reusable descriptive components, we enable a future where researchers can share their work as actionable resources rather than just static reports. Policymakers and research funders stand to gain through improved transparency and efficiency: the same workflow can answer new questions with minimal effort, and results come with built-in lineage and documentation.

This approach does not sacrifice technical depth or rigor – in fact, it demands rigorous documentation – but it encapsulates it in accessible formats. Non-technical stakeholders benefit from clearer high-level documentation (e.g., an executive summary of what the workflow does and requires), while technical users benefit from detailed, machine-readable metadata. Together, these practices move us closer to a more collaborative, reliable, and open scientific ecosystem, where knowledge is not just published, but also actively reused to drive further discovery.

5. Participation in this Pilot

This call defines two different types of platforms:

  • Primary Platform: The platform a bidder uses regularly for executing research workflows as a scientist, or the platform that a platform operator manages, or the platform that a platform developer works for.

  • Discovery Platform: A remote platform that allows the discovery of workflows that will then be ported to a primary platform.

This pilot offers opportunities for OGC members to participate in different roles. The pilot will try to support several participants per role. Bidders can bid for any number of roles. If selected, it cannot be guaranteed that all roles can be financially supported.

Role 1: Scientist. The Scientist defines workflows on the primary platform. In this pilot, the scientist supports the other participants by explaining all the steps of the workflow and supporting the redeployment and execution of the workflow on another platform. This other platform may not support the primary platform’s native execution environment.

Role 2: Platform Operator. The Platform Operator provides a platform that supports workflow discovery, execution, and result access. The operator ensures that the platform is accessible to other participants in this initiative and supports the deployment of new workflows on the platform as necessary.

Role 3: Platform Developer. The Platform Developer provides technical expertise and develops the necessary modifications that allow a workflow available on a discovery platform to be deployed and executed on the developer’s primary platform. The Platform Developer may support the transformation from one workflow model to another or concentrate on the redeployment and execution of workflows within the same workflow execution environment.

Role 4: Building Block Developer. The Building Block Developer analyzes the various challenges arising from the use of multiple platforms and integrates the findings into a single model. The building block developer abstracts the individual solutions and creates a universal model that follows the OGC building block concept and methodology. This model can serve as a blueprint for all platforms in the future to simplify the reusability of scientific processes. The building block developer works on OGC Building Blocks and is responsible for the definition of the minimum viable product, the OGC Workflow Building Block.

6. Pilot Organization

This pilot is organized in two parts, which are executed in parallel and partly connected. The first part concentrates on the discovery, redeployment, and execution of workflows across platforms. Roles 1 to 3 work closely together for that purpose.

The second part focuses on developing the OGC Workflow Building Blocks. This second part is the main field for the building block developers and builds on the common parts as illustrated in the figure below.

f2
Figure 2. Pilot organization with main threads for scientists, platform operators, and platform developers (left, orange); and OGC building block developers (purple, right)

Both roles 1-3 and role 4 work together in WP1 and WP2. Then, roles 1-3 focus on the workflow reuse experiments across platforms, whereas role 4 develops OGC Building Blocks for reusable workflows.

The following applies to both parts: The goal is to define a minimum viable product that can be extended in future initiatives. It is not possible to define all aspects of reusable workflows in a single initiative.

6.1. Thematic Scenarios

This pilot shall address different scenarios. We expect bidders to suggest meaningful scenarios. This applies in particular to bidders addressing role D100: Scientist. Other bids are invited to suggest scenarios as well.

The following scenario description provides guidance on what is expected. Technologies and tools are just used to illustrate the workflow description. Bidders are free to provide more detail on the workflow. The bidder shall be familiar with the workflow and explain why this workflow is suited for deployment and execution on another platform.

Example scenario description

The scenario focuses on a multi-stage Earth Observation processing workflow that ingests, processes, and analyzes large-scale satellite data archives to extract environmental indicators (e.g., land cover changes, geophysical signals). The original workflow was designed using a combination of open standards, modular tools, and reproducible environments and has the following key features:

  1. Data Access & Preparation

    1. Accesses large-volume EO datasets (e.g., Sentinel-1 SAR, gridded oceanic or atmospheric products) from open data hubs or platform archives.

    2. Uses OGC-compliant APIs or cloud-based storage interfaces for data discovery and retrieval.

    3. Applies spatial subsetting over a region of interest (e.g., a river basin or coastal area) and a defined time span (e.g., multi-year analysis).

  2. Workflow Architecture

    1. Built using openEO, OGC Application Package, or Jupyter Notebook or similar interface for chaining modular geospatial processing steps.

    2. May include containerized processing chains (e.g., Docker images with SNAP, Doris, or custom Python code).

    3. Incorporates third-party Python libraries (e.g., xarray, Dask, rioxarray) for advanced spatiotemporal analysis.

  3. Platform Portability

    1. The scenario qualifies to demonstrate deployment and successful execution of the workflow on different platforms, such as:

    2. Cloud-based EO exploitation platforms

    3. Kubernetes or Docker-based custom infrastructure

    4. HPC clusters with JupyterHub access

    5. Workflow must be provided as a reproducible Jupyter Notebook, with all dependencies managed via environments or containers.

  4. Reproducibility & Scalability

    1. Designed to handle large, multi-dimensional datasets, leveraging e.g., Dask for parallel computation and memory optimization.

    2. Resulting outputs should include metrics or visualizations that highlight the success of the workflow (e.g., land change maps, displacement fields, anomaly detection).

6.2. Timeline

The following table illustrates the timeline for all work packages. WP1 and WP2 run for the first five weeks, with final deliverables due at the end of week 5. WP3 starts as soon as the workflows identified for reuse are available.

f3

6.3. Work Packages 1-3

As described above, work packages one to three form a unit that deals with the transfer of workflows from one platform to another. In phase 2 of the OSPD, a scenario is implemented (or already exists) that focuses on the reuse of workflows that are native to one platform and are to be set up and executed on another platform. The scenario comprises the following steps:

  1. Finding and Accessing Workflows

    1. Use a combination of EarthCODE’s and national platforms’ cataloguing services to find workflows on external platforms that you want to re-run on your primary platform.

    2. Access detailed documentation and metadata to understand the workflow’s requirements and dependencies.

  2. Adapting Workflows

    1. Use the Open Science Ontology developed in this initiative, together with EarthCODE’s conversion tools, to adapt the workflow to the format supported by your target platform.

    2. Collaborate with the community to get insights and support on adapting workflows.

  3. Re-running Workflows

    1. Once adapted, re-run the workflow on the target platform using EarthCODE’s standardized interfaces.

    2. Publish the adapted workflow in a format that can serve as a best practice for others, and optionally, back to EarthCODE. The goal is to contribute to the community and enhance reproducibility and re-use.

    3. Publish/extend documentation and examples on EarthCODE

6.3.1. Research Questions

In this context, the following research questions must be addressed:

  • How do the various platforms describe their workflows?

  • What does the ontology look like that covers the different terminology and semantics used by the various platforms and underlying technical solutions? What is the minimum viable product ontology?

  • How can one platform execute a non-native workflow?

  • What aspects need to be addressed by integration platforms such as EarthCODE to facilitate cross-platform workflow execution?

In addition to these guiding research questions, OSPD shall address the following aspects:

  • Workflow Standardization and FAIR Principles

    • Standardized metadata and clearly documented workflows become easier to interpret, enabling users to re-create workflows even without native format support. What is the ideal model for workflow metadata and documentation?

    • Persistent identifiers (DOIs) for workflows and datasets help ensure that workflows are always findable and reproducible across different platforms. What mechanisms are used by the various platforms, and what integration support needs to be provided by integration platforms such as EarthCODE?

  • Containerization and Interoperability Tools

    • Many but not all platforms promote the use of containerization technologies (e.g., Docker, Singularity), which encapsulate software dependencies and workflow environments. This approach allows:

      • Workflows to run consistently across different platforms and environments, regardless of native format support.

      • Significant reduction in the complexity associated with setting up environments on platforms without native workflow support.

    • What role do these containerized computing environments play in daily operations? Is it possible to re-use an openEO workflow directly on a platform that supports OGC Application Packages or Jupyter Notebooks?

  • Reusable Open-Source Building Blocks

    • Open-source workflow components and reusable software modules that can be integrated into various platforms play an increasingly important role:

      • By providing modular, standardized building blocks, users can reconstruct workflows easily by combining these components.

      • This modular approach significantly reduces the barrier to porting workflows between different platforms.

    • What building blocks are most important? What best practices can be identified? What do these building blocks look like?

  • Adaptation Tools and Documentation

    • Comprehensive documentation and integrative ontologies for adapting and porting workflows are essential to supporting scientists in reusing existing workflows.

    • The ontologies should include:

      • Key aspects around reusable components

      • Support for the most prominent technologies, openEO, OGC application package, and Jupyter Notebooks

    • The documentation should include:

      • Code snippets and templates demonstrating how workflows should be structured to enhance portability.

      • Guidelines and best practices for workflow adaptation and migration, enabling users to easily convert or adapt workflows from one platform format to another.

  • Cloud and Hybrid Infrastructure Integration

    • Leveraging cloud-native principles significantly enhances workflows by enabling seamless integration with major cloud providers and hybrid infrastructures, including platforms such as national data processing environments.

      • Workflows designed with cloud-native standards inherently become cloud-ready, simplifying their deployment across diverse cloud-enabled platforms. What cloud-native aspects need to be addressed? Where do cloud-native approaches differ from traditional approaches?

      • Cloud-native approaches also ensure smooth connectivity with existing national or institutional platforms, providing a unified interoperability layer that facilitates straightforward cross-platform operations. What level of cloud-nativeness is already supported by the national platforms? Where can integration platforms such as EarthCODE facilitate further integration?

6.4. Work Package 4

Work Package 4 "OGC Workflow Building Blocks" complements the work done in work packages 1-3. The goal of this work package is address as many of the standardization challenges and implementation plan steps as possible. As previously stated, the implementation plan describes an iterative process and this pilot does not expect to address all aspects equally.

7. Work Packages

7.1. WP-1: Workflow discovery and evaluation

This work package addresses the discovery phase and identifies the minimal viable product for successful workflow discovery. It defines which elements need to be searchable and the necessary level of information regarding data, dependencies, and other requirements for successful re-use. The key is to understand how the various platforms describe the main aspects of input data, workflow, and workflow products.

Contribution of the various roles:

Scientist: Identifies suitable workflow(s) on discovery platforms that shall be reused on the primary platform and explains these to the other roles as necessary.

Platform provider: Supports the scientist in identifying suitable workflow(s) based on the primary platform’s capabilities.

Platform developer: Develops metadata model and metadata for the identified workflow(s) to the necessary level of detail that makes that workflow findable and “re-deployable”, i.e., all information necessary to understand the workflow’s requirements and dependencies for successful execution on another platform. Address the DOI challenge as part of this work package.

Building Block developer: This persona integrates the various approaches, works on a modular and harmonized metadata model, and defines the requirements for the OGC Workflow building block.

Final deliverables

  • Scientists: All findings are documented in the initiative report D001.

  • Platform operators: All findings are documented in the initiative report D001.

  • Platform developers: Create a metadata model and metadata instances for the identified workflows. Deliver the model and metadata instances in an appropriate format (e.g., JSON-LD, JSON Schema). Document all findings in D001.

  • Building Block developers: Homogenized metadata model for workflow description, execution environment, data, and miscellaneous dependencies. Document all findings in D010.

7.2. WP-2: Open Science Ontology

In this second work package, scientists, platform developers, and building block developers work together to define the minimum viable product for the Open Science Ontology. The ontology should include all aspects required in work package 1 to sufficiently understand each workflow with respect to function, requirements, and dependencies. Platform developers will implement the ontology for a specific platform, i.e., they will make a graph available that allows understanding workflows and their requirements:

f4
Figure 3. High-level view of the Open Science Ontology

The general model of scientific workflow is not dissimilar to any data supply pipeline. A workflow is designed to test a hypothesis, perform a specific analysis, or meet a particular business requirement.

In all cases, it is relevant to consider access to data sources and an infrastructure that provides (or requires) persistent access to curated data sources.

Each data source will have its own data structures and vocabularies that need to be understood to configure such a processing chain.

It is also necessary to consider the requirements to publish results for re-use, allowing such workflows to be combined to meet requirements.

A simple reference architecture for “reusable workflows” thus requires the separation of execution planning from the workflow execution itself. Thus, the workflow can be executed based on the execution plan, and execution parameters can be adjusted to meet the reuse context. This architecture is provided in the following figure. The Open Science Ontology shall ensure that all platforms use mappable terms. As an example, it should be easy for clients to understand that platform One uses the term “results”, whereas platform Two uses the term “outputs” for the same product.

f1
Figure 4. Reusable workflow architecture

Reading the previous figure from left to right, it starts with the formulation of a hypothesis, which shall be analyzed and requires data for this analysis (blue elements).

The data requirements are defined in the input data specification. Together with the data output specification and the execution parameter description, these form key elements of the execution step design. The execution plan defines all steps together (red elements).

Data is stored in data repositories, which use, like all other specifications, vocabularies for unique definitions. Data is available via data access APIs (all in purple).

The execution plan consists of the individual steps, each with its corresponding metadata and parameter value definitions (green elements).

The actual execution event produces an output, which will be stored in a data repository and made available from there. The provenance chain describes how this output was produced. This information is available as records via the corresponding API (yellow elements).

Together, scientists, platform operators, platform developers, and building block developers will:

  1. Define the OGC Open Science Ontology (OSO) as a minimum viable product. Ideally, the OSO draws on best practices and defines the main object classes and their relationships.

  2. Produce mappings of open science terms used in the various platforms to the OSO concepts as necessary.

  3. Instantiate the ontology as a knowledge graph for all relevant resources on a platform.

The main deliverable of WP2 is the open science ontology. Eventually, the ontology needs to become an integral part of each platform. For that reason, the platform developer and the building block developer may jointly explore how the ontology can be used by a platform and may develop the necessary solutions.

Final Deliverables

  • All roles 1-4 together: Create an Open Science Ontology with the core concepts identified in this initiative. A minimum viable product is envisioned here that can be extended in future initiatives. The open science ontology shall be documented in D001 and delivered in an appropriate, machine-readable format.

7.3. WP-3: Workflow reuse experiments

In WP-3, scientists, platform developers, and operators collaborate closely to implement the necessary steps for mapping workflows from one platform to another. The primary goal is to enable workflows, originally native to one specific computational environment, to be adapted and executed effectively on different, non-native platforms. The initiative will start with an initial set of standardized workflow types (e.g., Jupyter Notebook, OGC Application Package, openEO). It is envisioned that his list will be extended in the future.

This initiative aims to enable solid “conformsTo” elements that can be used in catalog entries such as OGC API Record instances. Ideally, the Open Science Ontology will be used in this context.

Key activities within this work package include:

  1. Discovery and Analysis: Scientists identify relevant workflows on a discovery platform that they wish to reuse on their primary platform. They work alongside platform operators and developers to ensure that the workflows are suitable and feasible for the intended reuse scenario. Essential metadata, dependencies, and execution requirements are jointly developed and documented.

  2. Adaptation and Conversion: Platform developers utilize the Open Science Ontology (OSO), developed in earlier work packages, along with supporting conversion tools to adapt and translate workflows into formats compatible with target platforms. This step may involve semantic mapping, conversion of data formats, integration of dependencies, and potentially encapsulation in containerized environments (such as Docker).

  3. Collaboration and Community Engagement: Participants actively engage with the broader community to ensure that workflow mappings are practically relevant, sharing insights and best practices. Scientists, in particular, provide critical domain-specific insights, ensuring that scientific validity and reproducibility are maintained during the adaptation process.

  4. Testing and Validation: Once workflows are adapted, testing occurs on target platforms. Results from the adapted workflows are compared against outputs from original implementations to ensure accuracy and reproducibility. Based on test outcomes, iterative refinements to mappings may occur.

  5. Documentation and Demonstrations: Detailed documentation of the workflow mapping process, including challenges encountered, lessons learned, and best practices developed, is produced. The final deliverables include comprehensive documentation, examples, and demonstrations illustrating the successful reuse and execution of workflows across multiple platforms.

Final Deliverables

  • Scientists, platform providers, and platform developers collaborate in live demonstrations that deliver the work in the form of a live demonstration. The demonstrations illustrate the successful reuse and execution of mapped workflows and are conducted using live screen share presentations. All demonstrations are complemented by comprehensive documentation detailing the steps taken, tools used, encountered challenges, and solutions developed throughout the workflow mapping process by all participants. The documentation is delivered as part of D001.

7.4. WP-4: OGC Workflow Building Blocks

While work packages 1 to 3 focus on the actual platforms and deal with mapping workflows from one platform to another, WP4 focuses more on exploring OGC building blocks. The aim here is to experiment extensively with existing building blocks or those currently under development.

OGC Building Blocks provide reusable content and mechanisms for:

  • OGC API request and response schemas

  • Data element descriptions (“semantic annotations”) for JSON

  • Vocabulary usage specifications

  • Composition of more complete descriptive schemas from simpler components

  • Provenance schema implementation patterns

  • Data cataloguing standards

The workflow in WP is very similar to the processes in WP1 to 3, except that the focus is now entirely on the Building Blocks. Optimizing reusability can start with incremental improvements using existing workflows. In WP4, it is not necessary to transfer a workflow from one platform to another. The following steps are planned.

  1. Identify candidate workflows (executable on a platform)

  2. Capture initial conditions:

    1. Workflow plans (Jupyter notebooks, OGC Application Packages, AiflowDAGs, FME, etc.)

    2. Identify data sources

    3. Acquire examples of data access queries and responses 2.. Acquire data access method documentation

  3. Analyse initial conditions to select key descriptive elements required

    1. Explore the use of AI to interpret workflow and publications to derive this information

  4. Select a small set of aspects that have high re-use value

  5. Define the context (workflow/domain) specific profiles of these needed

  6. Create Building Blocks to define sub-schemas

    1. Data inputs to one or more steps

    2. One or more key workflow output description aspects

    3. Document examples and perform testing

    4. Integrate into the application schema, target description package

  7. Qualify schema building blocks with vocabulary constraints

    1. Identify sources of vocabularies

    2. Publish vocabularies to a machine-readable resource (OGC has tools for this as part of the RAINBOW infrastructure)

    3. Create validation constraints to define and test vocabulary use.

  8. Ensure Data access services are deployed and citable

    1. These may pre-exist and be described in a more standardized way

    2. These may be established to demonstrate potential interoperability improvements (and reusability) of specific execution steps.

  9. Create a version of the workflow that:

    1. Captures configuration parameters of one or more execution steps

    2. Uses OGC API data access steps (in at least one case)

    3. (Optionally) uses OGC API Processes to execute steps

    4. Captures provenance of each step with any relevant data access details (e.g., “API access record”)

    5. Outputs data conforming to the “output specification”

    6. Outputs a provenance trace containing each execution step and all inputs, parameters, and outputs. (Note: outputs may not be persistent between steps)

  10. Validate outputs against the description standardization target (using the output Building Block specification)

The following list includes candidate aspects for standardization (applicable to all cases). Participants can concentrate on individual aspects; it is not necessary to address all aspects.

  • Terminology

  • Provenance

  • Data access service descriptions

  • API access parameters and invocation templates (template)

  • API access parameters and invocation (as used)

  • Data quality measures

  • Data dimensions

Contribution of the various roles:

Building Block Developer:

Deliverables (Building Block Developer):

  • Enhanced OGC Building Blocks documentation

  • OGC Building Blocks for the various candidate aspects for standardization

  • Example vocabularies

  • All results shall be documented in D010. As agreed during the initiative, OGC Building Block documentation and vocabularies will be delivered.

8. Master Schedule

The following table defines the master schedule for the initiative.

Table 1. Master Schedule
Milestone Date Description

M01

26 June 2025

Public Release: Call for Participation

M02

09 August 2025.

Final option to submit questions to the Questions Form. Questions submitted later than 23:59 EDT will not be answered anymore.

M03

14 August 2025

Proposals Due at 23:59 EDT

M04

9 September 2025

Kick-off meeting (virtual)

M05

10 October 2025

WP1 and WP2 completed

M06

12 December 2025

WP3 and WP4 completed, final demonstrations

M07

23 December 2025

Final reports and deliverables due

During the pilot, weekly or biweekly check-ins will be held for participants to discuss progress, highlight challenges, and share perspectives on key issues through video conferences.

9. Bidding and Selection Process

Bidders must indicate which role they wish to play in this initiative. Each role has been assigned a unique identifier:

  • D100: Scientist

  • D110: Platform operator

  • D120: Platform developer

  • D130: Building Block developer

  • D140: Lead architect

Each bidder may apply for any number of roles. When submitting a bid, please ensure that the costs for each role are listed separately. Further information on the bid submission process can be found below.

Bidders should anticipate that the funding allocated for each work item will typically range between $5,000 USD and $35,000 USD. While this range reflects the standard funding envelope, exceptions may be made in rare and well-justified cases, either above or below this range, subject to the specific scope and strategic importance of the proposed contributions.

Bidders are selected based on the evaluation criteria defined below.

9.1. Evaluation Criteria

Role D100, Scientist: Defines and explains workflows; supports redeployment and execution.

  • Demonstrated experience authoring and executing scientific workflows (e.g., Jupyter, openEO).

  • Strong domain expertise (e.g., Earth observation, ocean science).

  • Familiarity with ESA EarthCODE and EOEPCA initiatives.

  • Ability to document and explain workflow logic and data requirements to technical and non-technical audiences.

  • Experience with reproducibility, data provenance, and metadata.

  • Familiarity with cross-platform adaptation or open science best practices.

Role D110, Platform Operator: Provides access to and support for a workflow execution environment:

  • Operational responsibility for a platform supporting scientific workflow execution.

  • Proven support for open standards (e.g., OGC APIs, STAC, WMS/WFS).

  • Familiarity with ESA EarthCODE and EOEPCA initiatives or national platforms.

  • Technical capacity to host and expose workflows, data, and results via interoperable services.

  • Demonstrated ability to support external participants (user management, documentation, API access).

  • Availability of stable infrastructure and persistent data access mechanisms.

Role D120, Platform Developer: Adapts workflows for new platforms; creates descriptive metadata and ontological mappings:

  • Hands-on experience transforming workflows (e.g., reformatting, containerization, adapting metadata).

  • Familiarity with ESA EarthCODE and EOEPCA initiatives or national platforms.

  • Familiarity with platform-specific and cross-platform execution models.

  • Proficiency in JSON-LD, JSON Schema, or similar metadata modeling approaches.

  • Understanding of open science ontologies and semantic technologies.

  • Ability to generate and document metadata instances and mappings.

Role D130, Building Block Developer: Creates reusable schema components and harmonizes metadata models.

  • Deep understanding of modular standards development (e.g., OGC Building Blocks).

  • Experience designing schema components and semantic annotations.

  • Familiarity with controlled vocabularies, ontology engineering, and provenance modeling.

  • Proven ability to generalize from specific workflow implementations to reusable patterns.

  • Experience publishing and validating machine-readable specifications (e.g., using RAINBOW, JSON-LD, ShACL).

D140, Lead Architect: In addition to these roles, this initiative seeks a lead architect and report developer (role D140). The lead architect will lead the technical discussions and ensure that timelines are met and the discussions progress continuously. It is a multifaceted leadership role that spans technical coordination, architectural oversight, standards alignment, and cross-role collaboration. The lead architect will be selected based on the following selection criteria:

  • Technical and architectural expertise

    • Demonstrated leadership in past open science or standards-based research projects

    • Systems Architecture and Standards Proficiency, including deep understanding of workflow architecture concepts: execution planning, decoupled execution engines, containerization (Docker, Singularity), semantic interoperability, and provenance

    • Familiarity with OGC standards: OGC APIs (Features, Coverages, Processes, Records), Building Blocks, RAINBOW infrastructure, etc

    • Expertise in designing or integrating metadata models for workflows (e.g., JSON Schema, JSON-LD, W3C PROV)

    • Ability to harmonize heterogeneous workflow definitions across platforms (e.g., openEO, Jupyter, OGC Application Packages)

  • Implementation and Coordination Skills

    • Track record of deliverables in interdisciplinary settings

    • Ability to coordinate input from scientists, platform operators, platform developers, and building block developers

  • Leadership and Communication

    • Cross-team leadership skills, including strong facilitation skills to lead discussions across diverse stakeholder roles

    • Experience chairing working groups

    • Ability to synthesize technical results into impactful documentation for broad audiences, from scientists to policymakers

    • Lead author or contributor to open science-related publications or technical whitepapers

  • Community Engagement, active presence in open science, geospatial, or data standards communities

The lead architect will deliver reports D001 and D010. All participants contribute to these reports.

9.2. Bidders

A bidder is an OGC organizational member that holds membership in good standing and submits a proposal in response to this call. For "Extra Small Explorer" members, a funding cap of US$10,000 per initiative applies. If this limit is exceeded, a higher-level membership is required. All higher-level memberships have no cap. OGC Individual and Developer members can participate in initiatives but are not eligible for funding. In exceptional cases, membership can be assigned through the sponsor(s).

The OGC welcomes submissions from organizations representing a consortium. It emphasizes that OGC will enter into a contract with a single organization exclusively. It is up to the consortium representative to deal with its other consortium members. The contracted OGC member is fully accountable for the consortium.

10. OGC COSI Program Initiatives

This initiative is being conducted under the OGC Collaborative Solutions and Innovation (COSI) Policy and Procedures. COSI is part of OGC Research, our collaborative lab for experimentation, testing, and innovation. It’s where members come together to explore what works today, evaluate what’s emerging, and shape the future of trusted, interoperable location systems.

The Collaborative Solutions and Innovation Program (COSI) fosters cutting-edge solutions to global challenges by funding and supporting collaborative research and development Initiatives. It brings together broad stakeholders, including individuals, small and large businesses, government agencies, and research institutions, to address critical issues such as natural hazzard and disaster resilience.

OGC COSI initiatives promote rapid prototyping, testing, and validation of technologies, such as location standards or architectures. Within an initiative, OGC Members test and validate draft specifications to address geospatial interoperability requirements in real-world scenarios, business cases, and applied research topics. This approach not only encourages rapid technology development but also determines the technology maturity of potential solutions and increases technology adoption in the marketplace.

10.1. Information on bidding, selection, and key requirements

Responding to the Call For Participation (CFP):

To respond to the CFP as a bidder, you will submit an Online Form in which you describe your proposal. This proposal should include your (the bidder’s) technical solution(s) for each deliverable, cost sharing request(s) for funding, and proposed in-kind contribution(s) to the initiative.

The CFP includes a description of the deliverables against which bidders may submit proposals. Bidders may address technical deliverables, such as implementing a component of an infrastructure, or participatory deliverables, such as contributing to meetings and to writing documents. The timeline for completion of the deliverables is set out in the Master Schedule.

Proposal Evaluations will take place on a per-deliverable basis. Therefore, it is important that all proposals should all be entered into the form on a per-deliverable basis.

Proposals in response to the CFP should be submitted by the deadline listed in the Master Schedule.

Participant Selection and Agreements: Following the submission deadline, OGC will evaluate received proposals, review recommendations with Sponsors, and negotiate Participation Agreement (PA) contracts, including statements of work (SOWs). Participant selection will be complete once PA contracts have been signed with all Participants.

Required attendance at the Kickoff: The Kickoff is a meeting where Participants, guided by the Initiative Architect, will refine the Initiative architecture and settle on specifics to be used as a baseline for prototype component interoperability. Participants will be required to attend the Kickoff, including breakout sessions, and will be expected to use these breakouts to collaborate with other Participants.

Required attendance at Regular Telecons and Meetings: After the Kickoff, participants will meet frequently via weekly telecons (videoconferencing) and in person at OGC Member Meetings. As a minimum, participants are required to attend virtual meetings regularly.

Requirements for Development of Deliverables: Development of Components, Engineering Reports, Change Requests, and other deliverables will commence during or immediately after the Kickoff meeting.

Under the Participation Agreement contracts, ALL Participants will be responsible for contributing content to the documents / Engineering Reports (ERs), particularly regarding their component implementation experiences, findings, and future recommendations. Each participant will be required to provide at least one bullet point per week to the ER on work, progress, technical conversations and decisions, etc., while the ER Editor will be the primary compiler and author on the shared sections such as the Executive Summary. The ER editor is further responsible for capturing all design decisions and lessons learned during the whole initiative execution phase. Compiling the whole report at the end of the initiative does not work!

More detailed deliverable descriptions appear under Types of Deliverables.

Final Summary Reports, Demonstration Event and Other Stakeholder Meetings: Participant Final Summary Reports will constitute the close of funded activity. Further development work might take place to prepare and refine assets to be shown at webinars, demonstration events, and other meetings.

Assurance of Service Availability: Participants selected to implement service components must maintain availability for a period of no less than twelve months after the Participant Final Summary Report milestone.

10.2. Q&A option before the call closes

Questions and Requests for clarification: Bidders have the opportunity to submit questions about the CFP. Questions are due on the date listed in the Master Schedule Question submitters will remain anonymous, and answers will be compiled and published in the CFP clarifications.

Ongoing updates and answers to questions will be added to the CFP Corrigenda Table and the CFP Clarifications Table. The HTML version of the CFP will be updated automatically and appear at the same URL as the original version.

Bidders may submit questions using the Question form.

Appendix A: Pilot Organization and Execution

A.1. Initiative Policies and Procedures

This initiative will be conducted within the policy framework of OGC’s Bylaws and Intellectual Property Rights Policy ("IPR Policy"), as agreed to in the OGC Membership Agreement, and in accordance with the OGC COSI Program Policies and Procedures and the OGC Principles of Conduct, the latter governing all related personal and public interactions. Specifically:

Several key requirements are summarized below for ready reference:

  • Participants are defined in section Participants of the OGC COSI Policies and Procedures

  • Each selected Participant will agree to notify OGC staff if it is aware of any claims under any issued patents (or patent applications), which would likely impact an implementation of the specification or other work product, which is the subject of the initiative. Participant need not be the inventor of such patent (or patent application) in order to provide notice, nor will Participant be held responsible for expressing a belief which turns out to be inaccurate. Specific requirements are described under the "Necessary Claims" clause of the IPR Policy.

  • Each selected Participant will agree to provide more detailed requirements for its assigned deliverables, and to coordinate with other initiative Participants, at the Kickoff event.

A.2. Initiative Roles

The roles generally played in any OGC COSI Program initiative include Sponsors, Bidders, Participants, Observers, and the COSI Team. Explanations of the roles are provided in Tips for New Bidders.

The COSI Team may include an Initiative Director and Initiative Architect(s). Unless otherwise stated, the Initiative Director will serve as the primary point of contact (POC) for the OGC.

Initiative Architect(s) will work with Participants and Sponsors to ensure that Initiative activities and deliverables are properly assigned and performed. They are responsible for scope and schedule control, and will provide timely escalation to the Initiative Director regarding any high-impact issues or risks that might arise during execution.

A.3. Types of Deliverables

All activities in this pilot will result in a Deliverable. These Deliverables generally take the form of a documents, websites, or component implementations.

A.3.1. Documents

Initiative Reports or in short 'Reports', and Change Requests (CR) will be prepared in accordance with OGC published templates. Initiative Reports will be delivered by posting on the (members-only) OGC Pending directory when complete and the document has achieved a satisfactory level of consensus among interested participants, contributors and editors. Initiative Reports are the formal mechanism used to deliver results of the COSI Program to Sponsors and to the OGC Standards Program for consideration by way of Standards Working Groups and Domain Working Groups.

Tip

OGC Reports are produced by using a production pipeline. The base format is asciidoc. The asciidoc is eventually compiled to an OGC Report using the Metanorma production pipeline. The OGC team is available to help with that process. The easiest installation and execution requires a Docker environment. Once installed, it is very simple to use.

A typical Initiative Report shall not exceed 40 pages excluding annexes.

Document content should follow this OGC Document Editorial Guidance. File names for documents posted to Pending should follow this pattern "OGC <Initiative Name>: <Deliverable Name>". Example: "OGC Testbed-20: Integrity, Provenance, and Trust (IPT) Report".

A.3.2. Component Implementations

Component Implementations include services, clients, datasets, and tools.

A service component is typically delivered by deploying an endpoint via an accessible URL.

A client component typically exercises a service interface to demonstrate interoperability.

Participants shall either keep the component operational for at least 12 months after the end of the initiative, or deliver the component to OGC with the necessary license to operate the component on behalf of the participant. The concrete modalities for component delivery will be agreed during the initiative.

IMPORTANT

Under the Participation Agreement contracts, ALL Participants will be responsible for contributing content to the Initiative Reports, including component implementation experiences, findings, and future recommendations. The Lead Editor will be the primary author on shared sections such as the Executive Summary.

All components are delivered in the form of live demonstrations of the component. These live demonstrations will be screen captured and made available optionally in an OGC YouTube Playlist on the OGC YouTube channel.

Participants need to form teams for demonstrations.

A.3.3. Websites

Websites shall be developed so that they can easily be redeployed on OGC servers. Technical details will be discussed with the OGC Team.

The proposals should ideally state requirements for the execution environment. Please be aware that OGC does not allocate any funding or resources to the website during the implementation of this pilot, and hosting responsibilities fall on the proponent.

A.4. Proposals & Proposal Evaluation

Proposals are expected to be brief, broken down by deliverable, and precisely addressing the work items of interest to the bidder.

Proposals will be evaluated based on criteria in four areas: technical, compliance, experience, and management/cost. Further, more specific evaluation criteria, can be provided in the Call for Participation’s main sections.

A.4.1. Technical Evaluation Criteria

  • Technical Approach & Methodology: Clarity, feasibility, innovation, and appropriateness of the proposed approach are needed to effectively meet the objectives outlined in the tender.

A.4.2. Compliance Evaluation Criteria

  • Compliance with Requirements: Extent to which the proposal meets the functional and technical requirements specified in the tender documentation.

A.4.3. Experience Evaluation Criteria

  • Experience & Expertise: Demonstrated expertise, skills, and relevant experience of the team in handling similar projects or technologies, ensuring the team’s capability to deliver the solution.

A.4.4. Management/Cost Evaluation Criteria

  • Cost-Share Contribution: The proposed contribution and suggested in-kind contribution regarding the requested cost-share funding ensure alignment with financial expectations and the project’s goals.

Note that all Participants are required to provide some level of in-kind contribution (i.e. costs for which no cost-share compensation has been requested). As a rough guideline, a proposal should include at least one dollar of in-kind contribution for every dollar of cost-share compensation requested.

All else being equal, higher levels of in-kind contributions will be considered more favorably during evaluation. Participation may also take place by purely in-kind contributions (no cost-share request at all).

Once the proposals have been evaluated and cost-share funding decisions have been made, the COSI Team will begin notifying Bidders of their selection to enter negotiations to become an initiative Participant. Each selected Bidder will enter into a Participation Agreement (PA), which will include a Statement of Work (SOW) describing the assigned deliverables.

A.4.5. Reporting

Participants will be required to report the progress and status of their work regularly via phone conference; details will be discussed at the kick-off meeting. Additional administrative details such as invoicing procedures will also be included in the contract.

Monthly Reporting

The COSI Team will provide monthly progress reports to Sponsors. Ad hoc notifications may also occasionally be provided for urgent matters. To support this reporting, each testbed participant must submit (1) a Monthly Technical Report and (2) a Monthly Business Report by the first working day on or after the 3rd of each month. Templates and instructions for both of these report types will be provided. The reports shall not exceed one page.

The purpose of the Monthly Business Report is to provide initiative management with a quick indicator of project health from each Participant’s perspective. The COSI Team will review action item status on a weekly basis with assigned Participants.

Initiative Participants must remain available for the duration of the timeline so these contacts can be made.

Participant Final Summary Reports

Each Participant should submit a Final Summary Report by the milestone indicated in the Master Schedule. These reports should include the following information:

  1. Briefly summarize Participant’s overall contribution to the Initiative (for an executive audience),

  2. Describe, in detail, the work completed to fulfill the Participation Agreement Statement of Work (SOW) items (for a more technical audience), and

  3. Present recommendations on how we can better manage future OGC COSI Program initiatives.

This report may be in the form of email text or an attached document (at the Participant’s discretion).

Appendix B: Proposal Submission

B.1. General Proposal Submission Guidelines

This section presents general guidelines for submitting a CFP proposal. Detailed instructions for submitting a response proposal using the Bid Submission Form web page can be found in the Step-by-Step Instructions below.

Important

Please note that the content of the "Proposed Contribution" text box in the Bid Submission Form will be accessible to all Sponsors and should contain no confidential information such as labor rates. All financial information will not be shared beyond the OGC Team.

Similarly, no sensitive information should be included in the Attached Document of Explanation.

Proposals must be submitted before the deadline indicated in the Master Schedule.

Information submitted in response to this CFP will be accessible to OGC and Sponsor staff members. This information will remain in the control of these stakeholders and will not be used for other purposes without prior written consent of the Bidder. Once a Bidder has agreed to become a Participant, they will be required to release proposal content (excluding financial information) to all initiative stakeholders. Sensitive information other than labor-hour and cost-share estimates should not be submitted.

Bidders will be selected for cost share funds on the basis of adherence to the CFP requirements and the overall proposal quality. The general initiative objective is to inform future OGC standards development with findings and recommendations surrounding potential new specifications. Bidders not selected for cost share funds may still request to participate on a purely in-kind basis.

Bidders should avoid attempts to use the initiative as a platform for introducing new requirements not included in this Call for Participation. Any additional in-kind scope should be offered outside the formal bidding process, where an independent determination can be made as to whether it should be included in initiative scope or not. Out-of-scope items could potentially be included in another OGC COSI initiative.

Each selected Participant (even one not requesting any funding) will be required to enter into a Participation Agreement contract ("PA") with the OGC. The reason this requirement applies to purely in-kind Participants is that other Participants will likely be relying upon the delivery of their work. Each PA will include a Statement of Work ("SOW") identifying specific Participant roles and responsibilities.

B.2. Questions and Clarifications

Once the original CFP has been published, ongoing updates and answers to questions can be found in the CFP Corrigenda Table and the CFP Clarifications Table

Bidders may submit questions using the Question form.

All inquiries will be published anonymously. Answers will be regularly compiled and published in the CFP clarifications section.

B.3. Proposal Submission Procedures

The process for a Bidder to complete a proposal is set out in the online Bid Submission Form.

The Bid Submission form will be made available shortly after the CFP release. It will include a series of web forms, one for each deliverable of interest. Bidders should remember to submit one form for each deliverable for which they wish to submit a proposal.

New users must create an account before completing a form. Once an account is established, the user may log in and will be taken to a home page indicating the "Status of Your Proposal." The user can return to this page at any time by clicking the OGC logo in the upper left corner.

Any submitted bids will be treated as earnest submissions, even those submitted well before the response deadline. Be certain that you intend to submit your proposal before you click the Submit button on the Review page.

Important

Please consider making local backup copies of all text you are entering into the form in case anything needs to be re-entered. If you encounter any technical problems, please contact the OGC at techdesk@ogc.org.

B.3.1. High-Level Overview

Clicking on the “Propose” link will bring you to the Bid Submission Form.

To complete the form, new users should start by providing organizational information on the “Organizational Background” Page and click “Update” and “Continue”. Existing users should check and confirm the information on the “Organizational Background” Page is correct and click “Continue”.

This will navigate to an "Add Deliverable" page. The user should complete this form for each proposed deliverable.

Tip

For component implementations having multiple identical instances of the same deliverable, the bidder only needs to propose one instance. For simplicity, each bidder should just submit against the lowest-numbered deliverable ID. OGC will assign a unique deliverable ID to each selected Participant later (during negotiations). Example: If the CFP requests two clients with IDs D100 and D101, it is sufficient to submit a proposal against D100. The bid will automatically be considered for both deliverables.

A “Review” link, located on the far right of the screen, navigates to a page summarizing all the deliverables the Bidder is proposing. This Review tab won’t appear until the user has actually submitted at least one deliverable under the Propose tab first.

Tip

Consider regularly creating backup copies of this Review page at various points during proposal creation.

Once the “Submit” button is clicked, the user will receive an immediate confirmation on the website that their proposal has been received. The system will also send an email to the Bidder and to OGC staff.

Tip

In general, up until the time that the user clicks this Submit button, the proposal may be edited as many times as the user wishes. However, this initial version of the form contains no "undo" capability, so please use caution in over-writing existing information.

Under the “Done Adding Deliverables” section at the bottom of this page, a user may attach an additional document with further information and explanations. This document is optional.

Important

No sensitive information (such as labor rates) should be included in the Attached Document.

If this attachment is provided, it is limited to one per proposal and must be less than 5Mb.

This document could conceivably contain any specialized information that wasn’t suitable for entry into a Proposed Contribution field under an individual deliverable. It should be noted, however, that this additional documentation will only be read on a best-effort basis. There is no guarantee it will be used during evaluation to make selection decisions; rather, it could optionally be examined if the evaluation team feels that it might help in understanding any specialized (and particularly promising) contributions.

B.3.2. Step-by-Step Instructions

The “Propose” link takes the user to the first page of the proposal entry form. This form contains fields to be completed once per proposal such as names and contact information.

It also contains an optional Organizational Background field where Bidders (particularly those with no experience participating in an OGC initiative) may provide a description of their organization. It also contains a click-through check box where each Bidder will be required (before entering any data for individual deliverables) to acknowledge its understanding and acceptance of the requirements described in this appendix.

Clicking the Update and Continue button then navigates to the form for submitting deliverable by deliverable bids. On this page, existing deliverable bids can be modified or deleted by clicking the appropriate icon next to the deliverable name. Any attempt to delete a proposed deliverable will require scrolling down to click a Confirm Deletion button.

To add a new deliverable, the user would scroll down to the Add Deliverable section and click the Deliverable drop-down list to select the particular item.

The user would then enter the required information for each of the following fields (for this deliverable only). Required fields are indicated by an asterisk: * Estimated Projected Labor Hours* for this deliverable, * Funding Request*: total U.S. dollar cost-share amount being requested for this deliverable (to cover burdened labor only), * Estimated In-kind Labor Hours* to be contributed for this deliverable, and * Estimated In-Kind Contribution: total U.S. dollar estimate of the in-kind amount to be contributed for this deliverable (including all cost categories).

Tip

There’s no separate text box to enter a global in-kind contribution. Instead, please provide an approximate estimate on a per-deliverable basis.

Cost-sharing funds may only be used for the purpose of offsetting burdened labor costs of development, engineering, documentation, and demonstration related to the Participant’s assigned deliverables. By contrast, the costs used to formulate the Bidder’s in-kind contribution may be much broader, including supporting labor, travel, software licenses, data, IT infrastructure, and so on.

Theoretically there is no limit on the size of the Proposed Contribution for each deliverable (beyond the raw capacity of the underlying hardware and software). But bidders are encouraged to incorporate content by making reference or linking to external materials where possible (rather than inline copying and pasting). There is also a textbox on a separate page of the submission form for inclusion of Organizational Background information, so there is no need to repeat this information for each deliverable.

Important
  • A breakdown (by cost category) of the "In Kind Contribution" may be included in the Proposed Contribution text box for each deliverable.

  • However, please note that the content of this text box will be accessible to all Stakeholders and should contain no confidential information such as labor rates.

  • Similarly, no sensitive information should be included in the Attached Document of Explanation.

The “Proposed Contribution (Please include any proposed datasets)” field can be used to provide a succinct description of what the Bidder intends to deliver for this work item to meet the requirements expressed in the Technical Architecture. This could potentially include a brief elaboration on how the proposed deliverable will contribute to advancing the OGC standards baseline, or how implementations enabled by the specification embodied in this deliverable could add specific value to end-user experiences.

A Bidder proposing to deliver a Service Component Implementation can also use this field to identify what suitable datasets would be contributed (or what data should be acquired from another identified source) to support the proposed service.

Tip
  • In general, please try to limit the length of each Proposed Contribution to about one text page per deliverable.

  • Note that images cannot be pasted into the field Proposed Contribution textbox. Bidders should instead provide a link to a publicly available image.

A single bid may propose deliverables arising from any number of threads or tasks. To ensure that the full set of sponsored deliverables are made, OGC might negotiate with individual Bidders to drop and/or add selected deliverables from their proposals.

B.3.3. Tips for New Bidders

Bidders who are new to OGC initiatives are encouraged to review the following tips:

  • In general, the term "activity" is used to describe work to be performed in an initiative, and the term "deliverable" is used as a noun describing artifacts to be developed and delivered for inspection and use.

  • The roles generally played in any OGC COSI Program initiative are defined in the OGC COSI Program Policies and Procedures, from which the following definitions are derived and extended:

    • Sponsors are OGC member organizations that contribute financial resources to steer Initiative requirements toward rapid development and delivery of proven candidate specifications to the OGC Standards Program. These requirements take the form of the deliverables described herein. Sponsors' representatives help serve as "customers" during Initiative execution, helping ensure that requirements are being addressed and broader OGC interests are being served.

    • Bidder is an OGC organizational member that holds membership in good standing and submits a proposal in response to a COSI CFP. For "Extra Small Explorer" members, a funding cap of US$10,000 per initiative applies. If this limit is exceeded, a higher-level membership is required. All higher-level memberships have no cap. OGC Individual and Developer members can participate in initiatives but are not eligible for funding. A Bidder selected to participate will become a Participant through the execution of a Participation Agreement contract with OGC. Most Bidders are expected to propose a combination of cost-sharing request and in-kind contribution (though solely in-kind contributions are also welcomed).

    • Participants are selected OGC member organizations in good standing that generate empirical information through the definition of interfaces, implementation of prototype components, and documentation of all related findings and recommendations in (Engineering) Reports, Change Requests and other artifacts. They might be receiving cost-share funding, but they can also make purely in-kind contributions. Participants assign business and technical representatives to represent their interests throughout Initiative execution.

    • Observers are individuals from OGC member organizations that have agreed to OGC intellectual property requirements in exchange for the privilege to access Initiative communications and intermediate work products. They may contribute recommendations and comments, but the COSI Team has the authority to table any of these contributions if there’s a risk of interfering with any primary Initiative activities.

    • Supporters are OGC member organizations who make in-kind contributions aside from the technical deliverables. For example, a member could donate the use of their facility for the Kickoff event.

    • The COSI Team is the management team that will oversee and coordinate the Initiative. This team is comprised of OGC staff and OGC consultants.

    • The COSI Team communicates with Participants and other stakeholders during Initiative execution, provides Initiative scope and schedule control, and assist stakeholders in understanding OGC policies and procedures.

    • The term Stakeholders is a generic label that encompasses all Initiative actors, including representatives of Sponsors, Participants, and Observers, as well as the COSI Team.

    • Suppliers are organizations (not necessarily OGC members) who have offered to supply specialized resources such as cloud credits. OGCs role is to assist in identifying an initial alignment of interests and performing introductions of potential consumers to these suppliers. Subsequent discussions would then take place directly between the parties.

  • Proposals from non-members or individual members will be considered provided that a completed application for organizational membership is submitted prior to the proposal.

  • Any individual wishing to gain access to the Initiative’s intermediate work products in the restricted area of the tools used in the initiative (or attend private working meetings / telecons) must be an Observer.

  • Individuals from any OGC member organization that does not become an initiative Sponsor or Participant may still (as a benefit of membership) observe activities by registering as an Observer.

  • Prior initiative participation is not a direct bid evaluation criterion. However, prior participation could accelerate and deepen a Bidder’s understanding of the information presented in the CFP.

  • All else being equal, preference will be given to proposals that include a larger proportion of in-kind contribution.

  • All else being equal, preference will be given to proposed components that are certified OGC compliant.

  • All else being equal, a proposal addressing all of a deliverable’s requirements will be favored over one addressing only a subset. Each Bidder is at liberty to control its own proposal, of course. But if it does choose to propose only a subset for any particular deliverable, it might help if the Bidder prominently and unambiguously states precisely what subset of the deliverable requirements are being proposed.

  • The Sponsor(s) will be given an opportunity to review selection results and offer advice, but ultimately the Participation Agreement (PA) and Statement of Work (SOW) contracts will be formed bilaterally between OGC and each Participant organization. No multilateral contracts will be formed. Beyond this, there are no restrictions regarding how a Participant chooses to accomplish its deliverable obligations so long as these obligations are met in a timely manner (whether a 3rd-party subcontractor provides assistance is up to the Participant).

  • A Bidder may propose against one, several, or all deliverables. In past projects, more participants were assigned one deliverable, and fewer were assigned multiple deliverables.

  • In general, the Participant Agreements will not require delivery of any component source code to OGC.

    • What is delivered to OGC is the behavior of the component installed on the Participant’s machine, and the corresponding documentation of findings, recommendations, and technical artifacts contributed to the Initiative Report(s).

    • In some instances, a Sponsor might expressly require a component to be developed under open-source licensing, in which case the source code would become publicly accessible outside the Initiative as a by-product of implementation.

  • Results of other recent OGC initiatives can be found in the OGC completed initiatives list.

Appendix C: Abbreviations

The following table lists all abbreviations used in this CFP.

AI

Artifical Intelligence

API

Application Programming Interface

ARD

Analysis Ready Data

BiDS

Big Data from Space

CEOS

Committee on Earth Observation Satellites

CGDI

Canadian Geospatial Data Infrastructure

CNES

Centre National d’Études Spatiales

COSI

Collaborative Solutions and Innovation Program

CR

Change Request

DER

Draft Engineering Report

DGGS

Discrete Global Grid System

DWG

Domain Working Group

EGS

Emergency Geomatics Service

ER

Engineering Report

ESA

European Space Agency

FAIR

FAIR Findable Accessible Interoperable Reusable

IER

Initial Engineering Report

LLM

Large Language Model

ML

Machine Learning

NAPL

National Air Photo Library

NRB

Normalized Radar Backscatter

NRCan

Natural Resources Canada

OGC

Open Geospatial Consortium

PA

Participation Agreement

POL

Polarimetric

POC

Point of Contact

Q&A

Questions and Answers

RAG

Retrieval-Augmented Generation

RCM

RADARSAT Constellation Mission

SOW

Statement of Work

SWG

Standards Working Group

TBD

To Be Determined (at a later date)

TC

OGC Technical Committee

TEM

Technical Evaluation Meeting

TIE

Technology Integration / Technical Interoperability Experiment

URL

Uniform Resource Locator

USGS

United States Geological Survey

WCAG

Web Content Accessibility Guidelines

WG

Working Group (SWG or DWG)

Appendix D: Corrigenda & Clarifications

The following table identifies all corrections that have been applied to this CFP compared to the original release. Minor editorial changes (spelling, grammar, etc.) are not included.

Last update: 26 June 2025

Section Description

Appendix B, section B.1 .

Link to submission system fixed

Section 8, Master Schdeule

Q&A webinar replaced by online form

The following table identifies all clarifications that have been provided in response to questions received from organizations interested in this CFP.

Question Clarification

-

-

-

-