Use Case DataHub Earth and Environment
Compare changes
Some changes are not shown
For a faster browsing experience, some files are collapsed by default.
Files
3
_posts/2022/2022-10-04-use-case-datahub.md
0 → 100644
+ 89
− 0
The [DataHub](https://datahub.erde-und-umwelt.de/en/) of the Research Field [Earth and Environment](https://www.helmholtz.de/en/research/research-fields/earth-and-environment/) is a research data initiative across the seven centres of the Helmholtz Programme [Changing Earth - Sustaining our Future](https://earthenvironment.helmholtz.de/changing-earth/program/). Within its three compartemental Hubs ATMO, MARE and TERRA, it aims at the development of a distributed and interoperable research data infrastructure as well as operational procedures and data products for the whole Research Field and the Earth System Sciences in general.
In this context, the DataHub is responsible for implementing state-of-the-art policies, tools and best-practices particularly in the fields of research data management (RDM) and research software development across all participatng centres. It hence contributes to the fullfilment of the [FAIR principles](https://www.go-fair.org/fair-principles/) (Findability, Accessibility, Interoperability, Reusability) and fosters open science in the Research Field Earth and Environment.
One of the main objectives of the DataHub is the joint development of thematic viewers, through which products and information from the participating centres are presented to end-users from the scientific, economic, and governmental sectors. As an example, the so-called [Stakeholer-View](https://datahub.erde-und-umwelt.de/en/) integrates prominent and widely-used products like the [German Drought Monitor](https://datahub.erde-und-umwelt.de/en/s/drought) or the [Global Earthquake Viewer](https://datahub.erde-und-umwelt.de/en/s/earthquakes).
Apart from those software services, the DataHub is heavily promoting trainings through [HIFIS Education](https://hifis.net/services/software/training) for collaborative software development via GitLab and other fundamental [software engineering skills](https://events.hifis.net/category/4/) as well as consultation and support for software licensing issues through the [Software Engineering Consulting](https://www.hifis.net/services/software/consulting.html).
One major DataHub project is the joint development of an Earth Data Portal. This portal will serve as a central access point for searching and exploring data and thematic ressources from the research field Earth & Environment and beyond. This will be realized by tailor-made harvesting and mapping from a wide range of existing data archives and repositories like [Pangaea](https://www.pangaea.de), the [World Data Center for Climate](https://www.wdc-climate.de) at the [DKRZ](https://www.dkrz.de/de) or [RADAR4KIT](http://radar.kit.edu) as well as institutional repositories into a single portal infrastructure. The technical basis for the portal is the well-established [Marine Data Portal](https://marine-data.de/), which is now enhanced by community-driven functionalities.
The missing features and functionalities for a data portal covering the entire research field were identified during a requirement analysis, for which the developers made use of the HIFIS Codebase. In particular, by providing dedicated and well-defined [issue templates](https://gitlab.hzdr.de/help/user/project/description_templates.md#description-templates) for *feature requests* and *data / repository integration*, a clear and consistent structure is realized so that both the users and developers can easily add their suggestions, requirements and wishes.
The *feature request template* comes with a brief user story, in which a typical use case for a specific feature or functionality is described (e.g., I want to search for all environmental data in a certain region). By adding a set of pre-defined acceptance criteria, both the developers and the creator of the request can check the status of the feature implementation and close the request once all criteria are met.
The *data integration template* contains pre-defined tasks for providing all needed information like data sources (e.g., through a link to a repository), applied metadata schemas, interfaces as well as information about the data license. This, finally, allows for a convenient and efficient integration of new repositories and other data sources.