Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 135 Next »

The OpenEarth philosophy aims to collect and disseminate environmental and lab data sets in a project-superseding manner rather than on a project-by-project basis. We believe that science and engineering have become so data-intensive that data management is beyond the capabilities of* individual researchers. Data management needs to migrate from artisanal methods to 21st century technology. This implies data management needs to team up with IT-professionals, and v.v.. This belief is wide-spread, and is called the 4th paradigm. We recommend to read the 4th paradigm book. It illustrates the spreading belief that all sustainable solutions to manage data should be web-based and involve communities. OpenEarth aims to be a 4th paradigm workflow solutions to let scientist and engineers collaborate in communities over the web. Such communities should not only deal with data, but deal with numerical models and analysis tools as well. Data cannot be treated separately from the rest of science. Therefore OpenEarth aims to be an integral workflow for data, models and tools.

To be an effective and sustainable 4th paradigm solution, OpenEarth has identified the most promising international standards for exchange of data over the web. These standards come from different realms. These standards are shown in the scheme below. We aim to work with all of these standards, but currently only use the bold ones on a daily basis. These include subversion to store not only the raw data, but also the processing software (scripts, settings) under version control using the web 2.0 wikipedia approach: everyone can sign up for write access. This allows us to naturaly attribute versions to data, an aspect that lacks in most of today data management solutions.
For standardized data we use the netCDF format (NASA and OGC standard) with the THREDDS OPeNDAP for dissemination of TBs of data over the web. For ecological data with a overwhelming amount of meta-data, we use a plain-vanilla Relational DataBase Manegemtn System (RDBMS). We chose the powerful, open source PostgreSQL implementation with PostGIS spatio-temporal add-on. We are working on adopting dedicated spatio-temporal standard as well. These standard allow for live server-side processing on the data to meet the demands of the user. They deliver tailored data. The OGC consortium is the international body for secifications of these standards. The EU INSPIRE directive prescribes these standards. For typical GIS data (flat, 2D or 2.5D) we already work with postgis, geoserver and geonetwork. However, these so-called WxS protocols still lack implementation in operational software for many specific demands of time-dependent, 3D, curvi-linear data products in our field. We do not develop WxS software ourselves, but just wait for the open source implementations, most under OSGeo umbrela, to cover the demands of our our field. By far the most promising WxS client and server implementation in ADAGUC by the Dutch met office KNMI. ADAGUC not only implemented the WCS standard to request data over the web veyr fast, but also the WMS standard to request images. For exchange of graphics of data, we chose to start working with the KML standards, the standard behind Google Earth that was also adopted as standard by OGC.

The data collection procedure and the relation between those standards is explained in the [OpenEarth Data Standards] document, developed in the framework of the EU FP7 Project MICORE anfd Building with Nature. The basis the 3-step ETL procedure well-known in the database world. ETL describes the process to Extract data from somewhere, Transform it to the strict database datamodel requirements, and Load it into the database. We extend ETL with one crucial extra step: provide the data to users via the web. We believe that any effective data management solution should include the user both at the start of the ETL process and and the end. Loading data into the database and using data from the database should be possible from the work environment of the user. In the sketch above we explicitly included client and server to highlight the paramount importance of easy and immediate web-based Provide mechanisms of the data, that are not covered by ETL.

Numerous other datasets have been or are being uploaded continually in the MICORE and Building with Nature research programmes. And OpenEarth is not the only initiative to share and disseminate government-paid Earth science data freely on the web using open standards. We made an inventory of related initiatives. Our aim is to spread the use of the open standards and make them stick in our everyday work.

Unknown macro: {html}

<object width="425" height="344"><param name="movie" value=""></param><param name="allowFullScreen" value="true"></param><param name="allowscriptaccess" value="always"></param><embed src="" type="application/x-shockwave-flash" allowscriptaccess="always" allowfullscreen="true" width="425" height="344"></embed></object>

Unable to render content due to system error: java.lang.ClassCastException: com.adaptavist.confluence.contentformatting.LozengeMacro cannot be cast to com.atlassian.renderer.v2.macro.Macro
  • No labels