The MashMyData project Combining and comparing environmental science data on the web Alastair Gemmell 1, Jon Blower 1, Keith Haines 1, Stephen Pascoe 2,

Slides:



Advertisements
Similar presentations
Conversion of CPC Monitoring and Forecast Products to GIS Format Viviane Silva Lloyd Thomas, Mike Halpert and Wayne Higgins.
Advertisements

Seminar on Emerging Trends in Data Communication and Dissemination Statistical Data as a Service and Internet Mashups by Zoltan Nagy 19 February 2010.
The Reading e-Science Centre Jon Blower Reading e-Science Centre Environmental Systems Science Centre University of Reading United Kingdom.
South-east Regional e- Research Consortium (SeReRC) Keith Haines, Jon Blower, Dan Bretherton, Alastair Gemmell Reading e-Science Centre University of Reading.
CLADDIER project fundamentals Citation, Location and Deposition in Discipline and Institutional Repositories Sam Pepler Project Manager BADC CLADDIER workshop,
Samford University Virtual Supercomputer (SUVS) Brian Toone 4/14/09.
1 Richard White Design decisions: architecture 1 July 2005 BiodiversityWorld Grid Workshop NeSC, Edinburgh, 30 June - 1 July 2005 Design decisions: architecture.
H The MashMyData Project MashMyData [1] is a NERC (Natural Environment Research Council) funded Technology Proof of Concept project whose aim is to enable.
Extending the Capabilities of Geospatial One-Stop Through Partner-Developed Web-Services April 16, 2010 Federal Geographic Data Committee’s (FGDC) Cooperative.
Dynamic Quick View, interoperability and the future Jon Blower, Keith Haines, Chunlei Liu, Alastair Gemmell Environmental Systems Science Centre University.
Exploring large marine datasets using an interactive website and Google Earth Jon Blower, Dan Bretherton, Keith Haines, Chunlei Liu, Adit Santokhee Reading.
Cloud Computing for Chemical Property Prediction Paul Watson School of Computing Science Newcastle University, UK Microsoft Cloud.
TPAC Digital Library Talk Overview Presenter:Glenn Hyland Tasmanian Partnership for Advanced Computing & Australian Antarctic Division Outline: TPAC Overview.
The use of standard OGC web services in integrating distributed model, satellite and in-situ datasets Alastair Gemmell Jon Blower Keith Haines Environmental.
Discussion and conclusion The OGC SOS describes a global standard for storing and recalling sensor data and the associated metadata. The standard covers.
Athens Feb 2008 EuroMISS and EuroDess Viewing and Validation Services Keith Haines.
Metadata Creation with the Earth System Modeling Framework Ryan O’Kuinghttons – NESII/CIRES/NOAA Kathy Saint – NESII/CSG July 22, 2014.
GADS: A Web Service for accessing large environmental data sets Jon Blower, Keith Haines, Adit Santokhee Reading e-Science Centre University of Reading.
AYAN MITRA CHRIS HOFFMAN JANA HUTCHINS Arizona Geospatial Data Sharing Web Application Development April 10th, 2013.
M i SMob i S Mob i Store - Mobile i nternet File Storage Platform Chetna Kaur.
Climate Sciences: Use Case and Vision Summary Philip Kershaw CEDA, RAL Space, STFC.
Running Climate Models On The NERC Cluster Grid Using G-Rex Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre Environmental.
Introduction to Apache OODT Yang Li Mar 9, What is OODT Object Oriented Data Technology Science data management Archiving Systems that span scientific.
Taverna and my Grid Open Workflow for Life Sciences Tom Oinn
NOCS, PML, STFC, BODC, BADC The NERC DataGrid = Bryan Lawrence Director of the STFC Centre for Environmental Data Archival (BADC, NEODC, IPCC-DDC.
IPlant cyberifrastructure to support ecological modeling Presented at the Species Distribution Modeling Group at the American Museum of Natural History.
DELIVERING ENVIRONMENTAL WEB SERVICES (DEWS) Partners: UK Met Office (Lead Partner), British Atmospheric Data Centre (BADC), British Maritime Technology.
1 All-Hands Meeting 2-4 th Sept 2003 e-Science Centre The Data Portal Glen Drinkwater.
Page 1 Pacific THORPEX Predictability, 6-7 June 2005© Crown copyright 2005 The THORPEX Interactive Grand Global Ensemble David Richardson Met Office, Exeter.
June 20-22, nomads.ncdc.noaa.gov Being developed and integrated to provide one-stop.
Styx Grid Services: Lightweight, easy-to-use middleware for e-Science Jon Blower Keith Haines Reading e-Science Centre, ESSC, University of Reading, RG6.
Composing workflows in the environmental sciences using Web Services and Inferno Jon Blower, Adit Santokhee, Keith Haines Reading e-Science Centre Roger.
Large Scale Nuclear Physics Calculations in a Workflow Environment and Data Provenance Capturing Fang Liu and Masha Sosonkina Scalable Computing Lab, USDOE.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
InSPIRe Australian initiatives for standardising statistical processes and metadata Simon Wall Australian Bureau of Statistics December
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GO-ESSP Workshop, LLNL, Livermore, CA, Jun 19-21, 2006, Center for ATmosphere sciences and Earthquake Researches Construction of e-science Environment.
GAAIN Virtual Appliances: Virtual Machine Technology for Scientific Data Analysis Arihant Patawari USC Stevens Neuroimaging and Informatics Institute July.
GEON2 and OpenEarth Framework (OEF) Bradley Wallet School of Geology and Geophysics, University of Oklahoma
Next Steps: becoming users of the NGS Mike Mineter
Combining the strengths of UMIST and The Victoria University of Manchester “Use cases” Stephen Pickles e-Frameworks meets e-Science workshop Edinburgh,
Interoperability from the e-Science Perspective Yannis Ioannidis Univ. Of Athens and ATHENA Research Center
Metadata for structural science Workshop on research metadata in context Nijmegen, 7–8 September 2010 Simon Lambert STFC e-Science UK.
Next Steps.
Existing knowledge Grey literature and other documents, images, videos, and more represent the mission and work of an agency, but preserving and creating.
Alison Pamment 1, Steve Donegan 1, Calum Byrom 2, Oliver Clements 3, Bryan Lawrence 1, Roy Lowry 3 1 NCAS/BADC, Science and Technology Facilities Council,
© University of Reading 2008www.reading.ac.uk Reading e-Science Centre 9 September 2008 Harmonization of environmental data using the Climate Science Modelling.
Effects of Visualization and Interface Design on User Comprehensibility of Composite Data Asheem Chhetri, Apoorv Wairagade, Mahesh Gorantla, Hanye Xu,
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
Using Google Maps and other OpenSource GIS software for displaying geospatial data Jon Blower, Dan Bretherton, Keith Haines, Chunlei Liu, Adit Santokhee.
PLANETS, OPF & SCAPE A summary of the tools from these preservation projects, and where their development is heading.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
Grid Remote Execution of Large Climate Models (NERC Cluster Grid) Dan Bretherton, Jon Blower and Keith Haines Reading e-Science Centre
SHIWA: Is the Workflow Interoperability a Myth or Reality PUCOWO, June 2011, London Gabor Terstyanszky, Tamas Kiss, Tamas Kukla University of Westminster.
Holding slide prior to starting show. Lessons Learned from the GECEM Portal David Walker Cardiff University
1 2.5 DISTRIBUTED DATA INTEGRATION WTF-CEOP (WGISS Test Facility for CEOP) May 2007 Yonsook Enloe (NASA/SGT) Chris Lynnes (NASA)
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
The GridPP DIRAC project DIRAC for non-LHC communities.
A Collaborative e-Science Architecture towards a Virtual Research Environment Tran Vu Pham 1, Dr. Lydia MS Lau 1, Prof. Peter M Dew 2 & Prof. Michael J.
Using a Simple Knowledge Organization System to facilitate Catalogue and Search for the ESA CCI Open Data Portal EGU, 21 April 2016 Antony Wilson, Victoria.
Store and exchange data with colleagues and team Synchronize multiple versions of data Ensure automatic desktop synchronization of large files B2DROP is.
ECMWF 24 th November 2008 Deploying secure OGC services in front of a heterogeneous data archive. Bryan Lawrence, Phil Kershaw, Dominic Lowe, and Stephen.
Reading e-Science Centre Technical Director Jon Blower ESSC Director Rachel Harrison CS Director Keith Haines ESSC Associated Personnel External Collaborations.
Beyond simple features: Do complex feature types need complex service descriptions?' B.N. Lawrence (1,2), D. Lowe (1,2), S. Pascoe (1,2) and A. Woolf (1).
The future of Delft-FEWS
NERC DataGrid: Googling for Secure Data
Sharing models as social objects through HydroShare
Grid Portal Services IeSE (the Integrated e-Science Environment)
A Web-enabled Approach for generating data processors
Brian Matthews STFC EOSCpilot Brian Matthews STFC
Presentation transcript:

The MashMyData project Combining and comparing environmental science data on the web Alastair Gemmell 1, Jon Blower 1, Keith Haines 1, Stephen Pascoe 2, Phil Kershaw 2, Bryan Lawrence 2, Simon Woodman 3, Hugo Hiden 3 1.Reading e-Science Centre University of Reading 2.Centre for Environmental Data Archival British Atmospheric Data Centre 3.School of Computing University of Newcastle

Outline Background to MashMyData Motivation Challenges Interoperability and project architecture Current state of the project The future of the project

NERC-funded project under the ‘Technology proof of concept’ programme Commenced 1 st February Runs until 30 th June 2011 Aiming to present some of our later outputs at EGU 2011 Here we introduce the project and show its current status and plans for the future. Funded partners are Reading e-Science Centre (ReSC) and the Centre for Environmental Data Archival (CEDA) MashMyData Background

Motivation Environmental scientists use many diverse data sources including: in-situ measurements (e.g. ocean buoys, radiosondes) remotely-sensed data (e.g. satellite, radar) numerical simulations However this results in much heterogeneity of data formats, data access methods, and thus suitable software We want to allow scientists from different disciplines to bridge between a variety of datasets regardless of the underlying data formats etc.

Technical Challenges The MashMyData project is faced with a number of challenges in order to be successful Much overlap between these challenges and a number of important challenges in the wider e-Science community The solutions will potentially be widely applicable in the future Challenges: Dealing with data diversity Performing calculations remotely in a way that scales Accessing secure data, and the delegation problem Enabling traceability and reproducibility

Integrating web services and technologies Recent discussion on the gains of re-using existing e-Science We have identified a number of existing web services and technologies and integrated them in the MashMyData project: Reading e-Science Centre’s ncWMS/Godiva2 Web Map Service (displaying gridded environmental data) Centre for Environmental Data Archival’s Web Processing Service (number crunching for compute-intensive workflows) Newcastle University’s e-Science Central software (upload, workflows, versioning) University of Liege’s DIVA-on-web service (interpolating geospatial point data)

Architecture

Current project status First important step was to add multi-dataset capability to godiva2 viewing portal. As part of this we have added ability to view in-situ point data as well as gridded data This paves the way for mashing up datasets (e.g. produce average or difference of existing datasets) Security is being engineered currently, as is the Web Processing Service (required for mash-up workflows)

Web Interface

Click and drag a layer metadata box into a new position This alters the layer stacking on the map to enable layers to be moved towards the front or back The opacity of the layers can also be modified to reveal those underneath

Web Interface

Interface with e-Science Central User can upload data via the e- Science Central API. Thereafter they can view available data sources and workflows User can run a given workflow on data of choice and this will execute the workflow in e-Science Central This interface with e-Science Central will be invisible to the user – they just know they can upload data, view it and run workflows Files and workflows versioning and metadata are recorded by e-Science Central

Examples of work in progress

Further Work Finish integration with CEDA’s WPS (currently works with a simple test process) This in turn will pave the way for adding mash-up functionality to the web interface. Finish engineering the security solution. This will allow access to secure datasets (e.g. Met Office) for certain authorised users. Continue meetings with test case users to ensure that the system meets their needs (so far so good but relatively early days!)

Thanks! (Not live yet, but currently links to our project page including svn on Google Code)