JASMIN Success Stories

Slides:



Advertisements
Similar presentations
Cloud Computing at the RAL Tier 1 Ian Collier STFC RAL Tier 1 GridPP 30, Glasgow, 26th March 2013.
Advertisements

Tier-1 Evolution and Futures GridPP 29, Oxford Ian Collier September 27 th 2012.
VO Sandpit, November 2009 CEDA Storage Dr Matt Pritchard Centre for Environmental Data Archival (CEDA)
Take your CMS to the cloud to lighten the load Brett Pollak Campus Web Office UC San Diego.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 6 2/13/2015.
VO Sandpit, November 2009 NERC Big Data And what’s in it for NCEO? June 2014 Victoria Bennett CEDA (Centre for Environmental Data Archival)
Technology Steering Group January 31, 2007 Academic Affairs Technology Steering Group February 13, 2008.
Introduction to DoC Private Cloud
Duncan Fraiser, Adam Gambrell, Lisa Schalk, Emily Williams
Modelling and Data Centre Requirements: CEDA ESGF UV-CDAT Conference December 2014 Philip Kershaw, Centre for Environmental Data Archival, RAL Space,
JASMIN Petascale storage and terabit networking for environmental science Matt Pritchard Centre for Environmental Data Archival RAL Space Jonathan Churchill.
Effectively Explaining the Cloud to Your Colleagues.
VIRTUALIZATION AND CLOUD COMPUTING Dr. John P. Abraham Professor, Computer Engineering UTPA.
Ian Bird LHCC Referees’ meeting; CERN, 11 th June 2013 March 6, 2013
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 7 2/23/2015.
Steven Newhouse, Head of Technical Services European Bioinformatics Institute: ICT Challenges.
Virtualisation Cloud Computing at the RAL Tier 1 Ian Collier STFC RAL Tier 1 HEPiX, Bologna, 18 th April 2013.
CEMS: The Facility for Climate and Environmental Monitoring from Space Victoria Bennett, ISIC/CEDA/NCEO RAL Space.
JASMIN Overview UKMO Visit 24/11/2014 Matt Pritchard.
Climate Sciences: Use Case and Vision Summary Philip Kershaw CEDA, RAL Space, STFC.
Ceph Storage in OpenStack Part 2 openstack-ch,
VO Sandpit, November 2009 e-Infrastructure to enable EO and Climate Science Dr Victoria Bennett Centre for Environmental Data Archival (CEDA)
Jefferson Lab Site Report Kelvin Edwards Thomas Jefferson National Accelerator Facility HEPiX – Fall, 2005.
JASMIN and CEMS: The Need for Secure Data Access in a Virtual Environment Cloud Workshop 23 July 2013 Philip Kershaw Centre for Environmental Data Archival.
Virtualisation & Cloud Computing at RAL Ian Collier- RAL Tier 1 HEPiX Prague 25 April 2012.
RAL Site Report HEPiX FAll 2014 Lincoln, Nebraska October 2014 Martin Bly, STFC-RAL.
VO Sandpit, November 2009 e-Infrastructure for Climate and Atmospheric Science Research Dr Matt Pritchard Centre for Environmental Data Archival (CEDA)
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
Required Data Centre and Interoperable Services: CEDA
Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF.
Microsoft Cloud Solution.  What is the cloud?  Windows Azure  What services does it offer?  How does it all work?  How to go about using it  Further.
RAL Site Report HEPiX Spring 2015 – Oxford March 2015 Martin Bly, STFC-RAL.
Data Preservation at Rutherford Lab David Corney 9 th July 2010 KEK.
Using a Simple Knowledge Organization System to facilitate Catalogue and Search for the ESA CCI Open Data Portal EGU, 21 April 2016 Antony Wilson, Victoria.
Head in the Clouds, feet on the ground David Massey Chief Technology Officer.
IBERGRID as RC Total Capacity: > 10k-20K cores, > 3 Petabytes Evolving to cloud (conditioned by WLCG in some cases) Capacity may substantially increase.
EGI-InSPIRE RI EGI Compute and Data Services for Open Access in H2020 Tiziana Ferrari Technical Director, EGI.eu
RAL Site Report HEP SYSMAN June 2016 – RAL Gareth Smith, STFC-RAL With thanks to Martin Bly, STFC-RAL.
CLOUD COMPUTING Presented to Graduate Students Mechanical Engineering Dr. John P. Abraham Professor, Computer Engineering UTPA.
EGI-InSPIRE RI An Introduction to European Grid Infrastructure (EGI) March An Introduction to the European Grid Infrastructure.
EGI… …is a Federation of over 300 computing and data centres spread across 56 countries in Europe and worldwide …delivers advanced computing.
The EGI Federated Cloud
Cloud readiness assessment
Course: Cluster, grid and cloud computing systems Course author: Prof
Workshop on the Future of Big Data Management June 2013 Philip Kershaw
StratusLab First Periodic Review
Cloud Computing.
Cloud Adoption Framework
Introduction to Data Management in EGI
SCD Cloud at STFC By Alexander Dibbo.
Dagmar Adamova, NPI AS CR Prague/Rez
Current Resources LSST-DEV General Purpose Login/Compute Node. Supports developers and staff. Primary Linux Machine for this use.
Traditional Enterprise Business Challenges
Design and Implement Cloud Data Platform Solutions
University of Technology
WLCG Collaboration Workshop;
Connecting the European Grid Infrastructure to Research Communities
Dr. John P. Abraham Professor, Computer Engineering UTPA
Federated Identity Management: Status and perspectives of EGI
Concept of VLAN (Virtual LAN) and Benefits
Cloud computing mechanisms
Partition Starter Find out what disk partitioning is, state key features, find a diagram and give an example.
Increase and Improve your PC management with Windows Intune
Cloud Computing: Concepts
IT Infrastructure for a Data Science Campus
Joining the EOSC Ecosystem
EOSC-hub Contribution to the EOSC WGs
PayPal Cloud Journey & Architecture
Presentation transcript:

JASMIN Success Stories NERC Scientific Computing Forum 29 June 2017 Dr Matt Pritchard Centre for Environmental Data Analysis STFC / RAL Space

Outline What is JASMIN Success stories What are the services provided to users? How do they access them? Current facts / figures Success stories Service provider stories User stories Challenges Next steps

Logical View CEDA Archive Services Analysis Environment JASMIN Compute and Storage (Lotus + Private Cloud + Tape Store + Data Transfer Zone) Internal Helpdesk CEDA Archive Services Data Centres, Curation, DB systems User management, External Helpdesk CEDA Data Centres IPCC DDC ESGF etc Analysis Environment Compute Cloud: PaaS (JAP + Science VMs + User Management), IaaS, Group Workspaces: Fast Disk & Elastic Tape External Helpdesk NERC Managed Analysis Compute NERC Cloud Analysis Compute

Functional View Key Long Term Archive Storage CEDA JASMIN Long Term Archive Storage Short-Term Project Storage Archive Tape Elastic Tape Group Workspaces Archive gws1 gws2 gws3 NERC Managed Cloud Analysis Compute Unmanaged Cloud Tenancies Managed Cloud Tenancies Interactive Compute sci LOTUS IaaS Batch Compute login bastion xfer IaaS CEDA Service CEDA Services Data Transfer Zone Functional View xfer gridftp

Network View Key Archive Managed Compute Managed Cloud Tenancies CEDA JASMIN External Archive GWS Managed Compute Managed Cloud Tenancies Unmanaged Cloud sci LOTUS Firewall / Router CEDA Servicez Data Transfer Zone login bastion xfer CEDA Services xfer[23] xfer[23] JASMIN head router Other STFC Dept router JASMIN DTZ router gridftp, globus Optical Private Networks perfSONAR RAL Core Switch k9.leeds dtn02.rdf Archive FTP RAL STFC Firewall Met Office ESGF DN (gridftp) Catapult RAL Site access Routers JANET POP

Current facts & figures 1425 JASMIN user accounts 150 Group Workspaces (0.1 TB -> 700 TB) 10PB GWS capacity (>8.5 PB used) 5 PB Archive capacity (>4.5 PB used!) 5000 Compute nodes (LOTUS, virtualization, cloud) >>1000 Virtual machines >20 Cloud tenancies

JASMIN Evolution Phase Cost Storage Compute Network Other 1 £5 M 5 PB Panasas ? PB tape iSCSI arrays 700 cores Initial core network: Gnodal Virtualisation Prototype cloud Light paths JAP 1.5 £0.7 M 0.4 PB Panasas Tape drives, media ET service ? cores Gnodal switch upgrade Expansion of VM estate 2 £5.4 M 7 PB Panasas 6 PB tape 0.9 PB NetApp 3000 cores Major core network redesign & implementation CIS software Cloud management s/w 3 £2 M 2 PB Panasas Tape drives 800 cores “ 3.5 £1.2 M 1.2 PB Panasas 1.2 PB Object Store 5 PB tape 1000 cores misc Support / license renewals 4 TBC

JASMIN data growth

JASMIN Accounts Portal

JASMIN Cloud Portal

Data Transfer Zone “Science DMZ” concept Secure, friction-free path for science data Corporate firewall better able to handle “business” traffic

JASMIN User Stories

JASMIN User Stories COMET: Seismic Hazard monitoring with Sentinel-1 InSAR “We’ve had fantastic support from the team who have helped us to build a suitable system. The level of support provided has really helped us to achieve our goals and I don’t think a bespoke solution like this would be available anywhere else. Due to the enormous volumes of data we’re dealing with (each image is around 8GB when zipped), the collocation of the archive with the JASMIN system is essential.” Emma Hatton, University of Leeds, JAMSIN Conference June 2017

JASMIN User Stories

JASMIN User Stories

JASMIN User Stories

JASMIN User Stories

JASMIN User Stories

JASMIN User Stories

Challenges Scale Variety User expertise Effort Evolution Capital-heavy model Evolution Of Workflows Of Technology Of User expectations Of…

JASMIN Next steps JASMIN Phase 3.5 JASMIN Phase 4 2016/2017 1000 cores added to LOTUS (now in place) Object Store Proof of Concept Limited new fast disk integration 12/13 July JASMIN Phase 4 2017/18 Project now underway “limiting” case in JASMIN Science Case Challenges Storage JASMIN Phase 1 storage end-of-life March 2018 (5PB) Next: mixture of fast/parallel disk AND object store Compute Migration to OpenStack (cloud management infrastructure)

Further information JASMIN Centre for Environmental Data Analysis http://www.jasmin.ac.uk https://accounts.jasmin.ac.uk https://cloud.jasmin.ac.uk https://www.youtube.com/channel/UC11nPZVyjDLjYlS7NvbnlmQ Centre for Environmental Data Analysis http://www.ceda.ac.uk CEDA & JASMIN help documentation http://help.ceda.ac.uk STFC Scientific Computing Department http://www.stfc.ac.uk/SCD/ JASMIN paper Lawrence, B.N. , V.L. Bennett, J. Churchill, M. Juckes, P. Kershaw, S. Pascoe, S. Pepler, M. Pritchard, and A. Stephens. Storing and manipulating environmental big data with JASMIN. Proceedings of IEEE Big Data 2013, p68-75, doi:10.1109/BigData.2013.6691556