Dr. M.-C. Sawley IPP-ETH Zurich Nachhaltige Begegnungen Standing at the crossing point between data analysis and simulation Knowledge Discovery Panel.

Slides:



Advertisements
Similar presentations
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Why Grids Matter to Europe Bob Jones EGEE.
Advertisements

The Grid Professor Steve Lloyd Queen Mary, University of London.
Computing for LHC Dr. Wolfgang von Rüden, CERN, Geneva ISEF students visit CERN, 28 th June - 1 st July 2009.
Introduction to CMS computing CMS for summer students 7/7/09 Oliver Gutsche, Fermilab.
1 The EU e-Infrastructure Programme EUROCRIS Conference, Brussels, 17 th July 2007 Mário Campolargo European Commission - DG INFSO Head of Unit “GÉANT.
C3.ca in Atlantic Canada Virendra Bhavsar Director, Advanced Computational Research Laboratory (ACRL) Faculty of Computer Science University of New Brunswick.
1 Challenges and New Trends in Data Intensive Science Panel at Data-aware Distributed Computing (DADC) Workshop HPDC Boston June Geoffrey Fox Community.
Engenio 7900 HPC Storage System. 2 LSI Confidential LSI In HPC LSI (Engenio Storage Group) has a rich, successful history of deploying storage solutions.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
Client+Cloud The Future of Research Dr. Daniel A. Reed Corporate Vice President Extreme Computing Group & Technology Strategy and Policy.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
Jeffery Loo NLM Associate Fellow ’03 – ’05 chemicalinformaticsforlibraries.
NSF and Environmental Cyberinfrastructure Margaret Leinen Environmental Cyberinfrastructure Workshop, NCAR 2002.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
Introduction to Hadronic Final State Reconstruction in Collider Experiments Introduction to Hadronic Final State Reconstruction in Collider Experiments.
Data-Intensive Science (eScience) Ed Lazowska Bill & Melinda Gates Chair in Computer Science & Engineering University of Washington August 2011.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
WLCG/8 July 2010/MCSawley WAN area transfers and networking: a predictive model for CMS WLCG Workshop, July 7-9, 2010 Marie-Christine Sawley, ETH Zurich.
Mining Large Data at SDSC Natasha Balac, Ph.D.. A Deluge of Data Astronomy Life Sciences Modeling and Simulation Data Management and Mining Geosciences.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
The Grid Prof Steve Lloyd Queen Mary, University of London.
National Center for Supercomputing Applications Observational Astronomy NCSA projects radio astronomy: CARMA & SKA optical astronomy: DES & LSST access:
Frédéric Hemmer, CERN, IT DepartmentThe LHC Computing Grid – October 2006 LHC Computing and Grids Frédéric Hemmer IT Deputy Department Head October 10,
3D stereo scientific & information visualization environments NCSA Strategic Planning Presentation (April 20,2010) Donna Cox, Robert Patterson, Alex Betts,
Definition of Computational Science Computational Science for NRM D. Wang Computational science is a rapidly growing multidisciplinary field that uses.
Distributed EU-wide Supercomputing Facility as a New Research Infrastructure for Europe Gabrielle Allen Albert-Einstein-Institut, Germany Jarek Nabrzyski.
1 Kittikul Kovitanggoon*, Burin Asavapibhop, Narumon Suwonjandee, Gurpreet Singh Chulalongkorn University, Thailand July 23, 2015 Workshop on e-Science.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Computational Modelling Within Johnson Matthey Technology Centre Glenn Jones.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
European Organization for Nuclear Research Organisation Européenne pour la Recherche Nucléaire High-Energy Physics Data Delivering Data in Science ICSTI.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Pascucci-1 Valerio Pascucci Director, CEDMAV Professor, SCI Institute & School of Computing Laboratory Fellow, PNNL Massive Data Management, Analysis,
“Big Data” and Data-Intensive Science (eScience) Ed Lazowska Bill & Melinda Gates Chair in Computer Science & Engineering University of Washington July.
The DutchGrid Platform – An Overview – 1 DutchGrid today and tomorrow David Groep, NIKHEF The DutchGrid Platform Large-scale Distributed Computing.
Meeting, 5/12/06 CMS T1/T2 Estimates à CMS perspective: n Part of a wider process of resource estimation n Top-down Computing.
Mcs/ HPC challenges in Switzerland Marie-Christine Sawley General Manager CSCS SOS8, Charleston April,
1 European e-Infrastructure experiences gained and way ahead OGF 20 / EGEE User’s Forum 9 th May 2007 Mário Campolargo European Commission - DG INFSO Head.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
Power and Cooling at Texas Advanced Computing Center Tommy Minyard, Ph.D. Director of Advanced Computing Systems 42 nd HPC User Forum September 8, 2011.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
1 Why is Digital Curation Important for Workforce and Economic Development? Alan Blatecky Office of Cyberinfrastructure Symposium on Digital Curation in.
Software for the CMS Cosmic Challenge Giacomo BRUNO UCL, Louvain-la-Neuve, Belgium On behalf of the CMS Collaboration CHEP06, Mumbay, India February 16,
CMS Computing Model summary UKI Monthly Operations Meeting Olivier van der Aa.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
tons, 150 million sensors generating data 40 millions times per second producing 1 petabyte per second The ATLAS experiment.
GRIDSTART Brussels 20/9/02 1www.gridstart.org GRIDSTART and European activities Dr Francis Wray EPCC The University of Edinburgh.
Toward a Roadmap for Complex Systems Education Curriculum A Concept Mapping Project at the Santa Fe Institute, Santa Fe, New Mexico Derek Cabrera William.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
High throughput biology data management and data intensive computing drivers George Michaels.
1 June 11/Ian Fisk CMS Model and the Network Ian Fisk.
05/14/04Larry Dennis, FSU1 Scale of Hall D Computing CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Virtual Laboratory Amsterdam L.O. (Bob) Hertzberger Computer Architecture and Parallel Systems Group Department of Computer Science Universiteit van Amsterdam.
Emanuele Leonardi PADME General Meeting - LNF January 2017
Scientific Computing Department
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
A highly reliable data center network topology Tier 1 at JINR
Themes in Geosciences.
Opening Big Data; in small and large chunks
Workflows in archaeology & heritage sciences
Computing & Data Resources Application interface
Grid Application Model and Design and Implementation of Grid Services
GRIF : an EGEE site in Paris Region
The LHC Computing Grid Visit of Professor Andreas Demetriou
Presentation transcript:

Dr. M.-C. Sawley IPP-ETH Zurich Nachhaltige Begegnungen Standing at the crossing point between data analysis and simulation Knowledge Discovery Panel

Towards a better comprehension of complex phenomena and systems  Astronomy, astrophysics  Bioinformatics  Systems Biology  Oil & Gas  Meteorology, Oceanography, Volcanology  Bio medical engineering  Automotive, Combustion, Aeronautics  …………..(many more to come) SOS14/Knowledge Discovery/MCSawley2 Integration of multi scale, multimodal, multi disciplinary Technologies

Drawing a (far too simple) integration map of available technology Analysis Data intensiveCompute intensive Modelisation Simulation models/HPC Digital imaging Visualization Bioinformatics Sensor networks SOS14/Knowledge Discovery/3MCSawley

Unravelling the complexity of Nature  Challenges for data driven science  On-line filtering the deluge of experimental or observational data  Repacking (data reduction) into high level objects, validation, calibration, quality control  Analyzing at fine granularity  accessibility, network capacity, data curation, heterogeneity of the systems  Using data to enrich modelisation, simulation  Integrating new data, new knowledge on the way  Drives and opportunities  Driven by sensor physics, high resolution images, …  Requires setting up of highly complex and heterogeneous infrastructures  The chain is a strong as its weakest link  Balance between  collecting, filtering, simulating, distributing and interpreting very large amount of data which comes into large bursts  Role of HPC? SOS14/Knowledge Discovery/MCSawley4

SOS14/Knowledge Discovery/MCSawley5 Path of Discovery in Fundamental Particle Physics

At the detector site: Online computing  DELL cluster (33 racks, dualcore Harpertown)  230 TB disks acquisition system MCSawley6 High level trigger: 80 millions electronic channels X4 (each of them using 4 bytes) X40 millions (collision rate 40 MHz) X1/1000 (zero suppression) X1/ (on line event filtering)  O(10) PB/y sent to CERN IT SOS14/Knowledge Discovery/

Extracting scientific knowledge :CMS computing TIER-0 TIER-1 TIER-1 CAF TIER-2TIER-2TIER-2TIER-2 Prompt Reconstruction Re-Reconstruction Skims Simulation And Analysis Calibration Express-Stream Analysis 600MB/s MB/s ~ 20MB /s TIER-1 CASTOR MB / s SOS14/Knowledge Discovery/7MCSawley

Participants to the panel  Nagiza Samatova, Oak Ridge National Labs  Gerald Kneller, University of Orleans  Ron Oldfield, Sandia National Labs SOS14/Knowledge Discovery/MCSawley8

Questions for the panel  What are the challenges for bridging data analysis and simulation?  What is the role of HPC for mining scientific discovery?  How would you define the “P” of HPC : Performance, Productivity, Portability, Pain, …?  What would your wish list to the HPC community consist of?  How do you expect the data wealth and complexity to influence simulation models and methods? SOS14/Knowledge Discovery/MCSawley9