HEPIX spring 2013 Du 15 au 19 avril 2007 à Bologne

Slides:



Advertisements
Similar presentations
WLCG Cloud Traceability Working Group progress Ian Collier Pre-GDB Amsterdam 10th March 2015.
Advertisements

Testing as a Service with HammerCloud Ramón Medrano Llamas CERN, IT-SDC
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Michal Kwiatek, Juraj Sucik, Rafal.
CERN IT Department CH-1211 Genève 23 Switzerland t Integrating Lemon Monitoring and Alarming System with the new CERN Agile Infrastructure.
U.S. Department of the Interior U.S. Geological Survey David V. Hill, Information Dynamics, Contractor to USGS/EROS 12/08/2011 Satellite Image Processing.
Assessment of Core Services provided to USLHC by OSG.
Module 10 Configuring and Managing Storage Technologies.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
RAL Site Report HEPiX Fall 2013, Ann Arbor, MI 28 Oct – 1 Nov Martin Bly, STFC-RAL.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Tim Vander Kooi Systems
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
APC InfraStruxure TM Central Smart Plug-In for HP Operations Manager Manage Power, Cooling, Security, Environment, Rack Access and Physical Layer Infrastructure.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
HEPiX Summary Martin Bly HEPSysMan - RAL, June 2013.
Slide 1 Experiences with NMI R2 Grids Software at Michigan Shawn McKee April 8, 2003 Internet2 Spring Meeting.
Copyright © 2011 EMC Corporation. All Rights Reserved. MODULE – 6 VIRTUALIZED DATA CENTER – DESKTOP AND APPLICATION 1.
Take on messages from Lecture 1 LHC Computing has been well sized to handle the production and analysis needs of LHC (very high data rates and throughputs)
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
HEPiX Spring 2012 Highlights Helge Meinhard CERN-IT GDB 09-May-2012.
WLCG Cloud Traceability Working Group face to face report Ian Collier 11 February 2015.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Jose Castro Leon CERN – IT/OIS CERN Agile Infrastructure Infrastructure as a Service.
Virtualisation & Cloud Computing at RAL Ian Collier- RAL Tier 1 HEPiX Prague 25 April 2012.
WLCG operations A. Sciabà, M. Alandes, J. Flix, A. Forti WLCG collaboration workshop July , Barcelona.
GDB July 2015 Jeremy’s quick summary notes Also refer to the meeting minutes
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
HEPiX Spring 2013 CNAF Bologna / Italy Workshop wrap-up Helge Meinhard.
Trusted Virtual Machine Images a step towards Cloud Computing for HEP? Tony Cass on behalf of the HEPiX Virtualisation Working Group October 19 th 2010.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CERN IT Department CH-1211 Geneva 23 Switzerland t CF Computing Facilities Agile Infrastructure Monitoring CERN IT/CF.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Commissioning the CERN IT Agile Infrastructure with experiment workloads Ramón Medrano Llamas IT-SDC-OL
Tim Bell 04/07/2013 Intel Openlab Briefing2.
Computing Facilities CERN IT Department CH-1211 Geneva 23 Switzerland t CF Agile Infrastructure Monitoring HEPiX Spring th April.
Web Technologies Lecture 13 Introduction to cloud computing.
Evolving Security in WLCG Ian Collier, STFC Rutherford Appleton Laboratory Group info (if required) 1 st February 2016, WLCG Workshop Lisbon.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN IT Monitoring and Data Analytics Pedro Andrade (IT-GT) Openlab Workshop on Data Analytics.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
CC - IN2P3 Site Report Hepix Fall meeting 2010 – Ithaca (NY) November 1st 2010
A Service-Based SLA Model HEPIX -- CERN May 6, 2008 Tony Chan -- BNL.
Automated virtualisation performance framework 1 Tim Bell Sean Crosby (Univ. of Melbourne) Jan van Eldik Ulrich Schwickerath Arne Wiebalck HEPiX Fall 2015.
CERN IT Department CH-1211 Geneva 23 Switzerland t ES 1 how to profit of the ATLAS HLT farm during the LS1 & after Sergio Ballestrero.
Ian Collier, STFC, Romain Wartel, CERN Maintaining Traceability in an Evolving Distributed Computing Environment Introduction Security.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN Agile Infrastructure Monitoring Pedro Andrade CERN – IT/GT HEPiX Spring 2012.
HEPiX report Spring 2015 HEPiX meeting Oxford University, UK Helge Meinhard, CERN-IT Grid Deployment Board 10-Jun Helge Meinhard (at) CERN.ch.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
Platform & Engineering Services CERN IT Department CH-1211 Geneva 23 Switzerland t PES Agile Infrastructure Project Overview : Status and.
Instituto de Biocomputación y Física de Sistemas Complejos Cloud resources and BIFI activities in JRA2 Reunión JRU Española.
Trusted Virtual Machine Images the HEPiX Point of View Tony Cass October 21 st 2011.
Hepix spring 2012 Summary SITE:
CERN IT Department CH-1211 Genève 23 Switzerland t Monitoring: Present and Future Pedro Andrade (CERN IT) 31 st August.
CernVM-FS Operations in the CERN IT Storage Group Dan van der Ster (CERN IT-ST) CernVM Users Workshop 6-8 June D. van der Ster2.
Using the CMS Higher Level Trigger Farm as a Cloud Resource David Colling Imperial College London.
HEPiX spring 2013 report HEPiX Spring 2013 CNAF Bologna / Italy Helge Meinhard, CERN-IT Contributions by Arne Wiebalck / CERN-IT Grid Deployment Board.
Pilot Kafka Service Manuel Martín Márquez. Pilot Kafka Service Manuel Martín Márquez.
WLCG Workshop 2017 [Manchester] Operations Session Summary
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Highlights Spring 2013 April 15-19, Bologna, Italy
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
AGLT2 Site Report Shawn McKee/University of Michigan
Monitoring at a Multi-Site Tier 1
Helge Meinhard, CERN-IT Grid Deployment Board 10-May-2017
WLCG Collaboration Workshop;
Introduction to Cloud Computing
GGF15 – Grids and Network Virtualization
Cloud Management Mechanisms
Presentation transcript:

HEPIX spring 2013 Du 15 au 19 avril 2007 à Bologne

83 participants de 40 instituts 70 présentations dont 17 « site reports » Dans les locaux (historiques) de l’université de Bologne

Thèmes des présentations Computing: CPUs, batch Réseau: IPv6, monitoring, sécurité Infrastructure: salles (incidents, efficacité énergétique), services, outils, OS Stockage Grilles, clouds et virtualisation

Au CERN « Agile Infrastructure » BigData (File system CEPH) Nouvelle salle (« Wigner Data Center » à Budapest) Sécurité Nouveaux outils : Puppet, Git, Owncloud, Drupal Fédération d’identité Qualité Haute disponibilité

Parmi les autre sujets … IPV6 (Testbed) Avenir d’AFS Outils : monitoring, analyse des logs (Splunk), déploiement Conservation des données à long terme Qualité (CC: outil CMDB) (Manque de) fiabilité des matériels (firmware)

CERN Agile Infrastructure Luis FERNANDEZ ALVAREZ New resource & configuration management of IT infrastructure –No increase in staff members => manage the infrastructure more efficiently IaaS approach : –private cloud based on Openstack (nova) / configuration with puppet –Coll starting around Openstack with BNL, IN2P3, ATLAS/CMS, IHEP… LCG context : enable remote management of 2nd Tier-0 data center –unify the two CERN’s data centers located in Meyrin and in Wigner (Budapest) 90 % of hardware virtuali- zed In progress : Single source for accounting data 6

CERN Remote data center Wayne Salter Construction –Started 21st May 2012 First room operational January 2013 Two 100Gbps links are operational since late January –One commercial provider (T-Systems) and DANTE T-System RTT (Round Trip Time): 24ms DANTE RTT: 21ms First servers delivered and installed March 2013 Operations –Work still required to finalize operational procedures 7

CMSooooCloud Wojciech OZGA Use of HLT farm farm during LHC LS1 : additional computing resources –HLT farm :13312 cores 26 Tbytes RAM 195 kHS06 –CMS T0 121 kHS06 / CMS ∑ T1 150 kHS06 / CMS ∑ T2 399 kHS06 CMS specific computation on HLT farms –Minimal change, opportunistic usage –No reconfiguration no additional hardware Cloudify the CMS HLT cluster : Overlay Cloud layer deployed with zero impact on data taking Using Openstack –Nova compute service manages the VM lifecycle –Networking Virtualization (OpenVSwitch) CMS online network seperated from CERN network –Needs to increase the network connectivity to CERN Tier 0 8

Ceph as an option for Storage-as-a-Service Arne Wiebalck, Dan van der Ster Storage at CERN : AFS, CASTOR, EOD, NetApp filers, block storage for Virtual Machins… Looking for a consolidated generic storage system ? Ceph, distributed, open-source storage system being evaluated at CERN (not ready for production) –Unification : object store, block store, file system –Traditional storage management : file systems and blocks storage –Additional layer : Object store or Object storage –Decoupling the namespace from the underlying hardware No central table / single entry point / single point of failure Ceph uses instead an algorithm (Controlled Replication Under Scalable Hashing) to maps data to storage devices No central metata data server : Algorithmic data placement with data replication and redistribution capabilities Enhanced scalability of the storage system Looks promizing as a generic storage backend –For both image store/sharing and S3 storage service 9

Security update (1/2) Romain Wartel Citadel incident : cf. CERT Polska public report Putting in place a malware infrastructure & Business model … Still typical ssh-attacks in the academic community Back to the 90s : Ebury revisited : old style (1990s) sshd trojan –Actively used in 2011 ; found mostly on RHEL-base systems –Attacks can be discovered just by checking checksums of installed RPMs/DEBs. –Are we checking the integrity of binaries ? Which tools ? WLCG Operational security : incidents per year, – 2012 has been a quieter than usual –Attacks are more and more sophisticated Security paradigm shift 10

Security update (2/2) Romain Wartel The classic approach (strong controls mechanisms, well defined security parameters) … is to keep attackers outside the medieval approach New approach is to grant access to trust users –Security relies more on traceability & Ability to terminate access to users not following local policies Manageable security : –Attackers would never be allowed to….. –Malicious users will be isolated We will control the VMs.. BUT VMS need access to local ressources and will evolve dynamically Isolation almost impossible ….traceability remains the key point 11

Common LHC Network monitoring Shawn MC KEE Common to the four experiments Standardized network monitoring, –Std tool/framework :perfSONAR –Std measurement of network perf. Related metrics over time WLCG Ops pS task force wiki : US ATLAS wiki : PS-PS V3.3 (out very soon) will have all functionality for the mesh built-in WLCG mesh configurations are hosted in AFS LHC-FR Dashboard

Monitoring and log analysis at GSI : Monitoring –« HomeMade » dahsboard –Use SNMP (!! Security !!), many existing agents –Nagios config files generated from CMDB –Use interoperable tools (ex : OTRS tickets generated) Log management : –Bigdata problematics –Collect logs from syslog NG (Logstach) –RAM Buffering (Redis) –Aggregate /filter/index and store logs –ElasticSearch and Dashboard (KIBANA ) : parse Events and view trends

Monitoring and log analysis –At CERN : Monitoring HEP disk farms with COCKPIT : log repository + display + correlation engine. Bigdata solutions : storage with HBase on HDFS (Hadoop), analysis with MapReduce –At DESY : Monitoring GridEngine with SPLUNK collect log (job data) and filter (Syslog + Splunk Forwarder ) Nice (summary) graphs made by splunk (Web interface) Commercial tool

Divers Data center Energy Optimization –Contact : Wayne Salter (CERN) IT Computing Facilities Group Leader –HEPIX wishes to have a common effort on this topic ? –Dedicated track on measures adopted at the different sites? –Working group to share experience and advice on what measures could be taken by specific sites? Open AFS future : BOF session –still there is a community : Fermi, BNL, IN2P3, Beijing, DESY, Manchester, PSI … –Create HEP AFS inventory found useful => to be done before the Fall 2013 HEPix Meeting Site contact, mid-term plans, AFS use-cases, requirements –IPv6 & AFS : Towards a work plan Implementing it ourselves excluded What are our requirements ? Can we leave with private cells –Gather information – Get in touch with core developers - Set up a discussion and decide at the next HEPix meeting –Follow-up by Peter van der Reest (DESY), Arne Wiebalck (CERN), Andrei Maslenniko (CASPUR) 15

Prochains meetings Fall 2013 (28 oct au 1er nov.) : Université du Michigan (USA) Spring 2014 : du 19 au 23 mai au LAPP !!!!!