Lucas Fernandez Seivane Summer Student 2002 IT Group, DESY Hamburg

Slides:



Advertisements
Similar presentations
Beowulf Supercomputer System Lee, Jung won CS843.
Advertisements

HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Types of Parallel Computers
Benchmarking Parallel Code. Benchmarking2 What are the performance characteristics of a parallel code? What should be measured?
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Parallelized variational EM for Latent Dirichlet Allocation: An experimental evaluation of speed and scalability Ramesh Nallapati, William Cohen and John.
A Commodity Cluster for Lattice QCD Calculations at DESY Andreas Gellrich *, Peter Wegner, Hartmut Wittig DESY CHEP03, 25 March 2003 Category 6: Lattice.
A Comparative Study of Network Protocols & Interconnect for Cluster Computing Performance Evaluation of Fast Ethernet, Gigabit Ethernet and Myrinet.
Comparative Study of Beowulf Clusters and Windows 2000 Clusters By Seshendranath Pitcha.
Linux clustering Morris Law, IT Coordinator, Science Faculty, Hong Kong Baptist University.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Moving to Win 7 Considerations Dean Steichen A2CAT 2010.
Computer Systems 1 Fundamentals of Computing
The Study of Computer Science Chapter 0 Intro to Computer Science CS1510, Section 2.
Elements of a Computer System Dr Kathryn Merrick Thursday 4 th June, 2009.
CS492: Special Topics on Distributed Algorithms and Systems Fall 2008 Lab 3: Final Term Project.
The High Performance Cluster for QCD Calculations: System Monitoring and Benchmarking Lucas Fernandez Seivane Summer Student 2002.
So, Jung-ki Distributed Computing System LAB School of Computer Science and Engineering Seoul National University Implementation of Package Management.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Beowulf Cluster Jon Green Jay Hutchinson Scott Hussey Mentor: Hongchi Shi.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
Outline 3  PWA overview Computational challenges in Partial Wave Analysis Comparison of new and old PWA software design - performance issues Maciej Swat.
Chapter 2 Turning Data into Something You Can Use © The McGraw-Hill Companies, Inc., 2000 Processing Hardware.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
Distributed Programming CA107 Topics in Computing Series Martin Crane Karl Podesta.
What is QCD? Quantum ChromoDynamics is the theory of the strong force
Gravitational N-body Simulation Major Design Goals -Efficiency -Versatility (ability to use different numerical methods) -Scalability Lesser Design Goals.
Status and plans at KEK Shoji Hashimoto Workshop on LQCD Software for Blue Gene/L, Boston University, Jan. 27, 2006.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Exploring Parallelism with Joseph Pantoga Jon Simington.
Evaluating scientific software. Choosing appropriate software. Options. 1. Create your own from scratch. 2. Create your own with key bits from well-developed,
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
GRAPE 현황. Hardware Configuration Master –2 x Dual Core Xeon 5130 (2 GHz) –2 GB memory –500 GB hard disk 4 Slaves –2 x Dual Core Xeon 5130 (2 GHz) –1 GB.
Constructing a system with multiple computers or processors 1 ITCS 4/5145 Parallel Programming, UNC-Charlotte, B. Wilkinson. Jan 13, 2016.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Information Technology (IT). Information Technology – technology used to create, store, exchange, and use information in its various forms (business data,
CS120 Purchasing a Computer
Computer Hardware What is a CPU.
The Study of Computer Science Chapter 0
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Cache Memory.
Computational Requirements
Installation 1. Installation Sources
Supervisor: Andreas Gellrich
The Study of Computer Science Chapter 0
Constructing a system with multiple computers or processors
Genomic Data Clustering on FPGAs for Compression
Hadoop Clusters Tess Fulkerson.
The Study of Computer Science
Advanced Topics in Data Management
Constructing a system with multiple computers or processors
Constructing a system with multiple computers or processors
CSE8380 Parallel and Distributed Processing Presentation
Cache Memory.
Constructing a system with multiple computers or processors
The Study of Computer Science
Specifications Clean & Match
TOPIC: Computer-Aided Design
THE GOOGLE FILE SYSTEM.
Performance And Scalability In Oracle9i And SQL Server 2000
Designing a PC Farm to Simultaneously Process Separate Computations Through Different Network Topologies Patrick Dreher MIT.
Types of Parallel Computers
The Study of Computer Science Chapter 0
Cluster Computers.
Presentation transcript:

The High Performance Cluster for QCD Calculations: System Monitoring and Benchmarking Lucas Fernandez Seivane quevedin@mail.desy.de Summer Student 2002 IT Group, DESY Hamburg Supervisor: Andreas Gellrich Oviedo University (Spain)

Topics Some Ideas of QM The QFT Problem Lattice Field Theory What can we get? Approaches to the computing lattice.desy.de: Hardware Software The stuff we made: Clumon Possible improvements

Let’s do some physics… QM, “real behavior” of the world: ‘fuzzy world’ Relativity means causality (cause must precede consequence!) Any complete description of Nature must combine both ideas The only consistent way of doing this is … QUANTUM FIELD THEORY

The QFT Problem Impossible to solve it exactly PERTURBATIVE APPROACH Necessity of small coupling constant (like em = 1/137) Example: QED (the strange theory of light and matter) Taylor: em+2em/2 + 3em/6 +…

… but for QCD Not small coupling constant (at least at low energies) We cannot explain (at least analytically) a proton!!! We do need something exact (the LATTICE is EXACT*)

Lattice field theory Generic tool for approaching non perturbative QFT But more necessary in QCD (non perturbative aspects) Even pure theoretical interests (Wilson approach)

Lattice can calculate Path Integrals What can we get? We are interested in the spectra (bound states, masses of particles) We can do it by means of correlation functions: if we could calculate them exactly, we would have solved the theory They are extracted out of Path Integrals (foil1) The problem is calculate Path Integrals Lattice can calculate Path Integrals

(typical lattice sizes: a=0.05-0.1 fm, 1/a = 2GeV, L=32) A Naïve Approach Discretize space-time Monte-Carlo methods for choosing field configurations (Random generators) Numerical evaluation of Path Integrals and correlation functions!!! (typical lattice sizes: a=0.05-0.1 fm, 1/a = 2GeV, L=32) but…

…but Huge computer power Highly dimensional integrals The calculation requires to compute the inverse of an “infinite”-dimensional matrix, which takes a lot of CPU time and RAM. That’s why we need clusters, supercomputers or special machines (to divide the work) The amount of data transferred is not so important, the deciding factor is the LATENCY of the network and the scalability above 1TFlops

How can we get it? General Purpose Supercomputers: Very expensive Rigid (difficult upgrades on hardware) Fully customed parallel machines: Completely optimized Only this use (difficult recycling) Necessity of design, develop and build (or modify) the hard & soft Commodity clusters “Cheap PC” components Completely customizable Easy to upgrade / recycle

Machines Commercial Supercomputers: CrayT3E, Fujitsu VPP77, NECSx4, Hitachi SR8000… Parallel machines: APEmille/apeNEXT INFN/DESY QCDSP/QCDOC CU/UKQCD/Riken CP-PACS Tsukuba/Hitachi Commodity clusters + Fast Networking Low latency (Fast Networking) Fast Speed Standard software and programming environments

Lattice cluster@DESY Cluster bought from a company (Megware), Beowulf type (1 master, 32 slaves) Before upgrade (some weeks ago): 32 nodes: IntelXEONP4 1.7GHz 256 KB cache 1GB Rambus RAM 2  64 bit PCI slots 18 GB SCSI hard disks Fast Ethernet switch (normal networking, NFS disk mounting) Myrinet network (low latency) Upgrade (August 2002) 16 nodes: 2 IntelXEONP4 1.7GHz 256 KB cache 16 nodes: 2 IntelXEONP4 2.0GHz 512 KB cache

Lattice cluster@DESY(2) Software: SuSE Linux (modified by Megware) MPICH-GM (implementation of MPI-CHamaleon for Myrinet GM system) Megware Clustware (OpenSCE/SCMS modified): tool for monitoring and administration (but no logs)

Lattice cluster@DESY(3) http://lattice.desy.de/cgi-bin/clumon/cgi_clumon.pl Andreas Gellrich First Version: Provides logs and monitoring Perl written (customizable)

Lattice cluster@DESY(4) http://lattice.desy.de/cgi-bin/clumon/cgi_clumon.pl Me and Andreas Gellrich new version: Also graphical data and another log measure Uses MRTG to graph data

Clumon v2.0 (1)

Clumon v2.0 (2)

Work done (in progress) Getting the flavor of a really high-perf cluster Learning Perl (more or less) to understand Andreas tool Playing around with Andreas tool Search for how to graph this kind of data Learning how to use MRTG/RRDtool Some test and previous versions Only have to do last retouches (polishing): Time info of the cluster Better documentation of the tools Play around this last week with other stuff Prepare talk and document and write up

Possible Improvements The cluster is unplugged to AFS DESY Need for Backups / Archiving of the Data stored (dCash theoc01) Maybe reinstall the cluster with DESY Linux (to fully know what’s in it) Play around with other cluster stuff: OpenSCE, OSCAR, ROCKS…