Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots
Mission of Fermilab 2 Particle Physics Operations Capability Future Facilities Science and Technology R&D Producing world-class particle physics results Developing and operating advanced accelerators, detectors, and computing facilities Designing and building the next generation of facilities R&D into technologies for future facilities
Lab Scientific Roadmap 3
Long history of provisioning and operating storage for all Fermilab Users: approaching 100 Petabytes of data… 44
...and computing for experiments and scientific projects to which Fermilab is contributing: LHC: US CMS Tier-1 Computing Center Tevatron: continued analysis and preservation of CDF and D0 data. Computing for ~10 neutrino/dark matter/dark energy experiments. Significant contributor to the Open Science Grid. HPC systems for Lattice Quantum Chromodynamics (LQCD). Experiment data acquisition and accelerator controls systems. >30,000 cores of cluster based and distributed high throughput computing resources ~30 Petabytes of high performance disk cache. Heavily used 8x10Gigabit pipes to WAN; >100Gigabit data transfers on the data center LAN; Commissioning new 100Gigabit connectivity for wide area. 5
Typical US CMS Tier-1 Computing utilization: Jobs and Disk I/O 6
R&D and Support of Physics Software. For example: Accelerator modeling and simulation High speed data acquisition and accelerator controls. Distributed computing and databases, Frameworks for data simulation, processing and analysis. Scientific Linux added value to Redhat linux OS for HEP and wider community. Monte Carlo event generators, simulation packages, track finders, pattern recognition algorithms, high volume statistical analysis etc. 7 Plus many centralized Computing Services that achieved ISO20K Certification in 2012
Computing needs for CMS increasing significantly: Processing needs will increase by ~factor 5 in the next 3- 5 years. Computing budgets expected to remain constant. Data sizes will increase to exabyte scale by Aim to provide global federated data storage solutions to reduce total cost of storage. Must adapt to new multi-core technologies to preserve time to results. Must be agile to use of whatever CPU resources are available Leading to investigations into the use of HPC (DOE & NSF machines); private, commercial and public clouds. 8
Total computing needs for smaller experiments will outpace available resources on site in 2-3 more years: 9 Adapting to opportunistic use of OSG accessible resources Deploying Virtualization to increase efficiency of sharing across existing h/w. Exploring/evaluating bursting to off-site Cloud resources for peak and instantaneous demands.
10
OSG ramping up activity towards provisioning across heterogeneous resources 11
& sustaining the environment 12 & the Illinois Accelerator Research Center