Download presentation
Presentation is loading. Please wait.
Published byMadison Bond Modified over 8 years ago
1
ILC_3: DISTRIBUTED COMPUTING TOWARD ILC (PROPOSAL) CC-IN2P3 and KEK Computing Research Center (KEK-CRC) Hiroyuki Matsunaga (KEK) 2014 Joint Workshop of the TYL/FJPPL and FKPPL 26-28 May, 2014, in Bordeaux
2
Background CC-IN2P3 and KEK-CRC worked together on Grid computing within FJPPL 2007-2012 : Comp_3: Grid Interoperability and Data Management EGEE – EGI (gLite) in Europe, and NAGREI – RENKEI in Japan SAGA (Simple API for Grid Application) as a Grid Standard for a solution of the interoperability SRB / iRODS Data Grid The project was successful Useful to improve the development and deployment To exchange ideas and experiences is valuable Grid Computing works well in particle physics and other fields CC-IN2P3 is the Tier-1 center for WLCG KEK-CRC will be the “Tier-0” center for Belle-II experiment which starts taking data soon
3
Background (cont.) ILC project is progressing well Technical Design Report (TDR) was complete last year The candidate site (Kitakami) has been decided in Japan Not too early to start considering the computing design for ILC It should be sustainable to support the long-term project which will last for more than a decade (like LHC) It should adopt “standard” technologies as much as possible Grid is being replaced with Cloud Agile and (ideally) less deployment cost Network is still improving But it is not easy to exploit the available bandwidth efficiently More Virtualization Software-Defined Networking(SDN), Storage(SDS), and Data-Center(SDDC)?
4
The Aim of this project R&D of distributed computing for future large projects like ILC Evaluate and gain experience of emerging technologies such as cloud More studies on data storage and data transfer Might be useful for WLCG and Belle II To exchange ideas and knowledge between the two centers Re-establish the collaboration beyond the previous Comp_3 project Long-term collaboration has a benefit Especially when ILC is realized
5
Members French Group (CC-IN2P3) Japanese Group (KEK-CRC) P-E. MacchiH. Matsunaga B. DelaunayG. Iwai L. Caillat-ValletK. Murakami J-Y. NiefT. Sasaki V. HamarW. Takase M. PuelY. Watase S. Suzuki
6
Cloud computing Widely used in commercial sector Some HEP labs already deployed IaaS (Infrastructure as a Serivce) cloud IaaS cloud allows to Provision (compute) resources more dynamically Use resources more flexibly System environment can be adjusted according to use cases Optimize the resource usage (in rather opportunistic way) Ease the system administration Cloud management itself would not be easy Quite different from the conventional system management
7
Cloud experience at KEK No strong needs from users so far Set up small test system and performed tests ISF +LSF7 from Platform / IBM We chose a commercial solution for better support and integration of batch scheduler Will migrate to PCMAE (successor of ISF) + LSF9 this fall Another small instance using OpenStack Most popular software for free Actively developed Strong support by major IT companies Many HEP sites try to deploy OpenStack KEK’s main computer system will be replaced next summer Cloud will be integrated into the next system for production use
8
Cloud at CC-IN2P3 Already deployed using OpenStack For compute and testing/development systems Hundreds of CPU cores are provisioned Target projects for compute resources Large Synoptic Survey Telescope (LSST, http://www.lsst.org) Astronomy Euclid (http://www.euclid-ec.org/) Astronomy / Astrophysics (Dark Energy) Atlas MC simulation, HammerCloud validation
9
How to use the cloud Integrate with batch scheduler No more work needed for ISF / PCMAE & LSF Openstack & Univa Grid Engine Direct use of cloud interfaces Amazon EC2, OpenStack Nova Using middleware that has cloud interfaces DIRAC (VMDIRAC) ILC and Belle II go to this direction
10
Cloud R&D in this project Exchange experience and knowledge VM image creation and management Customizations VM-job scheduling Network configuration IP address management IPv6 migration/addition in the near future Many addresses available NaaS (Network as a Service) Storage integration Authentication/Authorization, Accounting, Information system Federation of the cloud resources Submitting jobs using DIRAC and direct API Interoperability between heterogeneous Cloud resources
11
iRODS iRODS (iRule Oriented Data Systems) is a light-weight data storage system Rule-based data management with a meta-data catalogue Versatile system that could be used by many communities according to their needs Various policies for data management and replication can be implemented Easier to administer than SRM storage systems iRODS lacks the SRM interface currently Developed by Prof. Reagan Moore of Univ. of North Carolina and his team Both CC-IN2P3 and KEK have expertise on iRODS (and also its predecessor SRB) Deployment as an additional data storage system
12
iRODS service at KEK 4 iRODS servers 2 DB (iCAT) servers PostgreSQL, active-standby Data Storage: GPFS + HPSS Clients Command line (iCommands) GUI (JUX, Java Universal eXplorer) Web (rodsweb) Supported projects MLF (Materials and Life Science Experimental Facility at J-PARC) 140TB, 10M files T2K (Tokai-to-Kamioka) experiment For Data Quality at the Near Detectors Federation with Queen Mary University of London (QMUL) 10TB, 800 files
13
Use Case in T2K Rules Bundle small files into large tar files Copy a tar file to GPFS, then replicate to HPSS disk DB GPFS HPSS Client iRODS server iRODS server file tar file
14
iRODS at CC-IN2P3 Used by multidisciplinary environment High Energy Physics BaBar (~2PB), double Chooz (600TB) Astrophysics AMS Biology and Biomedical applications Arts and Humanities 10+ data servers + 2 iCAT servers Oracle as the backend DB Mass Storage: HPSS Backup system: TSM
15
iRODS federation Federation of iRODS between KEK and CC-IN2P3 Basically worked in the previous FJPPL project Possible new federation with gravitational wave experiments KAGRA project will start the operation in the near future At Kamioka, Japan KEK helps set up distributed computing environment CC-IN2P3 supports Virgo, European experiment, by using iRODS For the gravitational wave detection, it is very important to cross check the results between the experiments Data should be shared by the experiments for that Good opportunity for iRODS federation Could be a precursor for ILC
16
Data transfer High-performance data transfer over WAN is a long-standing issue Necessary to tune OS/application configuration at both ends As for France-Japan network connection: High latency (300ms RTT) going through US Academic link shared by many users Directional asymmetry in the throughput Varying in time Network technology is improving Available bandwidth increasing 40 Gbit or 100 Gbit forseen for WAN in next few years SDN Important to check network performance continually Beneficial for other experiments too
17
Summary This project aims at studying some items of the distributed computing beyond the current Grid environment Cloud is getting mature and we want to evaluate it for the future large-scale deployment Data storage and data transfer remain important issues for a long time Of particular importance for the distributed computing is to work together and communicate with each other Face-to-face meeting is important
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.