Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.

Slides:



Advertisements
Similar presentations
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
Advertisements

LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
DPM Name Server (DPNS) Namespace Authorization Location of physical files DPM Server Requests queuing and processing Space Management SRM Servers v1.1,
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
RAL PPD Site Update and other odds and ends Chris Brew.
Data management for ATLAS, ALICE and VOCE in the Czech Republic L.Fiala, J. Chudoba, J. Kosina, J. Krasova, M. Lokajicek, J. Svec, J. Kmunicek, D. Kouril,
Tier1 Status Report Martin Bly RAL 27,28 April 2005.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
BNL Service Challenge 3 Site Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Site Report BEIJING-LCG2 Wenjing Wu (IHEP) 2010/11/21.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Steve Traylen PPD Rutherford Lab Grid Operations PPD Christmas Lectures Steve Traylen RAL Tier1 Grid Deployment
RAL DCache Status Derek Ross/Steve Traylen April 2005.
LCG Storage workshop at CERN. July Geneva, Switzerland. BNL’s Experience dCache1.8 and SRM V2.2 Carlos Fernando Gamboa Dantong Yu RHIC/ATLAS.
CERN Using the SAM framework for the CMS specific tests Andrea Sciabà System Analysis WG Meeting 15 November, 2007.
J Coles eScience Centre Storage at RAL Tier1A Jeremy Coles
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management and Interoperability Peter Kunszt (JRA1 DM Cluster) 2 nd EGEE Conference,
July 29' 2010INDIA-CMS_meeting_BARC1 LHC Computing Grid Makrand Siddhabhatti DHEP, TIFR Mumbai.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT DPM / LFC and FTS news Ricardo Rocha ( on behalf of the IT/GT/DMS.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
Rutherford Appleton Lab, UK VOBox Considerations from GridPP. GridPP DTeam Meeting. Wed Sep 13 th 2005.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
EGEE is a project funded by the European Union under contract IST VO box: Experiment requirements and LCG prototype Operations.
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Last update 29/01/ :01 LCG 1Maria Dimou- cern-it-gd Maria Dimou IT/GD CERN VOMS server deployment LCG Grid Deployment Board
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
1Maria Dimou- cern-it-gd LCG November 2007 GDB October 2007 VOM(R)S Workshop report Grid Deployment Board.
BaBar Cluster Had been unstable mainly because of failing disks Very few (
SRM-2 Road Map and CASTOR Certification Shaun de Witt 3/3/08.
GridKa December 2004 Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Doris Ressmann dCache Implementation at FZK Forschungszentrum Karlsruhe.
Last update 29/02/ :31 LCG 1Maria Dimou- cern-it-gd Maria Dimou IT/GD VOMS status IT GD Group Meeting
Scientific Computing in PPD and other odds and ends Chris Brew.
SESEC Storage Element (In)Security hepsysman, RAL 0-1 July 2009 Jens Jensen.
LCG Storage Workshop “Service Challenge 2 Review” James Casey, IT-GD, CERN CERN, 5th April 2005.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
ASCC Site Report Eric Yen & Simon C. Lin Academia Sinica 20 July 2005.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Mario Reale – GARR NetJobs: Network Monitoring Using Grid Jobs.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Martina Franca (TA), 07 November Installazione, configurazione, testing e troubleshooting di Storage Element.
RAL Plans for SC2 Andrew Sansum Service Challenge Meeting 24 February 2005.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
Security recommendations for dCache
EGEE Data Management Services
Status of the SRM 2.2 MoU extension
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
INFNGRID Workshop – Bari, Italy, October 2004
The LHCb Computing Data Challenge DC06
Presentation transcript:

Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005

Derek Ross E-Science Department DCache at RAL 1 Mid 2003 –We deployed a non grid version for CMS. –It was never used in production. End of 2003/Start of 2004 –RAL offered to package a production quality DCache. –Stalled due to bugs and and went back to developers and LCG developers.

Derek Ross E-Science Department DCache at RAL 2 Mid 2004 –Small deployment for EGEE JRA1 Intended for gLite i/o testing. End of 2004 –CMS instance 3 disk servers ~ 10TB disk space Disk served via nfs to pool nodes. Each pool node running a gridftp door. In LCG information system.

Derek Ross E-Science Department CMS Instance

Derek Ross E-Science Department DCache at RAL 4 Start of 2005 –New Production instance supporting CMS, DTeam,LHCb and Atlas VOs. 22TB disk space. CMS instance decommissioned and reused. Separate gdbm file for each VO. Uses directory-pool affinity to map areas of file system to VOs’ assigned disk.

Derek Ross E-Science Department

Derek Ross E-Science Department DCache at RAL 5 Early 2005 –Service Challenge 2 4 disk servers ~ 12TB disk space. UKLight connection to CERN. Pools directly on disk servers. Standalone Gridftp and SRM doors. SRM not used in Challenge due to software problems at CERN. Interfaced to Atlas Data Store.

Derek Ross E-Science Department SC2 instance gridftp Nortel 5510 Stack (80Gps) SRM Summit 7i UKLIGHT (2*1Gps) SJ4 2*1Gps dCache D/B 8 dCache pools 3 TB 3 TB 3 TB 3 TB Diskless GridFTP doors head

Derek Ross E-Science Department SC2 results Achieved 75MB/s to disk, 50MB/s to tape –Seen faster Mb/s to disk over LAN –Network delivered at last minute, under- provisioned Odd iperf results, high udp packet loss.

Derek Ross E-Science Department Future Developments Interface ADS to production dCache –Considering second srm door. –Implement script to propagate deletes from dCache to ADS Service Challenge 3 –Still planning. –Use production dCache. Experiments may want to retain data. –Avoid multi-homing if possible. Connect UKLight into site network.

Derek Ross E-Science Department Production Setup Proposed Testing: Dteam only for now

Derek Ross E-Science Department VO Support Bit of a hack – DCache has no concept of VOs –Gridmap periodically run through perl script to produce mapping of DN to Unix UID/GID. Each vo member mapped to first pool account of vo. All vo’s files owned by that account. –VOMS support coming…

Derek Ross E-Science Department Postgres Postgres SRM database is CPU hog –Being worked on. –Current recommendation is a separate host for PostgreSQL. Can use the database to store dCache transfer information for monitoring. In future may be possible to use for pnfs databases.

Derek Ross E-Science Department SRM requests Each SRM request lasts for (default) 24 hours if not finished properly. –Too many and the srm door queues new requests until slot available. –Educate users to use lcg-sd after an lcg- gt, don’t Ctrl-C lcg-rep…

Derek Ross E-Science Department SRM-SRM copies Pull mode –If dCache is the destination, then the destination pool initiates the gridftp transfer from the source srm. Need dcache-opt rpm installed (don’t need gridftp door running) on pools. Pool node need certificate and GLOBUS_TCP_PORT_RANGE accessible to incoming. –Lcg-utils don’t do this but srmcp does.

Derek Ross E-Science Department Quotas If two vo’s can access same pool, no way to stop one vo grabbing all of pool. No global quotas –Hard to do, pools can come and go Only way to restrict disk usage is limit pools a vo can write to. –But can’t get space available per vo.

Derek Ross E-Science Department Links