Working Meeting at FermiLab University of California Grid (UC Grid) January 19-20, 2010 Fermi Lab, Chicago, IL.

Slides:



Advertisements
Similar presentations
CSF4 Meta-Scheduler Tutorial 1st PRAGMA Institute Zhaohui Ding or
Advertisements

Scaling TeraGrid Access A Testbed for Attribute-based Authorization and Leveraging Campus Identity Management
Legacy code support for commercial production Grids G.Terstyanszky, T. Kiss, T. Delaitre, S. Winter School of Informatics, University.
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
MyProxy: A Multi-Purpose Grid Authentication Service
SALSA HPC Group School of Informatics and Computing Indiana University.
Academic Technology Services The UCLA Grid Portal - Campus Grids and the UC Grid Joan Slottow and Prakashan Korambath Research Computing Technologies UCLA.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
6th Biennial Ptolemy Miniconference Berkeley, CA May 12, 2005 Distributed Computing in Kepler Ilkay Altintas Lead, Scientific Workflow Automation Technologies.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Grid Services at NERSC Shreyas Cholia Open Software and Programming Group, NERSC NERSC User Group Meeting September 17, 2007.
UNICORE UNiform Interface to COmputing REsources Olga Alexandrova, TITE 3 Daniela Grudinschi, TITE 3.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
January, 23, 2006 Ilkay Altintas
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
Riccardo Bruno INFN.CT Sevilla, Sep 2007 The GENIUS Grid portal.
Grid ASP Portals and the Grid PSE Builder Satoshi Itoh GTRC, AIST 3rd Oct UK & Japan N+N Meeting Takeshi Nishikawa Naotaka Yamamoto Hiroshi Takemiya.
AliEn uses bbFTP for the file transfers. Every FTD runs a server, and all the others FTD can connect and authenticate to it using certificates. bbFTP implements.
A Web 2.0 Portal for Teragrid Fugang Wang Gregor von Laszewski May 2009.
CI Days: Planning Your Campus Cyberinfrastructure Strategy Russ Hobby, Internet2 Internet2 Member Meeting 9 October 2007.
Towards a Javascript CoG Kit Gregor von Laszewski Fugang Wang Marlon Pierce Gerald Guo
1. Introduction  The JavaScript Grid Portal is trying to find a way to access Grid through Web browser, while using Web 2.0 technologies  The portal.
GRAM: Software Provider Forum Stuart Martin Computational Institute, University of Chicago & Argonne National Lab TeraGrid 2007 Madison, WI.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Nimrod/K using Opal Services for Virtual Screening David Abramson, Ilkay Altintas, Daniel Crawl, Wilfred Li, Jane Ren, Jianwu Wang, Colin Enticott(presenter)
1 Overview of the Application Hosting Environment Stefan Zasada University College London.
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
National Computational Science National Center for Supercomputing Applications National Computational Science NCSA-IPG Collaboration Projects Overview.
Information Resources and Communications University of California, Office of the President UC-Wide Activities in Support of Research and Scholarship David.
Group 1 : Grid Computing Laboratory of Information Technology Supervisors: Alexander Ujhinsky Nikolay Kutovskiy.
1 Ilkay ALTINTAS - July 24th, 2007 Ilkay ALTINTAS Director, Scientific Workflow Automation Technologies Laboratory San Diego Supercomputer Center, UCSD.
NGS Innovation Forum, Manchester4 th November 2008 Condor and the NGS John Kewley NGS Support Centre Manager.
NUG 2004 Grid File Yanker Demo Shreyas Cholia Mass Storage Group, NERSC 06/24/2004.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
Holding slide prior to starting show. A Portlet Interface for Computational Electromagnetics on the Grid Maria Lin and David Walker Cardiff University.
BOINC: Progress and Plans David P. Anderson Space Sciences Lab University of California, Berkeley BOINC:FAST August 2013.
NEES Cyberinfrastructure Center at the San Diego Supercomputer Center, UCSD George E. Brown, Jr. Network for Earthquake Engineering Simulation NEES TeraGrid.
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University.
SAN DIEGO SUPERCOMPUTER CENTER Inca Control Infrastructure Shava Smallen Inca Workshop September 4, 2008.
VO Privilege Activity. The VO Privilege Project develops and implements fine-grained authorization to grid- enabled resources and services Started Spring.
Glite. Architecture Applications have access both to Higher-level Grid Services and to Foundation Grid Middleware Higher-Level Grid Services are supposed.
Shell Interface Shell Interface Functions Data. Graphical Interface Graphical Interface Command-line Interface Command-line Interface Experiments Private.
1 P-GRADE Portal tutorial at EGEE’09 Introduction to hands-on Gergely Sipos MTA SZTAKI EGEE.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
GSI: Security On Teragrid A Introduction To Security In Cyberinfrastructure By Dru Sepulveda.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Introduction to P-GRADE Portal hands-on Miklos Kozlovszky MTA SZTAKI
National Energy Research Scientific Computing Center (NERSC) Visportal : interface to grid enabled NERC resources Cristina Siegerist NERSC Center Division,
WLCG Authentication & Authorisation LHCOPN/LHCONE Rome, 29 April 2014 David Kelsey STFC/RAL.
The Gateway Computational Web Portal Marlon Pierce Indiana University March 15, 2002.
SAN DIEGO SUPERCOMPUTER CENTER Welcome to the 2nd Inca Workshop Sponsored by the NSF September 4 & 5, 2008 Presenters: Shava Smallen
PROGRESS: GEW'2003 Using Resources of Multiple Grids with the Grid Service Provider Michał Kosiedowski.
1 P-GRADE Portal hands-on Gergely Sipos MTA SZTAKI Hungarian Academy of Sciences.
InSilicoLab – Grid Environment for Supporting Numerical Experiments in Chemistry Joanna Kocot, Daniel Harężlak, Klemens Noga, Mariusz Sterzel, Tomasz Szepieniec.
Antonio Fuentes RedIRIS Barcelona, 15 Abril 2008 The GENIUS Grid portal.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
UCTrust Integration for UC Grid David Walker University of California, Davis ucdavis.edu Kejian Jin University of California, Los Angeles kjin.
Academic Technology Services The UC Grid Project OSG Consortium All-Hands Meeting Bill Labate & Joan Slottow Research Computing Technologies UCLA Academic.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
PLG-Data and rimrock Services as Building
Introduction to the Application Hosting Environment
LEAD-VGrADS Day 1 Notes.
Patrick Dreher Research Scientist & Associate Director
Information Technology Ms. Abeer Helwa
TeraGrid Identity Federation Testbed Update I2MM April 25, 2007
Grid Computing Software Interface
Presentation transcript:

Working Meeting at FermiLab University of California Grid (UC Grid) January 19-20, 2010 Fermi Lab, Chicago, IL

Krishna Muriki, Jackie Scoggins, Tammy Welcome, Shreyas Chola, Doug Olson, Gary Jung, Greg Kurtzer Lawrence Berkeley National Laboratory William Strossman, Russ Harvey, Bob Grant University of California, Riverside Paul Weakliem, Arlene Allen University of California, Santa Barbara Harry Mangalam, Joseph Farran, Frank Wessel University of California, Irvine Shava Smallen, Jianwu Wang, Ilkay Altinas, Richard Moore, David Hutches San Diego Super Computing Center Kejian Jin, Shao-Ching Huang, TV Singh, Sveta Mazurkova, Bill Labate, Qiyang Hu, Prakashan Korambath, Seonah Kim, Scott Johnson, Puneet Souda David Walker, Bill Broadley, Russ Hobby University of California, Davis LBNL

UC Grid Architecture UC Grid Portal UC MyProxy UCLA Portal UCSB PortalUCI PortalUCR Portal UCSD Portal C1C2 … C1C2… C1 C2… C1 C2… C1C2… UC Grid CA UC Register Service UC Sync Service Gaussian, Q-Chem, Matlab, Mathematica R, Stata, NWChem, CPMD, GAMESS Omssa, LAMMPS, Amber, Migrate Generic-Serial, Generic-Parallel Xmd,Octave Interactive-VNC

Single Campus Architecture Grid Portal Grid Appliance Credential mapped to local ID Grid Appliance https Uniform browser based interface Head node Cluster II Head node Cluster I ION Visualization Server Storage ServerMyProxy Server

The UC Grid Portal Provides Register Service (SOA): – user portal account creation and uniqueness of user name in the UC Grid, automatic CA sign, MyProxy push. Provides Sync Service (SOA) between campus portal and UC Grid portal: – Any change of an user account at a campus Grid Portal is automatically propagated to the UC Grid Portal. – Any change of new resources at a campus Grid Portal will also be reflected at UC Grid Portal From the UC Grid Portal, users can: – work with clusters from all the campuses those users are authorized to use. – transfer files across the campuses. – Use pooled resources across the campuses.

Campus Grid Portal Features: – Resource Discovery Cluster load, usage, queue information, job details – Data Manager File management such as create, remove, rename, edit, transfer etc. – Job Services Submit job as a cluster user Submit job as a pool user – Interactive X window for applications Matlab, Mathematica, Maple, Xstata etc. – Access to Teragrid

Job Submission General Job submission  Any user created executable Application job submission  Predefined executable o As Cluster user o As Pool user Pool user application job submission UGP Chooses the target cluster for the application job. Stages the input files to a guest user login id at the target cluster Submits the job to run under a guest login id at the target cluster. Local scheduler determines when those jobs are run. UGP facilitates the transfer of the output files back to the Pool User. Provide job monitoring and auditing

Interactive Xterm through VNC

Interactive GUI Application through VNC

File Transfer using grid-ftp

Generic Batch Job Submission Interface

Predefined Application Submission

Job Submission to Other Grid Resources User interface to fill in username and password to retrieve short lived credential for submitting jobs to Teragrid

Proteomics Workflow Project Binary Data/ Converter Pre-Processed Data (OMSSA/InsPect Format, Commercial Application) Mass Spectrometer Grid Temp Storage User Generated Query Cluster A Cluster B Cluster C OMSSA/InsPect Run NIH BLAST Library (Downloaded to Cluster) Grid Portal 1. Single job or job array 2. Output collection Data Merge/Split (based on job type) Meta-scheduler OMSSA Output to user’s desktop, OMSSA Viewer InsPect Output to user’s desktop, Web Viewer

Protein Design Processes Inside-out protocol to design novel enzyme catalysts Baker - Houk collaboration Transition structure (QM theozyme) Transition structure (QM theozyme) Match theozyme geometry to protein scaffold (RosettaMatch) Match theozyme geometry to protein scaffold (RosettaMatch) EDGE (Rank and Filter) EDGE (Rank and Filter) Experiments Redesign and repack Active site (RosettaDesign) & MD/EDGE Redesign and repack Active site (RosettaDesign) & MD/EDGE

Enzyme Design Using Kepler Workflow RosettaMatch Goal: Finds the enzyme Scaffolds for the input theozyme Approximately 250 Jobs per submission – Each creates around 2000 outputs Workflow Distribute the Jobs through Scheduler/Grid Edge Goal: Filter out Match results that are out of user defined range Least time consuming – Single node job RosettaDesign Goal: Repacks and Optimize the side chains Workflow Distribute 5 million jobs from Match run through Scheduler/Grid

17 Rosetta Match SGE Composite Actor Create Sun Grid Engine Jobs Submit Jobs and Wait for Finish Job Output Processing

Command line instructions for two cluster workflow (Hoffman2 and Hydro) Retrieve proxy from myproxy.ucgrid.org /u/local/apps/kepler/kepler.modules/build-area/kepler.sh -MyProxyFilePath /tmp/x509up_uid -GlobusTaskPath4Hoffman2 /u/scratch/ppk/rosetta/rosetta-grid-workflow-hoffman2 -GlobusTaskPath4Hydro /u/scratch/ppk/rosetta/rosetta-grid-workflow-hydro -ScaffoldPath4Hoffman2 /u/home2/ppk/rosettatest/scaffold-hoffman2-one-scaffold -ScaffoldPath4Hydro /u/home2/ppk/rosettatest/scaffold-hydro-one-scaffold /u/local/apps/kepler/rosetta-workflow/Rosetta-Match-BChain-Design- Globus-TopLevel-Parallel-v1-18.xml TaskPaths are scratch directories for the run and ScaffoldPath is the location of input files for each cluster pre-divided by the user. Workflow directives are in the xml file

19 Experimental data on Hoffman2 cluster Workflow nameWorkflow StructureInputsMain configurationsJob numberExecution time Rosetta- MatchSGE-v2- 4.xml MatchSGEJob226 Scaffold 1) Match command number for each job: 5 Generate 108,012 pdb files. 46 Match SGE jobs hours Rosetta- MatchSGE- BChain-v1-6.xml MatchSGEJob + BChain226 Scaffold 1) Match command number for each job: 5 Generate 216,024 pdb files. 46 Match SGE jobs hours Rosetta- DesignSGE-v1- 8.xml DesignSGEJob (each design includes cst, des and min) 10 Scaffold 1) Structure Number for Design Calculation: 10. 2) Command number for each job: 5 (cst) + 5 (des) + 50 (min) = 60. Generate pdb files. 286 Design SGE jobs hours Rosetta- MatchSGE- BChain- DesignSGE-v1- 9.xml SGEMatch, BChain, SGEDesign 10 scaffold 1) Command number for each Match job: 1. 2) Structure Number for Design Calculation: 10. 3) Command number for each Design job: 5 (cst) + 5 (des) + 50 (min) = Match Jobs and 286 Design Jobs hours. Rosetta- DesignSGE-v1- 8.xml DesignSGEJob (each design includes cst, des and min) 226 Scaffold 1) Structure Number for Design Calculation: ) Command number for each job: 10 (cst) + 10 (des) (min) = The whole cmd number could be 50 million and the whole execution could generate 7 million files: * * 10 = = 7,793,734,24. (2036 is the pdb file number for 10 Scaffold of Rosetta-MatchSGE-v2- 4.xml) Over 11,000 Design SGE jobs. One job may last for over 18 hours. Estimated whole execution time : (108012/2036) * * 10 = 7.79 months.

Workflow to Use Shibboleth Authentication to Sign X509 Certificate ePPN provides FN: LN: Org: Role: Username: Password: Username: Password: Identity Provider Certificate exist for the ePPN MyProxy Server ePPN/S P-ID Create Certificate Generate proxy Create Certificate Generate proxy Web Service NO Push YES SLC Service Provider Encrypt (RandomPassword, ePPN + SP- ID)) Two ways encryption/decryption Store the encrypted password to DB Encrypt (RandomPassword, ePPN + SP- ID)) Two ways encryption/decryption Store the encrypted password to DB Retrieve password from DB Decrypt (DBPassword, ePPN + SP-ID)) Two ways encryption/decryption Retrieve password from DB Decrypt (DBPassword, ePPN + SP-ID)) Two ways encryption/decryption Username/password

Cloud UC Cloud Architecture Public Network UC Cloud Controller Campus Resource Node controller Cluster controller Off Campus Resource Virtual OS

UC Grid Portal Contact Information Kepler: Jianwu Wang, SDSC, INCA, SDSC: Shava Smallen, UGP: Kejian Jin Rosetta: Proteomic: UCLA: Prakashan Korambath, UCSB: Paul Weakliem, UCR: Bill Strossman, UCI: Harry Mangalam,