National Center for Supercomputing Applications Production Cyberenvironment for a A Computational Chemistry Grid PRAGMA13, NCSA 26 Sep 07 Sudhakar Pamidighantam.

Slides:



Advertisements
Similar presentations
PIONIER 2003, Poznan, , PROGRESS Grid Access Environment for SUN Computing Cluster Poznań Supercomputing and Networking Center Cezary Mazurek.
Advertisements

DATE: 2008/03/11 NCHC-Grid Computing Portal (NCHC-GCE Portal) Project Manager: Dr. Weicheng Huang Developed Team: Chien-Lin Eric Huang Chien-Heng Gary.
March 6 th, 2009 OGF 25 Unicore 6 and IPv6 readiness and IPv6 readiness
Inetrconnection of CNGrid and European Grid Infrastructure Depei Qian Beihang University Feb. 20, 2006.
System Software Overview of China National Grid VEGA R&D Team, Research Centre for Grid and Service Computing, ICT, CAS.
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
C3.ca in Atlantic Canada Virendra Bhavsar Director, Advanced Computational Research Laboratory (ACRL) Faculty of Computer Science University of New Brunswick.
Web Services Darshan R. Kapadia Gregor von Laszewski 1http://grid.rit.edu.
High Performance Computing Course Notes Grid Computing.
Distributed components
G O B E Y O N D C O N V E N T I O N WORF: Developing DB2 UDB based Web Services on a Websphere Application Server Kris Van Thillo, ABIS Training & Consulting.
Workshop on Cyber Infrastructure in Combustion Science April 19-20, 2006 Subrata Bhattacharjee and Christopher Paolini Mechanical.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
Software – Part 3 V.T. Raja, Ph.D., Information Management College of Business Oregon State University.
Simo Niskala Teemu Pasanen
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Cluster Computing through an Application-oriented Computational Chemistry Grid Kent Milfeld and Chona Guiang, Sudhakar Pamidighantam, Jim Giuliani Supported.
GridChem Refactoring: Workflows in GridChem Sudhakar Pamidighantam
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
National Center for Supercomputing Applications GridChem: Integrated Cyber Infrastructure for Computational Chemistry Sudhakar.
GridChem-- User Support Kent Milfeld Supported by the NSF NMI Program under Award # Oct. 10, 2005.
Web Services An introduction for eWiSACWIS May 2008.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
GridChem A Computational Chemistry Cyber-infrastructure Sudhakar Pamidighantam NCSA, University of Illinois at Urabana Champaign
Web Services Description Language (WSDL) Jason Glenn CDA 5937 Process Coordination in Service and Computational Grids September 30, 2002.
COMP3019 Coursework: Introduction to GridSAM Steve Crouch School of Electronics and Computer Science.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
Web Services based e-Commerce System Sandy Liu Jodrey School of Computer Science Acadia University July, 2002.
© 2008 Open Grid Forum Independent Software Vendor (ISV) Remote Computing Primer Steven Newhouse.
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
The PROGRESS Grid Service Provider Maciej Bogdański Portals & Portlets 2003 Edinburgh, July 14th-17th.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Web Services BOF This is a proposed new working group coming out of the Grid Computing Environments Research Group, as an outgrowth of their investigations.
GEM Portal and SERVOGrid for Earthquake Science PTLIU Laboratory for Community Grids Geoffrey Fox, Marlon Pierce Computer Science, Informatics, Physics.
Application portlets within the PROGRESS HPC Portal Michał Kosiedowski
GridChem A Computational Chemistry Cyber-infrastructure Using Web services Sanibel Symposium 23 Feb 07 Sudhakar Pamidighantam NCSA, University of Illinois.
Web Services. Abstract  Web Services is a technology applicable for computationally distributed problems, including access to large databases What other.
San Diego Supercomputer Center National Partnership for Advanced Computational Infrastructure SRB + Web Services = Datagrid Management System (DGMS) Arcot.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
IODE Ocean Data Portal - ODP  The objective of the IODE Ocean Data Portal (ODP) is to facilitate and promote the exchange and dissemination of marine.
Grid Security: Authentication Most Grids rely on a Public Key Infrastructure system for issuing credentials. Users are issued long term public and private.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
GCRC Meeting 2004 BIRN Coordinating Center Software Development Vicky Rowley.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University.
Kemal Baykal Rasim Ismayilov
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
Some comments on Portals and Grid Computing Environments PTLIU Laboratory for Community Grids Geoffrey Fox, Marlon Pierce Computer Science, Informatics,
National Computational Science National Center for Supercomputing Applications National Computational Science Integration of the MyProxy Online Credential.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
Web services for Production Cyberenvironment for a A Computational Chemistry Grid University of Hyderabad, India 17 March 07 Sudhakar Pamidighantam NCSA,
GridChem Architecture Overview Rion Dooley. Presentation Outline Computational Chemistry Grid (CCG) Current Architectural Overview CCG Future Architectural.
GridChem Sciene Gateway and Challenges in Distributed Services Sudhakar Pamidighantam NCSA, University of Illinois at Urbaba- Champaign
A Desktop Client for HPC Chemistry Applications: GridChem Kent Milfeld Supported by the NSF NMI Program under Award #
GridChem Production Cyberenvironment for Computational Chemistry Pragma 12 Conference 21 March 07 Sudhakar Pamidighantam NCSA, University of Illinois at.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
GridChem Current Status
Enable computational and experimental  scientists to do “more” computational chemistry by providing capability  computing resources and services at their.
Some Basics of Globus Web Services
Introduction to XSEDE Resources HPC Workshop 08/21/2017
Patrick Dreher Research Scientist & Associate Director
Data Management Components for a Research Data Archive
Presentation transcript:

National Center for Supercomputing Applications Production Cyberenvironment for a A Computational Chemistry Grid PRAGMA13, NCSA 26 Sep 07 Sudhakar Pamidighantam NCSA, University of Illinois at Urbana-Champaign

National Center for Supercomputing Applications Acknowledgements

National Center for Supercomputing Applications Outline Historical Background Grid Computational Chemistry Production Environments Current Status Web Services Usage (Grid and Science Achievements) Brief Demo Future

National Center for Supercomputing Applications Motivation Software - Reasonably Mature and easy to use to address chemists questions of interest Community of Users - Need and capable of using the software Some are non traditional computational chemists Resources - Various in capacity and capability

National Center for Supercomputing Applications Background Qauntum Chemistry Remote Job Monitor ( Quantum Chemistry Workbench) 1998, NCSA Chemviz , NSF (USA) Technologies Web Based Client Server Models Visual Interfaces Distributed computing (Condor)

National Center for Supercomputing Applications GridChem NCSA Alliance was commissioned 1998 Diverse HPC systems deployed both at NCSA and Alliance Partner Sites Batch schedulers different at sites Policies favored different classes and modes of use at different sites/HPC systems

National Center for Supercomputing Applications Extended TeraGrid Facility

National Center for Supercomputing Applications NSF Petascale Road Map Track I Scheme Multi petaflop single site system to be deployed by 2010 Several Consortia Competing (Now under review) Track 2 Sub petaflop systems Several to be deployed until Track 1 is online First one will be at TACC ( 450 TFlops) Available Fall 2007 ( Processors/Cores) NCSA is deploying a 110 TFlops in April 2007 (10000 Processors/cores) Second subpetaflops systems being reviewed

National Center for Supercomputing Applications Grid and Gridlock Alliance lead to Physical Grid Grid lead to TeraGrid Homogenous Grid with predefined fixed software and system stack was planned (Teragrid) but it was difficult to keep it homogenous Local preferences and diversity leads to heterogeneous grids now! (Operating Systems, Schedulers, Policies, Software and Services ) Openness and standards that lead interoperability are critical for successful services

National Center for Supercomputing Applications Current Grid Status Grid Hardware Middleware Scientific Applications

National Center for Supercomputing Applications User Community Chemistry and Computational Biology User Base Sep 03 – Oct 04 NRAC AAB Small Allocations #PIs #SUs 5,953,100 1,374, ,000

National Center for Supercomputing Applications

Some User Issues Addressed by the new Services New systems meant learning new commands Porting Codes Learning new job submissions and monitoring protocols New proposals for time (time for new proposals) Computational modeling became more popular and number of users increased (User Management) Batch queues are longer / waiting increased Finding resources where to compute complicated - probably multiple distributed sites Multiple proposals/allocations/logins Authentication and Data Security Data management

National Center for Supercomputing Applications Computational Chemistry Grid This is a Virtual Organization Integrated Cyber Infrastructure for Computational Chemistry Integrates Applications, Middleware, HPC resources, Scheduling and Data management Allocations, User services and Training

National Center for Supercomputing Applications Resources System (Site)Procs Avail Total CPU Hours/Year Status Intel Cluster (OSC)36315,000 SMP and Cluster nodes HP Integrity Superdome (UKy) 33290,000 TB Replaced with an SMP/ Cluster nodes IA32 Linux Cluster (NCSA) 64560,000 Allocated Intel Cluster (LSU)10241,000,000 Allocated IBM Power4 (TACC)16140,000 Allocated Teragrid (Multiple Institutions) ,000New Allocations Expected The initial Acesss Grid Testbed Nodes (38) and Condor SGI resources (NCSA, 512 nodes) have been retired this year.

National Center for Supercomputing Applications Other Resources Extant HPC resources at various Supercomputer Centers (Interoperable) Optionally Other Grids and Hubs/local/personal resources These may require existing allocations/Authorization

National Center for Supercomputing Applications

Grid Middleware Proxy Server GridChem System user Portal Client Grid Services Grid applicationapplication Mass Storage

National Center for Supercomputing Applications Applications GridChem supports some apps already –Gaussian, GAMESS, NWChem, Molpro, ADF,QMCPack, Amber Schedule of integration of additional software –ACES-3 –Crystal –Q-Chem –Wein2K –MCCCS Towhee –Others...

National Center for Supercomputing Applications GridChem User Services Allocation Request

National Center for Supercomputing Applications GridChem User Services Consulting Ticketing System User View

National Center for Supercomputing Applications GridChem User Services Consulting Ticketing System Consultants View

National Center for Supercomputing Applications Gridchem Middleware Service (GMS)

National Center for Supercomputing Applications GrdiChem Web Services Quick Primer XML is used to tag the data, SOAP is used to transfer the data, WSDL is used for describing the services available and UDDI is used for listing what services are available. Web Services is different from Web Page Systems or Web Servers: There is no GUI Web Services Share business logic, data & processes through APIs with each other (not with user) Web Services describe Standard way of interacting with web based applications A client program connecting to a web service can read the WSDL to determine what functions are available on the server. Any special datatypesdatatypes used are embedded in the WSDL file in the form of XML Schema. Universal Description, Discovery, and Integration. WSRF Standards Compliant.

National Center for Supercomputing Applications GridChem Web Services Client Objects Database Interaction WS Resources DTO ObjectsHibernate Databasehb.xml Client DTO (Data Transfer Object) Serialize transfer through XML DAO (Data Access Object) How to get the DB objects hb.xml (Hibernate Data Map) describes obj/column data mapping Business Model DAO

National Center for Supercomputing Applications GridChem Data Models UsersProjectsResources UserProjectResource SoftwareResources ComputeResources NetworkResources StorageResources Resources resoruceID Type hostName IPAddress siteID userID projectID resourceID loginName SUsLocalUserUsed Jobs jobID jobName userID projID softID cost UsersResources

National Center for Supercomputing Applications Computational Chemistry Resource

National Center for Supercomputing Applications GMS_WS Use Cases Authentication Job Submission Resource Monitoring Job Monitoring File Retrieval …

National Center for Supercomputing Applications GetResourceProperty SetTerminationTime Destroy Create Login LoadVO RetrieveFiles LoadFiles DeleteFiles LoadParentFiles RefreshFiles MakeDirectory SubmitJob SubmitMultipleJobs PredictJobStartTime KillJob HideJob UnhideJob UnhideJobs DeleteJob FindJobs GetJobStatus RetrieveJobOutput RetrieveNextDataBlock StopFileAction GetUserPreferences PutUserPreferences GridChem Web Services Operations

National Center for Supercomputing Applications GMS_WS Authentication WSDL (Web Service Definition Language) is a language for describing how to interface with XML-based services. It describes network services as a pair of endpoints operating on messages with either document-oriented or procedure-oriented information. The service interface is called the port type WSDL FILE: <definitions name=GMS" targetNamespace= " xmlns=" … Contact GMS Creates Session, Session RP and EPR Sends EPR ( Like a Cookie, but more than that) Login Request (username:passwd) Validates, Loads UserProjects Sends acknowledgement Retrieve UserProjects (GetResourceProperty Port Type [PT]) GridChem ClientGMS

National Center for Supercomputing Applications GMS_WS Authentication Selects project LoadVO port type (w. MAC address) Verifies user/project/MACaddr Load UserResources RP Retrieve UserResources [as userVO/ Profile] (GetResourceProperty port Type PT) GridChem ClientGMS Validates, Loads UserProjects Sends acknowledgement

National Center for Supercomputing Applications GMS_WS Job Submission Create Job object PredictJobStartTime PT + JobDTO JobStart Prediction RP PT = portType RP = Resource Properties DTO = Data Transfer Object Completion: from batch system to GMS server DB Submission CoGKit GAT gsi-ssh If decision OK, SubmitJob PT + JobDTO Create Job object APISubmit Store Job Object Send Acknowledgement Need to check to make sure allocation-time is available. GC ClientGMS

National Center for Supercomputing Applications GMS_WS Monitoring Parse XML, Display PT = portType RP = Resource Properties DTO = Data Transfer Object DB = Data Base server Servers Job Launcher Notifications VO Admin parses DB (status + cost) Request for Job, Resource Status Alloc. Balance UserResource RP Updated from DB GC ClientGMSResources/Kits/DB Send info Discover Applications (Software Resources) Monitor System Monitor Queues

National Center for Supercomputing Applications GMS_WS Job Status Job Status jobDTO.status Job Launcher Status Update Estimate Start time Scheduler s/ notifications Notifications: Client, , IM GC ClientGMSResources/Kits/DB

National Center for Supercomputing Applications GMS_WS File Retrieval (MSS) GetResourceProperty PT FileDTO(?) LoadFile PT (project folder+job) Validates project folder owned by user. Send new listing PT = portType RP = Resource Properties DTO = Data Transfer Object MSS = Mass Storage System Job Completion: Send Output to MSS LoadFile PT MSS query UserFiles RP + FileDTO object Retrieve Root Dir. Listing on MSS with CoGKit or GAT or gsi-ssh API file request Store locally Create FileDTO Load into UserData RP RetrieveFiles PT (+file rel.path) Retrieve file: CoGKit or GAT or gsi-ssh GetResourceProperty PT GC ClientGMSResources/Kits/DB

National Center for Supercomputing Applications GMS_WS File Retrieval PT = portType RP = Resource Properties DTO = Data Transfer Object MSS = Mass Storage System Create FileDTO (?) Load into UserData RP RetrieveJobOutput PT (+JobDTO) Job Record from DB. Running: from Resource Complete: from MSS Retrieve file: CoGKit or GAT or gsiftp GetResourceProperty PT GC ClientGMSResources/Kits/DB

National Center for Supercomputing Applications GridChem Web Services WSRF (Web Services Resource Framework) Compliant WSRF Specifications: WS-ResourceProperties (WSRF-RP) WS-ResourceLifetime (WSRF-RL) WS-ServiceGroup (WSRF-SG) WS-BaseFaults (WSRF-BF) %ps -aux | grep ws /usr/java/jdk1.5.0_05/bin/java \ -Dlog4j.configuration=container-log4j.properties \ -DGLOBUS_LOCATION=/usr/local/globus \ -Djava.endorsed.dirs=/usr/local/globus/endorsed \ -DGLOBUS_HOSTNAME=derrick.tacc.utexas.edu \ -DGLOBUS_TCP_PORT_RANGE=62500,64500 \ -Djava.security.egd=/dev/urandom \ -classpath /usr/local/globus/lib/bootstrap.jar: /usr/local/globus/lib/cog-url.jar: /usr/local/globus/lib/axis-url.jar org.globus.bootstrap.Bootstrap org.globus.wsrf.container.ServiceContainer -nosec Logging Configuration Where to find Globus Where to get random seed for encryption key generation Classpath (required jars)

National Center for Supercomputing Applications GridChem Software Organization Open Source Distribution CVS for GridChem

National Center for Supercomputing Applications Package: org.gridchem.service.gms GMS_WS

National Center for Supercomputing Applications GMS_WS + Should these each be a separate package?

National Center for Supercomputing Applications model dto credential job notification filefile.task job.task user exceptions resource persistence synch query test util dao gpir crypt enumerators gat proxy GMS_WS client audit gms Classes for WSRF service implementation (PT) Cmd line tests to mimic client requests Data Access Obj – queries DB via persistent classes (hibernate) Data Transfer Obj – (job,File,Hardware,Software,User) XML How to handle errors (exceptions) CCG Service business mode (how to interact) Contains users credentials for job sub. file browsing,… Oversees correct handling of user data (get/putfile). Define Job & util & enumerations (SubmitTask, KillTask,…) CCGResource&Util, Synched by GPIR, abstract classes NetworkRes., ComputeRes., SoftwareRes., StorageRes., VisualizationRes. User (has attributes – Preference/Address) DB operations (CRUD), OR Maps, pool mgmt,DB session, Classes that communicate with other web services Periodically update DB with GPIR info (GPIR calls) JUnit service test (gms.properties): authen. VO retrieval, Res.Query,Synch, Job Mgmt, File Mgmt, Notification Contains utility and singleton classes for the service. Encryption of login password Mapping from GMS_WS enumeration classes DB GAT util classes: GATContext & GAT Preferences generation Classes deal with CoGKit configuration. Autonomous notification via , IM, textmesg.

National Center for Supercomputing Applications GMS_WS external jars Testing For XML Parsing Java Document Object Model –Lightweight –Reading/Writing XML Docs –Complements SAX (parser) & DOM –Uses Collections**

National Center for Supercomputing Applications GridChem Resources Monitoring

National Center for Supercomputing Applications GridChem Resources New Computing Systems SystemCapacity(Cpus/Cores)Capability Mercury(NCSA)1774Small/Large Parallel Runs Abe(NCSA)9600Massively Parallel Runs DataStar(SDSC)2368SharedMemory Large Runs Bluegene/L(SDSC)3456Cluster Large Parallel Runs TeragridCluster(SDSC)564Small/Large Parallel Runs BigRed(IU)1024SharedMemory Small/Large Runs BCX (UKy)1360Shared/Distributed Memory small/Large Parallel Runs

National Center for Supercomputing Applications Application Software Resources Currently Supported SuiteVersionLocation Gaussian 03C.02/D.01Many Platforms MolPro2006.1NCSA NWChem5.0/4.7Many Platforms GamessJan 06Many Platforms Amber8.0Many Paltforms QMCPack2.0NCSA

National Center for Supercomputing Applications GridChem Software Resources New Applications Integration Underway ADF Amsterdam Density Functional Theory Wien2K Linearized Augemented Plain wave (DFT) CPMD Car Parinello Molecular Dynamics QChem Molecular Energetics (Quantum Chemistry) Aces3 Parallel Coupled Cluster Quantum Chemistry Gromacs Nano/Bio Simulations (Molecular Dynamics) NAMD Molecular Dynamics DMol3 Periodic Molecular Systems ( Quantum Chemistry) Castep Quantum Chemistry MCCCS-Towhee Molecular Confirmation Sampling (Monte Carlo) Crystal98/06 Crystal Optimizations (Quantum Chemistry) ….

National Center for Supercomputing Applications GridChem User Services Allocation Community and External Registration Reviews, PI Registration and Access Creation Community User Norms Established Consulting/User Services Ticket tracking, Allocation Management Documentation, Training and Outreach FAQ Extraction, Tutorials, Dissemination Help is integrated into the GridChem client

National Center for Supercomputing Applications Users and Usage 242 Users under 128 Projects Include Academic PIs, two graduate classes And about 15 training users More than a CPU Wallhours since Jan 06 More than Jobs processed

National Center for Supercomputing Applications Science Enabled Azide Reactions for Controlling Clean Silicon Surface Chemistry: Benzylazide on Si(100)-2 x 1 Semyon Bocharov et al.. J. Am. Chem. Soc., 128 (29), , 2006 Chemistry of Diffusion Barrier Film Formation: Adsorption and Dissociation of Tetrakis(dimethylamino)titanium on Si(100)-2 × 1 Rodriguez-Reyes, J. C. F.; Teplyakov, A. V. J. Phys. Chem. C.; 2007; 111(12); Computational Studies of [2+2] and [4+2] Pericyclic Reactions between Phosphinoboranes and Alkenes. Steric and Electronic Effects in Identifying a Reactive Phosphinoborane that Should Avoid Dimerization Thomas M. Gilbert* and Steven M. Bachrach Organometallics, 26 (10), , 2007.*

National Center for Supercomputing Applications Science Enabled Chemical Reactivity of the Biradicaloid (HO...ONO) Singlet States of Peroxynitrous Acid. The Oxidation of Hydrocarbons, Sulfides, and Selenides. Bach, R. D et al. J. Am. Chem. Soc. 2005, 127, The "Somersault" Mechanism for the P-450 Hydroxylation of Hydrocarbons. The Intervention of Transient Inverted Metastable Hydroperoxides. Bach, R. D.; Dmitrenko, O. J. Am. Chem. Soc. 2006, 128(5), The Effect of Carbonyl Substitution on the Strain Energy of Small Ring Compounds and their Six-member Ring Reference Compounds Bach, R. D.; Dmitrenko, O. J. Am. Chem. Soc. 2006,128(14), 4598.

National Center for Supercomputing Applications GridChem Client Download Statistics

National Center for Supercomputing Applications Distribution of GridChem User Community

National Center for Supercomputing Applications Job Distribution

National Center for Supercomputing Applications System Wide Usage HPC SystemUsage (SUs) Tungsten(NCSA)5507 Copper(NCSA)86484 CCGcluster(NCSA)55709 Condor(NCSA)30 SDX(UKy) CCGCluster(UKy).5 Longhorn(TACC)54 CCGCluster(OSC)62000 TGCluster(OSC)36936 Cobalt(NCSA)2485 Champion(TACC)11 Mike4 (LSU)14537

National Center for Supercomputing Applications GridChem Client Enhancements New Molecular Editor JMolEditor (ANU) Integration VMD Is integrated Nanotube Generator (Tubegen) Will be available Gamess Graphical User Interphase

National Center for Supercomputing Applications Java Molecular Editor JMolEditor Three Dimensional Visual with Java 3D Intuitive Molecule Manipulation Interactive Bond, Angle and Dihedral Settings A Gaussian input generator Interface

National Center for Supercomputing Applications Nanotube Generator:Tubegen Courtesy : Doren Research Group at the University of Delaware Crystal Cell Types Output Formats

National Center for Supercomputing Applications GridChem Gamess GUI

National Center for Supercomputing Applications GridChem Post Processing IR/Raman Spectra now accessible from G03, MolPro, NWChem and Gamess Suites VCD/ROA To be Included

National Center for Supercomputing Applications GridChem Post Processing Normal Mode Viewing in 3D VRML Other Spectra With MO Integration NMR Electronic Spectra

National Center for Supercomputing Applications GridChem Usability Dynamic Information

National Center for Supercomputing Applications GridChem Usability Information on Potential Start and End Time for a given set of Job parameters Automated Resource Selection Possible Job Migration In case of dropped nodes or incomplete job Monitoring Multiple Jobs Automated Monitoring Job Output

National Center for Supercomputing Applications Implementation of GRMS resource management Service Moving toward Service based job submission eliminating gateway interfaces Infrastructure for multiple input files for single application Infrastructure for multiple inputs in High Throughput processing Integrated workflow for multi scale coupled modeling Meta-scheduling for High Throughput Processing Match Making, Round-robin scheduling, Preferred Host Set usage GridChem Middleware Infrastructure Implementation Currently underway

National Center for Supercomputing Applications GridChem In New Collaborations Resource Providers New Resource Providers Open Science Grid Initially for Bio-related applications (open source preferably) PRAGMA Partner sites University of Hyderabad ORNL (Could be via TeraGrid) International Partners KISTI, APAC, Daresbury Labs

National Center for Supercomputing Applications Scientific Collaborations GridChem Extension to Molecular Sciences (Bio, Nano, Geo and Materials Sciences) (NSF Proposal) Parameter Sweep for Potential Energy Hyper Surfaces (Faculty Fellows, NCSA) Automated Parameterization of Force fields (NSF Proposal) Ab initio Molecular Dynamics (Faculty Fellows, NCSA) Education (CI-TEAM) (NSF Proposals) Multi-Scale Modeling (IACAT, UIUC )

National Center for Supercomputing Applications Some New GridChem Infrastructure Workflow Editors Coupled Application Execution Large Scale Computing Metadata and Archiving Rich Client Platform Refactorization Intergrid Interactions Open Source Distribution Open Architecture and Implementation details

National Center for Supercomputing Applications Critical Gateways Issues Science Gateways compete with business as usual for the end user research scientist No direct access to HPC systems may be possible leading to apparent lack of control for users No End to end solutions If part of the research needs require old ways Gateways may be avoided Learning to use Gateways should provide substantial added benefit –Cost/Benefit Issues for users Flexibility to integrate new applications as needed by community quickly is critical to keep the user community engaged

National Center for Supercomputing Applications Authentication

National Center for Supercomputing Applications Resource Status

National Center for Supercomputing Applications Job Editor

National Center for Supercomputing Applications Job Submission

National Center for Supercomputing Applications Job Monitoring

National Center for Supercomputing Applications Gradient Monitoring

National Center for Supercomputing Applications Energy Monitoring

National Center for Supercomputing Applications Post Processing

National Center for Supercomputing Applications Visualization Molecular Visualization Electronic Properties Spectra Vibrational Modes

National Center for Supercomputing Applications Molecular Visualization Better molecule representations (Ball and Stick/VDW/MS) In Nanocad Molecular Editor Third party visualizer integration Chime/VMD Export Possibilities to others interfaces Deliver standard file formats (XML,SDF,MSF,Smiles etc…)

National Center for Supercomputing Applications Eigen Function Visualization Molecular Orbital/Fragment Orbital MO Density Visualization MO Density Properties Other functions Radial distribution functions

National Center for Supercomputing Applications Some example Visuals Arginine Gamess/6-31G* Total electronic density 2D - Slices

National Center for Supercomputing Applications Electron Density in 3D Interactive (VRML)

National Center for Supercomputing Applications Orbital 2D Displays N2 6-31g* Gamess

National Center for Supercomputing Applications Orbital 3D VRML

National Center for Supercomputing Applications Spectra IR/Raman Vibrotational Spectra UV Visible Spectra Spectra to Normal Modes Spectra to Orbitals

National Center for Supercomputing Applications Possible H-bonds network for P450 cam hydroperoxy intermediate Suggested: THR252 accepts an H-bond from the hydroperoxy (Fe(III)- OOH that promotes the second protonation on the distal oxygen, leading to the O-O bond cleavage Nagano, S.; Poulos, T.L. J. Biol. Chem. 2005, 250, p.1668 Auclair, K.; Hu, Z.; Little, D. M.; Ortiz de Montellano, P. R.; Groves, J. T. J. Am. Chem. Soc. 2002, 124, 6020.

National Center for Supercomputing Applications The Somersault Isomerization of Model Cpd0 Robert Bach and Olga Dmytrenko, 2006

National Center for Supercomputing Applications Energy Diagram for the Concerted Non-synchronous Hydroxylation of Isobutane Energy diagram (kcal/mol) for the oxidation of the isobutane with ground state, 24a (GS-8 hydrogen bonded to isobutane). MIN-24b [model oxidant MIN-10 (PorFe(SH)O HO) hydrogen bonded to isobutene] is not necessarily on the reaction pathway.

National Center for Supercomputing Applications Somersault Mechanism Summary for Isobutane Hydroxylation

National Center for Supercomputing Applications TetrakisDimethylAminoTitanium and its derivatives on Si(100)- 2x1 Surface: Diffusion Barrier Thinfilms on Silicon Rodrigues-Reyes and Teplyakov

National Center for Supercomputing Applications Benzylazide on Si(100)-2x1 Surface Deposition of Aromatic Moieties on Silicon for Lateral Electron Transfer Bocharov et al..

National Center for Supercomputing Applications [2+2] Cyclo Additions involving B=P Bonds Gilbert and Bachrach Dimerization Ethyne Addition Ethene Additions

National Center for Supercomputing Applications Possible H-bonds network for P450 cam hydroperoxy intermediate Suggested: THR252 accepts an H-bond from the hydroperoxy (Fe(III)- OOH that promotes the second protonation on the distal oxygen, leading to the O-O bond cleavage Nagano, S.; Poulos, T.L. J. Biol. Chem. 2005, 250, p.1668 Auclair, K.; Hu, Z.; Little, D. M.; Ortiz de Montellano, P. R.; Groves, J. T. J. Am. Chem. Soc. 2002, 124, 6020.

National Center for Supercomputing Applications The Somersault Isomerization of Model Cpd0 Robert Bach and Olga Dmytrenko, 2006

National Center for Supercomputing Applications Energy Diagram for the Concerted Non-synchronous Hydroxylation of Isobutane Energy diagram (kcal/mol) for the oxidation of the isobutane with ground state, 24a (GS-8 hydrogen bonded to isobutane). MIN-24b [model oxidant MIN-10 (PorFe(SH)O HO) hydrogen bonded to isobutene] is not necessarily on the reaction pathway.

National Center for Supercomputing Applications Somersault Mechanism Summary for Isobutane Hydroxylation

National Center for Supercomputing Applications Unsymmetrical Mo(CO) 4 Crown Ethers

National Center for Supercomputing Applications Dibenzaphosphepin based bis(phosphorous)polyether chelated Mo(CO) 4

National Center for Supercomputing Applications Crystal Structures CSD:XAPZAP cis-(6,6'-((1,1'-Binaphthyl)-2,2'- diylbis(oxy))bis(dibenzo(d,f)(1,3,2)dioxaphosp hepin))-tetracarbonyl-molybdenum(0) C48 H28 Mo1 O10 P2 CSD:DEQDOS cis-Tetracarbonyl-(P,P'-(6-(2'-oxy-2-biphenyl)-3,6- dioxa-hexanolato)-bis(dibenzo (d,f)(1,3,2)dioxaphosphepine)-P,P')-molybdenum C44 H32 Mo1 O12 P2

National Center for Supercomputing Applications Starting Structure

National Center for Supercomputing Applications Optimized Structure

National Center for Supercomputing Applications Reference Structure for Comparison 8 7

National Center for Supercomputing Applications Structural Comparisons C-C Torsion Angles for the OCH 2 CH 2 O Fragments and for the Axially Chiral Biaryl Groups Atoms PCMODEL* UFF Ab Initio Amber C37-C42-C43-C C1-C6-C7-C C13-C22-C23-C C32-O-C33-C O-C33-C34-O C33-C34-O-C C34-O-C35-C O-C35-C *Hariharasarma, et al. Organomet., , Ab Initio=B3LYP/3-21G* Amber9 ff03, GAFF, chloroform, 300K, median over 1ns MD

National Center for Supercomputing Applications MD OCH 2 CH 2 O Structure 8 7

National Center for Supercomputing Applications MD Biaryl Structure

National Center for Supercomputing Applications 1 H NMR Chemical Shift Comparison For Aromatic Protons Reference 32ppm (from TMS B3LYP/6-31g*) Atom Exp.AbinitioAtomExp.Abinitio H H H H H H H H H H H H H H H H H H H H H H H H H H H H

National Center for Supercomputing Applications Third Year Plans Post Processing Spectra and related entities New Application Support Aces3, Dmol3, Vasp,….. Expansion of Resources Teragrid, OSG, Pragma Systems and New resources at Partner Sites Extension Plan Two Proposals in review for Extension

National Center for Supercomputing Applications Future Plans Preparations for Petaflop computing High throughput massively parallel applications Complex workflows for integrating multiple interdependent applications Multiscale Computing Archiving and annotating data for future use Open Data initiatives by NIH and NSF

National Center for Supercomputing Applications Acknowledgments Rion Dooley, TACC Middleware Infrastructure Stelios Kyriacou, OSC Middleware Scripts Chona Guiang, TACC Databases and Applications Kent Milfeld, TACC Database Integration Kailash Kotwani, NCSA, Applications and Middleware Scott Brozell, OSC, Applications and Testing Michael Sheetz, UKy, Application Interfaces Vikram Gazula, UKy, Server Administration Tom Roney, NCSA, Server and Database Maintenance