APST Internals Sathish Vadhiyar. apstd daemon should be started on the local resource Opens a port to listen for apst client requests Runs on the host.

Slides:



Advertisements
Similar presentations
Generic MPI Job Submission by the P-GRADE Grid Portal Zoltán Farkas MTA SZTAKI.
Advertisements

1 Automated SFTP Windows and SUN Linux and SUN. 2 Vocabulary  Client = local=the machine generating the SFTP request  Server = remote = the machine.
Grid Resource Allocation Management (GRAM) GRAM provides the user to access the grid in order to run, terminate and monitor jobs remotely. The job request.
CERN LCG Overview & Scaling challenges David Smith For LCG Deployment Group CERN HEPiX 2003, Vancouver.
A Computation Management Agent for Multi-Institutional Grids
EGEE-II INFSO-RI Enabling Grids for E-sciencE Supporting MPI Applications on EGEE Grids Zoltán Farkas MTA SZTAKI.
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
Asynchronous Solution Appendix Eleven. Training Manual Asynchronous Solution August 26, 2005 Inventory # A11-2 Chapter Overview In this chapter,
Grids and Globus at BNL Presented by John Scott Leita.
Installing and running COMSOL on a Windows HPCS2008(R2) cluster
Lecture 8 Configuring a Printer-using Magic Filter Introduction to IP Addressing.
Resource Management Reading: “A Resource Management Architecture for Metacomputing Systems”
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Overview of TeraGrid Resources and Usage Selim Kalayci Florida International University 07/14/2009 Note: Slides are compiled from various TeraGrid Documentations.
Zach Miller Computer Sciences Department University of Wisconsin-Madison What’s New in Condor.
BaBar WEB job submission with Globus authentication and AFS access T. Adye, R. Barlow, A. Forti, A. McNab, S. Salih, D. H. Smith on behalf of the BaBar.
 Accessing the NCCS Systems  Setting your Initial System Environment  Moving Data onto the NCCS Systems  Storing Data on the NCCS Systems  Running.
AliEn uses bbFTP for the file transfers. Every FTD runs a server, and all the others FTD can connect and authenticate to it using certificates. bbFTP implements.
SAN DIEGO SUPERCOMPUTER CENTER Working with Inca Reporters Jim Hayes Inca Workshop September 4-5, 2008.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
Computational grids and grids projects DSS,
Grid Computing I CONDOR.
COMP3019 Coursework: Introduction to GridSAM Steve Crouch School of Electronics and Computer Science.
Client – Server Application Can you create a client server application: The server will be running as a service: does not have a GUI The server will run.
3-2.1 Topics Grid Computing Meta-schedulers –Condor-G –Gridway Distributed Resource Management Application (DRMAA) © 2010 B. Wilkinson/Clayton Ferner.
Linux Services Muhammad Amer. 2 xinetd Programs  In computer networking, xinetd, the eXtended InterNET Daemon, is an open-source super-server daemon.
Part 6: (Local) Condor A: What is Condor? B: Using (Local) Condor C: Laboratory: Condor.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
Some Design Notes Iteration - 2 Method - 1 Extractor main program Runs from an external VM Listens for RabbitMQ messages Starts a light database engine.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
Condor: High-throughput Computing From Clusters to Grid Computing P. Kacsuk – M. Livny MTA SYTAKI – Univ. of Wisconsin-Madison
Grid Compute Resources and Job Management. 2 Local Resource Managers (LRM)‏ Compute resources have a local resource manager (LRM) that controls:  Who.
Andrey Meeting 7 October 2003 General scheme: jobs are planned to go where data are and to less loaded clusters SUNY.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
Privilege separation in Condor Bruce Beckles University of Cambridge Computing Service.
Open Science Grid OSG CE Quick Install Guide Siddhartha E.S University of Florida.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
1 Short Course on Grid Computing Jornadas Chilenas de Computación 2010 INFONOR-CHILE 2010 November 15th - 19th, 2010 Antofagasta, Chile Dr. Barry Wilkinson.
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
Part Five: Globus Job Management A: GRAM B: Globus Job Commands C: Laboratory: globusrun.
Review of Condor,SGE,LSF,PBS
FTP File Transfer Protocol Graeme Strachan. Agenda  An Overview  A Demonstration  An Activity.
1 High-Performance Grid Computing and Research Networking Presented by David Villegas Instructor: S. Masoud Sadjadi
SAN DIEGO SUPERCOMPUTER CENTER Inca Control Infrastructure Shava Smallen Inca Workshop September 4, 2008.
Pilot Factory using Schedd Glidein Barnett Chiu BNL
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
Job Submission with Globus, Condor, and Condor-G Selim Kalayci Florida International University 07/21/2009 Note: Slides are compiled from various TeraGrid.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
WinSCP  Tool for accessing files on beaglebone system.
Configuring and Deploying Web Applications Lesson 7.
LSF Universus By Robert Stober Systems Engineer Platform Computing, Inc.
Grid Compute Resources and Job Management. 2 Grid middleware - “glues” all pieces together Offers services that couple users with remote resources through.
Jaime Frey Computer Sciences Department University of Wisconsin-Madison What’s New in Condor-G.
STAR Scheduling status Gabriele Carcassi 9 September 2002.
Open Science Grid Build a Grid Session Siddhartha E.S University of Florida.
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison Condor NT Condor ported.
EGEE 3 rd conference - Athens – 20/04/2005 CREAM JDL vs JSDL Massimo Sgaravatto INFN - Padova.
1 Figure 6-5: Turning Off Unnecessary Services Unnecessary Services  Operating system vendors used to install many services by default  This made them.
STAR Scheduler Gabriele Carcassi STAR Collaboration.
HTCondor’s Grid Universe Jaime Frey Center for High Throughput Computing Department of Computer Sciences University of Wisconsin-Madison.
Wouter Verkerke, NIKHEF 1 Using ‘stoomboot’ for NIKHEF-ATLAS batch computing What is ‘stoomboot’ – Hardware –16 machines, each 2x quad-core Pentium = 128.
+ Vieques and Your Computer Dan Malmer & Joey Azofeifa.
Reading e-Science Centre
Globus Job Management. Globus Job Management Globus Job Management A: GRAM B: Globus Job Commands C: Laboratory: globusrun.
Telnet/SSH Connecting to Hosts Internet Technology.
CUPS Print Services.
Michael P. McCumber Task Force Meeting April 3, 2006
Quick Tutorial on MPICH for NIC-Cluster
PolyAnalyst Web Report Training
Presentation transcript:

APST Internals Sathish Vadhiyar

apstd daemon should be started on the local resource Opens a port to listen for apst client requests Runs on the host where input files are located Input files can also be specified by running element apstd automatically copies output files from working directory to where apstd is started apst and apstd started by same user since apstd writes files on behalf of apst’s user

A APST run is associated with a XML file Task dependency can be enforced by APST XML <apst> </apst>

Some times there may not be file dependency but task dependency

XML example

security Some kind of security regarding what kind of commands will apstd accept over socket Given a description of the tasks to do and the resources (disks and machines) available, APST will assign individual tasks to available machines, copy the input files, run the tasks, and return the output files. APST also tries to assign tasks to machines intelligently, using information such as the load and speed of individual machines. The main APST program, apstd, handles all of the task assignment, application execution, and file copying. Splitting the control and user interface portions of APST like this allows you, for example, to run apstd on your main system but control it from your laptop.

Using local resources <apst> <task executable='perl' arguments='/home/${USER}/apst/Examples/charcount.pl /home/${USER}/apst/Examples/charcount0.dat' <task executable='perl' arguments='/home/${USER}/apst/Examples/charcount.pl /home/${USER}/apst/Examples/charcount0.dat' stdout='charcount0.out' /> stdout='charcount0.out' /> </apst> /home/${USER}/apst/bin/apstd -d --port 7890 first.xml

APST can use remote machines accessed through either a Globus GRAM or ssh, remote storage accessed through a Globus GASS server, scp, ftp, sftp, or an SRB server, and queueing systems controlled by Condor, DQS, LoadLeveler, LSF, PBS, or SGE.

Accessing remote resources – walk through <apst> </apst> Launches task on blueHost through ssh but assume files on local disk can be directly accessed

This tells apstd that blueHost can see files available on blueDisk, rather than those on the local disk.

The problem with this XML is that it requires that APST be installed on the remote machine in /home/${USER}/apst, since the arguments task attribute refers to files in this directory.

Equivalent to scp /home/${USER}/apst/Examples/charcount.pl blue.ufo.edu:/tmp/charcount.pl scp /home/${USER}/apst/Examples/charcount0.dat blue.ufo.edu:/tmp/charcount0.dat ssh blue.ufo.edu 'cd /tmp; perl./charcount.pl./charcount0.dat > charcount0.out' scp blue.ufo.edu:/tmp/charcount0.out /home/${USER}/apst/Examples/charcount0.out

Run the above example: /home/${USER}/apst/bin/apstd -d --port 7890 second.xml For globus: Scp -> gass Scp -> gass Ssh -> globus Ssh -> globus - i.e. machine and port where gatekeeper is running - i.e. machine and port where gatekeeper is running E.g.. E.g.. Run grid-proxy-init before starting apstd Run grid-proxy-init before starting apstd

Apst client program You can use apst to examine your application's state, add, stop, or restart tasks, and add or disable resources /home/${USER}/apst/bin/apst --host localhost:7890 command

Accessing batch systems Can replace pbs with lsf, condor, loadleveler

Gridinfo tag

Apstd daemon Can be started –heuristic= option. Default is wq Xml file has,,, Xml file has,,, <disk> Attributes – unique id, datadir Attributes – unique id, datadir Access method element Access method element Access method can be,,,,, or Access method can be,,,,, or

<host> Attributes – unique ID, cpus, disk, dnsname, memory, wd Attributes – unique ID, cpus, disk, dnsname, memory, wd Access method -,, or Access method -,, or Batch queuing system -,,,,, or Batch queuing system -,,,,, or Attributes – account, memory, node, nodetype, queue, stdin, stdout, stderr, time, option

<files> Specifies input, output and executable files Specifies input, output and executable files Contains one or more file attribute Contains one or more file attribute<file> Input files may have transfer attribute (yes or no) – whether files have to be transferred from submitting machine Input files may have transfer attribute (yes or no) – whether files have to be transferred from submitting machine Output files have analogously download attribute, may also have size attribute indicating the size of the output file – useful for scheduling decisions Output files have analogously download attribute, may also have size attribute indicating the size of the output file – useful for scheduling decisions

element may have element for input files element may have element for input files To indicate the placement of copies of the file that you have pre-staged to remote disks Will have disk attribute and copy attribute

<task> Attributes – executable, id, groups, wd, arguments, input, stdin, stdout, stderr, priority, host, memory, cost Attributes – executable, id, groups, wd, arguments, input, stdin, stdout, stderr, priority, host, memory, cost<infosource> Access method -,,, or. Access method -,,, or.