CE: compute element TP: CE & WN Compute Element Worker Node Installation configuration.

Slides:



Advertisements
Similar presentations
Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
Advertisements

12th EELA TUTORIAL - USERS AND SYSTEM ADMINISTRATOR E-infrastructure shared between Europe and Latin America CE + WN installation and configuration.
Instalación y configuración de CE+WN Angelines Alberto CIEMAT Grid Tutorial, Sept
Grid Resource Allocation Management (GRAM) GRAM provides the user to access the grid in order to run, terminate and monitor jobs remotely. The job request.
High Performance Computing
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
Basic Grid Job Submission Alessandra Forti 28 March 2006.
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
INFSO-RI Enabling Grids for E-sciencE Computing Element installation & configuration Giuseppe Platania INFN Catania EMBRACE Tutorial.
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
Accounting Update Dave Kant Grid Deployment Board Nov 2007.
1 Kolkata, Asia Joint CHAIN/EU-IndiaGrid2/EPIKH School for Grid Site Administrators, The EPIKH Project (Exchange Programme.
1 Worker Nodes Installation&Configuration Sara Bertocco INFN Padova 11 th International GridKa School 2013 – Big Data, Clouds and Grids.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) CE+WN+siteBDII Installation and configuration Bouchra
Enabling Grids for E-sciencE SGE J. Lopez, A. Simon, E. Freire, G. Borges, K. M. Sephton All Hands Meeting Barcelona, Spain 23 May 2007.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) Workload Management System + Logging&Bookkeeping Installation.
INFSO-RI Enabling Grids for E-sciencE Workload Management System Mike Mineter
Rochester Institute of Technology Job Submission Andrew Pangborn & Myles Maxfield 10/19/2015Service Oriented Cyberinfrastructure Lab,
The EDGeS project receives Community research funding 1 SG-DG Bridges Zoltán Farkas, MTA SZTAKI.
Grid job submission using HTCondor Andrew Lahiff.
9th EELA TUTORIAL - USERS AND SYSTEM ADMINISTRATORS E-infrastructure shared between Europe and Latin America CE + WN installation and configuration.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Feb. 06, Introduction to High Performance and Grid Computing Faculty of Sciences,
E-science grid facility for Europe and Latin America GridwWin: porting gLite to run under Windows Fabio Scibilia – Consorzio COMETA 30/06/2008.
CEOS WGISS-21 CNES GRID related R&D activities Anne JEAN-ANTOINE PICCOLO CEOS WGISS-21 – Budapest – 2006, 8-12 May.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America BDII Server Installation and Configuration Antonio Juan.
BDII Server Installation and Configuration Manuel Rubio del Solar Extremadura Advanced Research Center (CETA-CIEMAT) 11th EELA Tutorial for Users Sevilla,
4th EELA TUTORIAL - USERS AND SYSTEM ADMINISTRATORS E-infrastructure shared between Europe and Latin America CE + WN installation and configuration.
Enabling Grids for E-sciencE SGE J. Lopez, A. Simon, E. Freire, G. Borges, K. M. Sephton All Hands Meeting Dublin, Ireland 12 Dec 2007 Batch system support.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America Worker Node installation & configuration.
E-infrastructure shared between Europe and Latin America Introduction to the tutorial for site managers Vanessa Hamar Universidad de Los.
Proposal for a IS schema Massimo Sgaravatto INFN Padova.
gLite Information System UNIANDES OOD Team Daniel Alberto Burbano Sefair, Michael Angel.
1 Grid2Win: porting of gLite middleware to Windows Dario Russo INFN Catania
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
Shell Interface Shell Interface Functions Data. Graphical Interface Graphical Interface Command-line Interface Command-line Interface Experiments Private.
CREAM Installation&Configuration Sara Bertocco INFN Padova 11 th International GridKa School 2013 – Big Data, Clouds and Grids.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Grid2Win : gLite for Microsoft Windows Roberto.
Third EELA Tutorial for Managers and Users E-infrastructure shared between Europe and Latin America CE + WN installation and configuration.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Progress report from University of Cyprus.
12th EELA TUTORIAL - USERS AND SYSTEM ADMINISTRATORS FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin.
South African Grid Training COMPUTING ELEMENT Albert van Eck UFS - ICTS 18 November 2009 Slides by: GIUSEPPE PLATANIA.
TP: Grid site installation BEINGRID site installation.
E-science grid facility for Europe and Latin America COMPUTING ELEMENT GIUSEPPE PLATANIA INFN Catania 30 June - 4 July, 2008.
INFSO-RI Enabling Grids for E-sciencE Computing Element & Torque Server Installation Giuseppe La Rocca INFN Catania - Italy First.
User Interface UI TP: UI User Interface installation & configuration.
Presentation of the results khiat abdelhamid
GLite WN Installation Giuseppe LA ROCCA INFN Catania ACGRID-II School 2-14 November 2009 Kuala Lumpur - Malaysia.
RI EGI-TF 2010, Tutorial Managing an EGEE/EGI Virtual Organisation (VO) with EDGES bridged Desktop Resources Tutorial Robert Lovas, MTA SZTAKI.
Consorzio COMETA - Progetto PI2S2 UNIONE EUROPEA Grid2Win : gLite for Microsoft Windows Elisa Ingrà - INFN.
First South Africa Grid Training WORKER NODE Albert van Eck University of the Free State 25 July, 2008.
First South Africa Grid Training Installation and configuration of BDII Gianni M. Ricciardi Consorzio COMETA First South Africa Grid Training Catania,
SEE-GRID-SCI MON Hands-on Session Vladimir Slavnić Institute of Physics Belgrade Serbia The SEE-GRID-SCI initiative.
INFN/IGI contributions Federated Clouds Task Force F2F meeting November 24, 2011, Amsterdam.
Site BDII and CE Installation Muhammad Farhan Sjaugi, UPM 2009 November , UM Malaysia 1.
WORKER NODE Alfonso Pardo EPIKH School, System Admin Tutorial Beijing, 2010 August 30th – 2010 September 3th.
INFSO-RI Enabling Grids for E-sciencE Worker Node installation & configuration Giuseppe Platania INFN Catania EMBRACE Tutorial Clermont-Ferrand,
South African Grid Training WORKER NODE Albert van Eck UFS - ICTS 17 November 2009 Slides by GIUSEPPE PLATANIA.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America Worker Node & Torque Client Installation.
Regional SEE-GRID-SCI Training for Site Administrators
(Exchange Programme to advance e-Infrastructure Know-How)
lcg-infosites documentation (v2.1, LCG2.3.1) 10/03/05
CE+WN+siteBDII Installation and configuration
Installation and configuration of a Computing Element
Workload Management System
Grid Services Ouafa Bentaleb CERIST, Algeria
CMS report from FNAL demo week Marco Verlato (INFN-Padova)
gLite Job Management Christos Theodosiou
Grid Management Challenge - M. Jouvin
Presentation transcript:

CE: compute element TP: CE & WN Compute Element Worker Node Installation configuration

CE presentation The Computing Element is the central service of a site. Its main functionally are: – manage the jobs (job submission, job control) – update to WMS the status of the jobs – publish all site informations (about site, queue, number of total,free CPUs) It can run several kinds of batch system: –Torque + MAUI – LSF – Condor

TORQUE server presentation The Torque server is composed by a: – pbs_server – pbs_server which provides the basic batch services such as receiving/creating a batch job. The Torque client is composed by a: – pbs_mom – pbs_mom which places the job into execution. It is also responsible for returning the job’s output to the user The MAUI system is composed by a: – job_scheduler – job_scheduler which contains the site's policy to decide which job must be executed.

CE: site-info.def variables (1) Main variables of the site configuration file for the CE : CE_HOST=ce1.$MY_DOMAIN # Jobmanager specific settings JOB_MANAGER=lcgpbs CE_BATCH_SYS=torque BATCH_BIN_DIR=/usr/bin BATCH_VERSION=torque-1.0.1b BATCH_LOG_DIR=/var/spool/pbs/server_priv/accounting # Architecture and enviroment specific settings CE_CPU_MODEL=PIV CE_CPU_VENDOR=intel CE_CPU_SPEED=1001 CE_OS="Scientific Linux SL" CE_OS_RELEASE="SL" CE_OS_VERSION=3.0.5 CE_MINPHYSMEM=1024

CE : site-info.def variables (2) CE_MINVIRTMEM=2048 CE_SMPSIZE=1 CE_SI00=381 CE_SF00=0 CE_OUTBOUNDIP=TRUE CE_INBOUNDIP=FALSE CE_RUNTIMEENV=" LCG-2 LCG-2_1_0 … GLITE-3_0_0 R-GMA " # TORQUE - Change this if your torque server is not on the CE TORQUE_SERVER=$CE_HOST Worker Node list defined for the site “private.griprototype” : WN_LIST=/opt/glite/yaim/travail/wn-list.conf ce1.private.gridprototype se1.private.gridprototype

WN: worker node & Torque client presentation  The Torque client is composed by a:  pbs_mom  pbs_mom which places the job into execution. It is also responsible for returning the job’s output to the user The Worker Node is a service where the jobs run. Its main functionally are: execute the jobs update to Computing Element the status of the jobs It can run several kinds of client batch system: Torque LSF

CE certification: cd /etc/grid-security/ ln -s ce1.private.gridprototype.crt hostcert.pem ln -s ce1.private.gridprototype.key hostkey.pem chmod 644 hostcert.pem chmod 400 hostkey.pem For the CE1 machine, certificates are files named : ce1.private.gridprototype.crt ce1.private.gridprototype.key Certificates installation in /etc/grid-security directory on CE Get certificates from the BEINGRID CA Certification Authority: backup the certificate as a.p12 file and extract public and private keys openssl pkcs12 –nocert –in ce1.p12 –out ce1….cert openssl pkcs12 –nocert –in ce1.p12 –out ce1….key

List of mandatory configuration files : the WN list defined for the site “private.griprototype” : WN_LIST=/opt/glite/yaim/travail/wn-list.conf the mapped-users list defined for the site “private.griprototype” : /opt/glite/yaim/travail/users.conf the mapped-groups list defined for the site “private.griprototype” : /opt/glite/yaim/travail/groups.conf

CE installation and configuration gLite-yaim generic command: install_node site-info.def lcg-CE_torque glite-WN The CE is a certified machine, install certificates in the directory /etc/grid-security/ configure_node site-info.def CE_torque WN_torque BDII_site

CE publication test The CE should publish information to the BDII: lcg-infosites --vo egeode ce valor del bdii: rb1.private.gridprototype:2170 #CPU Free Total Jobs Running Waiting ComputingElement ce1.private.gridprototype:2119/jobmanager-lcgpbs-egeode The CE should publish status of jobs queues: As egeode005 user locally, it should match the WN list defined in /opt/glite/…/wn-list.conf pbsnodes -a se1.private.gridprototypece1.private.gridprototype state = free np = 1 properties = lcgpro ntype = cluster etc…

Local job submission on the CE To be able to submit jobs locally the user must be mapped egeode005 user on the new installed CE machine. cat test.sh #!/bin/sh /bin/hostname /bin/sleep 300 qsub -q egeode test.sh 35.ce1.private.gridprototype qstat -a ce1.private.gridprototype: Req'd Req'd Elap Job ID Username Queue Jobname SessID NDS TSK Memory Time ce1.private. egeode00 egeode test.sh :00 R

UI/GUI JAVA graphical interface commands : edj-wl-ui-jobmonitor.sh edj-wl-ui-jdleditor.sh …

CE Torque/Maui documentation  TORQUE ADMIN GUIDE  MAUI ADMIN GUIDE

Sample Image Questions on the CE ?