Installing Galaxy on a cluster :

Slides:



Advertisements
Similar presentations
Information Technology Center Introduction to High Performance Computing at KFUPM.
Advertisements

Network+ Guide to Networks, Fourth Edition Chapter 10 Netware-Based Networking.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
EduShib VA What is EduShib VA? EduShib VA (Virtual Appliance) is a image based implementation tool for eduroam and Shibboleth.
Installing and running COMSOL on a Windows HPCS2008(R2) cluster
Web Proxy Server Anagh Pathak Jesus Cervantes Henry Tjhen Luis Luna.
Testing PanDA at ORNL Danila Oleynik University of Texas at Arlington / JINR PanDA UTA 3-4 of September 2013.
Paper on Best implemented scientific concept for E-Governance Virtual Machine By Nitin V. Choudhari, DIO,NIC,Akola By Nitin V. Choudhari, DIO,NIC,Akola.
6/1/2001 Supplementing Aleph Reports Using The Crystal Reports Web Component Server Presented by Bob Gerrity Head.
TechEd /22/2017 5:40 AM © 2013 Microsoft Corporation. All rights reserved. Microsoft, Windows, and other product names are or may be registered trademarks.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
VIPBG LINUX CLUSTER By Helen Wang March 29th, 2013.
Module 10: Monitoring ISA Server Overview Monitoring Overview Configuring Alerts Configuring Session Monitoring Configuring Logging Configuring.
Module 1: Installing and Configuring Servers. Module Overview Installing Windows Server 2008 Managing Server Roles and Features Overview of the Server.
Module 9: Implementing Caching. Overview Caching Overview Configuring General Cache Properties Configuring Cache Rules Configuring Content Download Jobs.
Module 6: Managing Client Access. Overview Implementing Client Access Servers Implementing Client Access Features Implementing Outlook Web Access Introduction.
6/1/2001 Supplementing Aleph Reports Using The Crystal Reports Web Component Server Presented by Bob Gerrity Head.
Weekly Report By: Devin Trejo Week of June 14, 2015-> June 20, 2015.
Implementing ISA Server Caching
CITA 310 Section 4 Apache Configuration (Selected Topics from Textbook Chapter 6)
Galaxy in Production Nate Coraor Galaxy Team Penn State University.
Core and Framework DIRAC Workshop October Marseille.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
Nate Anderson So, You’ve Inherited an OnBase System.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Advanced Computing Facility Introduction
BY: SALMAN 1.
Workstations & Thin Clients
ETL Validator Deployment Options
Architecture Review 10/11/2004
Parrot and ATLAS Connect
Configure the intercom IP
Welcome to Indiana University Clusters
Open OnDemand: Open Source General Purpose HPC Portal
Assumptions What are the prerequisites? … The hands on portion of the workshop will be on the command-line. If you are not familiar with the command.
BY: SALMAN.
Welcome to Indiana University Clusters
Deploying Galaxy in a secure environment to analyse sensitive data
Configuring Attendant Console
Module Overview Installing and Configuring a Network Policy Server
Heterogeneous Computation Team HybriLIT
Getting Started with R.
MCTS Guide to Microsoft Windows 7
VI-SEEM Data Discovery Service
ASU Saguaro 09/16/2016 Jung Hyun Kim.
Hodor HPC Cluster LON MNG HPN Head Node Comp Node Comp Node Comp Node
R.Mashinistov (UTA) July
Processes The most important processes used in Web-based systems and their internal organization.
Mattias Wadenstein Hepix 2012 Spring Meeting , Prague
Integration of Singularity With Makeflow
Short Read Sequencing Analysis Workshop
Unit 27: Network Operating Systems
Helix - HPC/SLURM Tutorial
CCR Advanced Seminar: Running CPLEX Computations on the ISE Cluster
HC Hyper-V Module GUI Portal VPS Templates Web Console
Advanced Computing Facility Introduction
User Guide Subversion client TortoiseSVN
Moodle Scalability What is Scalability?
Footer.
SiteBuilder 2 Introduction.
High Performance Computing in Bioinformatics
The Neuronix HPC Cluster:
Jerald Overstreet, GISP Server Portal SQL Manager Admin
Creating and Managing Folders
HC VMware Module
H2020 EU PROJECT | Topic SC1-DTH | GA:
Plug-In Architecture Pattern
Short Read Sequencing Analysis Workshop
SharePoint 2013 Best Practices
Presentation transcript:

Installing Galaxy on a cluster : issues around the DB server, queue system, external authentication, etc. Nikolay Vazov University Center for Information Technologies (USIT) University of Oslo Norway

Swiss Galaxy Workshop, Wednesday, October 3rd, Bern The new UiO hpc-cluster: Abel http://www.uio.no/english/services/it/research/hpc/abel/index.html (in operation since October 1st 2012) Some facts about the Abel cluster: Ranked 96 in the world top 500 list with 178.6 TFlops 178.6 TFlops correspond roughly to ca 2700 PCs the fastest i Noway, 3rd in Scandinavia Ranked 68 in the Green500 list 652 compute nodes and 20 administration nodes All compute nodes on the cluster have a minimum 64 GB RAM, 16 physical CPU cores and are connected by FDR (56 Gbps) Infiniband. 10.432 cores used for computing: correspond to 2.600 quad-core PCs 400 TB shared disk Compute nodes with 350 TB local discs Compute nodes have a total of 48 TB RAM Power consumption 230KW (full load) Trivia: all the nodes were mounted in 14 hours (appr. 1'15” per node!) Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

The existing service – the Bioportal Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Bioportal features - jobs Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Bioportal features - files Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Galaxy in Abel - configuration cluster External authentication (FEIDE) Locally registered users node node node node node node Interface between Galaxy and SLURM – DRMAA job scheduler - SLURM Apache proxy PostgreSQL DB server Located on a different host Paster (WSGI) SSL connection Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Galaxy in Abel - configuration cluster External authentication (FEIDE) Locally registered users node node node node node node Interface between Galaxy and SLURM – DRMAA job scheduler - SLURM Apache proxy PostgreSQL DB server Located on a different host Paster (WSGI) SSL connection Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Job scheduling with Galaxy Galaxy – specifies the job runners DRMAA library - generic interface to various scheduling systems SLURM – schedules the jobs ( client/server) Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Job scheduling Galaxy -> DRMAA -> SLURM Galaxy server is outside the cluster. We prefer this situation to the Galaxy server being a part of the cluster. Galaxy, DRMAA and SLURM are located on an nfs mounted partition. Galaxy: universe_wsgi.ini # -- Job Execution # Comma-separated list of job runners to start. local is always started. If # ... The runners currently available are 'pbs' and 'drmaa'. start_job_runners = drmaa # The URL for the default runner to use when a tool doesn't explicitly define a # runner below. default_cluster_job_runner = drmaa:/// Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Job scheduling Galaxy -> DRMAA -> SLURM export DRMAA_PATH=/kevlar/proje cts/drmaa/lib/libdrmaa.so.1.0.2 export SLURM_DRMAA_CONF=/et c/slurm_drmaa.conf hpc-dev01 etc# cat slurm_drmaa.conf Job_categories: { default: "-A staff -p normal --mem-per- cpu=1000 -- comment=hello", } Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Job scheduling Galaxy -> DRMAA -> SLURM Plus a couple of changes (http://mdahlo.blogspot.no/2011/06/galaxy-on-uppmax.html) in the DRMAA egg (drmaa-0.4b3-py2.6.egg) Find munge Display the web form to specify node, cores, memory, partition, etc. Parse the data from the web form and set up a string into <path-to-galaxy>/database/pbs/slurm_settings.tmp Create a real sbatch file, add missing parameters, module load, etc, and send the job to the cluster Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Swiss Galaxy Workshop, Wednesday, October 3rd, Bern Job scheduling Galaxy -> DRMAA -> SLURM (thanks to Katerina Michalickova) SLURM (client has to be installed on the mounted partition) : /etc/slurm/slurm.conf hpc-dev01 slurm# cat slurm.conf ## slurm.conf: main configuration file for SLURM ## $Id: slurm_2.2.conf,v 1.30 2011/09/20 15:13:58 root Exp $ ## FIXME: check GroupUpdate*, TopologyPlugin, ## UnkillableStepProgram, UsePAM ### ### Cluster ClusterName=titan # NOW abel SlurmctldPort=6817 SlurmdPort=6818 TmpFs=/work TreeWidth=5 ## Timers: #default: MessageTimeout=10 ## FIXME: should be reduced when/if we see that slurmd is behaving: #SlurmdTimeout=36000 WaitTime=0 ### Slurmctld ControlMachine=blaster.teflon.uio.no SlurmUser=slurm StateSaveLocation=/tmp Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

SSL to the PostgreSQL server (thanks to Nate Coraor) Downloaded and recompiled an psycopg2-2.0.13 egg In universe_wsgi.ini database_connection = postgres://<dbuser>:<password>@<dbhost>:5432 /<dbname>?sslmode=require Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Authentication (thanks to Roland Hedberg) pysaml2-0.4.0/ Modify lib/galaxy/web/controllers/user.py Authentication working, but can not capture the POST from the IdP any help is appreciated :) Swiss Galaxy Workshop, Wednesday, October 3rd, Bern

Swiss Galaxy Workshop, Wednesday, October 3rd, Bern Thank you http://www.uio.no/english/services/it/research/ hpc/abel/index.html n.a.vazov@usit.uio.no Swiss Galaxy Workshop, Wednesday, October 3rd, Bern