LOFAR - Calibration, Analysis and Modelling of Radio-Astronomy Data EGI Conference 2015 19 May 2015, Lisbon Daniele Lezzi – Barcelona Supercomputing.

Slides:



Advertisements
Similar presentations
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 6 2/13/2015.
Advertisements

Susana Sánchez Expósito Instituto de Astrofísica de Andalucía - CSIC Pablo Martin, Jose Enrique Ruiz, Lourdes Verdes-Montenegro, Julian Garrido, Raül Sirvent,
Customized cloud platform for computing on your terms !
Building service testbeds on FIRE D5.2.5 Virtual Cluster on Federated Cloud Demonstration Kit August 2012 Version 1.0 Copyright © 2012 CESGA. All rights.
Ceph Storage in OpenStack Part 2 openstack-ch,
European Grid Initiative Federated Cloud update Peter solagna Pre-GDB Workshop 10/11/
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
ArcGIS Server for Administrators
ServiceSs, a new programming model for the Cloud Daniele Lezzi, Rosa M. Badia, Jorge Ejarque, Raul Sirvent, Enric Tejedor Grid Computing and Clusters Group.
Features Of SQL Server 2000: 1. Internet Integration: SQL Server 2000 works with other products to form a stable and secure data store for internet and.
European Grid Initiative Data Services and Solutions Part 2: Data in the cloud Enol Fernández Data Services.
Windows Azure poDRw_Xi3Aw.
ALL INFORMATION PRESENTED AS WELL AS ALL SESSIONS ARE MICROSOFT CONFIDENTIAL AND UNDER YOUR NON-DISCLOSURE AGREEMENT (NDA) AND\OR TECHNOLOGY PREVIEW.
EGI Technical Forum Madrid COMPSs in the EGI Federated Cloud Daniele Lezzi – BSC EGI Technical Forum Madrid.
ETICS An Environment for Distributed Software Development in Aerospace Applications SpaceTransfer09 Hannover Messe, April 2009.
The EUBrazilOpenBio-BioVeL Use Case in EGI Daniele Lezzi, Barcelona Supercomputing Center EGI-TF September 2013.
EGI-InSPIRE RI EGI Webinar EGI-InSPIRE RI Porting your application to the EGI Federated Cloud 17 Feb
EGI-Engage is co-funded by the Horizon 2020 Framework Programme of the European Union under grant number Marios Chatziangelou, et al.
Cloud interoperability and elasticity with COMPSs Federated Cloud F2F Jan , Amsterdam Daniele Lezzi – Barcelona Supercomputing Center.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
StratusLab is co-funded by the European Community’s Seventh Framework Programme (Capacities) Grant Agreement INFSO-RI Demonstration StratusLab First.
Instituto de Biocomputación y Física de Sistemas Complejos Cloud resources and BIFI activities in JRA2 Reunión JRU Española.
WP5 – Infrastructure Operations Test and Production Infrastructures StratusLab kick-off meeting June 2010, Orsay, France GRNET.
Hecuba and integration with COMPSs Federated Cloud Big Data Daniele Lezzi, Guillem Alomar – Barcelona Supercomputing Center.
EGI Technical Forum Madrid The EUBrazilOpenBio-BioVeL Use Case in EGI Daniele Lezzi – BSC EGI Technical Forum Madrid.
Federated Cloud: Computing UPVLC-I3M Effort allocated: 6 pms. Proposal: Integration of an Infrastructure Broker with self-configuration and auto-scaling.
Daniele Lezzi Execution of scientific workflows on federated multi-cloud infrastructures IBERGrid Madrid, 20 September 2013.
EGI-InSPIRE RI EGI Compute and Data Services for Open Access in H2020 Tiziana Ferrari Technical Director, EGI.eu
EGI-InSPIRE RI EGI-InSPIRE RI EGI-InSPIRE Software provisioning and HTC Solution Peter Solagna Senior Operations Manager.
EGI-Engage is co-funded by the Horizon 2020 Framework Programme of the European Union under grant number Federated Cloud Update.
CLOUD COMPUTING Presented to Graduate Students Mechanical Engineering Dr. John P. Abraham Professor, Computer Engineering UTPA.
EGI-InSPIRE RI An Introduction to European Grid Infrastructure (EGI) March An Introduction to the European Grid Infrastructure.
Support to user communities in EGI with COMPSs Federated Cloud F2F Jan , Amsterdam Daniele Lezzi – Barcelona Supercomputing Center.
1 EGI Federated Cloud Architecture Matteo Turilli Senior Research Associate, OeRC, University of Oxford Chair – EGI Federated Clouds Task Force
EGI… …is a Federation of over 300 computing and data centres spread across 56 countries in Europe and worldwide …delivers advanced computing.
The EGI Federated Cloud
PaaS services for Computing and Storage
Ecological Niche Modelling in the EGI Cloud Federation
Use of HLT farm and Clouds in ALICE
Open OnDemand: Open Source General Purpose HPC Portal
StratusLab First Periodic Review
Introduction to Distributed Platforms
From VPH-Share to PL-Grid: Atmosphere as an Advanced Frontend
Tools and Services Workshop
Flowbster: Dynamic creation of data pipelines in clouds
Working With Azure Batch AI
Prepared by: Assistant prof. Aslamzai
Federated Cloud Computing
FedCloud Blueprint Update
Introduction to Data Management in EGI
Platform as a Service.
Logo here Module 3 Microsoft Azure Web App. Logo here Module Overview Introduction to App Service Overview of Web Apps Hosting Web Applications in Azure.
Tools and Services Workshop Overview of Atmosphere
Building Applications with Windows Azure and SQL Azure
Exam : Implementing Microsoft Azure Infrastructure Solutions
An easier path? Customizing a “Global Solution”
AWS COURSE DEMO BY PROFESSIONAL-GURU. Amazon History Ladder & Offering.
OpenNebula Offers an Enterprise-Ready, Fully Open Management Solution for Private and Public Clouds – Try It Easily with an Azure Marketplace Sandbox MICROSOFT.
Dr. John P. Abraham Professor, Computer Engineering UTPA
Dev Test on Windows Azure Solution in a Box
The EGI Federated Cloud
Case Study: Algae Bloom in a Water Reservoir
Dev and Test Solution reference architecture.
Module 01 ETICS Overview ETICS Online Tutorials
ELIXIR Competence Center
MDC-B203 Deploying Applications in Microsoft System Center Virtual Machine Manager Using Services John Messec Program Manager Microsoft.
Day 2, Session 2 Connecting System Center to the Public Cloud
PerformanceBridge Application Suite and Practice 2.0 IT Specifications
Microsoft Virtual Academy
Presentation transcript:

LOFAR - Calibration, Analysis and Modelling of Radio-Astronomy Data EGI Conference May 2015, Lisbon Daniele Lezzi – Barcelona Supercomputing Center Susana Sanchez-Expósito - Instituto de Astrofísica de Andalucía – CSIC José Sabater - University of Edinburgh

The LOw Frequency Array (LOFAR) Credits: ©ASTRON. Radio interferometer composed by thousands of antennas distributed among several countries Application: Astronomy Galaxy CIG292. Image from SDSS Calibration pipeline use case: Pipeline for transforming the raw data from the antennas to science-ready data Existing collaboration in the AMIGA4GAS project This pipeline would be use by tens of users Calibration pipeline use case: Pipeline for transforming the raw data from the antennas to science-ready data Existing collaboration in the AMIGA4GAS project This pipeline would be use by tens of users

Why to use Fedcloud Frequency Sub-band 1 Sub-band 2 Sub-band n X GB DATA PARALLELI SM Needs to process large datasets (several TBs) Datasets can be divided in pieces that can be processed independently. COMPSs useful for automatic scaling and elasticity features. Up to 488 independent tasks working on dataset pieces of several GBs Further details in poster titled: EGI Federated Cloud for calibrating and analysing Radio-Astronomy data

Development use case status Pipeline modelled in COMPSs, that provides a Python binding and takes care of the parallel execution of the tasks. Virtual Appliance customized with the LOFAR software and deployed through the AppDB. import subprocess import sys import os from pycompss.api.task import task from pycompss.api.parameter import script_name = FILE) def iter_calib(script_name): os.chmod(script_name,0744) subprocess.call(script_name) print "end executiong" if __name__ == "__main__": args = sys.argv[1:] DATA_PATH=args[0] TEMPLATE_FILE=args[1] f=open(TEMPLATE_FILE,'r') content=f.read() f.close() list_f=os.listdir(DATA_PATH) for directory in list_f: # Iterate over the data inputs if os.path.isdir(DATA_PATH+"/"+directory): new_content=content.replace("INPUTDATAPATH",directory) script_name="job"+directory+".sh" f=open(script_name,"w") f.write(new_content) f.close() iter_calib(script_name) Integration with the Fed Cloud through COMPSs for VM management (using rOCCI). The data is uploaded on a NFS storage (on each site) and shared amongst VMs. No object storage used.

User Experience/Wish list The AppDB Virtual Appliances facilitate the installation of the user application on Fedcloud. The users can directly install their software on a running Fedcloud VM, which has been previously started as an instance of a Virtual Appliance. The process of creation and deployment of the Virtual Appliances in the AppDB is not automatic. There should be a way to let users create snapshots. New software releases imply starting again the deployment process. Last year LOFAR software team published 8 releases. The computing power provided by the different sites federated in Fedcloud allows to configure virtual machines with different capabilities (memory+cpu) in order to match with the specific needs of the use case. Lack of consolidated solutions for federating storage. The user data are too large to be stored in the VM images. They should be stored in volumes easily mountable from several VMs and synchronized across different sites. The COMPSs framework facilitates the porting and deployment (through the PMES service) of the application. COMPSs allows the users to program their applications sequentially, in a transparent way by taking care of their parallelization as well as of the VMs orchestration for their execution. User software deployment Scalability / performance User interaction Friendly tools for working as interface for the command line client would facilitate the first steps in using Fedcloud: to configure the VMs and install the software, manage the proxy certificates and propagate the new releases of Virtual Appliances to the Cloud Marketplace. Blocker!!

Future plans Continue and refine the implementation of the use case. More scalability and elasticity tests with multi-providers runs. Integration with a better data management solution. Evaluate PMES as end users interface. Integrate the application in the Astrophysics Science Gateway (see demo #2)

Thank you!