Implementation of a small-scale desktop grid computing infrastructure in a commercial domain    

Slides:



Advertisements
Similar presentations
Ivan Pleština Amazon Simple Storage Service (S3) Amazon Elastic Block Storage (EBS) Amazon Elastic Compute Cloud (EC2)
Advertisements

Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.
IBM Energy & Environment © 2008 IBM Corporation Energy Efficiency in the Data Centre … and beyond Peter Richardson UK Green Marketing Leader.
BOINC The Year in Review David P. Anderson Space Sciences Laboratory U.C. Berkeley 22 Oct 2009.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 8 Introduction to Printers in a Windows Server 2008 Network.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 11 Managing and Monitoring a Windows Server 2008 Network.
New Challenges in Cloud Datacenter Monitoring and Management
CHAPTER OVERVIEW SECTION 5.1 – MIS INFRASTRUCTURE
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
QTIP Version 0.2 4th August 2015.
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
Volunteer Computing and Hubs David P. Anderson Space Sciences Lab University of California, Berkeley HUBbub September 26, 2013.
CHAPTER FIVE INFRASTRUCTURES: SUSTAINABLE TECHNOLOGIES
Mantychore Oct 2010 WP 7 Andrew Mackarel. Agenda 1. Scope of the WP 2. Mm distribution 3. The WP plan 4. Objectives 5. Deliverables 6. Deadlines 7. Partners.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
A Distributed Computing System Based on BOINC September - CHEP 2004 Pedro Andrade António Amorim Jaime Villate.
November , 2009SERVICE COMPUTATION 2009 Analysis of Energy Efficiency in Clouds H. AbdelSalamK. Maly R. MukkamalaM. Zubair Department.
Planning and Designing Server Virtualisation.
Liam Newcombe BCS Data Centre Specialist Group Secretary Modelling Data Centre Energy Efficiency and Cost.
Wenjing Wu Computer Center, Institute of High Energy Physics Chinese Academy of Sciences, Beijing BOINC workshop 2013.
07:44:46Service Oriented Cyberinfrastructure Lab, Introduction to BOINC By: Andrew J Younge
The Scalable Virtual Network Senior Design III – Spring Quarter Eric Ridder Sean Stroh.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
What is the cloud ? IT as a service Cloud allows access to services without user technical knowledge or control of supporting infrastructure Best described.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
IP-Over-USB Gateway Ben Greenberg Bartosz Mach Adviser: Prof. Vincenzo Liberatore Case Western Reserve University Dept. of Electrical Engineering and Computer.
CEOS WGISS-21 CNES GRID related R&D activities Anne JEAN-ANTOINE PICCOLO CEOS WGISS-21 – Budapest – 2006, 8-12 May.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
June 30 - July 2, 2009AIMS 2009 Towards Energy Efficient Change Management in A Cloud Computing Environment: A Pro-Active Approach H. AbdelSalamK. Maly.
Virtualization Supplemental Material beyond the textbook.
GridChem Developers Conference Focus For Final Year Sudhakar Pamidighantam NCSA 25 August 2006.
OpenNebula: Experience at SZTAKI Peter Kacsuk, Sandor Acs, Mark Gergely, Jozsef Kovacs MTA SZTAKI EGI CF Helsinki.
Scheduling a 100,000 Core Supercomputer for Maximum Utilization and Capability September 2010 Phil Andrews Patricia Kovatch Victor Hazlewood Troy Baer.
HPHC - PERFORMANCE TESTING Dec 15, 2015 Natarajan Mahalingam.
SCI-BUS project Pre-kick-off meeting University of Westminster Centre for Parallel Computing Tamas Kiss, Stephen Winter, Gabor.
Introduction to Mobile-Cloud Computing. What is Mobile Cloud Computing? an infrastructure where both the data storage and processing happen outside of.
Hall D Computing Facilities Ian Bird 16 March 2001.
CernVM and Volunteer Computing Ivan D Reid Brunel University London Laurence Field CERN.
Frontiers of Volunteer Computing David Anderson Space Sciences Lab UC Berkeley 30 Dec
The Limits of Volunteer Computing Dr. David P. Anderson University of California, Berkeley March 20, 2011.
An Overview of Volunteer Computing
HPC In The Cloud Case Study: Proteomics Workflow
Volunteer Computing and BOINC
The Future of Volunteer Computing
Status of WLCG FCPPL project
Experience of Lustre at QMUL
BEST CLOUD COMPUTING PLATFORM Skype : mukesh.k.bansal.
Volunteer Computing for Science Gateways
How to connect your DG to EDGeS? Zoltán Farkas, MTA SZTAKI
Building a Virtual Infrastructure
Clinton A Jones Eastern Kentucky University Department of Technology
<Name of Product>Pilot Closeout Meeting <Customer Name>
CHAPTER OVERVIEW SECTION 5.1 – MIS INFRASTRUCTURE
Design and Implementation
New developments for deploying
Migration Strategies – Business Desktop Deployment (BDD) Overview
Overview Introduction VPS Understanding VPS Architecture
Dev Test on Windows Azure Solution in a Box
AWS Cloud Computing Masaki.
Genre1: Condor Grid: CSECCR
Cloud Computing Architecture
The Design Process—Planning
Working With Cloud - 3.
Fundamental Concepts and Models
System Center Operations Manager 2007 – Technical Overview
Performance And Scalability In Oracle9i And SQL Server 2000
5/12/2019 2:57 PM © Microsoft Corporation. All rights reserved.
Exploring Multi-Core on
Presentation transcript:

Implementation of a small-scale desktop grid computing infrastructure in a commercial domain    

Background and use of BOINC in the academic context Overview  Background and use of BOINC in the academic context   Infrastructure constructed  Quantification of the additional heating produced and power consumed  Quantification of cost  Conclusion

Berkeley Open Infrastructure for Network Computing (BOINC) Middleware for volunteer computing Open-source, NSF-funded development Community-maintained Server: used by scientists to make “projects” Client: runs on consumer devices “attach” to projects fetches/runs jobs in background

Relationship between Project, Application, Workunit and Task

Steps undertaken Set up a test project in Oxford and successfully build application Replicate test project infrastructure in commercial partner. Test the sandboxed application with a sample configuration on the test project Set up a workunit submission script in the test project. Set up the external workunit submission server and set up a script on this server that communicates with the workunit submission script in the test project. Link a limited number of pre-determined desktop clients (between 5->50 desktops) to the test project. Conduct two limited tests of two use-cases

Desktop client

Application Structure

Test Infrastructure Machine Operating system Minimum specification [CPU, RAM, Local Storage] Functions Test server Linux 8 CPU, 8GB, 1TB upload service, application service, download service

Environmental Monitoring

Utilisation Tranche Start Date End Date 1 22/5/17 2/6/17 2 25/6/17 29/6/17 3 2/7/17 8/7/17 4 14/7/17 5 17/7/17 18/7/17 6 23/7/17 24/7/17 7 29/8/17 13/9/17

WU Success and Failure

WU Success and failure per host

Environmental Analysis

Energy consumption Considering only WU calculated on the machine and hence only that systems consumption Baseline using period 15/07/18 to 29/08/17 Only consider tranches 4 & 7 Resulting consumption driven by BOINC utilisation 426.61W per WU €0.032 per WU (assuming 35 minute WU length)

Temperatures in exemplar rooms

Room temperature difference

138.93 Tflop overall PoC – Builds computed Date Submitted Date 90% complete Build # # of WU % #WU Success % CPU time Success TFlops Success Failure 30/6/17 3/7/17 113 1248 1.12 31 0.06 4.65 13/7/17 117 4016 71 97 10.29 4.46 8/7/17 125 4125 44 6.59 8.62 24/7/17 25/7/17 140 251 80 0.77 0.16 29/8/17 13/9/17 151 27081 42 78 41.84 61.49 138.93 Tflop overall

Build Completion Rate

Comparison of costs Contracted HPC costs €3856 for 1TFlop year Consider Build 151 as basis for analysis of capability for whole year Start time = 30/08/2017 15:20:00 (1504106454) End Time = 13/09/2017 23:59:00 (1505347140) Capability as reported by BOINC Benchmarking of resources and CPU time provided = 103.32 TFlop To run Build151 on the contracted HPC system has available would cost €398420 If this level of resources was used constantly for a whole year if would provide 2,290TFlop of resource On the contracted HPC system available this would cost €8.83M for a whole year at this capacity But systems in a final BOINC project would only be available from 8pm to 6pm Mon-Fri and 24 hours at weekends Therefore we must scale this by 58%, so you would have a maximum capability of 1335.8TFlop *note that this is not a true comparison due to additional capability of HPC system though as the service is not available without these features the partner would still need to pay for them even if they were unused. Earlier averaged energy calculations mean that a WU costs €0.032 per WU. Therefore to run the 27081 WU run to 13/09/2017 23:59:00 within Build151 cost €866.59 But… with the BOINC system we are already using existing systems to provide the capability and therefore incur no extra system costs other than establishing server capability which has cost €120k through both the Evaluation and PoC phases.

Further Work Implement automated deployment. Set up the production project. Link a larger set of desktop clients to the production project. Further extend the capability of the external workunit submission process, allowing user submission with relevant permissions. Conduct full-scale extended runs at large scale of both current and other relevant usecases. Document the project setup. Consider other possible applications for this BOINC infrastructure 

Conclusions Successfully run 47.5k workunits in a working BOINC infrastructure, Producing business relevant results for users, at a fraction of costs for comparable system, Makes no discernible impact on temperature within room studied (0.65C is within BMS error), Per WU energy cost of €0.032 (426.61W for ~35 minutes), No new infrastructure required, reusing existing computational systems improving ROI, Clear plan of required future activity to deploy in production, Directly scalable capability, With HPC -