The 6 th Annual Pangalactic BOINC Workshop. BOINC: The Year in Review David Anderson 31 Aug 2010.

Slides:



Advertisements
Similar presentations
Database Architectures and the Web
Advertisements

BOINC The Year in Review David P. Anderson Space Sciences Laboratory U.C. Berkeley 22 Oct 2009.
NFS. The Sun Network File System (NFS) An implementation and a specification of a software system for accessing remote files across LANs. The implementation.
INTRODUCTION TO CLOUD COMPUTING Cs 595 Lecture 5 2/11/2015.
The 9 th Annual Workshop September 2013 INRIA, Grenoble, France
Chapter 5 Roles and features. objectives Performing management tasks using the Server Manager console Understanding the Windows Server 2008 roles Understanding.
ATIF MEHMOOD MALIK KASHIF SIDDIQUE Improving dependability of Cloud Computing with Fault Tolerance and High Availability.
Volunteer Computing and Hubs David P. Anderson Space Sciences Lab University of California, Berkeley HUBbub September 26, 2013.
SNIA/SSIF KMIP Interoperability Proposal. What is the proposal? Host a KMIP interoperability program which includes: – Publishing a set of interoperability.
Achievements and Opportunities in Volunteer Computing David P. Anderson Space Sciences Lab U.C. Berkeley 18 April 2008.
A Guided Tour of BOINC David P. Anderson Space Sciences Lab University of California, Berkeley TACC November 8, 2013.
M i SMob i S Mob i Store - Mobile i nternet File Storage Platform Chetna Kaur.
Introduction to Cloud Technology StratusLab Tutorial (Orsay, France) 28 November 2012.
Exa-Scale Volunteer Computing David P. Anderson Space Sciences Laboratory U.C. Berkeley.
Introduction to the BOINC software David P. Anderson Space Sciences Laboratory University of California, Berkeley.
Wenjing Wu Computer Center, Institute of High Energy Physics Chinese Academy of Sciences, Beijing BOINC workshop 2013.
Javascript Cog Kit By Zhenhua Guo. Grid Applications Currently, most grid related applications are written as separate software. –server side: Globus,
Volunteer Computing with GPUs David P. Anderson Space Sciences Laboratory U.C. Berkeley.
and Citizen Cyber-Science David P. Anderson Space Sciences Laboratory U.C. Berkeley.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
BOINC: Progress and Plans David P. Anderson Space Sciences Lab University of California, Berkeley BOINC:FAST August 2013.
Evolving Interfaces to Impacting Technology: The Mobile TeraGrid User Portal Rion Dooley, Stephen Mock, Maytal Dahan, Praveen Nuthulapati, Patrick Hurley.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
TEMPLATE DESIGN © BOINC: Middleware for Volunteer Computing David P. Anderson Space Sciences Laboratory University of.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
TOPIC 7.0 LINUX SERVICES AND CONFIGURATION. ROOT USER Root user is called “super user” because it has power far beyond those of mortal user. As root,
BOINC: An Open Platform for Public-Resource Computing David P. Anderson Space Sciences Laboratory U.C. Berkeley.
David P. Anderson Space Sciences Laboratory University of California – Berkeley Public Distributed Computing with BOINC.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
CernVM and Volunteer Computing Ivan D Reid Brunel University London Laurence Field CERN.
Volunteer Computing and BOINC Dr. David P. Anderson University of California, Berkeley Dec 3, 2010.
Frontiers of Volunteer Computing David Anderson Space Sciences Lab UC Berkeley 30 Dec
The Future of Volunteer Computing David P. Anderson U.C. Berkeley Space Sciences Lab UH CS Dept. March 22, 2007.
Emulating Volunteer Computing Scheduling Policies Dr. David P. Anderson University of California, Berkeley May 20, 2011.
Volunteer Computing: Involving the World in Science David P. Anderson U.C. Berkeley Space Sciences Lab February 16, 2007.
Volunteer Computing: the Ultimate Cloud Dr. David P. Anderson University of California, Berkeley Oct 19, 2010.
The Limits of Volunteer Computing Dr. David P. Anderson University of California, Berkeley March 20, 2011.
All the computers in the world (~1 billion) BOINC: high-level goal Computational science biology, medicine Earth sciences, physics, astronomy, math, A.I.,...
Using volunteered resources for data-intensive computing and storage David Anderson Space Sciences Lab UC Berkeley 10 April 2012.
Volunteer Computing with BOINC: a Tutorial David P. Anderson Space Sciences Laboratory University of California – Berkeley May 16, 2006.
Frontiers of Volunteer Computing David Anderson Space Sciences Lab UC Berkeley 28 Nov
BOINC current work David Anderson 11 July Where we're at ● We've come a long way – Some successful projects – Progress on software ● The long road.
Web and Proxy Server.
Lecture 6: Cloud Computing
Volunteer Computing and BOINC
A new model for volunteer computing
Introduction to Cloud Technology
The Future of Volunteer Computing
Modularity Most useful abstractions an OS wants to offer can’t be directly realized by hardware Modularity is one technique the OS uses to provide better.
Improving Security Through Gunpowder Detection
By: Raza Usmani SaaS, PaaS & TaaS By: Raza Usmani
Volunteer Computing: SETI and Beyond David P
Volunteer Computing for Science Gateways
Affinity Depending on the application and client requirements of your Network Load Balancing cluster, you can be required to select an Affinity setting.
Designing a Runtime System for Volunteer Computing David P
Web Development Web Servers.
A Roadmap for Volunteer Computing in the U.S.
Exa-Scale Volunteer Computing
Securing the Network Perimeter with ISA 2004
Job Scheduling in a Grid Computing Environment
David Cameron ATLAS Site Jamboree, 20 Jan 2017
Database Architectures and the Web
The Application Lifecycle
Lecture 6: TCP/IP Networking 1nd semester By: Adal ALashban.
Cloud Web Filtering Platform
Technical Capabilities
University of California, Berkeley
Journey to the Cloud – Guidance and Lessons Learned
LO2 – Understand Computer Software
Exploring Multi-Core on
Presentation transcript:

The 6 th Annual Pangalactic BOINC Workshop

BOINC: The Year in Review David Anderson 31 Aug 2010

Credit: goals Device neutrality: a job should get the same credit no matter what device processes Project neutrality: a computer should get the same credit/day regardless of what project(s) it runs (easy to show that these can’t both be achieved)

1 st credit system ● CPU time x CPU benchmark ● not device neutral ● Replication and credit averaging ● granted credit depends on partner

2 nd credit system “Actual FLOPS”-based APIs for reporting FLOP counts publishes average credit/CPU sec, other projects scale to match Problems: most apps can’t count FLOPs doesn’t address GPUs no device neutrality doesn’t prevent cheating w/ single replication

Philosophy of 3 rd system Credits is based on peak FLOP count (PFC) PFC(J) = #CPUs * CPU benchmark + #GPUs * GPU rated FLOPS Reflects “opportunity cost”, not actual work Normalize in 2 ways

Statistics Maintain mean, variance of PFC(J) / WU.fpops_est for each: app app version (host, app version)

Normalize to most efficient app version CPU/Win mean PFC CPU/Mac CPU/Linux GPU/Win Note: this provides device neutrality at the expense of project neutrality

Host normalization Scale PFC for version V, host H by V.pfc_avg / H.pfc_avg Provides cheat-resistance even with single replication but need to prevent cherry-picking: don’t use host normalization unless host has returned N consecutive valid results

GPU-only projects On a project with both CPU and GPU versions, version normalization provides a measure of relative efficiency CPU vs. GPU Projects with only GPU apps don’t have this Solution: such projects scale by the weighted averages of projects that do

Experience ● New system tested in ● Works, but need to double credit (redefine Cobblestone) ● No project customization

Job runtime estimation Old system: R(est) = WU.fpops_est / CPU benchmark Maintain and scale by a project-wide “duration correction factor” Problems: bad if multiple versions scientists shouldn’t think about FLOPS doesn’t work for GPUs

New system ● Maintain mean, variance of normalized elapsed time for each (host, app version) ● Predicted runtime = mean * WU.fpops_est (per-app-version duration correction factor)

Other per-(host, app version) items ● Daily quota (for host punishment) ● Consecutive valid results: replaces error rate for ● “reliable” mechanism ● cherry-picking prevention

Notice system ● How does the BOINC client software communicate with volunteers? Currently: the Messages Tab. Problems: ● Requires user to look ● Non-prescriptive techno-babble ● Only bad news ● Only text ● Non-translatable

Notices architecture ● Multiple “notice” sources ● from client ● from schedulers ● RSS feeds from projects – project news – private messages – friend requests – messages in subscribed threads –...

Notice delivery ● System tray popup ● Notices tab

GPU support ● Exclusive apps ● Show GPU projects in attach wizard ● Snooze/suspend/resume GPU ● app_plan(): specify GPU RAM requirements ● use in scheduling; boinc_temporary_exit() ● Sample CUDA/OpenCL apps ● Support Fermi GPUs

Multithread app support ● boinc_init_parallel() ● suspend/resume multiple threads ● show projects in attach wizard

Other goodies ● GUI RPC as HTTP ● enable GUIs based on web technologies ● Web: project news as a message board ● easier to post ● users can discuss ● Preferences ● Transfer at most X MB every N days ● suspend if non-BOINC CPU load exceeds X

More goodies ● Stuff for Intel PtP ● web-based registration (manager finds cookie) ● HTTP proxy autodetect ● Server logging ● flags instead of -d 3 ● -d 4 means print DB queries

Upcoming ● Rewrite or replacement of Simple View ● or entire Manager? ● VM app support ● BOINC installer includes VirtualBox? ● Volpex ● IPC for BOINC apps ● virtual cluster ● Integration with Drupal

What we didn’t do ● Integrate remote job submission system from GPUGRID ● Accelerated batch completion

Adoption by scientists ● Single-scientist projects: a dead end ● Barriers to entry are too high ● Wrong marketing model ● Doesn’t handle sporadic requirements

Adoption by scientists ● Most scientists outsource HPC decisions to IT people ● IT people fear and loathe volunteer computing Napoleon: Volunteer computing just can’t handle the kinds of jobs that real scientists run. Me: What precisely is different about these jobs? Napoleon: THEY’RE JUST DIFFERENT, THAT’S ALL

A way forward Distinguish: ● Project operation ● operate servers ● port apps, interface with scientists ● Marketing ● branding/strategy ● mass media, online, non-traditional ● web development ● make bundling deals with computer/OS vendors

Project == existing HPC provider ● Supercomputer centers ● National grids (Teragrid, OSG) ● Hubs ● “Facebook + iPhone app store” for science area ● e.g. Nanohub ● HUBzero/BOINC integration proposal

ScienceUSA.org ● A consortium of funding agencies and HPC providers ● Unified brand, web site for scientific volunteer computing in U.S.; implemented using account manager mechanism ● Volunteers choose research areas, not projects ● Committee of consortium members allocates computing power among projects

● How to realize this? ● European/Asian counterparts?

Summary ● Volunteer computing has not approached its potential ● There are still many skeptics ● Let’s keep working