An Investigation into Implementations of DNA Sequence Pattern Matching Algorithms Peden Nichols Computer Systems Research April, 2004-2005.

Slides:



Advertisements
Similar presentations
Multiple Processor Systems
Advertisements

Condor use in Department of Computing, Imperial College Stephen M c Gough, David McBride London e-Science Centre.
Altiris for Desktop Management and More! Presented by: ITS (Scott Arnst and Kathleen Conover) January 16, 2004.
SLA-Oriented Resource Provisioning for Cloud Computing
A Dynamic World, what can Grids do for Multi-Core computing? Daniel Goodman, Anne Trefethen and Douglas Creager
IoP HEPP 2004 Birmingham, 7/4/04 David Cameron, University of Glasgow 1 Simulation of Replica Optimisation Strategies for Data.
The Who, What, Why and How of High Performance Computing Applications in the Cloud Soheila Abrishami 1.
Distributed Logging in Java with Constrained Resource Usage Sunil Brown Varghese, Daniel Andresen Dept. of Computing and Information Sciences Kansas State.
Gabrielle Allen*, Thomas Dramlitsch*, Ian Foster †, Nicolas Karonis ‡, Matei Ripeanu #, Ed Seidel*, Brian Toonen † * Max-Planck-Institut für Gravitationsphysik.
Lincoln University Canterbury New Zealand Evaluating the Parallel Performance of a Heterogeneous System Elizabeth Post Hendrik Goosen formerly of Department.
Large Scale File Distribution Troy Raeder & Tanya Peters.
Beowulf Cluster Computing Each Computer in the cluster is equipped with: – Intel Core 2 Duo 6400 Processor(Master: Core 2 Duo 6700) – 2 Gigabytes of DDR.
Cluster Computer For Bioinformatics Applications Nile University, Bioinformatics Group. Hisham Adel 2008.
Distributed Process Management1 Learning Objectives Distributed Scheduling Algorithms Coordinator Elections Orphan Processes.
PRESTON SMITH ROSEN CENTER FOR ADVANCED COMPUTING PURDUE UNIVERSITY A Cost-Benefit Analysis of a Campus Computing Grid Condor Week 2011.
Advanced Metering Infrastructure
University of Oklahoma Genome Center4/14/12.
MULTICOMPUTER 1. MULTICOMPUTER, YANG DIPELAJARI Multiprocessors vs multicomputers Interconnection topologies Switching schemes Communication with messages.
07/14/08. 2 Points Introduction. Cluster and Supercomputers. Cluster Types and Advantages. Our Cluster. Cluster Performance. Cluster Computer for Basic.
HPCC Mid-Morning Break Interactive High Performance Computing Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Sort-Last Parallel Rendering for Viewing Extremely Large Data Sets on Tile Displays Paper by Kenneth Moreland, Brian Wylie, and Constantine Pavlakos Presented.
THE AFFORDABLE SUPERCOMPUTER HARRISON CARRANZA APARICIO CARRANZA JOSE REYES ALAMO CUNY – NEW YORK CITY COLLEGE OF TECHNOLOGY ECC Conference 2015 – June.
VAP What is a Virtual Application ? A virtual application is an application that has been optimized to run on virtual infrastructure. The application software.
Simple Interface for Polite Computing (SIPC) Travis Finch St. Edward’s University Department of Computer Science, School of Natural Sciences Austin, TX.
Introduction to Parallel Programming MapReduce Except where otherwise noted all portions of this work are Copyright (c) 2007 Google and are licensed under.
Department of Computer Science Engineering SRM University
CS492: Special Topics on Distributed Algorithms and Systems Fall 2008 Lab 3: Final Term Project.
 What is an operating system? What is an operating system?  Where does the OS fit in? Where does the OS fit in?  Services provided by an OS Services.
Parallel Programming Models Jihad El-Sana These slides are based on the book: Introduction to Parallel Computing, Blaise Barney, Lawrence Livermore National.
Multiple Processor Systems. Multiprocessor Systems Continuous need for faster and powerful computers –shared memory model ( access nsec) –message passing.
DynamicBLAST on SURAgrid: Overview, Update, and Demo John-Paul Robinson Enis Afgan and Purushotham Bangalore University of Alabama at Birmingham SURAgrid.
WMU CS6260 Parallel Computations II Spring 2013 Presentation #2 Professor: Dr. de Doncker Name: Xuanyu Hu March/11/2013.
Performance Model & Tools Summary Hung-Hsun Su UPC Group, HCS lab 2/5/2004.
Design of a real time strategy game with a genetic AI By Bharat Ponnaluri.
Dynamic Resource Monitoring and Allocation in a virtualized environment.
Chapter 9: Virtual Memory Background Demand Paging Copy-on-Write Page Replacement Allocation of Frames Thrashing Memory-Mapped Files Allocating Kernel.
11 Overview Paracel GeneMatcher2. 22 GeneMatcher2 The GeneMatcher system comprises of hardware and software components that significantly accelerate a.
April 26, CSE8380 Parallel and Distributed Processing Presentation Hong Yue Department of Computer Science & Engineering Southern Methodist University.
Computer Systems Lab TJHSST Current Projects In-House, pt 3.
Server Virtualization
Chapter 8-2 : Multicomputers Multiprocessors vs multicomputers Multiprocessors vs multicomputers Interconnection topologies Interconnection topologies.
Understanding Networked Applications: A First Course Ideas and examples (Chapter 6) by David G. Messerschmitt.
Server Performance, Scaling, Reliability and Configuration Norman White.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
Running BLAST on the cluster system over the Pacific Rim.
Computing Simulation in Orders Based Transparent Parallelizing Pavlenko Vitaliy Danilovich, Odessa National Polytechnic University Burdeinyi Viktor Viktorovych,
Parallel Algorithm for Multiple Genome Alignment Using Multiple Clusters Nova Ahmed, Yi Pan, Art Vandenberg Georgia State University SURA Cyberinfrastructure.
Cluster Computing Applications for Bioinformatics Thurs., Sept. 20, 2007 process management shell scripting Sun Grid Engine running parallel programs.
Project18’s Communication Drawing Design By: Camilo A. Silva BIOinformatics Summer 2008.
Typing Pattern Authentication Techniques 3 rd Quarter Luke Knepper.
Future Research Web browsers are some of the most frequently used computer applications today. However, a large portion of their data cycles is wasted.
Optimizing Parallel Programming with MPI Michael Chen TJHSST Computer Systems Lab Abstract: With more and more computationally- intense problems.
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
December 13, G raphical A symmetric P rocessing Prototype Presentation December 13, 2004.
Background Computer System Architectures Computer System Software.
Using the VTune Analyzer on Multithreaded Applications
Harnessing the Power of Condor for Human Genetics
A. Rama Bharathi Regd. No: 08931F0040 III M.C.A
Grid Computing.
Performance Evaluation of Adaptive MPI
A Framework for Automatic Resource and Accuracy Management in A Cloud Environment Smita Vijayakumar.
What is Parallel and Distributed computing?
CSE8380 Parallel and Distributed Processing Presentation
Hybrid Programming with OpenMP and MPI
VDL Mode 4 Performance Simulator (DLS enhancements) presented by EUROCONTROL Montreal, 26 October 2004.
Identifying Slow HTTP DoS/DDoS Attacks against Web Servers DEPARTMENT ANDDepartment of Computer Science & Information SPECIALIZATIONTechnology, University.
Defining the Grid Fabrizio Gagliardi EMEA Director Technical Computing
Excursions into Parallel Programming
VDL Mode 4 Performance Simulator (DLS enhancements) presented by EUROCONTROL Montreal, 26 October 2004.
Excursions into Parallel Programming
Presentation transcript:

An Investigation into Implementations of DNA Sequence Pattern Matching Algorithms Peden Nichols Computer Systems Research April,

Abstract Purpose:  To investigate the relative practicality of the three main computational methods (supercomputing, clusters, and grid computing) as applications to the problem of DNA sequence pattern matching  To establish data for local unused processor time and the degree of invasiveness of various backgrounding methods, for possible applications to grid computing implementations of BLAST  To quantify the cost to efficiency when adding processors for various BLAST algorithms.

Abstract Methods  Grid Computing MPI (Message Passing Interface) implementations of BLAST algorithms Systems Lab Resources: 40+ networked desktop machines, 16 machines currently mpi-enabled  Backgrounding Algorithms run in the background while computer is in use To what extent do users notice performance change?  Dynamic load balancing Reallocate work as new processors become available/current processors become unavailable

Background Huge amounts of genetic data  Human Genome Project  The Institute for Genomic Research (TIGR) Current computational resources overwhelmed Decoding genome not profitable (yet) Independent/Government organizations don't have the money

Background Two approaches to solution Make the algorithms run faster - mpiBLAST with dynamic load balancing - Port algorithm implementations to clusters, supercomputers Increase utilization of current computational resources - Schools, labs across the country with lots of idle time - Most processors are nowhere near 100% load - Applications for other computationally intensive problems

Procedure 1 – Demonstrate potential for greater utilization of local computational resources - CPU Load Perl Script - Records one-minute load averages every second - Produces graphable results - Sample output:

Procedure As the graphs show, CPU usage on most processors is close to zero before I start running tests on other students' computers. These graphs show CPU load averages (a measure of processor use) of computers being used by students during a systems lab class.

Procedure But did students even notice when the processor use on their computers spiked over 100% above normal levels? No! Seven out of seven students tested reported no noticeable change in performance.

Procedure Conclusions: - Average Systems Lab students use nowhere near their processor's full capabilities, during class time - That processor time/power is effectively going to waste - We haven't even mentioned the time when there is no user logged in to a given computer! Think how much time is just wasted at the login screen...

Where it goes from here... Taking advantage of unused time -Develop automated backgrounding of BLAST algorithms -Long-term tests of CPU usage -Attempt to realize increase in long-term usage with new applications Optimizing the algorithms -Test BLAST algorithms against one another -Test algorithms on different machines -Investigate supercomputer, cluster implementations