TeraGrid Archival Migration of Data to the XD Era Phil Andrews et al The Moving Finger writes; and having writ, Moves on; nor all your Piety nor Wit Shall.

Slides:



Advertisements
Similar presentations
Re-factoring grid computing for usability Bruce Beckles University of Cambridge Computing Service.
Advertisements

Cross-site data transfer on TeraGrid using GridFTP TeraGrid06 Institute User Introduction to TeraGrid June 12 th by Krishna Muriki
User Services Transition To XD TG Quarterly Management Meeting, San Juan 12/7/2010 Amit & Sergiu.
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
Implementing A Simple Storage Case Consider a simple case for distributed storage – I want to back up files from machine A on machine B Avoids many tricky.
John’s Account of Jesus. Review Reasons people don’t believe Power of God’s Word Fear of Man.
Distributed Databases John Ortiz. Lecture 24Distributed Databases2  Distributed Database (DDB) is a collection of interrelated databases interconnected.
We make it easier for businesses of all sizes to safely accept checks transmodus offers clients automation utilizing our online processing platform for.
(e)Science-Driven, Production- Quality, Distributed Grid and Cloud Data Infrastructure for the Transformative, Disruptive, Revolutionary, Next-Generation.
Engineering Secure Software. Lottery Story A Threat We Can’t Ignore  Documented incidents are prevalent Carnegie Melon’s SEI has studied over 700 cybercrimes.
70-293: MCSE Guide to Planning a Microsoft Windows Server 2003 Network, Enhanced Chapter 7: Planning a DNS Strategy.
Preservasi Informasi Digital.  It will never happen here!  Common Causes of Loss of Data  Accidental Erasure (delete, power, backup)  Viruses and.
SM3121 Software Technology Mark Green School of Creative Media.
An Introduction to DuraCloud Carissa Smith, Partner Specialist Michele Kimpton, Project Director Bill Branan, Lead Software Developer Andrew Woods, Lead.
1 Chapter Overview Creating Sites and Subnets Configuring Intersite Replication Troubleshooting Active Directory Replication.
Core Services I & II David Hart Area Director, UFP/CS TeraGrid Quarterly Meeting December 2008.
How well do we communicate? Internal communications review February 2013.
What if you suspect a security incident or software vulnerability? What if you suspect a security incident at your site? DON’T PANIC Immediately inform:
Microsoft Active Directory(AD) A presentation by Robert, Jasmine, Val and Scott IMT546 December 11, 2004.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Module 7. Data Backups  Definitions: Protection vs. Backups vs. Archiving  Why plan for and execute data backups?  Considerations  Issues/Concerns.
Virus and Antivirus Team members: - Muzaffar Malik - Kiran Karki.
A SIR web based leave/absence management system. By Dave Doulton University of Southampton.
Royal Latin School. Spec Coverage: a) Explain the advantages of networking stand-alone computers into a local area network e) Describe the differences.
Day 8 Exporting Displays Cronjobs Mount. Chapter 5 Chapter 5 talks about X windows. –You should read the chapter. –However, you do not need to pay particular.
What if you suspect a security incident or software vulnerability? What if you suspect a security incident at your site? DON’T PANIC Immediately inform:
Preventing Common Causes of loss. Common Causes of Loss of Data Accidental Erasure – close a file and don’t save it, – write over the original file when.
Igor Gaponenko ( On behalf of LCLS / PCDS ).  An integral part of the LCLS Computing System  Provides:  Mid-term (1 year) storage for experimental.
UFP/CS Update David Hart. Highlights Sept xRAC results POPS Allocations RAT follow-up User News AMIE WebSphere transition Accounting Updates Metrics,
Where Cloud Storage Makes Sense For The Enterprise (And Where it Doesn’t) Andrew Reichman Senior Analyst Forrester Research January 22, 2009.
XP Explained Chapters 7-9. Primary Practices  Sit together Ideal Resistance Multi-site  Whole Team All the necessary skills in a single management structure.
TeraGrid Quarterly Meeting Dec 5 - 7, 2006 Data, Visualization and Scheduling (DVS) Update Kelly Gaither, DVS Area Director.
Introduction to Grid Computing Ed Seidel Max Planck Institute for Gravitational Physics
The Alternative Larry Moore. 5 Nodes and Variant Input File Sizes Hadoop Alternative.
ALMA Archive Operations Impact on the ARC Facilities.
Outlook 2013 & 2010 Lync Messanger. Outlook 2013 Ribbon Navigation Pane Message View Reading Pane Tabs Status Bar.
Update on the Grid Security Vulnerability Group Linda Cornwall, MWSG7, Amsterdam 14 th December 2005
Who Says Servers Can’t Crash? Rocky Mountain PBS Survives Multiple Server Crashes and Lives to tell about it! Presented By Michelle Nesmith Rocky Mountain.
TeraGrid Archival and Service Proposal Status Phil Andrews, Patricia Kovatch Cast thy bread upon the waters: for thou shalt find it after many days Ecclesiastes.
Feedback from the POOL Project User Feedback from the POOL Project Dirk Düllmann, LCG-POOL LCG Application Area Internal Review October 2003.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
AP-1 4. Agile Processes. AP-2 Agile Processes Focus on creating a working system Different attitude on measuring progress XP Scrum.
CS5103 Software Engineering Lecture 02 More on Software Process Models.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Distributed Data for Science Workflows Data Architecture Progress Report December 2008.
Data Area Report Chris Jordan, Data Working Group Lead, TACC Kelly Gaither, Data and Visualization Area Director, TACC April 2009.
TeraGrid & XD approach to long-term archival Phil Andrews et al The Moving Finger writes; and having writ, Moves on; nor all your Piety nor Wit Shall lure.
Threads. Readings r Silberschatz et al : Chapter 4.
Simulation Status for Year2 Running Charles F. Maguire Software Meeting May 8, 2001.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
File Transfer And Access (FTP, TFTP, NFS). Remote File Access, Transfer and Storage Networks For different goals variety of approaches to remote file.
Continuous Improvement. Start Simple and Continually Improve E.g., Gmail Labels 1.
Software Integration Highlights CY2008 Lee Liming, JP Navarro GIG Area Directors for Software Integration University of Chicago, Argonne National Laboratory.
CSE 486/586, Spring 2012 CSE 486/586 Distributed Systems Paxos Steve Ko Computer Sciences and Engineering University at Buffalo.
1 November 17, 2005 FTS integration at CMS Stefano Belforte - INFN-Trieste CMS Computing Integration Coordinator CMS/LCG integration Task Force leader.
System Architecture & Hardware Configurations Dr. D. Bilal IS 582 Spring 2008.
Page 1 Viruses. Page 2 What Is a Virus A virus is basically a computer program that has been written to perform a specific set of tasks. Unfortunately,
Visualization Update June 18, 2009 Kelly Gaither, GIG Area Director DV.
James Howison What should scientific software learn from “the open source way”? CC Credit: XSEDE Communities Symposium.
St. Mary’s Catholic School, Mayville Mrs. Kaiser, Technology Teacher.
Increasing Motivation Part 1. Three Motivation Problems Level of Concern Feeling Tone Success.
Guide By Phoebi Stewart. has changed the way we communicate with each other. Being able to send information quickly across the internet.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
TeraGrid Data Plan/Issues Phil Andrews In Xanadu did Kubla Khan A stately pleasure-dome decree: Where Alph, the sacred river, ran Through caverns measureless.
System Architecture & Hardware Configurations
Extreme Programming.
Distributed System Structures 16: Distributed Structures
Here are some top tips to help you bake responsible data into your project design:.
EECS 498 Introduction to Distributed Systems Fall 2017
Crowdfunding Let’s Grow State Getting Started
Presentation transcript:

TeraGrid Archival Migration of Data to the XD Era Phil Andrews et al The Moving Finger writes; and having writ, Moves on; nor all your Piety nor Wit Shall lure it back to cancel half a Line, Nor all your Tears wash out a Word of it. -Omar Khayyam Users’ view: When they give us their data, they expect it to be available even when the original recipient is not.

2 Significant Archival Data (~20PB) is at TG RP sites unfunded in XD TeraGrid Quarterly, Dec’08 What to do about data at current TeraGrid RP sites that do not yet have funds for the XD era? Do we have a communal obligation to continue data availability past the funding of centers that accepted it? The NSF thinks so! It’s Later Than You Think! – A Tale of Two Cities, Charles Dickens

3 Task Force to consider issue Members from most (maybe all) sites Send me if you want to NSF wants to see plan, encouraging idea of general replication approach at remote sites If replicate data at currently unfunded sites, then we are covered whatever happens Awkward funding implications TeraGrid Quarterly, Oct’08

4 More than one approach possible TeraGrid Quarterly, Dec’08 In the past, moved one archive (CTC) physically, another (PSC) across the network. Both moves successful, never replicated entire archive. Network moves require several months. Physical moves very concerning. Data offline or frozen during move. A merry road, a mazy road, and such as we did tread The night we went to Birmingham by way of Beachy Head! – G.K Chesterton

5 How much data are we talking? Approximately 10 PB total at each of SDSC and NCSA Other sites also have significant data 10 Gb/s = 10 PB/100 days Only TACC, NICS, PSC, continually funded into XD era at the moment NCSA -> track1 funding TeraGrid Quarterly, Oct’08

6 Option 1: Physical move TeraGrid Quarterly, Dec’08 Advantages: can wait until last minute, possibly funding neutral, doesn’t stress network, keeps physical resources in TG Disadvantages: dangerous, data unavailable for weeks, site could regain funding later, new host must handle format The Nuclear option: very awkward, mixed data tapes, lays waste to an existing archive. Forced upon us if we wait too long! Out of this nettle, danger, we pluck this flower, safety. – Shakespeare

7 Option 2: Network Transfer Cannot move 20 PB in any reasonable time Must rely on only 2-3 PB real data/site Advantages: Data checked during transfer. No danger of data loss. Site can recover. Disadvantages: ties up network, people resources. Long process. Doubles archival requirements. TeraGrid Quarterly, Oct’08 For though his body ’s under hatches, His soul has gone aloft. – Charles Dibdin

8 Option 3: Archival Replication Advantages: More general approach; increases TeraGrid value added. Intellectually stimulating rather than maudlin Disadvantages: more involved process. Could lead to drastically increased archival requirements. TeraGrid Quarterly, Oct’08 There is a tide in the affairs of men Which taken at the flood, leads on to fortune– Shakespeare

9 Replication Approaches: General Middleware: 1. iRODS can do replication, but must manage the data. Can’t import general or SRB data 2. SRB is slow Infrastructure: 1. HPSS archives can be connected via wide area GPFS. (HPSS 6.2.2, June’08) TeraGrid Quarterly, Oct’08

10 What to do now? The clock is ticking; if we are to investigate options, must do it soon SDSC, TACC, NCSA looking at iRODS SDSC runs HPSS as one archive, and exports GPFS Propose trying the GPFS-HPSS Integration (GHI) approach for replication between HPSS archives TeraGrid Quarterly, Oct’08

11 Using WAN GPFS to connect Archives

12 Will other file systems work? Can we use other approaches for Lustre? pNFS does have a proposed mechanism for replication via caching: Panache Will global file systems and HPSS come in pairs? Is a more general (but less efficient) middleware approach (iRODS?) preferable? TeraGrid Quarterly, Oct’08 Pay no attention to that man behind the curtain – L. Frank Baum

13 GHI status Some features already released Multiple HPSS archives not there yet; due next year Timing could be tricky, but could start with pre-release software Due for beta testing at NERSC, NCSA GPFS, HPSS guys interested (spoke at SC) TeraGrid Quarterly, Oct’08

14 Discussion: Is replication worth the effort? Will sites be prepared for physical move, if necessary? If no physical move, how do we fund resources? Do we let users say: “move everything”? We need an inventory of data! Are users rendering this discussion moot? TeraGrid Quarterly, Oct’08

15 Philosophy: Current funding approach allows continual ebb and flow of RP sites: we can handle Computational impact, but not Archival! Need Archival organization that allows for a frequent gain and loss of Data RP’s Hard to wait for XD to solve this problem TeraGrid Quarterly, Oct’08 I must go in and out – Bernard Shaw

16 Need to know what data is where: We don’t know which site has how much data, and on what media Different media can have major impacts on how quickly it can be moved or replicated Need a good story to go to NSF with funding request for a better Archival organization Need a Data census! TeraGrid Quarterly, Oct’08 In those days a decree went out from Caesar Augustus that the whole world should be counted – Luke, 2:1