Thursday AM, Lecture 1 Lauren Michael

Slides:



Advertisements
Similar presentations
Tivoli SANergy. SANs are Powerful, but... Most SANs today offer limited value One system, multiple storage devices Multiple systems, isolated zones of.
Advertisements

Lecture 12: MapReduce: Simplified Data Processing on Large Clusters Xiaowei Yang (Duke University)
© 2010 Quest Software, Inc. ALL RIGHTS RESERVED Quest vRanger / Architecture.
Buffers & Spoolers J L Martin Think about it… All I/O is relatively slow. For most of us, input by typing is painfully slow. From the CPUs point.
Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
Efficiently Sharing Common Data HTCondor Week 2015 Zach Miller Center for High Throughput Computing Department of Computer Sciences.
The Difficulties of Distributed Data Douglas Thain Condor Project University of Wisconsin
Intermediate HTCondor: Workflows Monday pm Greg Thain Center For High Throughput Computing University of Wisconsin-Madison.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
Data Deduplication in Virtualized Environments Marc Crespi, ExaGrid Systems
Utilizing Condor and HTC to address archiving online courses at Clemson on a weekly basis Sam Hoover 1 Project Blackbird Computing,
RAMCloud Design Review Recovery Ryan Stutsman April 1,
Volunteer Computing and Hubs David P. Anderson Space Sciences Lab University of California, Berkeley HUBbub September 26, 2013.
Maintaining a Microsoft SQL Server 2008 Database SQLServer-Training.com.
Introduction to Windows XP Professional Chapter 2 powered by dj.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
IntroductiontotoHTCHTC 2015 OSG User School, Monday, Lecture1 Greg Thain University of Wisconsin– Madison Center For High Throughput Computing.
Maintaining File Services. Shadow Copies of Shared Folders Automatically retains copies of files on a server from specific points in time Prevents administrators.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
Course ILT Basics of information technology Unit objectives Define “information technology” (IT), distinguish between hardware and software, and identify.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Workflows: from Development to Production Thursday morning, 10:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
Building a Real Workflow Thursday morning, 9:00 am Greg Thain University of Wisconsin - Madison.
Computer Guts and Operating Systems CSCI 101 Week Two.
Click here to download this powerpoint template : Colorful Networks Free Powerpoint TemplateColorful Networks Free Powerpoint Template For more : Powerpoint.
Turning science problems into HTC jobs Wednesday, July 29, 2011 Zach Miller Condor Team University of Wisconsin-Madison.
Building a Real Workflow Thursday morning, 9:00 am Lauren Michael Research Computing Facilitator University of Wisconsin - Madison.
APST Internals Sathish Vadhiyar. apstd daemon should be started on the local resource Opens a port to listen for apst client requests Runs on the host.
SAN DIEGO SUPERCOMPUTER CENTER Inca Control Infrastructure Shava Smallen Inca Workshop September 4, 2008.
1 MSRBot Web Crawler Dennis Fetterly Microsoft Research Silicon Valley Lab © Microsoft Corporation.
Experiences Running Seismic Hazard Workflows Scott Callaghan Southern California Earthquake Center University of Southern California SC13 Workflow BoF.
By: Joel Dominic and Carroll Wongchote 4/18/2012.
Five todos when moving an application to distributed HTC.
Getting the Most out of HTC with Workflows Friday Christina Koch Research Computing Facilitator University of Wisconsin.
Turning science problems into HTC jobs Tuesday, Dec 7 th 2pm Zach Miller Condor Team University of Wisconsin-Madison.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
THE ATLAS COMPUTING MODEL Sahal Yacoob UKZN On behalf of the ATLAS collaboration.
Hadoop Introduction. Audience Introduction of students – Name – Years of experience – Background – Do you know Java? – Do you know linux? – Any exposure.
Large Input in DHTC Thursday PM, Lecture 1 Lauren Michael CHTC, UW-Madison.
Christina Koch Research Computing Facilitators
Dynamic Extension of the INFN Tier-1 on external resources
Large Output and Shared File Systems
XNAT at Scale June 7, 2016.
By Chris immanuel, Heym Kumar, Sai janani, Susmitha
Overview of the Belle II computing
Demystifying Deduplication
Thursday AM, Lecture 2 Lauren Michael CHTC, UW-Madison
High Availability in HTCondor
Bernd Panzer-Steindel, CERN/IT
OSG Connect and Connect Client
Microsoft FrontPage 2003 Illustrated Complete
Integration of Singularity With Makeflow
Introduction of Week 3 Assignment Discussion
Composition and Operation of a Tier-3 cluster on the Open Science Grid
Overview Introduction VPS Understanding VPS Architecture
An Overview of the Computer System
湖南大学-信息科学与工程学院-计算机与科学系
Types of Computers Mainframe/Server
Map Reduce Workshop Monday November 12th, 2012
Introduction to High Throughput Computing and HTCondor
Brian Lin OSG Software Team University of Wisconsin - Madison
Communications & Computer Networks Resource Notes - Introduction
CS 345A Data Mining MapReduce This presentation has been altered.
The Challenge Collaboration Teachers PTO Students Anyone.
Quick Tips #1 – Wan accelerator seeding for backup jobs
Credential Management in HTCondor
Troubleshooting Your Jobs
Job Submission Via File Transfer
Brian Lin OSG Software Team University of Wisconsin - Madison
Thursday AM, Lecture 2 Brian Lin OSG
Presentation transcript:

Thursday AM, Lecture 1 Lauren Michael Data Considerations Thursday AM, Lecture 1 Lauren Michael

Like all things I always think of HTC/OSG usage as a spectrum: Mo Resources, Mo Planning Laptop Cluster OSG

Planning? Can’t control a cluster like your laptop, where you can install any software and place files (until they flat-out don’t fit) OSG: heterogeneity, borrowed resources (including network and disk), lack of on-the-fly troubleshooting

Benefits! On a cluster & OSG you can access 1000+ cores! Automate job tasks (with HTCondor)! Doesn’t burn up your laptop!

Overview – Data Handling Review of HTCondor Data Handling Data Management Tips What is ‘Large’ Data? Dealing with Large Data Next talks: OSG-wide methods for large-data handling, and when to stay ‘local’

Overview – Data Handling Review of HTCondor Data Handling Data Management Tips What is ‘Large’ Data? Dealing with Large Data Next talks: OSG-wide methods for large-data handling, and when to stay ‘local’

Review: HTCondor Data Handling exec server exec server exec server submit server exec server HTCondor submit file executable dir/ input output (exec dir)/ executable input output

Network bottleneck: the submit server exec server exec server exec server submit server exec server HTCondor submit file executable dir/ input output (exec dir)/ executable input output

Overview – Data Handling Review of HTCondor Data Handling Data Management Tips What is ‘Large’ Data? Dealing with Large Data Next talks: local and OSG-wide methods for large-data handling

Data Management Tips Determine your per-job needs minimize per-job data needs Determine your batch needs Leverage HTCondor and OSG data handling features!

Determining In-Job Needs “Input” includes any files transferred by HTCondor executable transfer_input_files data and software “Output” includes any files copied back by HTCondor output, error

First! Try to minimize your data split large input for better throughput eliminate unnecessary data file compression and consolidation job input: prior to job submission job output: prior to end of job moving data between your laptop and the submit server

Overview – Data Handling Review of HTCondor Data Handling Data Management Tips What is ‘Large’ Data? Dealing with Large Data Next talks: local and OSG-wide methods for large-data handling

What is big large data? In reality, “big data” is relative What is ‘big’ for you? Why?

What is big large data? In reality, “big data” is relative What is ‘big’ for you? Why? Volume, velocity, variety! think: a million 1-KB files, versus one 1-TB file

Network bottleneck: the submit server exec server exec server exec server submit server exec server HTCondor submit file executable dir/ input output (exec dir)/ executable input output

‘Large’ input data: The collaborator analogy What method would you use to send data to a collaborator? amount method of delivery words email body tiny – 100MB email attachment (managed transfer) 100MB – GBs download from Google Drive, Drop/Box, other web-accessible repository TBs ship an external drive (local copy needed)

Large input in HTC and OSG exec server amount method of delivery words within executable or arguments? tiny – 100MB per file HTCondor file transfer (up to 1GB total) 100MB – 1GB, shared download from web server (local caching) 1GB - 20GB, unique or shared StashCache (regional replication) 20 GB - TBs shared file system (local copy, local execute servers)

Transfers More Data HTCondor File Transfer HTTP Proxies StashCache Local Storage

Large input in HTC and OSG exec server amount method of delivery words within executable or arguments? tiny – 100MB per file HTCondor file transfer (up to 1GB total) 100MB – 1GB, shared download from web server (local caching) 1GB - 20GB, unique or shared StashCache (regional replication) 20 GB - TBs shared file system (local copy, local execute servers)

Network bottleneck: the submit server Input transfers for many jobs will coincide exec server exec server exec server submit server exec server HTCondor submit file executable dir/ input output (exec dir)/ executable input output

Network bottleneck: the submit server Input transfers for many jobs will coincide exec server exec server exec server submit server exec server HTCondor submit file executable dir/ input output (exec dir)/ executable input output Output transfers are staggered

Output for HTC and OSG exec server amount method of delivery words within executable or arguments? tiny – 1GB, total HTCondor file transfer 1GB+, total shared file system (local copy, local execute servers)

Output for HTC and OSG exec server Why are there fewer options than for input? exec server amount method of delivery words within executable or arguments? tiny – 1GB, total HTCondor file transfer 1GB+, total shared file system (local copy, local execute servers)

Overview – Data Handling Review of HTCondor Data Handling Data Management Tips What is ‘Large’ Data? Dealing with Large Data Next talks: local and OSG-wide methods for large-data handling

Exercises 1.1 Understanding a job’s data needs 1.2 Using data compression with HTCondor file transfer 1.3 Splitting input (prep for large run in 2.1)

Questions? Next: Exercises 1.1-1.3 Later: Handling large input data amount method of delivery words within executable or arguments? tiny – 100MB per file HTCondor file transfer (up to 1GB total) 100MB – 1GB, shared download from web server (local caching) 1GB - 20GB, unique or shared StashCache (regional replication) 20 GB - TBs shared file system (local copy, local execute servers)