Auspice: AUtomatic Service Planning in Cloud/Grid Environments David Chiu Dissertation Defense May 25, 2010 Committee: Prof. Gagan Agrawal, Advisor Prof.

Slides:



Advertisements
Similar presentations
Evaluating Caching and Storage Options on the Amazon Web Services Cloud Gagan Agrawal, Ohio State University - Columbus, OH David Chiu, Washington State.
Advertisements

Paging: Design Issues. Readings r Silbershatz et al: ,
SLA-Oriented Resource Provisioning for Cloud Computing
Data Grid Research Group Dept of Computer Science and Engineering The Ohio State University David Chiu and Gagan Agrawal Cost and Accuracy Sensitive Dynamic.
Cloud Computing Resource provisioning Keke Chen. Outline  For Web applications statistical Learning and automatic control for datacenters  For data.
1 An Empirical Study on Large-Scale Content-Based Image Retrieval Group Meeting Presented by Wyman
Presented by Zeehasham Rasheed
1 Exploring Data Reliability Tradeoffs in Replicated Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh Matei Ripeanu.
Cloud based Dynamic workflow with QOS for Mass Spectrometry Data Analysis Thesis Defense: Ashish Nagavaram Graduate student Computer Science and Engineering.
Computer Science and Engineering A Middleware for Developing and Deploying Scalable Remote Mining Services P. 1DataGrid Lab A Middleware for Developing.
MATE-EC2: A Middleware for Processing Data with Amazon Web Services Tekin Bicer David Chiu* and Gagan Agrawal Department of Compute Science and Engineering.
Adaptive Server Farms for the Data Center Contact: Ron Sheen Fujitsu Siemens Computers, Inc Sever Blade Summit, Getting the.
Basics of Operating Systems March 4, 2001 Adapted from Operating Systems Lecture Notes, Copyright 1997 Martin C. Rinard.
Query Planning for Searching Inter- Dependent Deep-Web Databases Fan Wang 1, Gagan Agrawal 1, Ruoming Jin 2 1 Department of Computer.
1 Exploring Data Reliability Tradeoffs in Replicated Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh Advisor: Professor.
Flashing Up the Storage Layer I. Koltsidas, S. D. Viglas (U of Edinburgh), VLDB 2008 Shimin Chen Big Data Reading Group.
A Metadata Based Approach For Supporting Subsetting Queries Over Parallel HDF5 Datasets Vignesh Santhanagopalan Graduate Student Department Of CSE.
Ohio State University Department of Computer Science and Engineering 1 Cyberinfrastructure for Coastal Forecasting and Change Analysis Gagan Agrawal Hakan.
An Autonomic Framework in Cloud Environment Jiedan Zhu Advisor: Prof. Gagan Agrawal.
Ohio State University Middleware Systems Driven by Sensing Scenarios Gagan Agrawal CSE (Joint Work with Qian Zhu, David Chiu, Ron Li, Keith Bedford ….
Master Thesis Defense Jan Fiedler 04/17/98
1 Time & Cost Sensitive Data-Intensive Computing on Hybrid Clouds Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
20 October 2006Workflow Optimization in Distributed Environments Dynamic Workflow Management Using Performance Data David W. Walker, Yan Huang, Omer F.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
1 A Framework for Data-Intensive Computing with Cloud Bursting Tekin Bicer David ChiuGagan Agrawal Department of Compute Science and Engineering The Ohio.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
April 14, 2003Hang Cui, Ji-Rong Wen and Tat- Seng Chua 1 Hierarchical Indexing and Flexible Element Retrieval for Structured Document Hang Cui School of.
SEEK EcoGrid l Integrate diverse data networks from ecology, biodiversity, and environmental sciences l Metacat, DiGIR, SRB, Xanthoria,... l EML is the.
Euro-Par, A Resource Allocation Approach for Supporting Time-Critical Applications in Grid Environments Qian Zhu and Gagan Agrawal Department of.
ICDL 2004 Improving Federated Service for Non-cooperating Digital Libraries R. Shi, K. Maly, M. Zubair Department of Computer Science Old Dominion University.
Mehmet Can Kurt, The Ohio State University Gagan Agrawal, The Ohio State University DISC: A Domain-Interaction Based Programming Model With Support for.
Computer Science and Engineering Predicting Performance for Grid-Based P. 1 IPDPS’07 A Performance Prediction Framework.
Data Replication and Power Consumption in Data Grids Susan V. Vrbsky, Ming Lei, Karl Smith and Jeff Byrd Department of Computer Science The University.
CCGrid 2014 Improving I/O Throughput of Scientific Applications using Transparent Parallel Compression Tekin Bicer, Jian Yin and Gagan Agrawal Ohio State.
Streamflow - Programming Model for Data Streaming in Scientific Workflows Chathura Herath.
Data Grid Research Group Dept. of Computer Science and Engineering The Ohio State University Columbus, Ohio 43210, USA David Chiu & Gagan Agrawal Enabling.
GEON2 and OpenEarth Framework (OEF) Bradley Wallet School of Geology and Geophysics, University of Oklahoma
1 Supporting Dynamic Migration in Tightly Coupled Grid Applications Liang Chen Qian Zhu Gagan Agrawal Computer Science & Engineering The Ohio State University.
Elastic Cloud Caches for Accelerating Service-Oriented Computations Gagan Agrawal Ohio State University Columbus, OH David Chiu Washington State University.
David Chiu and Gagan Agrawal Department of Computer Science and Engineering The Ohio State University 1 Supporting Workflows through Data-driven Service.
Computer Architecture Lecture 26 Past and Future Ralph Grishman November 2015 NYU.
ApproxHadoop Bringing Approximations to MapReduce Frameworks
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
Feb 24-27, 2004ICDL 2004, New Dehli Improving Federated Service for Non-cooperating Digital Libraries R. Shi, K. Maly, M. Zubair Department of Computer.
Ohio State University Department of Computer Science and Engineering Servicing Range Queries on Multidimensional Datasets with Partial Replicas Li Weng,
20 Copyright © 2008, Oracle. All rights reserved. Cache Management.
Sunpyo Hong, Hyesoon Kim
Data Grid Research Group Dept. of Computer Science and Engineering The Ohio State University Columbus, Ohio 43210, USA David Chiu and Gagan Agrawal Enabling.
Instance Discovery and Schema Matching With Applications to Biological Deep Web Data Integration Tantan Liu, Fan Wang, Gagan Agrawal {liut, wangfa,
Euro-Par, HASTE: An Adaptive Middleware for Supporting Time-Critical Event Handling in Distributed Environments ICAC 2008 Conference June 2 nd,
18 May 2006CCGrid2006 Dynamic Workflow Management Using Performance Data Lican Huang, David W. Walker, Yan Huang, and Omer F. Rana Cardiff School of Computer.
1 Supporting a Volume Rendering Application on a Grid-Middleware For Streaming Data Liang Chen Gagan Agrawal Computer Science & Engineering Ohio State.
Evaluating and Optimizing Indexing Schemes for a Cloud-based Elastic Key- Value Store Apeksha Shetty and Gagan Agrawal Ohio State University David Chiu.
GUILLOU Frederic. Outline Introduction Motivations The basic recommendation system First phase : semantic similarities Second phase : communities Application.
Efficient Multi-User Indexing for Secure Keyword Search
Parallel Programming By J. H. Wang May 2, 2017.
Liang Chen Advisor: Gagan Agrawal Computer Science & Engineering
Year 2 Updates.
The Composite-File File System: Decoupling the One-to-one Mapping of Files and Metadata for Better Performance Shuanglong Zhang, Helen Catanese, Andy An-I.
Li Weng, Umit Catalyurek, Tahsin Kurc, Gagan Agrawal, Joel Saltz
Data-Intensive Computing: From Clouds to GPU Clusters
Smita Vijayakumar Qian Zhu Gagan Agrawal
AWS Cloud Computing Masaki.
Declarative Transfer Learning from Deep CNNs at Scale
Chaitali Gupta, Madhusudhan Govindaraju
Cache - Optimization.
Tantan Liu, Fan Wang, Gagan Agrawal The Ohio State University
Scientific Workflows Lecture 15
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Auspice: AUtomatic Service Planning in Cloud/Grid Environments David Chiu Dissertation Defense May 25, 2010 Committee: Prof. Gagan Agrawal, Advisor Prof. Hakan Ferhatosmanoglu Prof. Christopher Stewart

2 Explosion of Scientific Data Sources The amount of scientific data has increased dramatically over the years In just one example, ‣ Large Hadron Collider (LHC) ‣ 15 petabytes annually ‣ 60 petabytes overall Management and processing have become challenging

3 Data Sources A Live Cyber Infrastructure

4 Computing & Storage Resources A Live Cyber Infrastructure

5 Shared/Proprietary Web Services = Web Service A Live Cyber Infrastructure

6... A Live Cyber Infrastructure

7 Service Interaction with Cyber Infrastructure... invoke results

8 Current GUI for Creating Workflows

9 Scientific Workflow Challenges ??? ‣ Difficulties for the scientist: ‣ How to identify which data sets to use, and from where to get them? ‣ Which services are available to me to use? ‣ What resources to utilize? ‣ How can I accelerate workflow execution? ‣ Do I really have to do all this myself?

10 Contributions Workflow System-- with the following support High-level scientific user querying ‣ D. Chiu and G. Agrawal. A Keyword Querying Interface for Invoking Scientific Workflows. (OSU-TR, submitting to ACM-GIS’10) ‣ D. Chiu and G. Agrawal. Enabling Ad Hoc Queries over Low-Level Scientific Data Sets. (SSDBM'09) Automatic workflow planning ‣ D. Chiu and G. Agrawal. Enabling Ad Hoc Queries over Low-Level Scientific Data Sets. (SSDBM'09) ‣ D. Chiu and G. Agrawal. Ad Hoc Scientific Workflows through Data-driven Service Composition. (eScience'07)

11 Contributions (continued) Quality of Service ‣ D. Chiu, S. Deshpande, G. Agrawal, and R. Li. A Dynamic Approach toward QoS- Aware Service Workflow Composition. (ICWS’09) ‣ D. Chiu, S. Deshpande, G. Agrawal, and R. Li. Cost and Accuracy Sensitive Dynamic Workflow Composition over Grid Environments. (GRID'08) ‣ D. Chiu, S. Deshpande, G. Agrawal, and R. Li. Composing Geoinformatics Workflows with User Preferences. (GIS’08) Accelerating Workflow Execution ‣ D. Chiu and G. Agrawal. Evaluating Caching and Storage Options on the Amazon Web Service Cloud. (OSU-TR, submitted to GRID’10) ‣ D. Chiu, A. Shetty, and G. Agrawal. Elastic Cloud Caches for Derived Data Reuse. (OSU-TR, submitted to SC’10) ‣ D. Chiu and G. Agrawal. Hierarchical Caches for Grid Workflows. (CCGrid’09)

12 Presentation Outline Motivation & Introduction Our Service Composition System: Auspice ‣ Metadata Framework ‣ Cost-Aware Service Planning ‣ Supporting Keyword Queries ‣ Elastic Cache Deployment Conclusion Auspice

13 Auspice System

14 Auspice System D. Chiu & G. Agrawal, eScience ’07 D. Chiu & G. Agrawal, SSDBM ’09

15 What known data or services can derive a coast line? Systematic Way to Plan Workflows? Goal-Driven, Recursive Concept Derivation Example User Goal: Coastline Extraction Coa st Line We are targeting some coastline concept in the geospatial domain

16 What known data or services can derive water level? Available ServicesAvailable Data What known data or services can derive a CTM? Available ServicesAvailable Data Coa st Line Coast Extrac t 1 Coast Data 1 Coast Data N Available ServicesAvailable Data Types What are its parameters? Systematic Way to Plan Workflows? Coast Extrac t K Wate r Leve l CTM

17 Coa st Line Systematic Way to Plan Workflows? Coast Extrac t K Wate r Leve l CTM Coast Extrac t 1 Coast Data 1 Coast Data N

18 Coa st Line Systematic Way to Plan Workflows? Workflow 1Workflow 2 Workflow 3...

19 Ontology for Applying Domain Information Domain concepts can be derived from executing a service Domain concepts can be derived from retrieving an existing data Service parameters can be represented by certain domain concepts

20 Example Subset of Some Ontology

21 Auspice Metadata Registration Given a data set or service, ‣ Ontology is applied to new resources ‣ Resources are indexed and immediately usable in workflow planner ‣ Non-intrusive

22 Registering Data Sets

23 Registering Services

24 Subset of Ontology, with Shoreline Target

25 Service Planning: An Example A Derived Execution Plan for shoreline concept

26 What Users Want Do what you can to provide me results in under 20 minutes. I want the fastest results with at least 75% accuracy - Exec time prediction, - Online data reduction - Domain-specific error modeling

27 Presentation Outline Motivation & Introduction Our Service Composition System: Auspice ‣ Metadata Framework ‣ Cost-Aware Service Planning ‣ Supporting Keyword Queries ‣ Elastic Cache Deployment Conclusion Auspice

28 Auspice System

29 Auspice System D. Chiu, S. Deshpande, G. Agrawal, & R. Li, GRID ’08 D. Chiu, S. Deshpande, G. Agrawal, & R. Li, ACM-GIS ’08 D. Chiu, S. Deshpande, G. Agrawal, & R. Li, ICWS ’09

30 Challenges We wish to project workflow execution time and workflow accuracy costs at planning time Allow input models per service We should prune all workflows unlikely to meet the user’s demands

31 Estimating Workflow Execution Time Service execution time (t x ) ‣ Each service is trained beforehand with various sized inputs Data output size (d size ) ‣ Known for files. But models are again trained for service output Network transmission time (t net ) ‣ Bandwidth between nodes are typically known Recall the workflow structure:

32 Estimating Workflow Error/Accuracy The recursive sum is similar for error propagation The errors,, attributed from services and data are implemented by domain scientists is an accuracy parameter, e.g., sampling rate, resolution,..

33 Cost Models Declared per Operation

34 Water Level Workflow Example Workflow Plan 1Workflow Plan 2 [t_total= t_x=1 t_d=0 o=47889 e=0.004] SRVC.getWL( X= Y= StnID= [t_total=2.5 t_x=0.5 t_d=0 o=0 e=0.004] SRVC.getKNearestStations( Longitude= Latitude= ListOfStations= [t_total=2 t_x=2 t_d=0 o=47889 e=0] SRVC.GetGSListGreatLakes() RadiusKM=100 K=3 ) time=00:06 date=01/30/2008 ) [t_total=2 t_x=2 t_d=0 o=47889 e=2.4997] SRVC.getWLfromModel( X= Y= time=00:06 date=01/30/2008 ) Total Projected Costs: Workflow Execution Time = Workflow Error = Total Projected Costs: Workflow Execution Time = Workflow Error =

35 On Meeting QoS? Users specify QoS accuracy with respect to domain, not data quality ‣ For instance, what does +/- 3 meters mean in terms of image resolution or sampling rate? But service planner is interested in data quality ‣ Inverse the error model? ‣ Adaptive precision logic

36 Adaptive Precision Logic Sampling Rate Time? Error? Time? Error? sample more sample less Time? Error? ‣ Often, the error model is read-only ‣ Suggest a new value for parameters via binary-search for the best possible value by repeatedly invoking the model

37 System Configuration Computing Environment ‣ Auspice (local)  Linux  Pentium IV 3.0GHz Dual Core  1GB RAM ‣ Service Node  Across OSU campus in Dept of Civil Engg and Geodetic Science  10MBps Interconnection ‣ Data Storage Node  Across state at Kent State University Dept of Computer Science

38 Cost Model Overheads

39 Experimented Workflow Shoreline Extraction Users can specify the following QoS Parameters: Allowed execution time Allowed error

40 On Meeting Time Constraints

41 On Meeting Error Constraints

42 Presentation Outline Motivation & Introduction Our Service Composition System: Auspice ‣ Metadata Framework ‣ Cost-Aware Service Planning ‣ Supporting Keyword Queries ‣ Elastic Cache Deployment Conclusion Auspice

43 Current GUI for Creating Workflows

44 Auspice System

45 Auspice System D. Chiu & G. Agrawal, SSDBM’09 D. Chiu & G. Agrawal, (submitting to GIS’10)

46 Supporting Keyword Querying Planning workflows is hard, while keyword search has become an extremely popular interface for information retrieval ‣ No need to know underlying structure of data ‣ No need to understand structured query languages like SQL Goal: Given set of key terms in the scientific domain, return ranked list of workflow plans to the user for execution

47 Keyword Decomposition coastCTM7/8/2003(41.30, -82.4)“”line Filter stopping/stemming/pattern-match map

48 Keyword Maximization coast CTM 7/8/ line C C C longitude C C date C latitude D D D Data-Substantiated Concepts Unsubstantiated Concepts Any combination of these is potentially what the query is targeting! Potential query parameters

49 Keyword Querying coast CTM line C C C Merged Super Concept Query Target CandidateRequisite Concepts 7/8/ C longitude C date C latitude D D D Query Parameters

50 Keyword Querying coast CTM line C C C Merged Super Concept Query Target CandidateRequisite Concepts 7/8/ C longitude C date C latitude D D D Query Parameters

51 Keyword Querying coast CTM line C C C Merged Super Concept Query Target CandidateRequisite Concepts 7/8/ C longitude C date C latitude D D D Query Parameters Enumerate Workflows

52 Ranking Workflow Plans by Relevance Method: ‣ Let be the set of input keyword-concepts ‣ Rank workflow plans on

53 A Case Study The following keyword queries were submitted to Auspice

54 Search Time

55 Precision

56 Result Set for QueryID 3 “( , ) 7/8/2003 wind CTM”

57 Presentation Outline Motivation & Introduction Our Service Composition System: Auspice ‣ Metadata Framework ‣ Cost-Aware Service Planning ‣ Supporting Keyword Queries ‣ Elastic Cache Deployment Conclusion Auspice

58 Problem: Query Intensive Circumstances...

59 Caching Intermediate Results Shoreline Extraction Time consuming! Can’t we cache the result from when it was last computed??

60 Caching Intermediate Results

61 Auspice System

62 Auspice System D. Chiu & G. Agrawal, CCGrid’09 D. Chiu, A. Shetty, & G. Agrawal, (submitted to SC’10) D. Chiu & G. Agrawal, (submitted to GRID’10)

63 Cloud Computing Pay as you go computing Elasticity ‣ Cloud applications can stretch and relax their resource requirements “Infinite” compute and storage resources

64 A Workflow Cache Compute Cloud... A B

65... A B Consistent Hashing

66... A B invoke: service(35) (35 mod 100) = 35 Which proxy has the page? h(k) = (k mod 100) h(35) Consistent Hashing

67 A B C Only records hashing into (25,50] need to be moved from A to C! Our algorithm for Scaling up GBA: Greedy Bucket Allocation

68 Experimental Configuration Workload ‣ Shoreline Extraction Workflow ‣ Takes 23 seconds to complete without benefits of cache ‣ Executed on a miss Amazon EC2 Cloud ‣ Each Cloud node:  Small Instances (Single core 1.2Ghz, 1.7GB, 32bits)  Ubuntu Linux ‣ Caches start out cold ‣ Cache stored in memory only

69 Experimental Configuration Our approach exploits a dynamic Cloud environment: ‣ Consistent Hashing: Greedy Bucket Allocation (GBA) ‣ Elastic number of nodes We compare GBA against statically allocated Cloud environments: ‣ 2 fixed nodes (static-2) ‣ 4 fixed nodes (static-4) ‣ 8 fixed nodes (static-8) ‣ Cache overflow --> LRU eviction

70 Relative Speedup Querying Rate: 255 invocations/sec Cost Savings

71 Maximum Execution Times (intensive rate) Querying Rate: 255 invocations/sec

72 That’s Not Completely Elastic What about relaxing the amount of nodes to help save Cloud save costs? First, we need an eviction scheme

73 Exponential Decay Eviction At eviction time: ‣ A value,, is calculated for each data record in the evicted slice ‣ is higher:  if was accessed more recently  if was accessed frequently ‣ If is lower than some fixed threshold, evict

74 Experimental Configuration Amazon EC2 Cloud ‣ Each Cloud node:  Small Instances (Single core 1.2Ghz, 1.7GB, 32bits)  Ubuntu Linux ‣ Caches start out cold ‣ Data stored in memory ‣ When 2 nodes become < 30% capacity, merge Sliding Window Configuration: ‣ Time Slice: 1 sec ‣ Size: 100 Time Slices

75 Data Eviction: 50/255/50 queries per sec Sliding Window Size = 100 sec 50 q/sec255 q/sec50 q/sec

76 Cache Contraction: 50/255/50 queries per sec

77 Cache Contraction: 50/255/50 queries per sec

78 Cache Contraction: 50/255/50 queries per sec Sliding Window Size = 100 sec 50 q/sec255 q/sec50 q/sec

79 Cache Hits over Varying Decay Sliding Window Size = 100 sec

80 Presentation Outline Motivation & Introduction Our Service Composition System: Auspice ‣ Metadata Framework ‣ Cost-Aware Service Planning ‣ Supporting Keyword Queries ‣ Elastic Cache Deployment Conclusion Auspice

81 Future Work Dynamic sliding window size Evaluate and model various Cloud infrastructure options to optimize cost for sustaining the cache Transparent remote data analysis over Clouds Deep Web Integration into querying framework

82 Summary and Conclusion Auspice is a workflow system, which ‣ Supports high-level keyword/NLP user queries ‣ Automatically composes workflows, and adapts to QoS Constraints ‣ Caches workflow results to accelerate workflow execution Questions? Auspice

83 Capturing Concept Derivability Domain concepts can be derived from executing a service Domain concepts can be derived from retrieving an existing data Service parameters represent different domain concepts

84 Indexing Data Sets

85 Applying Domain Information Domain concepts can be derived from executing a service Domain concepts can be derived from retrieving an existing data Service parameters represent different domain concepts

86 latitude A Case for Semantics Service Identification: ‣ Assume the following service retrieves a satellite image pertaining to (x,y) with resolution respective to r Questions to ask the system: ‣ How to deduce that this service can be used? ‣ How to determine what information is needed for input? ‣ Did the user provide enough information to invoke this service? get_image(double x, double y, double r) inputsTo longitudegrid_size outputsTo satellite image

87 Indexing Services Services (inputs, outputs) are also registered in much the same way

88 Systematic Service Planning Ontology, O Compose workflows in this form: data derivation service derivation

89 Presentation Outline Motivation & Introduction Our Service Composition System: Auspice ‣ Metadata Framework ‣ Cost-Aware Service Planning ‣ Supporting Keyword Queries ‣ Caching Intermediate Results ‣ Elastic Cache Deployment Conclusion Auspice

90 Caching Intermediate Results

91 A Hierarchical Cache

92 Misses Fast Slow Hits (Slow) Wouldn’t it be faster to centralize the index on the broker node? Do we really need the broker index? Isn’t hashing faster? Cache Access Types

93 Experimental Workflows Against Heterogeneous Bandwidths

94 Centralized on Broker vs. Hierarchical Out-of-core! In-core