SmartRE: An Architecture for Coordinated Network-Wide Redundancy Elimination Ashok Anand, Vyas Sekar, Aditya Akella University of Wisconsin, Madison Carnegie.

Slides:



Advertisements
Similar presentations
Software-defined networking: Change is hard Ratul Mahajan with Chi-Yao Hong, Rohan Gandhi, Xin Jin, Harry Liu, Vijay Gill, Srikanth Kandula, Mohan Nanduri,
Advertisements

REfactor-ing Content Overhearing to Improve Wireless Performance Shan-Hsiang Shen Aaron Gember Ashok Anand Aditya Akella abc 1d ab 1.
Cs/ee 143 Communication Networks Chapter 6 Internetworking Text: Walrand & Parekh, 2010 Steven Low CMS, EE, Caltech.
BY PAYEL BANDYOPADYAY WHAT AM I GOING TO DEAL ABOUT? WHAT IS AN AD-HOC NETWORK? That doesn't depend on any infrastructure (eg. Access points, routers)
The Network Layer Functions: Congestion Control
REDUNDANCY ELIMINATION AS A NETWORK-WIDE SERVICE Aditya Akella UW-Madison Shuchi Chawla Ashok Anand Chitra Muthukrishnan UW-Madison Srinivasan Seshan Vyas.
PFabric: Minimal Near-Optimal Datacenter Transport Mohammad Alizadeh Shuang Yang, Milad Sharif, Sachin Katti, Nick McKeown, Balaji Prabhakar, Scott Shenker.
Enabling Flow-level Latency Measurements across Routers in Data Centers Parmjeet Singh, Myungjin Lee Sagar Kumar, Ramana Rao Kompella.
Design and Implementation of a Consolidated Middlebox Architecture 1 Vyas SekarSylvia RatnasamyMichael ReiterNorbert Egi Guangyu Shi.
CacheCast: Eliminating Redundant Link Traffic for Single Source Multiple Destination Transfers Piotr Srebrny, Thomas Plagemann, Vera Goebel Department.
REDUNDANCY IN NETWORK TRAFFIC: FINDINGS AND IMPLICATIONS Ashok Anand Ramachandran Ramjee Chitra Muthukrishnan Microsoft Research Lab, India Aditya Akella.
Towards Virtual Routers as a Service 6th GI/ITG KuVS Workshop on “Future Internet” November 22, 2010 Hannover Zdravko Bozakov.
Packet Caches on Routers: The Implications of Universal Redundant Traffic Elimination Ashok Anand, Archit Gupta, Aditya Akella University of Wisconsin,
Packet Caches on Routers: The Implications of Universal Redundant Traffic Elimination Ashok Anand, Archit Gupta, Aditya Akella University of Wisconsin,
Rethinking NetFlow: A Case for a Coordinated “RISC” Architecture for Flow Monitoring Vyas Sekar Joint work with Mike Reiter, Hui Zhang David Andersen,
Wireless Mesh Networks 1. Architecture 2 Wireless Mesh Network A wireless mesh network (WMN) is a multi-hop wireless network that consists of mesh clients.
Traffic Engineering With Traditional IP Routing Protocols
Beneficial Caching in Mobile Ad Hoc Networks Bin Tang, Samir Das, Himanshu Gupta Computer Science Department Stony Brook University.
Jang, Donghyun 2011/4/4 1/21.
Traffic Engineering Jennifer Rexford Advanced Computer Networks Tuesdays/Thursdays 1:30pm-2:50pm.
Shadow Configurations: A Network Management Primitive Richard Alimi, Ye Wang, Y. Richard Yang Laboratory of Networked Systems Yale University.
December 20, 2004MPLS: TE and Restoration1 MPLS: Traffic Engineering and Restoration Routing Zartash Afzal Uzmi Computer Science and Engineering Lahore.
A Novel Approach for Transparent Bandwidth Conservation David Salyers, Aaron Striegel University of Notre Dame Department of Computer Science and Engineering.
In-Band Flow Establishment for End-to-End QoS in RDRN Saravanan Radhakrishnan.
CS 268: Lecture 12 (Router Design) Ion Stoica March 18, 2002.
S. Suri, M, Waldvogel, P. Warkhede CS University of Washington Profile-Based Routing: A New Framework for MPLS Traffic Engineering.
Building a Strong Foundation for a Future Internet Jennifer Rexford ’91 Computer Science Department (and Electrical Engineering and the Center for IT Policy)
1 Minimization of Network Power Consumption with Redundancy Elimination T. Khoa Phan* Joint work with: Frédéric Giroire*, Joanna Moulierac* and Frédéric.
Coordinated Sampling sans Origin-Destination Identifiers: Algorithms and Analysis Vyas Sekar, Anupam Gupta, Michael K. Reiter, Hui Zhang Carnegie Mellon.
Hash, Don’t Cache: Fast Packet Forwarding for Enterprise Edge Routers Minlan Yu Princeton University Joint work with Jennifer.
Not All Microseconds are Equal: Fine-Grained Per-Flow Measurements with Reference Latency Interpolation Myungjin Lee †, Nick Duffield‡, Ramana Rao Kompella†
NET-REPLAY: A NEW NETWORK PRIMITIVE Ashok Anand Aditya Akella University of Wisconsin, Madison.
MATE: MPLS Adaptive Traffic Engineering Anwar Elwalid, et. al. IEEE INFOCOM 2001.
Roadmap-Based End-to-End Traffic Engineering for Multi-hop Wireless Networks Mustafa O. Kilavuz Ahmet Soran Murat Yuksel University of Nevada Reno.
Redundancy elimination as a network service Aditya Akella UW-Madison.
SOAR: Simple Opportunistic Adaptive Routing Protocol for Wireless Mesh Networks Authors: Eric Rozner, Jayesh Seshadri, Yogita Ashok Mehta, Lili Qiu Published:
QoS-Driven Wireless Broadband Home Networking Based on Multihop Wireless Mesh Networks Authors : B. Park, W. Lee, S. Ahn, and Sungjoon Ahn Published :
Architecture and Algorithms for an IEEE based Multi-channel Wireless Mesh Network Ashish Raniwala, Tzi-cker Chiueh Stony Brook University Infocom2005.
Déjà Vu Switching for Multiplane NoCs NOCS’12 University of Pittsburgh Ahmed Abousamra Rami MelhemAlex Jones.
A Scalable, Commodity Data Center Network Architecture Jingyang Zhu.
RPT: Re-architecting Loss Protection for Content-Aware Networks Dongsu Han, Ashok Anand ǂ, Aditya Akella ǂ, and Srinivasan Seshan Carnegie Mellon University.
Computer Networks Performance Metrics. Performance Metrics Outline Generic Performance Metrics Network performance Measures Components of Hop and End-to-End.
Chi-Cheng Lin, Winona State University CS 313 Introduction to Computer Networking & Telecommunication Chapter 5 Network Layer.
Logical Topology Design
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
ECE466 - Tutorial Simple Delay and Throughput Analysis.
Aaron Gember, Theophilus Benson, Aditya Akella University of Wisconsin-Madison.
N. Hu (CMU)L. Li (Bell labs) Z. M. Mao. (U. Michigan) P. Steenkiste (CMU) J. Wang (AT&T) Infocom 2005 Presented By Mohammad Malli PhD student seminar Planete.
Measuring and Mitigating Web Performance Bottlenecks in Broadband Access Networks Srikanth Sundaresan, Nick Feamster (Georgia Tech) Renata Teixeira (Inria)
Piotr Srebrny 1.  Problem statement  Packet caching  Thesis claims  Contributions  Related works  Critical review of claims  Conclusions  Future.
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
CS 6401 Overlay Networks Outline Overlay networks overview Routing overlays Resilient Overlay Networks Content Distribution Networks.
Spring 2000CS 4611 Routing Outline Algorithms Scalability.
Optimization Problems in Wireless Coding Networks Alex Sprintson Computer Engineering Group Department of Electrical and Computer Engineering.
DECOR: A Distributed Coordinated Resource Monitoring System Shan-Hsiang Shen Aditya Akella.
Theophilus Benson*, Ashok Anand*, Aditya Akella*, Ming Zhang + *University of Wisconsin, Madison + Microsoft Research.
Internet Traffic Engineering Motivation: –The Fish problem, congested links. –Two properties of IP routing Destination based Local optimization TE: optimizing.
1 Scalability and Accuracy in a Large-Scale Network Emulator Nov. 12, 2003 Byung-Gon Chun.
Xin Li, Chen Qian University of Kentucky
Problem: Internet diagnostics and forensics
Architecture and Algorithms for an IEEE 802
FlowRadar: A Better NetFlow For Data Centers
CS 268: Router Design Ion Stoica February 27, 2003.
NOX: Towards an Operating System for Networks
任課教授:陳朝鈞 教授 學生:王志嘉、馬敏修
Columbia University in the city of New York
Edge computing (1) Content Distribution Networks
1 Multi-Protocol Label Switching (MPLS). 2 MPLS Overview A forwarding scheme designed to speed up IP packet forwarding (RFC 3031) Idea: use a fixed length.
Centralized Arbitration for Data Centers
Kyoungwoo Lee, Minyoung Kim, Nikil Dutt, and Nalini Venkatasubramanian
Presentation transcript:

SmartRE: An Architecture for Coordinated Network-Wide Redundancy Elimination Ashok Anand, Vyas Sekar, Aditya Akella University of Wisconsin, Madison Carnegie Mellon University 1

2 Redundancy Elimination (RE) for Increasing Network Capacity Enterprises Mobile Users Home Users Web content HTTP caches HTTP caches HTTP caches HTTP caches HTTP caches HTTP caches Data centers Wan Optimize r Other services (backup) Dedup/ archival RE: Leverage repeated transmissions Many “narrow” solutions to improve performance! Can we generalize this transparently? Benefit both users and ISPs?

In-Network RE as a Service 3 Routers keep a cache of recent pkts Routers keep a cache of recent pkts New packets get “encoded” or “compressed” w.r.t cached pkts New packets get “encoded” or “compressed” w.r.t cached pkts Encoded pkts are “decoded” or “uncompressed” downstream Encoded pkts are “decoded” or “uncompressed” downstream Key Issues: 1.Performance: Minimize Traffic FootPrint (“byte-hops”) 2. Cache Capacity: Can only provision finite DRAM 3. Processing Constraints: Enc/Dec are memory-access limited RE as a IP-layer service: Generalizes “narrow” deployments Transparent to users/apps: Democratizes benefits of RE Benefits ISPs: Better TE/Lower load

In-Network RE as a Service Hop-by-Hop (Anand et al, Sigcomm08) 4 Performance (Leverage all RE) ✔ Cache Constraints ✖ Processing Constraints ✖ Encode Decode Encode Decode Same packet encoded and decoded many times Same packet encoded and decoded many times Same packet cached many times Same packet cached many times Hop-by-hop RE is limited by encoding bottleneck Encoding: ~ 15 mem. accesses ~ 2.5 Gbps 50ns DRAM) Decoding: ~ 3-4 accesses > 10 Gbps 50ns DRAM)

5 Encode Decode Performance (Leverage all RE) ✖ Cache Constraints ✔ Processing Constraints ✔ Cannot leverage Inter-path RE Cannot leverage Inter-path RE Can leverage Intra-path RE Can leverage Intra-path RE In-Network RE as a Service At the Edge Doesn’t help ISPs (e.g., traffic engineering)

Motivating Question: How can we practically leverage the benefits of network-wide RE optimally? 6 EdgeHop-by-Hop Performance (Leverage all RE) ✖✔ Cache Constraints ✔✖ Processing Constraints ✔✖

Outline Background and Motivation High-level idea Design and Implementation Evaluation 7

SmartRE: High-level idea Don’t look at one-link-at-a-time Treat RE as a network-wide problem Cache Constraints: “Coordinated Caches” Each packet is cached only once downstream Cache Constraints: “Coordinated Caches” Each packet is cached only once downstream Processing Constraints: Ingress, Interior/Egress D ecode can occur multiple hops after encoder Performance: Network-Wide Optimization Account for traffic, routing, constraints etc. Performance: Network-Wide Optimization Account for traffic, routing, constraints etc. 8 SmartRE: Coordinated Network-wide RE

Cache Constraints Example 9 Packet arrivals: A, B, A,B Ingress can store 2pkts Interior can store 1pkt A,B B,A A,B BABBAB BABBAB After 2 nd pkt After 4 th pkt Total RE savings in network footprint (“byte hops”)? RE on first link No RE on interior RE on first link No RE on interior 2 * 1 = 2 Can we do better than this?

10 Cache Constraints Example Coordinated Caching Packet arrivals: A, B, A,B Ingress can store 2pkts Interior can store 1pkt A,B AAAAAA BBBBBB After 2 nd pkt 1 * * 3 = 5 RE for pkt A Save 2 hops RE for pkt A Save 2 hops RE for pkt B Save 3 hops RE for pkt B Save 3 hops After 4 th pkt Total RE savings in network footprint (“byte hops”)?

Dec 11 Processing Constraints Example Enc Dec Enc Dec 4 Mem Ops for Enc 2 Mem Ops for Dec 5 Enc/s 5 Dec/s 5 Enc/s 5 Dec/s 5 Enc/s 5Dec/s Total RE savings in network footprint (“byte hops”)? 5 * 6 = 30 units/s Note that even though decoders can do more work, they are limited by encoders 20 Mem Ops Enc 5 Enc/s Dec 5 D/s Enc 5 Dec/s 5 E/s Can we do better than this?

5 Dec/s 12 Processing Constraints Example: Smarter Approach 4 Mem Ops for Enc 2 Mem Ops for Dec 5 Enc/s 10 Dec/s Total RE savings in network footprint (“byte hops”)? 10*3 + 5 *2 = 40 units/s 20 Mem Ops 5 Dec/s 5 Enc/s edge core Many nodes are idle.Still does better! Good for partial deployment also Many nodes are idle.Still does better! Good for partial deployment also

Outline Background and Motivation High-level idea Design and Implementation Evaluation 13

SmartRE Overview 14 Network-Wide Optimization “Encoding Configs” To Ingresses “Encoding Configs” To NOC “Decoding Configs” To Interiors “Decoding Configs” To Interiors

Ingress/Encoder Operation 15 Encoding Config Encoding Config Packet Cache Check if this packet needs to be cached Identify candidate packets to encode Find “compressible” regions w.r.t cached packets Spring & Wetherall Sigcomm’00, Anand et al Sigcomm’08 Shim carries Info(matched pkt) MatchRegionSpec

Interior/Decoder Operation 16 Decoding Config Decoding Config Packet Cache Check if this packet needs to be cached Reconstruct “compressed” regions using reference packets Shim carries Info(matched pkt) MatchRegionSpec

Design Components 17 How do we specify coordinated caching responsibilities? Correctness: How do ingresses and interior nodes maintain cache consistency? How do ingresses identify candidate packets for encoding? What does the optimization entail?

18 How do we “coordinate” caching responsibilities across routers ? Non-overlapping hash-ranges per-path avoids redundant caching! (from cSamp, NSDI 08) [0.1,0.4] [0.7,0.9] [0.1,0.4] [0,0.3] [0.1,0.3] [0,0.1] 1.Hash (pkt.header) 2.Get path info for pkt 3.Cache if hash in range for path 1.Hash (pkt.header) 2.Get path info for pkt 3.Cache if hash in range for path

Design Components 19 How do we specify coordinated caching responsibilities? Correctness: How do ingresses and interior nodes maintain cache consistency? How do ingresses identify candidate packets for encoding? What does the optimization entail?

20 Network-wide optimization Traffic Patterns Traffic Matrix Redundancy Profile (intra + inter) Router constraints Processing (MemAccesses) Cache Size Encoding manifests Decoding manifests Objective: Max. Footprint Reduction (byte-hops) or any ISP objective (e.g., TE) Linear Program Inputs Output What does the “optimization” entail? Topology Routing Matrix Topology Map Path,HashRange

Design Components 21 How do we coordinate caching responsibilities across routers ? Correctness: How do ingresses and interior nodes maintain cache consistency? How do ingresses identify candidate packets for encoding? What does the optimization entail?

22 [0.1,0.4] [07,0.9] [0.7,0.9] [0.1,0.4] [0,0.3] [0.1,0.3] [0,0.1] How do ingresses and interior nodes maintain cache consistency? What if traffic surge on red path causes packets on black path to be evicted? Create “logical buckets” For every path-interior pair Evict only within buckets Create “logical buckets” For every path-interior pair Evict only within buckets

Network-Wide NOC RoutingRedundancy ProfileTrafficDevice Constraints SmartRE: Putting the pieces together 23 “Encoding Configs” To Ingresses “Encoding Configs” To Ingresses “Decoding Configs” To Interiors “Decoding Configs” To Interiors [0.1,0.4] [07,0.9] [0.7,0.9] [0.1,0.4] [0,0.3] [0.1,0.3] [0,0.1] Cache Consistency: Create “logical buckets” For every path-interior pair Evict only within buckets Cache Consistency: Create “logical buckets” For every path-interior pair Evict only within buckets Non-overlapping hash-ranges per-path avoids redundant caching! Candidate packets must be available on new packet’s path

Outline Background and Motivation High-level idea Design and Implementation Evaluation 24

Reduction in Network Footprint 25 SmartRE is 4-5X better than the Hop-by-Hop Approach SmartRE gets 80-90% of ideal unconstrained RE Results consistent across redundancy profiles, on synthetic traces Setup: Real traces from U.Wisc Emulated over tier-1 ISP topologies Processing constraints  MemOps & DRAM speed 2GB cache per RE device

More results … 26 Can we benefit even with partial deployment?  Even simple strategies work pretty well! What if redundancy profiles change over time?  Some “dominant” patterns which are stable  Get good performance even with dated configs

To Summarize.. RE as a network service is a promising vision – Generalizes specific deployments: benefit all users, apps, ISPs SmartRE makes this vision more practical – Look beyond link-local view; decouple encoding-decoding – Network-wide coordinated approach 4-5X better than current proposals – Works even with less-than-ideal/partial deployment Have glossed over some issues.. – Consistent configs, Decoding gaps, Packet losses, Routing dynamics Other domains: Data Center Networks, Multihop Wireless etc. 27

28 How do ingresses identify candidate packets for encoding? P1 P2 [0,0.2] [0,0.1] [0.2,0.5] [0.1,0.4] [0.5,0.7] [0.4,0.6] New [0,0.7] [0,0.6] Cached Always safe to encode w.r.t cached pkts on same path Always safe to encode w.r.t cached pkts on same path Cached If Cached in routers common to P1 and P2 i.e., Hash ε [0,0.4] Candidate packets must be available on this path Candidate packets must be available on this path

Decoding Gaps 29

Traffic Dynamics Changes in redundancy profiles – Ingress can track interior nodes.. – Ingress tracks “matches” between paths – Report back to NOC Routing dynamics – Recompute – Send signal to ingress to indicate invalid encodings 30

Packet Loss Problem with most RE, not just SmartRE – Some performance hit but not that bad Avoid encoding retransmissions – Otherwise will lead to repeated loss! 31

Memory Efficiency Example Packet arrival order: A, B, C, D, A, B, C, D Ingress link memories can store 3pkts Interior can only store 1pkt DABCDDABCD DABCDDABCD After 4 th pkt A B C D B C D A C D A B D A B C A B C D After 6 th pkt Hop-by-Hop: Total network footprint = 32-4= 28 DABCDDABCD A B C D B C D A C D A B D A B C A B C D B C D A C D A B D A B C A B C D “Coordinated”: Total network footprint = 32- ( ) =22 AAAAAAAAAA BBBBBBBBBB CCCCCCCCCC DDDDDDDDDD Can we do better than this? 32

Throughput Efficiency Example 33 10D 10E10D10E Hop-by-hop: Total savings = 5X * X * 2 = 40X 40 Mem Ops20 Mem Ops 5E 5D 5E 20D “Coordinated”: Total savings = 20X * 3 = 60X 40 Mem Ops20 Mem Ops 5E 0D Can we do better than this?

In-Network RE as a Service 34 Routers keep a cache of recent pkts Routers keep a cache of recent pkts New packets get “encoded” or “compressed” w.r.t cached pkts New packets get “encoded” or “compressed” w.r.t cached pkts Encoded pkts are “decoded” or “uncompressed” downstream Encoded pkts are “decoded” or “uncompressed” downstream In-Network RE as a Service Version1: At the Edge Key Issues: 1. Performance: Maximize Traffic FootPrint Reduction (“byte-hops”) Leverage all possible RE (e.g., inter-path) 2. Cache Capacity: Can only provision finite DRAM 3. Processing Constraints: Enc/Dec are memory-access limited Encode Decode Encode Decode Performance (Leverage all RE) ✖ Cache Constraints ✔ Processing Constraints ✔ Cannot leverage Inter-path RE Cannot leverage Inter-path RE Can leverage Intra-path RE Can leverage Intra-path RE Conceptually very appealing! Generalizes “narrow” deployments Transparent to users/apps: Democratizes benefits of RE Benefits ISPs: Better TE/Lower load

In-Network RE as a Service Version2: Hop-by-Hop (Anand et al, Sigcomm08) 35 Performance (Leverage all RE) ✔ Cache Constraints ✖ Processing Constraints ✖ Encode Decode Encode Decode Same packet encoded and decoded many times Same packet encoded and decoded many times Same packet cached many times Same packet cached many times Hop-by-hop RE is limited by encoding bottleneck Encoding: ~ 15 mem. accesses ~ 2.5 Gbps 50ns DRAM) Decoding: ~ 3-4 accesses > 10 Gbps 50ns DRAM)

Ingress/Encoder Algorithm Content store Send compressed packet Shim carries Info(matched pkt) MatchRegionSpec Check if this needs to be cached Encoding Config Encoding Config 36 Identify Candidate Packets to Encode i.e., cached along path of new packet Identify Candidate Packets to Encode i.e., cached along path of new packet Efficient Algorithms for Finding “Compressible Regions” Spring & Wetherall, Sigcomm’00 Anand et al, Sigcomm’08 Efficient Algorithms for Finding “Compressible Regions” Spring & Wetherall, Sigcomm’00 Anand et al, Sigcomm’08

Interior/Decoder Algorithm Content store Shim carries Info(matched pkt) MatchRegionSpec Check if this needs to be cached Decoding Config Decoding Config Reconstruct compressed regions Reconstruct compressed regions 37 Send uncompressed packet