Redundancy elimination as a network service Aditya Akella UW-Madison.

Slides:



Advertisements
Similar presentations
CCNA3: Switching Basics and Intermediate Routing v3.0 CISCO NETWORKING ACADEMY PROGRAM Switching Concepts Introduction to Ethernet/802.3 LANs Introduction.
Advertisements

Software-defined networking: Change is hard Ratul Mahajan with Chi-Yao Hong, Rohan Gandhi, Xin Jin, Harry Liu, Vijay Gill, Srikanth Kandula, Mohan Nanduri,
A Search Memory Substrate for High Throughput and Low Power Packet Processing Sangyeun Cho, Michel Hanna and Rami Melhem Dept. of Computer Science University.
A Scalable and Reconfigurable Search Memory Substrate for High Throughput Packet Processing Sangyeun Cho and Rami Melhem Dept. of Computer Science University.
COS 461 Fall 1997 Routing COS 461 Fall 1997 Typical Structure.
REfactor-ing Content Overhearing to Improve Wireless Performance Shan-Hsiang Shen Aaron Gember Ashok Anand Aditya Akella abc 1d ab 1.
REDUNDANCY ELIMINATION AS A NETWORK-WIDE SERVICE Aditya Akella UW-Madison Shuchi Chawla Ashok Anand Chitra Muthukrishnan UW-Madison Srinivasan Seshan Vyas.
SmartRE: An Architecture for Coordinated Network-Wide Redundancy Elimination Ashok Anand, Vyas Sekar, Aditya Akella University of Wisconsin, Madison Carnegie.
1 EL736 Communications Networks II: Design and Algorithms Class3: Network Design Modeling Yong Liu 09/19/2007.
CacheCast: Eliminating Redundant Link Traffic for Single Source Multiple Destination Transfers Piotr Srebrny, Thomas Plagemann, Vera Goebel Department.
REDUNDANCY IN NETWORK TRAFFIC: FINDINGS AND IMPLICATIONS Ashok Anand Ramachandran Ramjee Chitra Muthukrishnan Microsoft Research Lab, India Aditya Akella.
Packet Caches on Routers: The Implications of Universal Redundant Traffic Elimination Ashok Anand, Archit Gupta, Aditya Akella University of Wisconsin,
Packet Caches on Routers: The Implications of Universal Redundant Traffic Elimination Ashok Anand, Archit Gupta, Aditya Akella University of Wisconsin,
Prentice HallHigh Performance TCP/IP Networking, Hassan-Jain Chapter 10 TCP/IP Performance over Asymmetric Networks.
Hash-Based IP Traceback Best Student Paper ACM SIGCOMM’01.
1 Improving the Performance of Distributed Applications Using Active Networks Mohamed M. Hefeeda 4/28/1999.
Traffic Engineering With Traditional IP Routing Protocols
Shadow Configurations: A Network Management Primitive Richard Alimi, Ye Wang, Y. Richard Yang Laboratory of Networked Systems Yale University.
10 - Network Layer. Network layer r transport segment from sending to receiving host r on sending side encapsulates segments into datagrams r on rcving.
A Novel Approach for Transparent Bandwidth Conservation David Salyers, Aaron Striegel University of Notre Dame Department of Computer Science and Engineering.
Introduction to Management Information Systems Chapter 5 Data Communications and Internet Technology HTM 304 Fall 07.
7DS: Node Cooperation in Mostly Disconnected Networks Henning Schulzrinne (joint work with Arezu Moghadan, Maria Papadopouli, Suman Srinivasan and Andy.
Tradeoffs in CDN Designs for Throughput Oriented Traffic Minlan Yu University of Southern California 1 Joint work with Wenjie Jiang, Haoyuan Li, and Ion.
1 Minimization of Network Power Consumption with Redundancy Elimination T. Khoa Phan* Joint work with: Frédéric Giroire*, Joanna Moulierac* and Frédéric.
Coordinated Sampling sans Origin-Destination Identifiers: Algorithms and Analysis Vyas Sekar, Anupam Gupta, Michael K. Reiter, Hui Zhang Carnegie Mellon.
Hash, Don’t Cache: Fast Packet Forwarding for Enterprise Edge Routers Minlan Yu Princeton University Joint work with Jennifer.
Not All Microseconds are Equal: Fine-Grained Per-Flow Measurements with Reference Latency Interpolation Myungjin Lee †, Nick Duffield‡, Ramana Rao Kompella†
Computer Networks Layering and Routing Dina Katabi
NET-REPLAY: A NEW NETWORK PRIMITIVE Ashok Anand Aditya Akella University of Wisconsin, Madison.
Distributing Content Simplifies ISP Traffic Engineering Abhigyan Sharma* Arun Venkataramani* Ramesh Sitaraman*~ *University of Massachusetts Amherst ~Akamai.
1 Proceeding the Second Exercises on Computer and Systems Engineering Professor OKAMURA Laboratory. Othman Othman M.M.
Chapter 4. After completion of this chapter, you should be able to: Explain “what is the Internet? And how we connect to the Internet using an ISP. Explain.
CS An Overlay Routing Scheme For Moving Large Files Su Zhang Kai Xu.
Copyright © 2002 Pearson Education, Inc. Slide 3-1 CHAPTER 3 Created by, David Zolzer, Northwestern State University—Louisiana The Internet and World Wide.
1 BitHoc: BitTorrent for wireless ad hoc networks Jointly with: Chadi Barakat Jayeoung Choi Anwar Al Hamra Thierry Turletti EPI PLANETE 28/02/2008 MAESTRO/PLANETE.
David G. Andersen CMU Guohui Wang, T. S. Eugene Ng Rice Michael Kaminsky, Dina Papagiannaki, Michael A. Kozuch, Michael Ryan Intel Labs Pittsburgh 1 c-Through:
By Sylvia Ratnasamy, Andrey Ermolinskiy, Scott Shenker Presented by Fei Jia Revisiting IP Multicast.
An Efficient Approach for Content Delivery in Overlay Networks Mohammad Malli Chadi Barakat, Walid Dabbous Planete Project To appear in proceedings of.
RPT: Re-architecting Loss Protection for Content-Aware Networks Dongsu Han, Ashok Anand ǂ, Aditya Akella ǂ, and Srinivasan Seshan Carnegie Mellon University.
Aditya Akella The Performance Benefits of Multihoming Aditya Akella CMU With Bruce Maggs, Srini Seshan, Anees Shaikh and Ramesh Sitaraman.
EndRE: An End-System Redundancy Elimination Service Bhavish Aggarwal, Aditya Akella, Ashok Anand, Athula Balachandran, Pushkar Chitnis, Chitra Muthukrishnan,
Kiew-Hong Chua a.k.a Francis Computer Network Presentation 12/5/00.
Jennifer Rexford Princeton University MW 11:00am-12:20pm Measurement COS 597E: Software Defined Networking.
COP 5611 Operating Systems Spring 2010 Dan C. Marinescu Office: HEC 439 B Office hours: M-Wd 2:00-3:00 PM.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
Networking Fundamentals. Basics Network – collection of nodes and links that cooperate for communication Nodes – computer systems –Internal (routers,
Intradomain Traffic Engineering By Behzad Akbari These slides are based in part upon slides of J. Rexford (Princeton university)
Forwarding.
Distributed Denial-of-Service Attack Detection (and Mitigation?) Mukesh Agarwal, Aditya Akella, Ashwin Bharambe.
High-Speed Policy-Based Packet Forwarding Using Efficient Multi-dimensional Range Matching Lakshman and Stiliadis ACM SIGCOMM 98.
March 2001 CBCB The Holy Grail: Media on Demand over Multicast Doron Rajwan CTO Bandwiz.
Piotr Srebrny 1.  Problem statement  Packet caching  Thesis claims  Contributions  Related works  Critical review of claims  Conclusions  Future.
1 IEX8175 RF Electronics Avo Ots telekommunikatsiooni õppetool, TTÜ raadio- ja sidetehnika inst.
Performance Limitations of ADSL Users: A Case Study Matti Siekkinen, University of Oslo Denis Collange, France Télécom R&D Guillaume Urvoy-Keller, Ernst.
DECOR: A Distributed Coordinated Resource Monitoring System Shan-Hsiang Shen Aditya Akella.
Theophilus Benson*, Ashok Anand*, Aditya Akella*, Ming Zhang + *University of Wisconsin, Madison + Microsoft Research.
#16 Application Measurement Presentation by Bobin John.
Multi-protocol Label Switching
WAN Technologies. 2 Large Spans and Wide Area Networks MAN networks: Have not been commercially successful.
P4P: Proactive Provider Assistance for P2P Haiyong Xie Yale University.
Network layer (addendum) Slides adapted from material by Nick McKeown and Kevin Lai.
CIS 700-5: The Design and Implementation of Cloud Networks
Introduction Wireless devices offering IP connectivity
University of Maryland College Park
Mohammad Malli Chadi Barakat, Walid Dabbous Alcatel meeting
NOX: Towards an Operating System for Networks
A Comparison of Overlay Routing and Multihoming Route Control
Kalyan Boggavarapu Lehigh University
Bridges and Extended LANs
CS 6290 Many-core & Interconnect
Presentation transcript:

Redundancy elimination as a network service Aditya Akella UW-Madison

Growing traffic vs. network performance  Network traffic volumes growing rapidly  Annual growth: overall (45%), enterprise (50%), data center (125%), mobile (125%)*  Growing strain on installed capacity everywhere  Core (Asian ISPs – 80-90% core utilization), enterprise access, data center, cellular, wireless…  How to sustain robust network performance? * Interview with Cisco CEO, Aug 2007, Network world Enterprises Mobile users Home users Video Data centers Web content Other svcs (backup) ISP core Strain on installed link capacities 2

Enterprises Scale link capacities by eliminating redundancy Popular idea: duplicate suppression or redundancy elimination (RE) – Popular objects, partial content matches, backups, app headers – Effective capacity improves ~2X Many approaches to RE – Application-layer caches – Protocol-independent redundancy elimination (RE) Below app-layer WAN accelerators, de-duplication – Content distribution, bittorrent Point solutions  apply to specific link, protocol, or app Mobile users Home users Video Data centers Web content Other svcs (backup) Wan Opt Dedup/ archival ISP HTTP cache ISP HTTP cache CDN 3

Universal need to scale capacities 4 Wan Opt Dedup/ archival ISP HTTP cache ISP HTTP cache Point solutions inadequate Architectural support to address universal need to scale capacities? Bittorrent ✗ Point solutions: Little or no benefit in the core ✗ Point solutions: Little or no benefit in the core ✗ Point solutions: Other links must re-implement specific RE mechanisms ✗ Point solutions: Other links must re-implement specific RE mechanisms ✗ Point solutions: Only benefit system/app attached

Internet2 Packet cache at every router 5 Wisconsin Berkeley CMU Router upstream removes redundant bytes Router downstream reconstructs full packet Router upstream removes redundant bytes Router downstream reconstructs full packet IP layer RE using router packet caches Leverage rapidly declining storage/memory costs

RE as a network service: Why? Improved performance everywhere even if partially enabled – Generalizes point deployments and app-specific approaches Benefits all network end-points, applications, scales capacities universally – Benefits network core Improved switching capacity, responsiveness to sudden overload Other application domains: data centers, multi-hop wireless Architectural benefits – Enables new protocols and apps Min-entropy routing, RE-aware traffic engineering (intra- and inter-domain) Anomaly detection, in-network spam filtering – Improves apps: need not worry about using network efficiently App headers can be verbose  better diagnostics Controlling duplicate transmission in app-layer multicast is a non-issue 6

Internet2 7 Network RE  12 pkts (ignoring tiny packets) Network RE  12 pkts (ignoring tiny packets) Without RE  18 pkts 33% lower Without RE  18 pkts 33% lower Wisconsin Berkeley CMU Generalizes point deployments Benefits the network: improves effective switching capacity 6  2 packets 3  2 packets Implications example: Performance benefits

Wisconsin Internet2 8 RE + routing  10 pkts RE + routing  10 pkts Simple RE  12 pkts Berkeley CMU ✓ Verbose control messages ✓ New video adaptation algorithms ✓ Anomaly detectors ✓ Spam filtering ✓ Content distribution schemes ✓ Verbose control messages ✓ New video adaptation algorithms ✓ Anomaly detectors ✓ Spam filtering ✓ Content distribution schemes ✓ Minimum-entropy routing ✓ New, flexible traffic engineering mechanisms ✓ Inter-domain protocols ✓ Minimum-entropy routing ✓ New, flexible traffic engineering mechanisms ✓ Inter-domain protocols Implications example: New protocols

Talk outline Is there promise today? Empirical study of redundancy in network traffic – Extent, patterns – Implications for network RE Is an IP-level RE service achievable today? Network-wide RE architecture – Getting RE to work on ISP routers What next? Summary and future directions 9

Redundancy in network traffic * 10 * Joint work with: Ashok Anand, Chitra Muthukrishnan (UW-Madison) Ram Ramjee (MSR-India)

Empirical study of RE Upstream cache = content table + fingerprint index – RE algorithms Content-based names (“fingerprints”) for chunks of bytes in payload Fingerprints computed for content, looked up to identify redundancy – Downstream cache: content table Questions – How far are existing RE algorithms from optimal? Do better schemes exist? – Fundamental redundancy patterns and implications for packet caches Cache WAN link Data center Enterprise 11

Analysis approach 11 Enterprises (3 TB) – Small (10-50 IPs) – Medium ( IPs) – Large (100+ IPs) – Protocol composition HTTP (20-55%), File sharing (25- 70%) University link (1.6 TB) – Large university trace (10K IPs) – Outgoing /24, web server traffic – Protocol composition Incoming, HTTP 60% Outgoing, HTTP 36% 12  Emulate memory-bound (100 MB - 4GB) WAN optimizer  Emulate only redundancy elimination  Compute bandwidth savings as (saved bytes/total bytes)  Includes packet headers in total bytes  Includes overhead of shim headers used for encoding Packet traces Set-up

RE algorithms: MODP Spring et al. [Sigcomm 2000] Compute fingerprints Packet payload Window (w) Rabin fingerprinting Value sampling: sample those fingerprints whose value is 0 mod p Fingerprint table Packet store Payload-1 Payload-2 Lookup fingerprints in fingerprint table, derive maximal match across packets

RE algorithms: MAXP Similar to MODP More robust selection criteria 14 MAXP Choose fingerprints that are local maxima for p-byte region MODP Sample those fingerprints whose value is 0 mod p No fingerprint to represent the shaded region Gives uniform selection of fingerprints

Comparison of RE algorithms Trace is 68% redundant! MAXP outperforms MODP by 5-10% in most cases – Uniform sampling approach of MAXP – MODP loses due to non uniform clustering of fingerprints 15 (Store all FPs in a Bloom filter)

Comparison of RE algorithms GZIP offers 3-15% benefit – (10ms buffering) -> GZIP better by 5% MAXP significantly outperforms GZIP, offers 15-60% bandwidth savings – MAXP -> (10 ms) -> GZIP better by up to 8% 16

Zipf-like distribution for chunk matches Unique chunk matches sorted by their hit counts – Zip-fian distribution, slope = – Popular chunks are content fragments < 150B in size 17 80% of savings come from 20% of chunks Need to index 80% of chunks for remaining 20% of savings Small cache size should capture most benefits?

Cache size 18 Small caches can provide significant savings Diminishing returns for increasing cache size after 250 MB – Build packet caches today using DRAM on routers

Empirical study: Summary Significant redundancy in network traffic Careful selection of content fingerprints necessary Zipf distribution of content popularity; small matches important Relatively small caches sufficient 19

SmartRE: Effective router-level RE * 20 * Joint work with Ashok Anand (UW-Madison) and Vyas Sekar (CMU)

Realizing RE as a network service Building blocks to realize network RE service? Goal: optimal performance, or, maximal reduction in traffic footprint 1.Leverage all possible RE (e.g., inter-path) 2.Leverage resources optimally a)Cache capacity: finite memory (DRAM) on routers b)Processing constraints: enc/dec are memory-access limited – can only run at a certain maximum speed 21

Hop-by-hop RE revisited 22 Encode Decode Encode Decode Same packet encoded and decoded many times Same packet encoded and decoded many times Same packet cached many times Same packet cached many times Limited throughput: Encoding: ~15 mem. accesses ~2.5 Gbps 50ns DRAM) Decoding: ~3-4 accesses > 10 Gbps 50ns DRAM) Limited throughput: Encoding: ~15 mem. accesses ~2.5 Gbps 50ns DRAM) Decoding: ~3-4 accesses > 10 Gbps 50ns DRAM)

RE at the network edge 23 Encode Decode Encode Decode Cannot leverage Inter-path RE Cannot leverage Inter-path RE Can leverage Intra-path RE Can leverage Intra-path RE

How can we practically leverage the benefits of network-wide RE optimally? SmartRE: Motivating question 24 EdgeHop-by-Hop Leverage all RE ✖✔ Cache Constraints ✔✖ Processing Constraints ✔✖

SmartRE: Key ideas 25 Don’t look at one-link-at-a-time – Treat RE as a network-wide problem Cache Constraints: Routers coordinate caching; Each packet is cached only once downstream Cache Constraints: Routers coordinate caching; Each packet is cached only once downstream Processing Constraints: Ingress, Interior/Egress; Decode can occur multiple hops after encoder High Performance: Network-Wide Optimization; Account for traffic, routing, constraints etc. High Performance: Network-Wide Optimization; Account for traffic, routing, constraints etc.

Cache constraints 26 Packet arrivals: A, B, A,B Ingress can store 2pkts Interior can store 1pkt A,B B,A A,B BABBAB BABBAB After 2 nd pkt After 4 th pkt Total RE savings in network footprint? RE on first link No RE on interior RE on first link No RE on interior 2 * 1 = 2 Can we do better than this?

Coordinated caching 27 Packet arrivals: A, B, A,B Ingress can store 2pkts Interior can store 1pkt A,B AAAAAA BBBBBB After 2 nd pkt 1 * * 3 = 5 RE for pkt A Save 2 hops RE for pkt A Save 2 hops RE for pkt B Save 3 hops RE for pkt B Save 3 hops After 4 th pkt Total RE savings in network footprint?

Processing constraints 28 Dec Enc Dec Enc Dec 4 Mem Ops for Enc 2 Mem Ops for Dec 5 Enc/s 5 Dec/s 5 Enc/s 5 Dec/s 5 Enc/s 5Dec/s Total RE savings in network footprint? 5 * 5 = 25 units/s Note that even though decoders can do more work, they are limited by encoders 20 Mem Ops Enc 5 Enc/s 5 Dec/s Can we do better than this?

Coordinating processing 29 5 Dec/s 4 Mem Ops for Enc 2 Mem Ops for Dec 5 Enc/s 10 Dec/s Total RE savings in network footprint? 10*3 + 5*2 = 40 units/s 20 Mem Ops 5 Dec/s 5 Enc/s edge core Many nodes are idle. Still does better! Good for partial deployment also Many nodes are idle. Still does better! Good for partial deployment also

SmartRE system 30 Network-Wide Optimization “Encoding Configs” To Ingresses “Encoding Configs” To Ingresses “Decoding Configs” To Interiors “Decoding Configs” To Interiors

Ingress/Encoder Algorithm 31 Content store Send compressed packet Shim carries Info(matched pkt) MatchRegionSpec Check if this needs to be cached Encoding Config Encoding Config Identify Candidate Packets to Encode i.e., cached along path of new packet Identify Candidate Packets to Encode i.e., cached along path of new packet MAXP/MODP to find maximal compressible regions

Interior/Decoder Algorithm 32 Content store Shim carries Info(matched pkt) MatchRegionSpec Check if this needs to be cached Decoding Config Decoding Config Reconstruct compressed regions Reconstruct compressed regions Send uncompressed packet

Coordinating caching 33 Non-overlapping hash-ranges per-path avoids redundant caching! (from cSamp, NSDI 08) [0.1,0.4] [0.7,0.9] [0.1,0.4] 1.Hash (pkt.header) 2.Cache if hash in range 1.Hash (pkt.header) 2.Cache if hash in range [0,0.3] [0.1,0.3] [0,0.1] 1.Hash (pkt.header) 2.Get path info for pkt 3.Cache if hash in range for path 1.Hash (pkt.header) 2.Get path info for pkt 3.Cache if hash in range for path 33

Network-wide optimization 34 Network-wide optimization Traffic Patterns Traffic Matrix Redundancy Profile (intra + inter) Router constraints Processing (MemAccesses) Cache Size Encoding manifests Decoding manifests Objective: Max. footprint reduction or any network objective (e.g., TE)? Linear Program Inputs Output Topology Routing Matrix Topology Map Path, HashRange

Cache consistency 35 [0.1,0.4] [07,0.9] [0.7,0.9] [0.1,0.4] [0,0.3] [0.1,0.3] [0,0.1] What if traffic surge on red path causes packets on black path to be evicted? Create “logical buckets” for every path-interior pair; Evict only within buckets Create “logical buckets” for every path-interior pair; Evict only within buckets

P1 P2 [0,0.2] [0,0.1] [0.2,0.5] [0.1,0.4] [0.5,0.7] [0.4,0.6] New [0,0.7] [0,0.6] Cached Always safe to encode w.r.t cached pkts on same path Always safe to encode w.r.t cached pkts on same path Cached If cached in routers common to P1 and P2 i.e., Hash ε [0,0.4] Candidate packets must be available on this path Candidate packets must be available on this path Valid encodings 36

Network-Wide NOC/ central controller RoutingRedundancy ProfileTrafficDevice Constraints 37 “Encoding Configs” To Ingresses “Encoding Configs” To Ingresses “Decoding Configs” To Interiors “Decoding Configs” To Interiors [0.1,0.4] [07,0.9] [0.7,0.9] [0.1,0.4] [0,0.3] [0.1,0.3] [0,0.1] Cache Consistency: Create “logical buckets” For every path-interior pair Evict only within buckets Cache Consistency: Create “logical buckets” For every path-interior pair Evict only within buckets Non-overlapping hash-ranges per-path to avoid redundant caching Candidate packets must be available on new packet’s path

Results: Performance benchmarks Network#PoPsTime(s)#RoutersTime(s) Level Sprint Telstra # Match regions RedundancyThroughputThroughput (w/o overhead) 124%4.9Gbps8.7Gbps 232%4.5Gbps7.9Gbps 335%4.3Gbps7.7Gbps 435%4.3Gbps7.6Gbps Encoding: 2.2Gbps (w/o click overhead) Encoders  OC48; Decoders  OC192; Amenable to partial deployment For faster links: Fraction of traffic left unprocessed, to be acted on by other encoders/decoders  not optimal Encoders  OC48; Decoders  OC192; Amenable to partial deployment For faster links: Fraction of traffic left unprocessed, to be acted on by other encoders/decoders  not optimal

Network-wide benefits (ISPs) 39 SmartRE is 4-5X better than the hop-by-hop approach SmartRE gets 80-90% of ideal unconstrained RE Results consistent across redundancy profiles, on synthetic traces Setup: Real traces from U. Wisc Emulated over tier-1 ISP topologies Processing constraints  MemOps & DRAM speed 2GB cache per RE device

SmartRE: Other results 40 Can we benefit even with partial deployment?  Even simple strategies work pretty well! What if redundancy profiles change over time?  Some “dominant” patterns which are stable (for ISPs)  Get good performance even with dated configs

Summary and future directions RE service to scale link capacities everywhere Architectural niceties and performance benefits Substantial redundancy in traffic; High speed router RE seems feasible Future directions – End-host participation – Role of different memory technologies – DRAM, flash and PCM – Role of RE (+ OpenFlow) in energy management – TCP interactions with RE 41

Architectural implications: Enhancement to routing 42

Network RE: Impact on routing – RE-aware optimization – E.g.: minimize network-wide “traffic footprint” Traffic footprint on a link = latency * unique bytes on link – Min entropy routing – Or, control utilization of expensive cross-country links – Or, route all redundant traffic on low-capacity links 43 I1 R2R2 R3R3 R4R4 Network Operations Center Redundancy profile TM + Policy Redundancy Profile RE-aware routes ISP

Network-wide utilization Routed enterprise trace over Sprint topology (AS1239) – Average redundancy is 50% – 1GB cache per router Each point  reduction in network utilization with traffic entering at a given city RE: 12-35% reduction RE + Routing: 20-45% – Effectiveness depends on topology and redundancy profile % 20-45%

Responsiveness to flash crowds Volume increases at one of the border routers – Redundancy: 20%  50% – Inter-OD redundancy: 0.5  0.75 – Stale routes used RE, RE + Routing absorb sudden changes Staleness has little impact 45

End-hosts Network RE has limitations – Does not extend benefits into end-hosts Crucial for last hop links such as cellular, wireless links Energy savings, along with bandwidth and latency – Network-RE useless with encrypted traffic Some participation from end-hosts necessary – End-to-end RE (can be IP or higher-layer) Challenges and issues – Compression/decompression overhead on constrained end-points – Co-existence with network-RE: end-hosts signaling what network should/should not store? 46

Toward universal RE 47 Wan Opt Dedup/ archival Multicast ✗ Point solutions: No benefits in the core ✗ Point solutions: No benefits in the core ✗ Point solutions: Other links must re-implement specific RE mechanisms ✗ Point solutions: Other links must re-implement specific RE mechanisms ISP HTTP cache ISP HTTP cache Multiple point solutions for RE today Universal need to scale capacities? ✗ Point solutions: Only benefit system/app attached

Enterprises Traffic growth: Sustaining network performance  Network traffic growing rapidly  Annual growth: Enterprise (50%), backbone (45%), mobile settings (125%)  Strain on installed capacity: sustain network performance?  Core, enterprise/data center, last hop wireless links A key idea: leverage duplication – Popular objects, partial content matches, file backups, application headers,… Identify, remove data redundancies 48 ISPs Mobile users Home users Web content Data centers Video Other svcs (backup)

Toward universal RE 49 Wan Opt Dedup/ archival Multicast ✗ Point solutions: No benefits in the core ✗ Point solutions: No benefits in the core ✗ Point solutions: Other links must re-implement specific RE mechanisms ✗ Point solutions: Other links must re-implement specific RE mechanisms ISP HTTP cache ISP HTTP cache Ad hoc point deployments for RE at network edge ✗ Point solutions: Only benefit system/app attached Architectural support to address universal need to scale capacities? Implications?

Redundancy elimination (RE): Many solutions Application-layer approaches – E.g., Web and P2P caches, proxies – Protocol-specific; miss sub-object redundancies; dynamic objects Protocol-independent redundancy elimination elimination – Operate below app layer: remove duplicate bytes from any network flow – E.g.: WAN optimizers, de-duplication or single-copy storage – Much more effective and general than app-layer techniques Other approaches with similar goals – CDNs, peer-assisted download (e.g., Bittorrent), multicast protocols – Clean-slate: Data Oriented Transport and Similarity Enhanced Transfers All are point solutions  specific link, system, application or protocol 50

How? Ideas from WAN optimization Network must examine byte streams, remove duplicates, reinsert Building blocks from WAN optimizers: RE agnostic to application, ports or flow semantics Upstream cache = content table + fingerprint index – Fingerprint index: content-based names for chunks of bytes in payload – Fingerprints computed for content, looked up to identify redundant byte- strings – Downstream cache: content table 51 Cache WAN link Data center Enterprise

Growing traffic vs. network performance  Network traffic volumes growing rapidly  Annual growth: overall (45%), enterprise (50%), mobile (125%)*  Growing strain on installed capacity everywhere  Core, enterprise access, data center, cellular, wireless…  How to sustain robust network performance? 52 * Interview with Cisco CEO, Aug 2007, Network world Enterprises ISPs Mobile users Home users Video Data centers Web content Other svcs (backup)