Elsevier

Computer Networks

Volume 56, Issue 14, 28 September 2012, Pages 3233-3246
Computer Networks

P2P as a CDN: A new service model for file sharing

https://doi.org/10.1016/j.comnet.2012.06.010Get rights and content

Abstract

Traditional content distribution networks (CDNs), such as Akamai, replicate content at thousands of servers worldwide in an attempt to bring it closer to end users. Recent years have, however, brought a surge of peer-to-peer (p2p) systems that have demonstrated the ability both to help traditional CDN operations and to effectively disseminate content as independent applications. Unfortunately, this p2p surge has created significant problems to network operators by generating large volumes of inter-AS traffic. In this paper, we demonstrate that stepping back and applying traditional CDN design principles to p2p systems can help solve these emerging problems. In particular, focusing on the BitTorrent swarming protocol, we show that our new service model can, in the common case, reduce inter-AS traffic by 45–75%. Moreover, in scenarios when ISPs are shaping inter-AS traffic, it speeds up download times by 60% for the most popular torrents.

Our approach bases on disproving the common wisdom that the current peer altruism in p2p systems (BitTorrent in particular) is insufficient. We thus abandon the common approach of deploying novel incentives for cooperation, and focus on designing methods to effectively utilize existing system resources. We show that controlled regional-based content replication, common for the traditional CDN design, can effectively achieve this goal. We implement our system and demonstrate that it effectively scales. Moreover, it is incrementally deployable and brings significant benefits in partial deployment scenarios. ISPs and network regions in which the system gets deployed can resolve their inter-AS traffic problems instantly and autonomously, i.e., without any inter-ISP collaboration and without requiring that the system gets deployed in the entire Internet.

Introduction

Peer-to-peer (p2p) applications gained enormous popularity in recent years, generating a large fraction of wide-area Internet traffic [1], [2], [3]. This has created the well-known tussle between users, who strive for fast downloads, and Internet Service Providers (ISPs), whose flat-rate pricing business model is challenged by the enormous p2p traffic volumes. As a result, many ISPs started deploying middle-boxes (e.g., [2], [4]) capable of rate-limiting or blocking p2p protocols [5], [6], [7], [8]. This has in turn created an arms race with p2p developers who started designing methods to evade such restrictions (e.g., [9], [10]).

This problem has drawn a significant attention by the networking research community (e.g., [11], [12], [13], [14], [15], [9], [16], [3]), with several proposals claiming that the tussle between p2p applications and ISPs is unnecessary, and that a “win–win” solution can be achieved via ISP-friendly p2p design. In particular, some proposals show that proximity-based peer selection (e.g., [14]), can help reduce inter-AS traffic volumes. Other proposals (e.g., [15]) argue that cooperation among p2p applications and ISPs is the way to resolve the tussle. Unfortunately, recent research results, i.e., [3], demonstrate that such a win–win outcome is unlikely.

The problem with proximity-based approaches, which we confirm in our work here as well, is that not enough local peers exist in a given region, even for globally popular files. As a result, proximity-only approaches might have limited effects on reducing p2p traffic. The problem with approaches advocating collaboration between p2p applications and ISPs (e.g., [15]) is that ISPs can have divergent needs, and often no incentives to deploy such approaches. We detail more on both approaches in the related work section.

Our key contribution lies in demonstrating that a comprehensive and feasible win–win solution to this emerging problem does exist. In particular, we propose a new service model for p2p file sharing and demonstrate that when this model is applied to BitTorrent, it significantly decreases inter-AS traffic and achieves dramatic p2p performance improvements in the common case. Specifically, we demonstrate that our data delivery approach reduces the inter-AS traffic by 45–75% in the common case. Moreover, it reduces the inter-AS traffic generated by popular torrents by up to an order of magnitude and improves their download times by 60%. Our main goal is to reduce inter-AS traffic volumes without sacrificing peer performance measured by download time.

The common wisdom is that current peer behavior is characterized by limited altruism (willingness to share a file once peers downloaded it), which fundamentally limits the system performance (e.g., [17], [18], [3], [19]). We argue that this is not the case. In particular, existing measurements, including our own, show that the majority of peers stay connected and seed a previously downloaded file several hours after completing a download [3], [19]. We argue that this current level of peer altruism is enormous. Indeed, if peers on average spend two to three times longer as seeders than as leechers (downloaders), this implies that there should be enough resources to effectively serve all peers without any contention. This is true, in the common case, even when we account for asymmetric properties of access networks that typically constrain data transfer. Thus, we argue that the problems in p2p networks, including the inter-AS one, do not arise due to lack of resources, but rather due to the fact that these resources are unmanaged.

Motivated by the effectiveness in which ‘classical’ content distribution networks (CDNs), such as Akamai [20], bring content closer to clients, we propose a system-wide change that can enable effective content replication in swarming p2p systems such as BitTorrent. In particular, our service model enables peers that become seeders to serve the content popular in their own geographic region. This is similar to the way CDNs, such as Akamai, serve content popular in given geographic areas and redirect clients to close-by servers. Thus, instead of seeding a file to a distant peer (and likely getting throttled by intermediate ISPs [5]), the key idea is to reuse precious resources locally.

It is essential to understand that our approach requires no novel incentives for users to change or improve their current sharing behavior. Indeed, users have no incentives to share a file once they download it, yet they do. Our approach simply exploits this existing level of user altruism and further advances the p2p file sharing idea to the next level. Also, one should understand that our approach requires no changes in the way users use p2p file sharing applications. They can join any swarm they like, download any file they want, and disconnect whenever they want. At the same time, our approach does not change the amount of time that users upload, just which data is uploaded to which users. Finally, because our service model enables content to be uploaded to end-user machines, even if they did not originally request it, one concern might be that this can expose users to prosecutable content (for example child pornography). This is avoided in our system via filtering and scheduling policies, as we explain in detail below.

Section snippets

CDN-based p2p data delivery

Here, we first present the key ideas standing behind the ‘traditional’ CDN approach and explain how it improved the client–server communication paradigm. We then explain how the CDN concepts can be applied in the p2p context. We demonstrate that the available resources in p2p systems are plentiful, yet the real problem behind the p2p-generated inter-AS traffic ‘radiation’ problem is unbalanced content replication, i.e., the replication level does not follow content popularity.

P2p content replication

Here, we explain how a balanced content replication approach can be applied to p2p networks. In particular, we explain replication policies, i.e., what content to replicate, where to replicate, and when to replicate? In addition, we explain the necessary mechanisms needed by our approach to achieve this goal.

Dataset

We deploy MirrorPlane monitoring system at a cluster of 30 machines, to collect the necessary real-time swarm dynamics needed to evaluate our content replication approach. We provide MirrorPlane’s implementation details in Section 5. We first provide the dataset properties, that we later feed into a simulator.

We monitor 869 torrents selected from the PirateBay [21], IsoHunt [22] and Mininova [23] websites to gather individual swarm dynamics. We run MirrorPlane and monitor these torrents for

Implementation

Here we present the implementation details of MirrorPlane. Fig. 11 shows the architecture. MirrorPlane consists of two components: (i) a monitoring component that is a modified BitTorrent client capable of monitoring fine-grained swarm dynamics for a large number of torrents, and (ii) a scheduling component that implements the torrent replication scheduling algorithm described in Section 3.2.

Note that the monitoring component is fully implemented by modifying the Bitflu BitTorrent client and

Incentives for deployment

The current dispute between p2p applications, that ‘radiate’ large amounts of inter-AS traffic, and ISPs, who deploy middle boxes to shape or deny such traffic, is bringing nothing beneficial to either of the parties. P2p applications, and BitTorrent in particular, are perceived as wild, careless, network oblivious applications that create tremendous problems to network operators [36]. ISPs that tamper with p2p-generated traffic are getting bad press [6], and draw attention of net neutrality

Related work

There has been much past work related to our content delivery approach, which we can group roughly into work on (i) reducing inter-AS p2p traffic, (ii) peer-assisted content distribution, (iii) p2p incentives and altruism, and (iv) p2p performance optimization.

Conclusions

In this paper, we proposed a CDN-based content delivery approach for p2p applications, and designed policies and mechanisms for balanced content replication in p2p networks. Contrary to common belief, we showed that even after the “download-and-depart” behavior of a subset of users there are more than enough resources in p2p systems. We have shown that proximity-only approaches are limited not because of the protocol limitations (and limited endpoint views), but because the regional replication

Amit Mondal is a software engineer at Google. Previous to that he was a Ph.D. student in the Department of Electrical Engineering and Computer Science at Northwestern University. He received his B.S. degree in computer science and engineering from Indian Institute of Technology, Kanpur, India in 2004. His research interests are in the area of computer networking with emphasis on understanding the behavior of Internet protocols in diverse and extreme scenarios and exploring vulnerabilities in

References (57)

  • P2P Still Generates Most Traffic Worldwide,...
  • Managing Peer-To-Peer Traffic With Cisco Service Control Technology,...
  • M. Piatek, H. Madhyastha, J. John, A. Krishnamurthy, T. Anderson, Pitfalls for ISP-friendly P2P design, in: HotNets...
  • PRX Traffic Manager,...
  • M. Dischinger, A. Mislove, A. Haeberlen, K. Gummadi, Detecting BitTorrent blocking, in: ACM IMC ’08,...
  • Packet Forgery by ISPs: A Report on the Comcast Affair,...
  • DslReports: Comcast is using Sandvine to manage P2P connections,...
  • EFF: ‘Test Your ISP Project’,...
  • T. Karagiannis, A. Broido, N. Brownlee, K. Claffy, M. Faloutsos, Is P2P dying or just hiding? in: GLOBECOM ’04,...
  • Prevent traffic shaping on Azureus,...
  • V. Aggarwal, A. Feldmann, ISP-Aided Neighbor Selection in P2P Systems,...
  • V. Aggarwal et al.

    Can ISPs and P2P users cooperate for improved performance? ACM Computer Comm. Review

    (2007)
  • R. Bindal, P. Cao, W. Chan, J. Medved, G. Suwala, T. Bates, A. Zhang, Improving traffic locality in BitTorrent via...
  • D. Choffnes, F. Bustamante, Taming the torrent: a practical approach to reducing cross-ISP traffic in P2P systems, in:...
  • H. Xie, R. Yang, A. Krishnamurthy, Y. Liu, A. Silberschatz, P4P: provider portal for P2P applications, in: ACM SIGCOMM...
  • T. Karagiannis, P. Rodriguez, K. Papagiannaki, Should internet service providers fear peer-assited content...
  • M. Piatek, T. Isdal, T. Anderson, A. Krishnamurthy, A. Venkataramani, Do incentives build robustness in BitTorrent? in:...
  • M. Piatek, T. Isdal, A. Krishnamurthy, T. Anderson, One hop reputations for peer to peer file sharing workloads, in:...
  • L. Guo, S. Chen, Z. Xiao, E. Tan, X. Ding, X. Zhang, Measurement, analysis, and modeling of BitTorrent-like systems,...
  • Akamai,...
  • The Pirate Bay,...
  • IsoHunt,...
  • Mininova,...
  • G. Dan, N. Carlsson, Dynamic swarm management for improved BitTorrent performance, in: IPTPS ’09,...
  • DSL vs. Cable,...
  • K. Cho, K. Fukuda, H. Esaki, The impact and implications of the growth in residential user-to-user traffic, in: ACM...
  • BT to throttle P2P for faster broadband,...
  • UK ISP to Throttle BitTorrent,...
  • Cited by (13)

    • Survey on peer-assisted content delivery networks

      2017, Computer Networks
      Citation Excerpt :

      Historically peer-to-peer systems have been ISP-oblivious and could generate significant amount of cross-ISP traffic for what the ISPs can be charged. The loss of revenue among ISPs has started a tussle between ISPs and P2P systems resulting in blocking and discouraging P2P traffic by the former [74,75]. Achieving ISP-friendly peer-to-peer traffic is thus a major challenge for commercial deployment of PA-CDNs.

    • Analysis of health professional security behaviors in a real clinical setting: An empirical study

      2015, International Journal of Medical Informatics
      Citation Excerpt :

      The results of our survey reveal that 9.4% of those surveyed have at some time sent or received PHI by e-mail, in spite of the fact that this is one of the easiest-to-follow and clearest security policies. Peer-to-peer (P2P) applications generate a large fraction of Internet traffic [68]. Downloading files in a P2P environment may not only be immoral, but also illegal [69].

    • Performance analysis of peer-to-peer networks based on two-phase service queuing theory

      2021, International Journal of Communication Networks and Distributed Systems
    • AUCB: Accelerated Upper Confidence Bound Algorithm to Improve Caching Performance in Industrial Network

      2021, IEEE Advanced Information Technology, Electronic and Automation Control Conference (IAEAC)
    View all citing articles on Scopus

    Amit Mondal is a software engineer at Google. Previous to that he was a Ph.D. student in the Department of Electrical Engineering and Computer Science at Northwestern University. He received his B.S. degree in computer science and engineering from Indian Institute of Technology, Kanpur, India in 2004. His research interests are in the area of computer networking with emphasis on understanding the behavior of Internet protocols in diverse and extreme scenarios and exploring vulnerabilities in commonly used Internet protocols.

    Ionut Trestian received his B.S. degree in Computer Science from the Technical University of Cluj- Napoca, Romania, in 2007. He is currently working toward the Ph.D. degree at Northwestern University under the supervision of Aleksandar Kuzmanvoic. His research interests include network measurement, network security, overlay networks and social networks.

    Zhen Qin is a Ph.D. student in School of Computer Science and Engineering at University of Electronic Science and Technology of China. He received a Masters Degree in from Queen Mary University and is currently a visiting student at Northwestern University.

    Aleksandar Kuzmanovic is an Associate Professor in the Department of Electrical Engineering and Computer Science at Northwestern University. He received his B.S. and M.S. degrees from the University of Belgrade, Serbia, in 1996 and 1999 respectively. He received the Ph.D. degree from Rice University in 2004. His research interests are in the area of computer networking with emphasis on design, measurements, analysis, denial-of-service resiliency, and prototype implementation of protocols and algorithms for the Internet. He received the National Science Foundation CAREER Award in 2008.

    View full text