Product Code Database
Example Keywords: jelly -slippers $72
   » » Wiki: Peer-to-peer
Tag Wiki 'Peer-to-peer'.

Peer-to-peer ( P2P) computing or networking is a distributed application architecture that partitions tasks or workloads between peers. Peers are equally privileged, participants in the network. This forms a peer-to-peer network of nodes.

Peers make a portion of their resources, such as processing power, disk storage or network bandwidth, directly available to other network participants, without the need for central coordination by servers or stable hosts.Rüdiger Schollmeier, A Definition of Peer-to-Peer Networking for the Classification of Peer-to-Peer Architectures and Applications, Proceedings of the First International Conference on Peer-to-Peer Computing, IEEE (2002). Peers are both suppliers and consumers of resources, in contrast to the traditional client–server model in which the consumption and supply of resources are divided.

While P2P systems had previously been used in many application domains,

(2023). 9780970284679, Intel Press. .
the architecture was popularized by the file sharing system , originally released in 1999. The concept has inspired new structures and philosophies in many areas of human interaction. In such social contexts, peer-to-peer as a meme refers to the that has emerged throughout society, enabled by technologies in general.

Historical development
While P2P systems had previously been used in many application domains, the concept was popularized by systems such as the music-sharing application (originally released in 1999). The peer-to-peer movement allowed millions of Internet users to connect "directly, forming groups and collaborating to become user-created search engines, virtual supercomputers, and filesystems".
(2023). 9780596001100, O'Reilly. .
The basic concept of peer-to-peer computing was envisioned in earlier software systems and networking discussions, reaching back to principles stated in the first Request for Comments, RFC 1.RFC 1, Host Software, S. Crocker, IETF Working Group (April 7, 1969)

's vision for the World Wide Web was close to a P2P network in that it assumed each user of the web would be an active editor and contributor, creating and linking content to form an interlinked "web" of links. The early Internet was more open than the present day, where two machines connected to the Internet could send packets to each other without firewalls and other security measures. This contrasts to the -like structure of the web as it has developed over the years.

(2023). 9781595930453
As a precursor to the Internet, was a successful peer-to-peer network where "every participating node could request and serve content". However, ARPANET was not self-organized, and it lacked the ability to "provide any means for context or content-based routing beyond 'simple' address-based routing."
(2023). 9783540291923, Springer, Berlin, Heidelberg.

Therefore, , a distributed messaging system that is often described as an early peer-to-peer architecture, was established. It was developed in 1979 as a system that enforces a decentralized model of control.Horton, Mark, and Rick Adams. "Standard for interchange of USENET messages." Https:// The basic model is a client–server model from the user or client perspective that offers a self-organizing approach to newsgroup servers. However, communicate with one another as peers to propagate Usenet news articles over the entire group of network servers. The same consideration applies to SMTP email in the sense that the core email-relaying network of mail transfer agents has a peer-to-peer character, while the periphery of and their direct connections is strictly a client-server relationship.

In May 1999, with millions more people on the Internet, introduced the music and file-sharing application called Napster. Napster was the beginning of peer-to-peer networks, as we know them today, where "participating users establish a virtual network, entirely independent from the physical network, without having to obey any administrative authorities or restrictions".

A peer-to-peer network is designed around the notion of equal peer nodes simultaneously functioning as both "clients" and "servers" to the other nodes on the network. This model of network arrangement differs from the client–server model where communication is usually to and from a central server. A typical example of a file transfer that uses the client-server model is the File Transfer Protocol (FTP) service in which the client and server programs are distinct: the clients initiate the transfer, and the servers satisfy these requests.

Routing and resource discovery
Peer-to-peer networks generally implement some form of virtual on top of the physical network topology, where the nodes in the overlay form a of the nodes in the physical network. Data is still exchanged directly over the underlying TCP/IP network, but at the application layer peers are able to communicate with each other directly, via the logical overlay links (each of which corresponds to a path through the underlying physical network). Overlays are used for indexing and peer discovery, and make the P2P system independent from the physical network topology. Based on how the nodes are linked to each other within the overlay network, and how resources are indexed and located, we can classify networks as unstructured or structured (or as a hybrid between the two).
(2023). 9781420066043, Taylor & Francis. .
(2023). 9781616928339, IGI Global. .
(2023). 9783540726050, Springer.

Unstructured networks
Unstructured peer-to-peer networks do not impose a particular structure on the overlay network by design, but rather are formed by nodes that randomly form connections to each other.
(2023). 9783642230738, Springer.
(, , and are examples of unstructured P2P protocols).
(2023). 9781466507616, CRC Press.

Because there is no structure globally imposed upon them, unstructured networks are easy to build and allow for localized optimizations to different regions of the overlay.

(2023). 9780387784489, Springer.
Also, because the role of all peers in the network is the same, unstructured networks are highly robust in the face of high rates of "churn"—that is, when large numbers of peers are frequently joining and leaving the network.
(2023). 9780387097503, Springer.
(2023). 9783540441793, Springer. .

However, the primary limitations of unstructured networks also arise from this lack of structure. In particular, when a peer wants to find a desired piece of data in the network, the search query must be flooded through the network to find as many peers as possible that share the data. Flooding causes a very high amount of signaling traffic in the network, uses more /memory (by requiring every peer to process all search queries), and does not ensure that search queries will always be resolved. Furthermore, since there is no correlation between a peer and the content managed by it, there is no guarantee that flooding will find a peer that has the desired data. Popular content is likely to be available at several peers and any peer searching for it is likely to find the same thing. But if a peer is looking for rare data shared by only a few other peers, then it is highly unlikely that the search will be successful.

(2023). 9780387097503, Springer.

Structured networks
In structured peer-to-peer networks the overlay is organized into a specific topology, and the protocol ensures that any node can efficientlyTypically approximating O(log N), where N is the number of nodes in the P2P system search the network for a file/resource, even if the resource is extremely rare.

The most common type of structured P2P networks implement a distributed hash table (DHT),Other design choices include overlay rings and d-Torus. See for example R. Ranjan, A. Harwood, and R. Buyya, "Peer-to-peer based resource discovery in global grids: a tutorial," IEEE Commun. Surv., vol. 10, no. 2. and P. Trunfio, "Peer-to-Peer resource discovery in Grids: Models and systems," Future Generation Computer Systems archive, vol. 23, no. 7, Aug. 2007. in which a variant of consistent hashing is used to assign ownership of each file to a particular peer.

(2023). 9780769515823, IEEE Computer Society. .
(2023). 9783540407249
This enables peers to search for resources on the network using a : that is, ( key, value) pairs are stored in the DHT, and any participating node can efficiently retrieve the value associated with a given key.Moni Naor and Udi Wieder. Novel Architectures for P2P Applications: the Continuous-Discrete Approach . Proc. SPAA, 2003.Gurmeet Singh Manku. Dipsea: A Modular Distributed Hash Table . Ph. D. Thesis (Stanford University), August 2004.

However, in order to route traffic efficiently through the network, nodes in a structured overlay must maintain lists of neighbors that satisfy specific criteria. This makes them less robust in networks with a high rate of churn (i.e. with large numbers of nodes frequently joining and leaving the network).

(2023). 9780387097527, Springer. .
More recent evaluation of P2P resource discovery solutions under real workloads have pointed out several issues in DHT-based solutions such as high cost of advertising/discovering resources and static and dynamic load imbalance.

Notable distributed networks that use DHTs include , an alternative to BitTorrent's distributed tracker, the , the , and the . Some prominent research projects include the Chord project, , PAST storage utility, , a self-organized and emerging overlay network, and CoopNet content distribution system.

(2012). 9781461454823, Springer. .
DHT-based networks have also been widely utilized for accomplishing efficient resource discovery for systems, as it aids in resource management and scheduling of applications.

Hybrid models
Hybrid models are a combination of peer-to-peer and client–server models.
(2023). 9783540291923, Springer.
A common hybrid model is to have a central server that helps peers find each other. was an example of a hybrid model until. There are a variety of hybrid models, all of which make trade-offs between the centralized functionality provided by a structured server/client network and the node equality afforded by the pure peer-to-peer unstructured networks. Currently, hybrid models have better performance than either pure unstructured networks or pure structured networks because certain functions, such as searching, do require a centralized functionality but benefit from the decentralized aggregation of nodes provided by unstructured networks.

CoopNet content distribution system
CoopNet (Cooperative Networking) was a proposed system for off-loading serving to peers who have recently content, proposed by computer scientists Venkata N. Padmanabhan and Kunwadee Sripanidkulchai, working at Microsoft Research and Carnegie Mellon University.
(2023). 9783540441793, Springer. .
PDF (Microsoft, with addendum) PDF (Springer, original, fee may be required)
Project home page. When a server experiences an increase in load it redirects incoming peers to other peers who have agreed to the content, thus off-loading balance from the server. All of the information is retained at the server. This system makes use of the fact that the bottleneck is most likely in the outgoing bandwidth than the , hence its server-centric design. It assigns peers to other peers who are 'close in ' to its neighbors same in an attempt to use locality. If multiple peers are found with the same it designates that the node choose the fastest of its neighbors. is transmitted by having clients the previous stream, and then transmit it piece-wise to new nodes.

Security and trust
Peer-to-peer systems pose unique challenges from a computer security perspective.

Like any other form of , P2P applications can contain vulnerabilities. What makes this particularly dangerous for P2P software, however, is that peer-to-peer applications act as servers as well as clients, meaning that they can be more vulnerable to remote exploits.

(2023). 9783642035135, Springer.

Routing attacks
Since each node plays a role in routing traffic through the network, malicious users can perform a variety of "routing attacks", or denial of service attacks. Examples of common routing attacks include "incorrect lookup routing" whereby malicious nodes deliberately forward requests incorrectly or return false results, "incorrect routing updates" where malicious nodes corrupt the routing tables of neighboring nodes by sending them false information, and "incorrect routing network partition" where when new nodes are joining they bootstrap via a malicious node, which places the new node in a partition of the network that is populated by other malicious nodes.
(2023). 9783642035135, Springer.

Corrupted data and malware
The prevalence of varies between different peer-to-peer protocols. Studies analyzing the spread of malware on P2P networks found, for example, that 63% of the answered download requests on the network contained some form of malware, whereas only 3% of the content on contained malware. In both cases, the top three most common types of malware accounted for the large majority of cases (99% in gnutella, and 65% in OpenFT). Another study analyzing traffic on the network found that 15% of the 500,000 file sample taken were infected by one or more of the 365 different that were tested for.
(2023). 9783540736134, Springer.

Corrupted data can also be distributed on P2P networks by modifying files that are already being shared on the network. For example, on the network, the managed to introduce faked chunks into downloads and downloaded files (mostly MP3 files). Files infected with the RIAA virus were unusable afterwards and contained malicious code. The RIAA is also known to have uploaded fake music and movies to P2P networks in order to deter illegal file sharing. Consequently, the P2P networks of today have seen an enormous increase of their security and file verification mechanisms. Modern , chunk verification and different encryption methods have made most networks resistant to almost any type of attack, even when major parts of the respective network have been replaced by faked or nonfunctional hosts.

Resilient and scalable computer networks
The decentralized nature of P2P networks increases robustness because it removes the single point of failure that can be inherent in a client–server based system. As nodes arrive and demand on the system increases, the total capacity of the system also increases, and the likelihood of failure decreases. If one peer on the network fails to function properly, the whole network is not compromised or damaged. In contrast, in a typical client–server architecture, clients share only their demands with the system, but not their resources. In this case, as more clients join the system, fewer resources are available to serve each client, and if the central server fails, the entire network is taken down.

Distributed storage and search
There are both advantages and disadvantages in P2P networks related to the topic of data , recovery, and availability. In a centralized network, the system administrators are the only forces controlling the availability of files being shared. If the administrators decide to no longer distribute a file, they simply have to remove it from their servers, and it will no longer be available to users. Along with leaving the users powerless in deciding what is distributed throughout the community, this makes the entire system vulnerable to threats and requests from the government and other large forces. For example, has been pressured by the RIAA, MPAA, and entertainment industry to filter out copyrighted content. Although server-client networks are able to monitor and manage content availability, they can have more stability in the availability of the content they choose to host. A client should not have trouble accessing obscure content that is being shared on a stable centralized network. P2P networks, however, are more unreliable in sharing unpopular files because sharing files in a P2P network requires that at least one node in the network has the requested data, and that node must be able to connect to the node requesting the data. This requirement is occasionally hard to meet because users may delete or stop sharing data at any point.

In a P2P network, the community of users is entirely responsible for deciding which content is available. Unpopular files eventually disappear and become unavailable as fewer people share them. Popular files, however, are highly and easily distributed. Popular files on a P2P network are more stable and available than files on central networks. In a centralized network, a simple loss of connection between the server and clients can cause a failure, but in P2P networks, the connections between every node must be lost to cause a data-sharing failure. In a centralized system, the administrators are responsible for all data recovery and backups, while in P2P systems, each node requires its backup system. Because of the lack of central authority in P2P networks, forces such as the recording industry, RIAA, MPAA, and the government are unable to delete or stop the sharing of content on P2P systems.


Content delivery
In P2P networks, clients both provide and use resources. This means that unlike client–server systems, the content-serving capacity of peer-to-peer networks can actually increase as more users begin to access the content (especially with protocols such as that require users to share, refer a performance measurement studySharma P., Bhakuni A. & Kaushal R. "Performance Analysis of BitTorrent Protocol . National Conference on Communications, 2013 ). This property is one of the major advantages of using P2P networks because it makes the setup and running costs very small for the original content distributor.
(2023). 9783540258094, Springer.

File-sharing networks
Many peer-to-peer file sharing networks, such as , G2, and the popularized peer-to-peer technologies.
  • Peer-to-peer content delivery networks.
  • Peer-to-peer content services, e.g. caches for improved performance such as Correli CachesGareth Tyson, Andreas Mauthe, Sebastian Kaune, Mu Mu and Thomas Plagemann. Corelli: A Dynamic Replication Service for Supporting Latency-Dependent Content in Community Networks. In Proc. 16th ACM/SPIE Multimedia Computing and Networking Conference (MMCN), San Jose, CA (2009).
  • Software publication and distribution (Linux distribution, several games); via networks.

Copyright infringements
Peer-to-peer networking involves data transfer from one user to another without using an intermediate server. Companies developing P2P applications have been involved in numerous legal cases, primarily in the United States, over conflicts with law.
(2023). 9780387097503, Springer.
Two major cases are vs RIAA and MGM Studios, Inc. v. Grokster, Ltd.. In the last case, the Court unanimously held that defendant peer-to-peer file sharing companies Grokster and Streamcast could be sued for inducing copyright infringement.

  • The P2PTV and PDTP protocols.
  • Some proprietary multimedia applications use a peer-to-peer network along with streaming servers to stream audio and video to their clients.
  • for multicasting streams.
  • Pennsylvania State University, MIT and Simon Fraser University are carrying on a project called designed for facilitating file sharing among educational institutions globally.
  • Osiris is a program that allows its users to create anonymous and autonomous web portals distributed via P2P network.

Other P2P applications
  • Dat, a distributed version-controlled publishing platform.
  • I2P, an used to browse the Internet anonymously.
  • Unlike the related I2P, the Tor network is not itself peer-to-peer; however, it can enable peer-to-peer applications to be built on top of it via .
  • The InterPlanetary File System (IPFS) is a protocol and network designed to create a content-addressable, peer-to-peer method of storing and sharing distribution protocol. Nodes in the IPFS network form a distributed file system.
  • Jami, a peer-to-peer chat and SIP app.
  • , a peer-to-peer protocol designed for the .
  • , a Wireless community network designed to be independent from the Internet.
  • , connection sharing application that shares Internet access with other devices using Wi-Fi or Bluetooth.
  • , a directory-syncing app.
  • Research like the Chord project, the PAST storage utility, the , and the CoopNet content distribution system.
  • , a directory-syncing app.
  • and applications that power real-time marketplaces.
  • The U.S. Department of Defense is conducting research on P2P networks as part of its modern network warfare strategy. In May, 2003, , then director of , testified that the United States military uses P2P networks.
  • is a P2P streaming in for use in , as well as in the WebTorrent Desktop stand alone version that bridges WebTorrent and serverless networks.
  • in Windows 10 uses a proprietary peer-to-peer technology called "Delivery Optimization" to deploy operating system updates using end-users PCs either on the local network or other PCs. According to Microsoft's Channel 9 it led to a 30%-50% reduction in Internet bandwidth usage.Hammerksjold Andreas; Engler, Narkis, "Delivery Optimization - a deep dive" , Channel 9, 11 October 2017, Retrieved on 4 February 2019.
  • Artisoft's was built as a peer-to-peer operating system. Machines can be both servers and workstations at the same time.
  • Hotline Communications Hotline Client was built as decentralized servers with tracker software dedicated to any type of files and still operates today.
  • are peer-to-peer-based which use .
    • List of cryptocurrencies
    • List of blockchains

Social implications

Incentivizing resource sharing and cooperation
Cooperation among a community of participants is key to the continued success of P2P systems aimed at casual human users; these reach their full potential only when large numbers of nodes contribute resources. But in current practice, P2P networks often contain large numbers of users who utilize resources shared by other nodes, but who do not share anything themselves (often referred to as the "freeloader problem"). Freeloading can have a profound impact on the network and in some cases can cause the community to collapse.Krishnan, R., Smith, M. D., Tang, Z., & Telang, R. (2004, January). The impact of free-riding on peer-to-peer networks. In System Sciences, 2004. Proceedings of the 37th Annual Hawaii International Conference on (pp. 10-pp). IEEE. In these types of networks "users have natural disincentives to cooperate because cooperation consumes their own resources and may degrade their own performance".Feldman, M., Lai, K., Stoica, I., & Chuang, J. (2004, May). Robust incentive techniques for peer-to-peer networks. In Proceedings of the 5th ACM conference on Electronic commerce (pp. 102-111). ACM. Studying the social attributes of P2P networks is challenging due to large populations of turnover, asymmetry of interest and zero-cost identity. A variety of incentive mechanisms have been implemented to encourage or even force nodes to contribute resources.
(2023). 9783642035135, Springer.

Some researchers have explored the benefits of enabling virtual communities to self-organize and introduce incentives for resource sharing and cooperation, arguing that the social aspect missing from today's P2P systems should be seen both as a goal and a means for self-organized virtual communities to be built and fostered.P. Antoniadis and B. Le Grand, "Incentives for resource sharing in self-organized communities: From economics to social psychology," Digital Information Management (ICDIM '07), 2007 Ongoing research efforts for designing effective incentive mechanisms in P2P systems, based on principles from game theory, are beginning to take on a more psychological and information-processing direction.

Privacy and anonymity
Some peer-to-peer networks (e.g. ) place a heavy emphasis on and —that is, ensuring that the contents of communications are hidden from eavesdroppers, and that the identities/locations of the participants are concealed. Public key cryptography can be used to provide , , authorization, and authentication for data/messages. and other protocols (e.g. Tarzan) can be used to provide anonymity.
(2023). 9783642035135, Springer.

Perpetrators of live streaming sexual abuse and other have used peer-to-peer platforms to carry out activities with anonymity.

Political implications

Intellectual property law and illegal sharing
Although peer-to-peer networks can be used for legitimate purposes, rights holders have targeted peer-to-peer over the involvement with sharing copyrighted material. Peer-to-peer networking involves data transfer from one user to another without using an intermediate server. Companies developing P2P applications have been involved in numerous legal cases, primarily in the United States, primarily over issues surrounding law. Two major cases are vs RIAA and MGM Studios, Inc. v. Grokster, Ltd. In both of the cases the file sharing technology was ruled to be legal as long as the developers had no ability to prevent the sharing of the copyrighted material. To establish criminal liability for the copyright infringement on peer-to-peer systems, the government must prove that the defendant infringed a copyright willingly for the purpose of personal financial gain or commercial advantage.Majoras, D. B. (2005). Peer-to-peer file-sharing technology consumer protection and competition issues. Federal Trade Commission, Retrieved from exceptions allow limited use of copyrighted material to be downloaded without acquiring permission from the rights holders. These documents are usually news reporting or under the lines of research and scholarly work. Controversies have developed over the concern of illegitimate use of peer-to-peer networks regarding public safety and national security. When a file is downloaded through a peer-to-peer network, it is impossible to know who created the file or what users are connected to the network at a given time. Trustworthiness of sources is a potential security threat that can be seen with peer-to-peer systems.The Government of the Hong Kong Special Administrative Region, (2008). Peer-to-peer network. Retrieved from website:

A study ordered by the found that illegal downloading may lead to an increase in overall video game sales because newer games charge for extra features or levels. The paper concluded that piracy had a negative financial impact on movies, music, and literature. The study relied on self-reported data about game purchases and use of illegal download sites. Pains were taken to remove effects of false and misremembered responses.

Network neutrality
Peer-to-peer applications present one of the core issues in the network neutrality controversy. Internet service providers (ISPs) have been known to throttle P2P file-sharing traffic due to its high-bandwidth usage.Janko Roettgers, 5 Ways to Test Whether your ISP throttles P2P, Compared to Web browsing, e-mail or many other uses of the internet, where data is only transferred in short intervals and relative small quantities, P2P file-sharing often consists of relatively heavy bandwidth usage due to ongoing file transfers and swarm/network coordination packets. In October 2007, , one of the largest broadband Internet providers in the United States, started blocking P2P applications such as BitTorrent. Their rationale was that P2P is mostly used to share illegal content, and their infrastructure is not designed for continuous, high-bandwidth traffic. Critics point out that P2P networking has legitimate legal uses, and that this is another way that large providers are trying to control use and content on the Internet, and direct people towards a client–server-based application architecture. The client–server model provides financial barriers-to-entry to small publishers and individuals, and can be less efficient for sharing large files. As a reaction to this bandwidth throttling, several P2P applications started implementing protocol obfuscation, such as the BitTorrent protocol encryption. Techniques for achieving "protocol obfuscation" involves removing otherwise easily identifiable properties of protocols, such as deterministic byte sequences and packet sizes, by making the data look as if it were random. The ISP's solution to the high bandwidth is P2P caching, where an ISP stores the part of files most accessed by P2P clients in order to save access to the Internet.

Current research
Researchers have used computer simulations to aid in understanding and evaluating the complex behaviors of individuals within the network. "Networking research often relies on simulation in order to test and evaluate new ideas. An important requirement of this process is that results must be reproducible so that other researchers can replicate, validate, and extend existing work."Basu, A., Fleming, S., Stanier, J., Naicken, S., Wakeman, I., & Gurbani, V. K. (2013). The state of peer-to-peer network simulators. ACM Computing Surveys, 45(4), 46. If the research cannot be reproduced, then the opportunity for further research is hindered. "Even though new simulators continue to be released, the research community tends towards only a handful of open-source simulators. The demand for features in simulators, as shown by our criteria and survey, is high. Therefore, the community should work together to get these features in open-source software. This would reduce the need for custom simulators, and hence increase repeatability and reputability of experiments."

Popular simulators that were widely used in the past are NS2, OMNeT++, SimPy, NetLogo, PlanetLab, ProtoPeer, QTM, PeerSim, ONE, P2PStrmSim, PlanetSim, GNUSim, and Bharambe.

Besides all the above stated facts, there has also been work done on ns-2 open source network simulators. One research issue related to free rider detection and punishment has been explored using ns-2 simulator here.A Bhakuni, P Sharma, R Kaushal "Free-rider detection and punishment in BitTorrent based P2P networks" , International Advanced Computing Conference, 2014.

See also

External links

Page 1 of 1
Page 1 of 1


Pages:  ..   .. 
Items:  .. 


General: Atom Feed Atom Feed  .. 
Help:  ..   .. 
Category:  ..   .. 
Media:  ..   .. 
Posts:  ..   ..   .. 


Page:  .. 
Summary:  .. 
1 Tags
10/10 Page Rank
5 Page Refs
2s Time