Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2006, Computing Research Repository - CORR
Decentralized search aims to find the target node in a large network by using only local information. The applications of it include peer-to-peer file sharing, web search and anything else that requires locating a specific target in a complex system. In this paper, we examine the degree-based decentralized search method. Specifically, we evaluate the efficiency of the method in different cases with different amounts of available local information. In addition, we propose a simple refinement algorithm for significantly shortening the length of the route that has been found. Some insights useful for the future developments of efficient decentralized search schemes have been achieved. Index Terms—Decentralized search, complex network, scale-free network.
2005
Abstract We propose a new algorithm for finding a target node in a network whose topology is known only locally. We formulate this task as a problem of decision making under uncertainty and use the statistical properties of the graph to guide this decision. This formulation uses the homophily and degree structure of the network simultaneously, differentiating our algorithm from those previously proposed in the literature.
Knowledge and Information Systems, 2014
Borrowing from concepts in expander graphs, we study the expansion properties of real-world, complex networks (e.g. social networks, unstructured peer-to-peer or P2P networks) and the extent to which these properties can be exploited to understand and address the problem of decentralized search. We first produce samples that concisely capture the overall expansion properties of an entire network, which we collectively refer to as the expansion signature. Using these signatures, we find a correspondence between the magnitude of maximum expansion and the extent to which a network can be efficiently searched. We further find evidence that standard graph-theoretic measures, such as average path length, fail to fully explain the level of "searchability" or ease of information diffusion and dissemination in a network. Finally, we demonstrate that this high expansion can be leveraged to facilitate decentralized search in networks and show that an expansion-based search strategy outperforms typical search methods.
Internet Mathematics, 2008
We study a general framework for decentralized search in random graphs. Our main focus is on deterministic memoryless search algorithms that use only local information to reach their destination in a bounded number of steps in expectation. This class includes (with small modifications) the search algorithms used in Kleinberg's pioneering work on long-range percolation graphs and hierarchical network models. We give a characterization of searchable graphs in this model, and use this characterization to prove a monotonicity property for searchability.
Pdpta, 2006
Scalability in a peer-to-peer network is a challenging problem. Unstructured peer-to-peer networks inherently lack scalability, and structured networks are inefficient for a high churn rate. In this paper, we present a scalable search algorithm for a decentralized unstructured peer-to-peer network using a method to dynamically determine the number of nodes to forward a query to at once. The decision is based on the degree to which each neighbor has contributed to previous successful searches. The algorithm automatically creates a spanning graph of the high traffic links. Once a stable spanning graph is created, a query tends to travel along the edges of the spanning graph. This way, the number of hops required for a search is roughly bound by the diameter of the spanning graph. The simulation shows that our algorithm demonstrates significantly better performance in terms of the number of messages generated and hops required for a search over other popular algorithms.
Proceedings of the eleventh international conference on Information and knowledge management - CIKM '02, 2002
One important problem in peer-to-peer (P2P) networks is searching and retrieving the correct information. However, existing searching mechanisms in pure peer-to-peer networks are inefficient due to the decentralized nature of such networks. We propose two mechanisms for information retrieval in pure peer-to-peer networks. The first, the modified Breadth-First-Search (BFS) mechanism, is an extension of the current Gnuttela protocol, allows searching with keywords, and is designed to minimize the number of messages that are needed to search the network. The second, the Intelligent Search mechanism, uses the past behavior of the P2P network to further improve the scalability of the search procedure. In this algorithm, each peer autonomously decides which of its peers are most likely to answer a given query. The algorithm is entirely distributed, and therefore scales well with the size of the network. We implemented our mechanisms as middleware platforms. To show the advantages of our mechanisms we present experimental results using the middleware implementation.
Computer Communications, 2008
We propose a P2P search solution, called EZSearch, that enables efficient multidimensional search for remotely located contents that best match the search criteria. EZSearch is a hierarchical approach; it organizes the network into a hierarchy in a way fundamentally different from existing search techniques. EZSearch is based on Zigzag, a P2P overlay architecture known for its scalability and robustness under network growth and dynamics. The indexing architecture of EZSearch is built on top of the Zigzag hierarchy, that allows both k-nearest-neighbor and range queries to be answered with low search overhead and worst-case search time logarithmic with the network size. The indices are fairly distributed over a small number of nodes at a modest cost for index storage and update. The performance results of EZSearch drawn from our performance study are encouraging.
Lecture Notes in Computer Science, 2009
We present a decentralized self-X architecture for distributed neighborhood based search problems using an overlay network based on random graphs. This approach provides a scalable and robust architecture with low requirements for bandwidth and computational power as well as an adequate neighborhood topology, e.g. for several instances of parallel local search and distributed learning. Together with an adapted load balancing schema our architecture is self-organizing, self-healing and self-optimizing.
Europhysics Letters (EPL), 2006
The ability to perform an efficient search in a complex network is of great importance in real-world systems. We suggest a method for searching for nodes when the source does not possess full information about the shortest path to the destination. By assigning new short names to nodes we are able to reduce significantly the amount of information stored at the nodes, such that the required memory needed scales only logarithmically with the network size; yet we succeed in finding the destination node through paths very close in distance to the shortest ones. The method is shown to perform particularly well on scale-free networks, exploiting its unique characteristics. This, together with other properties, makes our method extremely useful for realistic systems such as the Internet.
2005
This paper presents an analytical framework to study search strategies in large-scale decentralized unstructured peer-to-peer (P2P) networks. The peers comprising the P2P network and their application-level connections are modeled as generalized random graphs (GRGs) whose simple and efficient analysis is accomplished using the generating function of the graph's degree distribution.
Sufficient uptime is the main concern in a distributed system where there is no centralized organization and control. Therefore searching in P2P file sharing systems is still a big problem due to the high churn rate of nodes. To find a particular piece of data within the network P2P systems explicitly or implicitly provide a lookup mechanism, which largely depends on the availability of nodes. In this paper we propose a stable node based ranking technique by detecting reliable nodes with adequate uptime by nature to ensure efficient searching and un-interruption during the data receiving. We have simulated an unstructured file-sharing environment to evaluate our proposed technique.
In order to use Internet resources efficiently we need to search and locate information efficiently. System performance diminishes by either duplicating a large quantity of data on each and every node or flooding query to all the nodes in the network. Firstly, this paper reviews various searching algorithms. Search techniques can be classified as blind search in which information about neighbors is not kept by the peer and informed search where peers store information for routing queries to other nodes. It discusses how range queries can be processed efficiently by rotating scheme over structured P2P systems and secure searching algorithm based on topology adaptation which penalizes the malicious peers. Genetic algorithm providing parallel search are also covered in the paper. Lastly, it focuses on merits, demerits and applicability of these algorithms in different situations.
Lecture Notes in Computer Science, 2004
In this paper, we report a decentralized algorithm, termed ImmuneSearch, for searching p2p networks. ImmuneSearch avoids query message flooding; instead it uses an immune-systems-inspired concept of proliferation and mutation for message movement. In addition, a protocol is formulated to change the neighborhoods of the peers based upon their proximity with the queried item. This results in topology evolution of the network whereby similar contents cluster together. The topology evolution help the p2p network to develop 'memory', as a result of which the search efficiency of the network improves as more and more individual peers perform searches. Moreover, the algorithm is extremely robust and its performance is stable even when peers are transient.
2005
Existing cluster-based searching schemes in unstructured peer-to-peer (P2P) networks employ flooding/random forwarding on connected dominating sets (CDS) of networks. There exists no upper bound on the size of CDS of a network. Both flooding and CDS hinder query efficiency. Random forwarding worsens the recall ratio. In this paper, we propose a cluster-based searching scheme that intelligently forward queries on the maximum independent sets (MIS) of networks. Our approach partitions the entire network into disjoint clusters with one clusterhead (CH) per cluster. CHs form a MIS and are connected through gateway nodes. Each node takes one role, a CH, a gateway, or an ordinary node. A CH looks up the data for the entire cluster using data summaries of cluster members, which are represented by bloom filters. Between clusters, CHs intelligently forward queries via gateways to the best neighbor CHs that are most likely to return query results. The experimental results demonstrate that our scheme greatly improves the query efficiency without degrading the quality of the query results, compared to existing approaches.
When searching for specific nodes in a network an agent hops from one node to another by traversing network links. If the network is large, the agent typically possesses partial background knowledge or certain intuitions about the network. This background knowledge steers the agent's decisions when selecting the link to traverse next. In previous research two types of background knowledge have been applied to design and evaluate search algorithms: homophily (node similarity) and node popularity (typically represented by the degree of the node). In this paper we present a method for evaluating the relative importance of those two features for an e cient network search. Our method is based on a probabilistic model that represents those two features as a mixture distribution, i.e. as a convex combination of link selection probabilities based on the candidate node popularity and similarity to a given target node in the network. We also demonstrate this method by analyzing four networks, including social as well as information networks. Finally, we analyze strategies for dynamically adapting the mixture distribution during navigation. The goal of our analysis is to shed more light into appropriate configurations of the background knowledge for e cient search in various networks. The preliminary results provide promising insights into the influence of structural features on network search e ciency.
Computing Research Repository, 2002
We review a number of message-passing algorithms that can be used to search through power-law networks. Most of these algorithms are meant to be improvements for peer-to-peer file sharing systems, and some may also shed some light on how unstructured social networks with certain topologies might function relatively efficiently with local information. Like the networks that they are designed for,
Proceedings IEEE INFOCOM 2005. 24th Annual …, 2005
Efficient discovery of information, based on partial knowledge, is a challenging problem faced by many large scale distributed systems. This paper presents a peer-to-peer search protocol that addresses this problem. The proposed system provides an efficient mechanism for advertising a binary pattern, and discovering it using any subset of its 1-bits. A pattern (e.g., Bloom filter) summarizes the properties (e.g., keywords or service description) associated with a shared object (e.g., document or service).
Lecture Notes in Computer Science, 2005
Performing efficient decentralized search is a fundamental problem in Peer-to-Peer (P2P) systems. There has been a significant amount of research recently on developing robust self-organizing P2P topologies that support efficient search. In this paper we discuss four structured and unstructured P2P models (CAN, Chord, PRU, and Hypergrid) and three characteristic search algorithms (BFS, k-Random Walk, and GAPS) for unstructured networks. We report on the results of simulations of these networks and provide measurements of search performance, focusing on search in unstructured networks. We find that the proposed models produce small-world networks, and yet none exhibit power-law degree distributions. Our simulations also suggest that random graphs support decentralized search more effectively than the proposed unstructured P2P models. We also find that on these topologies, the basic breadth-first search algorithm and its simple variants have the lowest search cost.
International Journal of Computer Applications, 2013
Peer-to-Peer (P2P) [1] are widely used for file sharing purposes. This type of usage provides decentralized solutions over centralized complex architecture. Peer-to-Peer networks are gaining attention from both the scientific perspective as well as the large Internet community. Popular applications utilizing this new technology offer many attractive features to a growing number of users. P2P is an architecture which is all-together a different class of applications that use the concept of distributed resources to perform an important crucial function in a decentralized manner. The popularity and bandwidth consumption attributed to current Peer-to-Peer filesharing applications makes the operation of these distributed systems very important for the Internet community. Efficiently discovering the queried resource is the initial and most important step in establishing an efficient peer-to-peer communication. Here, we will be describing and analyzing the performances of some existing search mechanisms deployed for the peer discovery and the content look up.
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.