Results 251 to 260 of about 3,180 (309)
CRATOR a CRAwler for TOR: Turning Dark Web Pages into Open Source INTelligence
Daniel De Pascale +3 more
openalex +2 more sources
Some of the next articles are maybe not open access.
Related searches:
Related searches:
UbiCrawler: a scalable fully distributed Web crawler
Software - Practice and Experience, 2004AbstractWe report our experience in implementing UbiCrawler, a scalable distributed Web crawler, using the Java programming language. The main features of UbiCrawler are platform independence, linear scalability, graceful degradation in the presence of faults, a very effective assignment function (based on consistent hashing) for partitioning the ...
Paolo Boldi +2 more
exaly +6 more sources
International Journal of Web Information Systems, 2006
Tools for the assessment of the quality and reliability of Web applications are based on the possibility of downloading the target of the analysis. This is achieved through Web crawlers, which can automatically navigate within a Web site and perform proper actions (such as download) during the visit.
C. Girardi, RICCA, FILIPPO, P. Tonella
openaire +2 more sources
Tools for the assessment of the quality and reliability of Web applications are based on the possibility of downloading the target of the analysis. This is achieved through Web crawlers, which can automatically navigate within a Web site and perform proper actions (such as download) during the visit.
C. Girardi, RICCA, FILIPPO, P. Tonella
openaire +2 more sources
ACM Transactions on Internet Technology, 2004
Topical crawlers are increasingly seen as a way to address the scalability limitations of universal search engines, by distributing the crawling process across users, queries, or even client computers. The context available to such crawlers can guide the navigation of links with the goal of efficiently locating highly relevant target pages.
Filippo Menczer +2 more
openaire +1 more source
Topical crawlers are increasingly seen as a way to address the scalability limitations of universal search engines, by distributing the crawling process across users, queries, or even client computers. The context available to such crawlers can guide the navigation of links with the goal of efficiently locating highly relevant target pages.
Filippo Menczer +2 more
openaire +1 more source
Proceedings of the International Conference and Workshop on Emerging Trends in Technology, 2010
As the number of Internet users and the number of accessible Web pages grows, it is becoming increasingly difficult for users to find documents that are relevant to their particular needs. Users must either browse through a large hierarchy of concepts to find the information for which they are looking or submit a query to a publicly available search ...
P. Dahiwale +2 more
openaire +1 more source
As the number of Internet users and the number of accessible Web pages grows, it is becoming increasingly difficult for users to find documents that are relevant to their particular needs. Users must either browse through a large hierarchy of concepts to find the information for which they are looking or submit a query to a publicly available search ...
P. Dahiwale +2 more
openaire +1 more source
Reducing web crawler overhead using mobile crawler
2011 International Conference on Emerging Trends in Electrical and Computer Technology, 2011To search any information on the web users extensively use the search engines. As the growth of the World Wide Web exceeded all expectations, the search engines rely on web crawlers to maintain the index of billions of pages for efficient searching. The web crawlers have to interact with millions of hosts and retrieve the pages continuously to keep the
S Anbukodi, K Muthu Manickam
openaire +1 more source
ACM SIGMETRICS Performance Evaluation Review, 2000
In this paper we study how to make web servers (e.g., Apache) more crawler friendly. Current web servers offer the same interface to crawlers and regular web surfers, even though crawlers and surfers have very different performance requirements. We evaluate simple and easy-to-incorporate modifications to web servers so that there are significant ...
Onn Brandman +3 more
openaire +1 more source
In this paper we study how to make web servers (e.g., Apache) more crawler friendly. Current web servers offer the same interface to crawlers and regular web surfers, even though crawlers and surfers have very different performance requirements. We evaluate simple and easy-to-incorporate modifications to web servers so that there are significant ...
Onn Brandman +3 more
openaire +1 more source
2015 2nd International Conference on Electronics and Communication Systems (ICECS), 2015
Users and uses of internet is growing tremendously these days which causing an extreme trouble and efforts at user side to get web pages searched which are as per concern and relevant to user's requirement Generally users approach to search web pages from a large available hierarchy of concepts or use a query to browse web pages from available search ...
Gunjan H. Agre, Nikita V. Mahajan
openaire +1 more source
Users and uses of internet is growing tremendously these days which causing an extreme trouble and efforts at user side to get web pages searched which are as per concern and relevant to user's requirement Generally users approach to search web pages from a large available hierarchy of concepts or use a query to browse web pages from available search ...
Gunjan H. Agre, Nikita V. Mahajan
openaire +1 more source
International Conference on Information Technology: Coding and Computing, 2004. Proceedings. ITCC 2004., 2004
The requirement of a Web crawler that downloads most relevant pages is still a major challenge in the field of information retrieval systems. The use of link analysis algorithms like page rank and other importance-metrics have shed a new approach in prioritizing the URL queue for downloading higher relevant pages. The combination of these metrics along
S. Ganesh +4 more
openaire +1 more source
The requirement of a Web crawler that downloads most relevant pages is still a major challenge in the field of information retrieval systems. The use of link analysis algorithms like page rank and other importance-metrics have shed a new approach in prioritizing the URL queue for downloading higher relevant pages. The combination of these metrics along
S. Ganesh +4 more
openaire +1 more source

