Results 21 to 30 of about 3,180 (309)

A focused crawler based on semantic disambiguation vector space model

open access: yesComplex & Intelligent Systems, 2022
The focused crawler grabs continuously web pages related to the given topic according to priorities of unvisited hyperlinks. In many previous studies, the focused crawlers predict priorities of unvisited hyperlinks based on the text similarity models ...
Wenjun Liu   +8 more
doaj   +1 more source

Public opinion data collection of power network using topic crawler

open access: yesXi'an Gongcheng Daxue xuebao, 2022
The traditional public opinion data collection methods of power network have some problems, such as low recall rate, low calculation accuracy and being time-consuming. Therefore, the topic crawler technology was used to improve the data collection method.
XI Zenghui   +3 more
doaj   +1 more source

Web Crawler and Web Crawler Algorithms: A Perspective

open access: yesInternational Journal of Engineering and Advanced Technology, 2020
A web crawler is also called spider. For the intention of web indexing it automatically searches on the WWW. As the W3 is increasing day by day, globally the number of web pages grown massively. To make the search sociable for users, searching engine are mandatory. So to discover the particular data from the WWW search engines are operated. It would be
K Velkumar, P Thendral
openaire   +1 more source

APLIKASI WEB CRAWLER UNTUK WEB CONTENT PADA MOBILE PHONE

open access: yesJUTI: Jurnal Ilmiah Teknologi Informasi, 2009
Crawling is the process behind a search engine, which served through the World Wide Web in a structured and with certain ethics. Applications that run the crawling process is called Web Crawler, also called web spider or web robot.
Sarwosri Sarwosri   +2 more
doaj   +1 more source

Discovering Land Cover Web Map Services from the Deep Web with JavaScript Invocation Rules

open access: yesISPRS International Journal of Geo-Information, 2016
Automatic discovery of isolated land cover web map services (LCWMSs) can potentially help in sharing land cover data. Currently, various search engine-based and crawler-based approaches have been developed for finding services dispersed throughout the ...
Dongyang Hou, Jun Chen, Hao Wu
doaj   +1 more source

A Hybrid Revisit Policy For Web Search

open access: yesJournal of Advances in Information Technology, 2012
A crawler is a program that retrieves and stores pages from the Web, commonly for a Web search engine. A crawler often has to download hundreds of millions of pages in a short period of time and has to constantly monitor and refresh the downloaded pages.
Vipul Sharma, Mukesh Kumar, Renu Vig
doaj   +1 more source

GeoWeb Crawler: An Extensible and Scalable Web Crawling Framework for Discovering Geospatial Web Resources

open access: yesISPRS International Journal of Geo-Information, 2016
With the advance of the World-Wide Web (WWW) technology, people can easily share content on the Web, including geospatial data and web services. Thus, the “big geospatial data management” issues start attracting attention.
Chih-Yuan Huang, Hao Chang
doaj   +1 more source

Nuevos retos de la tecnología web crawler para la recuperación de información

open access: yesMétodos de Información, 2014
El web crawler constituye una parte importante de la cadena documental en la recuperación de información, dado que genera el corpus documental necesario sobre el que aplicar los distintos algoritmos de recuperación.
Blázquez Ochando, Manuel
doaj   +1 more source

Institutional Repository Keyword Analysis with Web Crawler

open access: yesCentral European Journal of Educational Research, 2022
This study aims at investigating procedures of semantic and linguistic extraction of keywords from metadata of documents indexed in the Institutional Repository Unesp.
Mariângela Spotti Lopes Fujita   +2 more
doaj   +1 more source

RCrawler: An R package for parallel web crawling and scraping

open access: yesSoftwareX, 2017
RCrawler is a contributed R package for domain-based web crawling and content scraping. As the first implementation of a parallel web crawler in the R environment, RCrawler can crawl, parse, store pages, extract contents, and produce data that can be ...
Salim Khalil, Mohamed Fakir
doaj   +1 more source

Home - About - Disclaimer - Privacy