THE CONCEPTION OF INTEGRATING MUTITHREDED CRAWLER WITH PAGE RANK TECHNIQUE :A SURVEY
Journal Title: International Journal of Engineering Sciences & Research Technology - Year 30, Vol 4, Issue 6
Abstract
Web Crawler also well - known as “Web Robot”, “Web Spider” or merely “Bot” is software for downloading pages from the Web by design. Contrasting what the name may propose, a Web crawler does not in reality stir around computers connected to the Internet – as viruses or intelligent agents do – but only sends requests for documents on Web servers. The input to this software is starting or seed page. As the volume of th e World Wide Web (WWW) grows, it became essential to parallelize a web crawling process, with the intention of finish downloading pages in a rational amount of time. Web crawler which employs multi - processing to permit multiple crawler processes running in concurrent manner. There are a lot of programs out there for web crawling but it required a WebCrawler that allowed trouble - free customization. In this paper we have discussed on crawling technique and how Page Rank can increase the efficiency of web craw ling.
Authors and Affiliations
Ms. Amrita Banjare , Mr. Rohit Miri
HETEROGENEOUS FACE RECOGNITION USING KERNEL LDA METHOD
Here we propose the novel method to recognize the heterogeneous face recognition. Initially we remove the noise from the image. To remove the noise present in the image we use median filter. The system involves us...
A STRUCTURED ANALYSIS ON LEXICONS FOR OPEN SOURCE CLOUD TECHNOLOGIES
Cloud computing is a smart computing model since it allows for resources to be provisioned rendering on a demand basis, i.e., cloud users can lease resources as they become essential. Cloud Computing refers to both t...
Estimation of Weibull Parameters In Accelerated Life Testing Using Geometric Process With Type-Ii Censored Data
In Accelerated life testing (ALT), generally, the log linear function between life and stress is used to obtain the estimates of original parameters of the life. The log linear is just a simple re-parameterization of th...
CLOUD COMPUTING LOAD BALANCING MODEL WITH HETEROGENEOUS PARTITION
Cloud computing is an on demand service in which shared resources, information, software and other devices are provided according to the clients requirement at specific time . In the cloud computing paradigm, the sc...
Improving Search Strategy of Search Engine Using Probabilistic Latent Semantic
Users on the internet uses search engine to find information of their interest. However current search engines on web return answer to a query of user independent of user’s requirement for the information. In this...