版權(quán)說明:本文檔由用戶提供并上傳,收益歸屬內(nèi)容提供方,若內(nèi)容存在侵權(quán),請進行舉報或認領(lǐng)
文檔簡介
1、外文資料原文Efficient URL Caching for World Wide Web CrawlingMarc NajorkBMJ (International Edition) 2009Crawling the web is deceptively simple: the basic algorithm is (a)Fetch a page (b) Parse it to extract all linked URLs (c
2、) For all the URLs not seen before, repeat (a)–(c). However, the size of the web (estimated at over 4 billion pages) and its rate of change (estimated at 7% per week) move this plan from a trivial programming exercise to
3、 a serious algorithmic and system design challenge. Indeed, these two factors alone imply that for a reasonably fresh and complete crawl of the web, step (a) must be executed about a thousand times per second, and thus t
4、he membership test (c) must be done well over ten thousand times per second against a set too large to store in main memory. This requires a distributed architecture, which further complicates the membership test.A cruci
5、al way to speed up the test is to cache, that is, to store in main memory a (dynamic) subset of the “seen” URLs. The main goal of this paper is to carefully investigate several URL caching techniques for web crawling. We
6、 consider both practical algorithms: random replacement, static cache, LRU, and CLOCK, and theoretical limits: clairvoyant caching and infinite cache. We performed about 1,800 simulations using these algorithms with vari
7、ous cache sizes, using actual log data extracted from a massive 33 day web crawl that issued over one billion HTTP requests. Our main conclusion is that caching is very effective – in our setup, a cache of roughly 50,000
8、 entries can achieve a hit rate of almost 80%. Interestingly, this cache size falls at a critical point: a substantially smaller cache is much less effective while a substantially larger cache brings little additional be
9、nefit. We conjecture that such critical points are inherent to our problem and venture an explanation for this phenomenon.1. INTRODUCTIONbillion pages. Various studies [3, 27, 28] have indicated that, historically, the w
10、eb has doubled every 9-12 months.2. Web pages are changing rapidly. If “change” means “any change”, then about 40% of all web pages change weekly [12]. Even if we consider only pages that change by a third or more, about
11、 7% of all web pages change weekly [17]. These two factors imply that to obtain a reasonably fresh and 679 complete snapshot of the web, a search engine must crawl at least 100 million pages per day. Therefore, step (a)
12、must be executed about 1,000 times per second, and the membership test in step (c) must be done well over ten thousand times per second, against a set of URLs that is too large to store in main memory. In addition, crawl
13、ers typically use a distributed architecture to crawl more pages in parallel, which further complicates the membership test: it is possible that the membership question can only be answered by a peer node, not locally.A
14、crucial way to speed up the membership test is to cache a (dynamic) subset of the “seen” URLs in main memory. The main goal of this paper is to investigate in depth several URL caching techniques for web crawling. We exa
15、mined four practical techniques: random replacement, static cache, LRU, and CLOCK, and compared them against two theoretical limits: clairvoyant caching and infinite cache when run against a trace of a web crawl that iss
16、ued over one billion HTTP requests. We found that simple caching techniques are extremely effective even at relatively small cache sizes such as 50,000 entries and show how these caches can be implemented very efficientl
17、y. The paper is organized as follows: Section 2 discusses the various crawling solutions proposed in the literature and how caching fits in their model. Section 3 presents an introduction to caching techniques and descri
18、bes several theoretical and practical algorithms for caching. We implemented these algorithms under the experimental setup described in Section 4. The results of our simulations are depicted and discussed in Section 5, a
溫馨提示
- 1. 本站所有資源如無特殊說明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請下載最新的WinRAR軟件解壓。
- 2. 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請聯(lián)系上傳者。文件的所有權(quán)益歸上傳用戶所有。
- 3. 本站RAR壓縮包中若帶圖紙,網(wǎng)頁內(nèi)容里面會有圖紙預(yù)覽,若沒有圖紙預(yù)覽就沒有圖紙。
- 4. 未經(jīng)權(quán)益所有人同意不得將文件中的內(nèi)容挪作商業(yè)或盈利用途。
- 5. 眾賞文庫僅提供信息存儲空間,僅對用戶上傳內(nèi)容的表現(xiàn)方式做保護處理,對用戶上傳分享的文檔內(nèi)容本身不做任何修改或編輯,并不能對任何下載內(nèi)容負責。
- 6. 下載文件中如有侵權(quán)或不適當內(nèi)容,請與我們聯(lián)系,我們立即糾正。
- 7. 本站不保證下載資源的準確性、安全性和完整性, 同時也不承擔用戶因使用這些下載資源對自己和他人造成任何形式的傷害或損失。
最新文檔
- 網(wǎng)絡(luò)爬蟲外文翻譯---基于網(wǎng)絡(luò)爬蟲的有效url緩存
- 外文翻譯---基于網(wǎng)絡(luò)爬蟲的有效url緩存
- 計算機外文翻譯---基于網(wǎng)絡(luò)爬蟲的有效url緩存
- 計算機外文翻譯---萬維網(wǎng)爬行的有效url緩存(節(jié)選)
- 基于URL及上下文的主題網(wǎng)絡(luò)爬蟲研究.pdf
- 基于URL規(guī)則的聚焦爬蟲及其應(yīng)用.pdf
- 畢業(yè)論文外文翻譯-網(wǎng)絡(luò)爬蟲
- 惡意URL檢測項目中基于PageRank算法的網(wǎng)絡(luò)爬蟲的設(shè)計和實現(xiàn).pdf
- 網(wǎng)絡(luò)爬蟲畢業(yè)設(shè)計(含外文翻譯)
- 網(wǎng)絡(luò)視頻爬蟲的緩存和更新策略.pdf
- 面向網(wǎng)絡(luò)爬蟲的海量URL數(shù)據(jù)管理技術(shù)研究.pdf
- 網(wǎng)絡(luò)爬蟲的設(shè)計與實現(xiàn)畢業(yè)論文(含外文翻譯)
- 網(wǎng)絡(luò)爬蟲的設(shè)計與實現(xiàn)畢業(yè)論文(含外文翻譯)
- 基于URL及上下文的并行分塊處理主題爬蟲研究.pdf
- 基于Python 的網(wǎng)絡(luò)爬蟲.docx
- 主題爬蟲URL分析模型與高度技術(shù)研究.pdf
- 網(wǎng)絡(luò)爬蟲
- 無線傳感器網(wǎng)絡(luò)中基于能量有效的按需緩存策略研究.pdf
- 基于網(wǎng)絡(luò)爬蟲技術(shù)的網(wǎng)絡(luò)新聞分析
- 基于網(wǎng)絡(luò)爬蟲的用戶行為分析.pdf
評論
0/150
提交評論