An effective and efficient Web content extractor for optimizing the crawling process

SUMMARY Classical Web crawlers make use of only hyperlink information in the crawling process. However, focused crawlers are intended to download only Web pages that are relevant to a given topic by utilizing word information before downloading the Web page. But, Web pages contain additional informa...

Full description

Saved in:
Bibliographic Details
Published in:Software, practice & experience Vol. 44; no. 10; pp. 1181 - 1199
Main Authors: Uzun, Erdinç, Serdar Güner, Edip, Kılıçaslan, Yılmaz, Yerlikaya, Tarık, Agun, Hayri Volkan
Format: Journal Article
Language:English
Published: Bognor Regis Blackwell Publishing Ltd 01-10-2014
Wiley Subscription Services, Inc
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:SUMMARY Classical Web crawlers make use of only hyperlink information in the crawling process. However, focused crawlers are intended to download only Web pages that are relevant to a given topic by utilizing word information before downloading the Web page. But, Web pages contain additional information that can be useful for the crawling process. We have developed a crawler, iCrawler (intelligent crawler), the backbone of which is a Web content extractor that automatically pulls content out of seven different blocks: menus, links, main texts, headlines, summaries, additional necessaries, and unnecessary texts from Web pages. The extraction process consists of two steps, which invoke each other to obtain information from the blocks. The first step learns which HTML tags refer to which blocks using the decision tree learning algorithm. Being guided by numerous sources of information, the crawler becomes considerably effective. It achieved a relatively high accuracy of 96.37% in our experiments of block extraction. In the second step, the crawler extracts content from the blocks using string matching functions. These functions along with the mapping between tags and blocks learned in the first step provide iCrawler with considerable time and storage efficiency. More specifically, iCrawler performs 14 times faster in the second step than in the first step. Furthermore, iCrawler significantly decreases storage costs by 57.10% when compared with the texts obtained through classical HTML stripping. Copyright © 2013 John Wiley & Sons, Ltd.
Bibliography:ark:/67375/WNG-CNHFMNXD-5
istex:97A7D88A2DA97800B55E898E22C1310394788D41
ArticleID:SPE2195
ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:0038-0644
1097-024X
DOI:10.1002/spe.2195