An architecture for SCS: A specialized web crawler on the topic of security

No Thumbnail Available

Date

2004

Authors

Journal Title

Journal ISSN

Volume Title

Publisher

Information Today

Abstract

Mining for correct and relevant information in the World Wide Web is a difficult task, handled by Web crawlers. This study outlines the components of a specialized crawler on the topic of security (SCS) that heavily makes use of artificial neural networks and rule-based expert systems to establish successful focused crawling on the topic of security. SCS is designed to find, index and follow the updates of Web pages of interest, and proposes new approaches for reaching relevant pages, which might stay hidden to other crawling approaches. SCS consists of four new page explorers, a database of relevant pages, a relevance evaluator using artificial neural networks and an updater using rule-based expert systems. SCS is a multi-threaded multi-object Java Applet and Application combination with embedded SOL and PHP elements and is applicable on single or multiple machines through parallel processing with its expandable and modular structure.

Description

Bu çalışma, 12-17 Kasım 2004 tarihleri arasında Rhode Island[Amerika Birleşik Devletleri]’de düzenlenen 67. Asist Annual Meeting'de bildiri olarak sunulmuştur.

Keywords

Computer science, Information science and library science, Web crawler, Focused crawling, Specialized crawling, Artificial intelligence, Artificial neural networks, Rule-based systems, Security, World-wide-web, Search engines, Information

Citation

Özmutlu, H.C. ve Özmutlu, S. (2004). “An architecture for SCS: A specialized web crawler on the topic of security”. ASIST 2004: Proceedings of the 67. Asist Annual Meeting, Proceedings of the ASIST Annual Meeting, 41, 317-326.