𝔖 Bobbio Scriptorium
✦   LIBER   ✦

An architecture for SCS: A specialized web crawler on the topic of security

✍ Scribed by H. Cenk Özmutlu; Seda Özmutlu


Publisher
Wiley (John Wiley & Sons)
Year
2005
Tongue
English
Weight
1021 KB
Volume
41
Category
Article
ISSN
0044-7870

No coin nor oath required. For personal study only.

✦ Synopsis


Abstract

Mining for correct and relevant information in the World Wide Web is a difficult task, handled by Web crawlers. This study outlines the components of a specialized crawler on the topic of security (SCS) that heavily makes use of artificial neural networks and rule‐based expert systems to establish successful focused crawling on the topic of security. SCS is designed to find, index and follow the updates of Web pages of interest, and proposes new approaches for reaching relevant pages, which might stay hidden to other crawling approaches. SCS consists of four new page explorers, a database of relevant pages, a relevance evaluator using artificial neural networks and an updater using rule‐based expert systems. SCS is a multi‐threaded multi‐object Java Applet and Application combination with embedded SQL and PHP elements and is applicable on single or multiple machines through parallel processing with its expandable and modular structure.


📜 SIMILAR VOLUMES


Finding governmental statistical data on
✍ Irina Ceaparu; Ben Shneiderman 📂 Article 📅 2004 🏛 John Wiley and Sons 🌐 English ⚖ 613 KB

## Abstract More than 100 U.S. governmental agencies offer links through FedStats, a centralized Web site that facilitates access to statistical tables, reports, and agencies. This and similar large collections need appropriate interfaces to guide the general public to easily and successfully find