It almost seems like a fantasy book or maybe the prefect combination of fantasy and science fiction, this weeks reading really delves into the dark places of the web. The web below the web. It is almost like the multiple layers of the web.
First there is the searching and crawling of the web (Hawking). That the coded algorithms for crawling effects speed of the searches, keeps the request for don't clog the web and requests are answered systemically. In addition, that only the web site you want or close to the website is the one that is pulled form the super highway and rejects bad request or spam sites, The basis of this is Javascript, with hyperlinks de-coding to meet end users requests.
Then we explore the deep web, or the web below the web. Which is500 times larger than what we think of and can hold huge amounts of information. It contains 550 billion individual documents compared to the 1billion on the traditional web. More than 200,000 deep websites current on deep web. The deep web receives twice as much traffic than traditional web sites with deeper content than the conventional web. The deep web can easily meet the needs of every known information community. It is very topic specific database.
This leads to the Open Archives Initiative. The OAI protocols were widely adapted in 2001, with the test format being organized by U Of I. The goal of OAI is to develop and promote interoperability stands that aim to facilitate the effective use of content. Goal is also to better relationships to data providers (repositories) and service providers (harvesters). It capitalizes on Dublin Core metadata cataloging so that users can search effectively. The protocol can provide access to parts of the "invisible web" that are not easily accessible to search engines.
No comments:
Post a Comment