Deep Web, Dark Internet and Darknets

From IT

Revision as of 15:06, 5 January 2011 by Erica Prato (Talk | contribs)
Jump to: navigation, search

Contents

Introduction

Deep Web, Dark Internet and Darknet represent three concepts that might seem similar and that are sometimes used as synonimous. Really, these three concepts represent three very different worlds.

Hearing terms like these, they may seem fascinating, perhaps because are subjects that does not usually get talk about on the Web. More in deep, when a user surfs the Web, he thinks that all the information available all over the world are also available on it. There are, on the other hand, areas of the World Wide Web that cannot be easily accessed by the public.

In the following sections, these three concepts will be explained, in a way in which the differences between them can be clear.

The Deep Web

The Deep Web it's simply that part of the Internet which is not indexed by search engines.
Before to begin, we have to consider a preliminary difference between the Surface Web and the Deep Web. [1]

Surface Web is the term used to identify that portion of the World Wide Web that is indexed by conventional search engines: in other words, is what you can find by using general web search engines.

On the contrary, the Deep Web is defined as that portion of the Wolrd Wide Web that is not accessible through a research executed using general search engines, and is much bigger than the previous one (http://www.internettutorials.net/deepweb.asp).

Beyond the trillion pages a search engines such as Google knows, there is a really vast world of hidden data. This content could be:

  • the content of database, that is accessible only by query;
  • files such as multimedia ones, images, software;
  • the content on web sites protected by passwords or other kinds of restrictions;
  • the content of "full text" articles and books;
  • the content of social networks;
  • financial information;
  • medical research.

Nowadays, we have to consider also other kinds of content, such as:

  • blog postings;
  • bookmarks and citations in bookmarking sites;
  • flight schedules.

Search engines and the Deep Web

Search engines rely on crawlers, or spiders, that wander the Web following the trails of hyperlinks that link the Web together: it means that spidersindex the addresses of the pages they discover.

The negative aspect referable to this indiscriminate crawl approach had been replaced with the so called "popularity of pages" in a search engine like Google: in other words, the most popular pages, and so those that register the highest frequence of research, have priority both for crawling and displaying results.

In the Deep Web, happens that spiders, when finding a page, don't know what to do with it: it means that spiders can record these pages, but aren't able to display the content of them. The most frequent reasons can be refereable to technical barriers (database driven content, for instance) or decisions taken by the owners of web sites (the necessity to be register with a password to access the site, for instance), that make impossible for spiders to do their work. (http://websearch.about.com/od/invisibleweb/a/invisible_web.htm)

Another important reason refers to the linkage: if a web document is not linked to another, it will never be discovered.

Commercial search engineselaborated some methods that permit to navigate the Deep Web and find resources on specific Web Servers. Sitemap Protocol is one of them, and allows search engines to crawl the sites in a more intelligent way (http://en.wikipedia.org/wiki/Deep_Web).

To search the Deep Web

Most users tend to use search engines in a very elementary way, and perhaps for this reason the navigation on the Deep Web is limited to a smaller part of Web users. Despite of this fact, the Deep Web is considered to be the " paradigm for the next generation Internet " (2005, Deep Web FAQ, par. 35). In fact, a good use of the Deep Web can drastically reduce the time necessary for a research and give back high quality information: let consider only the fact that, if the Surface Web contains 1% of the information available on the Web, the Deep Web contains the 99% of them.
Even if Google can be considered one of the best existent search engines, it isn't able, or better, crawlers aren't able to rank documents contained in the Deep Web: they don't know how to search the Deep Web collections, as said before.
But the most serious problem according to Google regards the fact that not always it ranks pages in the way in which you would rank documents. Google ranks pages according to the assumpion related to the popularity of pages, but they are not necessary the most relevant for the users.
When a user is looking for resources related to science, technologies, business and so on it's so better to navigate the Deep Web instead of the Surface Web.

With the evolution of the Web and with the portals describe below, the Deep Web will become easier to use and surf.

In order to search Deep Web content, it's necessary to use specific portals, such as for instance CompletePlanet (http://www.completeplanet.com/). Because of the presence of thousands of databases that contain Deep Web content, CompletePlanet offers the possibility to navigate these Deep Web databases.
The linked image [2] shows the home page of the portal, with the list of all the available dynamic searcheable databases. it's in fact possible to to go to various topic areas (medicine, art&design, science, politics, and so on) and find content that are not display by using conventional search engines [3].
Other Deep Web search engines are (http://websearch.about.com/od/invisibleweb/tp/deep-web-search-engines.htm):

  • Clusty, that is a meta search engine able to combine results from different sources and give back the best possible result;
  • SurfWax, that gives the possibility to obtain results from different search engines at the same time, and to create personalized set of sources;
  • InternetArchive, that gives access to specific searcheable topics such as live music, audio and printed materials;
  • Scirus, that is dedicated only to scientific material;
  • USA.gov for access information and databases from the USA government.</p>

Search engines may find deep data, but their coverage sometimes give back only less relevant content. In addition, the research of this kind of data needs a certain ability in navigating the web. A possible technique to find Deep Web resources without using specific portals like the above mentioned ones is for instance this:

  • using a search engine that contains both Surface and Deep Web resources, it's possible to make researches in databases by usgin specific research terms:
    • on Google and InfoMine: what you are looking for (database OR repository OR archive)
    • on Teoma: what you are looking for (resources OR meta site OR portal OR pathfinder) (http://techdeepweb.com/4.html).

Deep Web Technologies

Deep Web explorers usually work, search, according to two ways:

  • They harvest documents. In this first case, we have to consider that harvesting has both positive and negative points: it's an ideal technique if there are adequate infrastructures to make contents available to users, but isn't the best one if the search interface is not intuitive and does not permit to the user to retrieve a huge amount of documents in an easy way. The Open Archive Initiative (http://www.openarchives.org/) is an harvesting protocol. It develops standards to facilitate the dissemination of content. Through it it's especially possible to collect multiple collections and aggregate them into a single, sentralized one.

  • They search collections on the fly. Deep Web Technologies Explorit (http://www.deepwebtech.com/index.html) presents this real-time search approach, also this with its weak and potisive points. The most considerable positive point is related to the fact that also most Deep Web collections led themselves to real-time searching, even if they do not led themselves to harvesting: in fact, by not implementing a harvesting protocol, the owner of a document does not have to do something to allow this document to be searched. According to this approach, the client uses a simple HTTP protocol and a web form that initiates the query to enter the database: the content will be processed and results will be displayed to the user. The weak point is especially related to remot collection, and in particular to the fact that there are ongoing demands on it. Another possible negative aspect is, on the contrary with conventional search engines, that rank results in an ordered way, related to the inability to rank documents in an effective way.

How big is it?

It seems that the Deep Web is 500 times bigger than the Surface Web, containing 7.500 terabytes of data and 550 billion of documents (August 2001, Bergman, Michael K., "The Deep Web: Surfacing Hidden Value". The Journal of Electronic Publishing 7).

Why the Deep Web is important?

Usually we trust only to what we find in search engines such as Google or Yahoo!. But sometimes is not so easy to find what we are really looking for and what we specifically need on them, especially if we are trying to find obscure and complicated topics.
If we compare the Web to a vast, a very big library, can be easier to understand the problem.
In a vast library, we wouldn't expect to find what we are looking for simply walking on the first bookshelf, but we have to go more in deep with our research, searching on a number of shelves and so on, especially if we are searching for a rare document or book.
It's the same with the Web: sometimes, search engines are not able to help us with our search because they do not present all the information about that topic, while the Deep Web will.

To make some real examples, let consider the CIA edition of "The Chemical and Biological Walfare Thread", that is a public document, but that cannot be accessed through a search engines like Google, or the documents of the U.S. Federal Government, and so millions of documents related to law, economics and so on. These data, that may be visible to everyone, are stored in databases, and for this reason cannot be accessed by using search engines. These outputs can be penetrated, on the other hand, by the Deep Web search engines: they are able to solve this situation, in this case caused by bureaucratic inertia, by displaying those documents.
While traditional search engines work weel on HTML pages, the Deep Web gathers information contained in databases, and so remains unseen to a lot of users (http://www.salon.com/technology/feature/2004/03/09/deep_web/index.html).

The Deep Web is a "contenitor" of useful content on the Internet that is usually of high quality, is very specific in nature and that is well managed.
Let consider, for instance, the PubMed Database. PubMed, that is managed by the U.S. National Library of Medicine, contains 20 million citations from sources such as MEDLINE, authorized journals, online books, it provides access to websites and other additional resources. The content is authored by professionals authors, writers, and is contained in professional sources: in fact, the National Center for Biotechnology Information and the National Library of Medicine spend time and resources to take the content available and to manage it. In particular, if we enter the section called "All Databases" - see the linked image [4] - we are able to see all the possible resources made available by PubMed, and typing what you are looking for in the search box, you receive back all the resources dedicated to that specific topic (http://www.ncbi.nlm.nih.gov/pubmed).

So, why the Deep Web, and especially the content of the Deep Web is so important? Because in fields such as business, sicence, and other worlds like these, time is very important and the marketplace is competitive, because it's important for a student who has to write his thesis to find authored and high quality material.

Personal tools