- This article is about using the Internet for research; for the field of research about the Internet, see Internet studies.
Internet research is the practice of using the Internet, especially the World Wide Web, for research. To the extent that the Internet is widely and readily accessible to hundreds of millions of people in many parts of the world, it can provide practically instant information on most topics, and is having a profound impact on the way in which ideas are formed and knowledge is created.
Research is a broad term. Here, it is used to mean "looking something up (on the Web)". It includes any activity where a topic is identified, and an effort is made to actively gather information for the purpose of furthering understanding. Common applications of Internet research include personal research on a particular subject (something mentioned on the news, a health problem, etc), students doing research for academic projects and papers, and journalists and other writers researching stories. It should be distinguished from scientific research - research following a defined and rigorous process - carried out on the Internet; from straightforward finding of specific info, like locating a name or phone number; and from research about the Internet.
Compared to the Internet, print physically limits access to information. A book has to be identified, then actually obtained. On the Net, the Web can be searched, and typically hundreds or thousands of pages can be found with some relation to the topic, within seconds. In addition, email (including mailing lists), online discussion forums (aka message boards, BBS's), and other personal communication facilities (instant messaging, IRC, newsgroups, etc) can provide direct access to experts and other individuals with relevant interests and knowledge. However, difficulties persist in verifying a writer's credentials, and therefore the accuracy or pertinence of the information obtainedâ€”see also the article Criticism of Wikipedia and its section Difficulty of fact-checking.
Further difficulties in internet research center around search tool bias and whether the searcher has sufficient skill to draw meaningful results from the abundance of material typically available. The first resources retrieved may not be the most suitable resources to answer a particular question. For example, popularity is often a factor used in structuring internet search results, but the most popular information is not always the most correct, or representative of the breadth of knowledge and opinion on a topic.
 Search tools
The most popular search tools for finding information on the internet include Web search engines, metasearch engines, Web directories, and specialty search services. A Web search engine uses software known as a Web crawler to follow the hyperlinks connecting the pages on the World Wide Web. The information on these Web pages is indexed and stored by the search engine. To access this information, a user enters keywords in a search form and the search engine queries its indices to find Web pages that contain the keywords and displays them in search engine results page (SERP). The SERP list typically includes hyperlinks and brief descriptions of the content found. Search results are ranked using complex algorithms, which take into consideration the location and frequency of keywords on a Web page, along with the quality and number of external hyperlinks pointing at the Web page.
A Metasearch engine enables users to enter a search query once and have it run against multiple search engines simultaneously, creating a list of aggregated search results. Since no single search engine covers the entire web, a metasearch engine can produce a more comprehensive search of the web. Most metasearch engines automatically eliminate duplicate search results. However, metasearch engines have a significant limitation because the most popular search engines, such as Google, are not included because of legal restrictions.
A Web directory organizes subjects in a hierarchical fashon that lets users investigate the breadth of a specific topic and drill down to find relevant links and content. Web directories can be assembled automatically by algorithms or handcrafted. Human-edited Web directories have the distinct advantage of higher quality and reliability, while those produced by alogrithms can offer more comprehensive coverage. The scope of Web directories are generally broad, such as DMOZ, Yahoo! and The WWW Virtual Library, covering a wide range of subjects, while others focus on specific topics.
Specialty search tools enable users to find information that conventional search engines and metasearch engines cannot access because the content is stored in databases. In fact, the vast majority of information on the web is stored in databases that require users to go to a specific site and access it through a search form. Often, the content is generated dynamically. As a consequence, Web crawlers are unable to index this information. In a sense, this content is "hidden" from search engines, leading to the term invisible or deep Web. Specialty search tools have evolved to provide users with the means to quickly and easily find deep Web content. These specialty tools rely on advanced bot and intelligent agent technologies to search the deep Web and automatically generate specialty Web directories, such as the Virtual Private Library.
 See also
- ^ Hargittai, E. (April 2002). "Second-Level Digital Divide: Differences in Peopleâ€™s Online Skills". First Monday. http://firstmonday.org/htbin/cgiwrap/bin/ojs/index.php/fm/article/view/942/864. Retrieved February 5, 2010.
 External links