The World Wide Web ("WWW" or simply the "Web") is a global, read-write information space. Text documents, images, multimedia and many other items of information, referred to as resources, are identified by short, unique, global identifiers called Uniform Resource Identifiers (URIs) so that each can be found, accessed and cross referenced in the simplest possible way. The term is often mistakenly used as a synonym for the Internet itself, but the Web is actually something that is available via the Internet, just like e-mail and many other Internet services.
The Web was first created around 1990 by Tim Berners-Lee working at CERN in Europe. As its inventor, Berners-Lee conceptualised the Web to be the Semantic Web where all its contents should be descriptively marked-up.
Contents
* 1 Basic terms
* 2 How the Web works
o 2.1 Caching
* 3 Origins
* 4 Web standards
Basic terms
The World Wide Web is the combination of 4 basic ideas:
* Hypertext: a format of information which allows, in a computer environment, one to move from one part of a document to another or from one document to another through internal connections among these documents (called "hyperlinks");
* Resource Identifiers: unique identifiers used to locate a particular resource (computer file, document or other resource) on the network;
* The Client-server model of computing: a system in which client software or a client computer makes requests of server software or a server computer that provides the client with resources or services, such as data or files; and
* Markup language: characters or codes embedded in text which indicate structure, semantic meaning, or advice on presentation.
On the World Wide Web, a client program called a user agent retrieves information resources, such as Web pages and other computer files, from Web servers using their URLs. If the user agent is a kind of Web browser, it displays the resources on a user's computer. The user can then follow hyperlinks in each web page to other World Wide Web resources, whose location is embedded in the hyperlinks. It is also possible, for example by filling in and submitting web forms, to post information back to a Web server for it to save or process in some way. Web pages are often arranged in collections of related material called "Web sites." The act of following hyperlinks from one Web site to another is referred to as "browsing" or sometimes as "surfing" the Web.
The phrase "surfing the Internet" was first popularized in print by Jean Armour Polly, a librarian, in an article called Surfing the INTERNET, published in the University of Minnesota Wilson Library Bulletin in June, 1992. Although Polly may have developed the phrase independently, slightly earlier uses of similar terms have been found on the Usenet from 1991 and 1992, and some recollections claim it was also used verbally in the hacker community for a couple years before that. Polly is famous as "NetMom" in the history of the Internet.
For more information on the distinction between the World Wide Web and the Internet itself—as in everyday use the two are sometimes confused—see Dark internet where this is discussed in more detail.
Although the English word worldwide is normally written as one word (without a space or hyphen), the proper name World Wide Web and abbreviation WWW are now well-established even in formal English. The earliest references to the Web called it the WorldWideWeb (an example of computer programmers' fondness for CamelCase) or the World-Wide Web (with a hyphen, this version of the name is the closest to normal English usage).
Ironically, the abbreviation "WWW" is somewhat impractical as it contains two or three times as many syllables (depending on accent) as the full term "World Wide Web", and thus takes longer to say.
How the Web works
Viewing a Web page or other resource on the World Wide Web normally begins either by typing the URL of the page into a Web browser, or by following a hypertext link to that page or resource. The first step, behind the scenes, is for the server-name part of the URL to be resolved into an IP address by the global, distributed Internet database known as the Domain name system or DNS.
The next step is for an HTTP request to be sent to the Web server at that IP address, for the page required. In the case of a typical Web page, the HTML text, graphics and any other files that form a part of the page will be requested and returned to the client (the Web browser) in quick succession.
The Web browser then renders the page as described by the HTML, CSS and other files received, incorporating the images and other resources as necessary. This produces the on-screen 'page' that the viewer sees.
Most Web pages will themselves contain hyperlinks to other relevant and informative pages and perhaps to downloads, source documents, definitions and other Web resources.
Such a collection of useful, related resources, interconnected via hypertext links, is what has been dubbed a 'web' of information. Making it available on the Internet created what Tim Berners-Lee first called the World Wide Web in the early 1990s [2] [3].
Caching
If the user returns to a page fairly soon, it is likely that the data will not be retrieved from the source Web server, as above, again. By default, browsers cache all web resources on the local hard drive. An HTTP request will be sent by the browser that asks for the data only if it has been updated since the last download. If it has not, the cached version will be reused in the rendering step.
This is particularly valuable in reducing the amount of web traffic on the internet. The decision about expiry is made independently for each resource (image, stylesheet, JavaScript file etc., as well as for the HTML itself). Thus even on sites with highly dynamic content, many of the basic resources are only supplied once per session or less. It is worth it for any Web site designer to collect all the CSS and JavaScript into a few site-wide files so that they can be downloaded into users' caches and reduce page download times and demands on the server.
There are other components of the Internet that can cache Web content. The most common in practice are often built into corporate and academic firewalls where they cache web resources requested by one user for the benefit of all. Some search engines such as Google also store cached content from Web sites.
Apart from the facilities built into Web servers that can ascertain when physical files have been updated, it is possible for designers of dynamically generated web pages to control the HTTP headers sent back to requesting users, so that pages are not cached when they should not be — for example Internet banking and news pages.
This helps with understanding the difference between the HTTP 'GET' and 'POST' verbs — data requested with a GET may be cached, if other conditions are met, whereas data obtained after POSTing information to the server usually will not.
Origins
The underlying ideas of the Web can be traced as far back as 1980, when, at CERN in Switzerland, the Englishman Tim Berners-Lee built ENQUIRE (referring to Enquire Within Upon Everything, a book he recalled from his youth). While it was rather different from the Web we use today, it contained many of the same core ideas (and even some of the ideas of Berners-Lee's next project after the WWW, the Semantic Web).
In March 1989, Tim Berners-Lee wrote Information Management: A Proposal, which referenced ENQUIRE and described a more elaborate information management system. With help from Robert Cailliau, he published a more formal proposal for the World Wide Web on November 12, 1990.
A NeXTcube was used by Berners-Lee as the world's first web server and also to write the first web browser, WorldWideWeb in 1990. --61.0.172.232 14:25, 19 November 2006 (UTC) By Christmas 1990, Berners-Lee had built all the tools necessary for a working Web [4]: the first Web browser (which was a Web editor as well), the first Web server and the first Web pages which described the project itself.
On August 6, 1991, he posted a short summary of the World Wide Web project on the alt.hypertext newsgroup. This date also marked the debut of the Web as a publicly available service on the Internet.
The crucial underlying concept of hypertext originated with older projects from the 1960s, such as Ted Nelson's Project Xanadu and Douglas Engelbart's oN-Line System (NLS). Both Nelson and Engelbart were in turn inspired by Vannevar Bush's microfilm-based "memex," which was described in the 1945 essay "As We May Think".
Berners-Lee's breakthrough was to marry hypertext to the Internet. In his book Weaving The Web, he explains that he had repeatedly suggested that a marriage between the two technologies was possible to members of both technical communities, but when no one took up his invitation, he finally tackled the project himself. In the process, he developed a system of globally unique identifiers for resources on the Web and elsewhere: the Uniform Resource Identifier.
The World Wide Web had a number of differences from other hypertext systems that were then available:
* The WWW required only unidirectional links rather than bidirectional ones. This made it possible for someone to link to another resource without action by the owner of that resource. It also significantly reduced the difficulty of implementing Web servers and browsers (in comparison to earlier systems), but in turn presented the chronic problem of broken links.
* Unlike predecessors such as HyperCard, the World Wide Web was non-proprietary, making it possible to develop servers and clients independently and to add extensions without licensing restrictions.
On April 30, 1993, CERN announced[1] that the World Wide Web would be free to anyone, with no fees due. Coming two months after the announcement that gopher was no longer free to use, this produced a rapid shift away from gopher and towards the Web.
An early popular Web browser was ViolaWWW which was based upon HyperCard. The World Wide Web, however, only gained critical mass with the 1993 release of the graphical Mosaic web browser by the National Center for Supercomputing Applications developed by Marc Andreessen. Prior to the release of Mosaic, graphics were not commonly mixed with text in Web pages and its popularity was less than older protocols in use over the Internet, such as Gopher protocol and Wide area information server. Mosaic's graphical user interface allowed the Web to become by far the most popular Internet protocol.
Web standards
At its core, the Web is made up of three standards:
* the Uniform Resource Identifier (URI), which is a universal system for referencing resources on the Web, such as Web pages;
* the HyperText Transfer Protocol (HTTP), which specifies how the browser and server communicate with each other; and
* the HyperText Markup Language (HTML), used to define the structure and content of hypertext documents.
Berners-Lee now heads the World Wide Web Consortium (W3C), which develops and maintains these and other standards that enable computers on the Web to effectively store and communicate different forms of information.
source
*en.wikipedia.org/wiki/World_Wide_Web