1.5
Example Networks
The subject of computer networking
covers many different kinds of networks, large and small, well known and less
well known. They have different goals, scales, and technologies. In the
following sections, we will look at some examples, to get an idea of the
variety one finds in the area of computer networking.
We will start with the Internet,
probably the best known network, and look at its history, evolution, and technology.
Then we will consider ATM, which is often used within the core of large
(telephone) networks. Technically, it is quite different from the Internet,
contrasting nicely with it. Next we will introduce Ethernet, the dominant local
area network. Finally, we will look at IEEE 802.11, the standard for wireless
LANs.
The Internet is not a network at
all, but a vast collection of different networks that use certain common
protocols and provide certain common services. It is an unusual system in that
it was not planned by anyone and is not controlled by anyone. To better
understand it, let us start from the beginning and see how it has developed and
why. For a wonderful history of the Internet, John Naughton's (2000) book is
highly recommended. It is one of those rare books that is not only fun to read,
but also has 20 pages of ibid.'s and op. cit.'s for the serious historian. Some
of the material below is based on this book.
Of course, countless technical books
have been written about the Internet and its protocols as well. For more
information, see, for example, (Maufer, 1999).
The story begins in the late 1950s.
At the height of the Cold War, the DoD wanted a command-and-control network
that could survive a nuclear war. At that time, all military communications
used the public telephone network, which was considered vulnerable. The reason
for this belief can be gleaned from Fig. 1-25(a). Here the black dots represent
telephone switching offices, each of which was connected to thousands of
telephones. These switching offices were, in turn, connected to higher-level
switching offices (toll offices), to form a national hierarchy with only a
small amount of redundancy. The vulnerability of the system was that the
destruction of a few key toll offices could fragment the system into many
isolated islands.
Figure 1-25. (a) Structure of the telephone system. (b)
Baran's proposed distributed switching system.
Around 1960, the DoD awarded a
contract to the RAND Corporation to find a solution. One of its employees, Paul
Baran, came up with the highly distributed and fault-tolerant design of Fig. 1-25(b). Since the paths between any two
switching offices were now much longer than analog signals could travel without
distortion, Baran proposed using digital packet-switching technology throughout
the system. Baran wrote several reports for the DoD describing his ideas in
detail. Officials at the Pentagon liked the concept and asked AT&T, then
the U.S. national telephone monopoly, to build a prototype. AT&T dismissed
Baran's ideas out of hand. The biggest and richest corporation in the world was
not about to allow some young whippersnapper tell it how to build a telephone
system. They said Baran's network could not be built and the idea was killed.
Several years went by and still the
DoD did not have a better command-and-control system. To understand what
happened next, we have to go back to October 1957, when the Soviet Union beat
the U.S. into space with the launch of the first artificial satellite, Sputnik.
When President Eisenhower tried to find out who was asleep at the switch, he
was appalled to find the Army, Navy, and Air Force squabbling over the
Pentagon's research budget. His immediate response was to create a single
defense research organization, ARPA, the Advanced Research Projects Agency.
ARPA had no scientists or laboratories; in fact, it had nothing more than an
office and a small (by Pentagon standards) budget. It did its work by issuing
grants and contracts to universities and companies whose ideas looked promising
to it.
For the first few years, ARPA tried
to figure out what its mission should be, but in 1967, the attention of ARPA's
then director, Larry Roberts, turned to networking. He contacted various
experts to decide what to do. One of them, Wesley Clark, suggested building a
packet-switched subnet, giving each host its own router, as illustrated in Fig. 1-10.
After some initial skepticism,
Roberts bought the idea and presented a somewhat vague paper about it at the
ACM SIGOPS Symposium on Operating System Principles held in Gatlinburg,
Tennessee in late 1967 (Roberts, 1967). Much to Roberts' surprise, another
paper at the conference described a similar system that had not only been
designed but actually implemented under the direction of Donald Davies at the
National Physical Laboratory in England. The NPL system was not a national system
(it just connected several computers on the NPL campus), but it demonstrated
that packet switching could be made to work. Furthermore, it cited Baran's now
discarded earlier work. Roberts came away from Gatlinburg determined to build
what later became known as the ARPANET.
The subnet would consist of
minicomputers called IMPs (Interface Message Processors) connected by 56-kbps
transmission lines. For high reliability, each IMP would be connected to at
least two other IMPs. The subnet was to be a datagram subnet, so if some lines
and IMPs were destroyed, messages could be automatically rerouted along
alternative paths.
Each node of the network was to
consist of an IMP and a host, in the same room, connected by a short wire. A
host could send messages of up to 8063 bits to its IMP, which would then break
these up into packets of at most 1008 bits and forward them independently
toward the destination. Each packet was received in its entirety before being
forwarded, so the subnet was the first electronic store-and-forward
packet-switching network.
ARPA then put out a tender for
building the subnet. Twelve companies bid for it. After evaluating all the
proposals, ARPA selected BBN, a consulting firm in Cambridge, Massachusetts,
and in December 1968, awarded it a contract to build the subnet and write the
subnet software. BBN chose to use specially modified Honeywell DDP-316
minicomputers with 12K 16-bit words of core memory as the IMPs. The IMPs did
not have disks, since moving parts were considered unreliable. The IMPs were
interconnected by 56-kbps lines leased from telephone companies. Although 56
kbps is now the choice of teenagers who cannot afford ADSL or cable, it was
then the best money could buy.
The software was split into two
parts: subnet and host. The subnet software consisted of the IMP end of the
host-IMP connection, the IMP-IMP protocol, and a source IMP to destination IMP
protocol designed to improve reliability. The original ARPANET design is shown
in Fig. 1-26.
Outside the subnet, software was
also needed, namely, the host end of the host-IMP connection, the host-host
protocol, and the application software. It soon became clear that BBN felt that
when it had accepted a message on a host-IMP wire and placed it on the host-IMP
wire at the destination, its job was done.
Roberts had a problem: the hosts
needed software too. To deal with it, he convened a meeting of network
researchers, mostly graduate students, at Snowbird, Utah, in the summer of
1969. The graduate students expected some network expert to explain the grand
design of the network and its software to them and then to assign each of them
the job of writing part of it. They were astounded when there was no network
expert and no grand design. They had to figure out what to do on their own.
Nevertheless, somehow an
experimental network went on the air in December 1969 with four nodes: at UCLA,
UCSB, SRI, and the University of Utah. These four were chosen because all had a
large number of ARPA contracts, and all had different and completely
incompatible host computers (just to make it more fun). The network grew
quickly as more IMPs were delivered and installed; it soon spanned the United
States. Figure 1-27 shows how rapidly the ARPANET grew in
the first 3 years.
Figure 1-27. Growth of the ARPANET. (a) December 1969. (b)
July 1970. (c) March 1971. (d) April 1972. (e) September 1972.
In addition to helping the fledgling
ARPANET grow, ARPA also funded research on the use of satellite networks and
mobile packet radio networks. In one now famous demonstration, a truck driving
around in California used the packet radio network to send messages to SRI,
which were then forwarded over the ARPANET to the East Coast, where they were
shipped to University College in London over the satellite network. This
allowed a researcher in the truck to use a computer in London while driving
around in California.
This experiment also demonstrated
that the existing ARPANET protocols were not suitable for running over multiple
networks. This observation led to more research on protocols, culminating with
the invention of the TCP/IP model and protocols (Cerf and Kahn, 1974). TCP/IP
was specifically designed to handle communication over internetworks, something
becoming increasingly important as more and more networks were being hooked up
to the ARPANET.
To encourage adoption of these new
protocols, ARPA awarded several contracts to BBN and the University of
California at Berkeley to integrate them into Berkeley UNIX. Researchers at
Berkeley developed a convenient program interface to the network (sockets) and
wrote many application, utility, and management programs to make networking
easier.
The timing was perfect. Many
universities had just acquired a second or third VAX computer and a LAN to
connect them, but they had no networking software. When 4.2BSD came along, with
TCP/IP, sockets, and many network utilities, the complete package was adopted
immediately. Furthermore, with TCP/IP, it was easy for the LANs to connect to
the ARPANET, and many did.
During the 1980s, additional
networks, especially LANs, were connected to the ARPANET. As the scale
increased, finding hosts became increasingly expensive, so DNS (Domain Name
System) was created to organize machines into domains and map host names onto
IP addresses. Since then, DNS has become a generalized, distributed database
system for storing a variety of information related to naming.
By the late 1970s, NSF (the U.S.
National Science Foundation) saw the enormous impact the ARPANET was having on
university research, allowing scientists across the country to share data and
collaborate on research projects. However, to get on the ARPANET, a university
had to have a research contract with the DoD, which many did not have. NSF's
response was to design a successor to the ARPANET that would be open to all
university research groups. To have something concrete to start with, NSF
decided to build a backbone network to connect its six supercomputer centers,
in San Diego, Boulder, Champaign, Pittsburgh, Ithaca, and Princeton. Each
supercomputer was given a little brother, consisting of an LSI-11 microcomputer
called a fuzzball. The fuzzballs were connected with 56-kbps leased lines and
formed the subnet, the same hardware technology as the ARPANET used. The
software technology was different however: the fuzzballs spoke TCP/IP right
from the start, making it the first TCP/IP WAN.
NSF also funded some (eventually
about 20) regional networks that connected to the backbone to allow users at
thousands of universities, research labs, libraries, and museums to access any
of the supercomputers and to communicate with one another. The complete
network, including the backbone and the regional networks, was called NSFNET.
It connected to the ARPANET through a link between an IMP and a fuzzball in the
Carnegie-Mellon machine room. The first NSFNET backbone is illustrated in Fig. 1-28.
NSFNET was an instantaneous success
and was overloaded from the word go. NSF immediately began planning its
successor and awarded a contract to the Michigan-based MERIT consortium to run
it. Fiber optic channels at 448 kbps were leased from MCI (since merged with
WorldCom) to provide the version 2 backbone. IBM PC-RTs were used as routers.
This, too, was soon overwhelmed, and by 1990, the second backbone was upgraded
to 1.5 Mbps.
As growth continued, NSF realized
that the government could not continue financing networking forever.
Furthermore, commercial organizations wanted to join but were forbidden by
NSF's charter from using networks NSF paid for. Consequently, NSF encouraged
MERIT, MCI, and IBM to form a nonprofit corporation, ANS (Advanced Networks and
Services), as the first step along the road to commercialization. In 1990, ANS
took over NSFNET and upgraded the 1.5-Mbps links to 45 Mbps to form ANSNET.
This network operated for 5 years and was then sold to America Online. But by
then, various companies were offering commercial IP service and it was clear
the government should now get out of the networking business.
To ease the transition and make sure
every regional network could communicate with every other regional network, NSF
awarded contracts to four different network operators to establish a NAP (Network
Access Point). These operators were PacBell (San Francisco), Ameritech
(Chicago), MFS (Washington, D.C.), and Sprint (New York City, where for NAP
purposes, Pennsauken, New Jersey counts as New York City). Every network
operator that wanted to provide backbone service to the NSF regional networks
had to connect to all the NAPs.
This arrangement meant that a packet
originating on any regional network had a choice of backbone carriers to get
from its NAP to the destination's NAP. Consequently, the backbone carriers were
forced to compete for the regional networks' business on the basis of service
and price, which was the idea, of course. As a result, the concept of a single
default backbone was replaced by a commercially-driven competitive infrastructure.
Many people like to criticize the Federal Government for not being innovative,
but in the area of networking, it was DoD and NSF that created the
infrastructure that formed the basis for the Internet and then handed it over
to industry to operate.
During the 1990s, many other
countries and regions also built national research networks, often patterned on
the ARPANET and NSFNET. These included EuropaNET and EBONE in Europe, which
started out with 2-Mbps lines and then upgraded to 34-Mbps lines. Eventually,
the network infrastructure in Europe was handed over to industry as well.
The number of networks, machines,
and users connected to the ARPANET grew rapidly after TCP/IP became the only
official protocol on January 1, 1983. When NSFNET and the ARPANET were
interconnected, the growth became exponential. Many regional networks joined
up, and connections were made to networks in Canada, Europe, and the Pacific.
Sometime in the mid-1980s, people
began viewing the collection of networks as an internet, and later as the
Internet, although there was no official dedication with some politician
breaking a bottle of champagne over a fuzzball.
The glue that holds the Internet
together is the TCP/IP reference model and TCP/IP protocol stack. TCP/IP makes
universal service possible and can be compared to the adoption of standard
gauge by the railroads in the 19th century or the adoption of common signaling
protocols by all the telephone companies.
What does it actually mean to be on
the Internet? Our definition is that a machine is on the Internet if it runs
the TCP/IP protocol stack, has an IP address, and can send IP packets to all
the other machines on the Internet. The mere ability to send and receive
electronic mail is not enough, since e-mail is gatewayed to many networks
outside the Internet. However, the issue is clouded somewhat by the fact that
millions of personal computers can call up an Internet service provider using a
modem, be assigned a temporary IP address, and send IP packets to other
Internet hosts. It makes sense to regard such machines as being on the Internet
for as long as they are connected to the service provider's router.
Traditionally (meaning 1970 to about
1990), the Internet and its predecessors had four main applications:
- E-mail. The ability to compose, send, and receive electronic mail has been around since the early days of the ARPANET and is enormously popular. Many people get dozens of messages a day and consider it their primary way of interacting with the outside world, far outdistancing the telephone and snail mail. E-mail programs are available on virtually every kind of computer these days.
- News. Newsgroups are specialized forums in which users with a common interest can exchange messages. Thousands of newsgroups exist, devoted to technical and nontechnical topics, including computers, science, recreation, and politics. Each newsgroup has its own etiquette, style, and customs, and woe betide anyone violating them.
- Remote login. Using the telnet, rlogin, or ssh programs, users anywhere on the Internet can log on to any other machine on which they have an account.
- File transfer. Using the FTP program, users can copy files from one machine on the Internet to another. Vast numbers of articles, databases, and other information are available this way.
Up until the early 1990s, the
Internet was largely populated by academic, government, and industrial
researchers. One new application, the WWW (World Wide Web) changed all that and
brought millions of new, nonacademic users to the net. This application,
invented by CERN physicist Tim Berners-Lee, did not change any of the
underlying facilities but made them easier to use. Together with the Mosaic
browser, written by Marc Andreessen at the National Center for Supercomputer
Applications in Urbana, Illinois, the WWW made it possible for a site to set up
a number of pages of information containing text, pictures, sound, and even
video, with embedded links to other pages. By clicking on a link, the user is
suddenly transported to the page pointed to by that link. For example, many
companies have a home page with entries pointing to other pages for product
information, price lists, sales, technical support, communication with
employees, stockholder information, and more.
Numerous other kinds of pages have
come into existence in a very short time, including maps, stock market tables,
library card catalogs, recorded radio programs, and even a page pointing to the
complete text of many books whose copyrights have expired (Mark Twain, Charles
Dickens, etc.). Many people also have personal pages (home pages).
Much of this growth during the 1990s
was fueled by companies called ISPs (Internet Service Providers). These are
companies that offer individual users at home the ability to call up one of
their machines and connect to the Internet, thus gaining access to e-mail, the
WWW, and other Internet services. These companies signed up tens of millions of
new users a year during the late 1990s, completely changing the character of
the network from an academic and military playground to a public utility, much
like the telephone system. The number of Internet users now is unknown, but is
certainly hundreds of millions worldwide and will probably hit 1 billion fairly
soon.
In this section we will attempt to
give a brief overview of the Internet today. Due to the many mergers between
telephone companies (telcos) and ISPs, the waters have become muddied and it is
often hard to tell who is doing what. Consequently, this description will be of
necessity somewhat simpler than reality. The big picture is shown in Fig. 1-29. Let us examine this figure piece by
piece now.
A good place to start is with a
client at home. Let us assume our client calls his or her ISP over a dial-up
telephone line, as shown in Fig. 1-29. The modem is a card within the PC that
converts the digital signals the computer produces to analog signals that can
pass unhindered over the telephone system. These signals are transferred to the
ISP's POP (Point of Presence), where they are removed from the telephone system
and injected into the ISP's regional network. From this point on, the system is
fully digital and packet switched. If the ISP is the local telco, the POP will
probably be located in the telephone switching office where the telephone wire
from the client terminates. If the ISP is not the local telco, the POP may be a
few switching offices down the road.
The ISP's regional network consists
of interconnected routers in the various cities the ISP serves. If the packet
is destined for a host served directly by the ISP, the packet is delivered to
the host. Otherwise, it is handed over to the ISP's backbone operator.
At the top of the food chain are the
major backbone operators, companies like AT&T and Sprint. They operate
large international backbone networks, with thousands of routers connected by
high-bandwidth fiber optics. Large corporations and hosting services that run
server farms (machines that can serve thousands of Web pages per second) often
connect directly to the backbone. Backbone operators encourage this direct
connection by renting space in what are called carrier hotels, basically
equipment racks in the same room as the router to allow short, fast connections
between server farms and the backbone.
If a packet given to the backbone is
destined for an ISP or company served by the backbone, it is sent to the
closest router and handed off there. However, many backbones, of varying sizes,
exist in the world, so a packet may have to go to a competing backbone. To
allow packets to hop between backbones, all the major backbones connect at the
NAPs discussed earlier. Basically, a NAP is a room full of routers, at least
one per backbone. A LAN in the room connects all the routers, so packets can be
forwarded from any backbone to any other backbone. In addition to being
interconnected at NAPs, the larger backbones have numerous direct connections
between their routers, a technique known as private peering. One of the many
paradoxes of the Internet is that ISPs who publicly compete with one another
for customers often privately cooperate to do private peering (Metz, 2001).
This ends our quick tour of the
Internet. We will have a great deal to say about the individual components and
their design, algorithms, and protocols in subsequent chapters. Also worth
mentioning in passing is that some companies have interconnected all their
existing internal networks, often using the same technology as the Internet.
These intranets are typically accessible only within the company but otherwise
work the same way as the Internet.
Since the beginning of networking, a
war has been going on between the people who support connectionless (i.e.,
datagram) subnets and the people who support connection-oriented subnets. The
main proponents of the connectionless subnets come from the ARPANET/Internet
community. Remember that DoD's original desire in funding and building the
ARPANET was to have a network that would continue functioning even after
multiple direct hits by nuclear weapons wiped out numerous routers and
transmission lines. Thus, fault tolerance was high on their priority list;
billing customers was not. This approach led to a connectionless design in
which every packet is routed independently of every other packet. As a
consequence, if some routers go down during a session, no harm is done as long
as the system can reconfigure itself dynamically so that subsequent packets can
find some route to the destination, even if it is different from that which
previous packets used.
The connection-oriented camp comes
from the world of telephone companies. In the telephone system, a caller must
dial the called party's number and wait for a connection before talking or
sending data. This connection setup establishes a route through the telephone
system that is maintained until the call is terminated. All words or packets
follow the same route. If a line or switch on the path goes down, the call is
aborted. This property is precisely what the DoD did not like about it.
Why do the telephone companies like
it then? There are two reasons:
- Quality of service.
- Billing.
By setting up a connection in
advance, the subnet can reserve resources such as buffer space and router CPU
capacity. If an attempt is made to set up a call and insufficient resources are
available, the call is rejected and the caller gets a kind of busy signal. In
this way, once a connection has been set up, the connection will get good
service. With a connectionless network, if too many packets arrive at the same
router at the same moment, the router will choke and probably lose packets. The
sender will eventually notice this and resend them, but the quality of service
will be jerky and unsuitable for audio or video unless the network is very
lightly loaded. Needless to say, providing adequate audio quality is something
telephone companies care about very much, hence their preference for
connections.
The second reason the telephone
companies like connection-oriented service is that they are accustomed to
charging for connect time. When you make a long distance call (or even a local
call outside North America) you are charged by the minute. When networks came
around, they just automatically gravitated toward a model in which charging by
the minute was easy to do. If you have to set up a connection before sending
data, that is when the billing clock starts running. If there is no connection,
they cannot charge for it.
Ironically, maintaining billing
records is very expensive. If a telephone company were to adopt a flat monthly
rate with unlimited calling and no billing or record keeping, it would probably
save a huge amount of money, despite the increased calling this policy would
generate. Political, regulatory, and other factors weigh against doing this,
however. Interestingly enough, flat rate service exists in other sectors. For
example, cable TV is billed at a flat rate per month, no matter how many
programs you watch. It could have been designed with pay-per-view as the basic
concept, but it was not, due in part to the expense of billing (and given the
quality of most television, the embarrassment factor cannot be totally discounted
either). Also, many theme parks charge a daily admission fee for unlimited
rides, in contrast to traveling carnivals, which charge by the ride.
That said, it should come as no
surprise that all networks designed by the telephone industry have had connection-oriented
subnets.
Our first example of a
connection-oriented network is X.25, which was the first public data network.
It was deployed in the 1970s at a time when telephone service was a monopoly
everywhere and the telephone company in each country expected there to be one
data network per country—theirs. To use X.25, a computer first established a
connection to the remote computer, that is, placed a telephone call. This
connection was given a connection number to be used in data transfer packets
(because multiple connections could be open at the same time). Data packets
were very simple, consisting of a 3-byte header and up to 128 bytes of data.
The header consisted of a 12-bit connection number, a packet sequence number,
an acknowledgement number, and a few miscellaneous bits. X.25 networks operated
for about a decade with mixed success.
In the 1980s, the X.25 networks were
largely replaced by a new kind of network called frame relay. The essence of
frame relay is that it is a connection-oriented network with no error control
and no flow control. Because it was connection-oriented, packets were delivered
in order (if they were delivered at all). The properties of in-order delivery,
no error control, and no flow control make frame relay akin to a wide area LAN.
Its most important application is interconnecting LANs at multiple company
offices. Frame relay enjoyed a modest success and is still in use in places
today.
Yet another, and far more important,
connection-oriented network is ATM (Asynchronous Transfer Mode). The reason for
the somewhat strange name is that in the telephone system, most transmission is
synchronous (closely tied to a clock), and ATM is not.
ATM was designed in the early 1990s
and launched amid truly incredible hype (Ginsburg, 1996; Goralski, 1995; Ibe,
1997; Kim et al., 1994; and Stallings, 2000). ATM was going to solve all the
world's networking and telecommunications problems by merging voice, data,
cable television, telex, telegraph, carrier pigeon, tin cans connected by
strings, tom-toms, smoke signals, and everything else into a single integrated
system that could do everything for everyone. It did not happen. In large part,
the problems were similar to those we described earlier concerning OSI, that
is, bad timing, technology, implementation, and politics. Having just beaten
back the telephone companies in round 1, many in the Internet community saw ATM
as Internet versus the Telcos: the Sequel. But it really was not, and this time
around even diehard datagram fanatics were aware that the Internet's quality of
service left a lot to be desired. To make a long story short, ATM was much more
successful than OSI, and it is now widely used deep within the telephone
system, often for moving IP packets. Because it is now mostly used by carriers
for internal transport, users are often unaware of its existence, but it is
definitely alive and well.
Since ATM networks are
connection-oriented, sending data requires first sending a packet to set up the
connection. As the setup packet wends its way through the subnet, all the routers
on the path make an entry in their internal tables noting the existence of the
connection and reserving whatever resources are needed for it. Connections are
often called virtual circuits, in analogy with the physical circuits used
within the telephone system. Most ATM networks also support permanent virtual
circuits, which are permanent connections between two (distant) hosts. They are
similar to leased lines in the telephone world. Each connection, temporary or
permanent, has a unique connection identifier. A virtual circuit is illustrated
in Fig. 1-30.
Once a connection has been
established, either side can begin transmitting data. The basic idea behind ATM
is to transmit all information in small, fixed-size packets called cells. The
cells are 53 bytes long, of which 5 bytes are header and 48 bytes are payload,
as shown in Fig. 1-31. Part of the header is the connection
identifier, so the sending and receiving hosts and all the intermediate routers
can tell which cells belong to which connections. This information allows each
router to know how to route each incoming cell. Cell routing is done in
hardware, at high speed. In fact, the main argument for having fixed-size cells
is that it is easy to build hardware routers to handle short, fixed-length
cells. Variable-length IP packets have to be routed by software, which is a
slower process. Another plus of ATM is that the hardware can be set up to copy
one incoming cell to multiple output lines, a property that is required for
handling a television program that is being broadcast to many receivers.
Finally, small cells do not block any line for very long, which makes
guaranteeing quality of service easier.
All cells follow the same route to
the destination. Cell delivery is not guaranteed, but their order is. If cells
1 and 2 are sent in that order, then if both arrive, they will arrive in that
order, never first 2 then 1. But either or both of them can be lost along the
way. It is up to higher protocol levels to recover from lost cells. Note that
although this guarantee is not perfect, it is better than what the Internet
provides. There packets can not only be lost, but delivered out of order as
well. ATM, in contrast, guarantees never to deliver cells out of order.
ATM networks are organized like
traditional WANs, with lines and switches (routers). The most common speeds for
ATM networks are 155 Mbps and 622 Mbps, although higher speeds are also
supported. The 155-Mbps speed was chosen because this is about what is needed
to transmit high definition television. The exact choice of 155.52 Mbps was
made for compatibility with AT&T's SONET transmission system, something.
The 622 Mbps speed was chosen so that four 155-Mbps channels could be sent over
it.
ATM has its own reference model,
different from the OSI model and also different from the TCP/IP model. This
model is shown in Fig. 1-32. It consists of three layers, the
physical, ATM, and ATM adaptation layers, plus whatever users want to put on
top of that.
The physical layer deals with the
physical medium: voltages, bit timing, and various other issues. ATM does not
prescribe a particular set of rules but instead says that ATM cells can be sent
on a wire or fiber by themselves, but they can also be packaged inside the
payload of other carrier systems. In other words, ATM has been designed to be
independent of the transmission medium.
The ATM layer deals with cells and
cell transport. It defines the layout of a cell and tells what the header
fields mean. It also deals with establishment and release of virtual circuits.
Congestion control is also located here.
Because most applications do not
want to work directly with cells (although some may), a layer above the ATM
layer has been defined to allow users to send packets larger than a cell. The
ATM interface segments these packets, transmits the cells individually, and
reassembles them at the other end. This layer is the AAL (ATM Adaptation Layer).
Unlike the earlier two-dimensional
reference models, the ATM model is defined as being three-dimensional, as shown
in Fig. 1-32. The user plane deals with data
transport, flow control, error correction, and other user functions. In
contrast, the control plane is concerned with connection management. The layer
and plane management functions relate to resource management and interlayer
coordination.
The physical and AAL layers are each
divided into two sublayers, one at the bottom that does the work and a
convergence sublayer on top that provides the proper interface to the layer
above it. The functions of the layers and sublayers are given in Fig. 1-33.
The PMD (Physical Medium Dependent)
sublayer interfaces to the actual cable. It moves the bits on and off and
handles the bit timing. For different carriers and cables, this layer will be
different.
The other sublayer of the physical
layer is the TC (Transmission Convergence) sublayer. When cells are
transmitted, the TC layer sends them as a string of bits to the PMD layer.
Doing this is easy. At the other end, the TC sublayer gets a pure incoming bit
stream from the PMD sublayer. Its job is to convert this bit stream into a cell
stream for the ATM layer. It handles all the issues related to telling where
cells begin and end in the bit stream. In the ATM model, this functionality is
in the physical layer. In the OSI model and in pretty much all other networks,
the job of framing, that is, turning a raw bit stream into a sequence of frames
or cells, is the data link layer's task.
As we mentioned earlier, the ATM
layer manages cells, including their generation and transport. Most of the
interesting aspects of ATM are located here. It is a mixture of the OSI data
link and network layers; it is not split into sublayers.
The AAL layer is split into a SAR (Segmentation
And Reassembly) sublayer and a CS (Convergence Sublayer). The lower sublayer
breaks up packets into cells on the transmission side and puts them back
together again at the destination. The upper sublayer makes it possible to have
ATM systems offer different kinds of services to different applications (e.g.,
file transfer and video on demand have different requirements concerning error
handling, timing, etc.).
As it is probably mostly downhill
for ATM from now on, we will not discuss it further in this book. Nevertheless,
since it has a substantial installed base, it will probably be around for at
least a few more years. For more information about ATM, see (Dobrowski and
Grise, 2001; and Gadecki and Heckart, 1997).
Both the Internet and ATM were
designed for wide area networking. However, many companies, universities, and
other organizations have large numbers of computers that must be connected.
This need gave rise to the local area network. In this section we will say a
little bit about the most popular LAN, Ethernet.
The story starts out in pristine
Hawaii in the early 1970s. In this case, ''pristine'' can be interpreted as
''not having a working telephone system.'' While not being interrupted by the
phone all day long makes life more pleasant for vacationers, it did not make
life more pleasant for researcher Norman Abramson and his colleagues at the
University of Hawaii who were trying to connect users on remote islands to the
main computer in Honolulu. Stringing their own cables under the Pacific Ocean
was not in the cards, so they looked for a different solution.
The one they found was short-range
radios. Each user terminal was equipped with a small radio having two
frequencies: upstream (to the central computer) and downstream (from the
central computer). When the user wanted to contact the computer, it just
transmitted a packet containing the data in the upstream channel. If no one
else was transmitting at that instant, the packet probably got through and was
acknowledged on the downstream channel. If there was contention for the
upstream channel, the terminal noticed the lack of acknowledgement and tried
again. Since there was only one sender on the downstream channel (the central
computer), there were never collisions there. This system, called ALOHANET,
worked fairly well under conditions of low traffic but bogged down badly when
the upstream traffic was heavy.
About the same time, a student named
Bob Metcalfe got his bachelor's degree at M.I.T. and then moved up the river to
get his Ph.D. at Harvard. During his studies, he was exposed to Abramson's
work. He became so interested in it that after graduating from Harvard, he
decided to spend the summer in Hawaii working with Abramson before starting
work at Xerox PARC (Palo Alto Research Center). When he got to PARC, he saw
that the researchers there had designed and built what would later be called
personal computers. But the machines were isolated. Using his knowledge of
Abramson's work, he, together with his colleague David Boggs, designed and
implemented the first local area network (Metcalfe and Boggs, 1976).
They called the system Ethernet
after the luminiferous ether, through which electromagnetic radiation was once
thought to propagate. (When the 19th century British physicist James Clerk
Maxwell discovered that electromagnetic radiation could be described by a wave
equation, scientists assumed that space must be filled with some ethereal
medium in which the radiation was propagating. Only after the famous Michelson-Morley
experiment in 1887 did physicists discover that electromagnetic radiation could
propagate in a vacuum.)
The transmission medium here was not
a vacuum, but a thick coaxial cable (the ether) up to 2.5 km long (with
repeaters every 500 meters). Up to 256 machines could be attached to the system
via transceivers screwed onto the cable. A cable with multiple machines
attached to it in parallel is called a multidrop cable. The system ran at 2.94
Mbps. A sketch of its architecture is given in Fig. 1-34. Ethernet had a major improvement over
ALOHANET: before transmitting, a computer first listened to the cable to see if
someone else was already transmitting. If so, the computer held back until the
current transmission finished. Doing so avoided interfering with existing
transmissions, giving a much higher efficiency. ALOHANET did not work like this
because it was impossible for a terminal on one island to sense the
transmission of a terminal on a distant island. With a single cable, this
problem does not exist.
Despite the computer listening
before transmitting, a problem still arises: what happens if two or more
computers all wait until the current transmission completes and then all start
at once? The solution is to have each computer listen during its own
transmission and if it detects interference, jam the ether to alert all
senders. Then back off and wait a random time before retrying. If a second
collision happens, the random waiting time is doubled, and so on, to spread out
the competing transmissions and give one of them a chance to go first.
The Xerox Ethernet was so successful
that DEC, Intel, and Xerox drew up a standard in 1978 for a 10-Mbps Ethernet,
called the DIX standard. With two minor changes, the DIX standard became the
IEEE 802.3 standard in 1983.
Unfortunately for Xerox, it already
had a history of making seminal inventions (such as the personal computer) and
then failing to commercialize on them, a story told in Fumbling the Future
(Smith and Alexander, 1988). When Xerox showed little interest in doing
anything with Ethernet other than helping standardize it, Metcalfe formed his
own company, 3Com, to sell Ethernet adapters for PCs. It has sold over 100
million of them.
Ethernet continued to develop and is
still developing. New versions at 100 Mbps, 1000 Mbps, and still higher have
come out. Also the cabling has improved, and switching and other features have
been added.
In passing, it is worth mentioning
that Ethernet (IEEE 802.3) is not the only LAN standard. The committee also
standardized a token bus (802.4) and a token ring (802.5). The need for three
more-or-less incompatible standards has little to do with technology and
everything to do with politics. At the time of standardization, General Motors
was pushing a LAN in which the topology was the same as Ethernet (a linear
cable) but computers took turns in transmitting by passing a short packet
called a token from computer to computer. A computer could only send if it
possessed the token, thus avoiding collisions. General Motors announced that
this scheme was essential for manufacturing cars and was not prepared to budge
from this position. This announcement notwithstanding, 802.4 has basically
vanished from sight.
Similarly, IBM had its own favorite:
its proprietary token ring. The token was passed around the ring and whichever
computer held the token was allowed to transmit before putting the token back
on the ring. Unlike 802.4, this scheme, standardized as 802.5, is still in use
at some IBM sites, but virtually nowhere outside of IBM sites. However, work is
progressing on a gigabit version (802.5v), but it seems unlikely that it will
ever catch up with Ethernet. In short, there was a war between Ethernet, token
bus, and token ring, and Ethernet won, mostly because it was there first and
the challengers were not as good.
Almost as soon as notebook computers
appeared, many people had a dream of walking into an office and magically
having their notebook computer be connected to the Internet. Consequently,
various groups began working on ways to accomplish this goal. The most
practical approach is to equip both the office and the notebook computers with
short-range radio transmitters and receivers to allow them to communicate. This
work rapidly led to wireless LANs being marketed by a variety of companies.
The trouble was that no two of them
were compatible. This proliferation of standards meant that a computer equipped
with a brand X radio would not work in a room equipped with a brand Y base
station. Finally, the industry decided that a wireless LAN standard might be a
good idea, so the IEEE committee that standardized the wired LANs was given the
task of drawing up a wireless LAN standard. The standard it came up with was
named 802.11. A common slang name for it is WiFi. It is an important standard
and deserves respect, so we will call it by its proper name, 802.11.
The proposed standard had to work in
two modes:
- In the presence of a base station.
- In the absence of a base station.
In the former case, all
communication was to go through the base station, called an access point in
802.11 terminology. In the latter case, the computers would just send to one
another directly. This mode is now sometimes called ad hoc networking. A
typical example is two or more people sitting down together in a room not
equipped with a wireless LAN and having their computers just communicate
directly. The two modes are illustrated in Fig. 1-35.
The first decision was the easiest:
what to call it. All the other LAN standards had numbers like 802.1, 802.2,
802.3, up to 802.10, so the wireless LAN standard was dubbed 802.11. The rest
was harder.
In particular, some of the many
challenges that had to be met were: finding a suitable frequency band that was
available, preferably worldwide; dealing with the fact that radio signals have
a finite range; ensuring that users' privacy was maintained; taking limited
battery life into account; worrying about human safety (do radio waves cause
cancer?); understanding the implications of computer mobility; and finally,
building a system with enough bandwidth to be economically viable.
At the time the standardization
process started (mid-1990s), Ethernet had already come to dominate local area
networking, so the committee decided to make 802.11 compatible with Ethernet
above the data link layer. In particular, it should be possible to send an IP
packet over the wireless LAN the same way a wired computer sent an IP packet
over Ethernet. Nevertheless, in the physical and data link layers, several
inherent differences with Ethernet exist and had to be dealt with by the
standard.
First, a computer on Ethernet always
listens to the ether before transmitting. Only if the ether is idle does the
computer begin transmitting. With wireless LANs, that idea does not work so
well. To see why, examine Fig. 1-36. Suppose that computer A is
transmitting to computer B, but the radio range of A's transmitter is too short
to reach computer C. If C wants to transmit to B it can listen to the ether
before starting, but the fact that it does not hear anything does not mean that
its transmission will succeed. The 802.11 standard had to solve this problem.
The second problem that had to be
solved is that a radio signal can be reflected off solid objects, so it may be
received multiple times (along multiple paths). This interference results in
what is called multipath fading.
The third problem is that a great
deal of software is not aware of mobility. For example, many word processors
have a list of printers that users can choose from to print a file. When the
computer on which the word processor runs is taken into a new environment, the
built-in list of printers becomes invalid.
The fourth problem is that if a
notebook computer is moved away from the ceiling-mounted base station it is
using and into the range of a different base station, some way of handing it
off is needed. Although this problem occurs with cellular telephones, it does
not occur with Ethernet and needed to be solved. In particular, the network
envisioned consists of multiple cells, each with its own base station, but with
the base stations connected by Ethernet, as shown in Fig. 1-37. From the outside, the entire system
should look like a single Ethernet. The connection between the 802.11 system
and the outside world is called a portal.
After some work, the committee came
up with a standard in 1997 that addressed these and other concerns. The
wireless LAN it described ran at either 1 Mbps or 2 Mbps. Almost immediately,
people complained that it was too slow, so work began on faster standards. A
split developed within the committee, resulting in two new standards in 1999.
The 802.11a standard uses a wider frequency band and runs at speeds up to 54
Mbps. The 802.11b standard uses the same frequency band as 802.11, but uses a
different modulation technique to achieve 11 Mbps. Some people see this as
psychologically important since 11 Mbps is faster than the original wired
Ethernet. It is likely that the original 1-Mbps 802.11 will die off quickly,
but it is not yet clear which of the new standards will win out.
To make matters even more
complicated than they already were, the 802 committee has come up with yet
another variant, 802.11g, which uses the modulation technique of 802.11a but
the frequency band of 802.11b.
That
802.11 is going to cause a revolution in computing and Internet access is now
beyond any doubt. Airports, train stations, hotels, shopping malls, and
universities are rapidly installing it. Even upscale coffee shops are
installing 802.11 so that the assembled yuppies can surf the Web while drinking
their lattes. It is likely that 802.11 will do to the Internet what notebook
computers did to computing: make it mobile.
No comments:
Post a Comment
silahkan membaca dan berkomentar