summaryrefslogtreecommitdiffstats
path: root/contrib/awk/doc/gawkinet.texi
diff options
context:
space:
mode:
Diffstat (limited to 'contrib/awk/doc/gawkinet.texi')
-rw-r--r--contrib/awk/doc/gawkinet.texi5075
1 files changed, 0 insertions, 5075 deletions
diff --git a/contrib/awk/doc/gawkinet.texi b/contrib/awk/doc/gawkinet.texi
deleted file mode 100644
index 2ffb581..0000000
--- a/contrib/awk/doc/gawkinet.texi
+++ /dev/null
@@ -1,5075 +0,0 @@
-\input texinfo @c -*-texinfo-*-
-@c %**start of header (This is for running Texinfo on a region.)
-@setfilename gawkinet.info
-@settitle TCP/IP Internetworking With @command{gawk}
-@c %**end of header (This is for running Texinfo on a region.)
-
-@c inside ifinfo for older versions of texinfo.tex
-@ifinfo
-@dircategory GNU Packages
-@direntry
-* Gawkinet: (gawkinet). TCP/IP Internetworking With @command{gawk}.
-@end direntry
-@end ifinfo
-
-@iftex
-@set DOCUMENT book
-@set CHAPTER chapter
-@set SECTION section
-@set DARKCORNER @inmargin{@image{lflashlight,1cm}, @image{rflashlight,1cm}}
-@end iftex
-@ifinfo
-@set DOCUMENT Info file
-@set CHAPTER major node
-@set SECTION node
-@set DARKCORNER (d.c.)
-@end ifinfo
-@ifhtml
-@set DOCUMENT web page
-@set CHAPTER chapter
-@set SECTION section
-@set DARKCORNER (d.c.)
-@end ifhtml
-
-@set FSF
-
-@set FN file name
-@set FFN File Name
-
-@c merge the function and variable indexes into the concept index
-@ifinfo
-@synindex fn cp
-@synindex vr cp
-@end ifinfo
-@iftex
-@syncodeindex fn cp
-@syncodeindex vr cp
-@end iftex
-
-@c If "finalout" is commented out, the printed output will show
-@c black boxes that mark lines that are too long. Thus, it is
-@c unwise to comment it out when running a master in case there are
-@c overfulls which are deemed okay.
-
-@iftex
-@finalout
-@end iftex
-
-@smallbook
-
-@c Special files are described in chapter 6 Printing Output under
-@c 6.7 Special File Names in gawk. I think the networking does not
-@c fit into that chapter, thus this separate document. At over 50
-@c pages, I think this is the right decision. ADR.
-
-@set TITLE TCP/IP Internetworking With @command{gawk}
-@set EDITION 1.1
-@set UPDATE-MONTH March, 2001
-@c gawk versions:
-@set VERSION 3.1
-@set PATCHLEVEL 0
-
-@ifinfo
-This file documents the networking features in GNU @command{awk}.
-
-This is Edition @value{EDITION} of @cite{@value{TITLE}},
-for the @value{VERSION}.@value{PATCHLEVEL} (or later) version of the GNU
-implementation of AWK.
-
-Copyright (C) 2000, 2001 Free Software Foundation, Inc.
-
-Permission is granted to copy, distribute and/or modify this document
-under the terms of the GNU Free Documentation License, Version 1.1 or
-any later version published by the Free Software Foundation; with the
-Invariant Sections being ``GNU General Public License'', the Front-Cover
-texts being (a) (see below), and with the Back-Cover Texts being (b)
-(see below). A copy of the license is included in the section entitled
-``GNU Free Documentation License''.
-
-@enumerate a
-@item
-``A GNU Manual''
-
-@item
-``You have freedom to copy and modify this GNU Manual, like GNU
-software. Copies published by the Free Software Foundation raise
-funds for GNU development.''
-@end enumerate
-@end ifinfo
-
-@setchapternewpage odd
-
-@titlepage
-@title @value{TITLE}
-@subtitle Edition @value{EDITION}
-@subtitle @value{UPDATE-MONTH}
-@author J@"urgen Kahrs
-@author with Arnold D. Robbins
-
-@c Include the Distribution inside the titlepage environment so
-@c that headings are turned off. Headings on and off do not work.
-
-@page
-@vskip 0pt plus 1filll
-Copyright @copyright{} 2000, 2001 Free Software Foundation, Inc.
-@sp 1
-@b{User Friendly} Copyright @copyright{} 2000 J.D.@: ``Iliad'' Frazier.
-Reprinted by permission.
-@sp 2
-
-This is Edition @value{EDITION} of @cite{@value{TITLE}},
-for the @value{VERSION}.@value{PATCHLEVEL} (or later) version of the GNU
-implementation of AWK.
-
-@sp 2
-Published by:
-@sp 1
-
-Free Software Foundation @*
-59 Temple Place --- Suite 330 @*
-Boston, MA 02111-1307 USA @*
-Phone: +1-617-542-5942 @*
-Fax: +1-617-542-2652 @*
-Email: @email{gnu@@gnu.org} @*
-URL: @uref{http://www.gnu.org/} @*
-
-ISBN 1-882114-93-0 @*
-
-Permission is granted to copy, distribute and/or modify this document
-under the terms of the GNU Free Documentation License, Version 1.1 or
-any later version published by the Free Software Foundation; with the
-Invariant Sections being ``GNU General Public License'', the Front-Cover
-texts being (a) (see below), and with the Back-Cover Texts being (b)
-(see below). A copy of the license is included in the section entitled
-``GNU Free Documentation License''.
-
-@enumerate a
-@item
-``A GNU Manual''
-
-@item
-``You have freedom to copy and modify this GNU Manual, like GNU
-software. Copies published by the Free Software Foundation raise
-funds for GNU development.''
-@end enumerate
-@c @sp 2
-@c Cover art by ?????.
-@end titlepage
-
-@iftex
-@headings off
-@evenheading @thispage@ @ @ @strong{@value{TITLE}} @| @|
-@oddheading @| @| @strong{@thischapter}@ @ @ @thispage
-@end iftex
-
-@ifinfo
-@node Top, Preface, (dir), (dir)
-@top General Introduction
-@comment node-name, next, previous, up
-
-This file documents the networking features in GNU Awk (@command{gawk})
-version 3.1 and later.
-@end ifinfo
-
-@menu
-* Preface:: About this document.
-* Introduction:: About networkiing.
-* Using Networking:: Some examples.
-* Some Applications and Techniques:: More extended examples.
-* Links:: Where to find the stuff mentioned in this
- document.
-* GNU Free Documentation License:: The license for this document.
-* Index:: The index.
-
-@detailmenu
-* Stream Communications:: Sending data streams.
-* Datagram Communications:: Sending self-contained messages.
-* The TCP/IP Protocols:: How these models work in the Internet.
-* Basic Protocols:: The basic protocols.
-* Ports:: The idea behind ports.
-* Making Connections:: Making TCP/IP connections.
-* Gawk Special Files:: How to do @command{gawk} networking.
-* Special File Fields:: The fields in the special file name.
-* Comparing Protocols:: Differences between the protocols.
-* File /inet/tcp:: The TCP special file.
-* File /inet/udp:: The UDB special file.
-* File /inet/raw:: The RAW special file.
-* TCP Connecting:: Making a TCP connection.
-* Troubleshooting:: Troubleshooting TCP/IP connections.
-* Interacting:: Interacting with a service.
-* Setting Up:: Setting up a service.
-* Email:: Reading email.
-* Web page:: Reading a Web page.
-* Primitive Service:: A primitive Web service.
-* Interacting Service:: A Web service with interaction.
-* CGI Lib:: A simple CGI library.
-* Simple Server:: A simple Web server.
-* Caveats:: Network programming caveats.
-* Challenges:: Where to go from here.
-* PANIC:: An Emergency Web Server.
-* GETURL:: Retrieving Web Pages.
-* REMCONF:: Remote Configuration Of Embedded Systems.
-* URLCHK:: Look For Changed Web Pages.
-* WEBGRAB:: Extract Links From A Page.
-* STATIST:: Graphing A Statistical Distribution.
-* MAZE:: Walking Through A Maze In Virtual Reality.
-* MOBAGWHO:: A Simple Mobile Agent.
-* STOXPRED:: Stock Market Prediction As A Service.
-* PROTBASE:: Searching Through A Protein Database.
-@end detailmenu
-@end menu
-
-@contents
-
-@node Preface, Introduction, Top, Top
-@unnumbered Preface
-
-In May of 1997, J@"urgen Kahrs felt the need for network access
-from @command{awk}, and, with a little help from me, set about adding
-features to do this for @command{gawk}. At that time, he
-wrote the bulk of this @value{DOCUMENT}.
-
-The code and documentation were added to the @command{gawk} 3.1 development
-tree, and languished somewhat until I could finally get
-down to some serious work on that version of @command{gawk}.
-This finally happened in the middle of 2000.
-
-Meantime, J@"urgen wrote an article about the Internet special
-files and @samp{|&} operator for @cite{Linux Journal}, and made a
-networking patch for the production versions of @command{gawk}
-available from his home page.
-In August of 2000 (for @command{gawk} 3.0.6), this patch
-also made it to the main GNU @command{ftp} distribution site.
-
-For release with @command{gawk}, I edited J@"urgen's prose
-for English grammar and style, as he is not a native English
-speaker. I also
-rearranged the material somewhat for what I felt was a better order of
-presentation, and (re)wrote some of the introductory material.
-
-The majority of this document and the code are his work, and the
-high quality and interesting ideas speak for themselves. It is my
-hope that these features will be of significant value to the @command{awk}
-community.
-
-@sp 1
-@noindent
-Arnold Robbins @*
-Nof Ayalon, ISRAEL @*
-March, 2001
-
-@node Introduction, Using Networking, Preface, Top
-@chapter Networking Concepts
-
-This @value{CHAPTER} provides a (necessarily) brief intoduction to
-computer networking concepts. For many applications of @command{gawk}
-to TCP/IP networking, we hope that this is enough. For more
-advanced tasks, you will need deeper background, and it may be necessary
-to switch to lower-level programming in C or C++.
-
-There are two real-life models for the way computers send messages
-to each other over a network. While the analogies are not perfect,
-they are close enough to convey the major concepts.
-These two models are the phone system (reliable byte-stream communications),
-and the postal system (best-effort datagrams).
-
-@menu
-* Stream Communications:: Sending data streams.
-* Datagram Communications:: Sending self-contained messages.
-* The TCP/IP Protocols:: How these models work in the Internet.
-* Making Connections:: Making TCP/IP connections.
-@end menu
-
-@node Stream Communications, Datagram Communications, Introduction, Introduction
-@section Reliable Byte-streams (Phone Calls)
-
-When you make a phone call, the following steps occur:
-
-@enumerate
-@item
-You dial a number.
-
-@item
-The phone system connects to the called party, telling
-them there is an incoming call. (Their phone rings.)
-
-@item
-The other party answers the call, or, in the case of a
-computer network, refuses to answer the call.
-
-@item
-Assuming the other party answers, the connection between
-you is now a @dfn{duplex} (two-way), @dfn{reliable} (no data lost),
-sequenced (data comes out in the order sent) data stream.
-
-@item
-You and your friend may now talk freely, with the phone system
-moving the data (your voices) from one end to the other.
-From your point of view, you have a direct end-to-end
-connection with the person on the other end.
-@end enumerate
-
-The same steps occur in a duplex reliable computer networking connection.
-There is considerably more overhead in setting up the communications,
-but once it's done, data moves in both directions, reliably, in sequence.
-
-@node Datagram Communications, The TCP/IP Protocols, Stream Communications, Introduction
-@section Best-effort Datagrams (Mailed Letters)
-
-Suppose you mail three different documents to your office on the
-other side of the country on two different days. Doing so
-entails the following.
-
-@enumerate
-@item
-Each document travels in its own envelope.
-
-@item
-Each envelope contains both the sender and the
-recipient address.
-
-@item
-Each envelope may travel a different route to its destination.
-
-@item
-The envelopes may arrive in a different order from the one
-in which they were sent.
-
-@item
-One or more may get lost in the mail.
-(Although, fortunately, this does not occur very often.)
-
-@item
-In a computer network, one or more @dfn{packets}
-may also arrive multiple times. (This doesn't happen
-with the postal system!)
-
-@end enumerate
-
-The important characteristics of datagram communications, like
-those of the postal system are thus:
-
-@itemize @bullet
-@item
-Delivery is ``best effort;'' the data may never get there.
-
-@item
-Each message is self-contained, including the source and
-destination addresses.
-
-@item
-Delivery is @emph{not} sequenced; packets may arrive out
-of order, and/or multiple times.
-
-@item
-Unlike the phone system, overhead is considerably lower.
-It is not necessary to set up the call first.
-@end itemize
-
-The price the user pays for the lower overhead of datagram communications
-is exactly the lower reliability; it is often necessary for user-level
-protocols that use datagram communications to add their own reliabilty
-features on top of the basic communications.
-
-@node The TCP/IP Protocols, Making Connections, Datagram Communications, Introduction
-@section The Internet Protocols
-
-The Internet Protocol Suite (usually referred as just TCP/IP)@footnote{
-It should be noted that although the Internet seems to have conquered the
-world, there are other networking protocol suites in existence and in use.}
-consists of a number of different protocols at different levels or ``layers.''
-For our purposes, three protocols provide the fundamental communications
-mechanisms. All other defined protocols are referred to as user-level
-protocols (e.g., HTTP, used later in this @value{DOCUMENT}).
-
-@menu
-* Basic Protocols:: The basic protocols.
-* Ports:: The idea behind ports.
-@end menu
-
-@node Basic Protocols, Ports, The TCP/IP Protocols, The TCP/IP Protocols
-@subsection The Basic Internet Protocols
-
-@table @asis
-@item IP
-The Internet Protocol. This protocol is almost never used directly by
-applications. It provides the basic packet delivery and routing infrastructure
-of the Internet. Much like the phone company's switching centers or the Post
-Office's trucks, it is not of much day-to-day interest to the regular user
-(or programmer).
-It happens to be a best effort datagram protocol.
-
-@item UDP
-The User Datagram Protocol. This is a best effort datagram protocol.
-It provides a small amount of extra reliability over IP, and adds
-the notion of @dfn{ports}, described in @ref{Ports, ,TCP and UDP Ports}.
-
-@item TCP
-The Transmission Control Protocol. This is a duplex, reliable, sequenced
-byte-stream protocol, again layered on top of IP, and also providing the
-notion of ports. This is the protocol that you will most likely use
-when using @command{gawk} for network programming.
-@end table
-
-All other user-level protocols use either TCP or UDP to do their basic
-communications. Examples are SMTP (Simple Mail Transfer Protocol),
-FTP (File Transfer Protocol) and HTTP (HyperText Transfer Protocol).
-@cindex SMTP
-@cindex FTP
-@cindex HTTP
-
-@node Ports, , Basic Protocols, The TCP/IP Protocols
-@subsection TCP and UDP Ports
-
-In the postal system, the address on an envelope indicates a physical
-location, such as a residence or office building. But there may be
-more than one person at the location; thus you have to further quantify
-the recipient by putting a person or company name on the envelope.
-
-In the phone system, one phone number may represent an entire company,
-in which case you need a person's extension number in order to
-reach that individual directly. Or, when you call a home, you have to
-say, ``May I please speak to ...'' before talking to the person directly.
-
-IP networking provides the concept of addressing. An IP address represents
-a particular computer, but no more. In order to reach the mail service
-on a system, or the FTP or WWW service on a system, you have to have some
-way to further specify which service you want. In the Internet Protocol suite,
-this is done with @dfn{port numbers}, which represent the services, much
-like an extension number used with a phone number.
-
-Port numbers are 16-bit integers. Unix and Unix-like systems reserve ports
-below 1024 for ``well known'' services, such as SMTP, FTP, and HTTP.
-Numbers above 1024 may be used by any application, although there is no
-promise made that a particular port number is always available.
-
-@node Making Connections, , The TCP/IP Protocols, Introduction
-@section Making TCP/IP Connections (And Some Terminology)
-
-Two terms come up repeatedly when discussing networking:
-@dfn{client} and @dfn{server}. For now, we'll discuss these terms
-at the @dfn{connection level}, when first establishing connections
-between two processes on different systems over a network.
-(Once the connection is established, the higher level, or
-@dfn{application level} protocols,
-such as HTTP or FTP, determine who is the client and who is the
-server. Often, it turns out that the client and server are the
-same in both roles.)
-
-@cindex server
-The @dfn{server} is the system providing the service, such as the
-web server or email server. It is the @dfn{host} (system) which
-is @emph{connected to} in a transaction.
-For this to work though, the server must be expecting connections.
-Much as there has to be someone at the office building to answer
-the phone@footnote{In the days before voice mail systems!}, the
-server process (usually) has to be started first and waiting
-for a connection.
-
-@cindex client
-The @dfn{client} is the system requesting the service.
-It is the system @emph{initiating the connection} in a transaction.
-(Just as when you pick up the phone to call an office or store.)
-
-In the TCP/IP framework, each end of a connection is represented by a pair
-of (@var{address}, @var{port}) pairs. For the duration of the connection,
-the ports in use at each end are unique, and cannot be used simultaneously
-by other processes on the same system. (Only after closing a connection
-can a new one be built up on the same port. This is contrary to the usual
-behavior of fully developed web servers which have to avoid situations
-in which they are not reachable. We have to pay this price in order to
-enjoy the benefits of a simple communication paradigm in @command{gawk}.)
-
-@cindex blocking
-@cindex synchronous communications
-Furthermore, once the connection is established, communications
-are @dfn{synchronous}. I.e., each end waits on the other to finish
-transmitting, before replying. This is much like two people in a phone
-conversation. While both could talk simultaneously, doing so usually
-doesn't work too well.
-
-In the case of TCP, the synchronicity is enforced by the protocol when
-sending data. Data writes @dfn{block} until the data have been received on the
-other end. For both TCP and UDP, data reads block until there is incoming
-data waiting to be read. This is summarized in the following table,
-where an ``X'' indicates that the given action blocks.
-
-@ifnottex
-@multitable {Protocol} {Reads} {Writes}
-@item TCP @tab X @tab X
-@item UDP @tab X @tab
-@item RAW @tab X @tab
-@end multitable
-@end ifnottex
-@tex
-\centerline{
-\vbox{\bigskip % space above the table (about 1 linespace)
-% Because we have vertical rules, we can't let TeX insert interline space
-% in its usual way.
-\offinterlineskip
-\halign{\hfil\strut# &\vrule #& \hfil#\hfil& \hfil#\hfil\cr
-Protocol&&\quad Reads\quad &Writes\cr
-\noalign{\hrule}
-\omit&height 2pt\cr
-\noalign{\hrule height0pt}% without this the rule does not extend; why?
-TCP&&X&X\cr
-UDP&&X&\cr
-RAW&&X&\cr
-}}}
-@end tex
-
-@node Using Networking, Some Applications and Techniques, Introduction, Top
-@comment node-name, next, previous, up
-@chapter Networking With @command{gawk}
-
-@cindex network
-The @command{awk} programming language was originally developed as a
-pattern-matching language for writing short programs to perform
-data manipulation tasks.
-@command{awk}'s strength is the manipulation of textual data
-that is stored in files.
-It was never meant to be used for networking purposes.
-To exploit its features in a
-networking context, it's necessary to use an access mode for network connections
-that resembles the access of files as closely as possible.
-
-@cindex Perl
-@cindex Python
-@cindex Tcl/Tk
-@command{awk} is also meant to be a prototyping language. It is used
-to demonstrate feasibility and to play with features and user interfaces.
-This can be done with file-like handling of network
-connections.
-@command{gawk} trades the lack
-of many of the advanced features of the TCP/IP family of protocols
-for the convenience of simple connection handling.
-The advanced
-features are available when programming in C or Perl. In fact, the
-network programming
-in this @value{CHAPTER}
-is very similar to what is described in books like
-@cite{Internet Programming with Python},
-@cite{Advanced Perl Programming},
-or
-@cite{Web Client Programming with Perl}.
-But it's done here without first having to learn object-oriented ideology, underlying
-languages such as Tcl/Tk, Perl, Python, or all of the libraries necessary to
-extend these languages before they are ready for the Internet.
-
-This @value{CHAPTER} demonstrates how to use the TCP protocol. The
-other protocols are much less important for most users (UDP) or even
-untractable (RAW).
-
-@menu
-* Gawk Special Files:: How to do @command{gawk} networking.
-* TCP Connecting:: Making a TCP connection.
-* Troubleshooting:: Troubleshooting TCP/IP connections.
-* Interacting:: Interacting with a service.
-* Setting Up:: Setting up a service.
-* Email:: Reading email.
-* Web page:: Reading a Web page.
-* Primitive Service:: A primitive Web service.
-* Interacting Service:: A Web service with interaction.
-* Simple Server:: A simple Web server.
-* Caveats:: Network programming caveats.
-* Challenges:: Where to go from here.
-@end menu
-
-@node Gawk Special Files, TCP Connecting, Using Networking, Using Networking
-@comment node-name, next, previous, up
-@section @command{gawk} Networking Mechanisms
-@cindex network
-
-The @samp{|&} operator introduced in @command{gawk} 3.1 for use in
-communicating with a @dfn{co-process} is described in
-@ref{Two-way I/O, ,Two-way Communications With Another Process, gawk, GAWK: Effective AWK Programming}.
-It shows how to do two-way I/O to a
-separate process, sending it data with @code{print} or @code{printf} and
-reading data with @code{getline}. If you haven't read it already, you should
-detour there to do so.
-
-@command{gawk} transparently extends the two-way I/O mechanism to simple networking through
-the use of special @value{FN}s. When a ``co-process'' is started that matches
-the special files we are about to describe, @command{gawk} creates the appropriate network
-connection, and then two-way I/O proceeds as usual.
-
-At the C, C++ (and basic Perl) level, networking is accomplished
-via @dfn{sockets}, an Application Programming Interface (API) originally
-developed at the University of California at Berkeley that is now used
-almost universally for TCP/IP networking.
-Socket level programming, while fairly straightforward, requires paying
-attention to a number of details, as well as using binary data. It is not
-well-suited for use from a high-level language like @command{awk}.
-The special files provided in @command{gawk} hide the details from
-the programmer, making things much simpler and easier to use.
-@c Who sez we can't toot our own horn occasionally?
-
-The special @value{FN} for network access is made up of several fields, all
-of them mandatory, none of them optional:
-
-@example
-/inet/@var{protocol}/@var{localport}/@var{hostname}/@var{remoteport}
-@end example
-
-The @file{/inet/} field is, of course, constant when accessing the network.
-The @var{localport} and @var{remoteport} fields do not have a meaning
-when used with @file{/inet/raw} because ``ports'' only apply to
-TCP and UDP. So, when using @file{/inet/raw}, the port fields always have
-to be @samp{0}.
-
-@menu
-* Special File Fields:: The fields in the special file name.
-* Comparing Protocols:: Differences between the protocols.
-@end menu
-
-@node Special File Fields, Comparing Protocols, Gawk Special Files, Gawk Special Files
-@subsection The Fields of the Special @value{FFN}
-This @value{SECTION} explains the meaning of all the other fields,
-as well as the range of values and the defaults.
-All of the fields are mandatory. To let the system pick a value,
-or if the field doesn't apply to the protocol, specify it as @samp{0}.
-
-@table @var
-@item protocol
-Determines which member of the TCP/IP
-family of protocols is selected to transport the data across the
-network. There are three possible values (always written in lowercase):
-@samp{tcp}, @samp{udp}, and @samp{raw}. The exact meaning of each is
-explained later in this @value{SECTION}.
-
-@item localport
-Determines which port on the local
-machine is used to communicate across the network. It has no meaning
-with @file{/inet/raw} and must therefore be @samp{0}. Application level clients
-usually use @samp{0} to indicate they do not care which local port is
-used---instead they specify a remote port to connect to. It is vital for
-application level servers to use a number different from @samp{0} here
-because their service has to be available at a specific publicly-known
-port number. It is possible to use a name from @file{/etc/services} here.
-
-@item hostname
-Determines which remote host is to
-be at the other end of the connection. Application level servers must fill
-this field with a @samp{0} to indicate their being open for all other hosts
-to connect to them and enforce connection level server behavior this way.
-It is not possible for an application level server to restrict its
-availability to one remote host by entering a host name here.
-Application level clients must enter a name different from @samp{0}.
-The name can be either symbolic
-(e.g., @samp{jpl-devvax.jpl.nasa.gov}) or numeric (e.g., @samp{128.149.1.143}).
-
-@item remoteport
-Determines which port on the remote
-machine is used to communicate across the network. It has no meaning
-with @file{/inet/raw} and must therefore be 0.
-For @file{/inet/tcp} and @file{/inet/udp},
-application level clients @emph{must} use a number
-other than @samp{0} to indicate which port on the remote machine
-they want to connect to. Application level servers must not fill this field with
-a @samp{0}. Instead they specify a local port for clients to connect to.
-It is possible to use a name from @file{/etc/services} here.
-@end table
-
-Experts in network programming will notice that the usual
-client/server asymmetry found at the level of the socket API is not visible
-here. This is for the sake of simplicity of the high-level concept. If this
-asymmetry is necessary for your application,
-use another language.
-For @command{gawk}, it is
-more important to enable users to write a client program with a minimum
-of code. What happens when first accessing a network connection is seen
-in the following pseudo-code:
-
-@smallexample
-if ((name of remote host given) && (other side accepts connection)) @{
- rendez-vous successful; transmit with getline or print
-@} else @{
- if ((other side did not accept) && (localport == 0))
- exit unsuccessful
- if (TCP) @{
- set up a server accepting connections
- this means waiting for the client on the other side to connect
- @} else
- ready
-@}
-@end smallexample
-
-The exact behavior of this algorithm depends on the values of the
-fields of the special @value{FN}. When in doubt, the following table
-gives you the combinations of values and their meaning. If this
-table is too complicated, focus on the three lines printed in
-@strong{bold}. All the examples in
-@ref{Using Networking, ,Networking With @command{gawk}},
-use only the
-patterns printed in bold letters.
-
-@multitable {12345678901234} {123456} {123456} {1234567} {1234567890123456789012345}
-@item @sc{protocol} @tab @sc{local port} @tab @sc{host name}
-@tab @sc{remote port} @tab @sc{Resulting connection level behavior}
-@item @strong{tcp} @tab @strong{0} @tab @strong{x} @tab @strong{x} @tab
- @strong{Dedicated client, fails if immediately connecting to a
- server on the other side fails}
-@item udp @tab 0 @tab x @tab x @tab Dedicated client
-@item raw @tab 0 @tab x @tab 0 @tab Dedicated client, works only as @code{root}
-@item @strong{tcp, udp} @tab @strong{x} @tab @strong{x} @tab @strong{x} @tab
- @strong{Client, switches to dedicated server if necessary}
-@item @strong{tcp, udp} @tab @strong{x} @tab @strong{0} @tab @strong{0} @tab
- @strong{Dedicated server}
-@item raw @tab 0 @tab 0 @tab 0 @tab Dedicated server, works only as @code{root}
-@item tcp, udp, raw @tab x @tab x @tab 0 @tab Invalid
-@item tcp, udp, raw @tab 0 @tab 0 @tab x @tab Invalid
-@item tcp, udp, raw @tab x @tab 0 @tab x @tab Invalid
-@item tcp, udp @tab 0 @tab 0 @tab 0 @tab Invalid
-@item tcp, udp @tab 0 @tab x @tab 0 @tab Invalid
-@item raw @tab x @tab 0 @tab 0 @tab Invalid
-@item raw @tab 0 @tab x @tab x @tab Invalid
-@item raw @tab x @tab x @tab x @tab Invalid
-@end multitable
-
-In general, TCP is the preferred mechanism to use. It is the simplest
-protocol to understand and to use. Use the others only if circumstances
-demand low-overhead.
-
-@node Comparing Protocols, , Special File Fields, Gawk Special Files
-@subsection Comparing Protocols
-
-This @value{SECTION} develops a pair of programs (sender and receiver)
-that do nothing but send a timestamp from one machine to another. The
-sender and the receiver are implemented with each of the three protocols
-available and demonstrate the differences between them.
-
-@menu
-* File /inet/tcp:: The TCP special file.
-* File /inet/udp:: The UDB special file.
-* File /inet/raw:: The RAW special file.
-@end menu
-
-@node File /inet/tcp, File /inet/udp, Comparing Protocols, Comparing Protocols
-@subsubsection @file{/inet/tcp}
-@cindex @file{/inet/tcp} special files
-@cindex TCP
-Once again, always use TCP.
-(Use UDP when low-overhead is a necessity, and use RAW for
-network experimentation.)
-The first example is the sender
-program:
-
-@example
-# Server
-BEGIN @{
- print strftime() |& "/inet/tcp/8888/0/0"
- close("/inet/tcp/8888/0/0")
-@}
-@end example
-
-The receiver is very simple:
-
-@example
-# Client
-BEGIN @{
- "/inet/tcp/0/localhost/8888" |& getline
- print $0
- close("/inet/tcp/0/localhost/8888")
-@}
-@end example
-
-TCP guarantees that the bytes arrive at the receiving end in exactly
-the same order that they were sent. No byte is lost
-(except for broken connections), doubled, or out of order. Some
-overhead is necessary to accomplish this, but this is the price to pay for
-a reliable service.
-It does matter which side starts first. The sender/server has to be started
-first, and it waits for the receiver to read a line.
-
-@node File /inet/udp, File /inet/raw, File /inet/tcp, Comparing Protocols
-@subsubsection @file{/inet/udp}
-@cindex @file{/inet/udp} special files
-@cindex UDP
-The server and client programs that use UDP are almost identical to their TCP counterparts;
-only the @var{protocol} has changed. As before, it does matter which side
-starts first. The receiving side blocks and waits for the sender.
-In this case, the receiver/client has to be started first:
-
-@page
-@example
-# Server
-BEGIN @{
- print strftime() |& "/inet/udp/8888/0/0"
- close("/inet/udp/8888/0/0")
-@}
-@end example
-
-The receiver is almost identical to the TCP receiver:
-
-@example
-# Client
-BEGIN @{
- "/inet/udp/0/localhost/8888" |& getline
- print $0
- close("/inet/udp/0/localhost/8888")
-@}
-@end example
-
-UDP cannot guarantee that the datagrams at the receiving end will arrive in exactly
-the same order they were sent. Some datagrams could be
-lost, some doubled, and some out of order. But no overhead is necessary to
-accomplish this. This unreliable behavior is good enough for tasks
-such as data acquisition, logging, and even stateless services like NFS.
-
-@node File /inet/raw, , File /inet/udp, Comparing Protocols
-@subsubsection @file{/inet/raw}
-@cindex @file{/inet/raw} special files
-@cindex RAW
-
-This is an IP-level protocol. Only @code{root} is allowed to access this
-special file. It is meant to be the basis for implementing
-and experimenting with transport level protocols.@footnote{This special file
-is reserved, but not otherwise currently implemented.}
-In the most general case,
-the sender has to supply the encapsulating header bytes in front of the
-packet and the receiver has to strip the additional bytes from the message.
-
-@cindex dark corner
-RAW receivers cannot receive packets sent with TCP or UDP because the
-operating system does not deliver the packets to a RAW receiver. The
-operating system knows about some of the protocols on top of IP
-and decides on its own which packet to deliver to which process.
-@value{DARKCORNER}
-Therefore, the UDP receiver must be used for receiving UDP
-datagrams sent with the RAW sender. This is a dark corner, not only of
-@command{gawk}, but also of TCP/IP.
-
-@cindex SPAK utility
-For extended experimentation with protocols, look into
-the approach implemented in a tool called SPAK.
-This tool reflects the hierarchical layering of protocols (encapsulation)
-in the way data streams are piped out of one program into the next one.
-It shows which protocol is based on which other (lower-level) protocol
-by looking at the command-line ordering of the program calls.
-Cleverly thought out, SPAK is much better than @command{gawk}'s
-@file{/inet} for learning the meaning of each and every bit in the
-protocol headers.
-
-The next example uses the RAW protocol to emulate
-the behavior of UDP. The sender program is the same as above, but with some
-additional bytes that fill the places of the UDP fields:
-
-@example
-@group
-BEGIN @{
- Message = "Hello world\n"
- SourcePort = 0
- DestinationPort = 8888
- MessageLength = length(Message)+8
- RawService = "/inet/raw/0/localhost/0"
- printf("%c%c%c%c%c%c%c%c%s",
- SourcePort/256, SourcePort%256,
- DestinationPort/256, DestinationPort%256,
- MessageLength/256, MessageLength%256,
- 0, 0, Message) |& RawService
- fflush(RawService)
- close(RawService)
-@}
-@end group
-@end example
-
-Since this program tries
-to emulate the behavior of UDP, it checks if
-the RAW sender is understood by the UDP receiver but not if the RAW receiver
-can understand the UDP sender. In a real network, the
-RAW receiver is hardly
-of any use because it gets every IP packet that
-comes across the network. There are usually so many packets that
-@command{gawk} would be too slow for processing them.
-Only on a network with little
-traffic can the IP-level receiver program be tested. Programs for analyzing
-IP traffic on modem or ISDN channels should be possible.
-
-Port numbers do not have a meaning when using @file{/inet/raw}. Their fields
-have to be @samp{0}. Only TCP and UDP use ports. Receiving data from
-@file{/inet/raw} is difficult, not only because of processing speed but also
-because data is usually binary and not restricted to ASCII. This
-implies that line separation with @code{RS} does not work as usual.
-
-@node TCP Connecting, Troubleshooting, Gawk Special Files, Using Networking
-@section Establishing a TCP Connection
-
-Let's observe a network connection at work. Type in the following program
-and watch the output. Within a second, it connects via TCP (@file{/inet/tcp})
-to the machine it is running on (@samp{localhost}), and asks the service
-@samp{daytime} on the machine what time it is:
-
-@cindex @code{|&} I/O operator
-@cindex @code{getline} built-in function
-@example
-BEGIN @{
- "/inet/tcp/0/localhost/daytime" |& getline
- print $0
- close("/inet/tcp/0/localhost/daytime")
-@}
-@end example
-
-Even experienced @command{awk} users will find the second line strange in two
-respects:
-
-@itemize @bullet
-@item
-A special file is used as a shell command that pipes its output
-into @code{getline}. One would rather expect to see the special file
-being read like any other file (@samp{getline <
-"/inet/tcp/0/localhost/daytime")}.
-
-@item
-The operator @samp{|&} has not been part of any @command{awk}
-implementation (until now).
-It is actually the only extension of the @command{awk}
-language needed (apart from the special files) to introduce network access.
-@end itemize
-
-The @samp{|&} operator was introduced in @command{gawk} 3.1 in order to
-overcome the crucial restriction that access to files and pipes in
-@command{awk} is always unidirectional. It was formerly impossible to use
-both access modes on the same file or pipe. Instead of changing the whole
-concept of file access, the @samp{|&} operator
-behaves exactly like the usual pipe operator except for two additions:
-
-@itemize @bullet
-@item
-Normal shell commands connected to their @command{gawk} program with a @samp{|&}
-pipe can be accessed bidirectionally. The @samp{|&} turns out to be a quite
-general, useful, and natural extension of @command{awk}.
-
-@item
-Pipes that consist of a special @value{FN} for network connections are not
-executed as shell commands. Instead, they can be read and written to, just
-like a full-duplex network connection.
-@end itemize
-
-In the earlier example, the @samp{|&} operator tells @code{getline}
-to read a line from the special file @file{/inet/tcp/0/localhost/daytime}.
-We could also have printed a line into the special file. But instead we just
-read a line with the time, printed it, and closed the connection.
-(While we could just let @command{gawk} close the connection by finishing
-the program, in this @value{DOCUMENT}
-we are pedantic, and always explicitly close the connections.)
-
-@node Troubleshooting, Interacting, TCP Connecting, Using Networking
-@section Troubleshooting Connection Problems
-It may well be that for some reason the above program does not run on your
-machine. When looking at possible reasons for this, you will learn much
-about typical problems that arise in network programming. First of all,
-your implementation of @command{gawk} may not support network access
-because it is
-a pre-3.1 version or you do not have a network interface in your machine.
-Perhaps your machine uses some other protocol
-like DECnet or Novell's IPX. For the rest of this @value{CHAPTER},
-we will assume
-you work on a Unix machine that supports TCP/IP. If the above program does
-not run on such a machine, it may help to replace the name
-@samp{localhost} with the name of your machine or its IP address. If it
-does, you could replace @samp{localhost} with the name of another machine
-in your vicinity. This way, the program connects to another machine.
-Now you should see the date and time being printed by the program.
-Otherwise your machine may not support the @samp{daytime} service.
-Try changing the service to @samp{chargen} or @samp{ftp}. This way, the program
-connects to other services that should give you some response. If you are
-curious, you should have a look at your file @file{/etc/services}. It could
-look like this:
-
-@ignore
-@multitable {1234567890123} {1234567890123} {123456789012345678901234567890123456789012}
-@item Service @strong{name} @tab Service @strong{number}
-@item echo @tab 7/tcp @tab echo sends back each line it receivces
-@item echo @tab 7/udp @tab echo is good for testing purposes
-@item discard @tab 9/tcp @tab discard behaves like @file{/dev/null}
-@item discard @tab 9/udp @tab discard just throws away each line
-@item daytime @tab 13/tcp @tab daytime sends date & time once per connection
-@item daytime @tab 13/udp
-@item chargen @tab 19/tcp @tab chargen infinitely produces character sets
-@item chargen @tab 19/udp @tab chargen is good for testing purposes
-@item ftp @tab 21/tcp @tab ftp is the usual file transfer protocol
-@item telnet @tab 23/tcp @tab telnet is the usual login facility
-@item smtp @tab 25/tcp @tab smtp is the Simple Mail Transfer Protocol
-@item finger @tab 79/tcp @tab finger tells you who is logged in
-@item www @tab 80/tcp @tab www is the HyperText Transfer Protocol
-@item pop2 @tab 109/tcp @tab pop2 is an older version of pop3
-@item pop2 @tab 109/udp
-@item pop3 @tab 110/tcp @tab pop3 is the Post Office Protocol
-@item pop3 @tab 110/udp @tab pop3 is used for receiving email
-@item nntp @tab 119/tcp @tab nntp is the USENET News Transfer Protocol
-@item irc @tab 194/tcp @tab irc is the Internet Relay Chat
-@item irc @tab 194/udp
-@end multitable
-@end ignore
-
-@smallexample
-# /etc/services:
-#
-# Network services, Internet style
-#
-# Name Number/Protcol Alternate name # Comments
-
-echo 7/tcp
-echo 7/udp
-discard 9/tcp sink null
-discard 9/udp sink null
-daytime 13/tcp
-daytime 13/udp
-chargen 19/tcp ttytst source
-chargen 19/udp ttytst source
-ftp 21/tcp
-telnet 23/tcp
-smtp 25/tcp mail
-finger 79/tcp
-www 80/tcp http # WorldWideWeb HTTP
-www 80/udp # HyperText Transfer Protocol
-pop-2 109/tcp postoffice # POP version 2
-pop-2 109/udp
-pop-3 110/tcp # POP version 3
-pop-3 110/udp
-nntp 119/tcp readnews untp # USENET News
-irc 194/tcp # Internet Relay Chat
-irc 194/udp
-@dots{}
-@end smallexample
-
-@cindex Linux
-@cindex GNU/Linux
-@cindex Microsoft Windows
-Here, you find a list of services that traditional Unix machines usually
-support. If your GNU/Linux machine does not do so, it may be that these
-services are switched off in some startup script. Systems running some
-flavor of Microsoft Windows usually do @emph{not} support such services.
-Nevertheless, it @emph{is} possible to do networking with @command{gawk} on
-Microsoft
-Windows.@footnote{Microsoft prefered to ignore the TCP/IP
-family of protocols until 1995. Then came the rise of the Netscape browser
-as a landmark ``killer application.'' Microsoft added TCP/IP support and
-their own browser to Microsoft Windows 95 at the last minute. They even back-ported
-their TCP/IP implementation to Microsoft Windows for Workgroups 3.11, but it was
-a rather rudimentary and half-hearted implementation. Nevertheless,
-the equivalent of @file{/etc/services} resides under
-@file{c:\windows\services} on Microsoft Windows.}
-The first column of the file gives the name of the service,
-the second a unique number, and the protocol that one can use to connect to
-this service.
-The rest of the line is treated as a comment.
-You see that some services (@samp{echo}) support TCP as
-well as UDP.
-
-@node Interacting, Setting Up, Troubleshooting, Using Networking
-@section Interacting with a Network Service
-
-The next program makes use of the possibility to really interact with a
-network service by printing something into the special file. It asks the
-so-called @command{finger} service if a user of the machine is logged in. When
-testing this program, try to change @samp{localhost} to
-some other machine name in your local network:
-
-@c system if test ! -d eg ; then mkdir eg ; fi
-@c system if test ! -d eg/network ; then mkdir eg/network ; fi
-@example
-@c file eg/network/fingerclient.awk
-BEGIN @{
- NetService = "/inet/tcp/0/localhost/finger"
- print "@var{name}" |& NetService
- while ((NetService |& getline) > 0)
- print $0
- close(NetService)
-@}
-@c endfile
-@end example
-
-After telling the service on the machine which user to look for,
-the program repeatedly reads lines that come as a reply. When no more
-lines are coming (because the service has closed the connection), the
-program also closes the connection. Try replacing @code{"@var{name}"} with your
-login name (or the name of someone else logged in). For a list
-of all users currently logged in, replace @var{name} with an empty string
-@code{""}.
-
-@cindex Linux
-@cindex GNU/Linux
-The final @code{close} command could be safely deleted from
-the above script, because the operating system closes any open connection
-by default when a script reaches the end of execution. In order to avoid
-portability problems, it is best to always close connections explicitly.
-With the Linux kernel,
-for example, proper closing results in flushing of buffers. Letting
-the close happen by default may result in discarding buffers.
-
-@ignore
-@c Chuck comments that this seems out of place. He's right. I dunno
-@c where to put it though.
-@cindex @command{finger} utility
-@cindex RFC 1288
-In the early days of the Internet (up until about 1992), you could use
-such a program to check if some user in another country was logged in on
-a specific machine.
-RFC 1288@footnote{@uref{http://www.cis.ohio-state.edu/htbin/rfc/rfc1288.html}}
-provides the exact definition of the @command{finger} protocol.
-Every contemporary Unix system also has a command named @command{finger},
-which functions as a client for the protocol of the same name.
-Still today, some people maintain simple information systems
-with this ancient protocol. For example, by typing
-@samp{finger quake@@seismo.unr.edu}
-you get the latest @dfn{Earthquake Bulletin} for the state of Nevada.
-
-@cindex Earthquake Bulletin
-@smallexample
-$ finger quake@@seismo.unr.edu
-
-[@dots{}]
-
-DATE-(UTC)-TIME LAT LON DEP MAG COMMENTS
-yy/mm/dd hh:mm:ss deg. deg. km
-
-98/12/14 21:09:22 37.47N 116.30W 0.0 2.3Md 76.4 km S of WARM SPRINGS, NEVA
-98/12/14 22:05:09 39.69N 120.41W 11.9 2.1Md 53.8 km WNW of RENO, NEVADA
-98/12/15 14:14:19 38.04N 118.60W 2.0 2.3Md 51.0 km S of HAWTHORNE, NEVADA
-98/12/17 01:49:02 36.06N 117.58W 13.9 3.0Md 74.9 km SE of LONE PINE, CALIFOR
-98/12/17 05:39:26 39.95N 120.87W 6.2 2.6Md 101.6 km WNW of RENO, NEVADA
-98/12/22 06:07:42 38.68N 119.82W 5.2 2.3Md 50.7 km S of CARSON CITY, NEVAD
-@end smallexample
-
-@noindent
-This output from @command{finger} contains the time, location, depth,
-magnitude, and a short comment about
-the earthquakes registered in that region during the last 10 days.
-In many places today the use of such services is restricted
-because most networks have firewalls and proxy servers between them
-and the Internet. Most firewalls are programmed to not let
-@command{finger} requests go beyond the local network.
-
-@cindex Coke machine
-Another (ab)use of the @command{finger} protocol are several Coke machines
-that are connected to the Internet. There is a short list of such
-Coke machines.@footnote{@uref{http://ca.yahoo.com/Computers_and_Internet/Internet/Devices_Connected_to_the_Internet/Soda_Machines/}}
-You can access them either from the command-line or with a simple
-@command{gawk} script. They usually tell you about the different
-flavors of Coke and beer available there. If you have an account there,
-you can even order some drink this way.
-@end ignore
-
-When looking at @file{/etc/services} you may have noticed that the
-@samp{daytime} service is also available with @samp{udp}. In the earlier
-example, change @samp{tcp} to @samp{udp},
-and change @samp{finger} to @samp{daytime}.
-After starting the modified program, you see the expected day and time message.
-The program then hangs, because it waits for more lines coming from the
-service. However, they never come. This behavior is a consequence of the
-differences between TCP and UDP. When using UDP, neither party is
-automatically informed about the other closing the connection.
-Continuing to experiment this way reveals many other subtle
-differences between TCP and UDP. To avoid such trouble, one should always
-remember the advice Douglas E.@: Comer and David Stevens give in
-Volume III of their series @cite{Internetworking With TCP}
-(page 14):
-
-@cindex TCP
-@cindex UDP
-@quotation
-When designing client-server applications, beginners are strongly
-advised to use TCP because it provides reliable, connection-oriented
-communication. Programs only use UDP if the application protocol handles
-reliability, the application requires hardware broadcast or multicast,
-or the application cannot tolerate virtual circuit overhead.
-@end quotation
-
-@node Setting Up, Email, Interacting, Using Networking
-@section Setting Up a Service
-The preceding programs behaved as clients that connect to a server somewhere
-on the Internet and request a particular service. Now we set up such a
-service to mimic the behavior of the @samp{daytime} service.
-Such a server does not know in advance who is going to connect to it over
-the network. Therefore we cannot insert a name for the host to connect to
-in our special @value{FN}.
-
-Start the following program in one window. Notice that the service does
-not have the name @samp{daytime}, but the number @samp{8888}.
-From looking at @file{/etc/services}, you know that names like @samp{daytime}
-are just mnemonics for predetermined 16-bit integers.
-Only the system administrator (@code{root}) could enter
-our new service into @file{/etc/services} with an appropriate name.
-Also notice that the service name has to be entered into a different field
-of the special @value{FN} because we are setting up a server, not a client:
-
-@cindex @command{finger} utility
-@cindex server
-@example
-BEGIN @{
- print strftime() |& "/inet/tcp/8888/0/0"
- close("/inet/tcp/8888/0/0")
-@}
-@end example
-
-Now open another window on the same machine.
-Copy the client program given as the first example
-(@pxref{TCP Connecting, ,Establishing a TCP Connection})
-to a new file and edit it, changing the name @samp{daytime} to
-@samp{8888}. Then start the modified client. You should get a reply
-like this:
-
-@example
-Sat Sep 27 19:08:16 CEST 1997
-@end example
-
-@noindent
-Both programs explicitly close the connection.
-
-@cindex Microsoft Windows
-@cindex reserved ports
-Now we will intentionally make a mistake to see what happens when the name
-@samp{8888} (the so-called port) is already used by another service.
-Start the server
-program in both windows. The first one works, but the second one
-complains that it could not open the connection. Each port on a single
-machine can only be used by one server program at a time. Now terminate the
-server program and change the name @samp{8888} to @samp{echo}. After restarting it,
-the server program does not run any more and you know why: there already is
-an @samp{echo} service running on your machine. But even if this isn't true,
-you would not get
-your own @samp{echo} server running on a Unix machine,
-because the ports with numbers smaller
-than 1024 (@samp{echo} is at port 7) are reserved for @code{root}.
-On machines running some flavor of Microsoft Windows, there is no restriction
-that reserves ports 1 to 1024 for a privileged user; hence you can start
-an @samp{echo} server there.
-
-Turning this short server program into something really useful is simple.
-Imagine a server that first reads a @value{FN} from the client through the
-network connection, then does something with the file and
-sends a result back to the client. The server-side processing
-could be:
-
-@example
-BEGIN @{
- NetService = "/inet/tcp/8888/0/0"
- NetService |& getline
- CatPipe = ("cat " $1) # sets $0 and the fields
- while ((CatPipe | getline) > 0)
- print $0 |& NetService
- close(NetService)
-@}
-@end example
-
-@noindent
-and we would
-have a remote copying facility. Such a server reads the name of a file
-from any client that connects to it and transmits the contents of the
-named file across the net. The server-side processing could also be
-the execution of a command that is transmitted across the network. From this
-example, you can see how simple it is to open up a security hole on your
-machine. If you allow clients to connect to your machine and
-execute arbitrary commands, anyone would be free to do @samp{rm -rf *}.
-
-@node Email, Web page, Setting Up, Using Networking
-@section Reading Email
-@cindex POP
-@cindex SMTP
-@cindex RFC 1939
-@cindex RFC 821
-The distribution of email is usually done by dedicated email servers that
-communicate with your machine using special protocols. To receive email, we
-will use the Post Office Protocol (POP). Sending can be done with the much
-older Simple Mail Transfer Protocol (SMTP).
-@ignore
-@footnote{RFC 1939 defines POP.
-RFC 821 defines SMTP. See
-@uref{http://rfc.fh-koeln.de/doc/rfc/html/rfc.html, RFCs in HTML}.}
-@end ignore
-
-When you type in the following program, replace the @var{emailhost} by the
-name of your local email server. Ask your administrator if the server has a
-POP service, and then use its name or number in the program below.
-Now the program is ready to connect to your email server, but it will not
-succeed in retrieving your mail because it does not yet know your login
-name or password. Replace them in the program and it
-shows you the first email the server has in store:
-
-@example
-BEGIN @{
- POPService = "/inet/tcp/0/@var{emailhost}/pop3"
- RS = ORS = "\r\n"
- print "user @var{name}" |& POPService
- POPService |& getline
- print "pass @var{password}" |& POPService
- POPService |& getline
- print "retr 1" |& POPService
- POPService |& getline
- if ($1 != "+OK") exit
- print "quit" |& POPService
- RS = "\r\n\\.\r\n"
- POPService |& getline
- print $0
- close(POPService)
-@}
-@end example
-
-@cindex RFC 1939
-The record separators @code{RS} and @code{ORS} are redefined because the
-protocol (POP) requires CR-LF to separate lines. After identifying
-yourself to the email service, the command @samp{retr 1} instructs the
-service to send the first of all your email messages in line. If the service
-replies with something other than @samp{+OK}, the program exits; maybe there
-is no email. Otherwise, the program first announces that it intends to finish
-reading email, and then redefines @code{RS} in order to read the entire
-email as multiline input in one record. From the POP RFC, we know that the body
-of the email always ends with a single line containing a single dot.
-The program looks for this using @samp{RS = "\r\n\\.\r\n"}.
-When it finds this sequence in the mail message, it quits.
-You can invoke this program as often as you like; it does not delete the
-message it reads, but instead leaves it on the server.
-
-@node Web page, Primitive Service, Email, Using Networking
-@section Reading a Web Page
-@cindex HTTP
-@cindex RFC 2068
-@cindex RFC 2616
-
-Retrieving a web page from a web server is as simple as
-retrieving email from an email server. We only have to use a
-similar, but not identical, protocol and a different port. The name of the
-protocol is HyperText Transfer Protocol (HTTP) and the port number is usually
-80. As in the preceding @value{SECTION}, ask your administrator about the
-name of your local web server or proxy web server and its port number
-for HTTP requests.
-
-@ignore
-@c Chuck says this stuff isn't necessary
-More detailed information about HTTP can be found at
-the home of the web protocols,@footnote{@uref{http://www.w3.org/pub/WWW/Protocols}}
-including the specification of HTTP in RFC 2068. The protocol specification
-in RFC 2068 is concise and you can get it for free. If you need more
-explanation and you are willing to pay for a book, you might be
-interested in one of these books:
-
-@enumerate
-
-@item
-When we started writing web clients and servers with @command{gawk},
-the only book available with details about HTTP was the one by Paul Hethmon
-called
-@cite{Illustrated Guide to HTTP}.@footnote{@uref{http://www.browsebooks.com/Hethmon/?882}}
-Hethmon not only describes HTTP,
-he also implements a simple web server in C++.
-
-@item
-Since July 2000, O'Reilly offers the book by Clinton Wong called
-@cite{HTTP Pocket Reference}.@footnote{@uref{http://www.oreilly.com/catalog/httppr}}
-It only has 75 pages but its
-focus definitely is HTTP. This pocket reference is not a replacement
-for the RFC, but I wish I had had it back in 1997 when I started writing
-scripts to handle HTTP.
-
-@item
-Another small booklet about HTTP is the one by Toexcell Incorporated Staff,
-ISBN 1-58348-270-9, called
-@cite{Hypertext Transfer Protocol Http 1.0 Specifications}
-
-@end enumerate
-@end ignore
-
-The following program employs a rather crude approach toward retrieving a
-web page. It uses the prehistoric syntax of HTTP 0.9, which almost all
-web servers still support. The most noticeable thing about it is that the
-program directs the request to the local proxy server whose name you insert
-in the special @value{FN} (which in turn calls @samp{www.yahoo.com}):
-
-@example
-BEGIN @{
- RS = ORS = "\r\n"
- HttpService = "/inet/tcp/0/@var{proxy}/80"
- print "GET http://www.yahoo.com" |& HttpService
- while ((HttpService |& getline) > 0)
- print $0
- close(HttpService)
-@}
-@end example
-
-@cindex RFC 1945
-@cindex HTML
-@cindex Yahoo!
-Again, lines are separated by a redefined @code{RS} and @code{ORS}.
-The @code{GET} request that we send to the server is the only kind of
-HTTP request that existed when the web was created in the early 1990s.
-HTTP calls this @code{GET} request a ``method,'' which tells the
-service to transmit a web page (here the home page of the Yahoo! search
-engine). Version 1.0 added the request methods @code{HEAD} and
-@code{POST}. The current version of HTTP is 1.1,@footnote{Version 1.0 of
-HTTP was defined in RFC 1945. HTTP 1.1 was initially specified in RFC
-2068. In June 1999, RFC 2068 was made obsolete by RFC 2616. It is an update
-without any substantial changes.} and knows the additional request
-methods @code{OPTIONS}, @code{PUT}, @code{DELETE}, and @code{TRACE}.
-You can fill in any valid web address, and the program prints the
-HTML code of that page to your screen.
-
-Notice the similarity between the responses of the POP and HTTP
-services. First, you get a header that is terminated by an empty line, and
-then you get the body of the page in HTML. The lines of the headers also
-have the same form as in POP. There is the name of a parameter,
-then a colon, and finally the value of that parameter.
-
-@cindex CGI
-@cindex @file{gif} image format
-@cindex @file{png} image format
-Images (@file{.png} or @file{.gif} files) can also be retrieved this way,
-but then you
-get binary data that should be redirected into a file. Another
-application is calling a CGI (Common Gateway Interface) script on some
-server. CGI scripts are used when the contents of a web page are not
-constant, but generated instantly at the moment you send a request
-for the page. For example, to get a detailed report about the current
-quotes of Motorola stock shares, call a CGI script at Yahoo! with
-the following:
-
-@example
-get = "GET http://quote.yahoo.com/q?s=MOT&d=t"
-print get |& HttpService
-@end example
-
-You can also request weather reports this way.
-@ignore
-@cindex Boutell, Thomas
-A good book to go on with is
-the
-@cite{HTML Source Book}.@footnote{@uref{http://www.utoronto.ca/webdocs/HTMLdocs/NewHTML/book.html}}
-There are also some books on CGI programming
-like @cite{CGI Programming in C & Perl},
-by Thomas Boutell@footnote{@uref{http://cseng.aw.com/bookdetail.qry?ISBN=0-201-42219-0&ptype=0}},
-and @cite{The CGI Book}.@footnote{@uref{http://www.cgibook.com}}
-Another good source is @cite{The CGI Resource Index}}.@footnote{@uref{http://www.cgi-resources.com}}
-@end ignore
-
-@node Primitive Service, Interacting Service, Web page, Using Networking
-@section A Primitive Web Service
-Now we know enough about HTTP to set up a primitive web service that just
-says @code{"Hello, world"} when someone connects to it with a browser.
-Compared
-to the situation in the preceding @value{SECTION}, our program changes the role. It
-tries to behave just like the server we have observed. Since we are setting
-up a server here, we have to insert the port number in the @samp{localport}
-field of the special @value{FN}. The other two fields (@var{hostname} and
-@var{remoteport}) have to contain a @samp{0} because we do not know in
-advance which host will connect to our service.
-
-In the early 1990s, all a server had to do was send an HTML document and
-close the connection. Here, we adhere to the modern syntax of HTTP.
-The steps are as follows:
-
-@enumerate 1
-@item
-Send a status line telling the web browser that everything
-is OK.
-
-@item
-Send a line to tell the browser how many bytes follow in the
-body of the message. This was not necessary earlier because both
-parties knew that the document ended when the connection closed. Nowadays
-it is possible to stay connected after the transmission of one web page.
-This is to avoid the network traffic necessary for repeatedly establishing
-TCP connections for requesting several images. Thus, there is the need to tell
-the receiving party how many bytes will be sent. The header is terminated
-as usual with an empty line.
-
-@item
-Send the @code{"Hello, world"} body
-in HTML.
-The useless @code{while} loop swallows the request of the browser.
-We could actually omit the loop, and on most machines the program would still
-work.
-First, start the following program:
-@end enumerate
-
-@example
-@c file eg/network/hello-serv.awk
-BEGIN @{
- RS = ORS = "\r\n"
- HttpService = "/inet/tcp/8080/0/0"
- Hello = "<HTML><HEAD>" \
- "<TITLE>A Famous Greeting</TITLE></HEAD>" \
- "<BODY><H1>Hello, world</H1></BODY></HTML>"
- Len = length(Hello) + length(ORS)
- print "HTTP/1.0 200 OK" |& HttpService
- print "Content-Length: " Len ORS |& HttpService
- print Hello |& HttpService
- while ((HttpService |& getline) > 0)
- continue;
- close(HttpService)
-@}
-@c endfile
-@end example
-
-Now, on the same machine, start your favorite browser and let it point to
-@uref{http://localhost:8080} (the browser needs to know on which port
-our server is listening for requests). If this does not work, the browser
-probably tries to connect to a proxy server that does not know your machine.
-If so, change the browser's configuration so that the browser does not try to
-use a proxy to connect to your machine.
-
-@node Interacting Service, Simple Server, Primitive Service, Using Networking
-@section A Web Service with Interaction
-@cindex GUI
-@ifinfo
-This node shows how to set up a simple web server.
-The subnode is a library file that we will use with all the examples in
-@ref{Some Applications and Techniques}.
-@end ifinfo
-
-@menu
-* CGI Lib:: A simple CGI library.
-@end menu
-
-Setting up a web service that allows user interaction is more difficult and
-shows us the limits of network access in @command{gawk}. In this @value{SECTION},
-we develop a main program (a @code{BEGIN} pattern and its action)
-that will become the core of event-driven execution controlled by a
-graphical user interface (GUI).
-Each HTTP event that the user triggers by some action within the browser
-is received in this central procedure. Parameters and menu choices are
-extracted from this request and an appropriate measure is taken according to
-the user's choice.
-For example:
-
-@cindex HTTP server, core logic
-@example
-BEGIN @{
- if (MyHost == "") @{
- "uname -n" | getline MyHost
- close("uname -n")
- @}
- if (MyPort == 0) MyPort = 8080
- HttpService = "/inet/tcp/" MyPort "/0/0"
- MyPrefix = "http://" MyHost ":" MyPort
- SetUpServer()
- while ("awk" != "complex") @{
- # header lines are terminated this way
- RS = ORS = "\r\n"
- Status = 200 # this means OK
- Reason = "OK"
- Header = TopHeader
- Document = TopDoc
- Footer = TopFooter
- if (GETARG["Method"] == "GET") @{
- HandleGET()
- @} else if (GETARG["Method"] == "HEAD") @{
- # not yet implemented
- @} else if (GETARG["Method"] != "") @{
- print "bad method", GETARG["Method"]
- @}
- Prompt = Header Document Footer
- print "HTTP/1.0", Status, Reason |& HttpService
- print "Connection: Close" |& HttpService
- print "Pragma: no-cache" |& HttpService
- len = length(Prompt) + length(ORS)
- print "Content-length:", len |& HttpService
- print ORS Prompt |& HttpService
- # ignore all the header lines
- while ((HttpService |& getline) > 0)
- ;
- # stop talking to this client
- close(HttpService)
- # wait for new client request
- HttpService |& getline
- # do some logging
- print systime(), strftime(), $0
- # read request parameters
- CGI_setup($1, $2, $3)
- @}
-@}
-@end example
-
-This web server presents menu choices in the form of HTML links.
-Therefore, it has to tell the browser the name of the host it is
-residing on. When starting the server, the user may supply the name
-of the host from the command line with @samp{gawk -v MyHost="Rumpelstilzchen"}.
-If the user does not do this, the server looks up the name of the host it is
-running on for later use as a web address in HTML documents. The same
-applies to the port number. These values are inserted later into the
-HTML content of the web pages to refer to the home system.
-
-Each server that is built around this core has to initialize some
-application-dependent variables (such as the default home page) in a procedure
-@code{SetUpServer}, which is called immediately before entering the
-infinite loop of the server. For now, we will write an instance that
-initiates a trivial interaction. With this home page, the client user
-can click on two possible choices, and receive the current date either
-in human-readable format or in seconds since 1970:
-
-@example
-function SetUpServer() @{
- TopHeader = "<HTML><HEAD>"
- TopHeader = TopHeader \
- "<title>My name is GAWK, GNU AWK</title></HEAD>"
- TopDoc = "<BODY><h2>\
- Do you prefer your date <A HREF=" MyPrefix \
- "/human>human</A> or \
- <A HREF=" MyPrefix "/POSIX>POSIXed</A>?</h2>" ORS ORS
- TopFooter = "</BODY></HTML>"
-@}
-@end example
-
-On the first run through the main loop, the default line terminators are
-set and the default home page is copied to the actual home page. Since this
-is the first run, @code{GETARG["Method"]} is not initialized yet, hence the
-case selection over the method does nothing. Now that the home page is
-initialized, the server can start communicating to a client browser.
-
-@cindex RFC 2068
-@cindex CGI
-It does so by printing the HTTP header into the network connection
-(@samp{print @dots{} |& HttpService}). This command blocks execution of
-the server script until a client connects. If this server
-script is compared with the primitive one we wrote before, you will notice
-two additional lines in the header. The first instructs the browser
-to close the connection after each request. The second tells the
-browser that it should never try to @emph{remember} earlier requests
-that had identical web addresses (no caching). Otherwise, it could happen
-that the browser retrieves the time of day in the previous example just once,
-and later it takes the web page from the cache, always displaying the same
-time of day although time advances each second.
-
-Having supplied the initial home page to the browser with a valid document
-stored in the parameter @code{Prompt}, it closes the connection and waits
-for the next request. When the request comes, a log line is printed that
-allows us to see which request the server receives. The final step in the
-loop is to call the function @code{CGI_setup}, which reads all the lines
-of the request (coming from the browser), processes them, and stores the
-transmitted parameters in the array @code{PARAM}. The complete
-text of these application-independent functions can be found in
-@ref{CGI Lib, ,A Simple CGI Library}.
-For now, we use a simplified version of @code{CGI_setup}:
-
-@example
-function CGI_setup( method, uri, version, i) @{
- delete GETARG; delete MENU; delete PARAM
- GETARG["Method"] = $1
- GETARG["URI"] = $2
- GETARG["Version"] = $3
- i = index($2, "?")
- # is there a "?" indicating a CGI request?
-@group
- if (i > 0) @{
- split(substr($2, 1, i-1), MENU, "[/:]")
- split(substr($2, i+1), PARAM, "&")
- for (i in PARAM) @{
- j = index(PARAM[i], "=")
- GETARG[substr(PARAM[i], 1, j-1)] = \
- substr(PARAM[i], j+1)
- @}
- @} else @{ # there is no "?", no need for splitting PARAMs
- split($2, MENU, "[/:]")
- @}
-@end group
-@}
-@end example
-
-At first, the function clears all variables used for
-global storage of request parameters. The rest of the function serves
-the purpose of filling the global parameters with the extracted new values.
-To accomplish this, the name of the requested resource is split into
-parts and stored for later evaluation. If the request contains a @samp{?},
-then the request has CGI variables seamlessly appended to the web address.
-Everything in front of the @samp{?} is split up into menu items, and
-everything behind the @samp{?} is a list of @samp{@var{variable}=@var{value}} pairs
-(separated by @samp{&}) that also need splitting. This way, CGI variables are
-isolated and stored. This procedure lacks recognition of special characters
-that are transmitted in coded form@footnote{As defined in RFC 2068.}. Here, any
-optional request header and body parts are ignored. We do not need
-header parameters and the request body. However, when refining our approach or
-working with the @code{POST} and @code{PUT} methods, reading the header
-and body
-becomes inevitable. Header parameters should then be stored in a global
-array as well as the body.
-
-On each subsequent run through the main loop, one request from a browser is
-received, evaluated, and answered according to the user's choice. This can be
-done by letting the value of the HTTP method guide the main loop into
-execution of the procedure @code{HandleGET}, which evaluates the user's
-choice. In this case, we have only one hierarchical level of menus,
-but in the general case,
-menus are nested.
-The menu choices at each level are
-separated by @samp{/}, just as in @value{FN}s. Notice how simple it is to
-construct menus of arbitrary depth:
-
-@example
-function HandleGET() @{
- if ( MENU[2] == "human") @{
- Footer = strftime() TopFooter
- @} else if (MENU[2] == "POSIX") @{
- Footer = systime() TopFooter
- @}
-@}
-@end example
-
-@cindex CGI
-The disadvantage of this approach is that our server is slow and can
-handle only one request at a time. Its main advantage, however, is that
-the server
-consists of just one @command{gawk} program. No need for installing an
-@command{httpd}, and no need for static separate HTML files, CGI scripts, or
-@code{root} privileges. This is rapid prototyping.
-This program can be started on the same host that runs your browser.
-Then let your browser point to @uref{http://localhost:8080}.
-
-@cindex @file{xbm} image format
-@cindex image format
-@cindex GNUPlot utility
-It is also possible to include images into the HTML pages.
-Most browsers support the not very well-known
-@file{.xbm} format,
-which may contain only
-monochrome pictures but is an ASCII format. Binary images are possible but
-not so easy to handle. Another way of including images is to generate them
-with a tool such as GNUPlot,
-by calling the tool with the @code{system} function or through a pipe.
-
-@node CGI Lib, , Interacting Service, Interacting Service
-@subsection A Simple CGI Library
-@quotation
-@i{HTTP is like being married: you have to be able to handle whatever
-you're given, while being very careful what you send back.}@*
-Phil Smith III,@*
-@uref{http://www.netfunny.com/rhf/jokes/99/Mar/http.html}
-@end quotation
-
-In @ref{Interacting Service, ,A Web Service with Interaction},
-we saw the function @code{CGI_setup} as part of the web server
-``core logic'' framework. The code presented there handles almost
-everything necessary for CGI requests.
-One thing it doesn't do is handle encoded characters in the requests.
-For example, an @samp{&} is encoded as a percent sign followed by
-the hexadecimal value---@samp{%26}. These encoded values should be
-decoded.
-Following is a simple library to perform these tasks.
-This code is used for all web server examples
-used throughout the rest of this @value{DOCUMENT}.
-If you want to use it for your own web server, store the source code
-into a file named @file{inetlib.awk}. Then you can include
-these functions into your code by placing the following statement
-into your program:
-
-@example
-@@include inetlib.awk
-@end example
-
-@noindent
-on the first line of your script. But beware, this mechanism is
-only possible if you invoke your web server script with @command{igawk}
-instead of the usual @command{awk} or @command{gawk}.
-Here is the code:
-
-@example
-@c file eg/network/coreserv.awk
-# CGI Library and core of a web server
-@c endfile
-@ignore
-@c file eg/network/coreserv.awk
-#
-# Juergen Kahrs, Juergen.Kahrs@@vr-web.de
-# with Arnold Robbins, arnold@@gnu.org
-# September 2000
-
-@c endfile
-@end ignore
-@c file eg/network/coreserv.awk
-# Global arrays
-# GETARG --- arguments to CGI GET command
-# MENU --- menu items (path names)
-# PARAM --- parameters of form x=y
-
-# Optional variable MyHost contains host address
-# Optional variable MyPort contains port number
-# Needs TopHeader, TopDoc, TopFooter
-# Sets MyPrefix, HttpService, Status, Reason
-
-BEGIN @{
- if (MyHost == "") @{
- "uname -n" | getline MyHost
- close("uname -n")
- @}
- if (MyPort == 0) MyPort = 8080
- HttpService = "/inet/tcp/" MyPort "/0/0"
- MyPrefix = "http://" MyHost ":" MyPort
- SetUpServer()
- while ("awk" != "complex") @{
- # header lines are terminated this way
- RS = ORS = "\r\n"
- Status = 200 # this means OK
- Reason = "OK"
- Header = TopHeader
- Document = TopDoc
- Footer = TopFooter
- if (GETARG["Method"] == "GET") @{
- HandleGET()
- @} else if (GETARG["Method"] == "HEAD") @{
- # not yet implemented
- @} else if (GETARG["Method"] != "") @{
- print "bad method", GETARG["Method"]
- @}
- Prompt = Header Document Footer
- print "HTTP/1.0", Status, Reason |& HttpService
- print "Connection: Close" |& HttpService
- print "Pragma: no-cache" |& HttpService
- len = length(Prompt) + length(ORS)
- print "Content-length:", len |& HttpService
- print ORS Prompt |& HttpService
- # ignore all the header lines
- while ((HttpService |& getline) > 0)
- continue
- # stop talking to this client
- close(HttpService)
- # wait for new client request
- HttpService |& getline
- # do some logging
- print systime(), strftime(), $0
- CGI_setup($1, $2, $3)
- @}
-@}
-
-function CGI_setup( method, uri, version, i)
-@{
- delete GETARG
- delete MENU
- delete PARAM
- GETARG["Method"] = method
- GETARG["URI"] = uri
- GETARG["Version"] = version
-
- i = index(uri, "?")
- if (i > 0) @{ # is there a "?" indicating a CGI request?
- split(substr(uri, 1, i-1), MENU, "[/:]")
- split(substr(uri, i+1), PARAM, "&")
- for (i in PARAM) @{
- PARAM[i] = _CGI_decode(PARAM[i])
- j = index(PARAM[i], "=")
- GETARG[substr(PARAM[i], 1, j-1)] = \
- substr(PARAM[i], j+1)
- @}
- @} else @{ # there is no "?", no need for splitting PARAMs
- split(uri, MENU, "[/:]")
- @}
- for (i in MENU) # decode characters in path
- if (i > 4) # but not those in host name
- MENU[i] = _CGI_decode(MENU[i])
-@}
-@c endfile
-@end example
-
-This isolates details in a single function, @code{CGI_setup}.
-Decoding of encoded characters is pushed off to a helper function,
-@code{_CGI_decode}. The use of the leading underscore (@samp{_}) in
-the function name is intended to indicate that it is an ``internal''
-function, although there is nothing to enforce this:
-
-@example
-@c file eg/network/coreserv.awk
-function _CGI_decode(str, hexdigs, i, pre, code1, code2,
- val, result)
-@{
- hexdigs = "123456789abcdef"
-
- i = index(str, "%")
- if (i == 0) # no work to do
- return str
-
- do @{
- pre = substr(str, 1, i-1) # part before %xx
- code1 = substr(str, i+1, 1) # first hex digit
- code2 = substr(str, i+2, 1) # second hex digit
- str = substr(str, i+3) # rest of string
-
- code1 = tolower(code1)
- code2 = tolower(code2)
- val = index(hexdigs, code1) * 16 \
- + index(hexdigs, code2)
-
- result = result pre sprintf("%c", val)
- i = index(str, "%")
- @} while (i != 0)
- if (length(str) > 0)
- result = result str
- return result
-@}
-@c endfile
-@end example
-
-This works by splitting the string apart around an encoded character.
-The two digits are converted to lowercase and looked up in a string
-of hex digits. Note that @code{0} is not in the string on purpose;
-@code{index} returns zero when it's not found, automatically giving
-the correct value! Once the hexadecimal value is converted from
-characters in a string into a numerical value, @code{sprintf}
-converts the value back into a real character.
-The following is a simple test harness for the above functions:
-
-@example
-@c file eg/network/testserv.awk
-BEGIN @{
- CGI_setup("GET",
- "http://www.gnu.org/cgi-bin/foo?p1=stuff&p2=stuff%26junk" \
- "&percent=a %25 sign",
- "1.0")
- for (i in MENU)
- printf "MENU[\"%s\"] = %s\n", i, MENU[i]
- for (i in PARAM)
- printf "PARAM[\"%s\"] = %s\n", i, PARAM[i]
- for (i in GETARG)
- printf "GETARG[\"%s\"] = %s\n", i, GETARG[i]
-@}
-@c endfile
-@end example
-
-And this is the result when we run it:
-
-@c artificial line wrap in last output line
-@example
-$ gawk -f testserv.awk
-@print{} MENU["4"] = www.gnu.org
-@print{} MENU["5"] = cgi-bin
-@print{} MENU["6"] = foo
-@print{} MENU["1"] = http
-@print{} MENU["2"] =
-@print{} MENU["3"] =
-@print{} PARAM["1"] = p1=stuff
-@print{} PARAM["2"] = p2=stuff&junk
-@print{} PARAM["3"] = percent=a % sign
-@print{} GETARG["p1"] = stuff
-@print{} GETARG["percent"] = a % sign
-@print{} GETARG["p2"] = stuff&junk
-@print{} GETARG["Method"] = GET
-@print{} GETARG["Version"] = 1.0
-@print{} GETARG["URI"] = http://www.gnu.org/cgi-bin/foo?p1=stuff&
-p2=stuff%26junk&percent=a %25 sign
-@end example
-
-@node Simple Server, Caveats, Interacting Service, Using Networking
-@section A Simple Web Server
-@cindex GUI
-In the preceding @value{SECTION}, we built the core logic for event driven GUIs.
-In this @value{SECTION}, we finally extend the core to a real application.
-No one would actually write a commercial web server in @command{gawk}, but
-it is instructive to see that it is feasible in principle.
-
-@iftex
-@image{uf002331,4in}
-@end iftex
-
-@cindex ELIZA program
-@cindex Weizenbaum, Joseph
-The application is ELIZA, the famous program by Joseph Weizenbaum that
-mimics the behavior of a professional psychotherapist when talking to you.
-Weizenbaum would certainly object to this description, but this is part of
-the legend around ELIZA.
-Take the site-independent core logic and append the following code:
-
-@example
-@c file eg/network/eliza.awk
-function SetUpServer() @{
- SetUpEliza()
- TopHeader = \
- "<HTML><title>An HTTP-based System with GAWK</title>\
- <HEAD><META HTTP-EQUIV=\"Content-Type\"\
- CONTENT=\"text/html; charset=iso-8859-1\"></HEAD>\
- <BODY BGCOLOR=\"#ffffff\" TEXT=\"#000000\"\
- LINK=\"#0000ff\" VLINK=\"#0000ff\"\
- ALINK=\"#0000ff\"> <A NAME=\"top\">"
- TopDoc = "\
- <h2>Please choose one of the following actions:</h2>\
- <UL>\
- <LI>\
- <A HREF=" MyPrefix "/AboutServer>About this server</A>\
- </LI><LI>\
- <A HREF=" MyPrefix "/AboutELIZA>About Eliza</A></LI>\
- <LI>\
- <A HREF=" MyPrefix \
- "/StartELIZA>Start talking to Eliza</A></LI></UL>"
- TopFooter = "</BODY></HTML>"
-@}
-@c endfile
-@end example
-
-@code{SetUpServer} is similar to the previous example,
-except for calling another function, @code{SetUpEliza}.
-This approach can be used to implement other kinds of servers.
-The only changes needed to do so are hidden in the functions
-@code{SetUpServer} and @code{HandleGET}. Perhaps it might be necessary to
-implement other HTTP methods.
-The @command{igawk} program that comes with @command{gawk}
-may be useful for this process.
-
-When extending this example to a complete application, the first
-thing to do is to implement the function @code{SetUpServer} to
-initialize the HTML pages and some variables. These initializations
-determine the way your HTML pages look (colors, titles, menu
-items, etc.).
-
-@cindex GUI
-The function @code{HandleGET} is a nested case selection that decides
-which page the user wants to see next. Each nesting level refers to a menu
-level of the GUI. Each case implements a certain action of the menu. On the
-deepest level of case selection, the handler essentially knows what the
-user wants and stores the answer into the variable that holds the HTML
-page contents:
-
-@smallexample
-@c file eg/network/eliza.awk
-function HandleGET() @{
- # A real HTTP server would treat some parts of the URI as a file name.
- # We take parts of the URI as menu choices and go on accordingly.
- if(MENU[2] == "AboutServer") @{
- Document = "This is not a CGI script.\
- This is an httpd, an HTML file, and a CGI script all \
- in one GAWK script. It needs no separate www-server, \
- no installation, and no root privileges.\
- <p>To run it, do this:</p><ul>\
- <li> start this script with \"gawk -f httpserver.awk\",</li>\
- <li> and on the same host let your www browser open location\
- \"http://localhost:8080\"</li>\
- </ul>\<p>\ Details of HTTP come from:</p><ul>\
- <li>Hethmon: Illustrated Guide to HTTP</p>\
- <li>RFC 2068</li></ul><p>JK 14.9.1997</p>"
- @} else if (MENU[2] == "AboutELIZA") @{
- Document = "This is an implementation of the famous ELIZA\
- program by Joseph Weizenbaum. It is written in GAWK and\
-/bin/sh: expad: command not found
- @} else if (MENU[2] == "StartELIZA") @{
- gsub(/\+/, " ", GETARG["YouSay"])
- # Here we also have to substitute coded special characters
- Document = "<form method=GET>" \
- "<h3>" ElizaSays(GETARG["YouSay"]) "</h3>\
- <p><input type=text name=YouSay value=\"\" size=60>\
- <br><input type=submit value=\"Tell her about it\"></p></form>"
- @}
-@}
-@c endfile
-@end smallexample
-
-Now we are down to the heart of ELIZA, so you can see how it works.
-Initially the user does not say anything; then ELIZA resets its money
-counter and asks the user to tell what comes to mind open heartedly.
-The subsequent answers are converted to uppercase and stored for
-later comparison. ELIZA presents the bill when being confronted with
-a sentence that contains the phrase ``shut up.'' Otherwise, it looks for
-keywords in the sentence, conjugates the rest of the sentence, remembers
-the keyword for later use, and finally selects an answer from the set of
-possible answers:
-
-@smallexample
-@c file eg/network/eliza.awk
-function ElizaSays(YouSay) @{
- if (YouSay == "") @{
- cost = 0
- answer = "HI, IM ELIZA, TELL ME YOUR PROBLEM"
- @} else @{
- q = toupper(YouSay)
- gsub("'", "", q)
- if(q == qold) @{
- answer = "PLEASE DONT REPEAT YOURSELF !"
- @} else @{
- if (index(q, "SHUT UP") > 0) @{
- answer = "WELL, PLEASE PAY YOUR BILL. ITS EXACTLY ... $"\
- int(100*rand()+30+cost/100)
- @} else @{
- qold = q
- w = "-" # no keyword recognized yet
- for (i in k) @{ # search for keywords
- if (index(q, i) > 0) @{
- w = i
- break
- @}
- @}
- if (w == "-") @{ # no keyword, take old subject
- w = wold
- subj = subjold
- @} else @{ # find subject
- subj = substr(q, index(q, w) + length(w)+1)
- wold = w
- subjold = subj # remember keyword and subject
- @}
- for (i in conj)
- gsub(i, conj[i], q) # conjugation
- # from all answers to this keyword, select one randomly
- answer = r[indices[int(split(k[w], indices) * rand()) + 1]]
- # insert subject into answer
- gsub("_", subj, answer)
- @}
- @}
- @}
- cost += length(answer) # for later payment : 1 cent per character
- return answer
-@}
-@c endfile
-@end smallexample
-
-In the long but simple function @code{SetUpEliza}, you can see tables
-for conjugation, keywords, and answers.@footnote{The version shown
-here is abbreviated. The full version comes with the @command{gawk}
-distribution.} The associative array @code{k}
-contains indices into the array of answers @code{r}. To choose an
-answer, ELIZA just picks an index randomly:
-
-@example
-@c file eg/network/eliza.awk
-function SetUpEliza() @{
- srand()
- wold = "-"
- subjold = " "
-
- # table for conjugation
- conj[" ARE " ] = " AM "
- conj["WERE " ] = "WAS "
- conj[" YOU " ] = " I "
- conj["YOUR " ] = "MY "
- conj[" IVE " ] =\
- conj[" I HAVE " ] = " YOU HAVE "
- conj[" YOUVE " ] =\
- conj[" YOU HAVE "] = " I HAVE "
- conj[" IM " ] =\
- conj[" I AM " ] = " YOU ARE "
- conj[" YOURE " ] =\
- conj[" YOU ARE " ] = " I AM "
-
- # table of all answers
- r[1] = "DONT YOU BELIEVE THAT I CAN _"
- r[2] = "PERHAPS YOU WOULD LIKE TO BE ABLE TO _ ?"
-@c endfile
- @dots{}
-@end example
-@ignore
-@c file eg/network/eliza.awk
- r[3] = "YOU WANT ME TO BE ABLE TO _ ?"
- r[4] = "PERHAPS YOU DONT WANT TO _ "
- r[5] = "DO YOU WANT TO BE ABLE TO _ ?"
- r[6] = "WHAT MAKES YOU THINK I AM _ ?"
- r[7] = "DOES IT PLEASE YOU TO BELIEVE I AM _ ?"
- r[8] = "PERHAPS YOU WOULD LIKE TO BE _ ?"
- r[9] = "DO YOU SOMETIMES WISH YOU WERE _ ?"
- r[10] = "DONT YOU REALLY _ ?"
- r[11] = "WHY DONT YOU _ ?"
- r[12] = "DO YOU WISH TO BE ABLE TO _ ?"
- r[13] = "DOES THAT TROUBLE YOU ?"
- r[14] = "TELL ME MORE ABOUT SUCH FEELINGS"
- r[15] = "DO YOU OFTEN FEEL _ ?"
- r[16] = "DO YOU ENJOY FEELING _ ?"
- r[17] = "DO YOU REALLY BELIEVE I DONT _ ?"
- r[18] = "PERHAPS IN GOOD TIME I WILL _ "
- r[19] = "DO YOU WANT ME TO _ ?"
- r[20] = "DO YOU THINK YOU SHOULD BE ABLE TO _ ?"
- r[21] = "WHY CANT YOU _ ?"
- r[22] = "WHY ARE YOU INTERESTED IN WHETHER OR NOT I AM _ ?"
- r[23] = "WOULD YOU PREFER IF I WERE NOT _ ?"
- r[24] = "PERHAPS IN YOUR FANTASIES I AM _ "
- r[25] = "HOW DO YOU KNOW YOU CANT _ ?"
- r[26] = "HAVE YOU TRIED ?"
- r[27] = "PERHAPS YOU CAN NOW _ "
- r[28] = "DID YOU COME TO ME BECAUSE YOU ARE _ ?"
- r[29] = "HOW LONG HAVE YOU BEEN _ ?"
- r[30] = "DO YOU BELIEVE ITS NORMAL TO BE _ ?"
- r[31] = "DO YOU ENJOY BEING _ ?"
- r[32] = "WE WERE DISCUSSING YOU -- NOT ME"
- r[33] = "Oh, I _"
- r[34] = "YOU'RE NOT REALLY TALKING ABOUT ME, ARE YOU ?"
- r[35] = "WHAT WOULD IT MEAN TO YOU, IF YOU GOT _ ?"
- r[36] = "WHY DO YOU WANT _ ?"
- r[37] = "SUPPOSE YOU SOON GOT _"
- r[38] = "WHAT IF YOU NEVER GOT _ ?"
- r[39] = "I SOMETIMES ALSO WANT _"
- r[40] = "WHY DO YOU ASK ?"
- r[41] = "DOES THAT QUESTION INTEREST YOU ?"
- r[42] = "WHAT ANSWER WOULD PLEASE YOU THE MOST ?"
- r[43] = "WHAT DO YOU THINK ?"
- r[44] = "ARE SUCH QUESTIONS IN YOUR MIND OFTEN ?"
- r[45] = "WHAT IS IT THAT YOU REALLY WANT TO KNOW ?"
- r[46] = "HAVE YOU ASKED ANYONE ELSE ?"
- r[47] = "HAVE YOU ASKED SUCH QUESTIONS BEFORE ?"
- r[48] = "WHAT ELSE COMES TO MIND WHEN YOU ASK THAT ?"
- r[49] = "NAMES DON'T INTEREST ME"
- r[50] = "I DONT CARE ABOUT NAMES -- PLEASE GO ON"
- r[51] = "IS THAT THE REAL REASON ?"
- r[52] = "DONT ANY OTHER REASONS COME TO MIND ?"
- r[53] = "DOES THAT REASON EXPLAIN ANYTHING ELSE ?"
- r[54] = "WHAT OTHER REASONS MIGHT THERE BE ?"
- r[55] = "PLEASE DON'T APOLOGIZE !"
- r[56] = "APOLOGIES ARE NOT NECESSARY"
- r[57] = "WHAT FEELINGS DO YOU HAVE WHEN YOU APOLOGIZE ?"
- r[58] = "DON'T BE SO DEFENSIVE"
- r[59] = "WHAT DOES THAT DREAM SUGGEST TO YOU ?"
- r[60] = "DO YOU DREAM OFTEN ?"
- r[61] = "WHAT PERSONS APPEAR IN YOUR DREAMS ?"
- r[62] = "ARE YOU DISTURBED BY YOUR DREAMS ?"
- r[63] = "HOW DO YOU DO ... PLEASE STATE YOUR PROBLEM"
- r[64] = "YOU DON'T SEEM QUITE CERTAIN"
- r[65] = "WHY THE UNCERTAIN TONE ?"
- r[66] = "CAN'T YOU BE MORE POSITIVE ?"
- r[67] = "YOU AREN'T SURE ?"
- r[68] = "DON'T YOU KNOW ?"
- r[69] = "WHY NO _ ?"
- r[70] = "DON'T SAY NO, IT'S ALWAYS SO NEGATIVE"
- r[71] = "WHY NOT ?"
- r[72] = "ARE YOU SURE ?"
- r[73] = "WHY NO ?"
- r[74] = "WHY ARE YOU CONCERNED ABOUT MY _ ?"
- r[75] = "WHAT ABOUT YOUR OWN _ ?"
- r[76] = "CAN'T YOU THINK ABOUT A SPECIFIC EXAMPLE ?"
- r[77] = "WHEN ?"
- r[78] = "WHAT ARE YOU THINKING OF ?"
- r[79] = "REALLY, ALWAYS ?"
- r[80] = "DO YOU REALLY THINK SO ?"
- r[81] = "BUT YOU ARE NOT SURE YOU _ "
- r[82] = "DO YOU DOUBT YOU _ ?"
- r[83] = "IN WHAT WAY ?"
- r[84] = "WHAT RESEMBLANCE DO YOU SEE ?"
- r[85] = "WHAT DOES THE SIMILARITY SUGGEST TO YOU ?"
- r[86] = "WHAT OTHER CONNECTION DO YOU SEE ?"
- r[87] = "COULD THERE REALLY BE SOME CONNECTIONS ?"
- r[88] = "HOW ?"
- r[89] = "YOU SEEM QUITE POSITIVE"
- r[90] = "ARE YOU SURE ?"
- r[91] = "I SEE"
- r[92] = "I UNDERSTAND"
- r[93] = "WHY DO YOU BRING UP THE TOPIC OF FRIENDS ?"
- r[94] = "DO YOUR FRIENDS WORRY YOU ?"
- r[95] = "DO YOUR FRIENDS PICK ON YOU ?"
- r[96] = "ARE YOU SURE YOU HAVE ANY FRIENDS ?"
- r[97] = "DO YOU IMPOSE ON YOUR FRIENDS ?"
- r[98] = "PERHAPS YOUR LOVE FOR FRIENDS WORRIES YOU"
- r[99] = "DO COMPUTERS WORRY YOU ?"
- r[100] = "ARE YOU TALKING ABOUT ME IN PARTICULAR ?"
- r[101] = "ARE YOU FRIGHTENED BY MACHINES ?"
- r[102] = "WHY DO YOU MENTION COMPUTERS ?"
- r[103] = "WHAT DO YOU THINK MACHINES HAVE TO DO WITH YOUR PROBLEMS ?"
- r[104] = "DON'T YOU THINK COMPUTERS CAN HELP PEOPLE ?"
- r[105] = "WHAT IS IT ABOUT MACHINES THAT WORRIES YOU ?"
- r[106] = "SAY, DO YOU HAVE ANY PSYCHOLOGICAL PROBLEMS ?"
- r[107] = "WHAT DOES THAT SUGGEST TO YOU ?"
- r[108] = "I SEE"
- r[109] = "IM NOT SURE I UNDERSTAND YOU FULLY"
- r[110] = "COME COME ELUCIDATE YOUR THOUGHTS"
- r[111] = "CAN YOU ELABORATE ON THAT ?"
- r[112] = "THAT IS QUITE INTERESTING"
- r[113] = "WHY DO YOU HAVE PROBLEMS WITH MONEY ?"
- r[114] = "DO YOU THINK MONEY IS EVERYTHING ?"
- r[115] = "ARE YOU SURE THAT MONEY IS THE PROBLEM ?"
- r[116] = "I THINK WE WANT TO TALK ABOUT YOU, NOT ABOUT ME"
- r[117] = "WHAT'S ABOUT ME ?"
- r[118] = "WHY DO YOU ALWAYS BRING UP MY NAME ?"
-@c endfile
-@end ignore
-
-@example
-@c file eg/network/eliza.awk
- # table for looking up answers that
- # fit to a certain keyword
- k["CAN YOU"] = "1 2 3"
- k["CAN I"] = "4 5"
- k["YOU ARE"] =\
- k["YOURE"] = "6 7 8 9"
-@c endfile
- @dots{}
-@end example
-@ignore
-@c file eg/network/eliza.awk
- k["I DONT"] = "10 11 12 13"
- k["I FEEL"] = "14 15 16"
- k["WHY DONT YOU"] = "17 18 19"
- k["WHY CANT I"] = "20 21"
- k["ARE YOU"] = "22 23 24"
- k["I CANT"] = "25 26 27"
- k["I AM"] =\
- k["IM "] = "28 29 30 31"
- k["YOU "] = "32 33 34"
- k["I WANT"] = "35 36 37 38 39"
- k["WHAT"] =\
- k["HOW"] =\
- k["WHO"] =\
- k["WHERE"] =\
- k["WHEN"] =\
- k["WHY"] = "40 41 42 43 44 45 46 47 48"
- k["NAME"] = "49 50"
- k["CAUSE"] = "51 52 53 54"
- k["SORRY"] = "55 56 57 58"
- k["DREAM"] = "59 60 61 62"
- k["HELLO"] =\
- k["HI "] = "63"
- k["MAYBE"] = "64 65 66 67 68"
- k[" NO "] = "69 70 71 72 73"
- k["YOUR"] = "74 75"
- k["ALWAYS"] = "76 77 78 79"
- k["THINK"] = "80 81 82"
- k["LIKE"] = "83 84 85 86 87 88 89"
- k["YES"] = "90 91 92"
- k["FRIEND"] = "93 94 95 96 97 98"
- k["COMPUTER"] = "99 100 101 102 103 104 105"
- k["-"] = "106 107 108 109 110 111 112"
- k["MONEY"] = "113 114 115"
- k["ELIZA"] = "116 117 118"
-@c endfile
-@end ignore
-@example
-@c file eg/network/eliza.awk
-@}
-@c endfile
-@end example
-
-@cindex Humphrys, Mark
-@cindex ELIZA program
-@cindex Yahoo!
-Some interesting remarks and details (including the original source code
-of ELIZA) are found on Mark Humphrys' home page. Yahoo! also has a
-page with a collection of ELIZA-like programs. Many of them are written
-in Java, some of them disclosing the Java source code, and a few even
-explain how to modify the Java source code.
-
-@node Caveats, Challenges, Simple Server, Using Networking
-@section Network Programming Caveats
-
-By now it should be clear
-that debugging a networked application is more
-complicated than debugging a single-process single-hosted application.
-The behavior of a networked application sometimes looks non-causal because
-it is not reproducible in a strong sense. Whether a network application
-works or not sometimes depends on the following:
-
-@itemize @bullet
-@item
-How crowded the underlying network is.
-
-@item
-If the party at the other end is running or not.
-
-@item
-The state of the party at the other end.
-@end itemize
-
-@cindex network
-The most difficult problems for a beginner arise from the hidden states of the
-underlying network. After closing a TCP connection, it's often necessary to wait
-a short while before reopening the connection. Even more difficult is the
-establishment of a connection that previously ended with a ``broken pipe.''
-Those connections have to ``time out'' for a minute or so
-before they can reopen.
-Check this with the command @samp{netstat -a}, which
-provides a list of still ``active'' connections.
-
-@node Challenges, , Caveats, Using Networking
-@section Where To Go From Here
-
-@cindex Loebner, Hugh
-@cindex Contest
-Now, you have learned enough to build your own application. You could,
-for example, take part in the
-Loebner Contest
-to win $100,000. The Loebner Prize is the first formal instantiation of
-a Turing Test. Hugh Loebner agreed with The Cambridge Center for Behavioral
-Studies to underwrite a contest designed to implement the Turing Test.
-Dr.@: Loebner pledged a Grand Prize of $100,000 for the first computer whose
-responses were indistinguishable from a human's. Each year an annual prize
-of $2000 and a bronze medal is awarded to the @emph{most} human computer.
-The winner of the annual contest is the best entry relative to other entries
-that year, irrespective of how good it is in an absolute sense. Here is
-an example of a conversation with the winning program of 1997:
-
-@smallexample
-(c)1997 Cambridge Center For Behavioral Studies all rights reserved [CONVERSE]
-[Intelligent Research Ltd.] Start at: [1997/04/29 08:59:40]
-PROGRAM[10:39:20]Hi, my name is Catherine, what's yours?
-JUDGE02[10:39:36]My name is Nathan.
-PROGRAM[10:39:37]Well, it's nice to meet you, Nathan.
-PROGRAM[10:39:45]Did you see that story on CNN last night about the lesbian
-PROGRAM[10:40:03]couple who came out at a White House party on Sunday?
-JUDGE02[10:40:45]No. I'm sorry I missed that. How did the Prez take it?
-PROGRAM[10:40:47]Ellen Degeneres was one of them - she was kissing her lover
-PROGRAM[10:41:08]in front of Bill Clinton. Do you think he was very
-PROGRAM[10:41:21]embarrassed?
-JUDGE02[10:41:37]You tell me. I didn't see it.
-PROGRAM[10:41:38]Are you still a Friend of Bill Clinton?
-JUDGE02[10:42:22]Yes. I would like to talk about something else. Do you know anything about sculpture.
-PROGRAM[10:42:24]Well, I suppose he's doing his best like most of us, but he
-PROGRAM[10:42:35]does seem to attract an UNUSUAL amount of bad publicity,
-PROGRAM[10:42:51]wouldnt you say--can it all be bad luck?
-JUDGE02[10:43:20]I asked you if you knew anything about sculpture.
-PROGRAM[10:43:21]OK, but what happens now--does it all drift on, with
-PROGRAM[10:43:36]everyone waiting to see how much his friends have said about
-PROGRAM[10:43:47]him under threat of lots and lots of jail-or will something
-PROGRAM[10:44:01]dramatic actually happen?
-JUDGE02[10:44:19]Do you know what sculpture is?
-PROGRAM[10:44:20]Well, youre more a fan of his than I am I guess---though I
-PROGRAM[10:44:34]cant SWEAR what Id have done in that locked hotel room, I
-PROGRAM[10:44:51]must say.
-JUDGE02[10:45:26]Have you ever been to the Salmagundi Club?
-PROGRAM[10:45:29]Never. to go back to the original discussion Time to get
-PROGRAM[10:45:43]off this one I think and onto something more down to earth!
-JUDGE02[10:46:20]What is more down to earth than art?
-PROGRAM[10:46:21]I don't know what -- to be honest. -
-PROGRAM[10:46:30]I suppose it depends on where you're coming from, but as the
-PROGRAM[10:46:46]song goes in My Fair Lady, "Why can't a woman be more like a
-PROGRAM[10:47:03]man? "
-JUDGE02[10:47:52]I liked that song but I prefer "Why can't children be more like we are"
-JUDGE02[10:48:10]from another play the name of which I cannot remember.
-PROGRAM[10:48:12]What, if anything, do you enjoy about being a man?
-JUDGE02[10:49:01]I would enjoy all aspects of life no matter what my gender.
-PROGRAM[10:49:02]Do you think of yourself as being attractive?
-@end smallexample
-
-@cindex Clinton, Bill
-This program insists on always speaking about the same story around Bill
-Clinton. You see, even a program with a rather narrow mind can behave so
-much like a human being that it can win this prize. It is quite common to
-let these programs talk to each other via network connections. But during the
-competition itself, the program and its computer have to be present at the
-place the competition is held. We all would love to see a @command{gawk}
-program win in such an event. Maybe it is up to you to accomplish this?
-
-Some other ideas for useful networked applications:
-@itemize @bullet
-@item
-Read the file @file{doc/awkforai.txt} in the @command{gawk} distribution.
-It was written by Ronald P.@: Loui (Associate Professor of
-Computer Science, at Washington University in St. Louis,
-@email{loui@@ai.wustl.edu}) and summarizes why
-he teaches @command{gawk} to students of Artificial Intelligence. Here are
-some passages from the text:
-
-@cindex AI
-@cindex PROLOG
-@cindex Loui, Ronald P.
-@cindex agent
-@quotation
-The GAWK manual can
-be consumed in a single lab session and the language can be mastered by
-the next morning by the average student. GAWK's automatic
-initialization, implicit coercion, I/O support and lack of pointers
-forgive many of the mistakes that young programmers are likely to make.
-Those who have seen C but not mastered it are happy to see that GAWK
-retains some of the same sensibilities while adding what must be
-regarded as spoonsful of syntactic sugar.@*
-@dots{}@*
-@cindex robot
-There are further simple answers. Probably the best is the fact that
-increasingly, undergraduate AI programming is involving the Web. Oren
-Etzioni (University of Washington, Seattle) has for a while been arguing
-that the ``softbot'' is replacing the mechanical engineers' robot as the
-most glamorous AI testbed. If the artifact whose behavior needs to be
-controlled in an intelligent way is the software agent, then a language
-that is well-suited to controlling the software environment is the
-appropriate language. That would imply a scripting language. If the
-robot is KAREL, then the right language is ``turn left; turn right.'' If
-the robot is Netscape, then the right language is something that can
-generate @samp{netscape -remote 'openURL(http://cs.wustl.edu/~loui)'} with
-elan.@*
-@dots{}@*
-AI programming requires high-level thinking. There have always been a few
-gifted programmers who can write high-level programs in assembly language.
-Most however need the ambient abstraction to have a higher floor.@*
-@dots{}@*
-Second, inference is merely the expansion of notation. No matter whether
-the logic that underlies an AI program is fuzzy, probabilistic, deontic,
-defeasible, or deductive, the logic merely defines how strings can be
-transformed into other strings. A language that provides the best
-support for string processing in the end provides the best support for
-logic, for the exploration of various logics, and for most forms of
-symbolic processing that AI might choose to call ``reasoning'' instead of
-``logic.'' The implication is that PROLOG, which saves the AI programmer
-from having to write a unifier, saves perhaps two dozen lines of GAWK
-code at the expense of strongly biasing the logic and representational
-expressiveness of any approach.
-@end quotation
-
-Now that @command{gawk} itself can connect to the Internet, it should be obvious
-that it is suitable for writing intelligent web agents.
-
-@item
-@command{awk} is strong at pattern recognition and string processing.
-So, it is well suited to the classic problem of language translation.
-A first try could be a program that knows the 100 most frequent English
-words and their counterparts in German or French. The service could be
-implemented by regularly reading email with the program above, replacing
-each word by its translation and sending the translation back via SMTP.
-Users would send English email to their translation service and get
-back a translated email message in return. As soon as this works,
-more effort can be spent on a real translation program.
-
-@item
-Another dialogue-oriented application (on the verge
-of ridicule) is the email ``support service.'' Troubled customers write an
-email to an automatic @command{gawk} service that reads the email. It looks
-for keywords in the mail and assembles a reply email accordingly. By carefully
-investigating the email header, and repeating these keywords through the
-reply email, it is rather simple to give the customer a feeling that
-someone cares. Ideally, such a service would search a database of previous
-cases for solutions. If none exists, the database could, for example, consist
-of all the newsgroups, mailing lists and FAQs on the Internet.
-@end itemize
-
-@node Some Applications and Techniques, Links, Using Networking, Top
-@comment node-name, next, previous, up
-
-@chapter Some Applications and Techniques
-In this @value{CHAPTER}, we look at a number of self-contained
-scripts, with an emphasis on concise networking. Along the way, we
-work towards creating building blocks that encapsulate often needed
-functions of the networking world, show new techniques that
-broaden the scope of problems that can be solved with @command{gawk}, and
-explore leading edge technology that may shape the future of networking.
-
-We often refer to the site-independent core of the server that
-we built in
-@ref{Simple Server, ,A Simple Web Server}.
-When building new and non-trivial servers, we
-always copy this building block and append new instances of the two
-functions @code{SetUpServer} and @code{HandleGET}.
-
-This makes a lot of sense, since
-this scheme of event-driven
-execution provides @command{gawk} with an interface to the most widely
-accepted standard for GUIs: the web browser. Now, @command{gawk} can even rival
-Tcl/Tk.
-
-@cindex Tcl/Tk
-@cindex JavaScript
-Tcl and @command{gawk} have much in common. Both are simple scripting languages
-that allow us to quickly solve problems with short programs. But Tcl has Tk
-on top of it and @command{gawk} had nothing comparable up to now. While Tcl
-needs a large and ever changing library (Tk, which was bound to the X Window
-System until recently), @command{gawk} needs just the networking interface
-and some kind of browser on the client's side. Besides better portability,
-the most important advantage of this approach (embracing well-established
-standards such HTTP and HTML) is that @emph{we do not need to change the
-language}. We let others do the work of fighting over protocols and standards.
-We can use HTML, JavaScript, VRML, or whatever else comes along to do our work.
-
-@menu
-* PANIC:: An Emergency Web Server.
-* GETURL:: Retrieving Web Pages.
-* REMCONF:: Remote Configuration Of Embedded Systems.
-* URLCHK:: Look For Changed Web Pages.
-* WEBGRAB:: Extract Links From A Page.
-* STATIST:: Graphing A Statistical Distribution.
-* MAZE:: Walking Through A Maze In Virtual Reality.
-* MOBAGWHO:: A Simple Mobile Agent.
-* STOXPRED:: Stock Market Prediction As A Service.
-* PROTBASE:: Searching Through A Protein Database.
-@end menu
-
-@node PANIC, GETURL, Some Applications and Techniques, Some Applications and Techniques
-@section PANIC: an Emergency Web Server
-@cindex PANIC program
-At first glance, the @code{"Hello, world"} example in
-@ref{Primitive Service, ,A Primitive Web Service},
-seems useless. By adding just a few lines, we can turn it into something useful.
-
-The PANIC program tells everyone who connects that the local
-site is not working. When a web server breaks down, it makes a difference
-if customers get a strange ``network unreachable'' message, or a short message
-telling them that the server has a problem. In such an emergency,
-the hard disk and everything on it (including the regular web service) may
-be unavailable. Rebooting the web server off a diskette makes sense in this
-setting.
-
-To use the PANIC program as an emergency web server, all you need are the
-@command{gawk} executable and the program below on a diskette. By default,
-it connects to port 8080. A different value may be supplied on the
-command line:
-
-@example
-@c file eg/network/panic.awk
-BEGIN @{
- RS = ORS = "\r\n"
- if (MyPort == 0) MyPort = 8080
- HttpService = "/inet/tcp/" MyPort "/0/0"
- Hello = "<HTML><HEAD><TITLE>Out Of Service</TITLE>" \
- "</HEAD><BODY><H1>" \
- "This site is temporarily out of service." \
- "</H1></BODY></HTML>"
- Len = length(Hello) + length(ORS)
- while ("awk" != "complex") @{
- print "HTTP/1.0 200 OK" |& HttpService
- print "Content-Length: " Len ORS |& HttpService
- print Hello |& HttpService
- while ((HttpService |& getline) > 0)
- continue;
- close(HttpService)
- @}
-@}
-@c endfile
-@end example
-
-@node GETURL, REMCONF, PANIC, Some Applications and Techniques
-@section GETURL: Retrieving Web Pages
-@cindex GETURL program
-@cindex robot
-GETURL is a versatile building block for shell scripts that need to retrieve
-files from the Internet. It takes a web address as a command-line parameter and
-tries to retrieve the contents of this address. The contents are printed
-to standard output, while the header is printed to @file{/dev/stderr}.
-A surrounding shell script
-could analyze the contents and extract the text or the links. An ASCII
-browser could be written around GETURL. But more interestingly, web robots are
-straightforward to write on top of GETURL. On the Internet, you can find
-several programs of the same name that do the same job. They are usually
-much more complex internally and at least 10 times longer.
-
-At first, GETURL checks if it was called with exactly one web address.
-Then, it checks if the user chose to use a special proxy server whose name
-is handed over in a variable. By default, it is assumed that the local
-machine serves as proxy. GETURL uses the @code{GET} method by default
-to access the web page. By handing over the name of a different method
-(such as @code{HEAD}), it is possible to choose a different behavior. With
-the @code{HEAD} method, the user does not receive the body of the page
-content, but does receive the header:
-
-@example
-@c file eg/network/geturl.awk
-BEGIN @{
- if (ARGC != 2) @{
- print "GETURL - retrieve Web page via HTTP 1.0"
- print "IN:\n the URL as a command-line parameter"
- print "PARAM(S):\n -v Proxy=MyProxy"
- print "OUT:\n the page content on stdout"
- print " the page header on stderr"
- print "JK 16.05.1997"
- print "ADR 13.08.2000"
- exit
- @}
- URL = ARGV[1]; ARGV[1] = ""
- if (Proxy == "") Proxy = "127.0.0.1"
- if (ProxyPort == 0) ProxyPort = 80
- if (Method == "") Method = "GET"
- HttpService = "/inet/tcp/0/" Proxy "/" ProxyPort
- ORS = RS = "\r\n\r\n"
- print Method " " URL " HTTP/1.0" |& HttpService
- HttpService |& getline Header
- print Header > "/dev/stderr"
- while ((HttpService |& getline) > 0)
- printf "%s", $0
- close(HttpService)
-@}
-@c endfile
-@end example
-
-This program can be changed as needed, but be careful with the last lines.
-Make sure transmission of binary data is not corrupted by additional line
-breaks. Even as it is now, the byte sequence @code{"\r\n\r\n"} would
-disappear if it were contained in binary data. Don't get caught in a
-trap when trying a quick fix on this one.
-
-@node REMCONF, URLCHK, GETURL, Some Applications and Techniques
-@section REMCONF: Remote Configuration of Embedded Systems
-@cindex REMCONF program
-@cindex Linux
-@cindex GNU/Linux
-@cindex Yahoo!
-Today, you often find powerful processors in embedded systems. Dedicated
-network routers and controllers for all kinds of machinery are examples
-of embedded systems. Processors like the Intel 80x86 or the AMD Elan are
-able to run multitasking operating systems, such as XINU or GNU/Linux
-in embedded PCs. These systems are small and usually do not have
-a keyboard or a display. Therefore it is difficult to set up their
-configuration. There are several widespread ways to set them up:
-
-@itemize @bullet
-@item
-DIP switches
-
-@item
-Read Only Memories such as EPROMs
-
-@item
-Serial lines or some kind of keyboard
-
-@item
-Network connections via @command{telnet} or SNMP
-
-@item
-HTTP connections with HTML GUIs
-@end itemize
-
-In this @value{SECTION}, we look at a solution that uses HTTP connections
-to control variables of an embedded system that are stored in a file.
-Since embedded systems have tight limits on resources like memory,
-it is difficult to employ advanced techniques such as SNMP and HTTP
-servers. @command{gawk} fits in quite nicely with its single executable
-which needs just a short script to start working.
-The following program stores the variables in a file, and a concurrent
-process in the embedded system may read the file. The program uses the
-site-independent part of the simple web server that we developed in
-@ref{Interacting Service, ,A Web Service with Interaction}.
-As mentioned there, all we have to do is to write two new procedures
-@code{SetUpServer} and @code{HandleGET}:
-
-@smallexample
-@c file eg/network/remconf.awk
-function SetUpServer() @{
- TopHeader = "<HTML><title>Remote Configuration</title>"
- TopDoc = "<BODY>\
- <h2>Please choose one of the following actions:</h2>\
- <UL>\
- <LI><A HREF=" MyPrefix "/AboutServer>About this server</A></LI>\
- <LI><A HREF=" MyPrefix "/ReadConfig>Read Configuration</A></LI>\
- <LI><A HREF=" MyPrefix "/CheckConfig>Check Configuration</A></LI>\
- <LI><A HREF=" MyPrefix "/ChangeConfig>Change Configuration</A></LI>\
- <LI><A HREF=" MyPrefix "/SaveConfig>Save Configuration</A></LI>\
- </UL>"
- TopFooter = "</BODY></HTML>"
- if (ConfigFile == "") ConfigFile = "config.asc"
-@}
-@c endfile
-@end smallexample
-
-The function @code{SetUpServer} initializes the top level HTML texts
-as usual. It also initializes the name of the file that contains the
-configuration parameters and their values. In case the user supplies
-a name from the command line, that name is used. The file is expected to
-contain one parameter per line, with the name of the parameter in
-column one and the value in column two.
-
-The function @code{HandleGET} reflects the structure of the menu
-tree as usual. The first menu choice tells the user what this is all
-about. The second choice reads the configuration file line by line
-and stores the parameters and their values. Notice that the record
-separator for this file is @code{"\n"}, in contrast to the record separator
-for HTTP. The third menu choice builds an HTML table to show
-the contents of the configuration file just read. The fourth choice
-does the real work of changing parameters, and the last one just saves
-the configuration into a file:
-
-@smallexample
-@c file eg/network/remconf.awk
-function HandleGET() @{
- if(MENU[2] == "AboutServer") @{
- Document = "This is a GUI for remote configuration of an\
- embedded system. It is is implemented as one GAWK script."
- @} else if (MENU[2] == "ReadConfig") @{
- RS = "\n"
- while ((getline < ConfigFile) > 0)
- config[$1] = $2;
- close(ConfigFile)
- RS = "\r\n"
- Document = "Configuration has been read."
- @} else if (MENU[2] == "CheckConfig") @{
- Document = "<TABLE BORDER=1 CELLPADDING=5>"
- for (i in config)
- Document = Document "<TR><TD>" i "</TD>" \
- "<TD>" config[i] "</TD></TR>"
- Document = Document "</TABLE>"
- @} else if (MENU[2] == "ChangeConfig") @{
- if ("Param" in GETARG) @{ # any parameter to set?
- if (GETARG["Param"] in config) @{ # is parameter valid?
- config[GETARG["Param"]] = GETARG["Value"]
- Document = (GETARG["Param"] " = " GETARG["Value"] ".")
- @} else @{
- Document = "Parameter <b>" GETARG["Param"] "</b> is invalid."
- @}
- @} else @{
- Document = "<FORM method=GET><h4>Change one parameter</h4>\
- <TABLE BORDER CELLPADDING=5>\
- <TR><TD>Parameter</TD><TD>Value</TD></TR>\
- <TR><TD><input type=text name=Param value=\"\" size=20></TD>\
- <TD><input type=text name=Value value=\"\" size=40></TD>\
- </TR></TABLE><input type=submit value=\"Set\"></FORM>"
- @}
- @} else if (MENU[2] == "SaveConfig") @{
- for (i in config)
- printf("%s %s\n", i, config[i]) > ConfigFile
- close(ConfigFile)
- Document = "Configuration has been saved."
- @}
-@}
-@c endfile
-@end smallexample
-
-@cindex MiniSQL
-We could also view the configuration file as a database. From this
-point of view, the previous program acts like a primitive database server.
-Real SQL database systems also make a service available by providing
-a TCP port that clients can connect to. But the application level protocols
-they use are usually proprietary and also change from time to time.
-This is also true for the protocol that
-MiniSQL uses.
-
-@node URLCHK, WEBGRAB, REMCONF, Some Applications and Techniques
-@section URLCHK: Look for Changed Web Pages
-@cindex URLCHK program
-Most people who make heavy use of Internet resources have a large
-bookmark file with pointers to interesting web sites. It is impossible
-to regularly check by hand if any of these sites have changed. A program
-is needed to automatically look at the headers of web pages and tell
-which ones have changed. URLCHK does the comparison after using GETURL
-with the @code{HEAD} method to retrieve the header.
-
-Like GETURL, this program first checks that it is called with exactly
-one command-line parameter. URLCHK also takes the same command-line variables
-@code{Proxy} and @code{ProxyPort} as GETURL,
-because these variables are handed over to GETURL for each URL
-that gets checked. The one and only parameter is the name of a file that
-contains one line for each URL. In the first column, we find the URL, and
-the second and third columns hold the length of the URL's body when checked
-for the two last times. Now, we follow this plan:
-
-@enumerate
-@item
-Read the URLs from the file and remember their most recent lengths
-
-@item
-Delete the contents of the file
-
-@item
-For each URL, check its new length and write it into the file
-
-@item
-If the most recent and the new length differ, tell the user
-@end enumerate
-
-It may seem a bit peculiar to read the URLs from a file together
-with their two most recent lengths, but this approach has several
-advantages. You can call the program again and again with the same
-file. After running the program, you can regenerate the changed URLs
-by extracting those lines that differ in their second and third columns:
-
-@c inspired by URLCHK in iX 5/97 166.
-@smallexample
-@c file eg/network/urlchk.awk
-BEGIN @{
- if (ARGC != 2) @{
- print "URLCHK - check if URLs have changed"
- print "IN:\n the file with URLs as a command-line parameter"
- print " file contains URL, old length, new length"
- print "PARAMS:\n -v Proxy=MyProxy -v ProxyPort=8080"
- print "OUT:\n same as file with URLs"
- print "JK 02.03.1998"
- exit
- @}
- URLfile = ARGV[1]; ARGV[1] = ""
- if (Proxy != "") Proxy = " -v Proxy=" Proxy
- if (ProxyPort != "") ProxyPort = " -v ProxyPort=" ProxyPort
- while ((getline < URLfile) > 0)
- Length[$1] = $3 + 0
- close(URLfile) # now, URLfile is read in and can be updated
- GetHeader = "gawk " Proxy ProxyPort " -v Method=\"HEAD\" -f geturl.awk "
- for (i in Length) @{
- GetThisHeader = GetHeader i " 2>&1"
- while ((GetThisHeader | getline) > 0)
- if (toupper($0) ~ /CONTENT-LENGTH/) NewLength = $2 + 0
- close(GetThisHeader)
- print i, Length[i], NewLength > URLfile
- if (Length[i] != NewLength) # report only changed URLs
- print i, Length[i], NewLength
- @}
- close(URLfile)
-@}
-@c endfile
-@end smallexample
-
-Another thing that may look strange is the way GETURL is called.
-Before calling GETURL, we have to check if the proxy variables need
-to be passed on. If so, we prepare strings that will become part
-of the command line later. In @code{GetHeader}, we store these strings
-together with the longest part of the command line. Later, in the loop
-over the URLs, @code{GetHeader} is appended with the URL and a redirection
-operator to form the command that reads the URL's header over the Internet.
-GETURL always produces the headers over @file{/dev/stderr}. That is
-the reason why we need the redirection operator to have the header
-piped in.
-
-This program is not perfect because it assumes that changing URLs
-results in changed lengths, which is not necessarily true. A more
-advanced approach is to look at some other header line that
-holds time information. But, as always when things get a bit more
-complicated, this is left as an exercise to the reader.
-
-@node WEBGRAB, STATIST, URLCHK, Some Applications and Techniques
-@section WEBGRAB: Extract Links from a Page
-@cindex WEBGRAB program
-@c Inspired by iX 1/98 157.
-@cindex robot
-Sometimes it is necessary to extract links from web pages.
-Browsers do it, web robots do it, and sometimes even humans do it.
-Since we have a tool like GETURL at hand, we can solve this problem with
-some help from the Bourne shell:
-
-@example
-@c file eg/network/webgrab.awk
-BEGIN @{ RS = "http://[#%&\\+\\-\\./0-9\\:;\\?A-Z_a-z\\~]*" @}
-RT != "" @{
- command = ("gawk -v Proxy=MyProxy -f geturl.awk " RT \
- " > doc" NR ".html")
- print command
-@}
-@c endfile
-@end example
-
-Notice that the regular expression for URLs is rather crude. A precise
-regular expression is much more complex. But this one works
-rather well. One problem is that it is unable to find internal links of
-an HTML document. Another problem is that
-@samp{ftp}, @samp{telnet}, @samp{news}, @samp{mailto}, and other kinds
-of links are missing in the regular expression.
-However, it is straightforward to add them, if doing so is necessary for other tasks.
-
-This program reads an HTML file and prints all the HTTP links that it finds.
-It relies on @command{gawk}'s ability to use regular expressions as record
-separators. With @code{RS} set to a regular expression that matches links,
-the second action is executed each time a non-empty link is found.
-We can find the matching link itself in @code{RT}.
-
-The action could use the @code{system} function to let another GETURL
-retrieve the page, but here we use a different approach.
-This simple program prints shell commands that can be piped into @command{sh}
-for execution. This way it is possible to first extract
-the links, wrap shell commands around them, and pipe all the shell commands
-into a file. After editing the file, execution of the file retrieves
-exactly those files that we really need. In case we do not want to edit,
-we can retrieve all the pages like this:
-
-@smallexample
-gawk -f geturl.awk http://www.suse.de | gawk -f webgrab.awk | sh
-@end smallexample
-
-@cindex Microsoft Windows
-After this, you will find the contents of all referenced documents in
-files named @file{doc*.html} even if they do not contain HTML code.
-The most annoying thing is that we always have to pass the proxy to
-GETURL. If you do not like to see the headers of the web pages
-appear on the screen, you can redirect them to @file{/dev/null}.
-Watching the headers appear can be quite interesting, because
-it reveals
-interesting details such as which web server the companies use.
-Now, it is clear how the clever marketing people
-use web robots to determine the
-market shares
-of Microsoft and Netscape in the web server market.
-
-Port 80 of any web server is like a small hole in a repellent firewall.
-After attaching a browser to port 80, we usually catch a glimpse
-of the bright side of the server (its home page). With a tool like GETURL
-at hand, we are able to discover some of the more concealed
-or even ``indecent'' services (i.e., lacking conformity to standards of quality).
-It can be exciting to see the fancy CGI scripts that lie
-there, revealing the inner workings of the server, ready to be called:
-
-@itemize @bullet
-@item
-With a command such as:
-
-@example
-gawk -f geturl.awk http://any.host.on.the.net/cgi-bin/
-@end example
-
-some servers give you a directory listing of the CGI files.
-Knowing the names, you can try to call some of them and watch
-for useful results. Sometimes there are executables in such directories
-(such as Perl interpreters) that you may call remotely. If there are
-subdirectories with configuration data of the web server, this can also
-be quite interesting to read.
-
-@item
-@cindex apache
-The well-known Apache web server usually has its CGI files in the
-directory @file{/cgi-bin}. There you can often find the scripts
-@file{test-cgi} and @file{printenv}. Both tell you some things
-about the current connection and the installation of the web server.
-Just call:
-
-@smallexample
-gawk -f geturl.awk http://any.host.on.the.net/cgi-bin/test-cgi
-gawk -f geturl.awk http://any.host.on.the.net/cgi-bin/printenv
-@end smallexample
-
-@item
-Sometimes it is even possible to retrieve system files like the web
-server's log file---possibly containing customer data---or even the file
-@file{/etc/passwd}.
-(We don't recommend this!)
-@end itemize
-
-@strong{Caution:}
-Although this may sound funny or simply irrelevant, we are talking about
-severe security holes. Try to explore your own system this way and make
-sure that none of the above reveals too much information about your system.
-
-@node STATIST, MAZE, WEBGRAB, Some Applications and Techniques
-@section STATIST: Graphing a Statistical Distribution
-@cindex STATIST program
-
-@cindex GNUPlot utility
-@cindex image format
-@cindex @file{gif} image format
-@cindex @file{png} image format
-@cindex @file{ps} image format
-@cindex Boutell, Thomas
-@iftex
-@image{statist,3in}
-@end iftex
-In the HTTP server examples we've shown thus far, we never present an image
-to the browser and its user. Presenting images is one task. Generating
-images that reflect some user input and presenting these dynamically
-generated images is another. In this @value{SECTION}, we use GNUPlot
-for generating @file{.png}, @file{.ps}, or @file{.gif}
-files.@footnote{Due to licensing problems, the default
-installation of GNUPlot disables the generation of @file{.gif} files.
-If your installed version does not accept @samp{set term gif},
-just download and install the most recent version of GNUPlot and the
-@uref{http://www.boutell.com/gd/, GD library}
-by Thomas Boutell.
-Otherwise you still have the chance to generate some
-ASCII-art style images with GNUPlot by using @samp{set term dumb}.
-(We tried it and it worked.)}
-
-The program we develop takes the statistical parameters of two samples
-and computes the t-test statistics. As a result, we get the probabilities
-that the means and the variances of both samples are the same. In order to
-let the user check plausibility, the program presents an image of the
-distributions. The statistical computation follows
-@cite{Numerical Recipes in C: The Art of Scientific Computing}
-by William H.@: Press, Saul A.@: Teukolsky, William T.@: Vetterling, and Brian P. Flannery.
-Since @command{gawk} does not have a built-in function
-for the computation of the beta function, we use the @code{ibeta} function
-of GNUPlot. As a side effect, we learn how to use GNUPlot as a
-sophisticated calculator. The comparison of means is done as in @code{tutest},
-paragraph 14.2, page 613, and the comparison of variances is done as in @code{ftest},
-page 611 in @cite{Numerical Recipes}.
-@cindex Numerical Recipes
-
-As usual, we take the site-independent code for servers and append
-our own functions @code{SetUpServer} and @code{HandleGET}:
-
-@smallexample
-@c file eg/network/statist.awk
-function SetUpServer() @{
- TopHeader = "<HTML><title>Statistics with GAWK</title>"
- TopDoc = "<BODY>\
- <h2>Please choose one of the following actions:</h2>\
- <UL>\
- <LI><A HREF=" MyPrefix "/AboutServer>About this server</A></LI>\
- <LI><A HREF=" MyPrefix "/EnterParameters>Enter Parameters</A></LI>\
- </UL>"
- TopFooter = "</BODY></HTML>"
- GnuPlot = "gnuplot 2>&1"
- m1=m2=0; v1=v2=1; n1=n2=10
-@}
-@c endfile
-@end smallexample
-
-Here, you see the menu structure that the user sees. Later, we
-will see how the program structure of the @code{HandleGET} function
-reflects the menu structure. What is missing here is the link for the
-image we generate. In an event-driven environment, request,
-generation, and delivery of images are separated.
-
-Notice the way we initialize the @code{GnuPlot} command string for
-the pipe. By default,
-GNUPlot outputs the generated image via standard output, as well as
-the results of @code{print}(ed) calculations via standard error.
-The redirection causes standard error to be mixed into standard
-output, enabling us to read results of calculations with @code{getline}.
-By initializing the statistical parameters with some meaningful
-defaults, we make sure the user gets an image the first time
-he uses the program.
-
-@cindex JavaScript
-Following is the rather long function @code{HandleGET}, which
-implements the contents of this service by reacting to the different
-kinds of requests from the browser. Before you start playing with
-this script, make sure that your browser supports JavaScript and that it also
-has this option switched on. The script uses a short snippet of
-JavaScript code for delayed opening of a window with an image.
-A more detailed explanation follows:
-
-@smallexample
-@c file eg/network/statist.awk
-function HandleGET() @{
- if(MENU[2] == "AboutServer") @{
- Document = "This is a GUI for a statistical computation.\
- It compares means and variances of two distributions.\
- It is implemented as one GAWK script and uses GNUPLOT."
- @} else if (MENU[2] == "EnterParameters") @{
- Document = ""
- if ("m1" in GETARG) @{ # are there parameters to compare?
- Document = Document "<SCRIPT LANGUAGE=\"JavaScript\">\
- setTimeout(\"window.open(\\\"" MyPrefix "/Image" systime()\
- "\\\",\\\"dist\\\", \\\"status=no\\\");\", 1000); </SCRIPT>"
- m1 = GETARG["m1"]; v1 = GETARG["v1"]; n1 = GETARG["n1"]
- m2 = GETARG["m2"]; v2 = GETARG["v2"]; n2 = GETARG["n2"]
- t = (m1-m2)/sqrt(v1/n1+v2/n2)
- df = (v1/n1+v2/n2)*(v1/n1+v2/n2)/((v1/n1)*(v1/n1)/(n1-1) \
- + (v2/n2)*(v2/n2) /(n2-1))
- if (v1>v2) @{
- f = v1/v2
- df1 = n1 - 1
- df2 = n2 - 1
- @} else @{
- f = v2/v1
- df1 = n2 - 1
- df2 = n1 - 1
- @}
- print "pt=ibeta(" df/2 ",0.5," df/(df+t*t) ")" |& GnuPlot
- print "pF=2.0*ibeta(" df2/2 "," df1/2 "," \
- df2/(df2+df1*f) ")" |& GnuPlot
- print "print pt, pF" |& GnuPlot
- RS="\n"; GnuPlot |& getline; RS="\r\n" # $1 is pt, $2 is pF
- print "invsqrt2pi=1.0/sqrt(2.0*pi)" |& GnuPlot
- print "nd(x)=invsqrt2pi/sd*exp(-0.5*((x-mu)/sd)**2)" |& GnuPlot
- print "set term png small color" |& GnuPlot
- #print "set term postscript color" |& GnuPlot
- #print "set term gif medium size 320,240" |& GnuPlot
- print "set yrange[-0.3:]" |& GnuPlot
- print "set label 'p(m1=m2) =" $1 "' at 0,-0.1 left" |& GnuPlot
- print "set label 'p(v1=v2) =" $2 "' at 0,-0.2 left" |& GnuPlot
- print "plot mu=" m1 ",sd=" sqrt(v1) ", nd(x) title 'sample 1',\
- mu=" m2 ",sd=" sqrt(v2) ", nd(x) title 'sample 2'" |& GnuPlot
- print "quit" |& GnuPlot
- GnuPlot |& getline Image
- while ((GnuPlot |& getline) > 0)
- Image = Image RS $0
- close(GnuPlot)
- @}
- Document = Document "\
- <h3>Do these samples have the same Gaussian distribution?</h3>\
- <FORM METHOD=GET> <TABLE BORDER CELLPADDING=5>\
- <TR>\
- <TD>1. Mean </TD>
- <TD><input type=text name=m1 value=" m1 " size=8></TD>\
- <TD>1. Variance</TD>
- <TD><input type=text name=v1 value=" v1 " size=8></TD>\
- <TD>1. Count </TD>
- <TD><input type=text name=n1 value=" n1 " size=8></TD>\
- </TR><TR>\
- <TD>2. Mean </TD>
- <TD><input type=text name=m2 value=" m2 " size=8></TD>\
- <TD>2. Variance</TD>
- <TD><input type=text name=v2 value=" v2 " size=8></TD>\
- <TD>2. Count </TD>
- <TD><input type=text name=n2 value=" n2 " size=8></TD>\
- </TR> <input type=submit value=\"Compute\">\
- </TABLE></FORM><BR>"
- @} else if (MENU[2] ~ "Image") @{
- Reason = "OK" ORS "Content-type: image/png"
- #Reason = "OK" ORS "Content-type: application/x-postscript"
- #Reason = "OK" ORS "Content-type: image/gif"
- Header = Footer = ""
- Document = Image
- @}
-@}
-@c endfile
-@end smallexample
-
-@cindex PostScript
-As usual, we give a short description of the service in the first
-menu choice. The third menu choice shows us that generation and
-presentation of an image are two separate actions. While the latter
-takes place quite instantly in the third menu choice, the former
-takes place in the much longer second choice. Image data passes from the
-generating action to the presenting action via the variable @code{Image}
-that contains a complete @file{.png} image, which is otherwise stored
-in a file. If you prefer @file{.ps} or @file{.gif} images over the
-default @file{.png} images, you may select these options by uncommenting
-the appropriate lines. But remember to do so in two places: when
-telling GNUPlot which kind of images to generate, and when transmitting the
-image at the end of the program.
-
-Looking at the end of the program,
-the way we pass the @samp{Content-type} to the browser is a bit unusual.
-It is appended to the @samp{OK} of the first header line
-to make sure the type information becomes part of the header.
-The other variables that get transmitted across the network are
-made empty, because in this case we do not have an HTML document to
-transmit, but rather raw image data to contain in the body.
-
-Most of the work is done in the second menu choice. It starts with a
-strange JavaScript code snippet. When first implementing this server,
-we used a short @code{@w{"<IMG SRC="} MyPrefix "/Image>"} here. But then
-browsers got smarter and tried to improve on speed by requesting the
-image and the HTML code at the same time. When doing this, the browser
-tries to build up a connection for the image request while the request for
-the HTML text is not yet completed. The browser tries to connect
-to the @command{gawk} server on port 8080 while port 8080 is still in use for
-transmission of the HTML text. The connection for the image cannot be
-built up, so the image appears as ``broken'' in the browser window.
-We solved this problem by telling the browser to open a separate window
-for the image, but only after a delay of 1000 milliseconds.
-By this time, the server should be ready for serving the next request.
-
-But there is one more subtlety in the JavaScript code.
-Each time the JavaScript code opens a window for the image, the
-name of the image is appended with a timestamp (@code{systime}).
-Why this constant change of name for the image? Initially, we always named
-the image @code{Image}, but then the Netscape browser noticed the name
-had @emph{not} changed since the previous request and displayed the
-previous image (caching behavior). The server core
-is implemented so that browsers are told @emph{not} to cache anything.
-Obviously HTTP requests do not always work as expected. One way to
-circumvent the cache of such overly smart browsers is to change the
-name of the image with each request. These three lines of JavaScript
-caused us a lot of trouble.
-
-The rest can be broken
-down into two phases. At first, we check if there are statistical
-parameters. When the program is first started, there usually are no
-parameters because it enters the page coming from the top menu.
-Then, we only have to present the user a form that he can use to change
-statistical parameters and submit them. Subsequently, the submission of
-the form causes the execution of the first phase because @emph{now}
-there @emph{are} parameters to handle.
-
-Now that we have parameters, we know there will be an image available.
-Therefore we insert the JavaScript code here to initiate the opening
-of the image in a separate window. Then,
-we prepare some variables that will be passed to GNUPlot for calculation
-of the probabilities. Prior to reading the results, we must temporarily
-change @code{RS} because GNUPlot separates lines with newlines.
-After instructing GNUPlot to generate a @file{.png} (or @file{.ps} or
-@file{.gif}) image, we initiate the insertion of some text,
-explaining the resulting probabilities. The final @samp{plot} command
-actually generates the image data. This raw binary has to be read in carefully
-without adding, changing, or deleting a single byte. Hence the unusual
-initialization of @code{Image} and completion with a @code{while} loop.
-
-When using this server, it soon becomes clear that it is far from being
-perfect. It mixes source code of six scripting languages or protocols:
-
-@itemize @bullet
-@item GNU @command{awk} implements a server for the protocol:
-@item HTTP which transmits:
-@item HTML text which contains a short piece of:
-@item JavaScript code opening a separate window.
-@item A Bourne shell script is used for piping commands into:
-@item GNUPlot to generate the image to be opened.
-@end itemize
-
-After all this work, the GNUPlot image opens in the JavaScript window
-where it can be viewed by the user.
-
-It is probably better not to mix up so many different languages.
-The result is not very readable. Furthermore, the
-statistical part of the server does not take care of invalid input.
-Among others, using negative variances will cause invalid results.
-
-@node MAZE, MOBAGWHO, STATIST, Some Applications and Techniques
-@section MAZE: Walking Through a Maze In Virtual Reality
-@cindex MAZE
-@cindex VRML
-@c VRML in iX 11/96 134.
-@quotation
-@cindex Perlis, Alan
-@i{In the long run, every program becomes rococo, and then rubble.}@*
-Alan Perlis
-@end quotation
-
-By now, we know how to present arbitrary @samp{Content-type}s to a browser.
-In this @value{SECTION}, our server will present a 3D world to our browser.
-The 3D world is described in a scene description language (VRML,
-Virtual Reality Modeling Language) that allows us to travel through a
-perspective view of a 2D maze with our browser. Browsers with a
-VRML plugin enable exploration of this technology. We could do
-one of those boring @samp{Hello world} examples here, that are usually
-presented when introducing novices to
-VRML. If you have never written
-any VRML code, have a look at
-the VRML FAQ.
-Presenting a static VRML scene is a bit trivial; in order to expose
-@command{gawk}'s new capabilities, we will present a dynamically generated
-VRML scene. The function @code{SetUpServer} is very simple because it
-only sets the default HTML page and initializes the random number
-generator. As usual, the surrounding server lets you browse the maze.
-
-@smallexample
-@c file eg/network/maze.awk
-function SetUpServer() @{
- TopHeader = "<HTML><title>Walk through a maze</title>"
- TopDoc = "\
- <h2>Please choose one of the following actions:</h2>\
- <UL>\
- <LI><A HREF=" MyPrefix "/AboutServer>About this server</A>\
- <LI><A HREF=" MyPrefix "/VRMLtest>Watch a simple VRML scene</A>\
- </UL>"
- TopFooter = "</HTML>"
- srand()
-@}
-@c endfile
-@end smallexample
-
-The function @code{HandleGET} is a bit longer because it first computes
-the maze and afterwards generates the VRML code that is sent across
-the network. As shown in the STATIST example
-(@pxref{STATIST}),
-we set the type of the
-content to VRML and then store the VRML representation of the maze as the
-page content. We assume that the maze is stored in a 2D array. Initially,
-the maze consists of walls only. Then, we add an entry and an exit to the
-maze and let the rest of the work be done by the function @code{MakeMaze}.
-Now, only the wall fields are left in the maze. By iterating over the these
-fields, we generate one line of VRML code for each wall field.
-
-@smallexample
-@c file eg/network/maze.awk
-function HandleGET() @{
- if (MENU[2] == "AboutServer") @{
- Document = "If your browser has a VRML 2 plugin,\
- this server shows you a simple VRML scene."
- @} else if (MENU[2] == "VRMLtest") @{
- XSIZE = YSIZE = 11 # initially, everything is wall
- for (y = 0; y < YSIZE; y++)
- for (x = 0; x < XSIZE; x++)
- Maze[x, y] = "#"
- delete Maze[0, 1] # entry is not wall
- delete Maze[XSIZE-1, YSIZE-2] # exit is not wall
- MakeMaze(1, 1)
- Document = "\
-#VRML V2.0 utf8\n\
-Group @{\n\
- children [\n\
- PointLight @{\n\
- ambientIntensity 0.2\n\
- color 0.7 0.7 0.7\n\
- location 0.0 8.0 10.0\n\
- @}\n\
- DEF B1 Background @{\n\
- skyColor [0 0 0, 1.0 1.0 1.0 ]\n\
- skyAngle 1.6\n\
- groundColor [1 1 1, 0.8 0.8 0.8, 0.2 0.2 0.2 ]\n\
- groundAngle [ 1.2 1.57 ]\n\
- @}\n\
- DEF Wall Shape @{\n\
- geometry Box @{size 1 1 1@}\n\
- appearance Appearance @{ material Material @{ diffuseColor 0 0 1 @} @}\n\
- @}\n\
- DEF Entry Viewpoint @{\n\
- position 0.5 1.0 5.0\n\
- orientation 0.0 0.0 -1.0 0.52\n\
- @}\n"
- for (i in Maze) @{
- split(i, t, SUBSEP)
- Document = Document " Transform @{ translation "
- Document = Document t[1] " 0 -" t[2] " children USE Wall @}\n"
- @}
- Document = Document " ] # end of group for world\n@}"
- Reason = "OK" ORS "Content-type: model/vrml"
- Header = Footer = ""
- @}
-@}
-@c endfile
-@end smallexample
-
-Finally, we have a look at @code{MakeMaze}, the function that generates
-the @code{Maze} array. When entered, this function assumes that the array
-has been initialized so that each element represents a wall element and
-the maze is initially full of wall elements. Only the entrance and the exit
-of the maze should have been left free. The parameters of the function tell
-us which element must be marked as not being a wall. After this, we take
-a look at the four neighbouring elements and remember which we have already
-treated. Of all the neighbouring elements, we take one at random and
-walk in that direction. Therefore, the wall element in that direction has
-to be removed and then, we call the function recursively for that element.
-The maze is only completed if we iterate the above procedure for
-@emph{all} neighbouring elements (in random order) and for our present
-element by recursively calling the function for the present element. This
-last iteration could have been done in a loop,
-but it is done much simpler recursively.
-
-Notice that elements with coordinates that are both odd are assumed to be
-on our way through the maze and the generating process cannot terminate
-as long as there is such an element not being @code{delete}d. All other
-elements are potentially part of the wall.
-
-@smallexample
-@c file eg/network/maze.awk
-function MakeMaze(x, y) @{
- delete Maze[x, y] # here we are, we have no wall here
- p = 0 # count unvisited fields in all directions
- if (x-2 SUBSEP y in Maze) d[p++] = "-x"
- if (x SUBSEP y-2 in Maze) d[p++] = "-y"
- if (x+2 SUBSEP y in Maze) d[p++] = "+x"
- if (x SUBSEP y+2 in Maze) d[p++] = "+y"
- if (p>0) @{ # if there are univisited fields, go there
- p = int(p*rand()) # choose one unvisited field at random
- if (d[p] == "-x") @{ delete Maze[x - 1, y]; MakeMaze(x - 2, y)
- @} else if (d[p] == "-y") @{ delete Maze[x, y - 1]; MakeMaze(x, y - 2)
- @} else if (d[p] == "+x") @{ delete Maze[x + 1, y]; MakeMaze(x + 2, y)
- @} else if (d[p] == "+y") @{ delete Maze[x, y + 1]; MakeMaze(x, y + 2)
- @} # we are back from recursion
- MakeMaze(x, y); # try again while there are unvisited fields
- @}
-@}
-@c endfile
-@end smallexample
-
-@node MOBAGWHO, STOXPRED, MAZE, Some Applications and Techniques
-@section MOBAGWHO: a Simple Mobile Agent
-@cindex MOBAGWHO program
-@cindex agent
-@quotation
-@cindex Hoare, C.A.R.
-@i{There are two ways of constructing a software design: One way is to
-make it so simple that there are obviously no deficiencies, and the
-other way is to make it so complicated that there are no obvious
-deficiencies.} @*
-C. A. R. Hoare
-@end quotation
-
-A @dfn{mobile agent} is a program that can be dispatched from a computer and
-transported to a remote server for execution. This is called @dfn{migration},
-which means that a process on another system is started that is independent
-from its originator. Ideally, it wanders through
-a network while working for its creator or owner. In places like
-the UMBC Agent Web,
-people are quite confident that (mobile) agents are a software engineering
-paradigm that enables us to significantly increase the efficiency
-of our work. Mobile agents could become the mediators between users and
-the networking world. For an unbiased view at this technology,
-see the remarkable paper @cite{Mobile Agents: Are they a good
-idea?}.@footnote{@uref{http://www.research.ibm.com/massive/mobag.ps}}
-
-@ignore
-@c Chuck says to take all of this out.
-@cindex Tcl/Tk
-A good instance of this paradigm is
-@cite{Agent Tcl},@footnote{@uref{http://agent.cs.dartmouth.edu/software/agent2.0/}}
-an extension of the Tcl language. After introducing a typical
-development environment, the aforementioned paper shows a nice little
-example application that we will try to rebuild in @command{gawk}. The
-@command{who} agent takes a list of servers and wanders from one server
-to the next one, always looking to see who is logged in.
-Having reached the last
-one, it sends back a message with a list of all users it found on each
-machine.
-
-But before implementing something that might or might not be a mobile
-agent, let us clarify the concept and some important terms. The agent
-paradigm in general is such a young scientific discipline that it has
-not yet developed a widely-accepted terminology. Some authors try to
-give precise definitions, but their scope is often not wide enough
-to be generally accepted. Franklin and Graesser ask
-@cite{Is it an Agent or just a Program: A Taxonomy for Autonomous
-Agents}@footnote{@uref{http://www.msci.memphis.edu/~franklin/AgentProg.html}}
-and give even better answers than Caglayan and Harrison in their
-@cite{Agent Sourcebook}.@footnote{@uref{http://www.aminda.com/mazzu/sourcebook/}}
-
-@itemize @minus
-@item
-@i{An autonomous agent is a system situated within and a part of
-an environment that senses that environment and acts on it, over time, in
-pursuit of its own agenda and so as to effect what it senses in the future.}
-(Quoted from Franklin and Graesser.)
-@item
-A mobile agent is able to transport itself from one machine to another.
-@item
-The term @dfn{migration} often denotes this process of moving.
-But neither of the two sources above even mentions this term, while others
-use it regularly.
-@end itemize
-
-Before delving into the (rather demanding) details of
-implementation, let us give just one more quotation as a final
-motivation. Steven Farley published an excellent paper called
-@cite{Mobile Agent System Architecture},@footnote{This often
-cited text originally appeared as a conference paper here:
-@uref{http://www.sigs.com/publications/docs/java/9705/farley.html}
-Many bibliographies on the Internet point to this dead link. Meanwhile,
-the paper appeared as a contribution to a book called More Java Gems here:
-@uref{http://uk.cambridge.org/computerscience/object/catalogue/0521774772/default.htm}}
-in which he asks ``Why use an agent architecture?''
-
-@quotation
-If client-server systems are the currently established norm and distributed
-object systems such as CORBA are defining the future standards, why bother
-with agents? Agent architectures have certain advantages over these other
-types. Three of the most important advantages are:
-@cindex CORBA
-
-@enumerate
-@item
-An agent performs much processing at the server where local bandwidth
-is high, thus reducing the amount of network bandwidth consumed and increasing
-overall performance. In contrast, a CORBA client object with the equivalent
-functionality of a given agent must make repeated remote method calls to
-the server object because CORBA objects cannot move across the network
-at runtime.
-
-@item
-An agent operates independently of the application from which the
-agent was invoked. The agent operates asynchronously, meaning that the
-client application does not need to wait for the results. This is especially
-important for mobile users who are not always connected to the network.
-
-@item
-The use of agents allows for the injection of new functionality into
-a system at run time. An agent system essentially contains its own automatic
-software distribution mechanism. Since CORBA has no built-in support for
-mobile code, new functionality generally has to be installed manually.
-
-@end enumerate
-
-Of course a non-agent system can exhibit these same features with some
-work. But the mobile code paradigm supports the transfer of executable
-code to a remote location for asynchronous execution from the start. An
-agent architecture should be considered for systems where the above features
-are primary requirements.
-@end quotation
-@end ignore
-
-When trying to migrate a process from one system to another,
-a server process is needed on the receiving side. Depending on the kind
-of server process, several ways of implementation come to mind.
-How the process is implemented depends upon the kind of server process:
-
-@itemize @bullet
-@item
-HTTP can be used as the protocol for delivery of the migrating
-process. In this case, we use a common web
-server as the receiving server process. A universal CGI script
-mediates between migrating process and web server.
-Each server willing to accept migrating agents makes this universal
-service available. HTTP supplies the @code{POST} method to transfer
-some data to a file on the web server. When a CGI script is called
-remotely with the @code{POST} method instead of the usual @code{GET} method,
-data is transmitted from the client process to the standard input
-of the server's CGI script. So, to implement a mobile agent,
-we must not only write the agent program to start on the client
-side, but also the CGI script to receive the agent on the server side.
-
-@cindex CGI
-@cindex apache
-@item
-The @code{PUT} method can also be used for migration. HTTP does not
-require a CGI script for migration via @code{PUT}. However, with common web
-servers there is no advantage to this solution, because web servers such as
-Apache
-require explicit activation of a special @code{PUT} script.
-
-@item
-@cite{Agent Tcl} pursues a different course; it relies on a dedicated server
-process with a dedicated protocol specialized for receiving mobile agents.
-@end itemize
-
-Our agent example abuses a common web server as a migration tool. So, it needs a
-universal CGI script on the receiving side (the web server). The receiving script is
-activated with a @code{POST} request when placed into a location like
-@file{/httpd/cgi-bin/PostAgent.sh}. Make sure that the server system uses a
-version of @command{gawk} that supports network access (Version 3.1 or later;
-verify with @samp{gawk --version}).
-
-@example
-@c file eg/network/PostAgent.sh
-#!/bin/sh
-MobAg=/tmp/MobileAgent.$$
-# direct script to mobile agent file
-cat > $MobAg
-# execute agent concurrently
-gawk -f $MobAg $MobAg > /dev/null &
-# HTTP header, terminator and body
-gawk 'BEGIN @{ print "\r\nAgent started" @}'
-rm $MobAg # delete script file of agent
-@c endfile
-@end example
-
-By making its process id (@code{$$}) part of the unique @value{FN}, the
-script avoids conflicts between concurrent instances of the script.
-First, all lines
-from standard input (the mobile agent's source code) are copied into
-this unique file. Then, the agent is started as a concurrent process
-and a short message reporting this fact is sent to the submitting client.
-Finally, the script file of the mobile agent is removed because it is
-no longer needed. Although it is a short script, there are several noteworthy
-points:
-
-@table @asis
-@item Security
-@emph{There is none}. In fact, the CGI script should never
-be made available on a server that is part of the Internet because everyone
-would be allowed to execute arbitrary commands with it. This behavior is
-acceptable only when performing rapid prototyping.
-
-@item Self-Reference
-Each migrating instance of an agent is started
-in a way that enables it to read its own source code from standard input
-and use the code for subsequent
-migrations. This is necessary because it needs to treat the agent's code
-as data to transmit. @command{gawk} is not the ideal language for such
-a job. Lisp and Tcl are more suitable because they do not make a distinction
-between program code and data.
-
-@item Independence
-After migration, the agent is not linked to its
-former home in any way. By reporting @samp{Agent started}, it waves
-``Goodbye'' to its origin. The originator may choose to terminate or not.
-@end table
-
-@cindex Lisp
-The originating agent itself is started just like any other command-line
-script, and reports the results on standard output. By letting the name
-of the original host migrate with the agent, the agent that migrates
-to a host far away from its origin can report the result back home.
-Having arrived at the end of the journey, the agent establishes
-a connection and reports the results. This is the reason for
-determining the name of the host with @samp{uname -n} and storing it
-in @code{MyOrigin} for later use. We may also set variables with the
-@option{-v} option from the command line. This interactivity is only
-of importance in the context of starting a mobile agent; therefore this
-@code{BEGIN} pattern and its action do not take part in migration:
-
-@smallexample
-@c file eg/network/mobag.awk
-BEGIN @{
- if (ARGC != 2) @{
- print "MOBAG - a simple mobile agent"
- print "CALL:\n gawk -f mobag.awk mobag.awk"
- print "IN:\n the name of this script as a command-line parameter"
- print "PARAM:\n -v MyOrigin=myhost.com"
- print "OUT:\n the result on stdout"
- print "JK 29.03.1998 01.04.1998"
- exit
- @}
- if (MyOrigin == "") @{
- "uname -n" | getline MyOrigin
- close("uname -n")
- @}
-@}
-@c endfile
-@end smallexample
-
-Since @command{gawk} cannot manipulate and transmit parts of the program
-directly, the source code is read and stored in strings.
-Therefore, the program scans itself for
-the beginning and the ending of functions.
-Each line in between is appended to the code string until the end of
-the function has been reached. A special case is this part of the program
-itself. It is not a function.
-Placing a similar framework around it causes it to be treated
-like a function. Notice that this mechanism works for all the
-functions of the source code, but it cannot guarantee that the order
-of the functions is preserved during migration:
-
-@smallexample
-@c file eg/network/mobag.awk
-#ReadMySelf
-/^function / @{ FUNC = $2 @}
-/^END/ || /^#ReadMySelf/ @{ FUNC = $1 @}
-FUNC != "" @{ MOBFUN[FUNC] = MOBFUN[FUNC] RS $0 @}
-(FUNC != "") && (/^@}/ || /^#EndOfMySelf/) \
- @{ FUNC = "" @}
-#EndOfMySelf
-@c endfile
-@end smallexample
-
-The web server code in
-@ref{Interacting Service, ,A Web Service with Interaction},
-was first developed as a site-independent core. Likewise, the
-@command{gawk}-based mobile agent
-starts with an agent-independent core, to which can be appended
-application-dependent functions. What follows is the only
-application-independent function needed for the mobile agent:
-
-@smallexample
-@c file eg/network/mobag.awk
-function migrate(Destination, MobCode, Label) @{
- MOBVAR["Label"] = Label
- MOBVAR["Destination"] = Destination
- RS = ORS = "\r\n"
- HttpService = "/inet/tcp/0/" Destination
- for (i in MOBFUN)
- MobCode = (MobCode "\n" MOBFUN[i])
- MobCode = MobCode "\n\nBEGIN @{"
- for (i in MOBVAR)
- MobCode = (MobCode "\n MOBVAR[\"" i "\"] = \"" MOBVAR[i] "\"")
- MobCode = MobCode "\n@}\n"
- print "POST /cgi-bin/PostAgent.sh HTTP/1.0" |& HttpService
- print "Content-length:", length(MobCode) ORS |& HttpService
- printf "%s", MobCode |& HttpService
- while ((HttpService |& getline) > 0)
- print $0
- close(HttpService)
-@}
-@c endfile
-@end smallexample
-
-The @code{migrate} function prepares the
-aforementioned strings containing the program code and transmits them to a
-server. A consequence of this modular approach is that the @code{migrate}
-function takes some parameters that aren't needed in this application,
-but that will be in future ones. Its mandatory parameter @code{Destination} holds the
-name (or IP address) of the server that the agent wants as a host for its
-code. The optional parameter @code{MobCode} may contain some @command{gawk}
-code that is inserted during migration in front of all other code.
-The optional parameter @code{Label} may contain
-a string that tells the agent what to do in program execution after
-arrival at its new home site. One of the serious obstacles in implementing
-a framework for mobile agents is that it does not suffice to migrate the
-code. It is also necessary to migrate the state of execution of the agent. In
-contrast to @cite{Agent Tcl}, this program does not try to migrate the complete set
-of variables. The following conventions are used:
-
-@itemize @bullet
-@item
-Each variable in an agent program is local to the current host and does
-@emph{not} migrate.
-
-@item
-The array @code{MOBFUN} shown above is an exception. It is handled
-by the function @code{migrate} and does migrate with the application.
-
-@item
-The other exception is the array @code{MOBVAR}. Each variable that
-takes part in migration has to be an element of this array.
-@code{migrate} also takes care of this.
-@end itemize
-
-Now it's clear what happens to the @code{Label} parameter of the
-function @code{migrate}. It is copied into @code{MOBVAR["Label"]} and
-travels alongside the other data. Since travelling takes place via HTTP,
-records must be separated with @code{"\r\n"} in @code{RS} and
-@code{ORS} as usual. The code assembly for migration takes place in
-three steps:
-
-@itemize @bullet
-@item
-Iterate over @code{MOBFUN} to collect all functions verbatim.
-
-@item
-Prepare a @code{BEGIN} pattern and put assignments to mobile
-variables into the action part.
-
-@item
-Transmission itself resembles GETURL: the header with the request
-and the @code{Content-length} is followed by the body. In case there is
-any reply over the network, it is read completely and echoed to
-standard output to avoid irritating the server.
-@end itemize
-
-The application-independent framework is now almost complete. What follows
-is the @code{END} pattern that is executed when the mobile agent has
-finished reading its own code. First, it checks whether it is already
-running on a remote host or not. In case initialization has not yet taken
-place, it starts @code{MyInit}. Otherwise (later, on a remote host), it
-starts @code{MyJob}:
-
-@smallexample
-@c file eg/network/mobag.awk
-END @{
- if (ARGC != 2) exit # stop when called with wrong parameters
- if (MyOrigin != "") # is this the originating host?
- MyInit() # if so, initialize the application
- else # we are on a host with migrated data
- MyJob() # so we do our job
-@}
-@c endfile
-@end smallexample
-
-All that's left to extend the framework into a complete application
-is to write two application-specific functions: @code{MyInit} and
-@code{MyJob}. Keep in mind that the former is executed once on the
-originating host, while the latter is executed after each migration:
-
-@smallexample
-@c file eg/network/mobag.awk
-function MyInit() @{
- MOBVAR["MyOrigin"] = MyOrigin
- MOBVAR["Machines"] = "localhost/80 max/80 moritz/80 castor/80"
- split(MOBVAR["Machines"], Machines) # which host is the first?
- migrate(Machines[1], "", "") # go to the first host
- while (("/inet/tcp/8080/0/0" |& getline) > 0) # wait for result
- print $0 # print result
- close("/inet/tcp/8080/0/0")
-@}
-@c endfile
-@end smallexample
-
-As mentioned earlier, this agent takes the name of its origin
-(@code{MyOrigin}) with it. Then, it takes the name of its first
-destination and goes there for further work. Notice that this name has
-the port number of the web server appended to the name of the server,
-because the function @code{migrate} needs it this way to create
-the @code{HttpService} variable. Finally, it waits for the result to arrive.
-The @code{MyJob} function runs on the remote host:
-
-@smallexample
-@c file eg/network/mobag.awk
-function MyJob() @{
- # forget this host
- sub(MOBVAR["Destination"], "", MOBVAR["Machines"])
- MOBVAR["Result"]=MOBVAR["Result"] SUBSEP SUBSEP MOBVAR["Destination"] ":"
- while (("who" | getline) > 0) # who is logged in?
- MOBVAR["Result"] = MOBVAR["Result"] SUBSEP $0
- close("who")
- if (index(MOBVAR["Machines"], "/") > 0) @{ # any more machines to visit?
- split(MOBVAR["Machines"], Machines) # which host is next?
- migrate(Machines[1], "", "") # go there
- @} else @{ # no more machines
- gsub(SUBSEP, "\n", MOBVAR["Result"]) # send result to origin
- print MOBVAR["Result"] |& "/inet/tcp/0/" MOBVAR["MyOrigin"] "/8080"
- close("/inet/tcp/0/" MOBVAR["MyOrigin"] "/8080")
- @}
-@}
-@c endfile
-@end smallexample
-
-After migrating, the first thing to do in @code{MyJob} is to delete
-the name of the current host from the list of hosts to visit. Now, it
-is time to start the real work by appending the host's name to the
-result string, and reading line by line who is logged in on this host.
-A very annoying circumstance is the fact that the elements of
-@code{MOBVAR} cannot hold the newline character (@code{"\n"}). If they
-did, migration of this string did not work because the string didn't
-obey the syntax rule for a string in @command{gawk}.
-@code{SUBSEP} is used as a temporary replacement.
-If the list of hosts to visit holds
-at least one more entry, the agent migrates to that place to go on
-working there. Otherwise, we replace the @code{SUBSEP}s
-with a newline character in the resulting string, and report it to
-the originating host, whose name is stored in @code{MOBVAR["MyOrigin"]}.
-
-@node STOXPRED, PROTBASE, MOBAGWHO, Some Applications and Techniques
-@section STOXPRED: Stock Market Prediction As A Service
-@cindex STOXPRED program
-@cindex Yahoo
-@quotation
-@i{Far out in the uncharted backwaters of the unfashionable end of
-the Western Spiral arm of the Galaxy lies a small unregarded yellow sun.}
-
-@i{Orbiting this at a distance of roughly ninety-two million miles is an
-utterly insignificant little blue-green planet whose ape-descendent life
-forms are so amazingly primitive that they still think digital watches are
-a pretty neat idea.}
-
-@i{This planet has --- or rather had --- a problem, which was this:
-most of the people living on it were unhappy for pretty much of the time.
-Many solutions were suggested for this problem, but most of these were
-largely concerned with the movements of small green pieces of paper,
-which is odd because it wasn't the small green pieces of paper that
-were unhappy.} @*
-Douglas Adams, @cite{The Hitch Hiker's Guide to the Galaxy}
-@end quotation
-
-@cindex @command{cron}
-Valuable services on the Internet are usually @emph{not} implemented
-as mobile agents. There are much simpler ways of implementing services.
-All Unix systems provide, for example, the @command{cron} service.
-Unix system users can write a list of tasks to be done each day, each
-week, twice a day, or just once. The list is entered into a file named
-@file{crontab}. For example, to distribute a newsletter on a daily
-basis this way, use @command{cron} for calling a script each day early
-in the morning.
-
-@example
-# run at 8 am on weekdays, distribute the newsletter
-0 8 * * 1-5 $HOME/bin/daily.job >> $HOME/log/newsletter 2>&1
-@end example
-
-The script first looks for interesting information on the Internet,
-assembles it in a nice form and sends the results via email to
-the customers.
-
-The following is an example of a primitive
-newsletter on stock market prediction. It is a report which first
-tries to predict the change of each share in the Dow Jones Industrial
-Index for the particular day. Then it mentions some especially
-promising shares as well as some shares which look remarkably bad
-on that day. The report ends with the usual disclaimer which tells
-every child @emph{not} to try this at home and hurt anybody.
-@cindex Dow Jones Industrial Index
-
-@smallexample
-Good morning Uncle Scrooge,
-
-This is your daily stock market report for Monday, October 16, 2000.
-Here are the predictions for today:
-
- AA neutral
- GE up
- JNJ down
- MSFT neutral
- @dots{}
- UTX up
- DD down
- IBM up
- MO down
- WMT up
- DIS up
- INTC up
- MRK down
- XOM down
- EK down
- IP down
-
-The most promising shares for today are these:
-
- INTC http://biz.yahoo.com/n/i/intc.html
-
-The stock shares to avoid today are these:
-
- EK http://biz.yahoo.com/n/e/ek.html
- IP http://biz.yahoo.com/n/i/ip.html
- DD http://biz.yahoo.com/n/d/dd.html
- @dots{}
-@end smallexample
-
-@ignore
-@c Chuck suggests removing this paragraph
-If you are not into stock market prediction but want to earn money
-with a more humane service, you might prefer to send out horoscopes
-to your customers. Or, once every refrigerator in every household on this side
-of the Chinese Wall is connected to the Internet, such a service could
-inspect the contents of your customer's refrigerators each day and
-advise them on nutrition. Big Brother is watching them.
-@end ignore
-
-The script as a whole is rather long. In order to ease the pain of
-studying other people's source code, we have broken the script
-up into meaningful parts which are invoked one after the other.
-The basic structure of the script is as follows:
-
-@example
-@c file eg/network/stoxpred.awk
-BEGIN @{
- Init()
- ReadQuotes()
- CleanUp()
- Prediction()
- Report()
- SendMail()
-@}
-@c endfile
-@end example
-
-The earlier parts store data into variables and arrays which are
-subsequently used by later parts of the script. The @code{Init} function
-first checks if the script is invoked correctly (without any parameters).
-If not, it informs the user of the correct usage. What follows are preparations
-for the retrieval of the historical quote data. The names of the 30 stock
-shares are stored in an array @code{name} along with the current date
-in @code{day}, @code{month}, and @code{year}.
-
-All users who are separated
-from the Internet by a firewall and have to direct their Internet accesses
-to a proxy must supply the name of the proxy to this script with the
-@samp{-v Proxy=@var{name}} option. For most users, the default proxy and
-port number should suffice.
-
-@example
-@c file eg/network/stoxpred.awk
-function Init() @{
- if (ARGC != 1) @{
- print "STOXPRED - daily stock share prediction"
- print "IN:\n no parameters, nothing on stdin"
- print "PARAM:\n -v Proxy=MyProxy -v ProxyPort=80"
- print "OUT:\n commented predictions as email"
- print "JK 09.10.2000"
- exit
- @}
- # Remember ticker symbols from Dow Jones Industrial Index
- StockCount = split("AA GE JNJ MSFT AXP GM JPM PG BA HD KO \
- SBC C HON MCD T CAT HWP MMM UTX DD IBM MO WMT DIS INTC \
- MRK XOM EK IP", name);
- # Remember the current date as the end of the time series
- day = strftime("%d")
- month = strftime("%m")
- year = strftime("%Y")
- if (Proxy == "") Proxy = "chart.yahoo.com"
- if (ProxyPort == 0) ProxyPort = 80
- YahooData = "/inet/tcp/0/" Proxy "/" ProxyPort
-@}
-@c endfile
-@end example
-
-@cindex CSV format
-There are two really interesting parts in the script. One is the
-function which reads the historical stock quotes from an Internet
-server. The other is the one that does the actual prediction. In
-the following function we see how the quotes are read from the
-Yahoo server. The data which comes from the server is in
-CSV format (comma-separated values):
-
-@example
-@c file eg/network/stoxdata.txt
-Date,Open,High,Low,Close,Volume
-9-Oct-00,22.75,22.75,21.375,22.375,7888500
-6-Oct-00,23.8125,24.9375,21.5625,22,10701100
-5-Oct-00,24.4375,24.625,23.125,23.50,5810300
-@c endfile
-@end example
-
-Lines contain values of the same time instant, whereas columns are
-separated by commas and contain the kind of data that is described
-in the header (first) line. At first, @command{gawk} is instructed to
-separate columns by commas (@samp{FS = ","}). In the loop that follows,
-a connection to the Yahoo server is first opened, then a download takes
-place, and finally the connection is closed. All this happens once for
-each ticker symbol. In the body of this loop, an Internet address is
-built up as a string according to the rules of the Yahoo server. The
-starting and ending date are chosen to be exactly the same, but one year
-apart in the past. All the action is initiated within the @code{printf}
-command which transmits the request for data to the Yahoo server.
-
-In the inner loop, the server's data is first read and then scanned
-line by line. Only lines which have six columns and the name of a month
-in the first column contain relevant data. This data is stored
-in the two-dimensional array @code{quote}; one dimension
-being time, the other being the ticker symbol. During retrieval of the
-first stock's data, the calendar names of the time instances are stored
-in the array @code{day} because we need them later.
-
-@smallexample
-@c file eg/network/stoxpred.awk
-function ReadQuotes() @{
- # Retrieve historical data for each ticker symbol
- FS = ","
- for (stock = 1; stock <= StockCount; stock++) @{
- URL = "http://chart.yahoo.com/table.csv?s=" name[stock] \
- "&a=" month "&b=" day "&c=" year-1 \
- "&d=" month "&e=" day "&f=" year \
- "g=d&q=q&y=0&z=" name[stock] "&x=.csv"
- printf("GET " URL " HTTP/1.0\r\n\r\n") |& YahooData
- while ((YahooData |& getline) > 0) @{
- if (NF == 6 && $1 ~ /Jan|Feb|Mar|Apr|May|Jun|Jul|Aug|Sep|Oct|Nov|Dec/) @{
- if (stock == 1)
- days[++daycount] = $1;
- quote[$1, stock] = $5
- @}
- @}
- close(YahooData)
- @}
- FS = " "
-@}
-@c endfile
-@end smallexample
-
-Now that we @emph{have} the data, it can be checked once again to make sure
-that no individual stock is missing or invalid, and that all the stock quotes are
-aligned correctly. Furthermore, we renumber the time instances. The
-most recent day gets day number 1 and all other days get consecutive
-numbers. All quotes are rounded toward the nearest whole number in US Dollars.
-
-@smallexample
-@c file eg/network/stoxpred.awk
-function CleanUp() @{
- # clean up time series; eliminate incomplete data sets
- for (d = 1; d <= daycount; d++) @{
- for (stock = 1; stock <= StockCount; stock++)
- if (! ((days[d], stock) in quote))
- stock = StockCount + 10
- if (stock > StockCount + 1)
- continue
- datacount++
- for (stock = 1; stock <= StockCount; stock++)
- data[datacount, stock] = int(0.5 + quote[days[d], stock])
- @}
- delete quote
- delete days
-@}
-@c endfile
-@end smallexample
-
-Now we have arrived at the second really interesting part of the whole affair.
-What we present here is a very primitive prediction algorithm:
-@emph{If a stock fell yesterday, assume it will also fall today; if
-it rose yesterday, assume it will rise today}. (Feel free to replace this
-algorithm with a smarter one.) If a stock changed in the same direction
-on two consecutive days, this is an indication which should be highlighted.
-Two-day advances are stored in @code{hot} and two-day declines in
-@code{avoid}.
-
-The rest of the function is a sanity check. It counts the number of
-correct predictions in relation to the total number of predictions
-one could have made in the year before.
-
-@smallexample
-@c file eg/network/stoxpred.awk
-function Prediction() @{
- # Predict each ticker symbol by prolonging yesterday's trend
- for (stock = 1; stock <= StockCount; stock++) @{
- if (data[1, stock] > data[2, stock]) @{
- predict[stock] = "up"
- @} else if (data[1, stock] < data[2, stock]) @{
- predict[stock] = "down"
- @} else @{
- predict[stock] = "neutral"
- @}
- if ((data[1, stock] > data[2, stock]) && (data[2, stock] > data[3, stock]))
- hot[stock] = 1
- if ((data[1, stock] < data[2, stock]) && (data[2, stock] < data[3, stock]))
- avoid[stock] = 1
- @}
- # Do a plausibility check: how many predictions proved correct?
- for (s = 1; s <= StockCount; s++) @{
- for (d = 1; d <= datacount-2; d++) @{
- if (data[d+1, s] > data[d+2, s]) @{
- UpCount++
- @} else if (data[d+1, s] < data[d+2, s]) @{
- DownCount++
- @} else @{
- NeutralCount++
- @}
- if (((data[d, s] > data[d+1, s]) && (data[d+1, s] > data[d+2, s])) ||
- ((data[d, s] < data[d+1, s]) && (data[d+1, s] < data[d+2, s])) ||
- ((data[d, s] == data[d+1, s]) && (data[d+1, s] == data[d+2, s])))
- CorrectCount++
- @}
- @}
-@}
-@c endfile
-@end smallexample
-
-At this point the hard work has been done: the array @code{predict}
-contains the predictions for all the ticker symbols. It is up to the
-function @code{Report} to find some nice words to introduce the
-desired information.
-
-@smallexample
-@c file eg/network/stoxpred.awk
-function Report() @{
- # Generate report
- report = "\nThis is your daily "
- report = report "stock market report for "strftime("%A, %B %d, %Y")".\n"
- report = report "Here are the predictions for today:\n\n"
- for (stock = 1; stock <= StockCount; stock++)
- report = report "\t" name[stock] "\t" predict[stock] "\n"
- for (stock in hot) @{
- if (HotCount++ == 0)
- report = report "\nThe most promising shares for today are these:\n\n"
- report = report "\t" name[stock] "\t\thttp://biz.yahoo.com/n/" \
- tolower(substr(name[stock], 1, 1)) "/" tolower(name[stock]) ".html\n"
- @}
- for (stock in avoid) @{
- if (AvoidCount++ == 0)
- report = report "\nThe stock shares to avoid today are these:\n\n"
- report = report "\t" name[stock] "\t\thttp://biz.yahoo.com/n/" \
- tolower(substr(name[stock], 1, 1)) "/" tolower(name[stock]) ".html\n"
- @}
- report = report "\nThis sums up to " HotCount+0 " winners and " AvoidCount+0
- report = report " losers. When using this kind\nof prediction scheme for"
- report = report " the 12 months which lie behind us,\nwe get " UpCount
- report = report " 'ups' and " DownCount " 'downs' and " NeutralCount
- report = report " 'neutrals'. Of all\nthese " UpCount+DownCount+NeutralCount
- report = report " predictions " CorrectCount " proved correct next day.\n"
- report = report "A success rate of "\
- int(100*CorrectCount/(UpCount+DownCount+NeutralCount)) "%.\n"
- report = report "Random choice would have produced a 33% success rate.\n"
- report = report "Disclaimer: Like every other prediction of the stock\n"
- report = report "market, this report is, of course, complete nonsense.\n"
- report = report "If you are stupid enough to believe these predictions\n"
- report = report "you should visit a doctor who can treat your ailment."
-@}
-@c endfile
-@end smallexample
-
-The function @code{SendMail} goes through the list of customers and opens
-a pipe to the @code{mail} command for each of them. Each one receives an
-email message with a proper subject heading and is addressed with his full name.
-
-@smallexample
-@c file eg/network/stoxpred.awk
-function SendMail() @{
- # send report to customers
- customer["uncle.scrooge@@ducktown.gov"] = "Uncle Scrooge"
- customer["more@@utopia.org" ] = "Sir Thomas More"
- customer["spinoza@@denhaag.nl" ] = "Baruch de Spinoza"
- customer["marx@@highgate.uk" ] = "Karl Marx"
- customer["keynes@@the.long.run" ] = "John Maynard Keynes"
- customer["bierce@@devil.hell.org" ] = "Ambrose Bierce"
- customer["laplace@@paris.fr" ] = "Pierre Simon de Laplace"
- for (c in customer) @{
- MailPipe = "mail -s 'Daily Stock Prediction Newsletter'" c
- print "Good morning " customer[c] "," | MailPipe
- print report "\n.\n" | MailPipe
- close(MailPipe)
- @}
-@}
-@c endfile
-@end smallexample
-
-Be patient when running the script by hand.
-Retrieving the data for all the ticker symbols and sending the emails
-may take several minutes to complete, depending upon network traffic
-and the speed of the available Internet link.
-The quality of the prediction algorithm is likely to be disappointing.
-Try to find a better one.
-Should you find one with a success rate of more than 50%, please tell
-us about it! It is only for the sake of curiosity, of course. @code{:-)}
-
-@ignore
-@c chuck says to remove this
-Let us give you one final indication as to what one can expect from
-a prediction of stock data, which is sometimes said to contain much
-randomness. One theory says that all relevant information to be taken
-into account when estimating the price of a stock is contained in the
-stock quotes. Every bit of useful information has influenced the
-fair price. Therefore (the theory says) temporary changes (i.e., fluctuations
-within a minute) have to be purely random. But what is the cause of
-short-term changes in stock prices?
-
-Stock prices are fixed when supply and demand meet each other.
-What people are willing to pay reflects human expectations.
-Human expectations are not necessarily random. On the Internet,
-you can find an elucidating paper about predictability and human
-expectations:
-@uref{http://it.ucsd.edu/IT/Newsletter/archives/meir/05meir.html,
-@cite{Reflections on ``Universal Prediction of Individual Sequences''}}
-The authors (Feder, Merhav, Gutman) introduce the reader to the subject
-by telling a thrilling anecdote.
-@cindex Shannon, Claude
-@quotation
-In the early 50's, at Bell Laboratories, David Hagelbarger built a
-simple ``mind reading'' machine, whose purpose was to play the ``penny
-matching'' game. In this game, a player chooses head or tail, while a
-``mind reading'' machine tries to predict and match his choice.
-Surprisingly, as Robert Lucky tells in his book ``Silicon Dreams'',
-Hagelbarger's simple, 8-state machine, was able to match the ``pennies''
-of its human opponent 5,218 times over the course of 9,795 plays.
-Random guessing would lead to such a high success rate with a probability
-less than one out of 10 billion! Shannon, who was interested in prediction,
-information, and thinking machines, closely followed Hagelbarger's
-machine, and eventually built his own stripped-down version of the machine,
-having the same states, but one that used a simpler strategy at each state.
-As the legend goes, in a duel between the two machines, Shannon's machine
-won by a slight margin! No one knows if this was due to a superior algorithm
-or just a chance happening associated with the specific sequence at that game.
-In any event, the success of both these machines against ``untrained'' human
-opponents was explained by the fact that the human opponents cannot draw
-completely random
-bits.
-@end quotation
-@end ignore
-
-@node PROTBASE, , STOXPRED, Some Applications and Techniques
-@section PROTBASE: Searching Through A Protein Database
-@cindex PROTBASE
-@cindex NCBI, National Center for Biotechnology Information
-@cindex BLAST, Basic Local Alignment Search Tool
-@cindex Hoare, C.A.R.
-@quotation
-@i{Hoare's Law of Large Problems: Inside every large problem is a small
- problem struggling to get out.}
-@end quotation
-
-Yahoo's database of stock market data is just one among the many large
-databases on the Internet. Another one is located at NCBI
-(National Center for Biotechnology
-Information). Established in 1988 as a national resource for molecular
-biology information, NCBI creates public databases, conducts research
-in computational biology, develops software tools for analyzing genome
-data, and disseminates biomedical information. In this section, we
-look at one of NCBI's public services, which is called BLAST
-(Basic Local Alignment Search Tool).
-
-You probably know that the information necessary for reproducing living
-cells is encoded in the genetic material of the cells. The genetic material
-is a very long chain of four base nucleotides. It is the order of
-appearance (the sequence) of nucleotides which contains the information
-about the substance to be produced. Scientists in biotechnology often
-find a specific fragment, determine the nucleotide sequence, and need
-to know where the sequence at hand comes from. This is where the large
-databases enter the game. At NCBI, databases store the knowledge
-about which sequences have ever been found and where they have been found.
-When the scientist sends his sequence to the BLAST service, the server
-looks for regions of genetic material in its database which
-look the most similar to the delivered nucleotide sequence. After a
-search time of some seconds or minutes the server sends an answer to
-the scientist. In order to make access simple, NCBI chose to offer
-their database service through popular Internet protocols. There are
-four basic ways to use the so-called BLAST services:
-
-@itemize @bullet
-@item
-The easiest way to use BLAST is through the web. Users may simply point
-their browsers at the NCBI home page
-and link to the BLAST pages.
-NCBI provides a stable URL that may be used to perform BLAST searches
-without interactive use of a web browser. This is what we will do later
-in this section.
-A demonstration client
-and a @file{README} file demonstrate how to access this URL.
-
-@item
-Currently,
-@command{blastcl3} is the standard network BLAST client.
-You can download @command{blastcl3} from the
-anonymous FTP location.
-
-@item
-BLAST 2.0 can be run locally as a full executable and can be used to run
-BLAST searches against private local databases, or downloaded copies of the
-NCBI databases. BLAST 2.0 executables may be found on the NCBI
-anonymous FTP server.
-
-@item
-The NCBI BLAST Email server is the best option for people without convenient
-access to the web. A similarity search can be performed by sending a properly
-formatted mail message containing the nucleotide or protein query sequence to
-@email{blast@@ncbi.nlm.nih.gov}. The query sequence is compared against the
-specified database using the BLAST algorithm and the results are returned in
-an email message. For more information on formulating email BLAST searches,
-you can send a message consisting of the word ``HELP'' to the same address,
-@email{blast@@ncbi.nlm.nih.gov}.
-@end itemize
-
-Our starting point is the demonstration client mentioned in the first option.
-The @file{README} file that comes along with the client explains the whole
-process in a nutshell. In the rest of this section, we first show
-what such requests look like. Then we show how to use @command{gawk} to
-implement a client in about 10 lines of code. Finally, we show how to
-interpret the result returned from the service.
-
-Sequences are expected to be represented in the standard
-IUB/IUPAC amino acid and nucleic acid codes,
-with these exceptions: lower-case letters are accepted and are mapped
-into upper-case; a single hyphen or dash can be used to represent a gap
-of indeterminate length; and in amino acid sequences, @samp{U} and @samp{*}
-are acceptable letters (see below). Before submitting a request, any numerical
-digits in the query sequence should either be removed or replaced by
-appropriate letter codes (e.g., @samp{N} for unknown nucleic acid residue
-or @samp{X} for unknown amino acid residue).
-The nucleic acid codes supported are:
-
-@example
-A --> adenosine M --> A C (amino)
-C --> cytidine S --> G C (strong)
-G --> guanine W --> A T (weak)
-T --> thymidine B --> G T C
-U --> uridine D --> G A T
-R --> G A (purine) H --> A C T
-Y --> T C (pyrimidine) V --> G C A
-K --> G T (keto) N --> A G C T (any)
- - gap of indeterminate length
-@end example
-
-Now you know the alphabet of nucleotide sequences. The last two lines
-of the following example query show you such a sequence, which is obviously
-made up only of elements of the alphabet just described. Store this example
-query into a file named @file{protbase.request}. You are now ready to send
-it to the server with the demonstration client.
-
-@example
-@c file eg/network/protbase.request
-PROGRAM blastn
-DATALIB month
-EXPECT 0.75
-BEGIN
->GAWK310 the gawking gene GNU AWK
-tgcttggctgaggagccataggacgagagcttcctggtgaagtgtgtttcttgaaatcat
-caccaccatggacagcaaa
-@c endfile
-@end example
-
-@cindex FASTA/Pearson format
-The actual search request begins with the mandatory parameter @samp{PROGRAM}
-in the first column followed by the value @samp{blastn} (the name of the
-program) for searching nucleic acids. The next line contains the mandatory
-search parameter @samp{DATALIB} with the value @samp{month} for the newest
-nucleic acid sequences. The third line contains an optional @samp{EXPECT}
-parameter and the value desired for it. The fourth line contains the
-mandatory @samp{BEGIN} directive, followed by the query sequence in
-FASTA/Pearson format.
-Each line of information must be less than 80 characters in length.
-
-The ``month'' database contains all new or revised sequences released in the
-last 30 days and is useful for searching against new sequences.
-There are five different blast programs, @command{blastn} being the one that
-compares a nucleotide query sequence against a nucleotide sequence database.
-
-The last server directive that must appear in every request is the
-@samp{BEGIN} directive. The query sequence should immediately follow the
-@samp{BEGIN} directive and must appear in FASTA/Pearson format.
-A sequence in
-FASTA/Pearson format begins with a single-line description.
-The description line, which is required, is distinguished from the lines of
-sequence data that follow it by having a greater-than (@samp{>}) symbol
-in the first column. For the purposes of the BLAST server, the text of
-the description is arbitrary.
-
-If you prefer to use a client written in @command{gawk}, just store the following
-10 lines of code into a file named @file{protbase.awk} and use this client
-instead. Invoke it with @samp{gawk -f protbase.awk protbase.request}.
-Then wait a minute and watch the result coming in. In order to replicate
-the demonstration client's behaviour as closely as possible, this client
-does not use a proxy server. We could also have extended the client program
-in @ref{GETURL, ,Retrieving Web Pages}, to implement the client request from
-@file{protbase.awk} as a special case.
-
-@smallexample
-@c file eg/network/protbase.awk
-@{ request = request "\n" $0 @}
-
-END @{
- BLASTService = "/inet/tcp/0/www.ncbi.nlm.nih.gov/80"
- printf "POST /cgi-bin/BLAST/nph-blast_report HTTP/1.0\n" |& BLASTService
- printf "Content-Length: " length(request) "\n\n" |& BLASTService
- printf request |& BLASTService
- while ((BLASTService |& getline) > 0)
- print $0
- close(BLASTService)
-@}
-@c endfile
-@end smallexample
-
-The demonstration client from NCBI is 214 lines long (written in C) and
-it is not immediately obvious what it does. Our client is so short that
-it @emph{is} obvious what it does. First it loops over all lines of the
-query and stores the whole query into a variable. Then the script
-establishes an Internet connection to the NCBI server and transmits the
-query by framing it with a proper HTTP request. Finally it receives
-and prints the complete result coming from the server.
-
-Now, let us look at the result. It begins with an HTTP header, which you
-can ignore. Then there are some comments about the query having been
-filtered to avoid spuriously high scores. After this, there is a reference
-to the paper that describes the software being used for searching the data
-base. After a repitition of the original query's description we find the
-list of significant alignments:
-
-@smallexample
-@c file eg/network/protbase.result
-Sequences producing significant alignments: (bits) Value
-
-gb|AC021182.14|AC021182 Homo sapiens chromosome 7 clone RP11-733... 38 0.20
-gb|AC021056.12|AC021056 Homo sapiens chromosome 3 clone RP11-115... 38 0.20
-emb|AL160278.10|AL160278 Homo sapiens chromosome 9 clone RP11-57... 38 0.20
-emb|AL391139.11|AL391139 Homo sapiens chromosome X clone RP11-35... 38 0.20
-emb|AL365192.6|AL365192 Homo sapiens chromosome 6 clone RP3-421H... 38 0.20
-emb|AL138812.9|AL138812 Homo sapiens chromosome 11 clone RP1-276... 38 0.20
-gb|AC073881.3|AC073881 Homo sapiens chromosome 15 clone CTD-2169... 38 0.20
-@c endfile
-@end smallexample
-
-This means that the query sequence was found in seven human chromosomes.
-But the value 0.20 (20%) means that the probability of an accidental match
-is rather high (20%) in all cases and should be taken into account.
-You may wonder what the first column means. It is a key to the specific
-database in which this occurence was found. The unique sequence identifiers
-reported in the search results can be used as sequence retrieval keys
-via the NCBI server. The syntax of sequence header lines used by the NCBI
-BLAST server depends on the database from which each sequence was obtained.
-The table below lists the identifiers for the databases from which the
-sequences were derived.
-
-@ifinfo
-@example
-Database Name Identifier Syntax
-============================ ========================
-GenBank gb|accession|locus
-EMBL Data Library emb|accession|locus
-DDBJ, DNA Database of Japan dbj|accession|locus
-NBRF PIR pir||entry
-Protein Research Foundation prf||name
-SWISS-PROT sp|accession|entry name
-Brookhaven Protein Data Bank pdb|entry|chain
-Kabat's Sequences of Immuno@dots{} gnl|kabat|identifier
-Patents pat|country|number
-GenInfo Backbone Id bbs|number
-@end example
-@end ifinfo
-
-@ifnotinfo
-@multitable {Kabat's Sequences of Immuno@dots{}} {@code{@w{sp|accession|entry name}}}
-@item GenBank @tab @code{gb|accession|locus}
-@item EMBL Data Library @tab @code{emb|accession|locus}
-@item DDBJ, DNA Database of Japan @tab @code{dbj|accession|locus}
-@item NBRF PIR @tab @code{pir||entry}
-@item Protein Research Foundation @tab @code{prf||name}
-@item SWISS-PROT @tab @code{@w{sp|accession|entry name}}
-@item Brookhaven Protein Data Bank @tab @code{pdb|entry|chain}
-@item Kabat's Sequences of Immuno@dots{} @tab @code{gnl|kabat|identifier}
-@item Patents @tab @code{pat|country|number}
-@item GenInfo Backbone Id @tab @code{bbs|number}
-@end multitable
-@end ifnotinfo
-
-
-For example, an identifier might be @samp{gb|AC021182.14|AC021182}, where the
-@samp{gb} tag indicates that the identifier refers to a GenBank sequence,
-@samp{AC021182.14} is its GenBank ACCESSION, and @samp{AC021182} is the GenBank LOCUS.
-The identifier contains no spaces, so that a space indicates the end of the
-identifier.
-
-Let us continue in the result listing. Each of the seven alignments mentioned
-above is subsequently described in detail. We will have a closer look at
-the first of them.
-
-@smallexample
->gb|AC021182.14|AC021182 Homo sapiens chromosome 7 clone RP11-733N23, WORKING DRAFT SEQUENCE, 4
- unordered pieces
- Length = 176383
-
- Score = 38.2 bits (19), Expect = 0.20
- Identities = 19/19 (100%)
- Strand = Plus / Plus
-
-Query: 35 tggtgaagtgtgtttcttg 53
- |||||||||||||||||||
-Sbjct: 69786 tggtgaagtgtgtttcttg 69804
-@end smallexample
-
-This alignment was located on the human chromosome 7. The fragment on which
-part of the query was found had a total length of 176383. Only 19 of the
-nucleotides matched and the matching sequence ran from character 35 to 53
-in the query sequence and from 69786 to 69804 in the fragment on chromosome 7.
-If you are still reading at this point, you are probably interested in finding
-out more about Computational Biology and you might appreciate the following
-hints.
-
-@cindex Computational Biology
-@cindex Bioinformatics
-@enumerate
-@item
-There is a book called @cite{Introduction to Computational Biology}
-by Michael S. Waterman, which is worth reading if you are seriously
-interested. You can find a good
-book review
-on the Internet.
-
-@item
-While Waterman's book can explain to you the algorithms employed internally
-in the database search engines, most practicioners prefer to approach
-the subject differently. The applied side of Computational Biology is
-called Bioinformatics, and emphasizes the tools available for day-to-day
-work as well as how to actually @emph{use} them. One of the very few affordable
-books on Bioinformatics is
-@cite{Developing Bioinformatics Computer Skills}.
-
-@item
-The sequences @emph{gawk} and @emph{gnuawk} are in widespread use in
-the genetic material of virtually every earthly living being. Let us
-take this as a clear indication that the divine creator has intended
-@code{gawk} to prevail over other scripting languages such as @code{perl},
-@code{tcl}, or @code{python} which are not even proper sequences. (:-)
-@end enumerate
-
-@node Links, GNU Free Documentation License, Some Applications and Techniques, Top
-@chapter Related Links
-
-This section lists the URLs for various items discussed in this @value{CHAPTER}.
-They are presented in the order in which they appear.
-
-@table @asis
-
-@item @cite{Internet Programming with Python}
-@uref{http://www.fsbassociates.com/books/python.htm}
-
-@item @cite{Advanced Perl Programming}
-@uref{http://www.oreilly.com/catalog/advperl}
-
-@item @cite{Web Client Programming with Perl}
-@uref{http://www.oreilly.com/catalog/webclient}
-
-@item Richard Stevens's home page and book
-@uref{http://www.kohala.com/~rstevens}
-
-@item The SPAK home page
-@uref{http://www.userfriendly.net/linux/RPM/contrib/libc6/i386/spak-0.6b-1.i386.html}
-
-@item Volume III of @cite{Internetworking with TCP/IP}, by Comer and Stevens
-@uref{http://www.cs.purdue.edu/homes/dec/tcpip3s.cont.html}
-
-@item XBM Graphics File Format
-@uref{http://www.wotsit.org/download.asp?f=xbm}
-
-@item GNUPlot
-@uref{http://www.cs.dartmouth.edu/gnuplot_info.html}
-
-@item Mark Humphrys' Eliza page
-@uref{http://www.compapp.dcu.ie/~humphrys/eliza.html}
-
-@item Yahoo! Eliza Information
-@uref{http://dir.yahoo.com/Recreation/Games/Computer_Games/Internet_Games/Web_Games/Artificial_Intelligence}
-
-@item Java versions of Eliza
-@uref{http://www.tjhsst.edu/Psych/ch1/eliza.html}
-
-@item Java versions of Eliza with source code
-@uref{http://home.adelphia.net/~lifeisgood/eliza/eliza.htm}
-
-@item Eliza Programs with Explanations
-@uref{http://chayden.net/chayden/eliza/Eliza.shtml}
-
-@item Loebner Contest
-@uref{http://acm.org/~loebner/loebner-prize.htmlx}
-
-@item Tck/Tk Information
-@uref{http://www.scriptics.com/}
-
-@item Intel 80x86 Processors
-@uref{http://developer.intel.com/design/platform/embedpc/what_is.htm}
-
-@item AMD Elan Processors
-@uref{http://www.amd.com/products/epd/processors/4.32bitcont/32bitcont/index.html}
-
-@item XINU
-@uref{http://willow.canberra.edu.au/~chrisc/xinu.html }
-
-@item GNU/Linux
-@uref{http://uclinux.lineo.com/}
-
-@item Embedded PCs
-@uref{http://dir.yahoo.com/Business_and_Economy/Business_to_Business/Computers/Hardware/Embedded_Control/}
-
-@item MiniSQL
-@uref{http://www.hughes.com.au/library/}
-
-@item Market Share Surveys
-@uref{http://www.netcraft.com/survey}
-
-@item @cite{Numerical Recipes in C: The Art of Scientific Computing}
-@uref{http://www.nr.com}
-
-@item VRML
-@uref{http://www.vrml.org}
-
-@item The VRML FAQ
-@uref{http://www.vrml.org/technicalinfo/specifications/specifications.htm#FAQ}
-
-@item The UMBC Agent Web
-@uref{http://www.cs.umbc.edu/agents }
-
-@item Apache Web Server
-@uref{http://www.apache.org}
-
-@item National Center for Biotechnology Information (NCBI)
-@uref{http://www.ncbi.nlm.nih.gov}
-
-@item Basic Local Alignment Search Tool (BLAST)
-@uref{http://www.ncbi.nlm.nih.gov/BLAST/blast_overview.html}
-
-@item NCBI Home Page
-@uref{http://www.ncbi.nlm.nih.gov}
-
-@item BLAST Pages
-@uref{http://www.ncbi.nlm.nih.gov/BLAST}
-
-@item BLAST Demonstration Client
-@uref{ftp://ncbi.nlm.nih.gov/blast/blasturl/}
-
-@item BLAST anonymous FTP location
-@uref{ftp://ncbi.nlm.nih.gov/blast/network/netblast/}
-
-@item BLAST 2.0 Executables
-@uref{ftp://ncbi.nlm.nih.gov/blast/executables/}
-
-@item IUB/IUPAC Amino Acid and Nucleic Acid Codes
-@uref{http://www.uthscsa.edu/geninfo/blastmail.html#item6}
-
-@item FASTA/Pearson Format
-@uref{http://www.ncbi.nlm.nih.gov/BLAST/fasta.html}
-
-@item Fasta/Pearson Sequence in Java
-@uref{http://www.kazusa.or.jp/java/codon_table_java/}
-
-@item Book Review of @cite{Introduction to Computational Biology}
-@uref{http://www.acm.org/crossroads/xrds5-1/introcb.html}
-
-@item @cite{Developing Bioinformatics Computer Skills}
-@uref{http://www.oreilly.com/catalog/bioskills/}
-
-@end table
-
-@node GNU Free Documentation License, Index, Links, Top
-@unnumbered GNU Free Documentation License
-@center Version 1.1, March 2000
-
-@display
-Copyright (C) 2000 Free Software Foundation, Inc.
-59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
-
-Everyone is permitted to copy and distribute verbatim copies
-of this license document, but changing it is not allowed.
-@end display
-@sp 1
-@enumerate 0
-@item
-PREAMBLE
-
-The purpose of this License is to make a manual, textbook, or other
-written document ``free'' in the sense of freedom: to assure everyone
-the effective freedom to copy and redistribute it, with or without
-modifying it, either commercially or noncommercially. Secondarily,
-this License preserves for the author and publisher a way to get
-credit for their work, while not being considered responsible for
-modifications made by others.
-
-This License is a kind of ``copyleft'', which means that derivative
-works of the document must themselves be free in the same sense. It
-complements the GNU General Public License, which is a copyleft
-license designed for free software.
-
-We have designed this License in order to use it for manuals for free
-software, because free software needs free documentation: a free
-program should come with manuals providing the same freedoms that the
-software does. But this License is not limited to software manuals;
-it can be used for any textual work, regardless of subject matter or
-whether it is published as a printed book. We recommend this License
-principally for works whose purpose is instruction or reference.
-
-@sp 1
-@item
-APPLICABILITY AND DEFINITIONS
-
-This License applies to any manual or other work that contains a
-notice placed by the copyright holder saying it can be distributed
-under the terms of this License. The ``Document'', below, refers to any
-such manual or work. Any member of the public is a licensee, and is
-addressed as ``you''.
-
-A ``Modified Version'' of the Document means any work containing the
-Document or a portion of it, either copied verbatim, or with
-modifications and/or translated into another language.
-
-A ``Secondary Section'' is a named appendix or a front-matter section of
-the Document that deals exclusively with the relationship of the
-publishers or authors of the Document to the Document's overall subject
-(or to related matters) and contains nothing that could fall directly
-within that overall subject. (For example, if the Document is in part a
-textbook of mathematics, a Secondary Section may not explain any
-mathematics.) The relationship could be a matter of historical
-connection with the subject or with related matters, or of legal,
-commercial, philosophical, ethical or political position regarding
-them.
-
-The ``Invariant Sections'' are certain Secondary Sections whose titles
-are designated, as being those of Invariant Sections, in the notice
-that says that the Document is released under this License.
-
-The ``Cover Texts'' are certain short passages of text that are listed,
-as Front-Cover Texts or Back-Cover Texts, in the notice that says that
-the Document is released under this License.
-
-A ``Transparent'' copy of the Document means a machine-readable copy,
-represented in a format whose specification is available to the
-general public, whose contents can be viewed and edited directly and
-straightforwardly with generic text editors or (for images composed of
-pixels) generic paint programs or (for drawings) some widely available
-drawing editor, and that is suitable for input to text formatters or
-for automatic translation to a variety of formats suitable for input
-to text formatters. A copy made in an otherwise Transparent file
-format whose markup has been designed to thwart or discourage
-subsequent modification by readers is not Transparent. A copy that is
-not ``Transparent'' is called ``Opaque''.
-
-Examples of suitable formats for Transparent copies include plain
-ASCII without markup, Texinfo input format, LaTeX input format, SGML
-or XML using a publicly available DTD, and standard-conforming simple
-HTML designed for human modification. Opaque formats include
-PostScript, PDF, proprietary formats that can be read and edited only
-by proprietary word processors, SGML or XML for which the DTD and/or
-processing tools are not generally available, and the
-machine-generated HTML produced by some word processors for output
-purposes only.
-
-The ``Title Page'' means, for a printed book, the title page itself,
-plus such following pages as are needed to hold, legibly, the material
-this License requires to appear in the title page. For works in
-formats which do not have any title page as such, ``Title Page'' means
-the text near the most prominent appearance of the work's title,
-preceding the beginning of the body of the text.
-@sp 1
-@item
-VERBATIM COPYING
-
-You may copy and distribute the Document in any medium, either
-commercially or noncommercially, provided that this License, the
-copyright notices, and the license notice saying this License applies
-to the Document are reproduced in all copies, and that you add no other
-conditions whatsoever to those of this License. You may not use
-technical measures to obstruct or control the reading or further
-copying of the copies you make or distribute. However, you may accept
-compensation in exchange for copies. If you distribute a large enough
-number of copies you must also follow the conditions in section 3.
-
-You may also lend copies, under the same conditions stated above, and
-you may publicly display copies.
-@sp 1
-@item
-COPYING IN QUANTITY
-
-If you publish printed copies of the Document numbering more than 100,
-and the Document's license notice requires Cover Texts, you must enclose
-the copies in covers that carry, clearly and legibly, all these Cover
-Texts: Front-Cover Texts on the front cover, and Back-Cover Texts on
-the back cover. Both covers must also clearly and legibly identify
-you as the publisher of these copies. The front cover must present
-the full title with all words of the title equally prominent and
-visible. You may add other material on the covers in addition.
-Copying with changes limited to the covers, as long as they preserve
-the title of the Document and satisfy these conditions, can be treated
-as verbatim copying in other respects.
-
-If the required texts for either cover are too voluminous to fit
-legibly, you should put the first ones listed (as many as fit
-reasonably) on the actual cover, and continue the rest onto adjacent
-pages.
-
-If you publish or distribute Opaque copies of the Document numbering
-more than 100, you must either include a machine-readable Transparent
-copy along with each Opaque copy, or state in or with each Opaque copy
-a publicly-accessible computer-network location containing a complete
-Transparent copy of the Document, free of added material, which the
-general network-using public has access to download anonymously at no
-charge using public-standard network protocols. If you use the latter
-option, you must take reasonably prudent steps, when you begin
-distribution of Opaque copies in quantity, to ensure that this
-Transparent copy will remain thus accessible at the stated location
-until at least one year after the last time you distribute an Opaque
-copy (directly or through your agents or retailers) of that edition to
-the public.
-
-It is requested, but not required, that you contact the authors of the
-Document well before redistributing any large number of copies, to give
-them a chance to provide you with an updated version of the Document.
-@sp 1
-@item
-MODIFICATIONS
-
-You may copy and distribute a Modified Version of the Document under
-the conditions of sections 2 and 3 above, provided that you release
-the Modified Version under precisely this License, with the Modified
-Version filling the role of the Document, thus licensing distribution
-and modification of the Modified Version to whoever possesses a copy
-of it. In addition, you must do these things in the Modified Version:
-
-@enumerate A
-@item
-Use in the Title Page (and on the covers, if any) a title distinct
-from that of the Document, and from those of previous versions
-(which should, if there were any, be listed in the History section
-of the Document). You may use the same title as a previous version
-if the original publisher of that version gives permission.
-
-@item
-List on the Title Page, as authors, one or more persons or entities
-responsible for authorship of the modifications in the Modified
-Version, together with at least five of the principal authors of the
-Document (all of its principal authors, if it has less than five).
-
-@item
-State on the Title page the name of the publisher of the
-Modified Version, as the publisher.
-
-@item
-Preserve all the copyright notices of the Document.
-
-@item
-Add an appropriate copyright notice for your modifications
-adjacent to the other copyright notices.
-
-@item
-Include, immediately after the copyright notices, a license notice
-giving the public permission to use the Modified Version under the
-terms of this License, in the form shown in the Addendum below.
-
-@item
-Preserve in that license notice the full lists of Invariant Sections
-and required Cover Texts given in the Document's license notice.
-
-@item
-Include an unaltered copy of this License.
-
-@item
-Preserve the section entitled ``History'', and its title, and add to
-it an item stating at least the title, year, new authors, and
-publisher of the Modified Version as given on the Title Page. If
-there is no section entitled ``History'' in the Document, create one
-stating the title, year, authors, and publisher of the Document as
-given on its Title Page, then add an item describing the Modified
-Version as stated in the previous sentence.
-
-@item
-Preserve the network location, if any, given in the Document for
-public access to a Transparent copy of the Document, and likewise
-the network locations given in the Document for previous versions
-it was based on. These may be placed in the ``History'' section.
-You may omit a network location for a work that was published at
-least four years before the Document itself, or if the original
-publisher of the version it refers to gives permission.
-
-@item
-In any section entitled ``Acknowledgements'' or ``Dedications'',
-preserve the section's title, and preserve in the section all the
-substance and tone of each of the contributor acknowledgements
-and/or dedications given therein.
-
-@item
-Preserve all the Invariant Sections of the Document,
-unaltered in their text and in their titles. Section numbers
-or the equivalent are not considered part of the section titles.
-
-@item
-Delete any section entitled ``Endorsements''. Such a section
-may not be included in the Modified Version.
-
-@item
-Do not retitle any existing section as ``Endorsements''
-or to conflict in title with any Invariant Section.
-@end enumerate
-
-If the Modified Version includes new front-matter sections or
-appendices that qualify as Secondary Sections and contain no material
-copied from the Document, you may at your option designate some or all
-of these sections as invariant. To do this, add their titles to the
-list of Invariant Sections in the Modified Version's license notice.
-These titles must be distinct from any other section titles.
-
-You may add a section entitled ``Endorsements'', provided it contains
-nothing but endorsements of your Modified Version by various
-parties--for example, statements of peer review or that the text has
-been approved by an organization as the authoritative definition of a
-standard.
-
-You may add a passage of up to five words as a Front-Cover Text, and a
-passage of up to 25 words as a Back-Cover Text, to the end of the list
-of Cover Texts in the Modified Version. Only one passage of
-Front-Cover Text and one of Back-Cover Text may be added by (or
-through arrangements made by) any one entity. If the Document already
-includes a cover text for the same cover, previously added by you or
-by arrangement made by the same entity you are acting on behalf of,
-you may not add another; but you may replace the old one, on explicit
-permission from the previous publisher that added the old one.
-
-The author(s) and publisher(s) of the Document do not by this License
-give permission to use their names for publicity for or to assert or
-imply endorsement of any Modified Version.
-@sp 1
-@item
-COMBINING DOCUMENTS
-
-You may combine the Document with other documents released under this
-License, under the terms defined in section 4 above for modified
-versions, provided that you include in the combination all of the
-Invariant Sections of all of the original documents, unmodified, and
-list them all as Invariant Sections of your combined work in its
-license notice.
-
-The combined work need only contain one copy of this License, and
-multiple identical Invariant Sections may be replaced with a single
-copy. If there are multiple Invariant Sections with the same name but
-different contents, make the title of each such section unique by
-adding at the end of it, in parentheses, the name of the original
-author or publisher of that section if known, or else a unique number.
-Make the same adjustment to the section titles in the list of
-Invariant Sections in the license notice of the combined work.
-
-In the combination, you must combine any sections entitled ``History''
-in the various original documents, forming one section entitled
-``History''; likewise combine any sections entitled ``Acknowledgements'',
-and any sections entitled ``Dedications''. You must delete all sections
-entitled ``Endorsements.''
-@sp 1
-@item
-COLLECTIONS OF DOCUMENTS
-
-You may make a collection consisting of the Document and other documents
-released under this License, and replace the individual copies of this
-License in the various documents with a single copy that is included in
-the collection, provided that you follow the rules of this License for
-verbatim copying of each of the documents in all other respects.
-
-You may extract a single document from such a collection, and distribute
-it individually under this License, provided you insert a copy of this
-License into the extracted document, and follow this License in all
-other respects regarding verbatim copying of that document.
-@sp 1
-@item
-AGGREGATION WITH INDEPENDENT WORKS
-
-A compilation of the Document or its derivatives with other separate
-and independent documents or works, in or on a volume of a storage or
-distribution medium, does not as a whole count as a Modified Version
-of the Document, provided no compilation copyright is claimed for the
-compilation. Such a compilation is called an ``aggregate'', and this
-License does not apply to the other self-contained works thus compiled
-with the Document, on account of their being thus compiled, if they
-are not themselves derivative works of the Document.
-
-If the Cover Text requirement of section 3 is applicable to these
-copies of the Document, then if the Document is less than one quarter
-of the entire aggregate, the Document's Cover Texts may be placed on
-covers that surround only the Document within the aggregate.
-Otherwise they must appear on covers around the whole aggregate.
-@sp 1
-@item
-TRANSLATION
-
-Translation is considered a kind of modification, so you may
-distribute translations of the Document under the terms of section 4.
-Replacing Invariant Sections with translations requires special
-permission from their copyright holders, but you may include
-translations of some or all Invariant Sections in addition to the
-original versions of these Invariant Sections. You may include a
-translation of this License provided that you also include the
-original English version of this License. In case of a disagreement
-between the translation and the original English version of this
-License, the original English version will prevail.
-@sp 1
-@item
-TERMINATION
-
-You may not copy, modify, sublicense, or distribute the Document except
-as expressly provided for under this License. Any other attempt to
-copy, modify, sublicense or distribute the Document is void, and will
-automatically terminate your rights under this License. However,
-parties who have received copies, or rights, from you under this
-License will not have their licenses terminated so long as such
-parties remain in full compliance.
-@sp 1
-@item
-FUTURE REVISIONS OF THIS LICENSE
-
-The Free Software Foundation may publish new, revised versions
-of the GNU Free Documentation License from time to time. Such new
-versions will be similar in spirit to the present version, but may
-differ in detail to address new problems or concerns. See
-@uref{http://www.gnu.org/copyleft/}.
-
-Each version of the License is given a distinguishing version number.
-If the Document specifies that a particular numbered version of this
-License ``or any later version'' applies to it, you have the option of
-following the terms and conditions either of that specified version or
-of any later version that has been published (not as a draft) by the
-Free Software Foundation. If the Document does not specify a version
-number of this License, you may choose any version ever published (not
-as a draft) by the Free Software Foundation.
-
-@end enumerate
-
-@c fakenode --- for prepinfo
-@unnumberedsec ADDENDUM: How to use this License for your documents
-
-To use this License in a document you have written, include a copy of
-the License in the document and put the following copyright and
-license notices just after the title page:
-
-@smallexample
-@group
-
- Copyright (C) @var{year} @var{your name}.
- Permission is granted to copy, distribute and/or modify this document
- under the terms of the GNU Free Documentation License, Version 1.1
- or any later version published by the Free Software Foundation;
- with the Invariant Sections being @var{list their titles}, with the
- Front-Cover Texts being @var{list}, and with the Back-Cover Texts being @var{list}.
- A copy of the license is included in the section entitled ``GNU
- Free Documentation License''.
-@end group
-@end smallexample
-If you have no Invariant Sections, write ``with no Invariant Sections''
-instead of saying which ones are invariant. If you have no
-Front-Cover Texts, write ``no Front-Cover Texts'' instead of
-``Front-Cover Texts being @var{list}''; likewise for Back-Cover Texts.
-
-If your document contains nontrivial examples of program code, we
-recommend releasing these examples in parallel under your choice of
-free software license, such as the GNU General Public License,
-to permit their use in free software.
-
-@node Index, , GNU Free Documentation License, Top
-@comment node-name, next, previous, up
-
-@unnumbered Index
-@printindex cp
-@bye
-
-Conventions:
-1. Functions, built-in or otherwise, do NOT have () after them.
-2. Gawk built-in vars and functions are in @code. Also program vars and
- functions.
-3. HTTP method names are in @code.
-4. Protocols such as echo, ftp, etc are in @samp.
-5. URLs are in @url.
-6. All RFC's in the index. Put a space between `RFC' and the number.
OpenPOWER on IntegriCloud