Sunteți pe pagina 1din 386

This page intentionally left blank

An Executive’s Guide to Information Technology


Assessing the most valuable information technology for an organization is becoming a growing challenge
for business professionals, who are confronted with an expanding array of options. This book is an A–Z
compendium of technological terms written for the non-technical executive, allowing quick identification
of what the term is and why it is significant. This is more than a dictionary. It is a concise review of the
most important aspects of information technology from a business perspective: the major advantages,
disadvantages, and business value propositions of each technological term are discussed, sources for
further reading are provided, and there is cross-referencing with other terms where applicable. The
essential elements of each concept are covered in a succinct manner so that the reader can quickly obtain
the required knowledge without wading through exhaustive descriptions. With over 200 terms, this is a
valuable reference for non- and semi-technical managers, executives, and graduate students in business
and technology management.

Robert Plant obtained his Ph.D. in Computer Science at the University of Liverpool in 1987. He is currently
an associate professor for the School of Business Administration at the University of Miami, and specializes
in teaching MIS Strategy both there and at other universities and companies. His research interests focus
on the role of information systems in strategic management.

Stephen Murrell obtained his D.Phil. in Computation in 1986 from the Oxford University’s Programming
Research Group. He is currently a lecturer in Computer Engineering at the University of Miami, where he
specializes in teaching programming, algorithms, and operating systems. His primary area of research is
in programming languages.
An Executive’s Guide to
Information Technology:
Principles, Business Models, and Terminology

Robert Plant
and Stephen Murrell
University of Miami, Florida
CAMBRIDGE UNIVERSITY PRESS
Cambridge, New York, Melbourne, Madrid, Cape Town, Singapore, São Paulo

Cambridge University Press


The Edinburgh Building, Cambridge CB2 8RU, UK
Published in the United States of America by Cambridge University Press, New York
www.cambridge.org
Information on this title: www.cambridge.org/9780521853361

© R. Plant and S. Murrell 2007

This publication is in copyright. Subject to statutory exception and to the provision of


relevant collective licensing agreements, no reproduction of any part may take place
without the written permission of Cambridge University Press.
First published in print format 2007

ISBN-13 978-0-511-27939-3 eBook (NetLibrary)


ISBN-10 0-511-27939-6 eBook (NetLibrary)

ISBN-13 978-0-521-85336-1 hardback


ISBN-10 0-521-85336-2 hardback

Cambridge University Press has no responsibility for the persistence or accuracy of urls
for external or third-party internet websites referred to in this publication, and does not
guarantee that any content on such websites is, or will remain, accurate or appropriate.
Contents

Introduction 1

ACM (Association for Computing Machinery) 3


Advertising 3
Agent 5
AIS (Association for Information Systems) 7
Algorithm 7
Analog 10
Anti-virus software 11
Application-development methods 12
Application generator 13
Application server 14
Application service provider (ASP) 15
Architecture 16
Artificial intelligence (AI) 17
ASCII 19
Assembler 20
Audio 21
Backup 24
Bandwidth 26
Bar code 29
Batch processing 30
BCS (British Computer Society) 31
Benchmark 32
Binary 33
Biometrics 34
Bit 35
Bluetooth 36
Broadband 37
Bug, debugging 38
Bus 40
Business continuity service provider 41

v
Contents

Business intelligence (BI) 42


Business process re-engineering 44
C, C++, C# 46
Cable Communications Policy Act of 1984 (CCPA) 48
Cables and connectors 49
Cache 50
CAD/CAM (computer aided design/computer aided manufacturing) 51
Capacity planning 52
Cell computing 54
Chaos theory 56
Children’s Online Privacy Protection Act of 1998 (COPPA) 59
CIO (chief information officer) 59
Click-stream tracking 60
Client 61
Client–server 62
Clone 64
Cluster 65
Cobol 67
Collaborative commerce 68
Compiler 69
Complexity 71
Compression 73
Computability 76
Computer 77
Computer-based training (CBT) 79
Computer Fraud and Abuse Act, 1986 80
Computer Misuse Act of 1990 (UK only) 81
Computer Security Act of 1987 82
Connectivity standard 82
Cookies 83
CPU (central processing unit) 85
Cracking 86
Database 88
Database administrator (DBA) 90
Data-flow diagram 91
Data mining 93
Data pool 94
Data Protection Act (UK only) 95
Data-quality audit 96
Data warehouse 97
Decision support system (DSS) 98

vi
Contents

Denial-of-service attack 99
DHCP (Dynamic Host Control Protocol) 101
Digital 102
Digital cash 103
Digital certificate 105
Digital Millennium Copyright Act (DMCA) 107
Digital signature 107
Digital wallet 110
Disclosure and interception of communications laws 110
Disk 111
Distributed database 115
Domain name 117
Dvorak/QWERTY keyboards 119
Dynamic web pages 120
e-Commerce/e-business 123
Efficiency 124
Electronic data interchange (EDI) 126
Email 128
Encryption 131
End-user development 133
Enterprise information portal (EIP) 134
Enterprise resource planning (ERP) systems 135
Entity-relationship diagram (ERD) 137
Ethernet 138
ETL (extracting, transforming, and loading of data) 139
European Union Directive on Privacy and Electronic Commerce 2002 141
Fiber optics 143
File server 144
File system 146
Firewall 148
Flash memory 150
Formal methods 151
Fortran 155
Fourth generation 156
FTP (file transfer protocol) 158
Functional programming 159
Fuzzy logic 161
Global positioning system 163
Groupware 164
Hacker 166
Hardware, software, firmware, application, program 167

vii
Contents

Health Insurance Portability and Accountability Act of 1996 (HIPAA) 168


Host, host name, hosting 169
Human–computer interaction 170
Hypertext, HTML 171
ICANN (Internet Corporation for Assigned Names and Numbers) 174
IEEE (Institute of Electrical and Electronic Engineers) 174
Index 175
Information lifecycle management (ILM) 177
Information Technology Infrastructure Library (ITIL) 178
Instant messaging (IM) 180
Internet 182
Internet protocol (IP) 185
ISO/IEC 17799 187
ISP (Internet service provider) 189
Java 191
JavaScript 193
Joint application design (JAD) 194
Knowledge-based systems 196
Knowledge engineer 198
Knowledge management 198
LAN/WAN/subnet/internet/intranet 201
Law cross-reference 201
Legacy system 203
Logic programming 204
Machine learning 207
Maintenance 208
Memory 209
Memory and disk size 211
Middleware 213
MIS (management information systems) department 214
Modem 215
Motherboard 217
Multicast 219
Natural language processing (NLP) 221
Network 223
Network-address translation (NAT) 226
Network devices 227
Neural network 229
Normalization 231
Object-oriented 233
OLAP (online analytical processing) 235

viii
Contents

One-way hash 236


Online communities 237
Open source software 238
Operating system 240
Optical character recognition (OCR) 243
Optical storage 244
OSI seven-layer model 247
Outsourcing 248
Packet switching and circuit switching 250
Parallel processing 251
Password 254
Patent 257
Peer to peer 258
Person month 259
Phishing 260
Port 262
Power protection 264
Privacy Act of 1974 266
Privacy Protection Act of 1980 267
Programming language 267
Protocol 272
Proxy 273
Public key–private key 274
Quantum computing 277
RAID (redundant array of independent disks) 278
Rapid application development (RAD) 279
Reliability 280
RFID (Radio-frequency identity) tags 282
Robotics 285
RSS (Really simple syndication) 286
Sarbanes–Oxley Act of 2002 (SOX) 289
Scalability 292
Secure 293
Security 295
Server 297
SIIA (the Software and Information Industry Association) 298
Software-development lifecycle 299
Software metrics 301
Software package 302
Spam 304
Spoofing 307

ix
Contents

Spyware 308
Storage 310
Structured design methodologies 313
T-Carrier 315
TCP/IP (Transport Control Protocol for the internet protocol) 315
Telnet 317
Thirty-two bit 319
Transaction-processing system (TPS) 321
Trojan horse 322
UML (Unified Modeling Language) 324
Unix 324
URL (Uniform resource locator) 326
Value added network (VAN) 329
Video 329
Virtual machine 331
Virtual memory 333
Virtual organization 335
Virtual private network (VPN) 336
Virtual reality 338
Virus 339
Visual Basic 342
Voice over IP (VoIP) 343
W3C (the World Wide Web Consortium) 346
Walk-through review 346
Waterfall model 347
Web services 348
Wireless application protocol (WAP) 349
Wireless network 351
World Wide Web 352
WYSIWYG 356
X.12 358
X.25 359
XML 359
Y2K problem 362
Zip 363

Index 365

x
Introduction

In writing this book, we have drawn upon that they must understand technology at
our experiences as professors, consultants, two levels. Firstly, they must understand
and technologists to provide a resource for what the technology actually is, and this is
executives, students, and other readers who addressed in our text by the provision of an
have a desire to achieve a rapid understand- overview section for each term. This section
ing of a technology, a computer-related pro- aims to cut through the ‘‘technobabble”
cess methodology, or a technology-related and the ‘‘spin” and to provide a solid basis
law. of understanding the technology, such that
The book provides not only the defini- executives can appraise the role of a tech-
tions for over 200 terms, but also a con- nology within the IT organization and the
cise overview of the term, the associated organization as a whole.
business value proposition, and a summary A second aspect of an executive’s role is
of the positive and negative aspects of to understand the business value proposi-
the technology or processes underlying the tion behind the technologies, and this is
term. addressed in the second section presented
The book addresses a problem faced by for every term. This enables executives to
many executives working in the twenty- come up to speed quickly on the major
first century organization, that of under- issues and how a technology fits into the
standing technology without needing to larger role it plays in their enterprise and
become a technologist. Today’s executives beyond; again references are provided to
use or are responsible for technology in enable further understanding to be gained.
nearly every aspect of their organization’s A third aspect of today’s executives’ func-
operations; however, the pace of change in tion is to understand their obligations
technology, the misinformation provided with respect to legislation and regulatory
from informal sources, and general opaque- frameworks such as Sarbanes--Oxley and
ness of the terminology can be off-putting the HIPAA acts. The book addresses this by
for executives and managers. In order to including a set of UK and US legislative
help executives overcome these problems, requirements under which organizations
we have drawn upon over twenty years of and their executives in those jurisdictions
teaching at the executive level and our must operate.
backgrounds as technologists to provide Finally, each item is concluded with con-
clear, understandable descriptions for the cise summaries of the positive and negative
most important technology and terminol- aspects generally associated with the tech-
ogy in use within today’s organizations. nology. This allows executives to ascertain
The executives’ need to understand tech- very quickly the key points associated with
nology is undeniable, but their role dictates the technology, law, or process model.

1
Introduction

While no book on technology can be readers in achieving as rapid an under-


exhaustive, we have aimed to provide clear standing of the technologies, laws, and pro-
and technically sound descriptions for as cesses as possible.
large a set of the terms used by technolo-
gists, vendors, and users as possible, with Robert Plant, Ph.D., Eur.Ing., C.Eng., FBCS
the aim of assisting executives and other Stephen Murrell, D.Phil.

Please note that the content of all law-related articles in this book represents the views of the authors who are not legal
experts, and the articles should not be read as presenting definitive legal advice.

2
Advertising

fields of computer science and information


ACM (Association for Computing
technologies can be monitored.
Machinery)
Definition: The Association for Computing Machinery Summary of positive issues
was founded in 1947 to act as a professional organiza- The ACM provides a large set of high-
tion focused upon the advancement of information- quality resources for the technology com-
technology skills. munity. They promote ethical standards for
the IT community and provide resources
to assist in the governance issues of IT
Overview
organizations.
The ACM acts as a professional organization
within the technology industry. The society
provides a range of services, including pro- Summary of potentially negative issues
fessional development courses, scholarly Membership of a professional organization
journals, conferences, sponsorship of spe- such as the ACM is optional, not compul-
cial interest groups (SIGs), and sponsorship sory as is the case in the older professions
of public awareness activities that pertain of medicine (e.g., American Medical Asso-
to technology, and to scientific and edu- ciation, British Medical Council), law (e.g.,
cational activities. The conferences, peer- the American Bar Association), and tradi-
reviewed publications, and SIGs provide tional engineering. There is no equivalent
high-quality forums for the interchange sanction to being ‘‘disbarred” for comput-
of information in the area of information ing professionals.
technology. ACM provides an electronic References
portal through which a large digital library r http://acm.org
of technology-related information may be r ACM, 1515 Broadway, 17th Floor, New
accessed, including academic and profes- York, NY 10036, USA.
sional journals, conference proceedings,
and a repository of algorithms. Associated terminology: AIS, BCS, IEEE,
W3C.

Business value proposition


The ACM provides a range of resources
that can assist technology professionals in Advertising
developing their skills. These include con-
tinuing education programs, a high-quality Foundation concepts: Web, e-Commerce.
database of technology literature, profes- Definition: Web-based advertising covers many tech-
sional conferences, and networking events. niques and technologies, each of which is intended to
The peer-reviewed literature includes an attract the user to take notice of the advertising and,
extensive library of algorithms allowing where appropriate, act upon that advert.
technology professionals to use tested and
proven algorithms rather than having to Overview
reinvent them. The ACM also provides ethi- Advertising via the internet has been grow-
cal guidelines for its membership and mon- ing in popularity and sophistication since
itors public policy issues in the area of tech- the internet was deregulated in 1995 and
nology and science. Access to the resources has spawned a series of advertising initia-
of the ACM provides an IT professional with tives. An early form of advertising was the
a quick, effective, and efficient mechanism banner ad, a simple block at the top of a
through which the continuously changing web page announcing a service or product;

3
Advertising

when the user clicked upon it, it acted as a that vendors give away without charge but
hyperlink taking them to the new site and have embedded advertising into the code
presenting more information on the prod- so that it suddenly appears upon the screen
uct, and, of course, an opportunity to buy and may or may not be removable, or could
it. The advertiser would then pay the orig- be subject to a time-delayed closure. As a
inal web site for the traffic that ‘‘clicked condition of use, some adware programs
through,” and possibly pay more should a go on to install programs on the user’s
successful transaction occur. machine, known as Adbots, and these pro-
Banner ads proved in many cases to grams then act as a type of Spyware sending
be less effective than desired and so the back information to their advertising
industry started to create personalization source pertaining to the user’s behavior.
software based upon individuals’ online Often these additional installations are
web-surfing activities, the goal being to undisclosed.
present specific products and services based Illicitly installed software may redirect
upon users’ past preferences. These person- web browsers to access all pages through
alization systems did draw criticism from a proxy, which can add banner advertise-
civil liberties groups concerned that they ments to totally non-commercial pages that
could profile individuals, and selling the in reality bear no such content. Home-page
information would be an infringement of hijacking is also a common problem: a web
the right to personal privacy. browser’s default or home page may be
As the internet evolved, more sophisti- reset to a commercial site, so that, every
cated advertising developed, including the time the web browser is started up, it
infamous Pop-up ad, an intrusive message shows advertisements or installs additional
that suddenly appears upon the screen in adware.
its own window. Pop-ups are often embed-
ded Java applets or JavaScript programs Business value proposition
that are initiated from the web page as it Well-placed internet advertising can be of
is viewed; usually the pop-up window can significant benefit to the company or indi-
be closed, but some of the more persistent vidual sending it out. In the 2000 US presi-
pop-up advertising cascades so that when dential election the Republican Party’s use
one window is closed another opens, slowly of email to spur its registered voters in
taking over the machine’s resources. It may the State of Florida to go out and vote
be necessary to shut the machine down has been acknowledged as influential in
completely to escape this situation. Careful George W. Bush’s closely contested victory
control of the web browser’s security set- over Al Gore, and changed the nature of
tings can usually prevent the problem from political campaigning. The internet can
occurring in the first place. also be used as a non-intrusive distribution
A form of advertising that is also intru- model for advertising that is driven by
sive, and in some instances criminal, is user demand, an example of which was
Adware (the term is used to denote a certain the advertising campaign by BMW, which
type of intrusive advertising but is also the used top film directors, writers, and famous
registered and trademarked name of a com- actors to create short films in which their
pany that sells anti-spyware and anti-spam vehicles were as much the stars as the
software). Adware takes the form of pop- actors. The ‘‘BMW films” series proved to be
up advertisements or banner ads and hugely popular and reinforced the brand.
comes from several sources. One source is Many early dot-com-era (1995--2000) com-
certain ‘‘freeware” or ‘‘shareware” products panies attempted to finance their business

4
Agent

through the use of a Click Through adver- Reference


tising model, but were unsuccessful since r R. L. Zeff (1999). Advertising on the Internet
the revenues generated were in most cases (New York, John Wiley and Sons).
insufficient to sustain the business. Online
Associated terminology: Internet,
advertising has been used effectively by
JavaScript, Spam, Spyware, Web browser.
some companies, a prime example being
Google, as a revenue source aligned to
a successful business model and popular
product. Agent
Many companies now offer pop-up
blockers, spam blockers, and other anti- Foundation concepts: Artificial intelligence.
advertising software, and hence the abil- Definition: A self-contained software component
ity of advertisers to reach their audience acting on the behalf of a human controller.
legally has been curtailed. With the con-
tinuous development of new technologies, Overview
the opportunity to create new advertising An agent, sometimes also called a ‘‘bot,”
media continues to develop in tandem. is a software application intended to per-
form some service for a human controller
without requiring constant supervision. A
Summary of positive issues
simple application could be to carry out
Online advertising is a relatively low-cost
a particular bidding strategy in an online
mechanism for the dissemination of com-
auction. After being programmed with the
mercial messages. The basic technology
human controller’s desires, an agent could
behind internet advertising is well known
constantly monitor the state of the auction,
and continues to evolve, and consequently
reacting to competing bids as instructed,
fresh approaches are constantly being
without any intervention; the human con-
developed. Non-intrusive internet advertis-
troller would be free to get on with life in
ing, where consumers pull the advertising
the meantime.
rather than the advertising being pushed
Another valuable application of agents,
onto them, can be very successful. Forms of
that is only just entering the realms of
advertising that don’t work will (with any
possibility, is in information gathering and
luck) very quickly die out.
filtering. The internet provides a vast quan-
tity of news and information sources, con-
Summary of potentially negative issues stantly being renewed and updated. It is
Some forms of advertising are illegal, spam impossible for a human reader to moni-
for example. Internet advertising is fre- tor even a small fraction of it all; even
quently thought of as annoying and intru- after identifying a moderately sized set of
sive by the recipient. Anti-advertisement sources for a particular area of interest,
software and systems are continually being keeping up to date with new developments
improved to block intrusive advertising. requires the devotion of significant periods
It is difficult to imagine why any cus- of time. If an agent were programmed to
tomer would do business with a company recognize articles that would be of inter-
that took over their computer with cas- est, it could maintain a permanent moni-
cading pop-up advertisements, and corpo- toring presence, filtering out just the few
rations would be well advised to think very items of genuine interest, and presenting
carefully before embarking upon such a them to the human reader on demand.
campaign. This use of agent technology could provide

5
Agent

a major increase in personal productivity, of condition that can easily and reliably be
but requires that the natural language prob- detected and acted upon by very simple
lem be mostly solved first. The problem software. However, it is widely believed that
is that software must be capable of fully the extreme speed of these agents has a
understanding human prose, not just rec- destabilizing effect on the markets, and
ognizing the words, but determining the may even have been responsible for some
intended semantics, and that is still the minor market crashes. The slowness of
subject of much current research. human reactions, and the need for a real
It is generally envisaged that agents will person to make orders, allows common
not be static presences on the human sense to slip in where it could not be pro-
controller’s own computer, but should be vided by a simple automatic agent.
able to migrate through the network. This Agents are also used by some online
would enable an agent to run on the com- financial services such as automobile-
puter that contains the data it is read- insurance consolidators, who send their
ing, and would therefore make much better bots out to the sites of other companies.
use of network bandwidth. It does, how- The bots fill out the forms on those compa-
ever, introduce a serious security and cost nies, retrieve a quote and then these quotes
problem: agents must have some guar- are all displayed on the consolidator’s site.
antee of harmlessness before they would The unauthorized use of bots on third-party
be allowed to run remotely, and the sites is prohibited by many organizations;
expense of providing the computational however, preventing their access to a web
resources required would have to be borne site open on the internet can be almost
by somebody. impossible.
The idea of an agent as a meaningful
representative of a person, acting on their Summary of positive issues
behalf in business or personal transactions, Agents can be created and used to auto-
is still very much a matter for science mate processes. Agents can be sent out over
fiction. The problem of artificial intelli- a network to collect and return informa-
gence, providing software with the ability tion to their owner.
to interact with humans in a human-like
manner, is far from being solved. It is Summary of potentially negative issues
believed by some that a large number of Agent technology is primitive and only
relatively simple agents, in a hierarchy well-structured relatively simplistic tasks
involving higher-level control agents to can be performed safely. Bots and agents
coordinate and combine efforts, may be the are very unwelcome on many web sites and
best way to create an artificial intelligence. networks, and web site providers should be
This theory remains unproven. aware that other organizations may make
covert use of their online resources, pre-
Business value proposition senting the results as their own; that can
The use of agents in business has some result in an unexpected increase in band-
degree of controversy attached to it. Sim- width and server load when a popular
plistic agents have been used for some time service is provided.
to carry out stock market transactions. If
Reference
it is desired to sell a particular stock as r J. Bradshaw (1997). Software Agents
soon as it crosses a particular price thresh-
(Cambridge, MA, MIT Press).
old, constantly monitoring prices to catch
exactly the right moment would require a Associated terminology: Artificial
lot of human effort. This is exactly the kind intelligence, Natural language processing.

6
Algorithm

Communications of AIS, and has over 20


AIS (Association for Information
specialist groups, including IT in Health-
Systems) care, Enterprise Systems, E-Business, and
Definition: The Association for Information Systems Accounting Information Systems. The spe-
(AIS) was founded in 1994 and acts as a profes- cialty groups provide forums through
sional organization for academics who specialize in which academic--industrial research collab-
information systems. orations are undertaken.

Overview Summary of positive issues


The AIS acts as a professional organiza- The AIS provides a set of high-quality
tion within the technology industry. The resources for academics and practitioners
aim of the organization is ‘‘to advance knowl- within the domain of management infor-
edge in the use of information technology to mation systems.
improve organizational performance and indi-
vidual quality of work life.” (www.aisnet.org). Summary of potentially negative issues
The society provides a range of services, Membership of a professional organization
including professional development cour- such as the AIS is optional, not compulsory
ses, scholarly journals, conferences, and as is the case in the older professions of
the sponsorship of special interest groups medicine (e.g., American Medical Associa-
(SIGs). The AIS is primarily focused on the tion, British Medical Council) and law (e.g.,
role of information systems in a business the American Bar Association).
context, and is closely aligned to servic- References
ing the needs of management information r http://www.aisnet.org/.
systems (MIS) professionals and academics r Association for Information Systems,
who study the field of MIS. The AIS, which P.O. Box 2712, Atlanta, GA 30301-2712,
merged with ISWORLD (Information Systems USA.
World) in 1998, also merged its major con-
ference, the AIS Conference on Information Associated terminology: ACM, BCS, IEEE,
Systems, with the highly respected Interna- W3C.
tional Conference on Information Systems
(ICIS) in 2001. ICIS provides a forum for
the presentation of high-quality papers and Algorithm
acts as the major annual academic recruit-
ing conference for MIS faculty. Foundation concepts: Program, Formal methods.
Definition: A set of explicit, unambiguous instructions
Business value proposition for solving a problem or achieving a computational
The AIS provides MIS professionals with a result.
set of resources through which their pro-
fessional skills may be developed. These Overview
resources range from continuing educa- An algorithm is, in a way, a solution to a
tion courses and access to a high-quality problem. It is a complex of instructions,
database of technology literature to profes- which, if followed exactly, will produce
sional conferences. a particular result. The method for long
The AIS provides a link between acade- multiplication as taught in elementary
mic research and industrial practice within schools is a good example of an algorithm.
the field of management information sys- The overall task, multiplying together two
tems. The society publishes two peer- large numbers, is not something that a per-
reviewed journals, the Journal of AIS and son can normally do in one step; we need to

7
Algorithm

be taught a method. The particular way of Of course, it is necessary to know how to


performing the task, running through the ‘‘pick up” a form, and how to compare two
digits of the multiplier one by one, mul- forms to see which comes first in alphabet-
tiplying each by every digit of the multi- ical order. Office workers know how to per-
plicand, writing the resultant digits in a form these activities, but it is possible to
particular pattern, and adding them all up produce an even more detailed algorithm
in columns at the end, is an algorithm for for these subtasks too.
multiplication. That is the nature of an algorithm. Every
Since computers are simply devices that required action must be spelled out, so that
are good at obeying explicit unambigu- the task may be performed by a worker
ous sequences of instructions, algorithms who has no understanding of it. At some
are the logical constructs that programs point it must be assumed that the worker
embody. One of the key tasks of program- knows something. They must be capable
ming a solution to a problem is designing of following instructions for example, so
a good algorithm. algorithms do not regress into an infinite
Just as there are many different ways to fuzz of infinitesimal detail. Even computers
solve problems in real life, there are often already know how to do some things: they
many different algorithms available to a can add and subtract and follow instruc-
programmer for one task. An office worker tions. An algorithm for computer execu-
who has to alphabetize a pile of forms may tion simply instructs how to perform a task
quickly fall into a particular pattern, and in terms of subtasks that the computer
get through the job almost mechanically. already knows.
The five-step sorting algorithm detailed
above is a realistic one that people do use,
(1) Start with all the forms in a pile on the
and it works. Unfortunately, it is satisfac-
left, and an empty space for a new pile
tory only for small piles of forms. If you
on the right.
start with just ten forms, the first is set
(2) Pick up the first form and set it aside.
aside, and compared with all nine of the
(3) Scan through all of the remaining forms
others, to find the one that goes on top
in the first pile, comparing each with
of the second pile. Then, the first of the
the one that was set aside; when you
remaining nine is set aside and compared
find a form in the left pile which
with all eight of the others, to find the next.
belongs after the one set aside (in alpha-
Then the first of the remainder is compared
betical order) then swap them: set aside
with all seven of the rest, and so on and so
the one from the pile, and put the pre-
on. By the end of the procedure 45 indi-
viously set-aside one back in the pile.
vidual comparisons will have been needed,
(4) Put the set-aside form on top of the new
which does not seem so bad. However, if
(right) pile.
you started with 20 forms in the pile, 190
(5) If the original (left) pile is not empty, go
individual comparisons would be needed:
back to step (2) and repeat.
twice the number of forms, but four times
the work. For large piles of forms, this algo-
This is a sequence of instructions that rithm can be impossibly slow. A mere ten
could easily be followed by a worker who thousand forms would require nearly fifty
has no idea how to sort forms, or, indeed, million comparisons.
no idea of what ‘‘sorted” means. Without It would be an absolute disaster if pro-
requiring any intelligence, this procedure grammers had to rediscover this well-
allows anyone to perform the task. known fact every time a programming task

8
Algorithm

called for inputs to be sorted. Repeated design to solve the computational problem
experience tells us that most program- independently of the final implementation
mers do not notice that the simple sorting language. This description may be written
method is too slow for large data sets, and, in a style that is easier to read than finished
of those who do notice, very few indeed program code, and thus amendments to
are able to work out a significantly faster the design can be made in a more informed
alternative. One of the essential parts of a manner than would be possible by exam-
formal training in programming is a long ining the code. Knowledge of algorithms
and demanding study of the large collec- and associated data structures also enables
tion of algorithms that have already been programmers to determine the efficiency of
discovered and analyzed, together with the their solutions. A simple and obvious algo-
Data Structures (carefully tailored, seemingly rithm that works well in simple cases may
unnatural ways of organizing data for effec- be completely unsuitable for commercial
tive access) that go with them. As with any use. A professional programmer must thus
other engineering profession, it is impossi- understand that alternate algorithms exist,
ble to do a good job without a thorough and be able to select the appropriate one,
knowledge of what has been tried before. perhaps even invent a new one, for any
If a programmer starts the job fully armed given situation. Further, the computability of
with what is already known, they will have an algorithm must also be considered, to
some chance of finding something new. ensure that the algorithm is theoretically
Inventiveness is important: not all prob- sound.
lems have been seen before. A programmer
who does not already know the standard
Summary of positive issues
algorithms and data structures is doomed
Algorithms are an abstraction of a problem
to nothing more than rediscovering the
to be solved; thinking of a general algo-
basics.
rithm rather than a specific problem allows
a programmer to write cleaner, more main-
Business value proposition
tainable code that has a strong chance of
Many of the algorithms and data structures
being reusable in other projects. An enor-
for standard and interesting problems are
mous collection of existing fully analyzed
thoroughly documented and analyzed in
algorithms with well-tested implementa-
sources such as those provided by ACM, and
tions is freely available, and a vast range
the standard textbooks well known to all
of books is also available on the subject.
trained programmers. Knowledge of estab-
Design, analysis, and proof of new algo-
lished algorithms not only gives program-
rithms continues to be a major direction
mers the tools necessary for solving stan-
of research.
dard problems without having to ‘‘reinvent
the wheel” at every step, but also provides
an essential foundation for designing truly Summary of potentially negative issues
new solutions. As Isaac Newton in 1676 Algorithmic design needs qualified spe-
explained his success in inventing calculus cialists in computer science and soft-
and understanding gravity, ‘‘If I have seen ware engineering. Inappropriate algorithm
further it is by standing on the shoulders design can result in inefficient or unwork-
of giants.” able solutions being implemented. Lack of
Algorithms are an abstraction, a high- knowledge of algorithmic techniques by
level view of a method, that enables pro- software professionals is a major cause of
grammers to construct and investigate a inefficiency and poor performance.

9
Analog

References than flowing continuously. All of these phenom-


r N. Wirth (1978). Algorithms plus Data ena occur on a scale far too small to be relevant
Structures Equals Programs (Englewood in normal life, and some are still quite contro-
Cliffs, NJ, Prentice-Hall). versial even regarding their supposed existence.]
r D. E. Knuth (1975). Fundamental Traditional electronics built with resis-
Algorithms (New York, Addison-Wesley). tors, transistors, vacuum tubes, etc. also has
r M. Hofi (1975). Analysis of Algorithms a fundamentally analog nature. A voltage is
(Oxford, Oxford University Press). just as infinitely variable as the length of a
piece of string. Before the digital age, elec-
Associated terminology: Efficiency,
tronic and even mechanical analog comput-
Computability, Programming language.
ers were common. In an electronic analog
computer, all physical quantities, whether
they are lengths, weights, or anything, are
Analog directly represented by a related electronic
quantity: voltage, current flow, etc. A device
Definition: An analog system is one that operates on that can add voltages (e.g. inputs of 3.1 V
a continuous scale. and 2.6 V producing a single output of 5.7 V)
can be used to add lengths or weights. Even
Overview into the 1970s tide tables were still being
Most real-world information is not directly created by analog computers.
perceived in digital form. Sound is really Analog computers have inherent prob-
a complex and continuously varying pres- lems: controlling voltages and currents to a
sure wave in the air, which we ‘‘hear” very precise degree is exceptionally difficult
when it impinges upon small hair-like sen- and expensive; long-term storage of accu-
sors in our ears, setting them aflutter. A rate analog data is practically impossible.
visual look at the world reveals an infinitely Now, everything that could be done by ana-
detailed spectrum of shapes, colors, sizes, log computation can be done more accu-
and textures. Even simple measurements rately and reliably by encoding the analog
such as the length of a piece of string inputs in digital form and performing a
or the weight of a potato have unlimited digital computation on it.
precision: if a string is stated to be 293 mm The only concession made by a modern
long, clearly that is just an approximation computer to the analog ‘‘real world” is usu-
limited by the ruler used; more careful ally a few analog--digital converters. The
measurement might reveal it to be 293.1, term ‘‘A to D,” or A/D, is used for a device
or 293.097 632 63 mm long. that takes an analog input (such as sound
The contrast between analog and digital or voltage) and encodes it in digital form.
is that analog measurements are thought The term ‘‘D to A,” or D/A, is the reverse. A
of as corresponding naturally to real-world computer with a microphone has a simple
measurements and having unlimited nat- A/D device to convert the sound into digital
ural precision, whereas digital measure- form. A computer with a loudspeaker has a
ments are always encoded as a string of dig- simple D/A device to convert digital signals
its and are therefore limited in precision by into audible form.
the number of digits written.
[Side note: Modern physics is of the opinion that Business value proposition
the Universe actually has a fundamentally digital Analog computing has largely been margi-
nature: light travels in particles called quanta, nalized by the digital age. Except for very
space is somehow granular and can not be infi- small systems, nearly everything that could
nitely divided, even time may really tick rather be done by an analog system can be done

10
Anti-virus software

more easily and accurately with a digital systems, watching for unexpected activity,
system. The analog world does not gener- and catching new viruses as quickly as pos-
ally impact modern business computing. sible. Once a new virus has been caught, a
new signature can be generated, and up-
Summary of positive issues loaded to all operational anti-virus soft-
Analog processing is a requirement when ware.
interfacing a computer system with real- It is essential that anti-virus software be
world inputs and outputs. continually updated. Reputable anti-virus
producers provide frequent automatic up-
Summary of positive issues dates through the internet. Anti-virus soft-
Analog processing is inherently inaccurate ware can only detect a virus that was
and inconsistent, but all analog signals already known at the time of its last
may be converted to digital form before update. There is no protection against a
processing, thus negating such issues. new form of virus.
When anti-virus software detects a virus,
Associated terminology: Digital, Audio, it will generally attempt to remove it from
Video. the system cleanly, restoring affected files
to their proper state. Sometimes this repair
operation is not possible, and the last resort
Anti-virus software is to quarantine the damaged file. This
means simply moving it to a protected loca-
Foundation concepts: Virus, Security tion from whence it can not be executed,
Definition: An application that detects and removes thus rendering it harmless without destroy-
viruses and other illicit software. ing the potentially important file that it is
attached to.
Overview Anti-virus software generally works in
Anti-virus software has the ability to recog- two modes. Periodically it begins a ‘‘scan,”
nize viruses that have already been investi- actively inspecting every single file on disk
gated by the software authors. Generally, a or in memory, looking for signs of infec-
‘‘signature” for each known virus is created, tion. This is a long process, and is typi-
which consists of certain recognizable key cally configured to happen overnight. In
attributes. This means that slightly mod- addition to periodic scans, the software
ified viruses, or viruses embedded within will also inspect all incoming (and possi-
other files, will also be detectable. bly outgoing) data, including emails, down-
A virus, beyond consisting of executable loaded web content, inserted floppy disks,
code, may have any form. There is noth- etc. Anti-virus software can also be con-
ing about a virus per se that can be detec- figured to monitor currently running soft-
ted. Only already discovered forms may ware, as well as that resident on disk, and
be detected. Fortunately, most viruses are halt any active viral processes.
unimaginative copies of existent forms,
and anti-virus software can be extremely Business value proposition
effective. Anti-virus software is a vital aspect of a
However, when a truly new virus is corporate security framework. A plan must
released upon the world, there is nothing be in place to ensure that the latest anti-
that any anti-virus software could possibly virus software is active and loaded onto
do to detect it. Protection relies upon the all email servers to intercept emails with
authors of the anti-virus software contin- known viruses and take the appropriate
ually monitoring thousands of vulnerable action. Anti-virus software should also be

11
Application development methods

present upon client computers to intercept create process guidelines or ‘‘development


infected software that is brought to that methods” through which systems can be
computer through downloads, CDs, and created. These methods structure the devel-
other portable memory devices. The use of opment as a set of smaller stages or steps
self-updating software is the recommended that focus upon specific issues usually
approach since this reduces the possibility encountered at a specific stage in devel-
of human error in maintaining the anti- opment. The models act to ensure qual-
virus software. ity levels and documentation levels, and
to help managers develop their project-
Summary of positive issues management plans.
Anti-virus software is easily accessible, The range of application development
deployable, and maintainable. Reputable methods mirrors the needs of the user
vendors actively work to identify new and communities in which they are employed.
mutated viruses and maintain their soft- The most rigorous of these are the Formal
ware to defend against new threats. methods, which are based upon the use
of mathematical notations and use proof
Summary of potentially negative issues systems to ensure that the implementation
New viruses are continually being devel- matches the specification. Commercial
oped and deployed, therefore all new viru- methods include RAD (Rapid Application
ses need to be identified and new defenses Development) and JAD (Joint Application
created and built into the anti-virus soft- Development). RAD is an approach based
ware. However, until the anti-virus software upon prototyping and was developed to
updates have been received and (automat- facilitate team software building, while
ically) installed, there remains a degree of JAD is a methodology based upon the use
vulnerability. of focused meetings through which sys-
tems are developed. Traditional models of
Reference development include the Waterfall model,
r P. Szor (2005). The Art of Computer Virus
a stage-based lifecycle model used since
Research and Defense (New York, the 1970s to develop business applications.
Addison-Wesley Professional). The development of Object-oriented software
Associated terminology: Spam, Trojan in the 1980s led to methods that assist in
horse, Advertising, Spyware. the development of object-based systems;
these range across the formality spectrum
from formal methods such as Object-Z to
the more standard commercial approach
Application development methods
known as Object-Oriented Analysis and
Foundation concepts: Software development. Design (OOAD).
Definition: An application development method is a Methodologies have also been devised to
process model through which a software program or help developers with systems development
system is developed. in specialized areas such as knowledge-
based systems, neural networks, graphical
Overview user interfaces, parallel processing, and
The development of individual software web development.
programs and larger software systems is,
except for trivial problems, a complex task. Business value proposition
Thus, in order to help application develop- The role of applications methodologies is to
ers and project managers cope with this facilitate the development of higher-quality
complexity, attempts have been made to systems with the minimum of effort. The

12
Application generator

employment of specific methods enables Knowledge Engineering Review, Volume 18,


standards to be enforced across the devel- Issue 1.
opment effort.
Associated terminology: Formal methods,
Object-oriented, Waterfall model, Joint
Summary of positive issues application design, Rapid application
A wide range of methods are available and development.
some knowledge of the possibilities helps
developers select the most appropriate
method for the task to be performed. Use
of the appropriate methodology engenders Application generator
higher-quality software development, faster
Definition: An application generator is a tool that aids
development, and heightened maintain-
in the generation of executable or source code from a
ability over the life of the system.
high-level design or model.

Summary of potentially negative issues Overview


The wide range of methods available The term Application generator grew out of
requires the developer to select the most the older term Code generator, the name
appropriate method with care. Some, such of the part of a Compiler that actually
as Formal methods, require specialist train- produces the final executable code. The
ing and high degrees of mathematical notion of a program developing a solu-
knowledge and experience on the part of tion from a higher-level description then
the developer. Some methods are ‘‘weak” evolved throughout the 1980s into the
and, rather than being based upon math- area of Computer-aided software engineering
ematical proof of correctness, rely upon a (CASE), which focused upon the genera-
testing strategy for measuring their correct- tion of programs from high-level model-
ness. Some methods have more tools avail- ing environments that developers manip-
able to the developers using them than ulated through graphical user interfaces.
others. Each of these factors needs to be The CASE initiatives ranged from very
understood for the method being deployed large systems built by major vendors that
since they significantly impact the qual- attempted to cover the entire software
ity of the system developed, as well as development lifecycle to smaller systems
the time, effort, and resources necessary to that provided tools to enable developers to
develop software. Failure to select the cor- build and link reusable software modules
rect method can have significant negative or libraries. The 1980s and 1990s saw an
consequences on the development process. explosion of interest in the Object-oriented
paradigm of software development and
References efforts to capture this market resulted in
r I. M. Holland and K. J. Lieberherr (1996). application generators specifically aimed at
‘‘Object-oriented design,” A.C.M. the development of object systems.
Computing Surveys, Volume 28, Issue 1. Vendors have built a wide range of appli-
r E. Clarke and J. M. Wing (1996). ‘‘Formal cation generators that support a variety of
methods: state of the art and future development methods, models, and appli-
directions,” A.C.M. Computing Survey, cations, including Rapid application devel-
Volume 28, Issue 4. opment, UML, Object-oriented models, Formal
r R. Plant and R. Gamble (2003). methods such as B, and Structured analysis and
‘‘Methodologies for the development of design. These development methods pro-
knowledge-based systems: 1982--2002,” duce results in a variety of programming

13
Application server

languages and systems, such as Ada, Cobol, will need to be trained in the methodolo-
Java, C++, C, Visual Basic, PL/1, DB2, SQL, gies used by the systems and must under-
and DDL. stand the implications and the limitations
of the code being generated. The code
Business value proposition generated will be consistently based upon
The use of application-development envi- the design, but the performance of that
ronments enables organizations and indivi- code will need to be assessed if perfor-
dual developers to create executable code mance is critical, since the code is often
rapidly from a high-level design. The appli- not optimized. Furthermore, it is impor-
cation generators provide a uniform devel- tant to determine whether designs and
opment environment, e.g., a UML GUI, code from different systems can be shared
through which code of consistent quality across tools, architectures, and databases.
is generated. The code that is generated
Reference
can then be tested and maintained through r J. Herrington (2003). Code Generation in
other functional aspects of the system.
Action (Greenwich, CT, Manning
This style of development is especially
Publications).
suitable to situations in which repetitive
incremental designs that operate upon Associated terminology: UML,
the same database and structures need to Object-oriented, Formal methods,
be developed. Development through appli- Structured design methodologies, Cobol,
cation generators allows the programmers Java, C++, C, Visual Basic.
to focus upon more important problems,
testing, and add-on functionality.
Application server
Summary of positive issues Foundation concepts: Server, Client.
Application generators provide developers Definition: An application server is a computer upon
with the potential to create high-quality which applications reside and that allows those appli-
designs rapidly, using systems that enforce cations to be accessed by client computers.
design rules through the developers’ inter-
face, from which consistent documented Overview
source code is generated. The design meth- Network architectures are sometimes defi-
ods available for support range from formal ned in terms of tiers, and the terms Two-
methods to traditional lifecycle methods. tier and Three-tier architecture are frequently
Similarly the range of languages output used. Each tier represents a layer of tech-
by these systems is equally wide, rang- nology. In a two-tier architecture Clients
ing through Ada, Java, C++, Cobol, and typically represent one layer of the archi-
real-time systems. Many tools are available, tecture, and the Server or servers represent
some from open-source vendors and others the other. In three-tier architectures a mid-
from commercial organizations, with cor- dle tier is placed between the client and the
responding levels of support. server. This intermediate tier is sometimes
referred to as the Application server and
Summary of potentially negative issues controls the information flow and directs
Application generators are not intended to requests that are made upon other system
provide a complete substitute for human resources such as a Database server, which
programmers but can be used to relieve may be distributed across the third tier.
them of repetitive, unrewarding, and error- The three-tier model allows greater num-
prone programming tasks. The developer bers of clients to be connected to the

14
Application service provider (ASP)

network as the application server in the


Application service provider (ASP)
second layer works to balance the workload
and optimize data use and data availability. Foundation concepts: Internet, Server.
The three-tier model became popular in the Definition: A third-party provider of applications that
1990s and eased the adoption of a range of are accessed through the internet.
network architectures, including the thin- Not to be confused with: Active server pages, a
client model. Microsoft web product.
The term Application server is also asso-
ciated with providing services to applica- Overview
tions based upon development using the The concept of an ASP is based upon the
Java 2 Platform (Enterprise Edition), known concept of outsourcing. Just as an organi-
as J2EE, a system for creating component- zation may purchase its security services or
based multi-tier enterprise applications. its janitorial services from an outsourcer,
some companies have determined that they
should purchase software services from a
Business value proposition third party. The role of an ASP can be as
Application servers are used to provide opti-
simple as a web Hosting company or an Inter-
mal access to applications and resources
net service provider where basic Web services
on a network. They facilitate the adop-
are purchased, relieving the client com-
tion of thin-client network architectures
pany of the responsibility for maintaining
and relieve the network’s database servers
a web site and an external internet con-
from having to work on task and process
nection. More complex application services
management.
provided by ASPs range from monthly pay-
roll processing services in the style of a
Summary of positive issues 1970s computing Bureau, to the provision of
Application servers are well supported by full Enterprise resource planning systems and
major software vendors. Application servers services. ASPs operate these functions on
can relieve and balance the load on the behalf of their clients, owning and running
primary network servers. the software applications at secure loca-
tions that provide full redundancy in terms
of power sources, data backups, and remote
Summary of potentially negative issues duplicate application servers.
The use of an additional server in a net-
work requires the expenditure of capital, Business value proposition
physical, and personnel resources. ASPs provide companies with an alterna-
tive to developing and providing their own
References
r D. Sadoski (1997). Two Tier Software applications. The ASP concept has been a
part of computing for a long time, with
Architectures: Software Technology Roadmap
‘‘Computer bureaux” being a central part
(Pittsburgh, PA, Software Engineering
of data processing in the 1960s and 1970s
Institute, Carnegie Mellon University).
r D. Sadoski and S. Comella-Dora (2000). when resources such as specialist printers
and payroll processing were expensive to
Three Tier Software Architectures: Software
own and run internally. The CIOs of today’s
Technology Roadmap (Pittsburgh, PA,
organizations face the same cost of own-
Software Engineering Institute,
ership and return on investment issues
Carnegie Mellon University).
as their predecessors faced in the era of
Associated terminology: Client--server. bureaux and also share concerns over the

15
Architecture

control of access to proprietary informa- Summary of potentially negative issues


tion and ensuring flexibility in the solution There is a trade-off between the cost asso-
options. ciated with using an ASP and the flexibil-
The internet has allowed universal access ity offered by it. Core processes outsourced
to applications, which, combined with to an ASP are potentially at risk. Inter-
reliable technology, provides a versatile nal learning and process improvement may
and secure ASP environment. ASPs pro- potentially be restricted by the running of
vide small businesses with an option to processes through an ASP.
service their traditional processes such as
internet service provision, web hosting, tax- References
r M. Lacity and L. Willcocks (2000). Global
ation, and human resource management
(HRM), and, depending upon requirements, Information Technology Outsourcing: In
contracts can be obtained in a variety of Search of Business Advantage (New York,
ways. For example, they may be based upon John Wiley & Sons).
r J. Harney (2002). Application Service
transaction volumes, bandwidth require-
ments, and database requirements, and Providers (ASPs): A Manager’s Guide (New
pay-as-you-go payment scales are also avail- York, Addison-Wesley Professional).
able. The ASPs provide useful functional- Associated terminology: CIO, Enterprise
ity and maintain high skill levels for their resource planning, Internet, Web
employees, thus reducing the training and services.
HRM requirements for those companies
using them. ASPs also provide a flexible
solution option for CIOs in times of orga- Architecture
nizational change or re-engineering.
The dilemma facing CIOs is that of how Foundation concepts: Hardware, Software.
to balance costs and flexibility. ASPs pre- Definition: The overall design of a system.
fer to have a fixed set of requirements
and charge more for flexibility in contracts, Overview
whilst CIOs prefer high degrees of flexibil- The design of any system of any kind may
ity and control with low costs. Contract be viewed at a variety of different levels.
negotiation is thus a central component of The top-level view, concerned with how
the ASP engagement process. CIOs are also the components fit together and interact,
required to determine which functions are rather than how the individual compo-
core to their business and, no matter how nents are constructed, is known as the sys-
good the ASP, failure in an outsourced core tem’s architecture. This is a direct exten-
function at an ASP could be fatal to an sion of the plain-English meaning of the
organization. word: the overall design of a building, con-
cerned with the structure of the building
Summary of positive issues as a whole, not considering the equally
The ASP model offers a third-party solution important but lower-level details of how
for companies and their software appli- bricks are made or the color of the bath-
cation requirements. ASPs can be used room fittings. It is perhaps telling that
as backup mechanisms for ensuring con- the word is derived directly from a Greek
tinuous process operation. ASP contacts compound meaning ‘‘the most important
may be undertaken through a variety of craft.”
payment options. Non-core processes can Computer architecture is concerned with
safely be passed to ASPs and run through how the registers, arithmetic units, control
the internet. components, etc. are combined to make an

16
Artificial intelligence (AI)

efficient working CPU, not with how the The CIO and the IT organization within
registers and arithmetic units themselves an enterprise must develop, document, and
are made. maintain systems architectures with the
Network architecture refers to the over- aim of supporting and enabling the busi-
all ‘‘map” of a network: which computers, ness processes of that enterprise. Intimate
routers, servers, switches, etc. are connec- knowledge of the architecture at all levels
ted to which, in what local groupings, and enables technology and business decisions
through which protocols. to be made in such a way as to facilitate
Software architecture is concerned with the continued alignment of corporate strat-
how the major components of an applica- egy and the technological architecture that
tion, perhaps a database, a user interface, underpins it.
and a network service provider, are combi-
Reference
ned together to make a single, integrated, r S. Spewak (1993). Enterprise Architecture
well-engineered whole.
Planning: Developing a Blueprint for Data,
Applications, and Technology (New York,
Business value proposition
John Wiley and Sons).
The term ‘‘architecture” is used by tech-
nologists and systems professionals in Associated terminology: CPU, Enterprise
many ways: computer architecture, net- resource planning.
work architecture, software architecture.
For CIOs the term ‘‘systems architecture”
is frequently used to describe the over- Artificial intelligence (AI)
all corporate computing resource which
contains as subsystems the other archi- Definition: Computer-based technologies that sim-
tectures. When designing systems architec- ulate or exceed human levels of task-related perfor-
tures many issues are considered, ranging mance or expertise, or that duplicate some aspect of
from technical issues such as the techni- intelligent human behavior.
cal limitations and capabilities of various
competing architectures to the ability of a Overview
particular architecture to support business The origins of artificial intelligence stem
processes. from the research of workers such as Alan
The selection and implementation of an Turing, who in the 1930s asked ‘‘Can a
architecture can be further complicated in machine think?” and devised a test known
several ways, including by the nature of the as the Turing test as an understandable way
organization’s existing systems, which may to approach the question. The test basically
or may not be compatible with the strate- consists of a human judge undertaking a
gic and technical requirements of the envi- conversation via a terminal and keyboard
sioned architecture, and the influence of (as in a Telex system) with a selection of
external entities such as major customers other entities. Some of those others may
who may place pressure upon a supplier to be real humans, others will be computer
use a particular technology, which then has or robotic systems. If the human can not
to be integrated into the overall architec- reliably tell which is which, Turing’s claim
tural design. The selection of a new archi- is that we must accept that the non-human
tecture may be influenced by factors such participants have achieved intelligence. The
as regulatory frameworks that require a point of the test is to isolate judgment from
certain technology to be used, or the strate- irrelevant or ‘‘unfair” influences: many
gic decision to use open-source software as claim that machines can never really think
a core of the software architecture. because they have no souls, or are only

17
Artificial intelligence (AI)

obeying their programming. The Turing ently provides the only possible solutions,
test allows a purely objective assessment traditional approaches being unsuitable or
based on merit alone. nonexistent. The incorporation of AI solu-
Since its inception, AI has developed tions into an application can therefore
many branches and can be thought of result in a distinctive differentiated solu-
in a variety of ways. The branches of tion for an organization’s products, for
AI include Machine learning, Natural lan- example washing machines that incorpo-
guage understanding, Neural networks, general rate fuzzy logic may be superior in per-
problem solving, Robotics, Expert systems, formance to traditionally controlled wash-
vision processing, and speech recognition. ing machines since they may be able to
These sub-disciplines have evolved at vary- adapt better to unforeseen combinations of
ing rates, with some aspects still being con- circumstances.
fined to research laboratories, whilst others
have become mainstream software design Summary of positive issues
techniques. Artificial intelligence has matured marke-
In its modern sense, AI is not limited to dly since its early days, and many of
attempts to duplicate or simulate intelli- the early techniques and applications that
gent human behavior. The basic techniques were experimental and cutting-edge have
of AI are frequently used in solving prob- subsequently become part of a computer
lems for which no method of solution scientist’s general ‘‘tool kit.” For example,
is known. Heuristic and Minimax searching the development of expert ‘‘knowledge-
allow a computer application to search based” systems may now be carried out
effectively an infinite range of possible through mature established commercial
solutions and find one that is at least ade- applications and supported through well-
quate. Neural networks can learn to extrapo- researched development models and test-
late solutions to problems that have never ing techniques. Artificial intelligence tends
arisen before by self-training based on to refer to methods that are still in the
samples of similar problems and their research arena, and when the technology
known solutions. These and other tech- becomes mature it merges into the gen-
niques follow the human pattern of learn- eral set of techniques available to software
ing from and adapting past experiences; engineers.
they can not be relied upon to produce
perfect or optimal solutions, but when Summary of potentially negative issues
no Algorithm is known, and no program- The newer techniques and applications
mer has been capable of creating a direct available through this branch of com-
solution, they may be the only alternative puter science are advanced in nature and
available. require specialized knowledge to design,
implement, and test. Consequently they
Business value proposition are expensive to develop. Additionally, the
Artificial intelligence covers a wide variety research nature of many aspects of AI
of areas, and the formulation of a work- requires careful management considera-
ing solution frequently requires the incor- tion before adoption, as does the suitability
poration of a wide variety of disciplines, of AI systems in respect to high-risk imple-
techniques, and skills. While this can be mentations. The validation and verification
an extremely difficult task to manage and of AI systems requires high degrees of spe-
incorporate into a working application, the cialized knowledge and remains an area of
incorporation of these AI techniques frequ- active research.

18
ASCII

It is essential to be aware of the true This code is still used as the primary rep-
nature of any AI technique built into a resentation in nearly all computers, but it
product; many produce generally very good is rather primitive in concept. Providing
results, but with no guarantees, and sce- representations only for the keys that
narios at the edge of a system’s range appeared on a teletype, ASCII is adequate
may produce markedly poorer results than only for standard American-English orthog-
customers expect. raphy. It does not even support complete
British usage (words such as anæsthetic
Reference
and œstrogen, and the £ sign), let alone
A. Barr and E. Feigenbaum (eds.) (1981). The
the accents and diacritical marks required
Handbook of A.I., Volumes I, II, and III
by European languages (élève, Würze, mañ-
(London, Pitman).
ana), or the thousands of symbols required
Associated terminology: Application by Asian languages.
generator, Expert systems, Fuzzy logic, As a remedy, an international standard
Machine learning, Natural-language known as ISO-8859 has been introduced.
processing, Robotics, Algorithm. This is an extension of ASCII, leaving all
of the existing codes in place, but also
assigning meaning to the 128 codes that
ASCII ASCII (for historically valid technical rea-
sons) did not use. This allows the more com-
Foundation concepts: Digital, Bit. mon accents and marks used in West Euro-
Definition: American Standard Code for Information pean languages, but is still a source of great
Interchange: the standard encoding for text stored on frustration for what it leaves out.
computers. Another more significant extension is
known as Unicode. Using more than a
Overview single eight-bit byte, Unicode can represent
All information stored on a computer has many thousands of characters and obscure
to be in numeric form. ASCII provides a symbols. Programmers are encouraged to
numeric representation for every symbol use this Wide character representation for
that can be typed on a ‘‘standard” keyboard. new development, but, due to compati-
Each character is represented by a unique bility issues, extra complexity in program-
small number as follows: ming and processing, and simple human
0--31 invisible characters such as inertia, converted applications are still in
ENTER and TAB. a minority.
32 space
33--47 ! ‘‘ # & % $ ‘ ( ) * + , - . / Business value proposition
48--57 the digits 0 1 2 3 4 5 6 7 8 9 The ASCII standard is universally known
58--64 :;<=>?@ and almost universally used, and has been
65--90 the capital letters A B C D E F key technology for computing since its
G . . . XYZ invention. The character set provides a
91--96 [ \ ]ˆ-- common representation for interactions
97--122 the lower-case letters a b c d e f between computers and other equipment,
g . . . xyz and thus improves reliability of communi-
123--126 { | } ∼ cations, reducing the need for error-prone
conversions, and lowering the cost of tech-
thus any typed text may be represented by nology ownership. ISO-8859 is the extended
a sequence of small numbers. standard used by web browsers.

19
Assembler

Summary of positive issues region of memory to another. It is these


ASCII is a very widely accepted and univer- baby steps that are expressed in Assembly
sally known standard, even surviving intact language programming.
in the new ISO standards. In the early days of computing, before
the development of Compilers (q.v.), pro-
Summary of potentially negative issues grammers had to work at this micro-
The standard is limited in its ability to managing assembly language level. Compu-
incorporate a growing need for symbols ters can’t understand human languages, so,
and foreign characters. Even the Unicode in order to communicate, humans have to
standard leaves out many scientific sym- adopt the computer’s own native language.
bols, thus requiring the use of non- This native language, when expressed in
standard fonts for technical publications, its pure internal, extremely user-unfriendly
which results in a significant risk of incor- binary form, is called Machine code, or
rect reproduction. sometimes Machine language. Assembly lan-
guage is simply machine code with a few
References small concessions to human readers: each
r ISO-14962--1997 (Geneva, International
instruction is represented by a mnemonic
Standards Organization). word rather than in binary (e.g., ‘‘ADD”
r ISO-8859--1 (Geneva, International
instead of ‘‘10000011” on a Pentium), and
Standards Organization). memory locations may be given names
r http://www.unicode.org.
instead of numbers.
Associated terminology: Dvorak/QWERTY Assembly code made programming a
keyboard. very labor intensive and error-prone pro-
cess. Additionally, each computer family
has its own internal structure and assem-
Assembler bly language, so the instructions used by
a Pentium 3 to perform any given opera-
Foundation concepts: Compiler, Programming lan- tion will be completely different from the
guage instructions used by a PowerPC G4, or any
Definitions: Assembly code is the internal language of other kind of processor, to do the same
a CPU, rendered in human-readable form. An assem- thing.
bler is the system software that translates assembly As higher-level languages such as Algol,
code into the purely internal binary form used by the Cobol, and Fortran were introduced, the
CPU. need for assembly-language programming
diminished significantly. An Algol program-
Overview mer could write programs in a much more
Although C, C++, Java, and Basic are well- expressive language with automatic error-
known and widely supported program- detection facilities, and the compiler soft-
ming languages that programmers use to ware would translate that Algol program
express algorithms in a form suitable for into assembly language automatically.
processing by a computer, they are not Early compilers were not very efficient, so,
directly understood by computers at all. for speed-critical parts of a program, pro-
The instructions directly understood by grammers would still resort to assembly
computers, even the most modern ones, language in order to take complete con-
express only the exceptionally simple ‘‘baby trol. High-level languages often do not pro-
steps” of a computation. They might com- vide support for machine-level operations,
mand the addition of two numbers, or the so operating systems implementers would
copying of a single data value from one also frequently use assembly language.

20
Audio

A modern, carefully designed compiler bly language programmer. Developing code


can almost always do a better job of pro- in assembly language is extremely labor
ducing fast, efficient assembly code than intensive and in most cases has no bene-
any programmer could in any reasonable fit. Debugging assembly-level programs is
amount of time. The notion that assembly an exceptionally difficult process.
language should be used for those parts of
References
a program that require absolutely optimal r Intel (1990). i486 Microprocessor,
performance is now almost always false,
Programmer’s Reference Manual
and most often counter-productive. It is
(Emeryville, CA, Intel).
generally only the creators of the most r G. Schneider, R. Davis, and T. Mertz
basic parts of an operating system who
(1992). Computer Organization and
need to use assembly language at all, and
Assembly Language Programming for the
then only in small amounts.
VAX (Malabar, FL, Krieger).
However, understanding of assembly lan- r R. Paul (1999). SPARC Architecture,
guage is still an important skill for any pro-
Assembly Language Programming, and C
fessional programmer. Even though it isn’t
(Upper Saddle River, NJ, Prentice-Hall).
directly used, assembly language reveals
how a particular kind of computer works. Associated terminology: Compiler,
Even a designer of rocket ships has to know Programming language, C, Fortran.
how rivets work.

Business value proposition


The use of assembly-level programming has Audio
declined markedly in recent years as com-
Foundation concepts: Digital, Compression.
piler techniques have improved. Even for
special processors (e.g., those in cellular
telephones), developers will primarily write Overview
their coded instructions at a high level and Sound is nothing more than rapid small
compile the solution down to small exe- changes in air pressure. Vibrating objects
cutable files which are burned into the (such as larynxes and guitar strings) cause
ROM associated with the processor within these changes, and nerves in the ear detect
the device. Assembly languages do, how- them. A microphone detects the pres-
ever, allow for closer manipulation of the sure variations in a similar way to the
processor and are always an option in ear, and converts them into corresponding
extreme situations. changes in an electrical voltage. These volt-
age changes may be measured digitally, and
Summary of positive issues represented as a series of numbers, which
A well-understood technology that has a may of course be stored or processed by
mature and large literature associated with computer. The same sequence of numbers
it. Today the primary users of assembly- may be replayed, and, through a digital-to-
level programming are operating-system analog converter, changed back into vary-
and hardware developers. ing voltages and electric currents. When
applied to a loudspeaker, the varying volt-
Summary of potentially negative issues age creates a reproduction of the original
Each processor type has its own individ- sound.
ual instruction set, and learning to make The reproduced sound will never be
effective use of an instruction set takes a a perfect reproduction of the original:
long time even for an experienced assem- digitization produces a Quantization effect,

21
Audio

which means that there is a smallest repre- recordings occupy a huge amount of disk
sentable change and nothing smaller than space, and any application that requires
it will be detected, and the varying volt- audio signals to be transmitted over a net-
ages can be sampled only so many times work will consume a large amount of band-
per second. However, the precision of the width. For this reason, there has been a lot
human ear provides a very useful cap on of research into the compression of audio
requirements; there is no point in record- files, and a number of very effective meth-
ing aspects of a sound that nobody is capa- ods are available.
ble of hearing. If the voltage produced There is sufficient redundancy in typ-
by a microphone is measured only 50 000 ical sound recordings that lossless com-
times per second, it will capture every- pression (i.e., perfect compression: after
thing within the range of human hear- decompression the restored data will be
ing, and that is well within technological identical to the original) may sometimes
capabilities. halve the file size. Standard applications for
Once an audio signal is in digital form, producing ‘‘zip files” approach this com-
it can easily be manipulated according to pression ratio, and special-purpose loss-
complex mathematical formulæ to produce less audio-compression software may fare
effects that would otherwise require very slightly better. However, it is rarely neces-
expensive signal processing hardware. Fil- sary to reconstruct digital audio with per-
tering to remove noise, or to enhance a weak fect accuracy, since the human ear is not
voice, or even to remove one voice from a perfectly precise. Lossy compression meth-
group; pitch changes to enhance a voice ods allow the user to select the appropri-
or modify one aspect of a sound; and a ate compromise between high compression
wide variety of other transformations can ratio and high fidelity for any given sit-
all be applied with nothing more than a uation. The well-known MP3 (MPEG Audio
standard personal computer. Editing oper- Layer 3), AAC (Advanced Audio Coding), and
ations, in which parts of a recording are WMA (Windows Media Audio) formats can
cut out and pasted together in a different reduce an audio recording to about 20% of
order, removing any audible discontinu- its original size without great loss of qual-
ities, are also easily performed. Scrambling, ity, by progressively discarding the parts
or audio encryption, is also much simpli- of the signal that are least likely to be
fied after digitization. Software for Digi- detectable by human listeners. When the
tal signal processing is widely available for original audio contains no high-frequency
most computing platforms, and is relatively components (e.g., spoken-word recordings)
easy to implement for a software engineer or when low quality is acceptable (e.g.,
with some knowledge of the mathematics cheap telephony), much greater compres-
of Fourier transforms. sion ratios are possible.
Digitized audio signals stored as is, just
as a sequence of numbers directly repre-
senting measurements of the sound pres- Business value proposition
sure, are known as WAV or Wave files, Since the advent of the internet, the use of
although the correct technical term is PCM sound on computers has radically departed
or Pulse code modulation. This is the format from the past when computers were pri-
used on audio CDs, and is the most conve- marily silent devices. The primary use of
nient for digital signal processing and play- audio is in entertainment and internet-
back. Wave files are rather large; a high- related applications, and there are myriad
fidelity stereo recording requires about applications, ranging from online instruc-
10 MB per minute. This means that long tional courses for employees, through

22
Audio

entertainment, advertisements, and assis- materials. In some situations, there are


tance for the disabled members of society. clearly problems associated with protection
The basis of the technologies associated of intellectual property and copyrighted
with audio on a computer is dependent materials; however, for those who want to
upon the number of users in the interac- be heard far and wide without restriction,
tion. One typical application is the Web- e.g., politicians, it creates a low-cost means
cast, a unidirectional presentation made of distributing their materials to their audi-
for example by company executives to ana- ences.
lysts or shareholders. This requires that
the recipients have an audio player on Summary of positive issues
their computer that transforms the incom- Audio technologies are well developed and
ing signal into audible sounds; the web- typically require only a computing device
cast may or may not be accompanied by with a browser. The quality of audio trans-
a video signal. Webcasts may be short in mitted on the internet is generally accept-
duration, such as those found in a share- able for most purposes, and high-fidelity
holders’ meeting, or they may be an end- audio is transmitted regularly as Digital
less streaming audio signal as broadcast by audio broadcasts to digital radios in the UK
a radio station and ‘‘streamed” (distributed) and other countries. DAB combines MPEG
via the internet. Bi-directional audio data and COFDM (Coded Orthogonal Frequency
transmissions, such as take place in VoIP Division Multiplex) technologies to create a
telephone systems, require the sender to digital signal.
have a microphone--speaker set and pro-
grams to encode and decode the data.
Audio data (e.g., music) is also capable of Summary of potentially negative issues
being downloaded from the internet and Internet-based audio requires compression
stored on a portable computing device such to be used and has some loss in data qual-
as Apple Computer’s popular iPods. These ity. WAV and other formats have large data
devices can download audio (the terminol- storage requirements. Intellectual property
ogy has evolved such that a data stream is concerns surround the distribution of dig-
sometimes termed ‘‘Podcasting”) in many ital files.
formats. Reference
The ability to provide audio data to r M. Bosi and R. Goldberg (2002).
computer users and mobile device users Introduction to Digital Audio Coding and
presents an opportunity to all owners of Standards, Springer International Series
audio materials to provide their materi- in Engineering and Computer Science
als to a potentially paying audience, and (Norwell, MA, Springer).
substantially frees the owner from the tra-
ditional manufacturing and distribution Associated terminology: Digital,
costs associated with the delivery of audio Compression, Voice over IP.

23
Backup

In short, any data kept on a computer is


Backup
vulnerable to unpredictable total loss. Steps
Foundation concepts: Storage, Information lifecycle must be taken to protect all essential data.
management, Power protection. With a networked system, it is common
Definition: Secondary copies of data and applications practice to provide a centralized backup
kept to guard against catastrophic loss in the event of service: during non-peak times, data is
a system failure. copied over the network from individual
Not to be confused with: BackUPS, the trade name of computers to a second system, usually with
a variety of uninterruptible power supply. enlarged storage capacity especially for this
purpose. A network manager should be
Overview able to automate this process, so that it
Hardware failures do occur. They can be might occur in the middle of the night
made less common by using only the most when the network is normally idle, and not
reliable hardware components, employing require any human presence. Backup oper-
all available Power protection technologies, ations are often layered, so that perhaps
and regularly monitoring system perfor- every night only files that were created or
mance to catch problems before they modified the previous day are backed-up;
occur, but nothing can prevent unexpected every weekend, all files modified or created
failure. the previous week might then be re-backed-
When a computer system suffers a hard- up to make a consolidated archive; perhaps
ware failure, it will sometimes leave data every month the entire contents of every
stored on disks undamaged, but some fail- disk drive would be backed-up. This kind of
ures can destroy a disk drive, and some will arrangement reduces the amount of space
result in damage to files: if an application required for backup copies, and also min-
is actively modifying a data file when the imizes the search time required to restore
computer suddenly stops working, the file data when necessary.
can be left in an intermediate state with A central network-based backup system
just half of the change made (perhaps a bal- usually makes sense for any organization
ance transfer had credited one account but with more than one computer in use. It
not yet debited the other), leaving the file’s also requires even more stringent security
state invalid. than the other systems. Since this one com-
Sadly, amongst the most common forms puter holds a copy of all data from all other
of hardware failures are failures of the disk computers, it will be an obvious target for
system itself. Damage may range anywhere any break-in attempts. In addition, it must
from a slight degradation in performance not be forgotten that the computer used
to a complete loss of all data and the disk for backups is just as delicate and fallible
itself becoming unusable. as any other computer. Losing all backed-
Additionally, software problems are very up material due to a single hardware fail-
likely to cause loss of data. The presence ure could be crippling for any organization.
of Bugs in applications is extremely wide- Clearly, the archive files created by backup
spread, and, if an application is ‘‘buggy,” it procedures can not be left long-term on a
could do anything from occasionally failing regular disk storage system.
to update files completely to accidentally Simple computer or software failures are
deleting everything stored on the disk. It not the only causes of data loss. Fires,
must not be forgotten that Viruses are also floods, earthquakes, tornados, riots, and
a common problem; anything that a buggy terrorist attacks can all leave an entire
application could do accidentally, a virus installation completely destroyed. Any orga-
will happily do on purpose. nization that is to survive such a disaster

24
Backup

must store its backup archives in a loca- available, and auto-loading units that can
tion that is both secure and distant. The hold ten tapes at a time and automatically
same principle applies to non-networked eject one and load another when neces-
computers: the only safe solution is to sary extend the capacity to genuinely use-
make copies of the backup archives on ful sizes. Providing adequate backups of
some removable storage medium, which data is a technological problem that can
should then be removed to another not be ignored or solved by default.
location. There are also legal considerations.
Choosing a suitable removable storage Backup archives containing copies of con-
medium is a serious problem. The simplest fidential data must be protected at least
and cheapest solution is to use writable as carefully as the original. Old tapes that
CDs (known as CD-R, CD-RW, and CD-RAM). have reached the end of their working life-
High-speed drives typically cost less than time or that are no longer needed can not
$100, and, when bought in quantity, the simply be thrown away; they must be prop-
blank disks cost only a few tens of cents. erly destroyed and made unreadable, or the
However, the capacity of a CD is only data must be completely overwritten first.
650 MB. That can be stretched a little by the In nearly all circumstances it is permissi-
careful use of data Compression, but is only ble to make a copy for backup purposes of
really adequate for the frequent Incremental purchased copyrighted software (commer-
backups that save only those files that have cial applications and operating systems, for
changed recently. It would take 250 CDs example) because CDs, even those used by
to perform a full backup on a typical well- reputable companies to distribute expen-
used disk drive. sive software, do sometimes fail or break.
Another alternative is the writable DVD However, care must be taken that illicit
(known as DVD−R, DVD−RW, DVD+R, installations are not performed from the
DVD+RW, DVD−RAM, DVD−DL, and backups, and, if the original software is
DVD+DL). These are in effect simply CDs sold or given to another party, the backups
with higher capacities. They are a little of it must be destroyed or erased.
slower, and rather more expensive, but
from the user’s perspective fundamentally Business value proposition
the same. The capacity of a DVD is 4.7 GB, The correct and successful backup of data
or, for the much more expensive dual-layer is vital to any organization and thus the
(−DL and +DL) versions, 9.4 GB: seven (or ultimate responsibility to ensure compli-
fourteen) times the capacity of a CD. So ance falls on the CIO. The task itself may
DVDs can hold more incremental backup technically fall under the responsibilities of
material, but would still not be suitable the network manager with input from the
for backing up entire disks. chief security officer. Typically the backing
Magnetic tapes in various forms provide up of data is considered through a process
another alternative. Until quite recently, known as Information lifecycle management
magnetic tapes were seen as the perfect (ILM) in which the data is guided all the way
solution to the backup problem, since a from creation to disposal through a series
single tape could store the contents of a of processes and technologies that support
few complete disk drives. Unfortunately, the value and nature of the data at each
disk technology has improved far more point in time. For example, transactional
rapidly than tape technology, and the ratio data may be archived incrementally (only
has been reversed. Currently, tape car- the changes are archived) to a backup tape
tridge drives with capacities between 1 server until the data is extracted from the
and 400 GB and greatly varying prices are transactional data, transformed and then

25
Bandwidth

loaded into a data warehouse, at which backup processes surrounding the informa-
point the data may be backed up incre- tion lifecycle.
mentally onto tapes and these removed
to another location for safe keeping. The Summary of positive issues
mechanisms and forms of the data backup The backup requirements of organizations
will also depend upon the scale of data can be related to well-known and estab-
being backed up. For a small company the lished models such as the information
capacity requirements may be such that a lifecycle. Technical issues surrounding the
CD or DVD provides sufficient storage capa- establishment of backup servers, use of tan-
bility for a weekly or monthly backup. For dem systems, off-site disaster recovery, reg-
those companies which require a complete ulatory compliance, and security are well
100%-redundant systems capability there understood, documented, and supported.
are specialist computers available from var-
ious manufacturers, which contain two or Summary of potentially negative issues
more of everything with identical data pro- All data-storage media are fallible. Data
cessing and storage occurring in each of storage and backup systems require the
the two systems; thus, should one CPU or expenditure of resources and capital. The
disk drive fail, the system can continue to creation of local data that is not supported
run as normal in the other system while by a network server may not be subject to
the problem is corrected. RAID (redundant or adhere to policies that enable data to
array of inexpensive disks) techniques pro- be recovered through backup procedures.
vide similar concurrent backups for disks Some of the technologies such as CD and
only. DVD are limited in capacity and there-
The legal issues facing many corpora- fore unsuitable for large-scale backups at
tions, particularly those operating in the data-rich organizations. Some data storage
United States, dictate the need for strong media have a wide range of sub-categories
backup policies, acts such as Sarbanes-- and require care in the selection of devices.
Oxley and HIPAA require data not only to Failure to adhere to regulations pertain-
be stored correctly but in some instances ing to data backup and storage can lead to
also to be disposed of correctly. Again these prosecution.
issues need to be resolved by the organi-
Reference
zation in the form of processes and poli- r D. Cougias, E. L. Heiberger, and K. Koop
cies pertaining to data backup. Clearly legal
(eds.) (2003). The Backup Book: Disaster
issues are important, but the possibility of
Recovery from Desktop to Data Center, 3rd
loss of intellectual property or even gov-
edn. (Lecanto, FL, Schaser-Vartan Books).
ernment secrets dictates that data must be
stored in the most appropriate way (e.g., Associated terminology: Bug, RAID,
using encryption), and disposed of securely Information lifecycle management.
(e.g., by passing a large electromagnetic
pulse through the system or at the very
least writing to the data storage medium Bandwidth
several times until the whole data-storage
area has been covered and all prior data Foundation concept: Bit.
made unreadable and incapable of being Definition:Themaximumrateofinformationtransfer.
extracted through any special forensic pro-
grams). Overview
There are many third-party companies, The quantity of information is measured
vendors, and consultants to support the in bits. One bit is the smallest amount

26
Bandwidth

of information possible; it represents the 01000010, 01000011, 01000100, 01000101,


answer to a yes/no question. All data can respectively, so the word BAD is represen-
be represented as a sequence of answers to ted by the 24-bit sequence 01000010010-
yes/no questions. 0000101000100. In this way, the entire bible
For example, if someone is having a baby, can be represented by a sequence of approx-
and you want to be told whether it is a imately 35 000 000 1s and 0s.
boy or a girl, someone could transmit to The rate at which data is transmitted is
you one of the phrases ‘‘it’s a boy” or ‘‘it’s measured in bits per second, and is lim-
a girl,” or you could decide in advance that ited for any physical medium. The max-
you want the answer to the question ‘‘is it a imum rate for any medium is called its
boy?,” and then they could transmit to you Bandwidth. The term comes from radio. A
exactly the same information with either radio station is allocated not just a particu-
‘‘yes” or ‘‘no.” In computing of course, yes lar frequency, but a whole continuous band
is traditionally represented by the digit 1, of frequencies: an AM station nominally
and no by 0. If the question is decided in transmitting on 200 kHz would in reality be
advance, one single digit, one bit, gives all transmitting on all frequencies in a band
the information. between 195 kHz and 205 kHz. This gives it
On the other hand, the question may be a bandwidth of 10 kHz (i.e., 10 000 cycles per
more complex. Perhaps someone is buying second), which determines the maximum
a new car, and you want to know what color rate at which it could transmit data. Wider
it is. You already know that it has to be red, bandwidths give larger data rates.
blue, green, or yellow, and want to be told Frequently used derived units for band-
which. Again, you could decide in advance width are: Baud, equivalent to one bit per
that you want the answers to two questions: second including overhead, not just useful
first ‘‘is it red or blue?,” then ‘‘is it blue or data; Mbps, megabits per second, equivalent to
yellow?” If the car is red the answers are 1 000 000 bits per second; Gbps, gigabits per
‘‘yes, no”; if it is blue the answers are ‘‘yes, second, equivalent to 1 000 000 000 bits per
yes”; if it is green the answers are ‘‘no, no”; second; Kbps, kilobits per second, equivalent
if it is yellow, the answers are ‘‘no, yes.” If to 1000 bits per second (this is not generally
the language of communication (the ques- the same ‘‘K” as is used to measure memory
tions to be answered) is decided in advance, size). Confusingly, the ‘‘bps” abbreviation is
an intrinsically non-yes/no question, ‘‘what occasionally used to mean bytes per second
color is it?,” can be answered with two instead of bits per second, but this is not
yes/no answers. the normal usage. The phrase ‘‘per second”
Every possible piece of information can is often omitted: a ‘‘gigabit network” is one
be encoded as a sequence of yes/no answers. with a bandwidth of 1 000 000 000 bits per
In simple cases, when there are only two second.
possible answers, only a single bit is requi- If a communications medium has a par-
red. When there are only four possible ticular bandwidth, it is physically impos-
answers, two bits are required. In most sible to transmit more data per second.
human communications, there is a very Using a 100-megabit ethernet connection,
large number of possibilities, but still it is physically impossible for more than
everything can be covered by a sequence 100 million bits to be transmitted per
of yes/no transmissions. A predetermined second.
eight-bit encoding based on the ASCII table Data may be compressed, which simply
can be used for sending arbitrary text. means removing unnecessary or redundant
For example, the letters A, B, C, D, E are details, transmitting only what really mat-
represented by the sequences 01000001, ters, and adding the redundant parts back

27
Bandwidth

in after reception. Returning to the original Organizations must also fully deter-
example, the string ‘‘YES” is a 24-bit mes- mine the drivers of the external demands
sage, but if it is known in advance that the for bandwidth requirements. A business-
only possible strings are ‘‘YES” and ‘‘NO,” to-business (B2B) e-commerce marketplace
the transmission can be reduced to one bit, for example, wholly dependent upon its
1 for ‘‘YES” or 0 for ‘‘NO.” Just one bit is online customers, would typically be con-
transmitted, but the receiver converts that cerned with metrics such as the end-
one bit back to ‘‘YES” or ‘‘NO” before dis- to-end response time experienced by its
playing the result. There was only one bit users when accessing and interacting with
of real information in the original 24-bit the marketplace (end-to-end response time
message, so a compression ratio of 24 to 1 refers to the time it takes to send a request
was possible. to a site and then receive a response back).
Bandwidth limitation is one of the To ensure customer satisfaction the B2B
fundamental requirements of information company world need to invest in a network
theory; no reputable authority disputes it. that has a bandwidth provision in excess of
If an invention claims to give you a trans- that needed to accommodate the demands
mission rate of 100 MBits per second on of a peak user population interacting with
a 10 MBit line, it must be doing so by the marketplace.
compressing the data, removing redundant Network facilities are often purchased
information. If the data you are transmit- from third-party service providers such as
ting does not have a 90% redundancy rate, ISPs or web-hosting services. Care needs to
the invention can not work. be taken when choosing a service provider
to ensure that the bandwidths and ser-
Business value proposition vices contracted are actually provided. For
Organizations need to determine both example, a web-hosting service may offer
their internal and their external band- a ‘‘bandwidth” of 10 MBytes per day, but
width requirements. The determination of that does not mean that the access speed is
internal bandwidth requirements is based high, it simply means that your customers
upon the amount of data transmitted may download up to 10 MBytes from that
internally within the organization across site per day without incurring extra costs
the LAN. External bandwidth requirements for you. Hence contracts need careful man-
are determined by considering the data agement and examination.
requirements of the company in relation
to any customers, vendors, or other entities Summary of positive issues
with which it interacts electronically. The cost of bandwidth and networking
In reality, determination of bandwidth equipment is continuing to decrease in real
requirements for companies can be a com- terms. As the availability of bandwidth con-
plex task. Typically network managers and tinues to grow, the number of devices that
CIOs are forced to trade off performance will be capable of connecting over a net-
against cost. The total cost of ownership work will continue to grow.
will be calculated inclusive of issues such
as the future bandwidth requirements for a Summary of potentially negative issues
growing organization, the scalability of the Total cost of ownership needs to be com-
network connections, the type of data traf- pletely understood in order to determine
fic, and the growth rates associated with the system requirements correctly, since
the different types of data traffic placed underestimation can have an extremely
upon the networks. negative consequence for the organization.

28
Bar code

Reference information readable by VCRs with special


r C. Lu (1998). The Race for Bandwidth: scanners, and even to publish small pieces
Understanding Data Transmission of software. Applications such as these
(Redmond, WA, Microsoft Press). seem attractive because bar-code scanners
may be constructed very cheaply, espe-
Associated terminology: Compression,
cially when compared with the costs of a
Bit, Binary, Digital.
full OCR-quality flatbed scanner. Unfortu-
nately, only small amounts of data can be
encoded as bar codes before they become
Bar code prohibitively oversized.

Definition:Auniformproductcode(UPC)thatuniquely Business value proposition


identifies an object, encoded as a sequence of visible Bar codes representing UPC and EAN codes
stripes or bars that are easily read by machine. require a laser or other optical scanner.
When EAN devised their data format they
Overview expanded the scope of the codes and made
The origins of the bar code stretch back UPC a subset of EAN; this enables EAN
to 1949 when Bernard and Norman Silver scanners to read both formats but not vice
filed a patent application titled ‘‘Classify- versa. Bar codes have been very useful in
ing Apparatus and Method.” They devised the automation of manufacturing, logis-
a disk that had concentric circles on it, tics, and, of course, retail operations, where
which could be read under ultraviolet light. the ubiquitous point-of-sale (POS) scanner
The modern ‘‘bar code” formally known speeds customers through the checkout at
as the Uniform Product Code (UPC) was stores.
devised in 1973 by George J. Laurer, a
researcher at IBM. Summary of positive issues
Each digit 0--9 corresponds to a unique The technology associated with bar-code
pattern of narrow and wide bands; an format is well established and understood,
entire bar code simply represents a large linking into inventory-management sys-
number. The UPC system assigns large tems and ERP systems.
sequences of numbers to different manu-
facturers, and within their own sequences
Summary of potentially negative issues
each manufacturer assigns unique num-
UPC codes have scalability constraints and
bers to each of their products. This ensures
are used only in the United States and
that no two products will ever have the
Canada. The growth in usage of radio-
same UPC, and a simple and cheap method
frequency identity (RFID) tags is expected to
exists for scanning bar codes and retriev-
first complement and then possibly replace
ing information on the product it is
the bar-code labeling mechanism.
attached to.
Two forms of bar code exist; the United References
States and Canada use the UPC, while the r R. Palmer (2001). The Bar Code Book:
European Article Numbering (EAN) format Comprehensive Guide to Reading, Printing,
is used elsewhere. Specifying, and Applying Bar Code and
Bar codes can be used to represent any Other Machine-Readable Symbols
numeric data, and therefore anything that (Petersborough, NH, Helmers).
has a digital representation. They have been r S. Pearce and R. Bushnell (1997). The Bar
used experimentally to publish TV-guide Code Implementation Guide: Using Bar

29
Batch processing

Codes in Distribution (Philadelphia, PA, of steps is required, and a human opera-


Tower Hill Press). tor is likely to make mistakes, or when a
sequence of steps is to be performed repeat-
Associated terminology: RFID, Optical
edly or at inconvenient times (such as
character recognition.
night-time backup procedures), automated
control is valuable. In both of these cases,
the commands are not punched onto cards,
Batch processing but typed into command files, which the
operating system is directed to obey at a
Foundation concept: Operating system. specified time. These files are known as
Definition: The automatic execution of a series of Batch files (BAT files under Windows), Com-
predetermined commands by a computer system to mand files, or Scripts.
complete a job without human intervention.
Business value proposition
Overview Batch processing is still a viable option
The term ‘‘batch processing” originated in when attempting to optimize the process-
the days when all the inputs to comput- ing of certain data sets. Not all data needs
ers were in the form of Punched cards (also to be processed in real time and the use of
known as Hollerith cards, cards of just over batch methods allows the systems adminis-
7 × 3 with holes punched in them to rep- trator (who performs a similar set of tasks
resent typed characters, each card repre- and functions to the computer operator of
senting a line of input), paper tape, or the 1970s) to optimize the machine’s usage,
magnetic tape, often without any interac- such as running the batch process on the
tive input at all. For each Run or Job, a system at night when other usage is low.
Batch of cards was prepared, commanding Alternatively, the systems administrator in
the computer to run various applications conjunction with the CIO may determine
in a particular sequence, directing input that they wish to outsource the processing
and output to various places, and provid- of certain batch processes if this makes eco-
ing contingency instructions for handling nomic or operational sense (this again is
error conditions. Batch processing in this the same as when companies sent tapes or
form allows very efficient use of computer batches of cards to computer bureaus in the
resources, since the system is never left idle 1960s and 1970s). Finally, batch processing
waiting for human activity. still occurs in modern ERP systems, when
The technology has now changed signifi- they perform tasks on accumulated data in
cantly, and for most computer systems the a single intensive session, such as creating
time wasted when the computer is awaiting a consolidation of accounts at the end of a
user commands is not considered a valu- financial reporting period.
able resource. However, there are two cases
in which the principles of batch processing, Summary of positive issues
if not its original details, remain in use. Batch processing allows optimization of the
When an organization has invested signifi- computing environment for certain func-
cant capital in a large mainframe or super- tions, and may provide an organization
computer for special-purpose processing, with a simplified option for the outsourc-
its time may still be in great demand, and ing of a process.
having it follow a predetermined sequence
of commands, rather than operate under Summary of potentially negative issues
direct human control, is a sensible deci- Batch processing operations still need care-
sion. Secondly, when a complex sequence ful monitoring to ensure that they do not

30
BCS (British Computer Society)

tie up valuable resources during periods Business value proposition


of peak demand on the computing enviro- The BCS provides a range of services for
nment during their run, and to make sure the IT profession that are of the high-
that the job is completed successfully. est quality, including professional develop-
Associated terminology: Legacy system, ment courses, scholarly journals, confer-
Transaction processing. ences, and sponsorship of specialty groups.
The society has a large number of branches
throughout the UK and the world that pro-
vide academic and practitioner seminars
BCS (British Computer Society) and lectures to members and facilitate net-
working opportunities for members.
Definition: The British Computer Society (www. The society also facilities the ‘‘Elite
bcs.org) was founded in 1957 and acts as an indus- Group” meetings, a forum for UK IT direc-
try body to promote professional services within the tors and senior executives to exchange
IT sector and provide resources to encourage greater views and discuss their technology-related
public awareness of the social and economic dimen- experiences.
sions of information technology. The membership levels of the society can
be used as a mechanism for judging profes-
Overview sional merit of individuals during corpo-
The BCS is a professional society based in rate recruitment, the levels reflecting the
the UK that aims to provide services, advice, academic as well as practical experience
and a career development path to its mem- of the holder, although many highly com-
bers and the IT sector as a whole. petent professionals choose not to pursue
The BCS sees its mission as one of rais- membership. Members are held account-
ing awareness of IT as a profession and able to a professional code of conduct and
ensuring that technology development is the BCS acts as arbiter of that code. The BCS
undertaken with the highest levels of also maintains a Professional Advice Regis-
integrity and competence. The society has ter of consultants, security specialists, and
a well-developed pathway for professional expert witnesses.
development consisting of membership
levels that are based upon academic qual- Summary of positive issues
ifications and experience. Standard grades The society has a wide reach through its
include Companion (ComBCS) for students branches and its professional development
and Associate (AMBCS) for affiliates, with courses. Its membership levels are well
Member (MBCS) and Fellow (FBCS) being embedded in the IT profession within the
considered as professional grades. The UK.
awarding of Chartered Professional Status
is also available for senior members Summary of potentially negative issues
through the Chartered Engineer (C.Eng.), There is no formal requirement to join
Chartered Scientist (C.Sci.) and European a ‘‘professional” organization such as the
Engineer (Eur. Ing.) designations which con- BCS, and membership is still optional for
fer professional status levels recognized in computing and IT professionals, unlike
the UK and the EU. for the older professions of medicine
The BCS also accredits university courses (American Medical Association, General
in computer science and IT, which can Medical Council) and law (the American
lead to exemptions from BCS exams for Bar Association). The Pan-European quali-
those candidates aiming to gain profes- fications and Chartered Scientist titles are
sional status. not widely understood beyond the EU.

31
Benchmark

References ‘‘Netbench,” which measures the perfor-


r http://www.bcs.org/BCS/. mance of a file server; and ‘‘Dhrystone,”
r The British Computer Society, 1 Sanford which measures the speed at which simple
Street, Swindon, Wiltshire SN1 1HJ, UK. arithmetic operations are performed. There
is some popularity in expressing a CPU’s
Associated terminology: ACM, AIS, IEEE,
speed in terms of ‘‘Flops,” the number of
W3C.
‘‘floating-point operations” (i.e. more com-
plex arithmetical operations) it can per-
form in a second (the MegaFlop is one
Benchmark million Flops), but this measurement is pri-
marily of interest for scientific applications,
Definition: A standardized set of tests, performed uni- and is of limited value in the business and
formly on a range of different systems, to produce an data processing environment.
objective comparison of performance.

Overview Business value proposition


When selecting a new computer system for Organizations need to understand the real-
any performance-critical task, it is essen- ities associated with the benchmarking of
tial to have some clear, consistent, and any systems that they contemplate acquir-
objective measurement of the computing ing, and which will be specified in the pur-
power of each of the candidate systems. chase contract. Performance benchmarks
Contrary to common belief, the advertised supplied by vendors need to be examined
clock speed of the CPU (figures such as with respect to the true operational con-
2.88 GHz) does not provide a reliable mea- text of the corporate IT architecture as
surement of performance. One kind of pro- a whole when the system is interacting
cessor may be able to do more in a single with the live network, other applications,
clock cycle than another, and a poorly desi- etc. Benchmark ‘‘stress” testing is especially
gned motherboard or slow memory can useful since failure usually occurs when
ruin a nominally fast processor. the systems are working at their opera-
A Benchmark is a purpose-built applica- tional limits.
tion, or a predetermined sequence of oper-
ations to be performed by an existing
application, designed to get an accurate pic- Summary of positive issues
ture of the speed of the computer as a Software tools and consultant support are
whole while it is performing exactly the available to facilitate benchmarking.
kinds of tasks that it will be used for. The
results of benchmark tests are often called
Performance metrics. Trade magazines fre- Summary of potentially negative issues
quently perform their favored benchmarks The benchmarking process can be resource-
on new systems when they are released intensive and as complex as the scale of
for review, and may provide a very cheap the system being tested. Failure to under-
source of such information. take correct benchmark testing prior to
Some of the most popular benchmarks acquisition can lead to catastrophic situa-
are the ‘‘Business Winstone,” which mea- tions, e.g., when a critical ERP module can
sures the speed of the most popular Win- not handle the volume of data requests
dows applications in realistic situations; and database updates the whole ERP sys-
‘‘Winbench 99,” which measures the perfor- tem may fail, possibly even corrupting the
mance of the system’s major components; database.

32
Binary

Reference times one is one,” and that’s it! Electronic


r D. Lilja (2000). Measuring Computer circuitry is well over a hundred times sim-
Performance: A Practitioner’s Guide pler, and therefore cheaper, if it has to
(Cambridge, Cambridge University encapsulate only the four facts ‘‘0 × 0 = 0,
Press). 0 × 1 = 0, 1 × 0 = 0, 1 × 1 = 1,” instead of
the two hundred more complex facts
Associated terminology: Software metrics,
that school-children have to learn (e.g.,
Efficiency.
‘‘7 × 8 = 6 carry 5”).
So all modern computers use the binary
system, based on twos, instead of our famil-
Binary iar decimal system based on tens. In deci-
mal each digit is worth ten times as much
Foundation concepts: Bit, Digital. as the digit to its right, so the four ones
Definition: A data representation in base two. in 1111 are worth, respectively, one thou-
sand, one hundred, ten, and one; the whole
Overview number is one thousand one hundred and
People do not usually distinguish between eleven. In binary each number is worth
the abstract idea of a number and its rep- only two times as much as the digit to its
resentation. The connection between the right, so the four ones in 1111 are worth,
abstract idea of the number forty-two, as respectively, eight, four, two, and one; the
the total number of spots on a pair of dice, whole number is fifteen.
or the number of days between November 5 Every operation is easier in binary than
and December 17, and the numeric repre- in decimal, even for people. The one disad-
sentation ‘‘42” is logically quite tenuous. If vantage is that numbers need to be longer:
people didn’t have ten fingers, our num- with ten different digits to choose from,
ber system would be quite different; if we a seven-digit phone number gives ten mil-
didn’t have eyes, we would not use inked lion different combinations; with only two
shapes as the representation at all. different digits, it would give only 128
It is quite possible to build a digital com- different combinations. In order to serve
puter based on our familiar decimal num- ten million people, binary phone numbers
ber system (in the 1950s this approach was would need to be 24 digits long. People
not uncommon); it is easy to construct elec- have severe problems trying to remember
tronic circuits that can accurately distin- a sequence of 24 ones and zeros; comput-
guish among ten different voltage levels, ers are not similarly encumbered by their
and that is all that is required to represent mental architecture.
the digits. It is also possible to construct
circuits that perform additions and multi- Business value proposition
plications in this form. However, it is much, The binary system is the basis of all mod-
much easier if only two different levels ern (post-1950s) computing and all devices
need to be distinguished. It is very easy and systems are based upon this, the only
to tell the difference between ‘‘on” and exception being analog input devices that
‘‘off,” orders of magnitude easier than dis- pass their output through a D/A (digital/
tinguishing among ten different levels of analog) converter. This approach to techno-
activity. logy allows compatibility, reliability, and
If there were only two digits instead of data interchange between devices and
the familiar ten, all of arithmetic would relieves organizations of the need to check
be much easier. How long would it take for compatibility at the most basic level of
to learn the multiplication table? ‘‘One data representation.

33
Biometrics

Summary of positive issues venience, so that a user does not have to


The binary system is universally accepted waste time with self-identification or user-
as the basis of computing and computing names and passwords. It may be a mat-
devices. ter of security, where critical information
can not be released to somebody simply
because they managed to find out what the
Summary of potentially negative issues password is. When strong encryption meth-
The binary system is easy for computers to
ods are used, it is impossible for anyone
understand but very difficult for humans
to remember their own encryption keys;
to interpret and process on any scale other
they are numbers with many hundreds,
than trivial examples. This has caused com-
sometimes thousands, of digits. Strong
puter scientists to develop higher-level pro-
encryption and decryption keys have to
gramming languages and data encoding
be electronically recorded, and that makes
mechanisms.
the problem of protecting that recording,
so that only the true owner may access
their own encryption keys, absolutely
Biometrics
critical.
Definition: “Life measurements”: measurements of Simply comparing a captured digital
the physical aspects of a human subject’s appearance, image with a pre-recorded one is not a
used to enable automatic recognition and verification viable solution. The temporary changes
of identity. that happen to a person’s appearance as
a result of normal life, such as hair cuts,
Overview strong wind, rain, allergic reactions, wear-
Recognition of individual people is a very ing contact lenses, new lipstick, etc., are
complex operation, made more difficult by much greater than the permanent differ-
the fact that people do it so naturally and ences in appearance between two distinct
unconsciously that we can not analyze our individuals. An image-based system that is
own methods. Explaining how you recog- capable of recognizing the right person
nize a familiar face is like trying to explain under varying circumstances must have so
how you make your heart beat, and pro- much latitude that it will also falsely rec-
vides no insights that are of use to a soft- ognize many other people.
ware designer. Biometrics is the idea of making quan-
At first sight, the problem seems nonexis- titative measurements of a person’s phys-
tent. People normally think they know ical attributes for use in future recogni-
exactly how they recognize others. We tion or identity verification. It is usually
remember hair and eye color, height, not the actual measurements that are
moles, and facial expressions, and it is used, but the relationships between them.
tempting to assume that we recognize a For example, the distance between a per-
person by matching them with a set of son’s eyes may be useless, as it varies
known characteristics. Unfortunately, that with range and angle of viewing, but
isn’t the way it works. We still recognize the ratio of the distance between their
our friends when they wear hats and close eyes to the distance between the earlobes
their eyes. More tellingly, we are not fooled is almost invariable for any given per-
by disguises when an interloper tries to son, and very hard to fake. Fingerprint
look like a familiar person. recognition is another aspect of biometrics:
Establishing identity is an important exact matching of pre-recorded fingerprints
part of all computing enterprises. Auto- will be thwarted by small scratches and ink
mated recognition may be simply a con- blots, but the detection of major features

34
Bit

such as whorls and ridges, and their rel- technologies and solutions are continuing
ative sizes and positions, can be quite to develop and are not infallible.
successful.
Reference
Biometric systems are still very much
J. Woodward, M. Orlands, and P. Higgins
in their infancy. There is no established
(2002). Identity Assurance in the
accepted set of biometric parameters that
Information Age: Biometrics (New York,
may be used for reliable recognition, and
McGraw-Hill).
the details of many existing systems are
jealously guarded commercial secrets. A Associated terminology: Encryption,
low-cost biometric system is unlikely to be Security.
of any use except as a gimmick or symbolic
deterrent. Even high-cost ones need to be
thoroughly tested before any investment is Bit
made.
Foundation concepts: Binary, Digital
Business value proposition Definition: A single binary digit, zero or one.
Biometrics systems can be purchased and
used to secure a variety of products and Overview
processes. These include fingerprint sys- The smallest piece of information that can
tems for doors, computers, and any device exist or be communicated is the bit; a sin-
that can be loaded with software and con- gle symbol 0 or 1.
nected to a USB fingerprint reader. The The meaning of the symbol is com-
use of facial recognition biometric systems pletely context-dependent: it could repre-
provides a convenient, cost-effective mech- sent ‘‘no/yes,” ‘‘off/on,” ‘‘the number I am
anism for the provision of a high level thinking of is less than/not less than 1024,”
of security. Other biometric systems such or the resolution of any other dilemma.
as iris recognition, hand geometry, signa- Usually a single bit is just part of a larger
tures, and voice prints are developing in communication, but any information can
their levels of reliability and acceptance. be represented as a series of 0s and 1s
(see ASCII, Binary). Information quantity is
Summary of positive issues always measured in bits and the useful-
Biometric technologies have matured and ness or carrying capacity of any channel
‘‘smart cards” with photographic biomet- of information is known as its Bandwidth
ric and cryptographic technologies are and is measured in bits per second. This
now required by agencies such as the US is true whether the channel is a digital
government’s federal agencies for person- fiber-optic cable, a high-definition color-
nel identification (Homeland Security Pres- television broadcast, a person tapping on
idential Directive 12). Technologies such a morse-code key, or the human voice.
as fingerprint readers have become acces-
sible and low-cost solutions are available. Business value proposition
The biometrics industry is supported by Bits provide the common basis of informa-
a large consulting base and academic tion measurement and representation for
literature. all digital computing, networks, and com-
puting devices.
Summary of potentially negative issues
Biometric security solutions can require a Summary of potentially negative issues
significant investment in ongoing adminis- Computing when considered at the individ-
tration, monitoring, and support. Biometric ual bit level can be extremely data intense

35
Bluetooth

and has necessitated the development of Bluetooth occupies layers 1 and 2 of the
higher-level tools, representations, and OSI seven-layer model, and thus may be
languages in order to manipulate the data used in place of network cards and cables
efficiently and effectively. in a local area network. Normal TCP/IP net-
Associated terminology: ASCII, Bandwidth, work traffic may be carried over Bluetooth
Compression. devices. In this capacity, Bluetooth is a pop-
ular means of adding mobile internet con-
nectivity to a portable computer by inter-
facing it with a suitably enabled cellular
Bluetooth
telephone.
Foundation concept: Wireless network.
Definition: Bluetooth is a specification (protocol) for Business value proposition
wireless data transmission between devices that Bluetooth is an open standard adminis-
support that protocol. tered by the Bluetooth SIG (https://www.
bluetooth.org/) that has been adopted by
Overview manufacturers and employed in a wide
Bluetooth is a protocol for wireless commu- variety of devices, including wireless head-
nication between devices of various kinds, sets, wireless-enabled telephones, computer
and has been designed to be economic in mice and keyboards, printers, USB adap-
terms of processing power requirements. ters, ‘‘Bluetooth access points” that allow
The protocol has its origins in a Special connection to a wireless network, and wire-
Interest Group (SIG) created in 1988 that less office equipment such as electronic
took its name from the tenth-century whiteboards, as well as Bluetooth-enabled
Danish king Harald Blatand (Harold gaming consoles, medical equipment devi-
Bluetooth in English). The SIG was focused ces, and GPS devices.
upon the development of an open stan- A primary adopter of the technology
dard for wireless communication between has been the automobile industry, to faci-
devices operating within a short range litate device-to-device communication,
of each other and having a low power focusing upon hands-free cell-phone opera-
consumption requirement. tion through the vehicle’s voice-recognition
The system works by wirelessly connect- system.
ing two or more devices in what is known
as a Piconet, using a base band frequency
of 2.4 GHz. The system has been designed Summary of positive issues
to work in noisy radio-frequency environ- Bluetooth is an open standard that facili-
ments, where there is already a lot of radio- tates communication between devices at a
frequency energy being transmitted and short distance using a frequency-hopping
less robust systems may be unable to work technique that enables it to work in noisy
through the interference. It achieves this by radio-frequency environments such as an
employing a technique referred to as Fre- office or home. Bluetooth has three levels of
quency hopping whereby the master device security and authentication systems built
communicates to the other devices (known into it. Multiple piconets can be combined
as slaves) in its piconet first at one frequency into one Scatternet.
and then at another, hopping through the
frequencies during the communication. If Summary of potentially negative issues
some of the frequencies are too noisy, and The Bluetooth system works at a close
the communication fails, then it will still proximity (up to 100 m depending upon
be successful on others. the power class of the system) and has a

36
Broadband

limit on the number of devices that can Several technologies have evolved to pro-
communicate on one piconet (one master vide this service. One such technology was
and seven slaves). The bandwidth of the termed an Integrated Services Digital Network
specification is low (1 Mbps). There is a prac- (ISDN) and is an ITU-T standard. Two lev-
tical limit to the number of piconets and els of ISDN have been defined: the basic
scatternets that can be in very close prox- rate interface (BRI) allows for two channels
imity due to interference from overlapping of 64 Kbps or one of 128 Kbps; the second
signals. level of ISDN is known as a primary rate
interface (PRI) and can carry 30 channels
References
r http://www.bluetooth.com of data at 64 Kbps per channel, providing a
r R. Morrow (2002). Bluetooth: Operation total bandwidth of 1.92 Mbps. These chan-
nels known as ‘‘bearer” or ‘‘B” channels can
and Use (New York, McGraw-Hill
be leased wholly or in part.
Professional).
A second broadband technology known
Associated terminology: OSI seven as the T system (T−/+1, T−2, T−3, etc.) is
layer-model, TCP/IP. discussed in the T-Carrier article.
The Bluetooth brand is owned by A third broadband technology is known
Telefonaktiebolaget LM Ericsson. as the digital subscriber line (DSL), a mech-
anism for transmitting digital data over
the standard copper telephone wire found
Broadband in most residential areas. Since DSL works
over a telephone line, it has the advan-
Foundation concept: T-Carrier. tage of being a dedicated (not shared) line.
Definition: Broadband is a general term that refers to There are several varieties of DSL, based
a high-speed data-transmission link. on the technology used and the distance
of the end user from the service provider.
Overview The speed of DSL broadband connections
The term broadband does not imply any spe- is often asymmetrical (ADSL), meaning that
cific minimum bandwidth for data trans- transmissions to the user (downloads) have
mission over a physical cable. Broadband is a different bandwidth from transmissions
a term generally used to describe any high- from the user (uploads). Typically, residen-
speed link for the transmission of data or tial users download much more than they
voice in a residential setting, that has a upload, and the asymmetry reflects this.
bandwidth higher than a dial-up internet Typical downstream rates are speeds of
connection. up to 1.5 Mbps while upstream rates of
Dial-up internet connections are based 640 Kbps or less are common. At the other
on the use of analog modems to connect end of the DSL bandwidth spectrum is
a computer to an internet service provider very-high DSL (VDSL), which uses fiber-optic
(ISP) over the normal cables owned by the cables to the end user, providing down-
telephone-service provider. Dial-up services stream bandwidths of 55 Mbps for lines
have been available to the general public of up to 300 m in length and 13 Mbps
since before the internet was deregulated over lengths beyond 1500 m, with upstream
in 1995 at a variety of bandwidths depend- bandwidths of 1.6--2.3 Mbps.
ing upon the technology used. The base A broadband connection can typically
bandwidth for dial up is 56 Kbps (Kbps = also be obtained through residential cable
thousand bits per second). television service providers. This requires
Broadband technologies are based upon the use of a cable modem rather than a tele-
the transfer of data as a digital signal. phone modem. While bandwidth typically

37
Bug, debugging

ranges from 0.5 to 3.0 Mbps, cable modems Overview


use lines that may be shared by a very large 1: Software never goes wrong. If software
number of other customers, and, as with does not behave as it should, then it was
all networking systems, the available share wrong right from the beginning. Either it
of bandwidth can be drastically reduced at was installed incorrectly, or it was wrong
times of peak demand. when it left the factory. Apart from the
possibility of software being corrupted by
Business value proposition a computer virus or accidentally by some
The utilization of broadband provides its other piece of incorrect software, soft-
user with a higher-bandwidth connection ware does not change in use; it remains
to the internet than traditional dial-up immutable. It can not go wrong. These dis-
technologies provide. tinctions are, of course, of little help. It is
well known that software very frequently is
wrong. Knowing where to place the blame
Summary of positive issues does not solve the problem.
Broadband is available at a variety of band- It is almost true to say that all software is
widths to home and commercial users. wrong. Every large piece of software seems
to have something wrong with it, usually
Summary of potentially negative issues many things, many of them significant.
Broadband speeds need to be checked This is an entirely human problem. It is not
by programs that specifically examine part of the nature of software to have bugs,
upstream and downstream data rates. it is the nature of software development
Cable access to broadband is limited in methods to produce them.
commercial office buildings because these Human beings are exceptionally fallible
facilities tend not to have been wired for creatures. We may be capable of all sorts
television. of clever things, but our memories are
very idiosyncratic. Developing a large
References piece of software requires programmers
r http://www.itu.int/.
r http://www.dslforum.org/. to remember incredible amounts of detail
about hw each component is to be used,
Associated terminology: Bandwidth, and how they are intended to interact. If
Network. more than one programmer is involved,
the vicissitudes of memory are abetted by
the ambiguities of communication: how
do you know that what you understood
Bug, debugging someone to say is exactly the meaning
intended?
Foundation concept: Software development lifecycle.
Over the years, numerous software devel-
Definition:
opment methodologies have been devel-
Bug (1): An error in some software. oped, with the intent of putting some order
Bug (2): A remote surveillance device. into the process of programming, provid-
Bug (3): A movable pointer on an instrument dial, ing checks and balances and requirements
avoiding memorization of settings. for documentation (see Software development
Debugging (1): Attempting to detect and remove lifecycle). All of these methods still rely on
errors in software. humans doing what they are supposed to
Debugging (2): Attempting to detect and remove do. No matter how well set out the rules
surveillance devices. are, there is always the possibility that a

38
Bug, debugging

human being will actually be human and engineers investigating a computer failure
make a mistake, failing to follow the proce- found an actual dead bug (in the sense of
dure and introducing the very errors that an insect) stuck somewhere in the works
those procedures are supposedly guaran- and causing the problem. This is not neces-
teed to prevent. sarily a true story, but it is widely believed
There is one approach that provides and often repeated.
a theoretical chance of success: Formal 2: Remote surveillance devices in com-
methods together with Computer-aided design puting take two forms.
(CAD); for further details see those entries. (1) Software illicitly installed on a com-
Even those methods do not really guaran- puter that accesses stored data or records
tee success. Formal methods succeed only activities, transmitting records to the
if someone has a correct understanding installer or keeping them for later retrieval.
of the problem to be solved, and success- Spyware (q.v.) is an example of a soft-
fully represents that understanding in a ware bug, but much more sophisticated
mathematical formulation. CAD allows a forms exist. Good anti-virus software will
programmer to convert the mathemati- find many instances of software bugs, and
cal formulation into usable software with- Adware removal tools find some others, but
out error, because the human program- in critical situations where high-value data
mer provides only the intelligence required is at stake, there are no tools that may be
to decide how to proceed; the computer completely relied upon. A trusted techni-
itself performs the actual steps of program- cian, fully familiar with exactly what soft-
ming, and refuses to perform any step ware is supposed to be on a computer, and
that does not conform to provably cor- performing frequent regular checks, is a
rect pre-established rules. Unfortunately, good but expensive precaution. The strict
somebody has to create the CAD soft- use of a Firewall (connection blocking soft-
ware in the first place, and, if the CAD ware or hardware) also helps a lot, but
application has a single bug in it, then nothing offers complete protection. It is
every program it helps to create could be quite possible (although certainly unlikely)
wrong. that spying utilities could be built into the
Human beings make mistakes, and can operating system or even the firewall soft-
not be prevented from doing so. Comput- ware itself.
ers can not understand the requirements (2) Physical devices (the ‘‘bugs” of spy
for software projects, and can not program films) attached to the computer, which
themselves. Testing is not an adequate solu- may transmit information by radio or by
tion to the problem. Of course, testing subverting existing channels (such as a
is essential, but it is not sufficient. Test- network connection), or simply store infor-
ing can not possibly cover every case that mation internally until the device can be
may come up. All major software manufac- retrieved by the planter. Kits are available
turers have extensive testing facilities; the for ‘‘sweeping” work areas and detecting
fact that bugs are still common is proof any radio-frequency emissions, but they are
that testing does not solve the problem. of limited reliability, since bugs do not have
Good programming practice can certainly to transmit all of the time, there are many
drastically reduce the number of bugs in frequencies that carry untraceable electri-
programs, but their elimination is not a cal noise, and bugs are not limited to radio
practical possibility. transmissions. Again, frequent visits from a
It is claimed that the term ‘‘bug” origi- trusted technician who knows exactly what
nated in the early days of computing, when should be inside the computer will help,

39
Bus

but only partly. It is quite possible for hard- Summary of positive issues
ware manufacturers to build perhaps disk Development methodologies and tech-
drives or even CPUs with recording equip- niques are available to minimize the occur-
ment embedded in their design. That is not rence of errors in a software system; Formal
to say that it does happen, but it is certainly methods are a prominent path. The use of
possible. verification and validation techniques also
assists in the minimization of errors in sys-
Business value proposition tems. Software developers have tools such
1: The term bug has several uses within as debuggers available to assist them in
the technology community, the most com- locating errors in their code.
mon being the bugs or errors in a pro-
gram code. A famous example of bugs in Summary of potentially negative issues
the code is the ‘‘blue screen of death” that Errors and bugs are present in all but
occurs when a personal computer crashes. the most carefully developed systems, and
Errors in software code are made by the even careful development is no guarantee
programmers during the development pro- against them. The use of formal methods
cess. Errors occur anywhere the code does throughout development would theoreti-
not satisfy the specification or where the cally lead to perfect bug-free software, but
specification is in fact wrong and the code formal methods require exceptional levels
manifests the specification errors. Typically of specialist training, and available tools
both cases are referred to as bugs. There provide incomplete coverage of the soft-
are steps that can be taken to eradicate ware development lifecycle. Even with the
bugs and errors through the use of valida- most perfectly designed software, simple
tion (‘‘are we building the right product?”) errors in typing (either of the code itself,
and verification (‘‘are we building the prod- or in later data entry) may cause an error
uct right?”) techniques. Programmers some- that lies dormant and undetected for years.
times have to examine thousands of lines of
Reference
code to locate the error; however, software r S. Murrell and R. Plant (1977). ‘‘A survey
products known as debuggers are available
of tools for validation and verification
to assist them in this task.
1985--1995,” Decision Support Systems,
2: Bugs can also refer to surveillance
Volume 21, No. 4.
devices. In situations where secrecy is
required, countermeasures may need to be Associated terminology: Formal methods,
in place to overcome these issues. A bug Reliability, Programming language.
may be of the cold-war variety with micro-
phones in walls and telephones, but more
likely in a modern context is invasive
Bus
software known as Spyware. Key-stroke cap-
turing software could be placed on a vul- Foundation Concept: CPU.
nerable computer, and the use of software Definition: A common connector used to transfer data,
to break into data repositories containing commands, or signals between connected devices. A
important intellectual property is always common pathway.
a threat. Careful counter-surveillance mea-
sures need to be put into place, covering Overview
both physical and software threats, and In computing, the word Bus has the same
this is the responsibility of the company’s derivation as the behemoth of public
chief security officer and the CIO. transportation: omnibus, meaning ‘‘for all”

40
Business continuity service provider

(Latin, dative plural of omnis). A bus is standard for a few years, supporting high-
a single elongated pathway shared by a performance video cards, but now PCI
number of subsystems, and used by all for (Peripheral Component Interconnect) is the
communications. The alternative design is almost universal standard for PCs. IDE, EIDE,
to provide every subsystem with a direct and ATA (Advanced Technology Attach-
connection to every other subsystem; this ment) are the commonly used busses for
could produce a faster overall system, since connecting disk drives to the motherboard.
different pairs of subsystems could com- Off the motherboard, and usually out-
municate at the same time, but would side the computer, there is a third level
be prohibitively expensive and complex to of bus use. Instead of having a socket for
implement. A bus replaces a multitude of every possible device that might ever be
direct connections with one shared med- connected to a computer, have one socket
ium. Busses do require special design con- that connects to a bus, and many different
siderations, to make sure that two pairs of devices may in turn be connected to that
subsystems do not attempt to communicate bus. USB (Universal Serial Bus) is the best-
on the same bus at the same time, but this known current bus system, and, although
is only of concern to computer hardware many devices may be connected to a sin-
designers; the question of whether or not gle USB bus, it is so popular that comput-
to have a bus is irrevocably pre-decided for ers usually have two or four USB sockets.
everyone else. SCSI (Small Computer System Interface) is
Busses occur at three distinct levels in another common standard, now waning in
a computer system. Inside the CPU, busses popularity, for high-speed external devices.
are used to connect the individual parts See Port for more information on USB and
(arithmetic units, sequencers, memory con- SCSI.
trollers, etc.) to make the whole CPU work Minicomputer and mainframe manufac-
as effectively as possible. The design of turers are usually less concerned with com-
the bus inside a CPU is invisible to users patibility with third-party equipment, so
of the computer, and it can be effectively larger computers tend to have proprietary
judged only by measuring the actual deliv- internal bus systems (which are often called
ered performance of the CPU in controlled Backplanes), such as Unibus and Q-bus, but
tests or Benchmarks (q.v.) there are also independent standards such
The CPU and other major components as the IEEE-1014 VMEbus. Mainframes often
are connected by another level of busses on use proprietary busses for peripheral inter-
the Motherboard. Again, there is very little connection, such as IBM’s FICON (Fiber
choice available, motherboard designers connectivity) and ESCON (Enterprise System
provide the best bus they can, and it can Connection), which can provide commu-
not be changed. If a user takes a strong nication at 100 Mbits per second over dis-
dislike to the currently favored bus tances exceeding ten miles, and Digital’s
design (currently PCI), they will have a DSSI. There are also independent standards
hard job finding any alternatives. The such as ANSI’s FDDI (Fiber Distributed Data
favored bus design does change as techno- Interface).
logy improves. The original IBM PCs used
a 16-bit bus called ISA (Industry Standard
Architecture), which was improved by EISA Business continuity service provider
(Extended ISA), and the short-lived Micro-
Channel Architecture. VESA (Video Electro- Foundation concepts: Information lifecycle manage-
nics Standards Association) was a popular ment.

41
Business intelligence (BI)

Definition: Business continuity service providers are potential threats and can work with enti-
organizations that perform technology-related disas- ties to provide documentation, training,
ter planning and recovery services. resources, and processes to ensure continu-
ity of service.
Overview
Owing to the mission-critical nature of Summary of positive issues
technology within organizations, it is BCSPs provide independent assessments of
imperative for them to undertake disaster risks and threats to an organization. They
recovery assessments and planning. Business are capable of presenting current best-in-
continuity planning (BCP) commences with class process and technical solutions to cor-
a risk assessment of the current imple- porations.
mentation and covers all aspects of the
IT organization, assessing potential threats Summary of potentially negative issues
to the integrity of the organization and The development of BCP requires frequent
its systems. Risks include power failures, updating as threats and risks change. The
security breaches, failure of data backup processes, people, knowledge, key docume-
systems, hardware failure, environmental nts, technology, infrastructure, customers,
threats (e.g., hurricanes), and infrastruc- vendors, and other ‘‘critical resources”
ture threats. The second stage of BCP associated with corporations continuously
involves managing the risk and develop- change and this also forces BCP on a con-
ing responses to the threats through such tinuous basis.
means as backup power systems, UPSs, Reference
firewalls, improved physical and software r H. Wang (2004). ‘‘Contingency planning:
security measures, redundant hardware, emergency preparedness for terrorist
duplicate facilities (e.g., in environments attacks,” IEEE Aerospace and Electronic
not threatened by extreme weather), and Systems Magazine, Volume 19, Issue 3.
stronger infrastructures. The use of stress
analysis upon systems is a useful mecha- Associated terminology: Hosting, Backup,
nism employed to show the potential for Information Technology Infrastructure
weakness in the systems. Library.
The assessment of risk levels and the
planning processes surrounding this assess-
ment may be attuned to the business risk Business intelligence (BI)
assessment. Should the threat be deemed
high, an assessment of critical systems and Foundation concept: ERP.
emergency-level responses may be made for Definition: The use of analytic software applications
all technologies, people, and processes crit- to monitor business processes.
ical to the functioning of the business.
Overview
Business value proposition Business intelligence or Business analytics soft-
The use of BCP allows businesses to con- ware helps corporations to perform anal-
sider a variety of risks and threats that ysis of their business processes. Business
could impact their business. The business intelligence systems operate upon data
continuity service providers (BCSPs) are that has been extracted, transformed, and
third-party specialists who develop con- loaded into specialist databases from
tingency plans for businesses. As service online operational databases such as those
providers they act independently to assess found in ERP systems. These databases are
the risks associated with a wide variety of typically referred to as Data warehouses and,

42
Business intelligence (BI)

depending upon the architecture of the country, showing call volumes from each
business intelligence system, the data ware- state. The user could then drill down to
house may be a single Enterprise data ware- see the volumes from the counties within
house that contains within it specialist Data a state, then the cities, the zip codes and
marts, e.g., a finance data mart, a customer so on. At each level data relating to pro-
relationship management (CRM) data mart cess metrics will be available to the user.
and a human capital management (HCM) Business intelligence systems also allow
data mart, or alternatively individual spe- score cards to be created for processes and
cialist data marts may have been created sets of processes based upon performance
from, and exist outside of, the main enter- targets and the actual results as they
prise data warehouse. relate to those metrics. The aim of busi-
Business intelligence systems create an ness intelligence software is to help man-
environment in which a process or set of agers to make faster, better decisions based
processes may be carefully monitored. Data upon accurate, easy-to-interpret, and timely
is drawn from the operational databases data.
as required (e.g., every day or every hour)
and placed into the data marts. The soft-
Summary of positive issues
ware then uses the data from one or more
Business intelligence has emerged as a
data marts, depending upon the architec-
mature branch of the software industry
tural design, to generate reports. For exam-
and many vendors offer industry-specific
ple, a business intelligence system may be
solutions for a variety of processes within
used to monitor a call center, providing
an industry or function. Business intelli-
the management with a visual ‘‘dashboard”
gence systems can usually be built onto an
through which to monitor such parameters
existing data mart or warehouse. The sys-
as call volume, response time, length of
tems are intended to be highly graphical in
call, and average number of calls per hour.
nature and possess powerful user interface
Business intelligence systems simplify the
capabilities.
production of reports.

Business value proposition Summary of potentially negative issues


Business intelligence systems provide man- The deployment and use of a business intel-
agers with the ability to monitor pro- ligence system relies upon the availabil-
cesses in a way that traditional transac- ity of clean, well-structured data. Failure
tion systems and reports can not. Busi- to create a well-formed data warehouse
ness intelligence software is developed by or data mart will invalidate any results
vendors who can provide industry-specific presented by the system. Selection of the
solutions and tools that offer managers correct analytic tools and metrics is vital,
insights into their core business processes. and careful planning needs to be under-
The systems are configured to match the taken prior to the creation of the data
processes of the organization deploying warehouse and the business intelligence
them and are typically highly graphical in solution.
nature, depicting process metrics through
Reference
a variety of mechanisms such as graphs, r E. Vitt, M. Luckevich, and S. Misner
bar charts, and traffic lights. Depending on
(2002). Business Intelligence (Redmond,
the design of the system, the software may
WA, Microsoft Press).
provide drill-down capabilities for the user.
For example, a system monitoring calls to a Associated terminology: Database, Data
call center may display a map of the whole warehouse.

43
Business process re-engineering

frequently very resource-intensive and dif-


Business process re-engineering
ficult.
Foundation concept: Management information sys- In the 1980s, as the value of using infor-
tems. mation systems as a strategic tool became
Definition: Business process re-engineering refers to more apparent, it became increasingly
the modification of one or more business processes in important to fully re-engineer processes,
order to achieve a new process goal. linking together more systems, ranging
from warehouse management to human
Overview resource management. However, the chal-
Business process re-engineering (BPR) is the lenge involved in doing this with old legacy
task of examining a business process and systems was becoming more difficult, as
redefining it to be more efficient, to meet maintaining them from a technological
changing business requirements, or to be point of view and to meet business pro-
more cost-effective. The re-engineering of a cess needs continued to drain corporate
process involves many components, includ- resources.
ing the human capital, the physical items An answer to the problem of integrating
(if any) involved in a task, the information legacy systems began to emerge through
flow, and the resource requirements. Each a new type of computing environment,
of these needs to be modeled and the con- known as Enterprise resource planning (ERP).
sequences of the proposed changes under- These systems integrated many core busi-
stood in order for a process to be modified ness processes, and had only one database
successfully. rather than many separate ones as was
Processes have been a part of every busi- common with old Cobol-based legacy sys-
ness for centuries, but the advent of com- tems. ERP system vendors implemented
mercial computing in the 1960s enabled best-of-class business processes within their
a wide range of business processes to be software, typically specific to an industry
automated. Data processing, as it was then vertical such as healthcare, retail, or man-
termed, led to the direct automation of ufacturing, and released the organization
existing business processes almost exactly from the need to identify, design, and
as they were performed by humans. The sys- implement these processes. The advent of
tems developed were typically performed ERP systems with many configurable pro-
for a single process (e.g., payroll), and were cess options built into them allowed busi-
termed stand-alone systems. Since they were nesses to re-engineer their processes in
not designed to communicate directly with significant ways. The best practices devel-
other systems, they were also frequently oped by the vendors allowed businesses to
called islands of automation. focus upon their core process competencies
The computerization of business pro- rather than worrying about upgrading and
cesses helped to make industry aware that maintaining non-core processes.
quantifiable, specifiable, and analyzable
processes really were in place. During the Business value proposition
1970s systems managers worked to integ- The philosophy underlying BPR is that
rate (re-engineer) stand-alone systems, in businesses need to reconsider their pro-
essence creating bridges linking the isla- cesses from the physical and information-
nds. During this period systems analysts flow perspectives and re-engineer those pro-
and programmers continued to attempt cesses that do not add value. Michael Ham-
to make modifications to existing systems mer in his landmark 1990 Harvard Busi-
in order to accommodate changes in pro- ness Review paper ‘‘Re-engineering work:
cess requirements. This maintenance was don’t automate, obliterate” stressed the

44
Business process re-engineering

importance of radical redesign rather than manage their non-core processes. This
continuing to incrementally change pro- allows an organization to focus upon its
cesses that had originally been created core activities.
decades ago for a different technologi-
cal world. Re-evaluating and re-engineering Summary of potentially negative issues
processes can, if performed correctly, create BPR requires that organizations under-
smoother, faster, more flexible, lower-cost stand the implications of change. There
end-to-end business processes such as order- is a potential for serious negative conse-
to-cash, procure-to-pay, and plan-to-report quences if the BPR effort is performed
procedures. These end-to-end process cycles badly. Unsuccessful ERP implementations
are typically supported by ERP systems and that have been central to BPR efforts have
many large re-engineering efforts are thus in the past caused organizations to go out
centered on implementing an ERP and the of business (FoxMeyer, a drug distribution
best-in-class process model that it brings company, suffered such a fate). BPR efforts
with it. require careful implementation because
the workforce may be hostile to process
Summary of positive issues change and work against the implementa-
BPR may be undertaken in conjunction tion of new systems and processes. Careful
with a systems re-engineering project, usu- expectation management is required.
ally in the form of an ERP. ERP systems
Reference
are mature and supported by consultants
M. Hammer (1990). ‘‘Re-engineering work:
and vendors who offer solutions for com-
don’t automate, obliterate,” Harvard
panies of all sizes. BPR can enable orga-
Business Review, July--August.
nizations to determine what is core and
what is non-core to their business, and tran- Associated terminology: UML, ERP,
sition to vendor-supported systems that Data-flow diagram.

45
C, C++, C#

in safety features if those commands do not


C, C++, C#
satisfy some safety standard. The program-
Foundation concept: Programming language. mer takes full responsibility for ensuring
Definition: Standard, general-purpose programming the correctness of a program.
languages. C was designed between 1969 and 1973,
primarily by Dennis Ritchie of Bell Labs. It
draws heavily from the design of an ear-
Overview lier language called B, which itself draws
The evolution of programming languages heavily from an earlier language called
is a short but very complex story. By BCPL, which itself draws heavily from an
the mid 1970s the design of mainstream even earlier language called CPL. When
programming languages had branched in viewed in the context of systems program-
two directions: the general-purpose pro- ming languages, C was clearly a major pos-
gramming languages and the systems itive evolutionary step. It enabled program-
programming languages. General-purpose mers to benefit from many of the lessons
languages were strongly oriented towards learned in the development of general-
supporting reliable software development; purpose programming languages without
the structure of the languages corre- strictly enforcing any rules and without
sponded closely to the structure of a good restricting what could be done. C became
top-down design, providing programmers a convenient and much safer alternative
with a built-in framework for safe and to assembly language programming. Early
verified construction; they were able to on, C was adopted as the official language
provide useful warnings about some of of Unix systems: all programming for Unix
the most common things that could go was in C, and every Unix system came with
wrong in a program’s design before it is a standard C compiler at no extra charge.
ever run, and they made many of the This undoubtedly fueled its increasing
most troublesome of programmer errors popularity.
impossible. No matter how good a tool is, it becomes
Unfortunately, what was good for general a liability when used for the wrong pur-
software development was not always good poses. The popularity of C for systems pro-
for low-level, or systems, programming. gramming soon leaked into other areas,
Implementors of operating systems, com- and it soon became arguably the most
pilers, and hardware-monitoring software popular language for all programming
sometimes have to perform exactly the tasks. The inability of C compilers to check
kind of low-level machine-oriented opera- for the most basic of errors in usage results
tions that are so strongly discouraged by in incredible numbers of programmer
the standards of good software develop- errors that simply could not be made in
ment, and programming languages that other languages. Inexperienced program-
enforce those standards are made almost mers exploit the flexibility of the rules,
unusable for such purposes. producing incomprehensible and unstruc-
As a result, the systems programming tured programs, and delight in the clever
language branch of parallel evolution was but dangerous tricks that are possible. The
also constructed. The essential feature of buffer-over-run problem, which is still one of
a systems programming language is that the most used vectors for virus delivery
a programmer should be able to give any and breaking in to supposedly secure sys-
commands that the computer is capable of tems, may be directly traced to the design
obeying, and not have to fight against built- of C.

46
C, C++, C#

In the 1980s, some improvements and complexity of C++, and has proved to
additions were made to the design of C, be another exceptionally popular program-
which resulted in there being two versions ming language, today rivaling C++, the
of the language. The original is known as Microsoft corporation introduced C# (pro-
K&R C, in honor of the authors (Kernighan nounced ‘‘See Sharp,” like the musical
and Ritchie) of the standard reference. note). It is essentially a fusion of C++ and
K&R C is now obsolete. The new version Java. Like Java, it provides better safety both
was eventually adopted as an international for programmers and for users than does
standard (ISO/IEC 9899), and is now gener- C++, automates many of the difficult pro-
ally known as ANSI C. gramming tasks (such as memory manage-
Once C had been adopted as a general- ment), and provides an extensive library
purpose programming language, it made of predefined operations useful for gen-
sense to put back into the language some of eral and network programming. C# makes
the features that had been deliberately left many of the minor design decisions differ-
out when it was designed as a systems pro- ently from Java, but is seen by program-
gramming language, and to include new mers as being a very similar tool to work
features that were not then part of the esta- with.
blished paradigm for programming. C++
(pronounced ‘‘See Plus Plus”) is a major Business value proposition
extension of C, which was designed by The major benefit of C is that it is a pow-
Bjarne Stroustrup starting as early as 1979. erful language that can be used at a vari-
It too has been adopted as an international ety of levels, from systems programming to
standard (ISO/IEC 14882). C++ adds object- application development. For many major
oriented programming and a vast library of systems developments C has become the
predefined algorithms (known as the STL, de facto standard programming language.
or Standard Template Library) as well as a For example, several ERP-systems deve-
number of less major features. It also reas- lopers have re-written their code into C,
serts many of the safety rules that are allowing their systems to be more easily
inherent in general-purpose languages, but modified than would have been possible
tend to be left out of systems languages. when the systems were written in a proprie-
However, adherence to the rules remains tary or older procedural programming lan-
optional. guage. Use of the C language in these large
C++ is an exceptionally large and com- systems developments has also encouraged
plex language with many arcane rules. application developers to create comple-
Even the most experienced of programmers mentary applications, since the program-
are frequently surprised by the details. mers do not have to learn a new program-
There was some negative reaction, founded ming language, and the use of a common
on the reasonable notion that, if the language between applications reduces
language is so complex that it even has integration issues and problems. The use of
surprises for experienced experts, what C in conjunction with Unix also provides
chance have normal programmers of get- positive benefits due to their compatible
ting things right? But C++ continued design features.
undaunted, and has easily overtaken C in The popularity of C has led to a large pro-
popularity. grammer base from which organizations
Following on the popularity of Java, can draw. Numerous compilers and tools
which provides the full power of object- have been created to support program-
oriented programming without all of the mers, along with an enormous amount of

47
Cable Communications Policy Act of 1984 (CCPA)

industry knowledge on the programming


Cable Communications Policy Act of
language and the techniques required to
solve problems.
1984 (CCPA)
A similar situation has occurred with Foundation concept: Data.
C++ and to a certain extent C#, both of Definition: The US Cable Communications Policy Act
which are used by programmers to write of 1984 is intended to protect personal information
object-oriented code. These two languages related to cable service subscribers that is collected by
are seen as a natural expansion of a pro- cable service providers.
grammer’s skill set and thus programmers
tend to be multi-lingual in C and its vari- Overview
ants. C++ is also popular with developers The US Cable Communication Policy Act of
because there is a large library of func- 1984 (CCPA) (Title 47, Chapter 5, Subchapter
tions that save the developers from hav- V-A, Part IV, Section 551, 1984) aims to pro-
ing to ‘‘reinvent the wheel,” creating code tect the privacy of cable service subscribers.
that thousands of programmers may have The code covers the obligations of the ser-
created many times before. vice provider; what data can be collected;
what data can be disclosed; the rights of
subscribers to access their own data; the
Summary of positive issues
cable providers’ obligations pertaining to
C, C++, and C# are languages with a large
the destruction of data; guidance on civil
developer base. Developers using these lan-
action damages, costs, and fees; regulations
guages are supported by a wide array of
pertaining to actions by states or franchis-
tools, compilers, and industry knowledge. C
ing authorities; and disclosure issues per-
is designed to work in close unity with the
taining to governmental entities.
Unix operating system. C++ has an exten-
sive library of functionality from which a Business value proposition
programmer can draw. The act aims to limit the unauthorized
collection of personal data that could be
Summary of potentially negative issues connected to an individual. Although it
The inappropriate use of C can lead to was originally intended to cover the col-
poorly designed programs that may be very lection of data from television cable sub-
difficult to validate. C++ is a complex lan- scribers, the act also relates to ‘‘other ser-
guage that requires considerable effort to vices provided by the cable operator” and
learn and completely understand. C# and explicitly identifies categories such as ‘‘any
C++ are not identical and the differences wire or radio communication service” and
need to be clearly understood. categories such as broadband internet ser-
vice provision could be perceived as falling
References within the provision of this act.
r B. Kernighan and D. Ritchie (1988).
The C Programming Language, 2nd Summary of positive issues
edn. (Englewood Cliffs, NJ, Prentice- The right of an individual to data privacy
Hall). is protected. Providers can apply encryption
r B. Stroustrup (1997). The C++ technologies to any data stored pertaining
Programming Language (New York, to an individual to prevent an unautho-
Addison-Wesley). rized security violation.
r http://msdn.microsoft.com/vcsharp/
programming/language/. Summary of potentially negative issues
It may be difficult for an individual to
Associated terminology: Java, Unix. access or identify a carrier that offers

48
Cables and connectors

acceptable terms of data protection in the Inside a desktop computer, the shield-
subscriber contract since all providers may ing provided by the metal casing is usually
require detailed personal information from enough to protect the internal wiring, so
those wishing to subscribe to their service, long as that wiring is not unduly long. For
e.g., the individual’s social-security number, example, it is not unusual for technically
telephone number, bank-account details, minded computer owners to buy extra-
and driver’s license details. length IDE (motherboard-to-disk) cables so
that their disk drives may be placed more
Reference
r Cable Communications Policy Act 47 conveniently. The extra length picks up
extra interference, and can have a seri-
USC §551.
ous effect on system reliability. Mainframe
Associated terminology: Security, computers with larger enclosures usu-
Encryption. ally have very carefully designed internal
shielding, which should not be interfered
with.
Cables and connectors Outside of the computer, cables for
high-speed communications must be prop-
Definition: A flexible physical connection that acts as erly shielded; that is (part of the reason)
a conduit for power, control signals, or data. why SCSI and USB-2 connectors seem to
be so unreasonably expensive. Shielding
Overview may take the form of a grounded metal
In the early days of computing and elec- sheath completely surrounding the conduc-
tronic communications, data processing tors (as in coaxial cable), or may be pro-
and transmission speeds were so slow that vided by carefully pairing signal-carrying
the components of a computer could be and grounded conductors (as in twisted
connected together with simple insulated pair), or some combination of the two.
copper wires without any loss of perfor- The connectors at the ends of these cables
mance. In modern computing, the design usually provide some kind of metal fixing
of cables adequate for high-speed data mechanism; this is not just to hold the
transfer has become a science in itself. cables in place, but also to provide conti-
When a rapidly changing electrical sig- nuity in the grounding.
nal flows along a conductor, such as a wire, Non-electronic communications media
some of its energy is actually transmitted do not suffer from electronic interfer-
into the environment as electromagnetic ence. The prime example is Fiber optics, in
radiation (that is exactly how radio trans- which the conducting wires are replaced by
mitters work). The more rapid the signal, transparent filaments, and signals are sent
the more of it gets transmitted. Conversely, along them as rapidly modulated flashes
when a conductor passes through an elec- of light, somewhat in the style of Morse
tromagnetic radiation field, part of that code flashed on an Aldis lamp. Complete
field is picked up, and converted to elec- immunity to electronic interference means
tric current in the conductor (that is how that fiber optics may be used in environ-
radio receivers work). Modern office equip- ments where electrical connections will
ment, especially computers, emits electro- not work, but the equipment required to
magnetic radiation on a wide spectrum of send and receive signals along fiber optics
frequencies, and those very same emissions is much more expensive. Of course, light
are picked up by that same equipment as itself would be interference for a fiber-optic
interference that can be strong enough to cable, but the cable is easily shielded by
mask the true operating signals. providing an opaque covering.

49
Cache

Breaks and kinks in high-speed cables are Summary of potentially negative issues
also much more trouble. At low speed, a The most common type of cable covering
break in a conductor simply results in no is PVC casing and, should it catch fire, the
signal getting through, an unfortunate con- fumes can be hazardous. Many fire codes
dition, but one that is easily detected. At require that a special form of cable known
high speeds, the effect of self-interference as ‘‘Plenum rated” cable be used, which is
can allow some attenuated portion of the fire-resistant. This is especially important if
signal to cross the break, making detec- the cable is to be strung in the overhead
tion and repair more difficult. A kink in spaces and recesses above offices and not
a cable can cause a short-span change in enclosed in special ducting. Plenum cable
impedance (the cable’s resistance to the is more expensive than PVC-cased cable and
flow of electricity); at low speeds, this sim- budgets need to be adjusted to take account
ply reduces the signal strength slightly; of this expense.
at high speeds, it can cause reflections
and echoes of the signal that drastically Associated terminology: Ethernet,
increase the overall error rate, and reduce Bandwidth.
throughput to a fraction of its expec-
ted value. With coaxial and twisted-pair
cables (as used in broadband network con- Cache
nections), and especially with fiber-optic
cable, it is essential never to kink or spindle Foundation concepts: Storage, Memory, Disk.
the cable or force it into sharp turns, and, Definition: A small but frequently accessed part of a
unless communications speed is not impor- data set, kept conveniently at hand for rapid access.
tant, cables that have been kinked but still
seem to work should probably be replaced. Overview
Carelessness with cables can be an expen- With any kind of data retrieval system,
sive mistake in the long term. there are likely to be some data items
that can be identified as more likely to be
needed than others. For example, a corpo-
Business value proposition
rate web server may store thousands of
A few extra dollars for a well-designed,
pages, any of which could be accessed at
carefully manufactured cable or connector
any time, but the corporation’s ‘‘home page”
can easily save thousands of dollars in
and its search page will probably be acces-
‘‘troubleshooting” and repair costs, and
sed far more frequently than any others.
incalculable amounts in saved down-time.
When retrieval of data requires some
Flimsy-looking metal covers should never
time or expenditure of computational
be removed from computer parts; they
effort, it makes sense to set aside these
may be too flimsy to provide any protec-
most frequently accessed items in some spe-
tion from physical damage, and it may
cial location that allows faster access if
seem sensible to increase ventilation, but
such a place exists. The set-aside store of
metal covers provide essential electrical
frequently accessed entries is called a Cache.
shielding.
In a system where normal access is
through a network connection, a cache
Summary of positive issues may sensibly be kept on a local disk. For
A wide range of cables with a variety of example, a web browser usually keeps a
bandwidths and associated connectors is cache of recently visited pages, so that
available. pressing the ‘‘back” button does not require

50
CAD/CAM (computer aided design/computer aided manufacturing)

any network access. In a system where nor- larger one is superior. Only a complete mea-
mal access is to data on a disk, a cache is surement of the system’s speed in use will
likely to be kept in memory. For example, be meaningful.
a file-processing system may keep copies of
recently accessed files in memory, because Summary of positive issues
it is a common pattern to keep working Caches allow computers to act more quickly
on the same file until some task has been than would otherwise be possible.
completed.
Web search engines often keep a cache
of the results of popular searches, keeping
Summary of potentially negative issues
The use of a cache adds slightly to a
dynamic statistics so that changes in pop-
system’s complexity, but it is a well-
ularity (as with people frequently access-
understood technology, and is unlikely to
ing web sites that have been in the news
have any significant negative effects.
recently) can be taken into account. A
database system will often keep a cache Reference
of recently accessed records, or important r J. Handy (1998). Cache Memory Book, 2nd
indexes. edn. (San Diego, CA, Academic Press).
Caches also play an important part in
Associated terminology: CPU, Bus.
hardware design. A disk drive controller
will usually contain a few megabytes of
memory, which is used as a cache to store
frequently and recently accessed data, and, CAD/CAM (computer aided design/
because memory accesses are so much computer aided manufacturing)
faster than disk accesses, the presence of
such a cache speeds up many common Definition: Computer aided design and computer
tasks by significant factors. Computer CPUs aided manufacturing are two computer-based sys-
and motherboards usually contain a cache tems that provide a basis for product modeling and
made of exceptionally fast memory, so that automated manufacturing.
a similar speed-up can be achieved for
normal memory accesses. Overview
Computer aided manufacturing (CAM) as we
Business value proposition know it today evolved from early computers
The use of caches is an important mecha- that used paper tape to issue instructions
nism through which a computer’s abilities to numerically controlled devices (NCDs).
can be extended. A web server that stores An NCD is controlled by a machine-control
popular corporate web pages in a cache will unit (MCU), which is composed of two
be faster than one that has to read repeat- parts: a data-processing unit (DPU), which
edly from backing storage; similarly, a com- reads and processes the instructions in the
puter that stores a user’s previous web page control program, passing the instructions
access in a cache will give faster access to on to the other part, the control unit
it than will one that doesn’t. The cache is (CU), which converts those instructions
built into the system’s design and, for soft- into electro-mechanical control commands
ware systems, it is usually an easy matter to for the machine itself. In the early days
reconfigure the size. Hardware caches are this required special languages for each
not normally at all flexible. It is a mistake device because each manufacturer used a
to compare cache sizes for one CPU or disk different instruction set. This problem was
drive with another, and conclude that the overcome when, in 1958, a language was

51
Capacity planning

created by researchers at MIT to tackle it. with the styling processes rather than the
The language, known as APT, allowed pro- engineering processes that would underlie
grammers to write in a standard code and a design. These would be considered in the
then, through NCD-specific Compilers, trans- CAD stage of the development which fol-
form the code into instructions that the lows the CAS stage.
NCD machine would understand. The CAD/CAM systems allow organi-
The development of Computer aided design zations to capture their organizational
(CAD) is generally acknowledged to have knowledge within the systems and repro-
started when Ivan Sutherland created a sys- duce aspects of a design with low cost in
tem known as Sketchpad at MIT in 1963. future designs, thus avoiding the syndrome
Sketchpad was an interactive graphics app- of ‘‘reinventing the wheel.”
lication that utilized one of the first graph-
ical user interfaces and a Light pen to enable Summary of positive issues
programmers to input designs. During the CAD/CAM provides a fast, cost-effective
next twenty years the CAD research and mechanism to develop designs. The results
vendor community developed more sophis- of the system are precise and repeatable.
ticated applications to enable designers or The systems allow 3D images to be created,
programmers to create wire-frame repre- scaled, manipulated, and viewed. The sys-
sentations of objects, view models as solid tems can be linked to VR, CAS, and other
objects, rotate and scale objects, view cross tools used in the concept-to-production
sections of objects, etc. lifecycle.
As the CAD systems evolved it was natu-
ral for the output from CAD systems to be Summary of potentially negative issues
funneled into existing CAM systems, allow- The cost of the software itself together with
ing the production and design issues to the required training can be considerable.
be examined. CAD/CAM systems have been The integration of the components, the
used by organizations to design and manu- CAD and the CAM, needs to be carefully
facture nearly every type of product, with considered to ensure that the systems can
automotive and aeronautic systems being communicate with each other using com-
amongst the early adopters and driving patible, neutral data file formats such as
forces behind much of the technological the ANSI Initial Graphics Exchange Specifi-
progress. cation (IGES) standard.
Reference
Business value proposition r D. Schodek, M. Bechthold, J. Griggs, K.
CAD/CAM has the major advantage of Kao, and M. Steinberg (2004). Digital
allowing designers and production engi- Design and Manufacturing: CAD/CAM
neers to examine their models in virtual Applications in Architecture and Design
space without the intense commitment of (New York, John Wiley and Sons).
resources and time that traditional meth-
ods require. These technologies are also Associated terminology: Virtual reality.
tied in with other types of systems such as
virtual reality (VR) modeling and computer
aided styling (CAS) to enable designers to Capacity planning
extend the design processes, both by view-
ing their systems in more realistic ways and Definition: Capacity planning is the process of deter-
with industry design-specific software. For mining current and future resource requirements
example, the use of CAS in the automo- and developing cost-effective solutions to meet those
bile industry focuses on helping designers needs.

52
Capacity planning

Overview used in capacity planning since the prob-


lems are usually multivariate in nature.
Capacity planning can be considered as the Metrics including quality of service (QoS)
mechanism through which systems man- and end-to-end response time may also be
agers develop their systems to meet future important for a network manager in deter-
demands. Capacity planning has been in mining future peak load requirements. The
existence since the advent of corporate overriding concern of the capacity planner
computing, when data processing man- is to ensure that the network does not run
agers would work to understand the per- out of capacity prior to the next upgrade
formance capabilities of the mainframe sys- point.
tems that they were considering buying or Capacity planning can be undertaken
leasing, matching that capability with a for any systems component (physical or
forecast of the company’s future processing human), e.g., backup data storage capacity,
requirements. server performance, help-desk service lev-
The basis of capacity planning is the els, and local storage in clients. Recently
balance between two factors: cost of ser- there has been much interest in the use of
vice delivery and capacity. Clearly, with an capacity planning techniques in the area
unlimited budget the capacity to satisfy of the web and web services. While the
almost any service level could be reached. application of techniques and methodolo-
Unfortunately, apart from special projects gies to these areas is fundamentally the
at government funded research agencies, same as measuring CPU loads on main-
few organizations have access to these frames, in reality the complex nature of the
levels of resources, so the task of capacity technology that delivers the web services
planning becomes a trade-off between the and their interactions makes capacity plan-
quality of the service delivered and the cost ning in these areas much more demanding.
of the resources available that provide that
service. Business value proposition
To undertake the capacity planning task, The development of a capacity planning
planners have two main tools: metrics and model for an IT-based organization is one
statistics. Metrics are used to measure past of the most critical projects that a CIO
performance, for example network traffic can undertake. The CIO must build the
may be measured in terms of bandwidth model in order to be able to present accu-
utilization over the entire network, histor- rate capital budget requests. The CIO also
ically 24 hours a day for a year. This would needs to be able to look forward and back-
enable the network manager and capac- ward in time. Looking back in time requires
ity planner to determine statistically the that the organization has established and
average usage, absolute maximum usage, enacted data collection programs that pro-
average maximum and average minimum vide accurate data for desired metrics.
usages, and other metrics. These statistics These form the basis of the projections that
would enable a utilization analysis to be CIOs and their staff make to determine
undertaken and a determination of current future system utilization rates and pro-
capacity limits to be established. ject upgrade points. However, these up-
Having established the potential peak grade points will be based upon a set of
loads that could be placed upon the exist- corporate operating assumptions. For exam-
ing system, the future workload for the ple, projected system loads may be based
network may be projected and the next upon a workforce growth of 1% per year
upgrade point established. Of course, it for the next three years and the assump-
would be rare for one metric alone to be tion that each user will have the same

53
Cell computing

systems-workload demands. Such projec- r D. Menasce and V. Almeida (2002).


tions would make sense in the case of a Capacity Planning for Web Services: Metrics,
stable established company’s IT help desk, Models and Methods (Englewood Cliffs,
but, if the company has a wide range of NJ, Prentice-Hall).
systems with highly demanding uses and
Associated terminology: CIO.
is experiencing hyper-growth in employee
numbers (such as an internet start-up com-
pany in the late 1990s), these conservative
Cell computing
projections will be wrong and potentially
harmful to the company. Also known as: Grid computing.
Thus, CIOs require access to strategic Foundation concept: CPU.
decision making at the executive level of Definition:Theuseofaverylargenumberofveryweak,
the organization. They can contribute to low-power interconnected computers instead of one
the process as well as obtaining planning fast, high-power one.
and forecast information. The result is that
better technology planning decisions can Overview
be made. The research and development effort
expended on producing ever more power-
Summary of positive issues ful computers is staggering, and stagger-
Capacity planning enables organizations to ingly successful. Roughly every five years,
provide a stable basis for their IT organiza- top-of-the-line computers are ten times
tion and provide a platform on which cor- faster than they were before. There is no
porate strategy may be executed effectively. end in sight for this rapid development,
Capacity planning is a well-understood aca- but, as computers get bigger and faster,
demic discipline that has substantial liter- their power requirements increase dispro-
ature for practitioners and academics. Per- portionately, design considerations become
formance monitoring tools are built into much more complex, and the chip fabrica-
many systems and can provide data for tion process becomes more delicate. Natu-
capacity planning tools built by specialty rally, alternatives are sought.
vendors as well as by the major software The idea behind cell computing is very
vendors. simple: a hundred small, cheap, slow com-
puter processors working together could
Summary of potentially negative issues easily outperform one very expensive, very
The general metrics available from systems fast computer. In simple terms, a CPU that
might not always be the most appropriate, was state-of-the-art five years ago now only
so special metrics packages may need to has one tenth of the power of the current
be created or purchased. Demand forecast- models, and as a result may be purchased
ing can be complex, especially when data is very cheaply. Buying a hundred of them
limited or in newly emergent areas. Trans- and making them work together could pro-
lating the impact of strategic decisions vide ten times the computing power of the
onto systems and developing forecasts from fastest models available today.
them is a difficult process and requires Of course, the reality is not quite as sim-
strong statistical and technical skills on the ple as that. If it takes a person 40 minutes
part of the capacity planner. to drive to work in the morning, that does
References not mean that four people could drive to
r T. Browning (1995). Capacity Planning for work in 10 minutes. Some tasks in com-
Computer Systems (New York, Academic puting, just as in real life, have an essen-
Press). tially serial nature, and can not be sped up

54
Cell computing

by parallel processing. However, many com- In the 1980s, the semiconductor manu-
mercially useful computational tasks do facturer Inmos, in conjunction with the
have a very parallel nature. For example, if designers of CSP and Occam, produced
1 000 000 000 un-indexed documents stored and marketed a self-contained single-chip
on disks are to be searched for those that computer, called the Transputer, which
contain some key phrase, the task would was designed to support cell computing
take a very long time, at least a day, and systems. Although it was technically suc-
up to a month, depending on the sizes of cessful, the Transputer never approached
the documents. Two computers could per- the low prices that full-scale cell comput-
form the task in exactly half the time; ten ing requires, and was not a commercial
could do it in exactly one tenth of the time. success.
Some tasks have a very serial nature, Wolfram claims to have developed a
some have a very parallel nature, and most whole new basis for science, built on
are somewhere between the two extremes. the ideas of cell computing, in his self-
Gene Amdahl, in 1967, analyzed the sit- published 2002 book; the book is to say
uation and the result is now known as the least controversial, but does add to
‘‘Amdahl’s Law”; it is still the most fre- the development of the subject. These
quently cited reference for the speed-up of ideas grew originally from a computer
processing obtainable through the use of game/simulation known as Conway’s Game
multiple processors. of Life, which was invented by the mathe-
Cell computing really comes into its own matician J. H. Conway in 1970.
at a different scale, not using a few dozen The practical implementation of cell
obsolete computers picked up cheaply, but computing would have immediate appli-
using many thousands, perhaps even mil- cation to high-power graphics production
lions, of purpose-built computers. The Pen- (one computing cell responsible for every
tium 4 (introduced in 2001) has nearly pixel of the screen), and a brief inves-
45 000 000 transistors on a single chip; the tigation of this possibility reveals the
Intel 8080 microprocessor (introduced in important fact that some programming
1974) had only 4500 transistors, but was problems actually become much easier if
still computationally useful. It is already they are designed for large-scale cell com-
possible to construct a single chip that puting instead of the traditional uniproces-
contains 10 000 individual computationally sor computers. It is also expected that cell
useful computers. computing would be very helpful in mod-
Software development for cell comput- eling physical phenomena, especially the
ing would require major retooling in the atmospheric models that support weather
industry. The ability to design a program forecasting. It is equally clear that the full
for thousands of slow processors is a quite benefits of cell-computing will not be fully
different skill from those normally pos- known until cell-computing hardware is
sessed by software engineers. Unusually, generally available.
the area has been very well investigated; It is known that the Neuron, the basic
C. A. R. Hoare of Oxford University thor- brain cell, has exceptionally simple func-
oughly developed the computational the- tionality that is easy to duplicate with very
ory in a series of papers on communicating inexpensive circuitry. A neuron is incapable
sequential processes (CSP), culminating in of performing any calculation or harboring
his influential 1985 book. The program- any thought alone; it is only through the
ming language Occam implements the vast number (over 100 000 000 000) of neu-
essential parts of CSP and is available for rons in the brain, and their complex inter-
most programming platforms. connectivity, that intelligent thought is

55
Chaos theory

possible. The possibility of being able to fab- multiple processors programmed in tra-
ricate a brain’s worth of artificial neurons ditional ways. Several vendors are work-
is still in the very distant future, but the ing on and deploying new forms of cell
undeniable success of real neurons fuels technology.
copious academic research toward what is
perhaps the ultimate application of cell Summary of potentially negative issues
computing. The use of large-scale communicating pro-
Recently there has been some revival of cesses using a cell processing model is
interest in cell computing in the main- largely confined to the research laboratory
stream of the computing industry, in the and will require extensive research to
form of IBM’s CELL project, or broadband bring the technologies into the commercial
processor architecture. This has been fueled realm.
by the desire for ever more realistic graph- References
ics in video game consoles, which are amon- r G. Amdahl (1967). The validity of the
gst the most computationally demanding single processor approach to achieving
computer applications in common use. large-scale computing capabilities,”
in Proceedings AFIPS National Computer
Business value proposition Conference (Anaheim, CA, AFIPS),
The commercial potential for cell comput- pp. 483--485.
ing is undeniable, but the current reality r C. A. R. Hoare (1985). Communicating
of the technology is that it lags consider- Sequential Processes (Englewood Cliffs, NJ,
ably behind theory. Industry has for the last Prentice-Hall).
few decades pursued the goal of putting r G. Jones (1987). Programming in Occam
more and more transistors onto a single (Upper Saddle River, NJ, Prentice-Hall).
chip rather than working on simpler chips r S. Wolfram (2002). A New Kind of Science
that can communicate with each other in (Champaign, IL, Wolfram Media).
large numbers. As the limits of technology
are reached, it will become more impor- Associated terminology: Parallel
tant for the extremely high density chips processing.
to communicate with each other and sub-
divide the tasks to which they are set in
order to continue to make gains in perfor- Chaos theory
mance. The move to cellular systems will
also require a major change in the style of Definition:Achaoticsystemisonethatmayhavebroad
programming and the development of new areas of stable predictable behavior, but also has areas
software systems to support the technology in which immeasurably small changes in the initial
and its environment. conditions produce significant changes in the results,
thus making long-term predictions impossible.
Summary of positive issues
The theory of communicating sequential Overview
processing is well known and understood Contrary to beliefs current in popular cul-
in the computer science research com- ture (Jurassic Park to be precise), chaos the-
munity. Cell computing has been develo- ory does not predict that complex systems
ped in the past and applications based will fail. It has absolutely nothing to say on
upon those technologies were developed. the subject.
Amdahl’s Law governs the processing per- The truth about chaos theory is much
formances obtainable through the use of more interesting and surprising. Since

56
Chaos theory

Newton postulated the basic laws of physics ably small change in conditions makes the
in the mid seventeenth century, people difference. These catastrophe points always
have generally believed that the universe exist; no matter how accurately condi-
works like a big machine: everything that tions are measured, there are always transi-
happens is the direct result of other things tional points at which that accuracy is not
that happened before. If you know how the enough to predict correctly a major yes-or-
machine is initially set up, you can predict no event. If a forecast says ‘‘no hurricane”
exactly what it will do in the future. when there is a hurricane, then, of course,
Weather forecasting is a good example. everything else is going to be completely
The underlying science of the weather is wrong too.
well known: how a small volume of air, It is generally accepted that other sys-
land, or sea behaves when temperature, tems of a similar nature, namely interact-
pressure, or other attributes vary, and how ing components with conditions that can
neighboring small volumes of air, land, or not be perfectly controlled or measured,
sea interact with each other are all under- can be expected to exhibit chaotic behavior
stood. Until quite recently it was believed and resist long-term forecasts. The behavior
that if data could be gathered, accurately of human or animal populations and stock-
measuring the current state of the entire market prices are prime examples.
atmosphere, a sufficiently powerful com- The fine boundary between sets of con-
puter could use the known equations of ditions with drastically different outcomes
physics to determine how the system would is known to science as the Julia set
evolve over time, and produce accurate (after Gaston Julia, French mathematician,
long-range weather forecasts. 1893--1978), and is often a fractal (grossly
It was realized from the beginning that simplified, a fractal is a complex shape
imperfect measurements would produce whose length can not be measured). The
imperfect forecasts, but it was quite rea- most famous of these is the Mandelbrot
sonably expected that good-enough mea- set, which is produced from a very sim-
surements would produce good-enough ple mathematical formula, x  = x 2 + c, and
forecasts, and gradual improvements in demonstrates that even exceptionally sim-
measurement technology would produce ple, completely known mathematical equa-
corresponding improvements in forecast tions can exhibit chaotic behavior.
accuracy.
It is now known that this is not the Business value proposition
case. In many scenarios it does work out The study of chaos theory and fractals has
quite well: small errors in measurements led to the abandonment of projects now
produce only small errors in the forecast, known to be infeasible, the ability to pro-
allowing rain and drought to be predicted; duce realistic graphical renderings of nat-
these are known as stable conditions. ural scenes from very little data, and the
Unfortunately, unstable conditions abound. discovery of some exceptionally effective
Consider the formation of a hurricane: data compression methods.
there are conditions under which a hurri-
cane will form, and there are conditions Summary of positive issues
under which it won’t. Between those two The study of chaos theory has revealed
sets of conditions, what happens? Initial some fundamental truths of the universe
conditions are infinitely variable, but a hur- and produced some useful techniques as a
ricane either forms or it doesn’t. There is by-product, rather in the manner of NASA
some point at which the tiniest immeasur- moon missions being credited with the

57
Chaos theory

The famous Mandelbrot set and a related fractal.

development of new technologies that have


degrees of reliability. Research into the
found their ways into everyday life, except
use of chaos theory and fractals is very
that there is a rather different level of
demanding and theoretical, and is unlikely
expenditure involved.
to be successfully carried out by anyone
lacking a Ph.D. in mathematics and a great
deal of time.
Summary of potentially negative issues
Claims of the general applicability of chaos References
theory to problems such as modeling the r B. Mandelbrot and R. Hudson (2004).
stock market to predict future performance The Misbehavior of Markets: A Fractal View
need to be considered carefully, since such of Risk, Ruin and Reward (Cambridge,
models rarely produce results with high MA, Perseus).

58
CIO (chief information officer)

r M. Barnsley (2001). Fractals Everywhere web-site operators, and the parents, and
(San Diego, CA, Elsevier). their interactions. The act aims not to be
r G. Williams (1997). Chaos Theory Tamed onerous and has special provision for cases
(Washington, DC, Joseph Henry Press). of limited interaction (such as a site receiv-
ing an email from a child to which only a
single response is needed). In such a case
parental notice and consent are not requi-
Children’s Online Privacy Protection
red as long as all personal information
Act of 1998 (COPPA) obtained from the child is deleted upon
Foundation concept: Security. execution of the email.
Definition:TheUSChildren’s OnlinePrivacyProtection
Act of 1998 is intended to “prohibit unfair or deceptive Summary of potentially negative issues
acts or practices in the collection, use, or disclosure A problematic aspect of the act has been
of personally identifiable information from and about the mechanism through which ‘‘verifiable
children on the internet.” parental consent” is obtained. At the out-
set of the act the final rule devised a
Overview ‘‘sliding scale” that allowed the methods
The US Children’s Online Privacy Protection used by web sites to ‘‘be reasonably cal-
Act of 1998 (COPPA) (Title XIII, Section 1301, culated in light of available technology, to
1998) defines the privacy policies to which ensure that the person providing consent is
‘‘web sites or online services” aimed at chil- the child’s parent.” These methods include
dren under 13 must adhere. Key provisions postal mail, ‘‘print-and-send” faxes, credit-
of the act are the placement of prominent card usage, toll-free verification, and digital
policy notices on web sites; the notification signatures.
of parents that a site is going to collect References
personal information; the mechanisms for r The Federal Register, March 15, 2006,
obtaining parental consent prior to a site Part III, Federal Trade Commission,
obtaining or collecting personal data; to 16 CFR Part 312, Children’s Online
disallow collection of data as an incentive Privacy Protection Rule; Final Rule.
for further involvement in the site by the r 15 USC §§6501--6505 (2006).
child; to allow parents to review informa-
tion collected on their child and have that Associated terminology: Law
information deleted if they wish; to allow cross-reference.
parents to prohibit an online service or web
site from collecting any further data on
their child; the role of schools in providing CIO (chief information officer)
permission; the dissemination of data to
third parties; and the procedures by which Foundation concept: MIS.
online services maintain and delete data. Definition: The chief information officer is a senior-
level executive responsible for all corporate informa-
Business value proposition tion systems.
The act requires the operators of web sites
aimed at children of 12 and under to oper- Overview
ate within a defined framework. The role of a Chief Information Officer (CIO)
is to ensure that the systems and technol-
Summary of positive issues ogy within a company are aligned to the
The act is comprehensive, defining the company’s strategy and that all systems are
responsibilities of the service providers, the compliant to government regulations. The

59
Click-stream tracking

CIO is a member of the executive layer who transformers. As such, a visionary CIO who
typically reports to the CEO and the board can also execute technology solutions has
of directors or through the office of the become a major corporate asset as both
CFO. technology and corporate strategies con-
The CIO as head of the IT organization is tinue to evolve rapidly.
an integral part of the corporate strategic
Reference
planning group and helps formulate corpo- r M. Earl, M. Feeney, and D. Feeney (1994).
rate strategy by bringing an understanding
‘‘Is your CIO adding value?,” Sloan
of the role of technology within the com-
Management Review, Volume 35, Issue 3.
pany’s marketplace, the impact that new
technologies will have upon the industry Associated terminology: Sarbanes--Oxley
in which the company operates, and the Act.
technology-related regulations that impact
the environment in which the company
operates (e.g., HIPAA, Sarbanes--Oxley). Click-stream tracking
CIOs manage departments that encom-
pass complex business processes and tech- Foundation concept: e-Commerce.
nologies that support business processes. Definition: Click-stream tracking systems are appli-
Issues with which the CIO has to contend cations that monitor web site accesses and compile
include business process re-engineering, data based upon visitor activity.
systems developments, maintaining legacy
systems, training, technology assessment, Overview
developing contingency plans, ensuring A click stream defines a sequential list
legal compliance, technology deployments (‘‘stream”) of activities (also known as
(ERP systems, robotic systems, etc.), tech- ‘‘clicks” or ‘‘hits”) that an individual visitor
nology outsourcing, and technologies that performs on a web site. Click-stream track-
support specific functional aspects of the ing systems monitor the click streams, con-
business (e.g., customer relationship man- solidating the data and generating reports.
agement systems). The applications that monitor click
streams range from those that come as
Business value proposition part of an internet server package, to cus-
The role of the CIO has changed consider- tom solutions from vendors who specialize
ably over time as IT organizations within in tracking and monitoring internet and
companies have increased in visibility and network traffic. Systems typically offer a
value. MIS departments were, up until the variety of data analysis tools to monitor a
mid 1980s, frequently considered as data wide variety of parameters, including met-
processing operations, run as cost centers rics such as the total number of visits to
within the organization by data process- the site, total number of pages viewed, total
ing managers. Subsequently, MIS depart- number of hits to the site, total number of
ments became recognized as value centers unique visitors, total number of new visi-
providing ‘‘enabling” technologies to the tors, total number of repeat visitors, total
core business processes and offering the usage number, average number of visits
potential for competitive advantage to be per day (week, month, etc.), time period
gained through their strategic deployment. of highest activity levels (hour, day, month,
During this period data processing man- etc.), average length of visits, average num-
agers became replaced by CIOs, who began ber of times a site is visited by an individual
to view themselves not only as organiza- (assuming that the IP address is known),
tional enablers but also as organizational most frequent visitor, domain, host, or user,

60
Client

most popular content accessed, most pop- Summary of positive issues


ular paths, most frequent entry and exit Click-stream analysis is supported by a vari-
points, most frequent keyword search, most ety of application vendors and consultants.
popular links, most popular browser used The data collected can be customized to
to access the site, most popular platform, a particular organization and its require-
most popular screen resolution, error anal- ments, which can be a technical analysis
ysis, and server response analysis data. of their site (e.g., performance-related) or
Raw data on page-visit sequences is typ- marketing-related (e.g., which aspect of the
ically accumulated in session cookies (q.v.), site is popular and which is not).
which are updated and exchanged between
browser and server on each access, and
Summary of potentially negative issues
would not contain any personal data.
Click-stream analysis is often perceived as
Occasionally, hidden input elements and
related to spyware and other invasive soft-
javascript methods may be used as an
ware. However, when it is used in isola-
alternative implementation method.
tion this is not the case, since click-stream
analysis is passive data collection and all
Business value proposition activity occurs on the web server. Click-
The click-stream analysis packages allow
stream analysis is, however, frequently com-
corporations to monitor web traffic pro-
bined with data collected at the client
files and reposition their online strategy
or customer site through the use of less
accordingly. They can, for example, use
legitimate software.
applications that perform marketing anal-
ysis of their online channel. This can be Reference
performed by categorizing specific aspects r M. Sweiger, M. Madsen, J. Langston, and
of the web site as belonging to a particu- H. Lombard (2002). Clickstream Data
lar product segment, e.g., toys, books, and Warehousing (New York, John Wiley and
videos. These can then be monitored and Sons).
analysis performed to assess the impact
Associated terminology: Cookies, Spyware.
of marketing campaigns, for example the
impact of an advertisement in a partic-
ular search engine. Typical channel-based
metrics would include keyword searching, Client
hyperlink referrals (where the customer
came from prior to entering the site), robot Foundation concepts: Network, Software.
or spider activity, and customer technol- Definition: Clients are computing devices through
ogy profiling (e.g., their browser, platform, which end users interface with a network-accessible
screen resolution, and ISP). system. They do not in themselves provide network
The more sophisticated analysis suites resources.
can monitor browsing patterns, keeping
data on the sequence of pages viewed by Overview
site visitors. This information enables site A client can be any device through which
designers to check the quality of their an end user interacts with a networked
layouts and ensure that visitors are able computing environment. An example of
to find what they are looking for with- this type of device is a personal computer
out getting too annoyed. It can also help with a web browser that acts as a front end
informational sites to identify bottlenecks, to the resources of a network. Through the
and group frequently co-visited pages more client, the user interfaces with the network
efficiently. to download files, browse resources, or use

61
Client–server

an application such as instant messenger Summary of potentially negative issues


to communicate with another end user. Failure to understand the total cost of own-
Clients request services from other devices ership at the end-user level can lead to
on the network; those devices which deliver excessive recurring systems expenses and a
them are commonly called Servers (file great waste of resources.
servers, web servers, application servers,
etc.). References
r W. Stevens (1994). TCP/IP Illustrated,
Protocols, or intermediate languages of
communication, are designed to enable Volume 1 (New York, Addison-Wesley).
r R. Orfali, D. Harkey, and J. Edwards
clients to interface with servers regardless
of any operating system or software incom- (1999). Client/Server Survival Guide, 3rd
patibilities between the two: a Windows edn. (New York, John Wiley and Sons).
r J. Chellis, C. Perkins, and M. Strebe
client can successfully interact with a Unix
or Macintosh server because all communi- (2000). MCSE: Networking Essentials, Study
cations are through a system-independent Guide, 3rd edn. (Alameda, CA, Sybex
protocol. Of course, it is still necessary to Press).
have a client that can handle the appro- Associated terminology: Server,
priate protocols; a web-browsing client is Client--server.
unlikely to be able to communicate with a
database server, but the client/server divi-
sion prevents a lot of potential problems. Client–server
See Client--server for further details.
Foundation concepts: Client, Networks, Server.
Business value proposition
Overview
The client concept allows corporate net-
When computers are connected together
works to interact with many different types
on a network, it is common practice to
of device, from a large powerful worksta-
localize certain functions. Prior to the
tion on which business intelligence soft-
advent of local area networks and wide area
ware resides, and to which data is regularly
networks, a typical business office environ-
downloaded from the network for local
ment would have encompassed a number
analysis, to a personal digital-assistant-type
of general purpose computers, or worksta-
device through which employees can gain
tions, upon which would have been per-
access to email on a continuous basis.
formed self-contained tasks (word process-
Clients allow greater flexibility of connec-
ing, spread-sheet processing, etc.); should
tivity to corporate computing resources
there have been a need to share files, they
than does the traditional Dumb terminal
would have been transported via floppy
configuration that previous generations of
disks. It soon became apparent that this
systems deployed. Clients allow individual
was a limitation on workflow and that the
end-user cost-of-ownership decisions to be
ability to share files was central to organiza-
made and increase the potential for a
tional efficiency. This prompted the devel-
greater return on investment in system
opment of networking. The first evolution
architectures.
was in terms of a peer-to-peer configura-
tion whereby everyone in an office would
Summary of positive issues be connected to everyone else. This resul-
Clients facilitate flexibility of network ted in connectivity problems due to the
architectures at the end-user level. complexity of the networks required and

62
Client–server

the security relations between computers a request. This is in direct contrast to the
within that network. Thus it became clear peer-to-peer organization, in which the two
that the computers within an organization communicating systems have the same sta-
needed to be connected through a central- tus. With a client--server system, it is impor-
ized computer, the server, or a group of tant that the server(s) have known, rela-
servers acting in a coordinated manner yet tively static locations (usually IP addresses),
controlled by a central primary computer. but clients can be very mobile.
Then, when co-workers need to share a file, The client--server relationship does not
the file in question is copied from the file apply only to whole computers. Any given
server, over the network, onto one of the application that has network access may
client workstations where it is to be used. act as a server or as a client, or even
If any modifications are made, the changes both. On any computer at any time, there
are sent back to the file server so that all could be a number of applications acting as
other workers will be able to access the new servers, passively waiting for requests, and
version. a number of applications acting as clients,
Similarly, when email arrives from out- driven by the user actively making connec-
side the organization, there needs to be a tions with servers. FTP is an example of an
designated computer that is always online application that acts both as a server and
and ready to receive it. When an associate as a client concurrently.
wishes to read their email, their client
workstation makes contact with the email Business value proposition
server, and retrieves all relevant unread The three-tier client--server architecture is
messages. by far the most common computing con-
Having one computer (or more) spe- figuration found within organizations. It
cially designated as ‘‘the server” means that is predominantly centralized in nature,
important files can be kept safe in one with flexibility of client location and access
central location that is carefully protected, through the networking protocols utilized
and there is always an online presence to connect the client to the server (TCP/IP).
to receive email and handle web services, The network response can be improved
regardless of what the various workstations on large-scale global client--server networks
may be doing. Although the server is prob- by defining groups of clients and having
ably the most important of the comput- mirror-image or secondary domain control-
ers, its role is mostly passive. Client--server lers administer the users of those clients.
architectures are also known as three-tier The primary and secondary controllers co-
architectures from the fact that the clients coordinate in order to remain synchronized
(tier one) request information from the pri- in terms of user profiles, security levels,
mary server (tier two) and they in turn file structures, etc. This ability to group
request the service from the specialized users into domains and have localized con-
servers (tier three), which provide a service trol yet global access to resources has made
such as keeping files, emails, or web pages; this the standard choice for the majority
they respond only to requests from the pri- of organizations because it simultaneously
mary server. facilitates security and flexibility of the
This is by far the most common organi- network.
zation for network services. When commu- Several variations on the nature of client--
nications are made, it is because one com- server architectures exist, two major cate-
puter or system was passively waiting for gories of which are Thin client and Fat client.
requests, and another actively made such In thin-client architectures the client itself

63
Clone

is a computer with little or no local pro- References


cessing power, displaying only the images r W. Stevens (1994). TCP/IP Illustrated,
sent to it by the server and being capable Volume 1 (New York, Addison-Wesley).
of sending requests to the server input r R. Orfali, D. Harkey, and J. Edwards
by the end user. Fat clients, on the other (1999). Client/Server Survival Guide, 3rd
hand, are computing devices with process- edn. (New York, John Wiley and Sons).
ing power at the client side of the relation- r J. Chellis, C. Perkins, and M. Strebe
ship. This may facilitate the client being (2000). MCSE: Networking Essentials, Study
able to process data rather than making Guide, 3rd edn. (Alameda, CA, Sybex
the server process any and all data as Press).
would be the case in the thin-client archi-
Associated terminology: Network, Server,
tecture. Similarly, fat clients may also have
Client, TCP, Internet protocol.
applications residing upon them and this
also allows local processing of data upon
that application; thin clients would not
Clone
have this facility. The value proposition
for the business is that this architecture Foundation concept: CPU.
allows the organization to determine the Definition: A complete computer or just a CPU, built by
total cost of ownership of their comput- anindependentmanufacturer,butconformingexactly
ing resources in a number of configura- to the specification of a “famous brand,” usually the
tions and develop a best case for return on IBM PC or the Intel 80486, to allow direct substitution
investment. Furthermore, the majority of of the product.
large software systems such as ERP systems
are now written for client--server architec- Overview
tures and vendors are withdrawing sup- The term ‘‘PC” was originally a completely
port for mainframe or stand-alone-based generic term meaning ‘‘personal computer”
architectures. without reference to any particular man-
ufacturer. The term was in common use
many years before IBM produced its first PC
Summary of positive issues in 1981. However, IBM cleverly incorpo-
Client--server three-tier architectures facil-
rated the letters PC into their product
itate flexibility of architecture design
names (‘‘IBM PC/AT,” ‘‘IBM PC/XT,” etc.), and
and map comfortably upon organizational
since that moment the phrase PC has been
structures. This type of architecture pro-
taken almost universally to mean ‘‘personal
vides centralized control of the network,
computer designed by (or compatible with)
enhances its security, and is scalable. Soft-
IBM.”
ware vendors use the client--server architec-
Once IBM’s line of computers had achie-
ture as their standard model upon which
ved total market domination, it was inevi-
they develop their software.
table that others would want a share of
the vast profits available. At first, this resul-
Summary of potentially negative issues ted in many ‘‘third-party” software houses:
Client--server architecture requires exten- companies that do not produce computers
sive knowledge to manage effectively. at all, only software or applications. Soft-
Failure to understand the total cost of own- ware written for one type of computer will
ership at the end-user level can lead to very generally not work at all on another type of
expensive recurring systems expense and a computer, so software manufacturers con-
waste of resources. centrated their efforts on the most popular

64
Cluster

platform: the IBM PC. IBM-PC-compatible For most desktop/workstation uses, the dif-
software became as popular as IBM PCs. ferences between manufacturers and CPU
This gave rise to the ‘‘Clone.” A clone is families are almost irrelevant, so, although
a computer that is exactly compatible with the variety of choices may seem bewilder-
an IBM PC, so completely compatible that ing, it is not really a problem.
it can run all software designed for an IBM
PC without modification, and thus may be Summary of potentially negative issues
bought instead and substituted directly for Many very cheaply made computers simply
the more well-known product. do not work properly.
The IBM-PC design is based on the Intel
microprocessor architecture. Any CPU that
behaves exactly as the corresponding Intel Cluster
CPU would behave may be used in its place.
Foundation concepts: Server, Network, Disk.
Thus the second type of clone: a micropro-
Definitions:
cessor CPU not made by Intel, but designed
to be directly substitutable on PC moth- 1. A group of computers configured to act as one.
erboards. Many of the product lines that 2. A group of blocks on a disk.
began as Intel clones have developed into 3. Agroupofsimilardataitemsthatmayberelated.
totally independent designs that are no
longer clones in the normal sense. Overviews
1. In situations where a single computer is
Business value proposition powerful enough to support the needs of
The vast majority of computers today are a moderate number of concurrent users,
clones, so much so that the term is losing but not powerful enough to support the
its meaning. For a computer to be a clone large number actually expected, it is com-
is no longer a derogatory term. There are mon to use clustering. A Cluster is a group
many variables to consider in selecting a of computers, usually configured identi-
CPU, and no single product or manufac- cally as servers. They will generally have
turer can claim to be ‘‘the” best. the same software installed, and, either
However, designing a whole computer through disk-sharing or by providing each
is an expensive and complex undertaking. computer with its own copy, will have
Poor design, especially of the motherboard, access to the same data.
results in poor performance and unreliabil- The cluster appears to the outside world,
ity. It is important to select a manufacturer under casual inspection, to be a single com-
that can put enough resources into proper puter. A system of dynamic load balanc-
development of their products. A good clue ing is used to redirect all user requests to
is to start reading the ‘‘user manual”: if the the member of the cluster that is currently
text is of poor quality or incomprehensible least occupied. Two users ‘‘logging in” at
this may be an indication of other quality the same time may find themselves con-
issues associated with the computer itself, nected to different real computers, but still
or its manufacturer. having access to all the same files. A cluster
of servers is often called a Server farm.
Summary of positive issues With clusters, it is common to use a
Competition provided by alternate man- system of Load balancing. This is a soft-
ufacturers ensures healthy development ware consideration that attempts to keep
budgets, continuing improvement of prod- the processing load on each of the mem-
uct lines, more options, and lower prices. bers of a cluster reasonably balanced.

65
Cluster

Implementations range from the extremely Business value proposition


simple method of assigning incoming sim- Clustering can be a very cost-effective solu-
ple requests to individual members in turn, tion. Twenty normal computers often cost
in a ‘‘round-robin” fashion, so that they less than one super-computer of 20 times
will be balanced on average, to having one the power. More importantly, a cluster cre-
system with special responsibility for mon- ates a very robust system: if one computer
itoring the state of each of the others and fails, the 19 others continue unaffected,
actively redirecting requests to the least and the system simply runs at 95% of its
busy, or having each member provide con- original power. Maintenance of both soft-
tinually updated assessments of its own ware and hardware may be performed on
load to a much simpler central redirec- a rotating basis, one cluster member at a
tion service. It is also possible under some time, so there is never any need for down-
circumstances to have computations that time.
have already begun on one cluster member
migrate to another when load conditions Summary of positive issues
change, although this is a very complex The computers, networking, and other
procedure. hardware used in clusters can be treated
2. A Block is the smallest individually as ‘‘commercial off the shelf ” (COTS), pro-
accessible unit of storage on a disk system; viding a low-cost architectural solution and
it is normally 512 bytes in size. Operating an associated lower maintenance cost than
systems assign numbers to blocks to act as a specialized hardware solution. Clusters
their addresses, and use the Block number as can be created using proprietary operat-
a way of finding data. Many older operating ing systems as well as open-source sys-
systems use only small-range numbers for tems. Clusters provide reliable and scalable
identifying blocks (Windows 3.1 used num- solutions.
bers in the range 1--65 527), and this created
an unacceptably low limit on the usable Summary of potentially negative issues
size of a disk (65 527 × 512 bytes is only Clusters might not be suitable for all
32 MB). processing requirements, especially those
The solution was to consider groups of requiring close connectivity or interac-
neighboring blocks to be one single Super- tions amongst tasks. The bandwidth and
block, and to assign numbers to these super- response-time requirements of a high-
blocks rather than to individual real blocks. volume transaction-processing system may
Super-blocks are properly called Clusters. If a be too great for the intra-cluster network,
disk drive were configured to have clusters in which case a high-power mainframe
of 64 blocks each, the maximum accessible computer may be necessary.
size would be raised to 2 GB. The disadvan-
tage is that one cluster becomes the small- References
r K. Kopper (2005). Linux Enterprise Cluster:
est allocatable amount, and thus even a
file containing only one byte of data would Build a Highly Available Cluster with
occupy 32 768 bytes of disk space. Commodity Hardware and Free Software
3. In data analysis and Data mining (q.v.) (San Francisco, CA, No Starch Press).
r T. Sterling, J. Salmon, D. Becker, and D.
a cluster is a group of data items that are
more closely related to each other than Savarese (1999). How to Build a Beowulf:
can be expected from random chance, and A Guide to the Implementation and
are thus assumed to have some meaningful Application of PC Clusters (Cambridge,
inter-relationship. MA, MIT Press).

66
Cobol

file structures that Cobol could manipu-


Cobol
late. From the 1970s until the late 1980s
Foundation concepts: Programming language, Cobol became the most widely used pro-
Compiler. gramming language for business applica-
Definition: The Common Business-Oriented Lan- tions and compilers were developed for a
guage (Cobol) was developed to facilitate the creation wide variety of platforms, ranging from
of business applications. large mainframes to personal computers.
Cobol has continued to evolve and in 2002
Overview both ISO and ANSI released new standards
Cobol, a programming language proposed (ISO-1989).
in 1952 by computer pioneer Admiral Grace
Hopper, was formally specified in 1960 Business value proposition
at the Conference on Data System Lan- Cobol has been used for around 40 years in
guages (CODASYL), and became an Amer- developing business applications, and there
ican National Standards Institute (ANSI) is a large body of knowledge and experi-
standard in 1968. The language was ence on the programming language, and
intended to provide an accessible mecha- many structured development methodolo-
nism through which business applications gies have been created to support it.
could be developed.
The language was designed to be easy Summary of positive issues
for non-technical programmers to read and Cobol is a relatively easy language to learn
write, which was very much against the and use, and in which to develop applica-
style of the day. Commands in Cobol appear tions. Billions of lines of code have been
to be plain English instructions, with a written over the past 40 years and there
limited vocabulary and strangely stiff syn- is a deep pool of technical and human
tax, including sentences such as ‘‘s u b - resources available to support applications
t r ac t d i s co u n t f ro m t o ta l ,” ‘‘sort and their development. Many CAD tools,
c u s t o m e r - f i l e o n a s c e n d i ng k e y structured design methods, test suites, and
ac c t - n u m b e r ,” and ‘‘p e r f o r m i n p u t - systems have been developed to support
p ro c e s s u n t i l f i l e - co m p l e t e .” Cobol. Cobol is continuing to evolve and
Although the successful use of Cobol systems have been developed both to enable
for programming does require substantial Cobol programs to be converted to other
training, it did open up commercial soft- languages, such as Java, and to receive
ware development to a new generation of input from XML documents.
programmers without scientific or engi-
neering backgrounds. Cobol was very dif- Summary of potentially negative issues
ferent from the primarily mathematically Cobol has been declining in use since well
oriented languages that were available to before the year 2000 when the ‘‘Y2K” prob-
developers in the 1970s such as Fortran, lem was understood and technically solved.
which were difficult to use for the develop- Y2K was the result of representing years in
ment of business applications that involved files by two digits (e.g., 1966 as 66), and
a significant amount of file and record pro- consequently programs in Cobol (and other
cessing as opposed to computation. The languages) reading and manipulating two-
emergence of Cobol by the 1970s and its digit date fields. Owing to the length of
rapid adoption by developers were also due time over which Cobol has been used,
in part to the emergent nature of more some programs may have become unstruc-
powerful online database technologies and tured in nature over their lifespan as

67
Collaborative commerce

they were amended by many generations similar systems) to be sent across a network
of programmers. This leads to difficulties efficiently and effectively.
in understanding, modifying, and testing An example of collaborative commerce
aged Cobol systems. occurs when company A (e.g., a car manu-
facturer) sends to company B (a tier-1 com-
References ponent manufacturer) a data file (in XML)
r D. McCracken (1970). Guide to Cobol
that was generated by their ERP system’s
Programming (New York, John Wiley and ‘‘planning and forecasting module,” which
Sons). was itself generated in the ERP module that
r http://www.cobolstandard.info/wg4/
models customer demand. Upon receipt by
standard.html. company B the information is routed into
r H. Hinman (2004). Microsoft.NET for
their ERP system and their ‘‘planning and
COBOL Programmers, 2nd edn. forecasting module” re-aligns company B’s
(Sunnyvale, CA, Fujitsu Software production levels, inventory levels, human
Corporation). resource requirements, and capital require-
ments, and notifies their tier-2 supplier of
Associated terminology: C and C++, Java,
the changes. The majority of these chan-
XML.
ges and messages can occur without any
human intervention at all if the process
models are designed to allow such an event
Collaborative commerce to occur.
The advent of internet technologies such
Foundation concepts: ERP, Network. as XML has allowed the development of net-
Definition: Collaborative commerce is the sharing works that allow many companies to inter-
of information between two or more entities to connect and share data in large quantities
improve the collaborators’ service delivery or product effectively and efficiently. The data-sharing
development. concept that underpins collaborative com-
merce has developed the concept of a Data
Overview synchronization hub. Such a hub allows com-
Collaborative commerce has fundamentally panies to place data pertaining to items
existed in one form or another since in an accepted standard format such as
the beginning of modern manufacturing EAN.UUC. The data is then placed in a ‘‘data
when, for example, manufacturers would pool” (a database where many companies
share their schedules with a supplier. How- place their data). This data pool is usually
ever, with the advent of more sophisticated administered by a third party (but for large
technologies such as networking, CAD/CAM companies this can be done in-house) and
design applications, virtual reality models, they send information pertaining to the
materials requirements planning systems, dataset to the global synchronization net-
ERP systems, and messaging technologies, work (GSN) provider, which holds the infor-
collaborative commerce has evolved dra- mation and the location of each item’s Data
matically to the point where corporations pool (q.v.). Customers or other companies
link aspects of their value chains together then search the GSN registry through their
into a value system. The ability for compa- chosen data pool provider and notify the
nies to share data across corporate bound- data pool provider of the items they wish
aries has been facilitated by the adoption to subscribe to. The GSN provider then uses
of common base technologies such as XML a ‘‘synchronization engine” automatically
and TCP/IP protocols. These protocols allow and continuously to synchronize the data
information generated by ERP systems (or between the two companies. For example, a

68
Compiler

component’s manufacturer places its prod- Reference


uct specifications in a data pool, then, if r G. Norris, J. Hurley, K. Hartley,
they change the specification of a compo- J. Dunleavy, and J. Balls (2001). e-Business
nent (perhaps the strength of a washer), and ERP Transforming the Enterprise
all parties who have subscribed to the data (New York, John Wiley and Sons).
pool will be notified of this change.
Associated terminology: XML,
Client--server, Internet.
Business value proposition
Collaborative commerce allows organiza-
tions to streamline their operational costs
by integrating aspects of their value chain Compiler
with aspects of partner companies’ value
chains. Collaborative commerce can be Foundation concepts: Programming language, Oper-
undertaken in any area of joint interest, be ating system.
it supply chain, logistics, manufacturing, Definition: An application that automatically trans-
market research, or purchasing. The aim is latesinstructionswritteninahuman-understandable
to leverage technology to bring down costs language into the computer’s own internal elec-
and speed the delivery of information that tronic representation, so that they become a runnable
is critical to making shared decisions, thus application.
reducing cycle times.
Overview
Summary of positive issues Computers and humans do not work in the
Open standards, protocols, and technolo- same way. For a computer, every single step
gies have evolved to facilitate the entry has to be spelled out in explicit, perfect
of companies into collaborative commerce detail. Even the seemingly trivial task of
relationships. Collaborative commerce ena- typing the word ‘‘hello” so that it appears
bles companies to perform a variety of on a monitor, a task that a human could
strategic undertakings that range from perform without any further explanation,
just-in-time inventory models and lean requires literally thousands of microscop-
manufacturing to collaborative resource ically detailed instructions when carried
planning. out by a computer.
In the early days of computing, pro-
Summary of potentially negative issues grammers had to work out for themselves
The large companies leading the push to every single detailed step of every task to
collaborative commerce may force early be performed, and laboriously enter those
adoption of technology upon less techni- instructions into the computer. Even the
cally sophisticated companies, making task of adding 10% tax to a list of numbers
them absorb the learning and technology and printing the results would require a
costs that accompany early adoption. program dozens of pages long. Because
Not entering into a collaborative com- so much detail was required from fallible
merce operational model may not be an humans, errors were very common, and all
option in certain industries or sectors in programming was a very expensive propo-
order for a company to remain competitive. sition, requiring highly trained expert tech-
As technologies continue to change, all nicians working for long periods.
collaborators utilizing a particular technol- Soon it became obvious that certain sub-
ogy will have to assess the impact of those tasks were common to nearly all pro-
changes and determine their technology- grams: reading and printing data, calcu-
migration policy collectively. lating square roots, and sorting files are

69
Compiler

almost universal requirements. Program- their resources ‘‘reinventing the wheel,”


mers wrote partial programs for just these hence the use of library functions such
subtasks, and fed them into the computer as ‘‘sort” or ‘‘merge” embedded within the
in advance. When a new program was compiler.
required, only the tasks unique to the new The history of computing includes a
problem had to be programmed, and the large amount of time and resources devo-
computer could be instructed to take a ted to the development of ‘‘better” pro-
selection of pre-entered partial programs, gramming languages and this has resulted
and Compile them together into a single in literally thousands of programming lan-
application. guages and variants being produced since
After that first step, there was a flood the 1950s, each of which had either an
of development that still continues today. associated compiler or an interpreter e.g.,
More and more of the tasks of program- Cobol, Fujitsu Cobol, MS-Cobol, Micro Focus
ming that were tedious and error-prone Cobol, Dev-C/C++, Algol 60, Algol 68, For-
were automated. Gradually it became pos- tran IV, Fortran 77, LCC-Win32, Perl, Pascal,
sible to give the computer instructions in Modular, Java, C, C++, Basic, and Visual
a language that seems almost like English, Basic (VB). The main reason for the exis-
and allows programmers to express their tence of all these variants is the fact that
designs in clear and abstract ways that there is not just one perfect programming
other programmers can easily understand language for all situations and problems,
and assist with. The computer itself accepts just like there is no one means of trans-
these high-level descriptions from program- portation for all situations (motorbike,
mers, and translates them into the thou- car, van, and truck all suit different
sands of exquisitely detailed instructions needs). The US government attempted to
that the programmers used to produce. develop one special programming language
A compiler is a computer application named Ada that would encompass all its
that automatically performs this trans- needs and minimize the need for govern-
lation from a clear, high-level, human- ment software developers and contractors
understandable language into the arcane to learn and develop code in multiple
and detailed internal ‘‘language” of com- languages. However, this ‘‘one-size-fits-all
puter instructions. A compiler is one of approach” failed as more and more ‘‘spe-
the largest, most complex, and most reli- cial exemptions” were made so that devel-
able pieces of software on any computer. opers could build code for different types of
Using a compiler is effectively making the applications. For example, while Ada code
computer program itself; in the 1950s, may be suitable for payroll systems and
programming with a compiler was called accounting systems, it would not necessar-
automatic programming. The human-written ily be suitable for code to be embedded
program that is the input to a compiler in a missile’s navigation system. A second
is known as Source code; the computer- factor adding to the complexity of the situ-
generated program that comes out is ation is the nature of compilers: they take
known as Object code or Executable code. the programming language as written at
the human level and manipulate it such
Business value proposition that it can operate upon the hardware plat-
The use of compilers allows programmers form (the computer). Unfortunately, until
to be more efficient and effective with their recently a different compiler was needed
time. As programming is a labor-intensive for each platform, in that a Cobol compiler
process it is vital that the programmers for an IBM platform would not work for a
and the IT organization do not spend Fujitsu platform, which resulted in variants

70
Complexity

being produced for each machine and fre- being closely coupled to the hardware and
quently the language designers would not operating systems of the computer being
stick strictly to the ANSI definitions of the used.
‘‘standard” language, thus causing more
References
variance. r R. Wexelblatt (1981). History of
In order to resolve this situation IT
Programming Languages (New York,
organizations within commercial settings
Elsevier).
have focused on the set of languages and r T. Bergin and R. Gibson (1996). History of
their compilers that are portable or inter-
Programming Languages -- II (New York,
operable, that is they will work on any
ACM Press--Addison-Wesley, 1996.
platform. This movement started with the
marriage of Unix and C, which are often Associated terminology: C and C++, Java,
open source in nature, enabling the pro- Cobol.
grammers to understand the constraints
and workings of the systems in which they
were programming. By the late 1980s C Complexity
had become a favored and common devel-
opment language and evolved into C++, Foundation concept: Algorithm.
an object-oriented programming (OOP) envi- Definition:Complexityisameasureoftherelationship
ronment, which was more powerful than between the amount of data to be processed and the
C yet still portable across platforms. While time required to perform that processing. It is not a
C++ is the predominant choice of devel- simple matter.
opment language for the majority of pro-
grammers, other compiled languages that Overview
are heavily intertwined with the platform In computing, Complexity is a technical term
upon which they operate are still widely that describes the fundamental speed of an
used, e.g., Visual Basic and J++ are predom- algorithm. It is not directly connected with
inantly development languages for inter- the common meaning of the word, ‘‘how
facing with Microsoft applications and for complicated something is.”
developing web interfaces. Take for example a simple searching
application. It takes as its input a plain
Summary of positive issues text file, perhaps a corporate report, per-
Compilers shorten the development time haps a novel, it doesn’t matter. The applica-
for programmers to write programs. The tion’s job is to search the entire file to see
library features of compilers save program- whether a particular word or sequence of
mers from needlessly duplicating previ- symbols appears anywhere within it, a com-
ous standard functions every time they mon task. Obviously, the longer the file is,
write a program. Open-source platform- the longer the search will take. We could
independent compilers allow developers to perform a simple experiment and work out
write code that will run upon a variety of the relationship. If we make the applica-
devices. tion search a 10 MB file, and find that it
takes 1 second to do so, we could reason-
Summary of potentially negative issues ably conclude that it would take 2 seconds
There exist millions of lines of code written to search a 20 MB file, and 10 seconds to
in thousands of programming languages search a 100 MB file.
and their variants, few of which will com- People naturally expect a linear relation-
pile and run on any other systems. Com- ship between the size of a problem and the
pilers are frequently proprietary in nature, time required to solve it. If the data size is

71
Complexity

doubled, the time required is also doubled; ing curriculum, and it is a very complex
if the size is multiplied by 100, then the subject.
time required is also multiplied by 100. This The magnitude of the problem is simple
is so much a part of normal life that it to illustrate. Consider the testing phase of
is usually just taken for granted. Driving a new software application. It is tested on
200 miles takes twice as long as driving a data set of 1000 records, and found to
100 miles; reading two reports takes twice take just 1 second to perform its task. A
as long as reading one report. naive developer would assume that process-
Interestingly, the relationship between ing 10 000 records would take 10 seconds;
data size and processing time is often very 1 000 000 records would take 1000 seconds
far from linear, and the assumption of (17 minutes) and so on. The table below
linearity can lead to serious miscalcula- shows what the true completion times
tions. One simple example is ‘‘long multi- would be for some common algorithm com-
plication,” as learned in elementary school. plexities.
To multiply together a pair of two-digit
numbers, four individual simple multipli- Data
cations are required (to work out 63 × 75, size Linear Quadratic Cubic Logarithmic
one calculates 6 × 7, 6 × 5, 3 × 7, and 3 × 5,
1000 1s 1s 1s 1.00 s
then with adding, carrying, and luck, one
2000 2s 4s 8s 1.10 s
produces the correct answer). It took me 3000 3s 9s 27 s 1.17 s
12 seconds to multiply together those two- 4000 4s 16 s 1 min 1.20 s
digit numbers. With the assumption of 5000 5s 25 s 2 min 1.24 s
linearity, I would expect it to take me 10 000 10 s 100 s 17 min 1.31 s
24 seconds to multiply together a pair of 100 000 100 s 21/2 hours 12 days 1.70 s
1 000 000 17 min 12 days 31 years 2.00 s
four-digit numbers, 120 seconds to multi-
ply together a pair of twenty-digit numbers,
and so on. This is a dramatic illustration, but not a
However, such a prediction is completely dramatized one. After finding that a new
invalid. Closer analysis reveals that mul- application works properly, and seeing that
tiplying together a pair of twenty-digit it takes 1 second to process 1000 data items,
numbers requires 400 individual simple many programmers would not bother to
multiplications, not 40, so it would take me test it with 1 000 000 data items, especially
1200 seconds, not 120. The amount of work after deducing that the test would take 17
required grows with the square of the data minutes. However, if the algorithm turned
size. Different problems (and, sometimes, out to be cubic, it would in fact take 31
different solutions to the same problem) years, not 17 minutes, and large-scale cus-
produce a different relationship between tomers would be justifiably angry. As the
amount of data and required processing table shows, non-linear complexities can
time. Linearity is possible, but must not be also work the other way, giving unexpect-
assumed. edly fast responses.
Non-linear relationships between the
amount of data and the time required to Business value proposition
process it are very common in computing. Failure to understand the complexity of a
A professional programmer must be aware problem makes it impossible to estimate
of this, and must know how to perform a even roughly how long it will take to solve
basic complexity analysis for an algorithm. a problem. This in turn makes it impossi-
This is a major component of a comput- ble to select the appropriate solution for

72
Compression

any given circumstances. Examples of such constrained and determined by many fac-
problems include the algorithms used to tors, including budget, hardware, legacy
search and compare databases. If a com- software, and operating system constraints.
pany were to develop a new product, say
References
a search engine for the internet, the algo- r D. Knuth (1975). Fundamental Algorithms
rithmic complexity of the approach taken
(New York, Addison-Wesley).
needs to be considered by the organiza- r M. Hofi (1975). Analysis of Algorithms
tion’s programming team. This complexity
(Oxford, Oxford University Press).
would include the way the data is struc-
tured (the data structure) in the database, the Associated terminology: Algorithm,
way the data is placed on the server config- Database, Computability.
uration, and the search algorithm that oper-
ates over the data. In order for the new
product to be successful, it is thus imper- Compression
ative that all of the technical complexity
issues be considered and that these issues Foundation concepts: Bit, Binary, Digital.
are well understood and resolved such that Definition: Reducing the size of a data set (the amount
platform and software purchases such as of storage it requires) without losing any of the data.
the database upon which the data will
reside can be made in an informed man- Overview
ner. For example, a poorly selected database Compression and Encryption are very similar
environment may lead to the programmer operations. Both convert data into an alter-
using a sub-optimal data structure to store native form, from which the original may
the data, which forces the selection of a be fully recovered later. With encryption,
sub-optimal search algorithm to be used the aim is to make the recovery impossible
and leads to a sub-optimal product being for unauthorized third parties. With com-
produced as a result. Thus, complexity can pression, the aim is to make the alternative
be directly tied to the value proposition of form as small as possible.
products as well as the total cost of own- Data in its Raw form, as it is collected,
ership and return-on-investment decisions unprocessed, tends to contain a lot of
associated with a product. redundancy. That is, with some analysis it
would be possible to represent exactly the
Summary of positive issues same information in a much less bulky
The mathematics of computational com- form. Human speech is a clear example. To
plexity, while not easy, is formally devel- record one minute of speech faithfully, so
oped within the discipline of computa- that a play-back is indistinguishable from
tion and computer science. The algorithmic the original to the human ear, requires
complexities of many well-known func- about 5 000 000 bytes of data. If the only
tions are documented and understood by real information in that speech is the
computer scientists. words that were spoken, it will amount to
not much more than 100 words, or a total
Summary of potentially negative issues of 500--1000 characters, which require one
Failure to determine the complexity of an byte each. The same information can be
algorithm can easily make a software prod- recorded in one two-thousandth of the
uct completely worthless. Not understand- amount of data. It has lost all the sounds of
ing the whole picture of product complex- human speech, but retained all of the use-
ity can be fatal. Algorithmic complexity is ful information. That is one form of data

73
Compression

compression: extracting the true informa- behind Huffman encoding, which is the basis
tion from the surrounding raw data. of many commercial compression schemes.
Another form is illustrated by the rep- Further compression may be achieved by
resentation of graphical images in digital noticing that whole areas of the image
form. The most common representation for are the same color, and encoding a repre-
such images uses a triplet of small num- sentation of that area together with just
bers (in the range 0--255) for each pixel; one copy of its color. Detecting large areas
the first indicates the degree of redness in of a single color is very easy for human
the pixel, the second greenness, and the observers, but computationally difficult. A
third blueness. So a bright red pixel would simpler version, a sequence of pixels within
be represented by (255,0,0), a greenish-blue a single row that are all of the same
one by (0,120,240), a black one by (0,0,0), color, is much easier, and is known as RLE,
and a white one by (255,255,255). These Run-length encoding, a popular compression
numbers are simply stored as they are, in scheme for some limited applications.
top-to-bottom, left-to-right order, to make Far greater compression ratios may be
an uncompressed image file. A number achieved by noticing that certain patterns
in the range 0--255 requires exactly one of pixel color recur over and over again,
byte of storage (that is why that particu- and complex encodings that refer back to
lar range is used), so a small picture of size previous parts of an image to reconstruct
just 200 × 200 pixels would occupy 120 000 patterns can be devised. This is the basis
bytes uncompressed. of LZW (Lempel--Ziv--Welch) and GIF (Com-
If it is observed that a particular picture puServe Graphics Interchange Format) com-
involves just four different colors, each of pression. Another variation of this provides
those colors may be given its own binary the logic behind ‘‘Zip files.” Zip files are
encoding, perhaps 00 for white, 01 for most commonly used under Windows, but
red, 10 for blue, and 11 for green. Then a are also compatible with other systems;
sequence of four neighboring pixels can be they are single files that contain ‘‘archives”
written in a single 8-bit byte: 00000110 for for whole folders or directories all com-
white, white, red, blue. That results in a pressed. Individual files may be accessed
total size of 10 000 bytes for the 200 × 200 and decompressed or replaced either indi-
image, compressing it to just over 8% of its vidually or in groups. Zip files are a pop-
raw size. Recoding is the simplest form of ular and convenient means of distributing
compression. whole data collections and applications.
If it is further observed that, of the four If perfect reconstruction of the original
colors that appear in the image, white raw data is not required, then Lossy (as
is extremely common (perhaps 90% of all opposed to Lossless) compression methods
pixels are white), red is next (at 8% of may give even greater savings. In simple
all pixels), and blue and green are quite terms, it may be observed that a large
rare (at 1% each), a variable-length encod- region of an image consists not of exactly
ing will produce additional savings. In the same shade of blue, but of a small
this example, a one-bit code, 0, could be range of very similar shades. The whole
used for white, a two-bit code, 10, for red, region could be approximated by a sin-
and three-bit codes, 110 and 111, for blue gle shade region, and the departure from
and green. The sequence white, white, red, true might not be apparent to observers.
white would be encoded as 00100. For the More complex lossy methods can involve
whole image, we would have a total size discovering mathematical functions that
of 44 800 bits or 5600 bytes, less than 5% approximate the changes in pixel values
of the original size. This is the basic idea within a region to any desired degree of

74
Compression

accuracy. Lossy methods may be very good uments. Scanners can encode their images
for images, especially if they are only to be in a bitmap format, producing extremely
viewed by humans, but are not suitable for large but easy-to-process files, but, for stor-
critical data, where approximating a cus- age after processing, a compressed form of
tomer’s name or account balance would not the file should be acceptable. All popular
be acceptable. image file formats (except BMP) can provide
Since bandwidth limitations are a uni- a great deal of compression.
versal fact of life for the computing com- A further aspect of compression that is
munity, compression methods have been extremely useful for organizations is the
researched intensively. A great many very ability to send compressed files over net-
successful compression algorithms have works, in essence increasing the availabil-
already been fully worked out and imple- ity of bandwidth for other tasks.
mented. A programmer faced with the Network managers can specify processes
challenge of data compression would be and enforce policies for the application of
well advised to perform a careful literature compression methods pertaining to their
search before reinventing the wheel. Com- network and its users. They can also ensure
pression is possible only when the original that applications whose main focus is the
data contains some redundancy; compres- manipulation and storage of data files that
sion works by removing or reducing that are large and require significant resources
redundancy. Once data has been well com- are optimized. These applications include
pressed, further compression will not make document-management systems that facil-
it any smaller, and in some cases may pro- itate the use, linkage, version archiving,
duce a slight enlargement. and long-term storage of heterogeneous file
types (image files, text files, web files, XML
Business value proposition files, audio files, video, etc.)
The use of compression techniques for
the transmission and storage of data pro- Summary of positive issues
vides organizations and individuals with Compression allows large files to be stored
the ability to leverage their infrastructure, and transmitted in a smaller size than
incurring a relatively low overhead. A prin- their normal format. Decompression allows
cipal benefit of compression applications is files to be restored to their original format.
the ability to store data at a reduced size; There are many well-researched and well-
while this incurs overhead during compres- supported compression algorithms, tools,
sion and decompression, if the data is being and packages available for the purpose of
stored for the long term (say 7 years for compression. Compression techniques are
audit and compliance purposes) the stor- built into many software applications and
age savings can be considerable and the automatically generate compressed files for
work to compress the data may be a one- storage. Most files can be significantly com-
time expense. Data compression on a user’s pressed with lossless methods if exact
local computer system also allows consider- reconstruction of the original is needed,
able savings in storage requirements. A pop- and even further compressed with lossy
ular use of such technology is for the local methods if a good approximation to the
archival storage of emails in compressed original is sufficient.
format; this enables users to read old mes-
sages if they wish, after decompressing Summary of potentially negative issues
them, but reduces the local storage require- Compression and decompression impose a
ments. Similar processes can be associated resource overhead on the system when the
with the storage of images or scanned doc- operations are performed. If compressed

75
Computability

files are to be transmitted to other users, do it, but it is not known to be impossible.
the recipient must have the decompres- In fact, we know that it is possible, we just
sion software in order to read them. don’t know how. The ability might just be
Repeated compression and decompression round the corner, or we might never find
using high-ratio lossy methods results in it out.
Generation loss, in which the data is contin- In the physical world, most things clas-
uously degraded by each operation. sified as impossible may well not be. With
the speed-of-light example, it may be that
Reference
r K. Savood (2000). Introduction to Data new discoveries in physics will completely
change our beliefs; that does not make
Compression, Morgan Kaufmann Series
impossible things become possible, it sim-
in Multimedia and Information Systems
ply makes us realize that we were wrong in
(San Francisco, CA, Morgan Kaufmann).
the first place.
Associated terminology: Audio, Video, In computer technology, things called
Information lifecycle management. impossible are usually merely beyond our
reach. An expert may say that 2048-bit RSA
encryption is safe, meaning that, with cur-
rently foreseeable technology, encrypted
Computability messages will remain unreadable. But
nobody really knows how technology will
Foundation concept: Algorithm. develop in the future. Totally new, cur-
Definition:Whattasksarereallyimpossibletoperform rently unimagined, designs for computers
with a computer system? could render all these assumptions invalid.
The idea of a Quantum computer is currently
Overview no more than science fiction, but, if a prac-
The concept ‘‘impossible” is usually con- tical one were to be constructed, it would
fused with the concepts ‘‘I don’t know how have exactly this effect.
to do that” and ‘‘that isn’t within our tech- Additionally, there is the possibility of
nological grasp.” It is often said that it new developments in the theory of com-
is impossible for anything to travel faster putation. Continuing with the same exam-
than light, but the truth is simply that our ple, one of the great unknowns is the
current theories of physics do not allow ‘‘P equals NP” question, which many theo-
anything to accelerate beyond the speed of reticians have devoted their entire careers
light. We do not really know that these the- to trying (and failing) to answer. Nobody
ories are correct, and human understand- knows whether P equals NP or not, and
ing of physics has been completely turned currently it is a very dusty academic sub-
on its head any number of times in the ject. But if the answer is ever found, the
past. People may say that it is impossible to results will be striking. If it turns out that
fly to the moon for less than $300; we can’t P does not equal NP, then RSA encryption
do it now, but we have no idea what tech- really is safe. If it turns out that P equals
nological advances the future may hold. NP, then immediately, without any further
Similarly in programming, there are work, literally thousands of intractable
many things that we just don’t know how problems suddenly become easy; cracking
to do. True artificial intelligence, a com- RSA encryption (and a lot of other things)
puter mimicking a human well enough instantly becomes viable. What was called
that an impartial blind observer can not impossible is now possible.
tell that it is not really a human, has not To make matters more complex, theoret-
been achieved, and nobody knows how to ical computing does know of some things

76
Computer

that are genuinely impossible. Things that, have only a limited understanding of and
no matter what new technological leaps exposure to the discipline and its impact
or theoretical insights occur, must remain upon their work.
completely impossible; problems that if
References
solved would result in a contradiction in r J. Hopcroft and J. Ullman (1979).
logic itself. The study of computability, that
Automata Theory, Languages, and
is, possibility and impossibility in computer
Computation (New York, Addison-Wesley).
processes, is deeply complex and very easy r J. Brady (1977). The Theory of Computer
to misunderstand. As a general rule, only
Science (London, Chapman and Hall).
those with advanced degrees in computer r A. Turing (1936). ‘‘On computable
science or engineering are even aware that
numbers with an application to the
the subject exists, and far fewer have any
Entscheidungsproblem,” Proceedings of the
expertise in the matter.
London Mathematical Society, 2 (42),
230--265.
Business value proposition
Any attempt to do the impossible is likely Associated terminology: Algorithm,
to be a financial disaster. A belief that Complexity.
something is impossible when it isn’t can
result in a technological disaster. When-
ever a programmer claims that something Computer
is impossible, it is important to find out
what they really mean by that, and whether Definition: A calculating and data-manipulating
they really have a valid basis for the claim. device capable of storing its own program of
For example, the ability of a program to instructions and following those instructions without
check itself for correctness is theoretically intervention.
impossible and this impossibility derives
from Turing’s thesis, thus it is important Overview
that, should claims such as ‘‘self-checking Everybody who is reading this will already
software” be made by vendors, this should have a fairly good idea of what a com-
be thoroughly examined and understood puter is. Instead of discussing unnecessary
before a program is used in what could be technical details, this article will present
a critical process, such as one involving a the major varieties of computer currently
threat to human life. available.
Palmtop, PDA (portable digital assistant).
Summary of positive issues These are computers so small that they can
Computability is a theoretically sound be held in the palm of one hand. Gen-
field of study and significant amounts of erally they have much lower processing
research literature exist to help computer power than any other kind of computer
scientists and professionals examine the (because the very restrictive space require-
underlying mathematical and logical basis ments leave very little room for power-
of their programs and specifications. ful batteries) and very restricted storage
capacity. Some, but certainly not all, are
Summary of potentially negative issues restricted to running a small range of spe-
Computability theory is an extremely spe- cially reduced software applications. They
cialized and complex field of endeavor, have small displays and either miniatur-
which is primarily studied by research- ized keyboards or none at all (using a stylus
ers and academics. Programmers without instead), so they can be quite unsuitable for
advanced course work in computation may normal office work.

77
Computer

Notebook, laptop, portable. These terms ations. Since disks are a significant com-
have grown to mean almost exactly the ponent in the cost of a computer, diskless
same thing. Originally, a portable computer workstations can be financially very attrac-
was one that could be carried around, but tive. The performance of a diskless worksta-
doing so was not necessarily a pleasant tion can be very poor, because it is strictly
experience; such computers are generally bounded by available network bandwidth.
not made any more. When shades of mean- Microcomputer, minicomputer, main-
ing are intended, a notebook is usually an frame, super-computer. These are vaguely
extra-small laptop. defined points on the spectrum of overall
Desktop. A general term that now covers processing power for computers. A micro-
nearly all computers in common use. Any computer is normally taken to be any
computer that can fit on a normal desk and computer whose entire CPU is on a sin-
still leave enough of the desktop available gle integrated-circuit chip. This covers all
for other uses. desktop computers now, and the term has
Server. The term ‘‘server” properly fallen out of use. Originally microcomput-
applies to a software configuration: a net- ers were thought of as essentially toy com-
work application that provides some ser- puters that were not suitable for any seri-
vice to a set of client systems on demand. ous work. Minicomputers were computers
The word is often used to describe a that were really suitable for use by only one
desktop-style computer that is at the higher user (or a very few concurrent users), but
end of the performance range, and so were certainly not just toys. This term has
might reasonably be used for running also fallen out of use, since the realms of
server applications. The term is not strictly microcomputers and minicomputers have
defined, and servers range from a personal completely merged to create the desk-
computer that runs applications that do top computer. A mainframe is a high-power,
not require especially high performance high-price computer, usually very large and
to clusters of high-performance computers having a staffing and facilities requirement
that run enterprise-class systems such as to keep it running (e.g., air-conditioned
ERPs. rooms, automated non-destructive fire
Workstation. The counterpart to extinguishers, and heavy-duty uninterrupt-
‘‘server.” The term is not strictly defined, ible power supplies). Mainframes provide
but is generally taken to mean a desktop- much faster processing and storage access,
style computer intended for use by one and much higher input and output band-
person at a time. Workstations are not at width than other computers, typically
the top of the performance range. being able to support thousands of con-
Thin client. A very-low-power worksta- current users. A super-computer is simply
tion, often one that is used only for display a very fast mainframe, often supporting
purposes, all the real work being done by fewer users and having a lower access band-
a server that receives commands from, and width than a typical mainframe, but pro-
transmits responses to, the thin client over viding even higher computing speed.
a network connection. A thin client could Embedded system. Not all computers
be a reasonable means to provide restricted are clearly identifiable objects intended
public access, perhaps including just web- for direct use by people. Often comput-
browser capabilities. ers are built into other devices as part
Diskless workstation. A variety of thin of the control or monitoring system. Mod-
client, which has no disks or other perma- ern cars, microwave ovens, and even wash-
nent storage media of its own, and so is ing machines are computer-controlled. An
totally reliant upon a server for all oper- embedded system is a computer that is

78
Computer-based training (CBT)

built into something else, not running Summary of positive issues


general-purpose applications, but directly The wide variety of computer specifications
interfaced with the system that it helps to enables systems designers to create and
control. configure systems to meet their individual
systems specifications.
Business value proposition
Computers are not classified just by the Summary of potentially negative issues
above terms. The question ‘‘should we get Technology has a distinct ‘‘half life” and
a Windows, a Linux, or a Unix computer?” needs to be assessed continually regarding
is frequently asked. The important ques- its suitability to support the organization
tion to consider is ‘‘what does the com- and its processes. Technology is not univer-
puter need to be able to do?” In order sally compatible.
to answer this, it is important to under-
stand the required performance character- Reference
istics for the processes that the system is r M. Campbell-Kelly and W. Aspray (2004).
to support. However, note that we now Computer: A History of the Information
talk of ‘‘system” rather than ‘‘computer” Machine (New York, Basic Books).
and link the computer itself to the oper-
Associated terminology: Hardware,
ating system and the applications on top
Operating systems, Client--server, ERP.
of that, coupled with the databases, and
linked through networks.
The first layer of the ‘‘system” above
the hardware of the computer itself is the Computer-based training (CBT)
operating system. A similar assessment is
required for the choice of operating sys- Definition: Computer-based training is the use of
tem(s) to be deployed. An investigation of computers and related technologies to provide an
the applications, databases, and other sys- instructional learning environment through which
tems planned (or in place and to be integra- human subjects can interactively undertake an
ted into it) will need to be undertaken. For educational experience.
example, some database systems will not
run on specific popular operating systems, Overview
while some ERP systems will run on any Computer-based training (CBT) involves the
operating system and thus an assessment use of computers, software, and network
of the inter-operability of the overall sys- technologies to provide an educational
tems architecture needs to be undertaken. platform. Various system configurations
The technical assessment of inter- are utilized, including internet delivery,
component compatibility needs to con- CD-ROM-based delivery, and LAN delivery
sider the loads that are to be placed upon over a corporate intranet. The technologies
the system, e.g., is the system intended used for content include video- and audio-
to support a factory with relatively low delivery formats, video-assisted interactive
throughput or a factory that supports simulation training, interactive quizzes
high-volume mass production? How many and exams, and the delivery of case stud-
users will be using the system at peak load ies combined with exams.
times, what are the end-to-end response
times that are required of the system, etc? Business value proposition
The technical issues and capacity issues CBT is a popular tool used by organizations
will then enable a total cost-of-ownership wishing to provide an online educational
assessment to be undertaken. experience. The tools can be placed online,

79
Computer Fraud and Abuse Act, 1986

allowing employees to undertake training Multimedia Learning (San Francisco, CA,


in accordance with their schedules. Train- Pfeiffer).
ing can be repeated until satisfactory per-
Associated terminology: CD-ROM, LAN,
formance levels have been achieved. Corpo-
Internet.
rations can monitor employee performance
and compliance levels through CBT. CBT
allows flexible training strategies and test-
ing methods to be developed; certification Computer Fraud and Abuse Act, 1986
systems frequently use adaptive testing in
which the questions asked in a test get Foundation concept: Security.
harder or easier depending upon a stu- Definition: Federal law forbidding unauthorized
dent’s results up to that point in the test. access to data stored on computers owned by the fed-
The CBT model provides a low-cost means eral government and financial institutions, or access
to provide flexible training for the work- to further the commission of fraud.
force. CBT is of particular benefit to com-
panies that experience large turnover rates Overview
in training their workforce, since it allows The act forbids intentional access to a
for repetitive training in standard processes computer that either is unauthorized or
at a low cost. exceeds authorization (covering a user who
is authorized to use a computer and has
Summary of positive issues a valid account, but uses it to access data
The principal advantages of CBT are the that he or she is not authorized to access).
lowered delivery price for common courses, However, the act covers only accesses to
the ability to provide an educational experi- data that are covered by national security
ence for employees that can be undertaken concerns, accesses to data on computers
at the discretion of the student, and the owned by the federal government, and
ability to provide courses that can be taken cases in which financial information is
repeatedly by students until the level of accessed on a computer owned by a finan-
material comprehension is acceptable. The cial institution.
CBT method allows students to pace them- The act also forbids unauthorized access
selves through the materials or to use the to any computer when the intent is fraud
system to undertake training as it fits their and the access furthers that fraud, except
schedules. when the only fraud is the unauthorized
use of the computer and the value of that
Summary of potentially negative issues use is less than $5000 in a year.
Development of CBT systems can be expen- Section 5 of the act is of more general
sive and time-consuming. CBT systems still interest, since it forbids ‘‘the transmission
require human management and students of a program . . . or command” that ‘‘causes
need to have a mechanism through which damage without authorization” to a ‘‘pro-
questions may be addressed. Specialized tected computer.” This is a clear descrip-
courses or topics that are time-sensitive in tion of releasing a malicious worm, virus,
nature might not be applicable or cost- or Trojan horse. However, to be covered,
effective. the damage caused must add up to at
least $5000, result in some injury or threat
Reference to public safety, or involve a government
r R. Clark and R. Mayer (2002). e-Learning computer.
and the Science of Instruction: Proven Section 6 of the act forbids the unau-
Guidelines for Consumers and Designers of thorized dissemination of passwords, if the

80
Computer Misuse Act of 1990 (UK only)

intent is to defraud, and it affects inter- the intent of accessing any program or data
state or foreign commerce or a government on that computer, if they know that that
computer. access is unauthorized. The act explicitly
states that the intent does not have to be
Business value proposition to access any particular program or data,
The legislation primarily covers intentional or even to access any particular computer,
misuse or unauthorized access to federal so simply releasing a virus or Trojan horse
(US) computers, equipment, and resources. on the world at large is covered if it ever
The act also covers general misuse of works. The offense can earn up to five years’
computing resources (such as those at imprisonment.
corporations) resulting in ‘‘damage.” The Section 3 of the act adds that it is an
legislation acts as a deterrent to misuse offence to do ‘‘any act which causes an
and provides the judiciary with the abil- unauthorized modification of the contents
ity to impose penalties upon those found of any computer.” Section 17 clarifies the
guilty. position, stating that it is an offence just to
cause a program to be run without autho-
rization.
Summary of positive issues
Spyware (q.v.) is, in its most common
The act forbids intentional unauthorized
usage, software designed to provide infor-
access to a computer system and access
mation from a computer, without its
that exceeds authorization. The act covers
owner’s consent or knowledge, to a third
the misuse of access rights to inflict dam-
party. This is fully covered by Section 1
age upon other systems, including medical
of the act, which forbids simply accessing
and government systems, damage to indi-
data without authorization, regardless of
viduals in the form of physical injury, and
whether any fraud is committed or any
threats to public safety. Financial damage
harm is done, or even whether any files are
from misuse is set at a lower bound of
modified.
$5000.
Section 3 covers unauthorized modifica-
References tions of the desktop setup, changing a web-
r 18 USC 1030; Public Law 99--474. browser’s default ‘‘home page,” and intro-
r http://www.usdoj.gov/criminal/ ducing viruses and worms into the system.
cybercrime/1030NEW.htm.
Associated terminology: Virus, Trojan Business value proposition
horse, Law cross-reference. The legislation primarily covers the inten-
tional misuse of any computer within the
UK and its jurisdiction, either directly or
Computer Misuse Act of 1990 through remote means such as releasing a
(UK only) Trojan horse. The act also covers the unau-
thorized modification of a database or the
Foundation concept: Security. contents of a computer. The act also is pow-
Definition: The law of the UK which covers unautho- erful in that it covers the situation when a
rized computer use. user has accessed data to which they are
not authorized.
Overview The act provides a strong deterrent
Section 1 of the act clearly states that a against the misuse of computing resources
person is guilty of an offence if they cause at corporate level by malicious or even over-
a computer to perform any function with inquisitive individuals.

81
Computer Security Act of 1987

Summary of positive issues Reference


r Computer Security Act of 1987, Public
It is a wide-ranging act that protects com-
puter systems from misuse through attack Law 100--235 (H.R. 145).
by harmful software and from users who Associated terminology: Law cross-
access information to which they are not reference.
authorized. It covers all computers.
Reference
r ‘‘Computer Misuse Act 1990 Elizabeth
II. Chapter 18,” The Stationery Office Connectivity standard
Limited, 123 Kingsway, London WC2B Foundation concepts: Network, Client, Server.
6PQ, UK. Definition: A connectivity standard is any protocol
Associated terminology: Law cross- used by two or more devices to communicate.
reference.
Overview
The term Connectivity standard is used widely
to denote a variety of standards or proto-
Computer Security Act of 1987 cols that are utilized to enable two or more
devices to communicate and share data.
Foundation concept: Security. A set of standards is proposed and
Definition: The act aims to establish standards and administered through authorities and bod-
procedures “that improve the security and privacy of ies such as ANSI, which is responsible for
sensitive information in federal computer systems.” X.12, and the World Wide Web Consor-
tium, which is responsible for XML. These
Overview standards are aimed at worldwide cross-
The Computer Security Act of 1987 was industry coverage and acceptance.
enacted ‘‘to assign the National Bureau A second set of standards is proposed
of Standards responsibility for develop- and administered within specific indus-
ing standards and guidelines for Federal tries to cover special regulatory, techni-
computer systems, including responsibility cal, or data requirements. For example,
for developing standards and guidelines the National Committee for Clinical Labora-
needed to assure the cost-effective secu- tory Standards developed a ‘‘vendor-neutral
rity and privacy of sensitive information in standard” known as POCT1-A that permits
Federal computer systems.” medical point-of-contact devices from dif-
The act requires the NBS to produce ferent manufacturers to communicate with
guidelines for a wide range of areas, includ- laboratory and hospital information sys-
ing technical, managerial, physical, and tems (e.g., hand-held devices that perform
administrative areas, aiming to prevent analysis on blood for levels of electrolytes,
unauthorized access to federal data sys- proteins, or glucose).
tems. The act also calls for training to be A third set of standards is proprietary
provided in accordance with the goal of and developed by organizations that are
proving data security and assessing system associated with a product: data transmis-
vulnerability across agencies. The imple- sion over networks was subject to vendor-
mentation and status of the act are mon- devised standards prior to the adoption
itored by a Computer System Security and of TCP/IP. Proprietary standards shared
Privacy Advisory Board within the Depart- between two or more entities are also
ment of Commerce. developed to solve a particular technical,

82
Cookies

business, or data need but are not usu- to upgrade or change at the vendor’s pace
ally considered ‘‘standards” in the univer- rather than independently. Such standards,
sal sense of those supported by bodies such not developed by authorizing bodies, might
as ISO and ANSI. However, should the tech- not become adopted by a wide user base
nology developed by a vendor become pop- and could lead to problems in connect-
ular, indiscernible, or even forced upon ing with other entities that adopted what
customers, such protocols may be known turned out to be the ‘‘industry best prac-
as ‘‘industry standards,” but should still be tice” or de facto standard.
considered proprietary in nature. Some vendors do not always support all
There is a fourth set, termed ‘‘open stan- the connectivity standards, and a compro-
dards,” for which the specifications and mise may be required in order to enable
protocols are explicitly available to anyone communications. For example, the develop-
who wishes to adopt them or modify them. ment of a middleware solution or simply
the adoption of a second choice, which is a
Business value proposition more supported standard, may be required.
A wide variety of connectivity standards is
Associated terminology: Protocol,
available to facilitate connection between
Client--server, W3C, IEEE.
hardware, software, and network devices,
through both wired and wireless media.
The use of widely adopted standards allows
organizations to connect with a wide spec-
Cookies
trum of users who have also adopted that Foundation concept: e-Commerce.
standard. International and national stan- Definition: A cookie is a small packet of information
dards are typically adopted and used by created by a web server and stored on a web-browsing
technology vendors to enable their prod- computer without the knowledge of the computer’s
ucts to connect with other products and owner. It contains information about recent transac-
are developed to maintain compatibility tions with the server, and is used to identify returning
between vendors. Open standards allow browsers and provide some context for their visit.
organizations to amend and modify tech-
nology protocols to meet their particular Overview
requirements. Browsing the web is a Stateless activity.
This means that, when a browser visits a
Summary of positive issues web site, an internet connection is made
There exists a wide range of connectiv- between the browser and the server, the
ity standards that allow entities to com- browser transmits a request for a document
municate in a variety of ways and incur or page of information, the server trans-
the overhead that is appropriate for their mits the requested information, and the
requirements. International, national, and connection is terminated. If a second docu-
industry-specific standards tend to have ment is needed from the same server, a new
wide support and established consulting connection may be required. No informa-
resources are available to establish, main- tion survives from one request to the next.
tain, and support them. Even if a complex interaction is under way,
in which the user is completing a multi-
Summary of potentially negative issues part Form to supply essential data for a
The adoption of proprietary connectivity transaction, each stage is completely dis-
standards imposes a degree of dependency connected. A web server works by listening
upon the vendor and it may be necessary for anonymous requests for data, servicing

83
Cookies

those requests, and immediately forgetting eroding about cookies, and when they do
them. provide the best solution, some clear state-
This would make reliable commercial ment of policy may calm users.
transactions impossible. If the browser’s The ‘‘cookie” debate is centered upon
order for a product, the server’s response third-party privacy issues, where compa-
with the price, and the browser’s provision nies use cookies to develop ‘‘one-to-one”
of credit-card data and shipping address marketing campaigns aimed at the individ-
can not be kept together, the system is not ual computer and its user. Primarily a com-
going to work. Cookies are one possible pany will contract with a marketing com-
solution to the problem. pany to develop profiles of its customers
A web server can create a cookie that con- in order to develop customized responses
tains within it essential details of a transac- and advertising. Typically, the marketing
tion, and send it to the web browser along company places cookies on their clients’
with the information that is to be viewed. sites and, since a cookie can be associated
The browser will store the cookie, possibly with any object (e.g., a banner advertise-
permanently, on disk. Every time a connec- ment, download icon, or web page), the pro-
tion is made to a server, the cookies previ- file for individuals’ surfing habits can be
ously sent by that server are sent back to developed for that specific company’s web
it. Thus the server can reconstruct earlier site.
parts of the transaction. The server must Marketing companies also aggregate data
usually record all cookies that it has sent by placing cookies on multiple sites and
for validation and other purposes. thus, when a customer visits any of these
Cookies can persist on the browser’s com- sites, the information pertaining to a par-
puter for an unlimited time, and create ticular cookie is not actually sent directly
a record of the owner’s web-browsing his- back to the company whose site is being
tory that the owner may well rather not visited but rather the information is sent
have exist. There is a widespread percep- to the marketing company, allowing aggre-
tion that cookies can extract private infor- gated data to be stored. The aim of this
mation from a computer and secretly trans- practice is to create an overall score for an
mit it back to a server. There can be no individual that can then be passed on to a
doubt that some insecure systems have client company summarizing an individual
allowed cookies to be released to unautho- customer’s total web usage.
rized third parties, creating a potentially In light of the controversial practice of
severe violation of privacy and contributing cookie aggregation, the Internet Engineer-
to identity theft. Cookies are not active; ing Task Force, an influential body through
they can not ‘‘scan” a computer or erase which internet standards and policies are
its files. It is technically possible for a developed, is examining the issue of ‘‘third-
virus to be transported by a cookie, but party” cookie requests and is developing
some other means would be required to policy positions that may limit or stop such
activate it. practices being employed without the cus-
tomers’ consent.
Business value proposition
Bad publicity surrounding cookies gener- Summary of positive issues
ates dissatisfaction with sites that insist Cookies are a simple device that eases
on using them. Other solutions to the and facilitates internet and electronic-
stateless-server problem may be available commerce transactions. They are univer-
and should be considered. However, there sally available on all platforms and permit
is nothing inherently wrong or privacy- reliable transactions to take place.

84
CPU (central processing unit)

Summary of potentially negative issues rize and be tested on than for any usefully
The negative public perception of cookies descriptive purpose.
has grown from the use of cookies by third In a desktop or personal computer, the
parties who aggregate the cookie data and term CPU is now usually used for the sin-
profile the users, which can be perceived as gle chip (integrated circuit) that provides
an invasion of privacy by computer users the ‘‘brains” of the computer, such as a Pen-
or even an unacceptable security risk for tium IV, a G4, or an Alpha, to name but
many users. a few, even though it also contains a sig-
nificant amount of cache memory and I/O
References control. The term Microprocessor is usually
r D. Kristol and L. Montulli (2000).
used for any CPU that consists of a single
‘‘RFC2965: HTTP State Management integrated circuit.
Mechanism” (proposed standard), There are three distinct schools of
http://www.ietf.org/rfc/rfc2965.txt. thought on overall CPU design philosophy.
r http://www.cookiecentral.com/dsm.htm.
The traditional philosophy holds that CPUs
Associated terminology: Web server, should be designed with as many bells and
Security. whistles as possible, able to perform com-
plex operations such as evaluating polyno-
mials and rotating four-dimensional vec-
tors with single instructions. This is known
as the CISC (Complex Instruction Set Com-
CPU (central processing unit) puter) design, and probably reached its
acme in the now-discontinued DEC VAX
Foundation concepts: Hardware, Computer.
line of processors. An alternate philosophy
Definition: The functional unit in a computer respon-
is that, if the processors are made as sim-
sible for performing arithmetic and logical operations,
ple as possible, limiting their abilities to
carrying out the execution of programmed instruc-
performing simple arithmetic and logical
tions, and controlling the whole system.
tests, they will be able to work much more
quickly and efficiently. This is known as the
Overview RISC (Reduced Instruction Set Computer)
In the days of mainframe computers, the design, and probably reached its acme in
circuitry that performed the arithmetic the now-discontinued Inmos Transputers. A
and control functions of a computer would third philosophy is that CPUs should not
usually occupy at least one quite large be powerful or complex or fast; a large
and impressive looking box, adorned with number of very cheap processors cooperat-
flashing lights and switches. The memory ing on one task can do a better job than
would occupy another nearby box, and one very fast powerful processor. This is
there would be a lot of other large boxes the CSP (Communicating Sequential Proces-
also connected to it, responsible for con- sors) design, and is currently re-emerging
trolling disk drives, magnetic tapes, print- in IBM’s Cell Computers.
ers, and so on. These boxes had to be given Which of the three philosophies is cho-
names, and were called the Central process- sen may have little effect on the end
ing unit (or CPU), Memory unit, and I/O (input user, or even on applications programmers,
and output) unit, respectively. These desig- but, for organizations involved in low-
nations have survived as names for identifi- level (operating systems and languages),
able parts of modern computers, although real-time, or video-game programming, the
they are much less significant, and serve basic design of the CPUs to be used is a
more to give students something to memo- major consideration.

85
Cracking

Business value proposition As CPU capacity expands, the support sys-


The choice of CPU for most users is either tems need to be enhanced to enable the
irrelevant or beyond their control. A pur- CPU to function.
chaser may favor brand X over brand Y, but
References
typically purchasers buy on the basis of per- r C. Thimmannagari (2004). CPU Design:
ceived macro factors such as speed, mem-
Answers to Frequently Asked Questions (New
ory, and price. Unless they are especially
York, Springer).
sophisticated they will typically not be able r http://www.intel.com/technology/silicon/
to balance the trade-offs that are occurring
mooreslaw/index.htm.
even at the macro level, just equating more
speed and memory with higher quality Associated terminology: Cell computing,
and then trading off cost versus quality. Bus, Clone.
The issues surrounding the CPU’s ability to
function effectively for any particular appli-
cation and data are rarely considered. It Cracking
should be noted that CPU manufacturers
do market some chips as more suitable for Foundation concepts: Security, Password.
multi-media purposes than others, but the Definition: Revealing encrypted data, passwords,
same overall constraints on systems perfor- encryption keys, and secret product-activation codes
mance apply. by analysis.
In the future CPU designers may move
from one structure to another, say from Overview
RISC back to CISC systems, or they may The idea of ‘‘cracking codes” is well known.
vacillate somewhere in between. The pri- An analyst pores over encrypted message
mary impact this will have upon the cor- intercepts, theorizes, analyzes, and guesses
porate IT executive will be on the choices until the process is discovered, the code
that they will have for software--hardware is broken, and future intercepts can be
combinations. Should one design approach decoded with much less effort. Cracking is
become technically more advantageous, it a process of intellectual analysis and exper-
does not always follow that software ven- imentation. If one is fortunate enough to
dors will develop code for that platform, find a decrypted message, one can not
and, as consolidation amongst the mega- claim to have cracked the code. Cracking
vendors continues, this may limit a CIO’s in the world of computer security is exactly
technology-purchasing options. the same.
Computer systems use hidden secret
information for a wide variety of purposes.
Summary of positive issues Apart from the obvious reason of wanting
CPUs continue to evolve following Moore’s
to keep confidential data private, there are
law, which states that the number of tran-
more subtle operational reasons. When an
sistors on a chip doubles about every two
authorized user accesses a computer, they
years. CPUs are highly reliable under nor-
need to identify themselves in some way;
mal operating conditions. The CPU archi-
this is usually done with a non-secret user-
tectures are fairly well documented and
name coupled with a secret password. The
supported.
computer system must be able to verify
that the password entered is correct, so it
Summary of potentially negative issues must have some record of the correct pass-
CPUs are built by a limited number of word. Naturally, password files are kept in
manufacturers using proprietary designs. a strongly encrypted form.

86
Cracking

When an expensive software application direction for any of the modern strong
is sold, the customer must have some way methods).
of installing it on their computer, and re- The distinction between a true crack,
installing it after something goes wrong; software that actively discovers passwords
this is easily handled when software pack- or generates activation keys, and sim-
ages are distributed on CDs. Since the early ply publishing a known password or key,
1990s it has been very easy for anyone to is vitally important. Many ‘‘cracks” web
make copies of any CD. This causes a sig- sites do simply provide a list of known
nificant problem for the software manufac- keys, many or all of which have simply
turer, since one legitimate sale may result been copied from people who legitimately
in hundreds of black-market copies enter- bought the software in question. This is
ing into circulation. In the age of high- not a crack, and it is a simple matter for
speed personal internet connections, it is a savvy software manufacturer to make
very common to find the entire contents of each of the published keys invalid. Once a
popular CDs freely downloadable from pub- key-generation method has been discovered
lic web sites. A common solution is for soft- and made public, there is very little that a
ware manufacturers to design their prod- manufacturer can do about it without cut-
ucts so that they will work only after a ting off all of their legitimate customers.
secret identification code has been entered;
the product may communicate with head- Business value proposition
quarters (over the internet) to ensure that The ‘‘industry” of cracking is an aspect of
no two products have been activated with the software culture that developers and
the same key. Manufacturers must design a network owners need to be aware of. Under-
system that allows them to generate a very standing the risks associated with password
large number of different keys, all of which and access-control mechanisms is a vital
will be accepted by the software, whilst aspect of controlling intellectual-property
keeping it impossible for customers to gen- ownership and security.
erate those keys for themselves.
There is a major underground indus- Summary of positive issues
try devoted to cracking all of these appli- Security and control of passwords and
cations of software security. Anyone with access mechanisms is an aspect of systems
an internet connection can freely down- development and ownership that is control-
load cracking kits that do a surprisingly lable by the use of carefully designed and
good job of discovering any poorly chosen conscientiously implemented formal pro-
passwords. For most software that requires cesses to manage access, storage, and use
an activation key, there is software avail- of critical information.
able that will instantly generate new valid
Summary of potentially negative issues
ones. Of course, every major government
Cracking is endemic and a part of the soft-
has a department devoted to cracking all
ware culture; it is unlikely to go away.
of the major encryption algorithms (we
may take some comfort from the fact that Associated terminology: Hacker,
there is no evidence of success in this Encryption.

87
Database

should contain. For example, an organiza-


Database
tion may have a number of different paper
Definition: An organized collection of information. forms used to record basic employee infor-
mation, but they will all contain essen-
Overview tially the same information, with perhaps
A collection of files and records is just raw a few minor differences. It would make per-
data. Once it has been organized, rational- fect sense to design one kind of employee
ized, and formatted so that it can be effec- record that covers all cases, so that they
tively accessed in a uniform way, enabling may all be kept together, and searched with
general queries to be answered by a simple single simple queries. On the other hand,
procedure, it becomes a database. the information kept on customers is likely
Rational organization is of course essen- to be significantly different, and the infor-
tial for any effective record keeping, but mation on supply levels for raw materials
for a database system the demands are even more so.
far more rigorous. A database system will A formalized description is created
be expected to be able to answer queries (using a Data definition language, DDL) for
totally automatically. To answer even the each kind of record that will exist. This
simplest of queries, such as ‘‘what is the description specifies exactly what individ-
salary of the employee named Q. X. Smith ual items of data will be present (e.g., first
in the advertising department?,” mechan- name, last name, social-security number)
ically, without any access to intelligence, and their formats and ranges (a string of
requires an established format to the data. up to 30 letters, a nine-digit number, etc.).
A human searcher could simply leaf thro- Using this description, a database manage-
ugh all the employee records; they would ment system (DBMS) is able to set up the very
easily be able to pick out the employee’s large disk files that will eventually contain
name and their departmental assignment all of the real data, in a manner that makes
from each form, because everybody knows searching simple and fast. All of the data
what proper names look like. It is just descriptions, which are really data about
common sense; nobody would ever confuse data, are known as the meta-data, and are
a department name with a social-security stored as part of the database for future ref-
number even if they were written in unex- erence.
pected places on the form. In database terminology, the entire col-
Of course, computers don’t have com- lection of records of a single type is known
mon sense, and can’t be expected to know as a table (hence a query may specify
what makes a credible name for a human. a search of the ‘‘employee table”); many
For any such query to be answerable, every DBMSs keep each table as an individual
single employee record must be format- very large disk file, but this is not always
ted uniformly, so that the query applica- the case. The records within a table are
tion knows exactly where in each record actually known as records, and the individ-
to find the employee’s name, department, ual data items that make up the records
and salary. There must be no exceptions (names, numbers, etc.) are known as fields
or slip-ups; if an employee’s name is recor- or attributes. Each field has a name (e.g.
ded as ‘‘Research and Development,” there ‘‘social-security number”) and a type (e.g.,
is no way a computer system could react ‘‘nine-digit number”).
appropriately. Once the tables have been set up, they
The first step in setting up a database is must be populated with their data. If the
to decide what different kinds of records it organization already has the data recorded

88
Database

in some electronic form, this will require but by the relationship between data items
some simple but specialized programming in different tables. For example, to record
to perform the data conversion automat- the fact that Q. X. Smith sold 75 brown
ically. If the data does not exist in elec- cows to Amalgamated Meats Incorporated
tronic form, then a long period of manual for $15 000 on July 4, 1993, there might
data entry will be required. Paper forms can be an entry in the employees table indi-
be automatically scanned, and, although cating (amongst other things) that there
scanners now have very good resolution is an employee named Q. X. Smith with
and high reliability, the written informa- ID number 3636; there might be an entry
tion must be converted into a computer- in the products table indicating that there
recognizable digital form. Optical character is a product named Brown Cow with
recognition (OCR) technology is not suffi- UPC 2135712; there might be an entry in
ciently advanced, especially for decoding the customers table showing a customer
hand-written characters, that it could be named Amalgamated Meats Incorporated
relied upon for commercially or legally with reference number 25008, and there
essential data. The time spent verifying and might be a record in the sales table con-
correcting the results of automatic OCR can taining just the six numbers 3636 25008
be as much as the time taken to enter the 2135712 75 15000 19990704. This organiza-
data manually. tion saves a lot of space in the data files,
Once the data is entered, a DBMS appli- and goes a long way toward ensuring data
cation usually provides some default user consistency (see Normalization for further
interface in the form of an interactive explanation), but does mean that the pro-
Query language, the most popular of which cedure for answering simple queries like
is SQL, the Structured Query Language (pro- ‘‘What is the total value of sales made by
nounced ‘‘Ess Cue Ell”; Sequel is a slightly Q. X. Smith?” becomes quite complex.
different thing, a complete DBMS soft- DBMSs, even light-weight single-user ver-
ware package, rather than just the lan- sions for personal computers, are often
guage). SQL is a standardized language for not stand-alone applications, but a client--
all database operations; it covers database server pair. A client provides the front
creation, and record entry and deletion, end, interacting with the user and com-
not just queries, and it can be mastered posing correct queries, perhaps even ren-
by nearly all after some technical train- dering the user’s commands into SQL. The
ing, but it can not be reliably used by command is sent to the Database server, a
unskilled untrained employees, so most second application, not necessarily on a dif-
organizations also require some significant ferent computer, for processing. The even-
expenditure in software development to tual responses are then reprocessed by the
produce a safe and reliable user interface. client for display in a suitable format. The
Commercial DBMS software, especially that use of a database server even when only
intended for use on desktop and similar sys- one computer is involved simplifies design,
tems, often provides a more intuitive and and is a great aid to scalability and acces-
user-friendly graphical interface, but it still sibility. It is a simple matter to change
requires a fair amount of training before it the connection to a remote server when
can be used effectively. conditions demand, but continue to work
Nearly all database systems in use today as before. When a database becomes too
follow the Relational database model. That large to handle locally, multiple database
means that some real information is rep- servers are used, making a Distributed
resented not by records in single tables, database.

89
Database administrator (DBA)

Business value proposition References


r C. Date (2000). An Introduction to
The use of relational databases has become
almost universal in standard database file Database Systems (New York, Addison-
systems. The relational system is at the Wesley).
r R. Elmasri and S. Navathe (1994).
center of modern applications, and systems
such as enterprise resource planning (ERP) Fundamentals of Database Systems (New
systems take full advantage of their capa- York, Addison-Wesley).
r J. Bowman, S. Emerson, and M.
bilities. ERP systems, for example, use a
single database to store and manipulate Darnovsky (2001). The Practical SQL
their data. This is an efficient and effective Handbook (New York, Addison-Wesley).
approach to data management since they Associated terminology: Database
will, for example, contain a single occur- administrator, CIO, Distributed database,
rence of a customer’s name, address, ID, Application server.
etc., and this prevents problems such as
data redundancy, where different addresses
occur for a customer on two or more
databases. Database administrator (DBA)
The range and scope of relational
database technology have evolved signifi- Foundation concepts: Database, ERP, CIO.
cantly since it was formulated by Ted Codd Definition: A database administrator is an expert IT
in 1970. At one end of the spectrum, data- professional who maintains and optimizes the corpo-
base technology is available on a user’s rate databases.
desktop computer through a small data-
base application; at the other end, large Overview
corporate databases operate on specialized The use of data within a corporation is cen-
high-performance servers. To ensure that tral to any information system, and thus
the database created at an organization is the organization and use of that data are
technically correct and matches the needs critical to optimal performance of corpo-
of organization, a database administrator rate information systems. To ensure that
(DBA) is employed to oversee operations. the data within a company is designed
and utilized correctly, companies appoint
a professional database expert known as a
Summary of positive issues Database administrator (DBA), who reports to
Database theory is highly developed in
the CIO.
the academic and practitioner literatures.
The role of the DBA is to act as the autho-
Human resources to develop, run, and
rity pertaining to any data used in the cor-
maintain databases are widely available.
porate information systems. While in small
Certification by vendors is available to
companies the DBA may also be a part of
ensure the training levels of the employees.
the programming team, in larger organi-
zations the DBA heads a data-administra-
Summary of potentially negative issues tion group. The duties of a DBA include
Creation and execution of database systems planning, designing, and implementing
requires specific technical training, and the databases of the organization as well
the technology continues to evolve. Some as enforcing security measures and ensur-
database systems such as the older hierar- ing operational performance. Specific tech-
chical model have fallen out of favor. Conti- nical issues for which the DBA is respon-
nuous updating of skills may be necessary. sible include advising on the selection of

90
Data-flow diagram

database-related products, e.g., data ware- are self-configuring, requiring little outside
houses, and products that interact with intervention.
database systems, e.g., ERP systems. The
DBA is responsible for the physical data- Summary of positive issues
base design, including changing existing DBAs act as the controlling authority for
databases to accommodate modified or any corporate data, database, or system
new applications. Other duties include the relating to data. DBAs are highly trained
training of employees in database tech- and skilled in database technologies. DBAs
nologies and development of contingency- act to ensure the security of their database
planning activities with the CIO and chief systems and to enforce regulatory require-
security officer so that any data corrup- ments.
tion or loss can be remedied quickly and
effectively.
Summary of potentially negative issues
DBAs are required to be highly skilled and
Business value proposition must have on-going educational training.
The DBA performs a very important tech-
Employment of a weak or under-skilled
nical function within any IT organization,
DBA can lead to catastrophic database prob-
managing and developing the corporate
lems. Some systems require very specific
database so that it runs efficiently and
skills not typically found in a DBA, e.g.,
effectively. DBAs should be technically
knowledge of ERP systems.
highly skilled individuals who understand
the inner workings of database systems. Reference
This takes extensive training and knowl- r C. Mullins (2002). Database
edge of the system upon which they are Administration: The Complete Guide to
working. It is essential that the train- Practices and Procedures (New York,
ing of the DBA is regularly enhanced, Addison-Wesley).
since the technologies of the systems upon
which they work continue to evolve. Simi- Associated terminology: Data Protection
larly, the regulatory environment in which Act, Sarbanes--Oxley, HIPAA.
the corporation and its systems operate
also changes and DBAs need to be edu-
cated in their regulatory requirements, e.g., Data-flow diagram
Sarbanes--Oxley and HIPAA.
The role of the DBA is undergoing a Foundation concept: Software development lifecycle.
change in terms of the nature of the Definition: Data-flow diagrams are graphical models
databases upon which they operate. The thatshowtheflowofdatathroughasystem,theexter-
advent of complex packages such as ERP nalsourcesanddestinationsofdata,theprocessesthat
systems requires specialized skill sets in transform the data, and the locations where the data
order to be able to understand those is stored.
systems and their database functionality.
The databases of these systems are not Overview
intended to be managed in the same The concept of a logical data-flow model,
way as the traditional databases of the usually known simply as a data-flow dia-
past. ERP databases are highly specialized gram (DFD), is used in the analysis and
and extremely complicated, so customiza- design phases of the system’s lifecycle as a
tion of their structures, even by experts, communication tool to understand, verify,
is highly ill-advised because the systems and validate the system requirements.

91
Data-flow diagram

A DFD is constructed using the following eling are performed. Traditionally, using
symbols: structured techniques, process modeling
was performed prior to data modeling.
A process (e.g., check for stock availability,
However, today, due to the influence of
process credit-card payment):
information engineering, data modeling
precedes process modeling.

Business value proposition


DFDs have been a popular and successful
means used by systems analysts since the
late 1970s. They provide an intuitive visual
view of the system through ‘‘box-and-line”-
An external entity (e.g., customer, vendor):
based constructs as opposed to earlier text-
based descriptions. These constructs can be
easily interpreted by system users and facil-
itate the analyst--user discussion and com-
munication necessary to understand and
validate the system requirements.
The models can not only be used to cre-
A data store (e.g., customers, products, ate new systems but are also used in the
payments): redesign of business processes, and act as a
basis for a business process re-engineering
effort. DFDs are strongly associated both
with the ‘‘systems analysis and design” appro-
ach to capturing system requirements and
A data flow (e.g., credit-card information with the ‘‘informal methods” school of sys-
traveling between a customer and a payment tems development. The approach is well
process): known and widely used and has been a core
element in the design of transaction pro-
cessing systems since the 1970s.
When performing process modeling (cre-
ating DFDs), the systems analyst starts at Summary of positive issues
the highest level of abstraction, and once DFDs are a widely known means for mod-
all the processes and data flows have been eling used by systems analysts in analy-
mapped at that level, the analyst would sis and design. They are intuitive to use
create a lower-level DFD for each high- and are easily understood by non-technical
level process by ‘‘exploding” the process personnel. This modeling method allows a
to show greater detail (i.e., sub-processes high-level view of a system to be created
and their data flows). This is repeated until and then exploded down to lower levels of
all processes have been ‘‘exploded” and refinement. The model fits well with infor-
the lowest level of detail desired achieved, mal design methods such as the Waterfall
resulting in a set of ‘‘primitive” processes. method and is a traditional central compo-
Traditionally DFDs were used to model nent of methodologies for systems analysis
both existing and proposed systems. Cur- and design.
rently DFDs are used to model only the pro-
posed system. Another difference between Summary of potentially negative issues
traditional and current modeling concerns The data-flow method is becoming dated
the order in which data and process mod- and is being superseded by models such

92
Data mining

as Semantic data models, Object models, and the product consumed, or did they buy
the Unified modeling language, which facil- more of another product in association
itate the mapping of process models into with the promoted product?
constructs that are more rigorous and are The theory and practice of data mining
easier to map onto more modern program- continue to develop and evolve. A relatively
ming environments such as Object-oriented new branch of data mining is Text mining.
programming. Much of the data on the internet and in
databases is actually textual in nature
Reference
r T. DeMarco (1978). Structured Analysis and rather than numeric (e.g., web pages con-
taining normal English prose). This has led
System Specification (New York, Yourdon
to a wide variety of techniques being used
Press).
to analyze the text, its patterns, and its con-
Associated terminology: UML, tent. The goals for undertaking text mining
Normalization. are the same as those for quantitative data
mining, namely the identification of new
and novel data patterns within seemingly
Data mining heterogeneous data.

Foundation concept: Database. Business value proposition


Definition: Data mining is the technique of searching Data mining is the process of analyzing
data for “patterns of interest” to the miner. data by detecting patterns. For example,
data mining performed by the WalMart
Overview company revealed that, when a hurricane
Data mining, sometimes referred to as Knowl- warning is issued by the US Weather Ser-
edge discovery in databases (KDD), is the use vice, demand for strawberry-flavored break-
of algorithms to search through data sets fast tarts is dramatically heightened in
looking for interesting patterns or confir- stores in the affected area (‘‘Private FEMA,”
mation of a predicted property. Data min- Wall Street Journal, September 10, 2005).
ing uses a complex set of techniques drawn Armed with this knowledge, WalMart can,
from a variety of computer science dis- immediately following (or even preced-
ciplines, including Databases and Artificial ing) a hurricane warning, order, ship, and
intelligence, and from statistics. shelve more breakfast tarts, providing the
The data upon which the mining activ- customer with the product they desire
ity is performed can be in any form; how- in a just-in-time-delivery manner, driving
ever, a Data warehouse in which transac- greater profit for the company. Data mining
tional data has been cleaned, formalized, in this instance directly contributes to Wal-
and archived can make the process more Mart’s bottom line, allowing them to adjust
straightforward. In principle, data mining their product-demand levels and achieve
uses a variety of techniques from database high completed-sales rates with a very short
theory and machine learning to examine product shelf-life/cycle-time.
data sets looking for patterns that display Data mining (including text mining)
‘‘interesting relationships.” For example, a allows organizations to extract potentially
supermarket may establish the success or rewarding information from their histori-
failure of a promotion campaign by exam- cal data sets. These techniques enable CIOs
ining the sales data, answering questions and organizations to leverage data sets that
such as ‘‘did the target group buy more may have fixed storage costs associated with
or less of the product?” Did they switch them due to compliance and regulatory
brands? Did they increase the volume of issues.

93
Data pool

Summary of positive issues specifically for the purpose of allowing


Data mining enables new and potentially third parties data access. An example of
important and profitable information to be this is the data pool operated by the
extracted from a data set. The technolo- National Snow and Ice Data Center (NSIDC),
gies enable a wide range of data sets to which archives and distributes data about
be examined, including text. Data mining snow cover, ice sheets, and related topics.
can be used on traditional databases and The NSIDC has over 500 data sets that
data warehouses. Vendors are available to can be accessed online and retrieved via
provide consulting, tools, and support. The FTP, ranging from data on ‘‘Shallow bore-
area is supported by an academic and prac- hole temperatures, Ilulissat, Greenland” to
tical literature. ‘‘Summer surface hydrographic measure-
ments in the Laptev Sea, 1980 to 2000.”
A key aspect of data pools is that the
Summary of potentially negative issues
data is reliable, clean, and has well-defined
Data mining and text mining can be com-
fields.
plex and challenging from the statisti-
A second type of data pool can be defined
cal and technological perspectives. Data
as a large data superset comprised of data
mining can also be resource-intensive and
subsets containing the product meta-data
requires a mixed-staffing initiative consist-
of an individual company’s products. The
ing of technologists and staff knowledge-
pressure for organizations to become more
able in the processes captured in the data
connected in terms of their supply chains
sets. Data mining is still an emerging tech-
has led many to adopt common technolo-
nology.
gies for communicating, such as the use of
References XML over the internet. However, for a large
r K. Hirji (2001). ‘‘Exploring data mining manufacturer, the task of notifying all of
implementation,” Communications of the its customers of all its product changes or
A.C.M., Volume 44, No. 67. modifications each and every time a change
r I. Witten and F. Eibe (2005). Data is made would be a daunting task using a
Mining: Practical Machine Learning Tools peer-to-peer messaging system.
and Techniques (San Francisco, CA, The need to share data has led to the
Morgan Kaufmann). concept of a Data synchronization hub. Such
a hub allows companies to place data per-
Associated terminology: Business
taining to items in an accepted standard
intelligence, CIO.
format such as EAN.UCC. The data is then
placed in a data pool. This data pool is usu-
ally administered by a third party (but for
large companies this can be done in-house)
Data pool and information pertaining to the data
set is sent to a Master data registry (MDR)
Foundation concept: Database. such as the global synchronization net-
Definition: A data pool is a data set specifically created work GS1 or the WorldWide Retail Exchange
to allow the data to be retrieved by third parties. (WWRE), which holds the information and
records the location of each item’s data
Overview pool.
A Data pool differs from a standard database Customers or other companies search
in that it is a data set created by an orga- an MDR through their chosen data pool
nizational entity, separate from its inter- provider and notify the data pool provider
nally focused, live transactional systems, of the items they wish to subscribe to.

94
Data Protection Act (UK only)

The MDR operator uses a Synchronization r http://www.transora.com/.


engine to automatically and continuously r http://nsidc.org/.
synchronize the data between the two com-
Associated terminology: Client, Server,
panies. For example, a component man-
Data mining, XML.
ufacturer may place its product specifi-
cations in a data pool; then, when they
change the specification of one component,
all parties who have subscribed to the data
pool will be notified that this change has
Data Protection Act (UK only)
occurred. Foundation concept: Data.
Definition: The UK Data Protection Act defines the
Business value proposition obligations for owners of personal information (the
Data pools allow organizations to simplify data processors) and the rights of the data subjects.
their data collection, modification, and dis-
semination functions. For a single entity Overview
providing third-party access to clean data The Data Protection Act (DPA) has estab-
this can be done straightforwardly through lished a standard of good practice for ‘‘any-
an internet-based FTP connection. one processing personal information.” It is
The use of data pools through data syn- based upon eight principles: the data must
chronization hubs allows organizations to be
manage their vendor--customer data rela-
tionships in a cost-effective, efficient, and (1) fairly and lawfully processed;
timely way. (2) processed for limited purposes;
(3) adequate, relevant, and not excessive;
Summary of positive issues (4) accurate and up to date;
FTP-based data pools allowing third parties (5) not kept longer than necessary;
to download data upon demand are relativ- (6) processed in accordance with the
ely easy for individual entities to establish. individual’s rights;
For commercial organizations, standards (7) secure; and
such as ‘‘EAN.UCC” have been established, (8) not transferred to countries outside
to which companies can map their data. the European Economic Area unless
Having achieved standardized data for- the country has adequate protection
mats, there are various third-party data for the individual.
pool providers through which companies
can manage their data-pool requirements. These principles are combined with six
Several very large data synchronization conditions, one or more of which must
companies have been established to manage have been met in order for the per-
the data requirements of large commercial sonal information to be considered fairly
organizations. processed:

Summary of potentially negative issues (1) the individual has consented to the
The conversion of data to the UCC standards processing;
can be a challenging resource-intensive (2) processing is necessary for the
problem. performance of a contract with the
individual;
References (3) processing is required under a legal
r http://www.worldwideretailexchange. obligation (other than the one imposed
org/cs/en/index.htm. by the contract);

95
Data quality audit

(4) processing is necessary to protect the Associated terminology: Law cross-


vital interests of the individual; reference.
(5) processing is necessary to carry out
public functions, e.g., the
administration of justice; and Data quality audit
(6) processing is necessary in order to
pursue the legitimate interests of the Foundation concepts: Database, Information lifecycle
data controller or third parties (unless management.
it could unjustifiably prejudice the Definition: A data quality audit is a check to ensure
interests of the individual). that a data set actually matches the data definitions
originally established for it.
Sensitive personal data includes racial or
ethnic origin, political opinions, religious Overview
or other beliefs, trade-union membership, A data quality audit (DQA) is an assessment
physical or mental health concerns, sex of a data set with respect to the original
life, and criminal proceedings or convic- data definitions for that data set. Audits
tions. The data subjects have seven rights can be performed upon any data set, rang-
detailed under the DPA: ing from a spreadsheet to the database of
an ERP.
(1) the right to access their data;
Many organizations have been compil-
(2) the right to prevent processing;
ing data into databases for over 30 years,
(3) the right to prevent processing for
and many ‘‘legacy” databases continue to
direct marketing;
be used in corporate computing systems.
(4) rights in relation to automated
The quality of the data entered into these
decision-making;
systems is in many cases undetermined.
(5) the right to compensation;
For example, if a company has several
(6) the right to rectification, blocking,
databases scattered around its organiza-
erasure, and destruction; and
tions, it may be that customers have differ-
(7) the right to ask the Commissioner to
ent names or addresses recorded in differ-
assess whether the act has been
ent places: Bob Smith, R. Smith, and Robert
contravened.
Smith may all be the same person; Fort
Criminal offenses are detailed for cir- Lauderdale and Ft. Lauderdale may both be
cumventing the act and include noti- the same place. In order to prevent this
fication offenses, procuring and selling type of situation, a database administrator
offenses, and electronic-communications may undertake a DQA.
offenses. The act is overseen by the Infor- One methodology for performing a DQA
mation Commissioner’s Office, a branch of has been proposed by Gonzales, and it is
the UK government. representative of the methods in general.
His model has six stages that are iteratively
Business value proposition applied to different data categories (e.g.,
The act enables businesses to operate under customer data, account data, and trans-
a regulatory framework and clearly sets out actional data). Stage 1 is an examination
their obligations to the data subject. of the data definition language associated
with the data to be examined. Stage 2
References requires the analyst to identify a sample
r Data Protection Act 1998 Chapter 29. data set, whose size will reflect the com-
r http://www.informationcommissioner. plexity of the database and its structures;
gov.uk. this is necessary because some data sets

96
Data warehouse

can be many terabytes in size and it would the historical data to be dumped and writ-
clearly be impracticable to survey the whole ten off or to be scrubbed through the ETL
data set. Stage 3 involves the establish- process; this is a great opportunity to re-
ment of a base rule set to use in conjunc- establish the integrity of the data set.
tion with the sample data set (e.g., the
volume held in a tanker transporting fuel Summary of positive issues
oil can not be less than zero or greater DQAs facilitate the establishment of statis-
than the tanker’s capacity). Stage 4 involves tical data showing the quality of a data set.
expanding the rule set, establishing con- Tools, methodologies, and consultants are
fidence levels, and designing the output available to support the DQA process.
report structures. Stage 5 involves encour-
aging the user community to examine and Summary of potentially negative issues
modify rules, recalculate the confidence A large amount of resources may be needed
intervals, and adjust any reports that will to audit large old data sets and systems.
result from the audit. Finally, in stage 6 the Syntactic errors in data are easier to iden-
analysis is performed, and results are col- tify than semantic ones. Data sets are no
lected and examined. longer just numerically based and may
A poor result in a DQA should lead an include heterogeneous data sets, e.g., XML
organization to re-examine the data-entry files, pdf files, graphical images, and audio
and process models around which the sys- files, which are more difficult to audit.
tems are based. It is possible to clean up the
data and re-establish data integrity, but this Reference
r M. Gonzales (2004). ‘‘The data quality
can be a difficult and expensive proposi-
tion that may also involve cutting out data audit,” Intelligent Enterprise, July.
that is out of bounds and can not be cor-
rected through knowledge available from
other data sources. Data warehouse
An alternative to modifying a data set is
to establish a new data set through a pro- Foundation concepts: Database, ERP, File server.
cess known as ETL (extraction, transforma- Definition: A data warehouse is a file server that
tion, and loading of data), a process that contains data that has been extracted, transformed,
is usually related to the establishment of a and loaded into it from other databases. The data is
Data warehouse. then accessible for analytic programs, commonly
known as business intelligence software, to run upon
Business value proposition it. The results are frequently presented in a graphical
DQAs are essential components of a cor- format.
poration’s data management process. The
older the data set, the higher the proba- Overview
bility that it contains many errors. Modern The term Data warehouse is used to describe
information systems such as ERP systems, a dedicated facility for aggregation and
which have only one instance of a data item storage of data. Data warehouses are repos-
in its singular database, attempt to mini- itories of data that has been extracted from
mize the opportunity for data corruption corporate systems. The data undergoes a
to occur, both in the form of data-entry con- transformation process (or is Scrubbed as it
straints and through detecting errors in the is sometimes called) in which the data is
applications themselves. checked for type and consistency, before
The creation of a new data warehouse or it is loaded into the relational database
the establishment of an ERP system allows of the warehouse where the information

97
Decision support system (DSS)

is stored. This process is known as ETL way allows for focused one-time inquiries,
(extraction, transformation, and loading). with trends and exceptions being processed
Once the data has been loaded, it is avail- easily, and thus facilitates rapid decision
able to be manipulated and examined so making. This type of analysis would be
that analysis may be performed upon it. expensive and very difficult in traditional
Data warehouse analytic software allows databases.
the data to be examined from multiple
perspectives, termed Multi-dimensional data Summary of positive issues
views (or Slicing and dicing). This analysis is Data warehouses are secure data repos-
also frequently termed Cubical analysis since itories whose data has been ‘‘scrubbed”
the data is often manipulated into a three- and cleaned, resulting in a database with
dimensional cube, but more dimensions high integrity. The data structure used to
are possible. For example, a data warehouse store the data, namely the data cube, facil-
for a computer manufacturer may be struc- itates rapid data analysis and a highly
tured in three dimensions, of which the flexible approach to business intelligence
first is a dimension that is based upon the querying.
sales markets: north, west, south, and east.
A second dimension may be based upon Summary of potentially negative issues
product type (PC, PDA, laptop, and server), Data warehouses require dedicated servers
while the third dimension may be sales by if they are to be most effective and the
quarter. This would provide the basis for data extraction, transformation, and load-
data analysis by ‘‘slicing and dicing” these ing (into the data cube itself) can be diffi-
three dimensions. The programs that run cult, thus a careful return on investment
on the data warehouse allow the analyst analysis needs to be undertaken when
to drill down on the data, perhaps start- developing such systems.
ing with an analysis of sales for all the
References
markets the company serves over one year, r W. Inmom (1996). Building the Data
then drilling down to a quarter sales
Warehouse (New York, John Wiley and
period, then drilling down to a given
Sons).
month, and then again down to a given r E. Vitt, M. Luckevic, and S. Misner
week or day.
(2002). Business Intelligence (Redmond,
WA, Microsoft Press).
Business value proposition
The use of a data warehouse allows a busi- Associated terminology: Business
ness to have a central repository for his- intelligence, Data mining, Database, OLAP,
torical data. A data warehouse removes the ETL.
need to keep the data in other online sys-
tems such as the organization’s ERP. The
data warehouse facilitates the backup and Decision support system (DSS)
security of the data. The data after ETL will
be consistent, verified, and archived. The Foundation concept: Business intelligence.
executive information system software or Definition: A decision support system is an application
Business intelligence software that runs on through which managers and executives can analyze
the data warehouse allows the data to be corporate data.
examined at multiple levels of detail and
the results to be presented in a variety Overview
of formats, including graphical ones. The Decision support systems (DSSs) are the
ability to quickly manipulate data in this predecessors of what has become known

98
Denial-of-service attack

as Business intelligence or Business analytics. Summary of positive issues


They consist of three components, a dedi- DSSs allow individuals or groups to con-
cated data warehouse, a modeling system, sider a problem through the use of a
and an interactive user interface, usually computer system that provides the deci-
with graphical capabilities. DSSs are also sion makers with data, analysis tools, and
the basis of Executive information systems, reporting capabilities. This allows better
analytic systems developed to address the decision-making within information-rich
needs of senior executives, with the poten- environments or when critical analysis of
tial to display results in the form of an multiple data sets is required in light of
‘‘executive dashboard.” given decision criteria.
A collaborative form of DSS, known as
Group decision support systems (GDSSs) has Summary of potentially negative issues
also been developed. The GDSS concept While military and government institu-
evolved in the 1980s and has also been tions have adopted the technology, the
known as War rooms, Decision support labora- expense and dedicated resources some-
tories, and DDS conference rooms. GDSSs help times required have limited their adop-
managers or executives to solve a problem tion within the wider commercial world,
by bringing them together, either physi- so DSSs tend to reside in academic and
cally or virtually, to make decisions that are research institutions.
supported by their colleagues and by the
most appropriate data and analytic tools Reference
available. r D. Power (2003). ‘‘A brief history of
decision support systems,”
http://DSSResources.COM/history/
Business value proposition dsshistory.html.
DSSs have evolved from the academic-based
Associated terminology: Business
arena into the commercial world, and have
intelligence, Data warehouse, ERP.
become known as business intelligence or
business analytics; they are used in con-
junction with ERP and data warehouse sys-
tems. Denial-of-service attack
The field of GDSSs also continues to
evolve, but is mainly situated in academic Foundation concepts: Security, Network.
forums. The use of a GDSS allows a large Definition: An attack that is intended to prevent
group of people focused upon a particular legitimate users from accessing or having full use of a
problem to be connected together. Led by computer system, rather than attempting to destroy,
a human facilitator, the members of the steal, or modify information.
group interact with the system simultane-
ously, developing ideas that the system can Overview
filter, sort, and analyze. The group, in con- Viruses, worms, and Trojan horses have
junction with the facilitator and the sys- a clear malicious intent; they attempt to
tem, then acts to rank and develop the destroy or disrupt computer resources.
ideas further. The systems may have the Spyware attempts to steal information,
ability to make the input from each par- and adware hijacks a computer and uses it
ticipant anonymous, which can strengthen as an advertising outlet. A denial-of-service
the arguments, overcoming the problem (DoS) attack is a fourth form of digital
of rank or dominant group members over- assault. It does not attempt to cause per-
whelming a discussion. manent damage or to steal anything, just

99
Denial-of-service attack

to prevent proper use of a system. Because and is especially effective because hun-
of this, many self-righteous DoS attackers dreds of computers can be used in uni-
manage to convince themselves that they son to create an unsurvivable flood on a
are not really doing anything illegal. whole target network. The messages that
A DoS attack is often designed to exploit cause a flood attack can be traced back to
a known error in a popular operating sys- their source, but the source often turns out
tem, and, since there are so many known to be an innocent third party’s unsecured
errors in popular operating systems, there home computer, which may have seemed
is a very wide variety of attack methods to be running quite slowly, but would have
available to the attacker. A very popular shown no other sign of having been com-
attack a few years ago was known as the promised.
‘‘ping of death,” in which a deliberately
invalid message was sent using the ICMP Business value proposition
protocol, and, due to a bug in the operat- There is no positive aspect for businesses
ing system, receiving that message would regarding a DoS attack; the only thing that
immediately crash the target computer. No organizations and individuals can do is
permanent harm done, just reboot and it’s guard against them. This entails having a
running again, but all it took was a very strong security policy in place and under-
simple program, easily downloaded from standing what the attacks are and how they
the internet, to send the same message break into an organization or take over a
automatically every 5 minutes, rendering computer.
the target computer completely useless. To A strong and continuously updated fire-
this day, many installations have disabled wall system is the best form of defense,
their ‘‘ping” service even though the bug which should be combined with regular
has long since been fixed. Many responses systems checkups to ensure that the sys-
to attacks seem to be based more on super- tem has not been taken over and become
stition and voodoo than on sense. Some a zombie controlled from outside the orga-
attacks make use of known bugs in net- nization. While it is important that organi-
worked applications, where reception of a zations maintain an understanding of the
particular message might result in the tar- latest DoS strategies employed and defend
get application filling a whole disk with themselves against them, it is also impor-
bogus files, or embarking upon a non- tant that, when a technique has been
terminating computation. defeated, the company assess whether sys-
Another form of DoS attack requires no tems ‘‘lockdowns” are still necessary. Such
knowledge of any operating-system flaws was the case when the ‘‘ping of death”
at all. Simply flood the target computer was a peril to systems; organizations per-
with as many internet-connection requests manently closed down their ‘‘ping” ser-
as you can send; it will be unable to process vice, denying access to a useful utility, even
them all, so normal requests from legiti- though the bug through which the DoS
mate users will be crowded out. attack was being made has subsequently
A common means of mounting DoS been resolved.
attacks untraceably is to take over an
unwitting third-party computer, or prefer- Summary of potentially negative issues
ably a lot of them, through virus and DoS attacks can be problematic for organi-
Trojan-horse software, and use those com- zations and result in systems outages.
puters to act as an army of Zombie com- These outages may be prolonged and per-
puters sending the flood of messages. This sist until a fix has been found by the
is called a Distributed denial-of-service attack, operating systems vendor, security software

100
DHCP (Dynamic Host Control Protocol)

vendor, or other security specialist, depend- puters and other network devices that are
ing upon the type of attack. normally connected at any one time, but
not enough for every single device that
Reference
r J. Mirkovic, S. Dietrich, D. Dittrich, and they are responsible for. This is a very com-
mon situation for internet service providers
P. Reiher (2004). Internet Denial of Service:
(ISPs), for which only a fraction of their cus-
Attack and Defense Mechanisms
tomers will be actually using the internet
(Englewood Cliffs, NJ, Prentice-Hall).
at any given time.
Associated terminology: Virus, Trojan When a DHCP-enabled device first
horse. attempts to access the network, it must
negotiate with a DHCP server for an
address. If there is an IP address available,
DHCP (Dynamic Host Control Protocol) the DHCP server will allocate it temporar-
ily to the requesting device, giving it
Foundation concepts: Internet protocol, Network. a Lease on the number with a specified
Definition: A system for temporarily allocating IP expiration time. The device may then
addresses to computers on an as-needed basis. use that IP address as its own. When the
device disconnects, or when the lease
Overview expires, the DHCP server simply reclaims
Every computer or network-enabled device the allocated IP address, and the device’s
connected to the internet must have an network communications stop working. A
IP address (internet-protocol address). An IP device may request a lease renewal at any
address is four small numbers separated by time before expiration, and it is granted if
dots (such as 221.39.107.82), which uniquely demand is not too high.
identify the computer, distinguishing it DHCP allows a large set of computers to
from all others. All information transmit- share a not-so-large set of IP addresses, but
ted over the internet uses IP addresses to is successful only if demand for addresses
specify its destination; a computer without does not exceed the number available. For
an IP address will be unable to receive or end-user systems DHCP provides a generally
transmit anything. satisfactory service, but it is not at all suit-
Long runs of consecutive IP addresses able for systems that play host to perma-
are allocated by a central authority (see nent or long-term network services, such
Domain name) to individual organizations, as web servers and email servers. In order
and those organizations are responsible for to be found by their clients, servers must
allocating the individual IP addresses to the have a known, and therefore non-variable,
individual computers within their domain. IP address.
In the early days of the internet, this Network address translation (NAT) is ano-
scheme worked very smoothly, but demand ther technology that may be used either in
has reached such levels that there are conjunction with, or instead of, DHCP; it
not enough IP addresses to go round. IPv6 helps to solve the problem of having too
(the new version of the internet protocol, few IP addresses available to meet demand.
in which addresses are much longer) will
eventually relieve some of the problems, Business value proposition
but is still not in general use. Network administrators use DHCP to over-
DHCP, the Dynamic Host Control Proto- come the problem of not having sufficient
col, is a very popular solution. An orga- IP addresses available to them to allocate
nization may have enough IP addresses one IP address per device. Thus, assum-
allocated to it to cover the number of com- ing that the network is to be connected

101
Digital

to the outside world, the network adminis- Reference


trator will take the addresses available and r R. Droms and T. Lemon (2002). The DHCP
allocate one or more of these addresses to Handbook (Indianapolis, IN, Sams).
a specific server that is connected to the
Associated terminology: ISP, Protocol.
internet or external network. This server
then acts as the gateway for internet traf-
fic. The remaining IP addresses can then
be allocated to other servers, known as Digital
DHCP servers. These servers operate over
subnets, whose clients may request an IP Foundation concepts: Bit, Binary.
address. These are then given to the client Definition: A digital system is one that is based upon
on a temporary basis, known as a lease, a sequence of events, symbols, or data.
and, when the lease is about to expire, the
client can then request an extension to that Overview
lease. Digital data simply means data represented
The DHCP servers allocate, maintain, as a sequence of symbols, rather than as
and service requests across the network in a direct physical analog. For example, the
a way that is efficient, cost-effective and length of a pencil could be represented as
resource-preserving. Large IP address blocks a proportional voltage, with one volt for
are extremely difficult to obtain and are one inch; that would be an analog repre-
costly when available, especially in those sentation. Alternatively, one could measure
countries which were early IP adopters. the pencil and write down its length, 6.31
DHCP also frees the network administra- inches; that is a digital representation.
tor from having fixed network addresses Working directly with analog representa-
for each client in their network, thus facil- tions may seem more natural and accurate,
itating the management of the system, and and indeed this technology was invented
enabling devices and systems to be added long before digital computers, but analog
without worrying about re-allocating IP computation has severely limited accuracy,
addresses. speed, and reliability. Today, digital compu-
tation is universally superior.
Summary of positive issues Anything written as a series of symbols
DHCP is a cost-effective method of allocat- (e.g., 6.31, ‘‘cat,” )
ing IP addresses to devices on a corporate is already in digital form, but an addi-
network. DHCP is well known and easy to tional very simple encoding step (e.g.,
implement, and the operating systems that c → 01100011, a → 01100001, t → 01110100)
run servers usually have DHCP capabilities is usually taken to reduce the symbols to
within their toolsets. binary numbers, which are then concate-
nated to produce a pure bit-stream (e.g.,
Summary of potentially negative issues ‘‘cat” → 011000110110000101110100), which
Networks that have heavy traffic from a is the internal form used by a modern dig-
large number of devices that are continu- ital computer. Every measurable quantity
ally online may place demands upon the can thus be reduced to a stream of 0s and
DHCP servers that result in some devices 1s.
being without an IP address and slowed Even something as seemingly non-digital
networks due to the volume of device- as music is simple to digitize. Sound is
generated traffic. DHCP is not suitable nothing but a varying pressure wave in the
for devices that require specific static IP atmosphere; the changing pressure may
addresses. be monitored and digitally measured at

102
Digital cash

frequent intervals, producing a stream of cash is really just a token of barterable


numbers that represents the continuously value) are traceable, and leave a trail of
varying sound pressure. If the samples are records that allow minute details of a pri-
taken frequently enough (44 000 per second vate life to be reconstructed.
is usually considered sufficient), they can In many ways, the abolition of cash
be played back through a digital-to-analog might seem to be a good thing: the prac-
converter connected to a loudspeaker, and titioners of kidnapping, bribery, and black-
the human ear will be unable to detect the mail would be dealt a serious blow if there
difference. were no untraceable forms of currency, and
since 2001 the public has shown a remark-
Business value proposition able level of acceptance toward the loss
In practicality, all modern business com- of traditional privacy rights. However, it
puting systems are digital and based upon should be borne in mind that cash-like
the binary representation. This allows the things can not be abolished; if there were
builders of electronic devices to ensure no such thing as cash, kidnappers and
the successful interconnection and inter- crooked politicians would simply demand
operability of devices. payment in gold or some other substance
with intrinsic value.
Summary of positive issues Digital cash would certainly be a popu-
Digital computer systems based upon base- lar invention: something that exists only in
two computation are the standard mecha- digital form, so it can be carried around or
nism for representing data. transmitted instantaneously between com-
puters in any quantity without effort. It
Summary of potentially negative issues could be kept encrypted until it is needed,
Digital base-two representations, bits, have so the possibility of theft would be much
required a set of mechanisms to represent reduced. It could be used to buy things
and manipulate the data at a higher level without the payer’s identity being revealed.
such that the data can be understood by It could be backed up like any other data,
humans. and so made immune to destruction by fire
or other disasters.
Associated terminology: ASCII, Audio, Such benefits come at a price. Digital
Video, Compression. cash, in a form at least, has been success-
fully invented. It consists of a group of
exceptionally complex (and patented) pro-
Digital cash tocols based in part on public-key encryp-
tion and digital signatures. The fact that it
Foundation concepts: Encryption, e-Commerce, has been known since 1990, and has not
Security. yet caught on, and in fact is still almost
Definition:Ameansofprovidinganonymouspayment unheard of, is testament to the fact that it
electronically. is not very convenient to use.
The complexity of digital-cash implemen-
Overview tation is due to the seemingly impossi-
Even in this age of credit cards and elec- ble demands placed upon it. Once some-
tronic commerce, cash still plays a very thing is in digital form, it is impossible to
important role in society. Cash provides pri- prevent it from being duplicated. If some-
vacy, allowing people to buy and sell things body obtains a digital cash token, they
without anyone else finding out. All other could send that token to a hundred dif-
forms of payment (excluding barter, and ferent vendors in payment for different

103
Digital cash

things. Since copying can not be prevented, governments may enjoy the ability to iden-
security is instead based on preventing tify all taxable incomes, the impact that
double-spending. Somehow, when a vendor such a transaction trail would have upon
receives a digital cash payment, they must society has limited the political will to
be able to check immediately that it has not implement such systems.
already been spent, and at the same time The current limits of deployment of dig-
ensure that it can not be spent again in ital cash include the implementation of
the future. To be truly cash-like, it would systems for electronic payment from bank
still have to be possible for the receiving to bank (e.g., to pay a water bill or the
vendor to spend it legitimately, so illegiti- mortgage), which are in effect wire trans-
mate double-spending is an extremely dif- fers, systems for electronic payment from
ficult problem to overcome. bank to a credit card (e.g., this can be used
Also, to prevent double-spending, all to make payments related to online auc-
pieces of digital cash must in some way tions), digital-cash-based debit cards (e.g.,
be different from all others. If they were in the United States welfare recipients use
not unique, it would be logically impossible electronic benefits transfer (EBT) systems in
to distinguish genuinely different tokens which a debit card has replaced the use
from illicitly double-spent copies. But, if all of paper ‘‘food stamps”). However, none of
tokens are unique and digital, it would be these systems approach true digital cash
possible for the issuer to record all tokens and the probability of large-scale use of
issued together with the identity of the per- such systems in the near future is low.
son to whom they were issued, thus intro-
ducing traceability and missing the point
of digital cash. Summary of positive issues
Digital cash provides the ability to trans-
Business value proposition fer money electronically from one person
The processes associated with implement- to another. The use of identifiable digital
ing digital cash have been in existence cash allows governments and individuals
since the 1990s; however, they have to trace payment trails. Anonymous digital
received very little publicity due to their cash can reduce the complexity of transac-
complexity and the need to change the pay- tions, and increase customer confidence.
ment behavior of a large group of the popu-
lation. The two types of digital cash, anony- Summary of potentially negative issues
mous and identified cash, both have many With identified digital cash governments
issues associated with them. can monitor all digital transactions, and
Anonymous digital cash allows people small payments may incur disproportion-
to transfer monies without a trace being ate transaction costs. With anonymous cash
left, which would be a problematic issue the lack of a transaction trail may facilitate
for governments that rely upon a paper illegal activities. There are significant prob-
trail to prevent unlawful activities. Physical lems in developing processes to prevent
cash makes it hard for a criminal to carry double-spending, ensuring proof of owner-
around millions of dollars, pounds, or yen ship, and backing up cash, and as a result
without attracting attention. the digital cash protocols are extremely
Identified cash has problems associated complex.
with it that stem primarily from the fact
that every transaction would be traceable, References
and complete profiles of an individual’s r D. Chaum (1983). ‘‘Blind signatures for
spending habits could be created. While untraceable payments,” in Advances in

104
Digital certificate

Cryptology CRYPTO ’82, Lecture Notes in always be possible for someone to subvert a
Computer Science, ed. D. Chaum, R. communication channel, temporarily redi-
Rivest, and A. Sherman (New York, recting accesses to the public directory to
Plenum Press). a fraudulent one of their own.
r D. Chaum (1985). ‘‘Security without Digital certificates are an attempt to
identification,” Communications of the solve this most serious problem. The under-
A.C.M., October. lying idea is that, if there is one entity
anywhere that can be fully trusted (per-
Associated terminology: Cracking, haps a saintly person with a lot of secu-
Protocol. rity and computing power, perhaps a trust-
worthy corporation!, or even a trustworthy
government!), they would become a Certi-
Digital certificate fication authority (CA). Everybody, on first
receiving their own public-key--private-key
Foundation Concepts: Digital signature, Public key, pair, would register their public key with
Encryption. the CA. The CA would very thoroughly
Definition: A public key, digitally signed by a trusted investigate the individual to verify their
authority. identity, doing at least as much as modern
governments do before issuing passports.
Overview Then the CA would then issue the appli-
The only real problem with public-key cant with a digital certificate, a simple and
encryption, once the strength of the algo- short document saying essentially ‘‘This . . .
rithm has been established, is key manage- is the true public key for . . .” The certificate
ment. As well as the obvious need to keep is encrypted using the CA’s own private
private keys utterly private, there is a cor- key.
respondingly urgent need to make public There are two essential ingredients to
keys utterly public. this scheme. The first is that everybody
If a criminal can trick anyone into believ- in the world should know the CA’s public
ing that a public key generated by him is key; it should be stored indelibly and
in fact your public key, then, in the digital immutably on all systems, and never need
world, he can be you. Encrypted messages to be looked up from any directory service.
intended for you will be readable by him The second is that the CA must be perfectly
and only him, and, possibly worse, he will trustworthy and perfectly secure. A dishon-
be able to sign legal documents (contracts, est CA could get away with almost any-
bank drafts, confessions) as you. thing. If the CA’s private key is ever revealed
This is why wide publication of public then all identities are open to theft.
keys is essential. Everybody you do busi- If the CA is perfectly trustworthy and per-
ness with should ideally have a secure copy fectly secure, then digital certificates are
of your public key as provided directly by as reliable as the encryption system used.
you; everyone you might do business with Nobody can ever present a fraudulent pub-
in the future should have instant access to lic key to steal another’s identity, because
an incorruptible public-key directory ser- nobody would ever accept a public key that
vice. Of course, this ideal can not practi- does not come as part of a digital certifi-
cally be realized. If you had to deliver your cate. Nobody can make a false digital cer-
public key to everybody in person, elec- tificate without access to the CA’s private
tronic commerce would be pointless, and key.
how would they verify your identity even Secondary CAs may also be created, and
then? With a public directory, it would would most probably be necessary since the

105
Digital certificate

load on a single universal CA would be tion process also results in more efficient
overwhelming. Each secondary CA would certification management and a potential
have its own digital certificate of identity for reducing the total cost of ownership
signed by the one primary CA. Then every involved in the certification-related pro-
digital certificate issued by the secondary cesses.
authority would contain an extra clause Digital certificates and associated appli-
that is a copy of the secondary authority’s cations can be used for a variety of pur-
own digital certificate. That means that poses, including the certification of identi-
those receiving a certificate issued by a sec- ties, adding encryption, confirming sources
ondary authority may still validate it with of data/information, and restricting access
knowledge only of the primary CA’s public to information on the basis of certified
key. identify.
Currently, there are some organizations
offering competing services as trusted CAs,
Summary of positive issues
and their public keys are not securely built
Certificates are easy to create and use
into computer systems. Digital certificates
through CAs. They provide positive identi-
are offered for costs in the range of hun-
fication of data authors, individuals, com-
dreds of dollars. It is envisaged that each
panies, or other entities.
organization would obtain a single digi-
tal certificate from a central authority, and
then act as its own secondary certification Summary of potentially negative issues
authority for internal purposes. Great care must be taken in selecting a CA:
anybody can set up a web site, call them-
Business value proposition selves a CA, advertise on the internet, and
Digital certificates are a means that allows sell digital certificates. A CA that is not com-
companies or individuals to authenti- pletely trustworthy is much worse than no
cate the entities (e.g., companies, people, CA at all. A prospective customer must ask
and governments) that they interact with themselves what makes this private com-
online. There are two primary options pany so thoroughly worthy of trust, and
for businesses wishing to deploy digital how can they guarantee complete security
certificates; they can administer their own of their encryption keys?
certificates, or they can outsource the man- Digital certificates are currently a very
agement of the certificates. expensive proposition, especially for indi-
It would be difficult for most organiza- viduals and small organizations. It may be
tions to administer and issue their own hoped that the advent of secondary CAs
digital certificate. Also they would have could ease the situation, as could govern-
the problem that their certificate would ment regulation in this commercially vul-
be doubted by many, if not all, of those nerable area which impinges directly upon
who received it, hence the advent of CAs. national-security concerns.
The existence of a CA allows organiza-
References
tions and individuals to make use of an r B. Schneier (1996). Applied Cryptography
established certification process. Further,
(New York, John Wiley and Sons).
these authorities usually provide a suite r J. Feghhi and P. Williams (1998). Digital
of applications that work in conjunction
Certificates: Applied Internet Security (New
with software from other vendors (e.g., web-
York, Addison-Wesley).
server vendors) to provide a secure cost-
effective mechanism for certificate man- Associated terminology: Phishing,
agement. The outsourcing of the certifica- Cracking, Hacker.

106
Digital signature

Title IV covers special issues such as


Digital Millennium Copyright
‘‘Ephemeral Recordings for Broadcasters,”
Act (DMCA) ‘‘Distance Education Study,” ‘‘Exemption
Definition: The Digital Millennium Copyright Act of for Nonprofit Libraries and Archives,” ‘‘Web-
1998 is divided into five titles that implement two casting Amendments to the Digital Perfor-
1996WorldIntellectualPropertyOrganization(WIPO) mance Right in Sound Recordings,” and
treaties, covering the copyright of performances and the ‘‘Assumption of Contractual Obliga-
phonograms, liabilities for online copyright infringe- tions upon Transfers of Rights in Motion
ment, computer maintenance competition assurance, Pictures.”
and issues such as web-casts and the design of ships’ Title V pertains to the protection of hull
hulls. and deck designs of vessels no longer than
200 feet.
Overview
The five titles of the act cover a wide
Business value proposition
range of technology copyright areas. Title I
The act requires businesses to operate
is the ‘‘Copyright of Performances and
under a regulatory framework and within
Phonograms Treaties Implementation Act
the context of a state that also works to
of 1998.” Title II is the ‘‘Online Copyright
protect the interests of copyright owners.
Infringement Liability Limitation Act.” Title
III is the ‘‘Computer Maintenance Competi- Reference
tion Assurance Act. Title IV contains six r ‘‘The Digital Millennium Copyright Act
‘‘Miscellaneous provisions,” while Title V is of 1998,” The United States Copyright
the ‘‘Vessel Hull Design Protection Act.” Office, 1998.
Title I implements two WIPO treaties:
Associated terminology: Caching,
the WIPO Copyright Treaty (WCT) and
Network, Law cross-reference.
the WIPO Performances and Phonograms
Treaty (WPPT). The intent is to obligate
WIPO member states to protect copyright
works from being infringed upon by cit- Digital signature
izens of other member countries. Title I
also obligates member states ‘‘to prevent Foundation concepts: Security, Encryption, One-way
circumvention of technological measures hash, Public key.
used to protect copyrighted works, and to Definition: An annotation added to a digital document
prevent tampering with the integrity of by an individual, that can not be copied onto another
copyright management information.” document or created by any other entity, and does not
Title II acts to create limitations of permit the document to be altered afterwards. An ide-
liability for online service providers per- alized version of the traditional pen-and-ink personal
taining to copyright infringement; these signature.
include provision of data and materials in
terms of ‘‘Transitory Digital Network Com- Overview
munications,” ‘‘System Caching,” ‘‘Informa- The idea of a signature is well under-
tion Residing on Systems or Networks at stood. It is a personal mark that attests to
the Direction of Users,” and ‘‘Information the authenticity of a physical document,
Location Tools.” and which often constitutes some legally
Title III covers exemptions for the autho- enforceable guarantee. The problems are
rized copying or adaptation of software manifold: documents may be altered after
during the maintenance or repair of a com- they have been signed; a signature may be
puter by software owners or lessees. copied onto another document; the signer

107
Digital signature

may falsely claim that a true signature was lished. If the two versions match, the doc-
forged. ument is genuine.
The notion of a digital signature as a dig- The sender can not later deny having
itized version of a person’s normal signa- signed the document. The fact that the fin-
ture, which may be included as an image gerprint was successfully decrypted using
in a document, is patently absurd. Digital their public key is proof that it was origi-
documents are even easier to modify than nally encrypted using their private key, and
paper ones, and a signature graphic can nobody else has access to that.
be cut out of one document and added to The document can not be modified after
another seamlessly in seconds. ‘‘Digital sig- it has been signed, because any change
nature” should not be confused with ‘‘elec- to the document will result in a change
tronic signature,” which is simply the sys- to the fingerprint (one-way hash), and the
tem to capture digitally the pen strokes fraudster can not substitute a new finger-
made when signing for a paperless credit- print, because it must be encrypted with
card purchase with a stylus and electronic the sender’s private key.
pad. Digital signatures are useful for resolving
The true concept of a digital signa- intellectual property and non-disclosure
ture relies upon two essential technologies: disputes, since they enable individuals to
Public-key encryption (q.v.) and One-way-hash prove that they know something without
operations (q.v.). To summarize, in a public- revealing what it is until later. Essential
key encryption system, every individual information may be recorded in a doc-
has two encryption keys, one of which ument and digitally signed in the nor-
is completely private and secret, whereas mal way. The digital signature alone, with-
the other is totally open and public. Data out the document that it applies to, is
encrypted with one of the keys can only be sent to the other party. They can not
decrypted with the other. A one-way hash reconstruct the document from the signa-
is an operation applied to any data that ture alone, but the sender is also inca-
reduces it to a moderately sized number (a pable of constructing another document
few dozen digits); this number acts as a fin- later that would match that signature.
gerprint for the data, so that the slightest This mechanism also allows a third party
change to the data would produce a totally to witness a document with their own
different fingerprint number, but the fin- digital signature without being able to
gerprint does not contain enough informa- read any sensitive information it may
tion to reconstruct any part of the original contain.
data. The whole system of digital signatures
A digital signature for a document is cre- relies upon complete openness. All of
ated by applying a one-way hash to that the procedures must be public, otherwise
document, encrypting the resultant finger- nobody will be able to verify signatures;
print using your private key, and appen- every participant’s public key must be pub-
ding the result to the document. It is a lished as widely as possible, otherwise it
simple, fast, and secure procedure. becomes possible for fraudsters to extend
Anyone receiving this signed document fake public keys for their victims. The only
can easily calculate what the fingerprint thing that must be kept secret is the private
should be: they have the document, and key. There must be no possibility of a pri-
one-way hash procedures are always openly vate key ever being revealed under any cir-
published. They can also easily decrypt the cumstances. The advance of electronic com-
copy of the fingerprint appended by the merce almost equates private key with per-
sender: public keys are always openly pub- sonal identity.

108
Digital signature

The DSA (Digital Signature Algorithm) sole control; and (d) it is linked to the
is an embodiment of the DSS (Digital Sig- data to which it relates in such a manner
nature Standard) produced by the NIST that any subsequent change of the data is
(National Institute for Standards and Tech- detectable.” The Basic Electronic Signature
nology) under authority of the Computer is a general definition that covers digi-
Security Act of 1987. It is based on a vari- tally recorded hand-written signatures; it is
able key length of between 512 and 1024 the Advanced Electronic Signatures that
bits, which would be very large for normal correspond to the true digital signatures
encryption, but has attracted criticism for discussed above. Many EU countries have
not being long enough for such a critical subsequently created their own legal frame-
purpose. Some cryptanalysts are suspicious works in association with the EU directive,
of the DSA, because it was designed by the including the UK implementation of the
NSA (National Security Agency), and not directive termed the ‘‘Electronic Signatures
everyone is totally willing to trust them. Regulations 2002.”
The US and EU laws principally aim to
Business value proposition ensure two things: that the signatory is
The United States enacted a variety of laws who they say they are, and that the digi-
pertaining to digital signatures, including tally signed document is authentic. These
the Electronic Signatures in Global and two aspects of signed digital documents
National Commerce Act of 2000 which aims form the basis of a ‘‘non-repudiation” ser-
to ‘‘facilitate the use of electronic records and vice, a service that can provide proof of
signatures in interstate and foreign commerce data origin, transmission, and delivery, pro-
by ensuring the validity and legal effect of con- tecting the sender and receiver from false
tracts entered into electronically” (www.ftc.gov). claims.
At the state level, the Uniform Electronic
Transactions Act of 1999 is intended to Summary of positive issues
‘‘remove barriers to electronic commerce by Digital signature technology is well estab-
validating and effectuating electronic records lished and backed by legal frameworks in
and signatures,” and the Digital Signature many countries. Non-repudiation services
and Electronic Authentication Act of 1998 exist to validate data transfer.
(SEAL), which is an amendment to the Bank
Protection Act of 1968, is intended primar- Summary of potentially negative issues
ily to enable the use of electronic authenti- The terms ‘‘digital signature” and ‘‘elec-
cation techniques by financial institutions. tronic signature” are sometimes used inter-
In 1998 the European Union (EU) changeably when they may in reality rep-
issued an ‘‘Electronic Signature Directive” resent different levels of technology and
(1999/93/EC), which defines two different security. Not all countries have legislation
electronic signatures, namely the Basic to cover the use of digital signatures and
Electronic Signature which they define as country-specific contract law needs to be
‘‘Data in electronic form which are attached examined prior to commencing electronic
to or logically associated with other electronic transactions.
data and which serve as a method of authen-
tication” and an Advanced Electronic Sig- References
nature, which must meet the following r B. Schneier (1996). Applied Cryptography
requirements: ‘‘(a) it is uniquely linked to (New York, John Wiley and Sons).
the signatory; (b) it is capable of identifying r Directive 1999/93/EC of the European
the signatory; (c) it is created using means Parliament and of the Council of 13th
that the signatory can maintain under his December 1999 on a Community

109
Digital wallet

framework for electronic signatures, about having their personal financial data
Official Journal of the European held by a third party.
Communities, L13/12, January 2000.
Associated terminology: Digital certificate,
Associated terminology: Digital cash, Digital cash.
Digital certificate.

Disclosure and interception of


Digital wallet communications laws
Foundation concepts: Encryption, Security. Definition: The Unlawful Access to Stored Commu-
Definition: A secure software mechanism for storing nications Act (18 USC 2701), Voluntary Disclosure of
credit-card and other payment information on a com- Customer Communications or Records Act (18 USC
puter to ease online purchases and transactions. 2702), and Wiretap Act (18 USC 2511) provide a
general prohibition against intercepting communi-
Overview cations or disclosing communications that one has
A Digital wallet is an application that resides valid access to.
on a computer and manages personal infor-
mation such as name, address, credit-card Unlawful Access to Stored
details, and bank-account information. This Communications Act, 18 USC 2701
application is used to populate automati- This act makes it an offense for an indi-
cally the fields of an electronic web-based vidual who ‘‘intentionally accesses without
form such as those used when an online authorization a facility through which an
purchase is made. This relieves the user electronic communication service is provi-
of continually entering the same data into ded; or intentionally exceeds an authoriza-
forms. The data is encrypted and requires tion to access that facility and thereby obta-
the vendor to send a digital certificate to ins, alters, or prevents authorized access to
verify their identity. Some digital wallet ser- a wire or electronic communication while
vices are provided by third-party vendors it is in electronic storage in such system.”
and the wallet resides on their server.
Voluntary Disclosure of Customer
Communications or Records Act,
Business value proposition 18 USC 2702
Digital wallets provide users with a secure
This act covers electronic communications
and convenient mechanism by which to
and the contents of those communications,
populate the fields of any web service they
e.g., voice and data. Its general purpose is
wish to interact with.
to forbid those who provide public commu-
nications services from releasing communi-
Summary of positive issues cations to any but the intended recipient,
Digital wallet software is available from thus providing some guarantee of security,
several vendors, some of which hold the privacy, and confidentiality.
data on their servers for added security and It has three major sections. The first for-
ease of access; in that case users don’t have bids providers of public electronic-commu-
to be at their own personal machine to use nications services to knowingly divulge to
the resource. any person or organization the contents of
a communication while it is in electronic
Summary of potentially negative issues storage.
The data in a wallet must be very care- The second section extends coverage to
fully secured. Users may feel uncomfortable those who provide any remote computing

110
Disk

service to the public (such as web-hosting ted, such as the activities of an officer or
services and data processing bureaus), agent of the US government in the nor-
similarly forbidding the release of com- mal course of his official duty to conduct
munications either carried or stored by electronic intelligence, e.g., the FBI’s Car-
that service. The third adds a prohibition nivore system used from 1998 to January
against the providers of communications 2005, which sifted through email at ISPs,
and remote computing services releasing capturing emails for only those individuals
any information about subscribers or cus- named in a court order.
tomers to any government entity.
While these aspects of the act cover the The Cyber Security Enhancement Act,
privacy framework for electronic communi- Section 225 of the Homeland Security
cations and the contents of those com- Act of 2002, which amends various
munications, there is an extensive set of sections of 18 USC
exemptions built into the act, some of This act takes the form of a set of amend-
which are necessary for business to occur ments to previous legislation related to
(for instance, an email provider must be security and privacy on electronic (com-
able to send email out of a mail box to its puter) systems. The act strengthens exist-
intended recipient). Other exceptions allow ing laws in various ways, such as explicitly
federal, state or local-government access in making it an offense to advertise online
certain circumstances, such as the need for illegal devices, defined as ‘‘any electronic,
information in the case of life-threatening mechanical, or other device knowing or
emergencies. having reason to know that the design of
such device renders it primarily useful for
Wiretap Act, 18 USC 2511 the purpose of the surreptitious intercep-
The act is primarily intended to cover the tion of wire, oral, or electronic communica-
intentional interception, use, or disclosure tions” (18 USC 2512). Previously only ‘‘news-
of any communication transmitted over a paper, magazine, handbill, or other publi-
wire, orally, or electronically. The law also cation” were specified.
makes it an offense to use any device to Laws pertaining to ISPs and their ability
intercept communications. This aspect of to report issues concerning life-threatening
the act can be used to cover traditional behavior were amended to allow greater
‘‘bugs,” ‘‘wire taps,” and eavesdropping, flexibility in to whom they can disclose the
and to prosecute the malicious use of unau- information (18 USC 2702). The laws per-
thorized ‘‘packet sniffers” (programs that taining to hacking were strengthened and
can examine network traffic, looking for now cover fatal injury as a result of hack-
passwords or other information). The act ing (18 USC 1030). Sentencing guidelines
also makes it illegal to disclose any infor- were substantially strengthened for some
mation so found, for example placing a cor- offences.
porate password obtained through a packet
sniffer on a ‘‘cracker” website. Subsequent Associated terminology: Law cross-
to that it is also illegal for information to reference.
be taken from that cracker web site and
used, knowing that the information itself
was obtained illegally. Disk
The act does describe a large number
of situations in which the interception, Foundation concept: Storage.
use, and disclosure of information from Definition: The primary device for long-term data
wire, oral, or electronic sources are permit- storage.

111
Disk

Overview Access speeds are usually quoted as the


A disk drive consists of some number, usu- average random access time, which is very
ally one to five, of exceptionally smooth easy to compute: for the average access,
metal disks with magnetizable surfaces, the time must be the seek time plus (on
locked together and spinning at a high rate average) half the time for a single rotation.
(often over 10 000 revolutions per minute) For a disk that rotates at 7200 rpm (a very
on a single central axle. Very fine Read--write common speed, and the speed is usually
heads, one per surface, move rapidly in and prominently advertised), the rotation time
out and are able to reach any point on any must be 8.33 milliseconds; so any claimed
surface very quickly. The heads are able to average access time that is not notably
detect magnetic patterns already present greater than half of that figure must be
on the disk surface, and to write new mag- viewed with grave suspicion.
netic patterns on any region they pass over. When a disk is ‘‘powered down,” and not
In this way binary data is stored as micro- spinning, the read--write heads actually
scopic magnetic domains all over the disk rest on the surface of the disk. As the disk
surfaces. speeds up, friction causes air molecules to
The rate of rotation of the disks is closely be dragged around with the disk surface,
controlled, and never varies (changing the and at full speed (usually around 65 mph) a
spin speed would be a very slow opera- few molecules actually get dragged under
tion because of the rotational inertia to the read--write heads, and they ride on a
be overcome), but the read--write heads are cushion of air just a few molecules thick.
mounted on delicate arms (looking some- That is how they maintain the right dis-
thing like tweezers) that can move to any tance above the surface and move without
radial position over the disk surface on causing wear. It is also the main weakness
demand. Data stored on a disk is organized of disks. If a disk drive receives some physi-
into concentric circular tracks: to access cal shock while it is in operation, the heads
data, first the heads are moved to the right can crash into the surface and cause signif-
radial position for the desired track, then it icant damage. This is the condition known
is necessary to wait until the right portion as a Head crash. Usually the disk drive as a
of that track rotates itself under the heads. whole will survive, but the data stored in
A typical disk in a PC has a usable the damaged region can be irretrievably
radius of around one inch (or 3 cm), into lost. Modern disk drives contain an auto-
which hundreds of thousands of tracks are matic mechanism that ensures that the
packed. That requires exceptionally fine heads are moved to a designated Landing
control in the positioning of the heads, zone, a region of the disk not used for data
and that is a significant factor both in the storage, whenever power is lost, so they will
cost of disk drives and in the data access naturally and gently land in an unimpor-
time. Incredibly, the heads can move from tant region. This means that disks can with-
one track to another, and be ready to read stand much more violent shocks when they
or write, in just a few thousandths of a are powered down, but they are still quite
second. The speed of a disk is controlled delicate devices and must be treated gently.
by two components, the track-to-track Seek Because disk drives rely on moving parts,
time, plus the wait for the desired data to they do not have the almost infallible
rotate into position, known as the Rota- nature and unlimited lifetimes of semicon-
tional latency. This means that the access ductor circuitry. The minimal amounts of
time for any given piece of data is very damage caused by vibration, poorly con-
variable, depending on the disk’s current trolled power supplies, and just plain use
position at the time access is needed. (no moving part is completely frictionless,

112
Disk

so gradual erosion must occur) accumu- floppy disks (in the 8 and 51/4 inch for-
late, until inevitably the whole drive fails. mats) had less than rigid cases and were
It must be understood that any data stored indeed slightly floppy. The read--write heads
on disks is subject to unpredictable loss; do not hover microscopically above the
the only solution is to keep backup copies surface, but actually make close contact,
on different media of anything that is pushing into and bending the surface. For
essential. this reason, floppy disks must rotate much
The parameters of a disk that determine more slowly than hard disks, have a much
its capacity are the number of surfaces, the smaller capacity (generally only just over
number of tracks, and the amount of data 1 MB), and are not kept spinning full-time.
that can be packed into one track. Stating They also have a much more finite life-
the number of tracks can be ambiguous time, but are correspondingly more robust,
(does 100 000 tracks mean 100 000 tracks and can be dropped from a great height
per surface, or a total of 100 000 overall?). without harm. Floppy disks are encased in
Rather than simply resolving the ambigu- a square protective sleeve, which is also
ity with a decisive definition, a different used to hold them in place when they are
term is usually used. A Cylinder is the stack inserted into the drive. Floppy disks are
of tracks at the same radial position on all always exchangeable: any number of dif-
surfaces, so a disk with five surfaces and ferent disks may be inserted into a sin-
100 000 cylinders has an unambiguous total gle drive, used, then removed and replaced
of 500 000 tracks overall. by another. Hard disks on small comput-
On any single track, the data is not stored ers are never exchangeable, but the high-
as one long stream of bits, but is divided up performance disk drives used on main-
into a number of Blocks. A block is almost frames do still sometimes have an exchan-
universally 512 bytes or 4096 bits. Until geable disk pack as an option.
recently it was almost universal for there The first working disk drive in commer-
to be 63 blocks on each track, but now the cial production was the IBM 350, which was
number varies considerably, with the outer introduced in 1956 as a component of the
(longer) tracks holding hundreds of blocks. IBM 305 computer. It was the size of a large
For efficiency of organization and access, wardrobe, had a capacity of 4.5 MB, and was
and to ensure sufficiently accurate position- available for lease at $35 000 per year. The
ing, a block is the smallest amount of data first hard-disk drive available for an IBM PC,
that can be written onto a disk at one time. 26 years later, was the Seagate ST412, with
If a single byte is to be changed, it is neces- a capacity of only 10 MB. IBM also intro-
sary to read the whole 512 bytes that sur- duced the first floppy disk, in 1971. In the
round it, modify that one byte, and then 8 inch format, it had a capacity of approxi-
write the entire 512 back again. In modern mately 0.1 MB, and was originally a system
times, blocks are often erroneously called engineer’s tool, not a general data-storage
Sectors, but the term sector properly refers medium.
to the entire wedge-shaped collection of The ZipDrive disk, Jaz disk, and simi-
blocks at the same position on every track. lar devices are two kinds of compromise
Disk drives are often referred to as Hard disk technology. A lower-capacity version
disks to distinguish them from Floppy disks, (around 100 MB) uses the same technology
an alternative technology that through as floppy disks, but is much more finely
cheapness became very popular in the and expensively constructed, so it gives
1970s, but is now disappearing. Floppy higher capacity, access speeds, and relia-
disks are made of a flexible material and bility. The larger-capacity version uses the
are much less finely constructed; early essential technology of hard disks, but with

113
Disk

exchangeable media in small cartridges nificantly to price and power consumption;


that look somewhat like thick floppy disks. having one disk with three partitions adds
Before a disk can be used in a computer nothing but convenience.
system, it must be Formatted. This is the pro-
cess of creating all the logical structures Business value proposition
(such as empty directories or folders, lists Disks form the basis of secondary storage in
of available blocks, and lists of file names) all computers and the amount and type of
that the operating system expects to find. storage that a corporate system requires are
Formatting is an operating system depen- determined through the information life-
dent task: a disk formatted for one system cycle (see Information lifecycle management).
may be completely unusable by another. It Some small modern computing devices
is also a repeatable task: a used disk may such as PDAs and true Thin clients do not
be reformatted to make it appear to be as have disk storage, and rely on non-volatile
new, so one can start again with a blank solid-state (Flash) memory or network access
slate. Many operating systems provide two to disks attached to another system.
options for formatting: Quick or Full. A full The floppy disk as a storage medium has
format is usually a very slow procedure that shrunk in size since the 1970s, evolving
often truly erases all old data on the disk. from 8 to 51/4 to 31/2 inch, while the storage
A quick format saves time by overwriting capacity has grown; however, their evolu-
data only when necessary; old confidential tion seems to be coming to the end of the
data may be recoverable from a disk after line as they are replaced by memory devices
a quick format. with no moving parts that can be con-
When formatting a disk, most operating nected to USB ports (flash memory). These
systems also provide the option of Partition- devices have a significantly larger mem-
ing. A Disk partition is simply the division ory, and are smaller, more portable, and
of a large disk into a number of smaller far more robust and reliable than floppy
regions, each of which is treated as a com- disks. As this evolution continues, organi-
plete individual disk in its own right. This zations would be well advised to migrate to
can have three advantages. Some operat- the newer alternatives. Flash memory does
ing systems have an in-built maximum disk not yet approach the capacity of traditional
size; partitioning an oversized disk can hard disks, and continues to be many times
often reduce it to a usable size. Also, catas- more expensive per megabyte. The ability to
trophic disk failures, such as a head crash add external and even removable hard-disk
in an essential system directory, can easily drives remains a useful option for network
affect a whole partition, but other parti- and system designers.
tions on the same disk have a greater The price--performance ratio of disk-
chance of surviving unscathed. Finally, it based secondary storage devices continues
can greatly simplify system maintenance to improve for consumers; however, they
to have multiple disk drives. If there is should not be considered infallible and pro-
one disk for ‘‘system files” (the operating vision must be made by the network man-
system and its data), another for applica- ager (and the individual) for the inevitable
tions, and yet another for user data, sys- disk crash. While modern disk-based sys-
tem upgrades can be much easier, since the tems are more reliable than previous gen-
whole operating system can be erased and erations, they contain moving parts and
re-installed without interfering with appli- hence will fail in due course. Systems man-
cations or data. Similarly, whole applica- agers need to have a device-replacement
tions may be replaced without any risk to strategy built into their maintenance and
essential data. Having three disks adds sig- capacity planning strategies.

114
Distributed database

Summary of positive issues multiple database servers must be used


The cost of secondary storage in the form instead.
of disks continues to decrease while per- Designing a distributed database is a
formance parameters increase. Each gener- complex task, and minutely tests the
ation of disk-based memory is more robust database administrator’s and CIO’s skills.
and reliable than previous ones. Flash- Apart from the obvious problems of select-
memory devices are replacing removable ing suitable hardware, ensuring that it is
floppy-disk drives and enabling users to up and running close to 100% of the time,
have access to larger, more convenient, and providing secure and regular backups of
reliable portable data storage. External disk data, and ensuring constant network acces-
drives are available to allow users of desk- sibility, there are some fundamental tech-
top and laptop computer to extend their nical questions of distributed design. The
storage capabilities. most fundamental of these is exactly how
the database should be distributed; this is
Summary of potentially negative issues known as the problem of partitioning the
The hard drives in computers do crash and database.
it is imperative that a contingency plan Should the data be divided up so that dif-
should be put in place to backup and secure ferent tables (complete collections of data
data. Hard drives in laptop computers are items of the same kind) are stored on
by no means insensitive to movement and different computers, which does not help
can crash should sudden movements of the when there is one major table that is both
machine occur, especially if it is carried large and frequently accessed? Should the
around while active, or if a laptop com- tables be partitioned vertically, meaning that
puter is knocked off a desk; contingency each server holds a part of every record,
planning is thus also vital in these situa- or should they be partitioned horizontally,
tions. meaning that each server contains just
some of the records, but each record on
Associated terminology: Information each server is complete, or should a combi-
lifecycle management, Backup, Optical nation of vertical and horizontal partition-
storage. ing be used? Or should entire copies of the
whole database be stored on each of a num-
ber of servers?
Distributed database Whether and how a database is parti-
tioned has a major effect on its opera-
Foundation concepts: Database, Parallel processing. tions. If multiple copies are kept on dif-
Definition: A database that is spread over a number of ferent servers, it is much easier to spread
computers, distributing the load of storage and query the load of query processing, since any
processing. query could equally well be answered by
any server. It also becomes critical to keep
Overview the different copies of the database prop-
One of the most important applications of erly synchronized, so that, when a change
Distributed or Parallel processing is to database is made to a record, there is no period of
systems. Corporate databases can quickly inconsistency, during which a query will
become very large, and may need to sup- produce different answers depending upon
port multiple concurrent accesses from any which server answers it. An inconsistent
number of branch locations. In many cases database can have disastrous results. Copy-
there is no super-computer that is super ing the database onto multiple servers also
enough to handle the required load, so does nothing to reduce the problem of a

115
Distributed database

database being too large for one server to base system can be designed, its general
handle. modus operandi must be decided: is it
Horizontal partitioning does relieve the going to keep all of the data in a ‘‘live”
problem of too-large database tables. A transactional database that is immediately
database of 100 000 000 records could be accessible, or is the data going to be pre-
split amongst ten servers, each holding partitioned, with old data that is no longer
10 000 000 complete records. Any query will likely to be accessed by online systems
probably have to be sent to all ten servers, being placed in a Data warehouse, and only
and the individual partial results will then the live data remaining in the transactional
need to be combined before being returned database?
to the user, but that is one of the eas- When an organization’s overall require-
ier tasks. Horizontal partitioning usually ments have been defined, then the database
implies the three-tier model of application system can be designed. The design of a
servers, and often does little to relieve the database is a technically challenging task
problem of having too many queries to and for any non-trivial case it requires
answer quickly, since each query must be a highly skilled and trained database
processed by multiple servers. administrator (DBA). The DBA needs to be
Vertical partitioning may help to reli- skilled not only in database technology
eve both problems. With a database of 100 but also in the specific environment in
000 000 records, each of ten servers would which they are to operate. Databases differ
hold the more closely related parts of every from vendor to vendor, and from version
one of the 100 000 000 records. With care- to version, so very specific knowledge is
ful planning, it can be arranged that the needed.
most common queries can be answered by Central to the DBA’s task is the deter-
a single server, but for those uncommon mination of how and whether to partition
queries based on fields that never appear the data across a set of computers. There
together on any one server, processing can are several methods of partitioning, and
become exceptionally complex and ineffi- the choice of solution is complex, involving
cient. With vertical partitioning, the prob- a careful examination of the options. This
lems of temporary inconsistencies immedi- requires knowledge of the interactions that
ately following insertions and deletions are occur between the hardware, the operating
greatly magnified. system, the network, and the application
The decisions on how the data should be software, as well as understanding the prac-
distributed across multiple servers and the tical constraints surrounding the database
design of the algorithms for answering the system itself.
various possible queries go hand-in-hand. The overall design of the database needs
One can not effectively be done before the to be carefully considered, and the trade-
other. The design of a distributed database off amongst performance, ability to scale,
is one of the most complex challenges fac- and cost must be built into the technical
ing an IT professional. business case developed to help identify a
solution.
Business value proposition
As corporations grow, their databases grow Summary of positive issues
with them. It is not unknown for some Distributed databases and partitioning
organizations to have databases of multiple techniques allow for a variety of solu-
terabytes, and sources have reported that tions to managing very large databases. Dis-
some government agencies have databases tributed databases allow DBAs to attempt to
of over 100 terabytes. Before a large data- optimize their database systems.

116
Domain name

Summary of potentially negative issues names rests with ICANN (the Internet
Creating distributed databases is extremely Corporation for Assigned Names and
challenging technically. A poorly designed Numbers).
partition of the database can cause Once a domain name has been assigned to
database failures, instability, and unaccept- an organization, that organization is free
ably poor performance. to create any number of sub-names to refer
to individual servers or internal networks,
Reference by prefixing extra components to the begin-
r M. Ozsu and P. Valduriez (1999).
ning of the domain name. For example, the
Principles of Distributed Database Systems owner of ‘‘company.com” has complete
(Englewood Cliffs, NJ, Prentice-Hall). control over all names that end with
‘‘.company.com,” and can freely define
Associated terminology: Database
‘‘www.company.com,” ‘‘sales.texas.company.
administrator, CIO.
com,” and anything else of a similar nature.
When a web browser, or any other inter-
net client, attempts to access a server using
Domain name a name like ‘‘sales.texas.company.com,” a
system known as DNS (Domain Name Ser-
Foundation concept: Internet protocol, Network. vice) is invoked. This is a network of servers,
Definition: A readable and memorable name asso- embedded in the internet, which provides
ciated with an IP address, providing a convenient the necessary information for converting
reference to internet-connected computers. names back to the numeric IP addresses
that are required for internet communica-
Overview tions.
All computers connected to the internet There is a ‘‘Top Level DNS Server” for
must have a numeric IP address associated all names ending in ‘‘.com”; this server
with them. This is the only way that com- is the ultimate authority for all ‘‘.com”
munications are routed across the inter- addresses, which is why proper domain
net, so it is an absolute requirement. IP name registration is essential. The top-
addresses currently consist of four smallish level server is asked what it knows about
numbers (range 0-255) separated by dots, ‘‘sales.texas.company.com,” and responds
for example ‘‘192.168.45.234.” These num- with a partial answer: the address of
bers are not at all memorable, and are an another DNS server that is responsible for
extremely unsatisfactory way to publicize all addresses ending in ‘‘.company.com.”
the address of a web site. With the advent The question is then asked again of this
of IPv6, these addresses will become even server, and it may reply with the com-
longer sequences of numbers (see Internet plete answer, or it may instead reply
protocol for details). with the address of another DNS server
A Domain name is a more user-friendly, that is responsible for all names in the
human-oriented name associated with an ‘‘.texas.company.com” subdomain. Eventu-
IP address, having a more familiar and ally, if everything has been set up correctly,
memorable form such as ‘‘MyCompany- the question will reach a DNS server that
Name.com” or ‘‘AnotherCompany.co.jp.” knows the complete address, and commu-
Before a domain name can be used, it must nications may begin.
be registered with the internet authorities; When a domain name is registered, it is
there are numerous organizations provid- also necessary to make sure that there is at
ing domain name registration services for least one DNS server that can answer DNS
a fee; ultimate responsibility for domain queries about hosts within the domain.

117
Domain name

For an organization that does not have the United States passed the Federal Anti-
its own 24-hour computing service, this Cybersquatting Consumer Protection act of
will probably involve additional fees. It is 1999. However, for smaller organizations
normally considered essential to have two or individuals, obtaining a ‘‘meaningful”
DNS servers prepared to handle queries domain name can be problematic, since
about any domain, since some down-time many popular names and terms have been
is inevitable, and without a working DNS, used already, e.g., www.robertplant.com.
internet sites can not be found. This can cause the need for a re-branding
The last component of a domain name effort.
does have some significance, and can not The choice of domain name of course
be chosen freely. There are seven traditional does not have to be limited to local names,
top-level domains: ‘‘.com” was intended for e.g., a company in the UK can choose
multinational companies, but has come to to have a .com address, or even a co.us
be used by US companies; ‘‘.edu” is for US address. It is, of course, very difficult for
educational institutions; ‘‘.gov” is for US people to guess a domain name and even
government agencies; ‘‘.int” is for interna- powerful search engines might not easily
tional organizations; ‘‘.mil” is for the US help a potential customer locate the com-
military, ‘‘.net” is for networks, and ‘‘.org” pany they are looking for without specific
is for other US organizations. In order to information. There is no domain-name-
relieve demand on the exceptionally pop- index equivalent of the ‘‘Yellow Pages,” so
ular ‘‘.com” names, six new ones were for example, finding a hardware shop in
created: ‘‘.biz,” ‘‘.info,” ‘‘.name,” ‘‘.aero,” London that has not had much internet
‘‘.coop,” and ‘‘.museum.” Non-US individu- traffic or been in business long could be
als and organizations are expected to use difficult.
country-specific domain names; these end
in two dotted components, one indicating
Summary of positive issues
the kind of organization, and the other
The domain name-space is regulated by
(always two letters) indicating the coun-
ICANN. Domain names remove the need for
try it is based in; so, for example, ‘‘.co.uk”
people having to remember IP addresses.
is the UK-specific version of ‘‘.com,” and
Domain names are unique. The configu-
‘‘.co.jp” is the Japanese. There is a ‘‘.us” code
ration of a DNS server is relatively easy
for the United States, but it is rare for a
and connections through ISPs are available
US corporation to use ‘‘.co.us” instead of
almost universally.
‘‘.com.”

Business value proposition Summary of potentially negative issues


The choice of a domain name is a vital Domain names for popular words, phrases,
aspect of corporate branding, for large and names are frequently already regis-
established corporations their first choice tered. Top-level domains such as .biz and
of domain name was most likely secured .name are not popular and not widely rec-
within the first few years of the inter- ognized by the public.
net being deregulated to allow commer-
References
cial exploitation. There have been many r http:/www.icann.org.
instances of individuals Cybersquatting: buy- r ICANN, 4676 Admiralty Way, Suite 330,
ing up the domain names associated with
Marina del Rey, CA 90292--6601, USA.
famous brands and then attempting to
extort the legal holders of the brand Associated terminology: ISP, DHCP,
name. To prevent this from happening, Client--server.

118
Dvorak/QWERTY keyboards

used sequences involve alternating hands


Dvorak/QWERTY keyboards
and progress from little fingers to larger
Definition: QWERTY is the standard layout of letter fingers (a natural motion for humans), com-
keys on keyboards; Dvorak is an alternative layout. monly typed sequences involve minimal
movement of the hands and reaching with
Overview the fingers, and the load of keypresses is
The strange rendition of the alphabet shared more equitably between the fingers,
found on typewriters and other keyboards, slightly favoring the stronger ones.
Q W E R T Y U I O P A S D F G H J K
L Z X C V B N M, is neither random nor Business value proposition
arbitrary. It was deliberately designed (by The selection of a keyboard is usually made
Christopher Sholes, around 1870) with two by default, without any thought, by the
ends in mind. One was to reduce key jams vast majority of businesses or individuals;
by slowing down the rate at which common they simply take what comes with the com-
letter sequences could be typed by placing puter from the manufacturer. However, the
the constituent letters far apart. The other long-term productivity perspective could
was to ensure that the letters of the word suggest considering a change to a Dvo-
typewriter were all on the top row, as a sales rak or an even more ergonomic keyboard.
gimmick. It is clear that there must have Several manufacturers produce ergonomic
been other design criteria that were not keyboards that are more comfortable, with
recorded in the annals of history. lower key-impact requirements, separate
Now that mechanical typewriters are keyboards for each hand, and adjustable
things of the past, key jams are no longer key spacing. While rapid change is unlikely,
a concern, and the limited benefits of the the advent of cellular telephone text mes-
QWERTY layout are outweighed by its dis- saging, which is based on alphabetical key
advantages. The QWERTY layout is far from ordering, may result in future consumers
optimal: it requires frequent uncomfort- being more open to changing keyboard for-
able movements that do slow down typing, mats.
and are believed to contribute to repetitive-
stress injuries. Summary of positive issues
The Dvorak (often spelled Dvorjak) key- Tests do reveal that learning to type is eas-
board was designed by August Dvořák to ier and faster with the Dvorak keyboard,
optimize the key layout for fast and com- and that, once the keyboard has been mas-
fortable typing. He called his layout ASK tered, typing is faster and less stressful.
(the American Simplified Keyboard), but his
own more interesting name is always used Summary of potentially negative issues
instead. The main design features are that Dvorak keyboards are not very commonly
the most commonly used letters all appear used, and so are not produced in very
on the middle row, the most commonly large quantities, and as a result are more

119
Dynamic web pages

expensive than the QWERTY standard key- basic nature of web interactions. When a
boards. Typists need to be retrained (or web browser wishes to access a particular
at least to retrain themselves), and often page, it composes a simple command in
resist the change. Switching to Dvorak key- the HTTP language. This command usually
boards incurs some expense, and a period begins with the word ‘‘GET,” and contains
of much reduced productivity while typists an identifier for the desired page (its URL),
gain familiarity with them. together with any user inputs that may
have been provided in a form. This request
Reference is transmitted to the relevant server. When
r R. Soukoreff and I. MacKenzie (1995).
a web server receives a GET request, it
‘‘Theoretical upper and lower bounds translates the URL into a simple file name,
on typing speed using a stylus and soft and looks for that file. If the file con-
keyboard,” Behaviour & Information tains simple HTML or plain text, the entire
Technology, 14, 370--379. file is sent, unprocessed, to the requesting
browser, which then displays the content
on the screen.
Dynamic web pages If the server finds that the file actually
contains Server-parsed HTML (SHTML), then
Foundation concepts: Web page, Internet, HTML. the processing is slightly more complex.
Definition: Web-page content that is not static, but Instead of just transmitting the entire file
can vary according to circumstances, and react to user as-is, the server reads the file looking for
inputs and requirements. special tags. Most of the file is simply sent
to the browser as-is, but Server-side includes,
Overview surrounded by ‘‘<!--#” and ‘‘-->” are
The most basic web pages are simple text replaced by the appropriate text. For exam-
files with a fixed HTML content. Every time ple, if the text ‘‘The file xxx.html is
a particular web page is accessed by a <!-- fsize file = “xxx.html”-- >
browser, the content shown will be exactly bytes long” were to appear in an SHTML
the same. This is, of course, perfect for pro- document, and the server had a file
viding fixed information that does not vary, called xxx.html 7055 bytes in length,
but provides only limited possibilities for then the text ‘‘The file xxx.html is
user control and interaction (no more than 7055 bytes long” would be sent to the
the ability to click on a link to another browser. SHTML is of only limited power,
page), and makes truly interactive opera- but does increase the scope of simple web
tions such as online registration, form sub- pages.
mission, hit counters, shopping carts, and If the requested file is of the CGI (Com-
database queries impossible. mon Gateway Interface) type, then the server
Dynamic web pages provide content that assumes that it is an application pro-
is not created until it is needed; with each gram, and executes it. The running pro-
browser access, the page is created on-the- gram receives any inputs that were sent as
fly, and thus may be different each time part of the request, and all output created
and can even react to browser input. There by the running program is sent directly
are five very commonly used technologies to the requesting browser. Anything that
for dynamic web page creation: SHTML, the server is capable of doing can be pro-
CGI, JSP and ASP, and Scripting, and many grammed into a CGI application; using CGI
others that are just variations on a theme. is just like running an application over
To understand dynamic web page gen- the web. Programmers can use any lan-
eration, it is essential to understand the guage that the server supports, and have

120
Dynamic web pages

access to all operating-system services. CGI monly used to make the cursor or hyper-
is by far the most powerful and flexible of text links change their form depending on
the dynamic generation methods, and that the position of the mouse. Scripts are most
does have its drawbacks. Writing CGI code commonly written in JavaScript and Visual
requires real programming knowledge, and Basic.
is not a task for unskilled workers. There is
also a security risk: CGI can do anything,
Business value proposition
so a careless programmer could acciden-
Dynamic web pages created through the
tally open up the whole system to outside
SHTML, CGI, JSP, and ASP web technolo-
attack. In skilled hands, CGI is safe, simple,
gies allow businesses to build more sophis-
and powerful.
ticated web sites that facilitate a variety
JSP (Java Server Pages) and ASP (Active Server
of activities ranging from gaming to B2C
Pages) are very similar technologies. Both
e-commerce and B2B data exchanges. The
are based on the idea of SHTML: the server
web technologies allow adoption of a vari-
processes the file as it is being sent to the
ety of processing models; for example, in
browser, and substitutes information gen-
the scripting model the processing is at
erated on-the-fly for special tags in the text.
the client, whereas the JSP model per-
With both ASP and JSP, the special tags
forms the processing at the server, and
contain parts of (or even complete) pro-
this flexibility allows developers to create
grams that are executed to generate the
systems that are matched to their busi-
new content. These program segments can
ness need and their systems architecture
do anything from providing the current
and to take into account the technologies
date and time to substituting the results
deployed by their customers, partners, and
of complex cross-network database queries.
vendors.
In JSP the language of the program ele-
ments is based on Sun’s Java; in ASP it
is based on Microsoft’s Visual Basic. The Summary of positive issues
use of JSP requires Java support on the Dynamic web content is necessary if web
server; the use of ASP requires a Microsoft sites are to go beyond merely provid-
server. ing static information, and interact with
Scripting is a different kind of system. users. e-Commerce can not be conducted
When scripting is used, the file retrieved with merely static web pages; the closest
by the server contains program code, but is approach would be to provide customers
not processed by the server at all. The file, with forms that they must print, complete
together with any code it contains, is sent by hand, and fax in to make an order, and
directly back to the requesting browser, such a procedure would certainly discour-
and the program code is run locally by age many potential customers.
the browser. This has the positive effect CGI has unlimited power, allowing any-
of relieving the server of a lot of the pro- thing a programmer is capable of coding,
cessing load, but at a very high price. It but is lacking in convenience, and does
is the browser that has to execute the require some technical programming abil-
script, so it has no access to any informa- ity. Javascript provides much greater protec-
tion stored on the server. Of course, scripts tion against both accidents and malice, but
may open communications channels back has a much greater likelihood of arousing
to the server, but that adds a great deal of the suspicions of a user’s security system
complexity both to the script and to the and being blocked. Scripting allows simple
server, and somewhat defeats the purpose user interactions without any load on the
of scripting. Scripting elements are com- server.

121
Dynamic web pages

Summary of potentially negative issues grammers can do unlimited damage; CIOs


must be sure of the competence and loy-
Dynamically created pages put an extra
alty of any programmers selected to work
load on the server, the browser, or both.
on server and CGI implementations.
Careful capacity planning is required to
The more sophisticated dynamic web
ensure against systems being overwhelmed
pages become, the more likely they are to
at times of peak demand.
arouse the suspicions of client-side security
CGI and, to lesser extents, JSP and ASP
systems (particularly firewalls, web-browser
require a degree of technical program-
security settings, and anti-virus software).
ming ability from the originating organiza-
Content that gets blocked, even if it is com-
tion. CGI requires only very basic support
pletely innocent and harmless, can have a
and setup on a web server, but JSP and
negative impact on the corporate image in
ASP require more significant administra-
the customers’ minds.
tion, and generally put a heavier load on a
server. CGI’s power means that careless pro- Associated terminology: Javascript.

122
e-Commerce/e-business

net infrastructure as HTML and XML have


e-Commerce/
/e-business
become more established. Several models
Definition: e-Commerce/e-business are business have become well established, including
models that utilize the internet as a mechanism data-synchronization hubs, e-procurement
through which companies, vendors, and customers industry-specific systems, and specialist e-
interact. procurement systems. Data synchroniza-
tion hubs such as Transora.com provide
Overview the members of that hub with the ability
e-Commerce originated with the deregula- to synchronize data with other members
tion of the internet in 1995. It was orig- and notify specific members of events such
inally conceived as a business-to-customer as price changes or product-specification
(B2C) model of commerce, and corporations changes. Industry-specific e-procurement
were formed to provide services and prod- systems provide members with the ability
ucts over the internet without having any to perform a range of procurement activ-
physical retail presence (e.g., Amazon.com, ities through a central hub; a key exam-
the well-known online book seller). ple is COVISINT, which provides procure-
The term e-Business was coined by Lou ment services for the automobile indus-
Gerstner, then CEO of IBM, who looked try. Specialist e-procurement systems pro-
beyond the B2C model to use the internet vide specialized or niche services, such
for delivering services and supporting phys- as Chile Compra, which provides procure-
ical business channels. ment services for the Chilean government
The start-up companies that adopted (https://www.chilecompra.cl/).
pure B2C commerce, for which there is no
physical retail store, together with start- Business value proposition
up business-to-business (B2B) organizations, B2C models can provide businesses with an
which also owned no physical retailing or alternative retail channel if they are devel-
warehousing capacity, aimed to ‘‘disinter- oped in alignment with the overall cor-
mediate” the established companies by hav- porate business strategy. Four factors can
ing lower internal and distribution-channel be considered to help ‘‘position” a com-
costs. Unfortunately, this model proved to pany’s B2C online presence: brand, service,
be very expensive to support as the busi- market, and technology. The brand strat-
nesses attempted to scale up and reach pro- egy could be to reinforce an existing brand,
fitability, and the majority ran out of capi- create a new brand, or use the internet to
tal following the decline in the US stock reposition a brand. All of these strategies
market of 2000 and closed down. Those require that brand equity modeling be used
that survived the stock-market crash endu- in order to determine the most appropri-
red a market consolidation and continue to ate branding strategy to utilize. The online
attempt to build market share, drive down brand that an organization attempts to
costs, and become profitable or maintain project is driven by three other positional
profitability. factors: service levels must match the brand
Traditional companies with physical re- requirements, the market being addressed
tail and supply-chain facilities (also termed must offer products that are aligned with
‘‘clicks-and-bricks” companies) have contin- the company’s desired brand position, and
ued to develop their models and focus upon the technology must be adequate to support
those aspects of the business model that these conditions.
provide profitability (e.g., Walmart.com). B2B models provide businesses with the
The B2B business models have also potential both to lower transaction costs
evolved as essential elements of the inter- and to increase the speed of transaction

123
Efficiency

data exchange. Data-synchronization B2B to locate and use alternative sources of sup-
systems can be in the form of consortia, ply. Organizations need to evolve their sites
with the members benefiting from low- accordingly. Lack of alignment in corporate
ered development costs since they share the branding, service, or product offerings for
overhead associated with technical devel- bricks-and-clicks organizations can lead to
opment. e-Procurement portals also allow the loss of customers or confusion on the
members to reduce overall costs by avoid- part of customers as to their relationship
ing the search effort involved in access- with the company.
ing multiple corporate sites, hosting a Data-synchronization B2B systems requ-
site themselves, or competing against each ire a relatively high degree of technical,
other with proprietary systems. process, and organizational sophistication
of active participants. B2B procurement
Summary of positive issues hubs may evolve until a particular hub
Many successful B2C models have emerged, becomes dominant; selection of hub mem-
and traditional companies such as the UK bership is an important strategic sourcing
supermarket chain Tesco (www.tesco.com) issue.
understand the importance of building
upon established strengths and ensuring Reference
r R. Plant (2000). e-Commerce: Formulation
that the online business model is aligned to
the overall corporate strategy. Tesco’s busi- of Strategy (Englewood Cliffs, NJ,
ness model is such that physical orders Prentice-Hall).
taken over the web (e.g., groceries, clothes, Associated terminology: Internet,
DVDs) are filled and delivered from the Dynamic web pages, W3C.
store nearest to the customer, while service
products (e.g., car insurance, holidays, and
flights) are provided via the web.
The technical costs associated with Efficiency
B2C and B2B systems developments have
declined as tools, standards, support, and Foundation concept: Algorithm.
methodologies have been evolved both by Definition: A measure of how well resources are used
commercial vendors and by organizations by an application.
such as the World Wide Web Consortium
(W3C.org). Overview
B2B models such as data synchroniza- The general scientific definition of effi-
tion centers provide companies with the ciency is the output actually produced by a
ability to share and distribute data to a system expressed as a fraction of the maxi-
large number of businesses and entities at mum output any such system could theo-
a lower cost than would be possible with retically produce. An engine is 50% efficient
a proprietary system. Membership of a B2B if half of the energy put in (in the form
procurement exchange reduces the trans- of fuel) is converted into useful work (the
action costs associated with the procure- other half being lost through friction, pol-
ment process and reduces costs for mem- lution, etc.)
bers because they do not have to build the While this definition can be useful in
procurement system themselves. computing, it is not the definition nor-
mally used by computing personnel. In
Summary of potentially negative issues most cases there is no ‘‘theoretical best”
B2C models continue to evolve as emerging performance that an application may be
technologies change the customers’ ability measured against. In computing, the term

124
Efficiency

‘‘efficiency” is usually used as an umbrella gram run faster, it may need more mem-
term for absolute measures of perfor- ory, but to make a program run with less
mance. memory, it may have to be made slower.
For example, it may be necessary to pro- There is of course a fundamental differ-
cess a large amount of information to pro- ence between time efficiency and memory
duce a summarized report. Application A efficiency. Fast execution is usually a most
might load all of the information from disk desirable quality, but, if a program is slow,
into main memory before processing and all you have to do is wait a little longer and
thereby produce faster results than appli- the results will appear. If a program needs
cation B; which might perform exactly the too much memory, there is nothing to be
same task only taking the data in smaller done about it.
chunks, and therefore working much more The efficiency of human effort is another
slowly. A naive observer would conclude factor that must not be forgotten. If a pro-
that application A is better because it pro- grammer spends a week making one part
duces the same results but faster. A deeper of an application more efficient, so that
analysis is required. it now takes one second to run instead of
For example, tests may be performed, ten, is that an effective use of resources?
comparing the two applications on a num- The answer of course depends on circum-
ber of different data sets: stances. If this part of the application is run
many times by customers, then the total
improvement can be enormous.
Data-set Time Memory Time Memory
size for A for A for B for B Business value proposition
20 MB 0.15 s 21 MB 6.00 s 2 MB The methods used to determine the effi-
40 MB 0.30 s 41 MB 12.00 s 2 MB ciency of normal engineering problems
60 MB 0.45 s 61 MB 18.00 s 2 MB typically allow clear results to be deter-
mined. Computing systems do not always
provide such clear-cut decisions. In reality
From this (trivially simple) experiment, there are trade-offs between the different
one might conclude that application A ways of solving any given computing prob-
requires 0.0075 seconds per megabyte of lem; for example, using more processing
data, whereas application B requires 0.3 sec- power against using more memory or vice
onds per megabyte, so application A is 40 versa.
times faster than application B. But the A programmer who is aware of the con-
amount of memory occupied by applica- straints applicable to the system being
tion A grows alarmingly as the size of developed can attempt to construct appli-
the data set grows, whereas application cations that maximize some quantifiable
B uses only a constant amount of mem- measure of efficiency. The determination
ory. Of course, there are many other fac- of efficiency within large corporate systems
tors to be taken into consideration, such as environments is a very difficult problem
how much human intervention each appli- to solve, due to the interaction of many
cation needs, how much network band- systems-related variables. For example, a
width they require, and how much mem- corporate web server servicing simple web-
ory the application processing the data viewing requests involves many processes
requires. and sub-processes, including the network-
This illustrates an important concept: data-traffic profile, the efficiency of the
there is often a trade-off between one form web-page design, the hardware that sup-
of efficiency and another. To make a pro- ports the system, and the efficiency of the

125
Electronic data interchange (EDI)

firewall. Each factor has to be taken into


Electronic data interchange (EDI)
account when determining the efficiency
of the overall system. Typically, stress and Foundation concepts: Network, Database.
load tests are undertaken to determine Definition: Electronic data interchange is the trans-
whether a system will support the aver- mission of data or signals between two or more
age and maximum loads to be placed entities.
upon it.
Many applications that are highly sen- Overview
sitive in terms of their efficiency require- The origins of electronic data interchange
ments are carefully examined, and, when (EDI) go back as far as the beginning of com-
deployed, reside upon their own platforms mercial computing (or data processing as
in order to maintain the integrity of those it was commonly referred to in the 1960s
performance characteristics. and 1970s). EDI exists to connect two orga-
nizations that need to transfer or exchange
information.
Summary of positive issues In early EDI, data-exchange methods
The theoretical efficiency of a critical were created in an ad hoc manner to solve
strategic system can be determined. This is individual problems between, for example,
vital for systems whose failure to perform a manufacturer and a vendor, and scala-
at a certain level may involve serious prob- bility and compatibility of technology were
lems with processes or, more importantly, not seen as important. Typically a company
risk to human life, e.g., aircraft systems. would have many ‘‘standard” file struc-
Stress and load tests can be used to deter- tures, with different versions being cre-
mine performance and efficiency profiles ated for each client, buyer, or supplier as
empirically. needed.
Not only was managing the different
Summary of potentially negative issues file structures and data types problem-
The determination of systems efficiency is atic, with companies incurring large main-
a complex and time-consuming problem tenance overheads, but also it had the
and failure to do it for a strategic or crit- side effect of locking companies together,
ical system can be extremely detrimental trapped by the proprietary data formats
to an organization and its performance. It they shared. This had two effects; for sup-
is essential to decide exactly which aspects pliers it created a major barrier to entry
of efficiency are most important because against competitors who would usually not
one aspect might be increased only at the have the same file-format standards, and
expense of another decreasing. for customers it resulted in switching costs
each time they changed data format.
References The proprietary advantages enjoyed by
r N. Wirth (1978). Algorithms plus Data companies who forced their data formats
Structures Equals Programs (Englewood upon their business partners could only
Cliffs, NJ, Prentice-Hall). last so long, and it became apparent to
r D. E. Knuth (1975). Fundamental all parties, including the United States gov-
Algorithms (New York, Addison-Wesley). ernment, that this situation was unfair
r M. Hofi (1975). Analysis of Algorithms and anti-competitive. In January 1986 a fed-
(Oxford, Oxford University Press). eral mandate was issued, stipulating that
certain industries, including healthcare,
Associated terminology: Computability, could not use proprietary EDI formats and
Complexity. that only one standard messaging system

126
Electronic data interchange (EDI)

would be allowed, that being defined by Summary of positive issues


the American National Standards Insti- EDI is used to send millions of messages
tute and known as ASC (X.12). Several and files every day all over the globe; both
groups, including the United Nations and the X.12 and the XML approach are widely
member countries, have carried the devel- used and are undergoing constant revision.
opment of this data-exchange standard The X.12 standard was created in 1979 and
forward: UN/EDIFACT (United Nations Rules is widely used, with a large body of indus-
for Electronic Data Interchange for Admin- try knowledge and support to assist X.12
istration, Commerce and Transport), The users and developers. The standard cov-
European Board for EDI Standardization ers a very wide range of documents, with
(EBES) through the European Committee approximately 300 transaction sets having
for Standardization (CEN), and the Japan been developed. The X.12 standard from
Association for Simplification of Interna- Version 4010 onward is year-2000 compli-
tional Trade Procedures (JASTPRO) created ant and the US federal government man-
a series of EDI protocols and associated dated that healthcare providers be X.12
standards. compliant by August 1, 2003 in accordance
The rise in importance of electronic data with the 1996 Health Insurance Portabil-
interchange and the need to pass differ- ity and Accountability Act (HIPAA). XML,
ent data types between entities have caused a platform-independent approach to data
the term EDI to be refined, with EDI exchange, also continues to grow in pop-
being used to refer to data-processing-type ularity, with both open-source and propri-
transactions and inter-organizational sys- etary software available to support its user
tems (IOS), and ‘‘e-business” referring to the base. The use of X.12 and XML provides
transfer of more flexible data types between companies with the opportunity to reduce
entities. transaction costs, speed data exchange, and
IOS has been facilitated by the rise of simplify procedures.
open standards such as the Extensible
Markup Language, known as XML, which Summary of potentially negative issues
is a simple, flexible text format that orig- The dual approaches open to organizations
inated from the ISO 8879 standard and wishing to undertake EDI (X.12 and XML)
the advance of network protocols such as for data messaging can cause an increase
TCP/IP. in overhead if both mechanisms have to
be employed. The technical and human-
resources overhead associated with imple-
Business value proposition menting an X.12 solution can be consid-
EDI has evolved to allow highly flexible erable and it is less flexible than an XML
messaging between two or more entities. solution. However, X.12 is evolving toward
Two main approaches are followed: the the incorporation of an XML component.
exchange of fixed file formats such as those XML adoption by companies and software
found in the X.12 standard created by ANSI, vendors continues to increase, due in part
and the exchange of flexible file formats to its flexibility. As an open standard, XML
through the Extensible Markup Language may be amended and changed to accommo-
(XML) which has been developed under the date new requirements faster than other
auspices of the World Wide Web consor- formats.
tium (W3C.org) and the W3C XML working
groups. Both approaches allow entities to References
connect to each other and send data in an r http://www.x12.org/x12org/
efficient manner. international/index.html.

127
Email

r http://www.unece.org/trade/untdid/. off must still be accepted and kept some-


r http://www.cenorm.be/default.htm. where, and delivered as soon as the com-
puter is online again. That is why email
Associated terminology: XML, X.12,
servers came into existence, together with
Health Insurance Portability and
the internet email protocols SMTP, POP2,
Accountability Act.
POP3, and IMAP.
To receive email, a person must have
an account on an Email server. Any com-
Email puter can be an email server, the required
software is widely available and puts lit-
Foundation concepts: Internet, Protocol. tle load on the system. The only require-
Definition: A system for exchanging messages and ment for an email server is that it must
data in a non-interactive mode. be permanently on, and permanently con-
nected to the internet. To be useful, it must
Overview also have a Static IP Address, but that is not
Email has been around, in one form or strictly necessary. Having an account is not
another, for about 40 years. Originally, an onerous requirement; it simply means
most computer users had an account on a that the server must know about you (you
mainframe, a single large shared computer. have a user-name and a secret password)
If one user wanted to leave a message for and it must set aside a little disk space to
another, they could simply store some text keep your messages until you read them.
in a particular disk location associated with ISPs usually provide their customers with
that user’s account, and the next time the access to an email server as part of the
intended recipient logged in they would package.
be shown the message. Very simple soft- When email is sent to a remote com-
ware automated the tasks of sending and puter, an internet connection using TCP/IP
receiving email, and the resultant system is made to the SMTP (Simple Mail Trans-
was popular, low-maintenance, and easy fer Protocol) service application that must
to use. be running on that computer. After a short
The coming of the internet complicated exchange in which the sender and recipient
matters, because users now expected to be are identified, the message itself is trans-
able to send email messages to correspon- mitted and the connection is broken. An
dents on remote computers. This was not a SMTP server keeps a special file or folder,
particularly difficult enhancement to the known as a mail box, for each known user;
system; all it needs is a special-purpose the new message is appended to the end of
application running permanently in the the appropriate mail box.
background on each email-enabled com- Whenever an email-browser client is
puter, waiting for email messages to arrive started, it automatically makes an inter-
over the network connection, and dealing net connection to any email servers that
with them exactly as local emails had been it has an account on. This time, it does
dealt with before. The email-sending appli- not connect to the SMTP server, which is
cation simply had to know how to transmit used only for sending, but connects to the
outgoing messages. POP2 (Post Office Protocol version 2) server,
What complicated email was the coming the POP3 server, or the IMAP (Internet Mes-
of the personal computer. Personal comput- sage Access Protocol) server. All three of
ers spend a lot of time turned off or other- these provide the same basic services. IMAP
wise disconnected from the internet. Email is the most sophisticated, providing many
that arrives when a computer is turned additional features; POP2 is the simplest.

128
Email

Users are normally able to select which is a message to all members of a predefined
to be used in the options for their email- list.
browsing client.
When the email-browsing client success- Business value proposition
fully connects to the POP or IMAP server, it Email has become one of the most valu-
identifies itself by sending the user-name able of all corporate communication tools.
and secret password, and is given a list of Email is accessible almost universally on a
any unread emails waiting for it. Usually, it wide variety of devices, and has become
will then ask for each of those messages in almost indispensable in today’s organiza-
turn, and request that they be deleted, then tion.
display the list of messages to the user, and The technology of email has evolved suf-
allow him or her to do with them as they ficiently that end users are unaware of the
will. The whole procedure of connecting to majority of technical issues involved, since
the POP or IMAP server and retrieving the these fall upon the network manager. Users
list of new messages is repeated every few do, however, expect a seamless service from
minutes for as long as the browser contin- their systems and demand as close to 100%
ues to run. That is the only way that new up-time as possible, and, when the ‘‘email
messages are detected; the server has to be server” is down, they expect all unread
asked explicitly, it will never send a notifi- emails to be available when the system
cation. comes back on line. While the creation of
Email messages were originally nothing an email server is quite straightforward,
more than plain text, but modern systems and in an ideal world providing 100% relia-
allow any kind of data to be attached to the bility would be no problem, this is not the
message. Attachments are sent to the SMTP case in a real-world implementation.
server as a continuation of the main mes- The email systems of organizations are
sage, all in one unbroken transmission. It in essence the digital equivalent of a cor-
is the email-browsing client that picks out porate ‘‘front door” and everyone’s email
the attachments and treats them as sepa- comes and goes through it. This makes
rate entities. Modern systems allow whole it both vital from a strategic communi-
web pages to be sent as the main mes- cations perspective and simultaneously a
sage, so it is displayed to the recipient with key point of attack for those with evil
all typesetting and images in place. If the intent. The network manager may employ
email viewer has insufficient security set- a variety of techniques to prevent email
tings, such a message may also contain an from being attacked and these include
embedded program that will be run auto- proxy servers and firewalls to prevent spam
matically as the message is displayed. This attacks and emails with dangerous attach-
is a sufficient opening for a virus to be ments entering the system; Trojan horses,
installed and activated. viruses, worms, and other programs that
Broadcasting is not possible with current users unfortunately open by mistake are
email systems: every message must have a primary concern. It is problematic that
a specific recipient stated, so sending to these malevolent system attacks continue
multiple recipients requires multiple send- to evolve and change, forcing the network
ings. There are products known as ListServs manager and corporations (as well as indi-
(no error: there is no ‘‘e” after the ‘‘v”) that viduals) to deploy ever-increasing resources
automate the sending of a single message to defend against them. One of the worst
to a long list of multiple recipients. Email scenarios for network managers to con-
clients generally understand the idea of sider is the complete loss of the email
‘‘mailing lists” and will automatically send server with the discovery that there is

129
Email

no backup. Many system managers have emails from customers and vendors, some
rightly lost their jobs over this neglect and legitimate, some not. It is advisable for
it is vital that a backup procedure is in executives to have private email addresses
place and checked frequently to ensure as well as their more public ones; the
that it works; even a few hours’ worth non-private emails can then be filtered
of lost emails can be considerably damag- through an assistant. For example, it
ing to many organizations. The network is particularly easy for people to guess
manager needs to consider issues such as an executive’s email addresses if every-
capacity planning for the server and the one in the company has the same email
memory demands placed upon it by users. address form (e.g., psmith@company.com,
Many users save all their emails and wish jsmith@company.com). Some system man-
to access them all from their inbox and agers establish email address lists for pre-
are reluctant to archive them, their email validated correspondents and filter incom-
box acting as a ‘‘Rolodex.” This can place ing email for executives through that list.
significant storage demands upon the file
server and the network manager will usu-
Summary of positive issues
ally put in place quotas and other proce-
Email aids communication when used
dures to prevent users being wasteful of
in a filtered, secure environment. Email
storage.
allows users to access communication asyn-
Many companies have strict policies
chronously (when they want to, rather than
about the type of emails that can be sent
when the system or external forces decide
(both internally and externally), since they
it is time). Email servers are easy to estab-
are documents of the organization and as
lish and the technology protocols are well
such will be archived by the organization,
known. Third-party email systems are avail-
and are subject to discovery in legal pro-
able and accessible via the internet. Email
ceedings. Workers need to understand this
can carry attachments and can be sent out
basic point and must not send internal
to many addresses at once via a ListServ.
emails saying ‘‘don’t buy this stock, my
Email can provide an audit trail of cor-
research shows it is awful” and then send
respondence and may be required to be
out external emails to customers saying
archived for regulatory compliance. Email
‘‘buy this stock, it is great” because these
can be accessed through a wide variety of
conflicting positions may one day be used
devices and services.
in a lawsuit. For legal and other reasons
some people simply prefer not to use email.
President G. W. Bush stated ‘‘I don’t email,” Summary of potentially negative issues
adding ‘‘and there’s a reason. I don’t want Email can facilitate attacks on corporate
you reading my personal stuff.” (speech at networks and individuals through spam
the American Society of Newspaper Editors and malicious attachments. Email can
conference in Washington, DC, April 14, reduce productivity: the easier it is to
2005). ask a question, the more likely a person
For most people email is too critical is to ask that question without thinking
to day-to-day life for it to be neglected, about it first; since the days when access
but email can become very intrusive and to customer support required writing a
detract from workers’ productivity. Again letter or paying for a long-distance tele-
procedures and policies need to be put phone call, the number of totally unnec-
in place to assist productivity. Executives essary enquiries has grown enormously.
are also frequently the subject of direct Employees will often spend an inordinate

130
Encryption

amount of time sending and responding to techniques. The only way to be confident
personal emails if a proper policy is not that an encryption method is safe is to use
enacted and enforced. Archiving requires one that is well known and has already
the expenditure of resources by network withstood years of attack by the entire
managers to ensure regulatory compliance. cryptanalysis community.
Encryption always depends upon a Key,
References
r D. Wood and M. Stone (1999). something like a secret password or num-
ber that plays an essential part in the
Programming Internet Email (Sebastopol,
encryption process. The same encryption
CA, O’Reilly Press).
r S. Cobb (2002). Privacy for Business: Web scheme used on the same message, but
with a different key, will yield completely
Sites and Email (Saint Augustine, FL,
different results. The safety of encryption
Dreva Hill LLC).
is based on keeping the key secret, not
Associated terminology: Instant the method. The simplest of cracking tech-
messaging. niques, and in many cases the most success-
ful, simply involve trying out every possi-
ble key. If, for example, encryption keys are
Encryption four-digit PINs, then there are only 10 000
possible different keys. A computer could
Foundation concept: Security. try them all out in turn in a fraction of
Definition: Disguising information to make it inacces- a second, and be guaranteed to crack the
sible to certain others. message.
It is always essential that there should
Overview be so many possible keys that they could
Obscurity is not security. Hiding informa- not possibly all be tried. This is clearly
tion, or encoding it using a strange pro- illustrated by the example of 56-bit DES
cess that ‘‘nobody could ever guess,” pro- (Data Encryption Standard), a well-known
vides no security at all. One disgruntled worldwide standard. The 56-bit DES encryp-
employee, one inspired enemy, and all is tion scheme supports 256 (which is about
lost. Since the first electronic computer was 70 000 000 000 000 000) different keys. A
built in 1943 for the specific purpose of 4 GHz Pentium 4 could probably perform
breaking enemy codes, data security has about 7 000 000 experimental decryptions
moved firmly into the realm of abstract per second, which means that it would
mathematics. Anybody who tries making take around 300 years to try every pos-
up their own encryption schemes without sible key. That may sound secure, but it
extensive knowledge of cryptanalysis is set- means that any industrial rival could sim-
ting themselves up for disaster. ply buy 300 computers, leave them run-
The simple encryption schemes that ning for a year, and have everything. Or
have been tried and tested for centuries, 3000 computers could do it in about a
which are based on alphabetic substitu- month. Deploying 3000 computers would
tion (for example, A → K, B → F, C → W, cost some money, but there are economies
D → A, . . .) and permutation (for exam- in scale, and the value of being able to read
ple, hello → lhloe), can be cracked in a all rivals’ secrets would certainly be much
fraction of a second with the aid of a higher. To make matters worse, special-
computer. A new encryption scheme may purpose integrated circuits have been fab-
seem safe to its inventor, but could crum- ricated that are many times faster. DES is
ble quickly under attack from unsuspected expected to be superseded by AES (Advanced

131
Encryption

Encryption Standard); it has a more open ric system, it is necessary to have just one
design, and supports 256-bit keys. AES is key pair (an encryption key and a decryp-
the first encryption method approved by tion key) for each individual involved, and
the US government for ‘‘top secret” docu- all combinations of secure communication
ments to be released to the public. become possible. See Public key--private key
Increasing the key size has an inordi- for further information.
nate effect on the security of the system: Another encryption scheme of value is
56-bit DES keys are approximately 16-digit the One-way hash. This is a kind of encryp-
numbers; 112-bit DES keys are approxi- tion that can not be decrypted even if
mately 32-digit numbers, a fairly small the key is known. The original message is
increase in size, but a major increase in never recoverable even by legitimate read-
security. It increases the time required ers. Although it sounds pointless, this tech-
to try all possible keys on a farm of nique is very valuable, and makes possible
3000 computers from one month to Digital signatures and Digital certificates.
2 000 000 000 000 000 000 years. Of course, All encryption schemes (with one excep-
there is no practical limit to the size of a tion, the One-time pad) are in principle crack-
key, except that legitimate encryption and able. The only security lies in the fact that
decryption by those who do know the key the entire world’s cryptanalysts are work-
will also take a little longer. ing on them and have been for a long time.
The need for large key sizes brings about It would be very difficult to keep a flaw
its own problems. Nobody can be expected secret, and what discoveries one lab does
to remember a 32-digit number, but if ever manage to keep secret will probably soon
it is written down or recorded in any way, be repeated by others.
simple theft becomes an attractive means
of breaking the best encryption. Key secu- Business value proposition
rity and key exchange are serious problems There are many tried and tested encryp-
for the security-conscious organization. tion algorithms, and the choice of algo-
There are two basic forms of encryption rithm should not be made without com-
method: symmetric, in which the same key petent research. Amongst the best known
is used for encryption and decryption; and are DES, a symmetric algorithm that lies
asymmetric, in which case the key that under a cloud of suspicion: it contains
must be used to read a message must be unexplained steps and there is some belief
different from the one used to encode it. that it contains a Back door that would allow
Symmetric systems require that a different faster cracking by those in the know with-
key is produced for each possible pairing of out needing a key search. AES will proba-
communicants: if four people A, B, C, and D bly replace DES in the near future. RSA is a
wish to communicate securely, they could slower but highly trusted asymmetric algo-
pick a single key for their group, but that rithm. RC4 is simple, fast, and symmetric.
would mean that they have no internal pri- PGP has become almost an internet stan-
vacy. Instead six different keys are needed, dard, using a combination of symmetric
one for when A and B communicate, one and asymmetric techniques. Many encryp-
for when A and C communicate, one for A tion algorithms are subject to patents, the
and D, one for B and C, one for B and D, and status of which changes as time progresses.
one for C and D. To give N people private
communication, approximately 12 N 2 keys Summary of positive issues
are needed; the key-management problems A wide variety of encryption mecha-
can become substantial. With an asymmet- nisms and tools is available to implement

132
End-user development

encryption. A proactive encryption plan relies upon keys not being kept by any party
will result in greater data security, but the one owner.
customer confidence, and corporate-data It is easy for a programmer to over-
integrity. estimate the security of a new encryp-
tion scheme they have invented. Assessing
Summary of potentially negative issues the security of an encryption algorithm
In the United States there are restrictions is an exceptionally difficult task, requir-
on the export of some encryption tech- ing extensive training and experience well
nologies. In 1996 the US Department of beyond that provided by even the most
Commerce relaxed export restrictions, stat- advanced technical degrees. Management
ing that ‘‘Any encryption commodity or would be well advised not to give full rein
software, including components, of any to amateur cryptographers designing cor-
key length can now be exported under a porate security procedures.
license exception after a technical review
References
to any non-government end-user in any r R. Needham and M. Schroeder (1978).
country except for the seven state support-
‘‘Using encryption for authentication
ers of terrorism.” Encryption software can
in large networks of computers,”
not be exported from the United States
Communications of the A.C.M., Volume 21,
to Cuba, Iran, Iraq, Libya, North Korea,
No. 12.
Sudan or Syria. However, in 2004 several r B. Schneier (1996). Applied Cryptography
restrictions were re-implemented: ‘‘mass-
(New York, John Wiley and Sons).
market” encryption products with more r http://www.bxa.doc.gov/encryption/
than 64 bits of symmetric key require
MassMarket Keys64bitsNUp.html.
review before they may be exported and
re-exported without a license. The law cov- Associated terminology: Public
ers specific products and technologies and key--private key, One-way hash, Password.
US-based companies need to understand
their obligations under this legal frame-
work. The UK and the EU have similar laws End-user development
pertaining to symmetric algorithms over
64 bits and require the granting of an Open Definition: End-user development refers to any sys-
General Export License (OGEL) and prohibit tem that is either written or configured by an end user
export to Afghanistan, Iraq, Libya, North rather than being completely developed by a systems
Korea, and China. professional for that user.
Regardless of legal restrictions, strong
encryption algorithms are universally Overview
known, and even published in books, so The development of computing from the
a moderately competent programmer any- 1940s until the 1980s allowed users to
where in the world could recreate them obtain a wide variety of processing solu-
easily. tions and reports. These systems were gen-
The UK also has enacted the Regulation erally built and supported by a professional
of Investigatory Powers Act 2000, which staff of operators, systems analysts, systems
requires that anyone who has created an programmers, and applications program-
encryption key must be able to produce the mers. This led to delays in the produc-
key if required to do so by a legal author- tion of systems and in the implementation
ity. This is a very controversial act, since of changes, even relatively insignificant
the safety of public-key encryption systems changes on a report generated by the

133
Enterprise information portal (EIP)

system. As technology became more flexi- side the ability of end users to configure
ble and organizational pressures demanded internal systems to meet their own spe-
more flexible reporting, the delay from cific systems needs. Applications can usu-
request to delivery became difficult to jus- ally be configured to form a common basis
tify. In order to overcome this, technologies upon which all users develop their systems,
were deployed to allow users to develop thus avoiding the problems associated with
their own applications. Some of these ran every user developing disparate systems.
on the corporate data systems while oth-
ers were stand-alone application suites that Summary of potentially negative issues
allowed users to create their own databases It needs to be remembered that end users
and applications on the corporate client-- are usually not computer scientists or infor-
server system. For example, this allows a mation technologists. While they may be
user at a company that has an ERP sys- skilled in the processes around which
tem to create their own reports to their they develop their systems, they might not
own individual specification. End-user com- be fully aware of the issues and limita-
puting also occurs when users design (or tions that can have significant effects. This
more typically configure) applications that requires that any end-user development be
are then run upon a personal computer. done in a controlled and carefully managed
Spreadsheets are a very common example. environment. A further issue for organiza-
tions is that, while the technology depart-
Business value proposition ment of the organization is freed from the
The ability of users to develop their own obligation to perform all systems develop-
data sets and then design and run appli- ment, end users now need to be trained
cations to provide customized reports typ- and have their skill sets upgraded as the
ically reduces the demands upon the IT technology and the organization’s use of it
professionals and the IT organization over- change.
all. Usually these applications are devel- Reference
oped as configurable modules of commer- r M. Mahmood (2005). Advanced Topics in
cial applications. This saves the end user End User Computing Series, 2nd edn.
from having to become a technologist in (Hershey, PA, Idea Group Publishing).
addition to doing their own job, while
maintaining some degree of integrity in Associated terminology: ERP, Visual Basic.
the system being generated and used. This
is important because systems frequently
Enterprise information portal (EIP)
need to download data and communicate
with other users as well as with corpo- Definition: An enterprise information portal provides
rate offices, so a wide mix of data types, users with access to data, applications, and reporting
structures, and reporting would cause capabilities directly from their desktop, drawing from
more problems than end-user development corporate ERP, business intelligence, and data ware-
solves. house systems.

Summary of positive issues Overview


End-user development frees the corporate Enterprise information portals (EIPs), also
IT organization from having to perform known as corporate portals or Enterprise
each and every activity associated with user portals, allow corporate users direct access
demands. There are many common applica- to the data, applications, and functional-
tions that can be provided to users to help ity required to perform their job function
them develop their own solutions along- from their desktop computers.

134
Enterprise resource planning (ERP) systems

EIPs can be incorporated as a configura- interoperability,” Communications of the


ble customizable desktop into an ERP envi- A.C.M., Volume 47, No. 10.
ronment. These EIPs provide easy access
Associated terminology: ERP, Legacy
to all the users’ functional requirements
system.
pertinent to performing their tasks, for
example, a logistics manager providing the
user with direct access to the pertinent ERP
Enterprise resource planning (ERP)
tools, business analytics, data warehouses,
and communications tools required.
systems
The role of an EIP is to provide a mech- Foundation concept: Database.
anism for searching the data in all of Definition: An enterprise resource planning system is
the organization’s information systems, or an integrated software environment that integrates
as many of them as the user’s clearance processes and data using a single database.
allows, in order to solve a problem. This has
made EIPs suitable for use within the con- Overview
text of an ERP system, since ERP systems ERP systems are designed to integrate
tie all the corporate data together through business functions across an organiza-
a single database. It is possible to create tion using a single shared database. The
an EIP that works within a legacy system’s functions of an ERP are contained in a
environment involving disparate systems, set of modules whose processes can be
but this is a much more difficult and tech- configured to meet organizational needs.
nically challenging task. These modules typically include account-
ing, finance, logistics, procurement, man-
Business value proposition ufacturing, human capital management,
EIPs enable a user to have a desktop config- and customer-relationship management
ured especially for their job function. This functions. The modules in an ERP inter-
enables the creation of a very effective work act and communicate through a shared
environment that does not require users to database, typically using a database server,
go and search for the tools required to per- in which all of the transactional data asso-
form their tasks and then integrate them ciated with the ERP is stored. The strength
together. EIPs enable user-to-user commu- of the system is in part due to the nature of
nications to be tailored to the business the database structure in that a data item
function, reducing overhead and increas- is stored only once (for example, in an ERP
ing effectiveness. system, a vendor’s address is stored only
once, whereas in a traditional system com-
Summary of positive issues posed of multiple separate systems, a ven-
EIPs provide users with function-specific dor’s address may be stored in the database
desktops. EIPs can be built into an ERP sys- associated with the accounting system, in
tem’s configuration. EIPs facilitate access to another database associated with the logis-
all corporate data sources. tics system, and yet again in the customer-
relationship-system database). The tradi-
Summary of potentially negative issues tional approach was problematic because it
EIP systems are technically challenging to was highly inefficient and error-prone, with
establish in a legacy-systems environment. data items repeated in each database poten-
tially being inconsistent. The centralized
Reference nature of the ERP database allows the data
r M. Smith (2004). ‘‘Portals: toward an to be accessed by any module, and, should
application framework for any changes to the database be made, the

135
Enterprise resource planning (ERP) systems

new version of the data is immediately vis- frequency identity tags), relieving the cor-
ible and accessible by all modules. porate IT organizations of the task.
ERP systems enable end-to-end process The ERP environment allows greater
cycles such as the procurement-to-payment security and accountability as well as
cycle and are capable of interacting with better corporate governance (for example,
ERP systems in external organizations simplifying Sarbanes--Oxley and HIPAA
through XML or other file types. ERP system compliance). ERP solutions, vendors, and
modules are typically run on a single oper- consultants are available to support com-
ating system and the modules have a con- panies of all sizes from small businesses to
sistent user interface. the largest global enterprises. ERP systems
facilitate the development of data ware-
Business value proposition houses since the single database resolves
ERP systems are composed of a set of many of the data-conflict issues associ-
modules that perform a variety of busi- ated with extracting data from multiple
ness activities ranging from logistics to sources.
customer-relationship management. The
primary benefit of the ERP systems phi- Summary of potentially negative issues
losophy is that the system modules are ERP system implementations can be com-
designed to integrate together, interacting plex and expensive, depending upon the
through the single database. The modules scale of the implementation, the number
are configurable to allow companies to of modules, the number of users, and the
align the processes contained in the mod- number of critical functions being pro-
ule with their own needs. vided. Failure to implement them correctly
ERP systems vendors have traditionally can result in severe consequences, so the
attempted to encode best industry process decision to implement an ERP is a board-
practices into their systems. Consequently, level decision. An ERP system must be
ERP vendors have created industry-specific assessed with respect to the functionality
versions of their systems, e.g., healthcare, that a company will require from it. The
manufacturing, and retail. ERP systems inability to meet the critical core needs of a
allow organizations to focus upon their company might tempt organizations to cus-
core business processes and decommission tomize their ERP system, something that
old, hard-to-maintain, legacy systems. should generally be avoided.
Several vendors have written their ERP
Summary of positive issues systems in proprietary specialized program-
The design of ERP systems as integrated ming languages and this can affect the
software environments whose modules effort required to create any specialized
communicate through a single database modules or add-on functions. Some ERP sys-
relieves organizations of the burden of tems do not support common operating
maintaining multiple databases and mod- systems and databases; this can affect the
ules that may have run on disparate operat- selection of the systems architecture.
ing systems and architectures. ERP vendors
Reference
constantly revise their systems and work to r D. O’Leary (2000). Enterprise Resource
incorporate not only best practice at the
Planning Systems (Cambridge, Cambridge
process level but also the latest legal and
University Press).
regulatory requirements. ERP vendors also
work to incorporate the latest technolo- Associated terminology: XML, Operating
gies into their environments (e.g., radio- system.

136
Entity-relationship diagram (ERD)

must have at least one manufacturer asso-


Entity-relationship diagram (ERD)
ciated with it).
Foundation concept: Database. The ERD technique involves three major
Definition: An entity-relationship diagram is a dia- phases, beginning with the Conceptual data
grammaticformusedinthedesignofthelogicalstruc- model, the highest-level view of the domain,
ture of a database and as part of its documentation. containing only the entities and the rela-
tionships that exist between them. The con-
Overview ceptual model is then refined into a Logi-
Entity-relationship diagrams (ERDs) were cal data model in which attributes and Keys
first developed by Peter Chen in 1976. They (unique identifiers for individual entities)
use a set of diagrammatic forms to rep- are indicated and normalization is per-
resent the Entities, Attributes, and Relation- formed. The third level is the Physical data
ships in a database. An entity is a class of model, which again refines the design to
object for which the designer (systems ana- take into account physical implementation
lyst) wishes to store data. An entity has constraints. These three stages enable the
associated with it a set of attributes that systems analyst to create a design that is
define the form that the data takes. For workable and, whenever possible, optimal.
example, in a very simplified purchasing-
department database, there might be two Business value proposition
different kinds of data recorded: one to rep- ERDs provide the systems analyst with a
resent a manufacturer and one to repre- powerful tool through which systems can
sent a product. The attributes of the man- be modeled and refined, leading to the ulti-
ufacturer might be name, address, contact mate implementation. The ERD modeling
telephone number, etc.; the attributes of a technique has been used in various forms
product might be the product description, for 30 years and continues to be one of the
its UPC, and its price per unit. keys to successful systems design.
Relationships show the associations that
exist between entities; thus the fact that a
particular manufacturer produces a partic- Summary of positive issues
ular product is represented by a link bet- ERDs have been used by a large group of
ween individual entries in the two tables, or analysts and IT professionals for over 30
by a relationship between the two entities. years. They are supported by consultants,
The three main diagrammatic forms tools, and an extensive academic and prac-
used are a rectangular box for entities, titioner literature.
an ellipse for attributes, and a diamond
for relationships. Lines between the shapes Summary of potentially negative issues
indicate how they relate to one another; The ERD style of design is being superseded
the lines may be decorated in various ways by other techniques, primarily UML, that
to indicate particular details of a relation- are more amenable to the Object-oriented
ship (perhaps showing that each product style of development and programming.

name addr phone descr UPC price

manufacturer makes product

137
Ethernet

Reference the sender will notice that there was inter-


r P. Chen (1976). ‘‘The entity-relationship ference, and automatically attempt retrans-
model -- toward a unified view of data,” mission after some random delay.
A.C.M. Transactions on Database Systems, This kind of system is technically
Volume 1, No. 1. known as CSMA/CD: Carrier sense multiple-
access/collision detect. It works very well with
Associated terminology: UML.
only a moderate number of computers con-
nected to the LAN, or when overall network
usage is low. As network traffic increases,
Ethernet the likelihood of collisions also increases,
until it reaches a critical point at which
Foundation concepts: Network, Protocol. attempting to transmit more data results
Definition: A common connectivity medium for local in less data actually getting through. How-
area networks. ever, the ability to create a fully func-
tional network without any extra support
Overview hardware is very attractive, and ethernet
Ethernet was designed by the Digital Equip- remains the most popular set of standards.
ment Corporation, Xerox, and Intel in the On an ethernet LAN, each computer has a
mid 1970s to provide a means for connect- unique address (known as its MAC -- media
ing together a number of computers in a access control -- or Hardware address), and
Local area network (LAN), using very cheap all transmissions are strictly formatted to
hardware for the connections. The design include source and destination addresses,
was extremely successful, and is the basis plus error-detection codes to ensure that all
of most LANs today. Although the Network collisions are detected.
cards required to add ethernet functionality The original ethernet specification used
to a computer were originally not particu- transmission speeds of 10 Mbps (million
larly cheap, a LAN requires no other infras- bits per second) over a single Coaxial cable
tructure besides extremely low-cost cables (see Cables and connectors for details). Newer
and connectors. Ethernet is now defined by versions use speeds of 100 Mbps (Fast Eth-
the IEEE 802.3 standard. ernet) and 1000 Mbps (Gigabit Ethernet),
The basic idea behind an ethernet net- and different kinds of cables; coaxial cable
work is that all of the computers involved is not used above 10 Mbps, and is usu-
are connected to a single long strand of ally replaced by Twisted-pair cables. Coaxial
cable, which they use in much the same cables look like the cables used to bring
way as people using ‘‘walkie-talkies”; the cable television into a home, or to connect
‘‘ether” in ethernet refers to the (nonexis- television sets to rooftop antennæ; twisted-
tent) medium through which radio waves pair cables look like telephone cables,
are imagined to travel. When one computer but the connectors are of a slightly dif-
needs to transmit some data, it simply ferent size. Fiber optics provides a third
waits until the network is quiet -- no other alternative, the cable being distinguished
transmissions seem to be in progress -- by the complete absence of metallic
and transmits its own data. Inevitably this conductors.
will result in occasional collisions -- two The different versions of ethernet are dis-
computers simultaneously noticing that tinguished by a simple naming scheme,
the network is quiet and starting their own which provides the familiar but some-
transmissions -- but this is not a problem. what mystical designations 10-base-2 and
Every computer constantly monitors all 100-base-T. These names always have three
traffic on the network; if a collision occurs, parts; the first is the transmission speed in

138
ETL (extracting, transforming, and loading of data)

millions of bits per second; the second is nology in use is that of ethernet, due to its
always the word ‘‘base” (except in the rare low cost and the wide range of technical
case of 10-broad-36, where ‘‘broad” indi- options for those deploying the system.
cates that multiple channels of a cable- The different options available to the net-
television system are used); the third indi- work designer need to be carefully consid-
cates the type of cable used. When the third ered through a set of parameters, includ-
component of the name is a number, as in ing the speed and bandwidth requirements
10-base-2 or 10-base-5, it means that coaxial for the organization, the budget available,
cable is used, and the number itself is the and the possible future need of the organi-
maximum length for a run of cable in hun- zation for an increased bandwidth capacity.
dreds of meters. So 10-base-2 means rather
cheap narrow coaxial cable with no runs of Summary of positive issues
more than 200 m (to be pedantic, the max- Ethernet is an almost universal network
imum is 185 m), and a transmission rate of technology. The technology is easy to imple-
10 Mbps; 10-base-5, also known as Thickwire ment, scale, and adapt to meet organiza-
Ethernet, uses a thicker and more expensive tional needs. Ethernet technology is used
coaxial cable to allow longer runs of up to and universally understood by computer
500 m. systems capable of being networked.
When the third component of the name
is a letter, it indicates a specific kind of Summary of potentially negative issues
non-coaxial cable. ‘‘T” refers to twisted- Network systems need to be monitored and
pair cable (telephone-like), and ‘‘F” refers stress tests must be performed. As network
to fiber-optic cable. A second letter or traffic grows, there is a potential for bottle-
digit may be added to give more specific necks to be created and systems performa-
details: 100-base-F is Fast Ethernet using nce to be reduced.
fiber-optic cables; 1000-base-T is Gigabit Eth- References
ernet using twisted-pair cables. The letter r C. E. Spurgeon (2000). Ethernet: The
‘‘X” is used informally to mean ‘‘anything,” Definitive Guide (Sebastopol, CA, O’Reilly
so 1000-base-X means all kinds of Gigabit Press).
Ethernet. ‘‘L” and ‘‘S” refer to light of long r W. Stevens (1994). TCP/IP Illustrated (New
and short wavelength on fiber-optic cables. York, Addison-Wesley).
When any kind of cable other than coax-
ial is used, it is no longer possible simply
to connect multiple computers to the same ETL (extracting, transforming, and
strand of wire; additional items of hard- loading of data)
ware such as hubs and switches are needed
to connect multiple systems. With twisted- Foundation concepts: Database, Data warehouse,
pair and fiber-optic cables, each segment of ERP.
cable connects only two devices, but it still Definition: A set of processes used to ensure that data
works in fundamentally the same way as is of a high quality, consistent, and uniform.
with coaxial cable. See Network devices for
details. Overview
The effective use of a data warehouse or
Business value proposition ERP system requires that the data in its
Several types of LAN technologies exist, database is of a high quality and uniform
including token ring, fiber-distributed data in nature. In order to achieve this, certain
interface (FDDI), and ARCnet. By far the processes are undertaken when the data
most extensive and popular network tech- warehouse is initially created or when data

139
ETL (extracting, transforming, and loading of data)

is added to it. These are the data-extraction, Once the data warehouse or ERP has
data-transformation, and data-loading pro- been established, all future data that is
cesses of ETL. written to the database needs to adhere
The very nature of data warehouses, to the data structures and conditions that
in that they are read-only data reposito- have been established for the database, and
ries intended to be used for analytic pur- hence must pass through a transformation
poses and compliance archiving, requires process.
that the data contained within them be of
the highest quality. ERP systems similarly
require data that is consistent and of a high Business value proposition
quality. To ensure that these criteria are A data warehouse that has not been popu-
met, the data placed in the systems first has lated with consistent data is useless since
to be extracted from the source database, the whole purpose of a data warehouse
then examined for a variety of possible cor- is to be able to use the consolidated data
ruptions prior to incorporation, and finally to perform Business intelligence (also known
placed into the database of the ERP or data as business analytics). Without consistent
warehouse. data, the old adage ‘‘garbage in, garbage
The extraction process is usually per- out” applies and may render any analysis
formed by a software tool that is able to useless.
configure itself to adapt to the type (e.g., ETL is also used when the database for
relational or hierarchical) and data struc- an ERP system is being created. As with
tures of the originating database. The tool data warehouses, ERP systems require a
may need to be configured for a variety of cleansed, consistent database if they are to
types of legacy database that may be over work as designed.
30 years old in some cases. Inside these
databases there is an infinite number of Summary of positive issues
possible data structures in which the data The data-warehousing and ERP industries
could be encoded. are mature, and this has led to a wide
The data-extraction tool must also be range of vendor-supported ETL products to
capable of combining the data from a manage the process. Vendors range from
set of sources, each potentially different database vendors, data-warehouse vendors,
in nature (e.g., a relational database, a and ERP vendors to specialty third-party
spreadsheet, an indexed sequential file, vendors. There are many consultants and
and a flat file). The process of data database specialists available to support the
transformation (sometimes referred to as task. Once the data has been ported over
data cleaning or ‘‘scrubbing”) ensures that to the new database, this facilitates future
the data warehouse contains just one system-migration efforts since the data is
unique data element for each data item already clean and consistent.
in the database. This may be necessary
because the originating databases may
have used different data representations, Summary of potentially negative issues
for example the warehousing database The creation of a data warehouse or ERP
system may have used ‘‘US” to repre- requires the data to pass through an ETL
sent a country, the customer-relationship- process; failure to perform this process cor-
management systems may have used rectly will lead to poor and possibly catas-
‘‘U.S.A.,” and the human-resources systems trophic systems performance. The ETL pro-
may have used ‘‘United States,” all meaning cess can be a time-consuming activity and
the same thing. requires significant effort to ensure that

140
European Union Directive on Privacy and Electronic Commerce 2002

the data is correct and amenable to future electronic communications services in


corporate needs, and that it conforms to public communications networks in
the system requirements. the community,” and digital exchanges
and subscriber lines connected to
Reference analog exchanges.
r R. Kimball and J. Caserta (2004). The r Article 4 covers the security
Data Warehouse ETL Toolkit: Practical requirements and calls upon providers
Techniques for Extracting, Cleaning, not only to take the appropriate
Conforming and Delivering Data (New ‘‘technical and organizational” security
York, John Wiley and Sons). measures, but also to notify their
subscribers of the risks.
r Article 5 calls upon member states to
‘‘ensure the confidentiality of
European Union Directive on Privacy
communications and related data
and Electronic Commerce 2002
traffic.”
r Article 7 defines the rights of
Definition: The European Union Directive on Privacy
and Electronic Communications aims both to protect subscribers to have itemized billing for
the personal data of individuals and to encourage the services.
r Article 8 expresses the obligations of
free movement of such data by having EU member
countries ensure the rights and freedoms of the indi- service providers regarding the
vidual to data privacy during data processing. presentation of calling-line
identification.
r Article 9 discusses the issues
Overview
The European Union Directive on Privacy surrounding location data.
r Article 12 defines the rights of
and Electronic Commerce, which contains
21 articles, was adopted on July 12, 2002 subscribers in relation to the capture
and an implementation requirement was of subscriber data in public
adopted by member states on October directories.
r Article 13 covers unsolicited
31, 2003. The directive supplements the
Telecommunications Data Protection Direc- communications such as direct
tive (97/66/EC), also known as the Commu- marketing via email. The article states
nications Data Protection Directive (CDPD). that email in which the sender
According to article 1, ‘‘The directive disguises or conceals their
harmonizes the provisions of the Mem- identification is prohibited.
r Article 14 states that there are no
ber States required to ensure an equivalent
level of protection of fundamental rights mandatory equipment requirements
and freedoms, and in particular the right but that equipment should be
to privacy, with respect to the processing compatible with fulfilling the other
of personal information in the electronic articles of the directive.
communication sector and to ensure the
free movement of such data and electronic Business value proposition
communication equipment and services in The European Union directive is a power-
the country.” ful and wide reaching initiative. The direc-
Key articles include the following. tive has a different perspective from the US
legislation. In the EU the principal philoso-
r Article 2 covers ‘‘the processing of phy is that an individual’s rights come first
personal data in connection with and that they must opt in to any data set
provision of publicly available rather than be automatically included in a

141
European Union Directive on Privacy and Electronic Commerce 2002

data set and than have to opt out if they ing recourse for individuals who wish to
desire. This prevents the merger of data remove or modify their own data.
sets within the business community with- r Choice; individuals must be given the
out the prior consent of the individual. opportunity to choose (opt out of)
The 2002 directive built on an earlier whether their personal information will
directive (95/46/EC) that discusses the con- be disclosed to a third party or used for
troversial issue of inter-country data trans- a purpose incompatible with the pur-
fer. Article 25 of 95/46/EC states that ‘‘The pose for which it was originally col-
Member States shall provide that the trans- lected or subsequently authorized by the
fer to a third country of personal data individual. For sensitive information, an
which are undergoing processing or are explicit affirmative choice (opt in) must
intended for processing after transfer may be given if the information is to be dis-
take place only if, without prejudice to closed to a third party or used for a pur-
compliance with the national provisions pose other than its original purpose or
adopted pursuant to the other provisions of the purpose authorized subsequently by
this Directive, the third country in question the individual.
ensures an adequate level of protection.” So r Onward transfer pertains to the disclosure
far the EU Commission has recognized five of information to a third party, stating
countries as meeting the provisions of the that organizations must apply the notice
article: Switzerland, Canada, Argentina, and choice principles.
Guernsey, and the Isle of Man, plus the US r Access covers the rights that individuals
Department of Commerce (DOC) through have pertaining to their ability to cor-
its ‘‘Safe Harbor Privacy Principles,” and the rect, amend, or delete information when
US Bureau of Customs and Border Protec- it is inaccurate.
tion for the transfer of air-passenger name r Security pertains to the ‘‘reasonable pre-
records in order to provide adequate border cautions” safe harbors must take to pro-
protection. tect personal information from loss, mis-
The United States has proved a difficult use, and unauthorized access, disclo-
market for the EU to ignore with respect sure, alteration, and destruction.
to data traffic, since US corporations oper- r Data integrity pertains to the relevancy of
ate in a ‘‘sectoral” framework that ‘‘relies the use of data in terms of the reason for
on a mix of legislation, regulation, and which it was collected; and section viii
self regulation” (US DOC) and are clearly details enforcement (US DOC).
not subject to EU law. A ‘‘safe-harbor” com-
promise was developed, in which organiza- References
tions can either join a self-regulatory pri- r Official Journal of the European
vacy program that adheres to the safe har- Communities, L 201/37, 2002.
bor’s requirements or develop their own r US Department of Commerce (2000).
self-regulatory privacy policy that conforms Safe Harbor Privacy Principles, Issues
to the safe harbor (US DOC). (Washington, DC, US Department of
The safe-harbor policy includes compo- Commerce).
nents pertaining to the following issues.
r Notice, notifying and informing individ- Associated terminology: Law cross-
uals about data collected and provid- reference.

142
Fiber optics

but microwave radiation in the environ-


Fiber optics
ment is a major concern for many). Con-
Foundation concept: Cables and connectors. verting between electronic signals and
Definition: Data encoded as pulses of light conducted radio or microwave signals for transmission
along flexible transparent wires as a replacement for requires expensive equipment.
data sent as electrical pulses along a flexible metal Fiber optics provides another alternative.
wire. Instead of an electrically conductive wire,
a light-conductive (i.e., transparent) fiber
Overview is used. A light shone down one end will
The most familiar means for transmitt- follow curves in certain kinds of transpar-
ing signals or data is by sending electri- ent fibers, and appear at the other end. It
cal impulses along a flexible metal (nearly seems as though the light follows a curved
always copper) wire. This has many advan- path, although it is in fact following a large
tages: electrical signals are the native inter- number of short straight-line paths, grad-
nal infrastructure of all computing devices, ually being reflected around curves. For
so very little signal conversion is required; long-distance transmissions, the fibers are
wires can easily be insulated to prevent very narrow strands of glass, and the light
cross-over with other signals; electronic sig- used is infrared (invisible to the human
nals travel along wire at the speed of light, eye) laser light. Laser-light emitters can be
so speed is maximized. There is also one designed to switch on and off at exception-
major disadvantage: when electric currents ally rapid rates.
pass through metal wires they generate Fiber optics are successfully used to
magnetic fields, and when metal wires pass transmit signals at almost 100 000 000 000
through varying magnetic fields, they pick bits per second, a far greater bandwidth
up induced electric currents. This means than is available with metal wires, over
that, whenever two signal wires run close distances of up to 100 miles, a distance
together, they will interfere with each other, that would require a number of retransmis-
and whenever a wire passes near to any sions to maintain an electrical signal. Fiber
electronic device, it will pick up interfere- optics do not interact with magnetic fields,
nce from it. Heavy-duty electro-mechanical and do not create them to any detectable
devices, such as large electric motors, degree, so a fiber-optic cable could run
can easily generate enough interference to through the center of the largest motor in
swamp any real signal, and even enough to the world without picking up interference,
damage connected electronic equipment. and thousands of fibers can run together
Common alternatives, such as trans- for tens of miles without any signal cross-
mitting signals by radio or microwave, over. Fiber-optic cables are much lighter
suffer from similar problems. Radio and than metal wires, and it is much more dif-
microwave signals can not be insulated ficult to tap into them and intercept sig-
and can not be channeled along a ‘‘wire,” nals. Fiber optics are also used for short-
so multiple signals have to use signifi- distance communications, even between
cantly different wavelengths in order to components of the same computer system,
avoid complete interference; electrically because of their high bandwidth capac-
generated interference is still a problem ity and immunity to electrical noise. FDDI
for radio, and microwave signals need a (fiber-distributed data interface) is a well-
straight ‘‘line-of-sight” unobstructed path known implementation for LANs (local-area
to follow; and there may be significant envi- networks) that use fiber optics instead of
ronmental dangers associated with some metal wires, and allows systems on the
wireless signals (there is some controversy, same LAN to be over 100 miles apart.

143
File server

The disadvantages of fiber optics are that Summary of potentially negative issues
the cables are more expensive, and the Fiber-optic cable technology (the cable itself
equipment needed at either end of a cable and the light--electronic signal converters)
to convert between light and electronic sig- is more expensive than traditional metal
nals is much more expensive. Fiber optics cable technologies. Connecting two cables
can only carry signals; optical computing together is difficult. All devices connected
devices, although of great theoretical inter- to the cable need their own power supplies
est, do not yet exist. Connecting together because no useful power may be transmit-
two fiber-optic cables is very difficult. It ted through a fiber-optic cable.
is also impossible to transmit any useful
amount of power along a fiber-optic con- Reference
r D. Goff (2002). Fiber Optic Reference Guide,
nection, so all connected devices must have
their own power sources. 3rd edn. (Woburn, MA, Focal Press).
Associated terminology: LAN, Ethernet.
Business value proposition
Fiber-optic cables are advantageous for sev-
eral compelling reasons. The first is their File server
ability to carry a far greater bandwidth
Foundation concepts: Network, Client–server.
than is possible over any other medium.
Definition:Acomputerorgroupofcomputerssetaside
Second, they do not suffer from any electro-
to act as a data repository, storing files for many users,
magnetic interference (EMI) and hence can
and making them accessible over a network.
be incorporated into system designs where
metal cables would either be impractical
Overview
or require special and expensive treatment
When it was normal for organizations
(e.g., automotive and aeronautic engine-
to have one main computer, which was
management systems). Third, the cables
used by everyone who needed computer
permit data to be transmitted over large
resources, all files were kept on the same
distances without the need for signal boost-
system, and were, in principle at least,
ers. Fourth, the cables themselves are
accessible to all. Now that everybody has
much lighter than metal wire, and thus
their own computer, sharing files and
are favored in weight-sensitive applications
accessing one’s own files from a different
(e.g., in airliners).
location has become an important concern.
Although the price of the cable is higher
Portable removable media (such as flash
than that for metal-based cables, the versa-
memory cards) are a partial solution, but
tility and advantages of the technology will
require that the need for a particular file
increase its use and, as demand rises, costs
is anticipated. The various computers on
of supply will drop. Eventually fiber optics
a network could all be set up to run FTP
will probably become the de facto standard
(file transfer protocol) servers so that they
media for data transmission.
can be retrieved by the network from any-
where, but that adds to the administra-
Summary of positive issues tive load, and creates a significant security
Fiber-optic cables are light-weight, have a risk.
high-capacity bandwidth, are not subject to The other solution is to provide a File
EMI, and can carry signals for up to 100 server. This is one or more specially des-
miles without the need for retransmission ignated computers on a network that are
or a signal booster. responsible for storing files securely and

144
File server

making them available to authorized users. cult to find reliable implementations from
Normally, when a file server is in use, the a variety of sources. Using a file server can
users’ own computers cease to be the pri- also be an equally simple procedure. In pop-
mary storage site for their own files; every- ular systems, a Shared folder is used. This
thing of importance resides on the file appears as a perfectly normal folder on the
server, it is downloaded when work starts, desktop, which can be opened and accessed
worked on locally, then uploaded back to just like any other folder, but in fact repre-
the file server when finished. Many users sents an automated portal to a file server.
prefer to keep their own files on their own For example, adding a file to the shared
computers, perhaps wisely not trusting all folder is translated behind the scenes into
their eggs to one basket; they must sim- the correct sequence of network commands
ply remember to upload up-to-date copies to send that file to the server. Many users
of their files to the server at reasonable happily use a network file server for years
intervals. without ever realizing it. NFS (Network File
The use of a file server is a great aid to System, not to be confused with NSF, the
system maintenance: creating regular back- National Science Foundation), originally cre-
ups of all files is much easier if all the files ated by Sun in 1984, is one of the old-
are in one place. File servers can also pro- est and most popular of current systems.
vide financial gains: if all large files are kept Versions of it are often freely available
on a file server, then the individual work- for most versions of Unix and many other
station computers can have much smaller platforms.
and therefore cheaper disks than would
normally be possible. Some organizations Business value proposition
go all the way to using Diskless systems, in File servers are used in organizations for
which the individual workstations have no a variety of reasons, including central-
disks at all, relying on network access to a ized security, user management, applica-
file server for absolutely everything. tion license management, cost, backup pro-
There are some negative aspects to the cesses, efficiency, and integration. Having
use of file servers. If all files reside on all the files on one server allows that server
a network-accessible file server, then the to be at the center of all security efforts; all
security of the entire network and of requests for data can be examined equally
that server in particular becomes a major and verified, preventing localized security
concern. Network traffic will be greatly breaches and duplication of effort in secur-
increased if every file access requires com- ing many localized data storage facilities.
munication with a remote file server, so An extension of centralized security is the
higher-bandwidth network infrastructure centralized control of user access and the
may be required. Even the fastest of net- creation of user profiles (descriptions of
works can not provide the same access what resources each user is permitted to
speed as that available from a local disk access). This can help in the creation and
drive, so there may be an appreciable slow- enforcement of access policies and ensure
down in some applications. compliance with regulatory requirements
The term file server may be correctly such as Sarbanes--Oxley.
used to refer either to the computer that The use of file servers allows network
holds the files, or to the software applica- managers to ensure that the applications
tion that runs on that computer handling running on corporate systems are in accor-
the requests. File server software is tech- dance with the licenses held by the orga-
nologically fairly simple, so it is not diffi- nization. The centralization of licenses also

145
File system

allows managers to ensure that the licenses


File system
are being used optimally and effectively.
This is important because many appli- Foundation concept: Disk.
cation vendors charge customers on the Definition: The logical organization of portions of a
basis of the number of users accessing the diskorotherlong-termstoragedeviceneededtocreate
application or the number of processors data files.
used.
File systems allow centralized backup sys- Overview
tems to be efficient and effective, ensuring A standard disk drive contains hundreds
that the correct processes are performed. of thousands of millions of bytes of data,
The use of file servers allows network enough to store 100 000 copies of the
managers, CIOs, and chief security offi- Bible without compression. The only inher-
cers to manage the complete information- ent organization is that those bytes are
management lifecycle. grouped into hundreds of millions of Blocks,
File servers allow the creation of either each of exactly 512 bytes (enough for just
thin- or fat-client systems. This allows an one small paragraph of text). If users are
organization to design an efficient and inte- to be able to store a large number of data
grated system that reflects its own individ- files of various sizes, large and small, and
ual needs. be able to find them again, some additional
organization must be imposed upon a disk.
That is exactly what a file system does.
Summary of positive issues One of the major components of an oper-
File servers provide centralized data stor- ating system (such as Unix, Windows, DOS,
age and access capability. File servers allow etc.) is the built-in software that imple-
systems architects to develop a wide range ments the file system and gives users and
of system configurations and help network applications convenient and reliable access
mangers control costs, security levels, to it. An operating system can not really
users, licenses, backups, and the efficiency exist without a file system, since an oper-
of the system. Many freeware and open- ating system must reside on disk as a large
source file servers are available and provide collection of essential files. In general, each
a variety of functions. different operating system has its own kind
of file system, and a disk that was for-
matted for one system may be completely
Summary of potentially negative issues unreadable by another, although many sys-
File servers can be complex to manage and
tems do now provide some cross-system
require special training to ensure correct
support.
use. The problematic nature of ‘‘putting all
All file systems provide the same basic
your eggs in one basket” is a critical con-
functionality: creating and accessing files
cern. If all of a corporation’s data is stored
and folders (outside of Windows and
in one place, then that one place needs to
MacOs, folders are usually called Directories
be very carefully protected.
or Catalogs). However, different file systems
Reference do vary considerably in their provision of
r M. Tulloch (2003). Windows Server 2003 in the more advanced features which may be
a Nutshell (Sebastopol, CA, O’Reilly considered essential by any business ven-
Press). ture or serious user, such as control of
access by multiple users, and recovery after
Associated terminology: FTP (file-transfer disk ‘‘crashes.” The major file systems com-
protocol), Client--server, Network. monly in use are listed below.

146
File system

FAT (File Allocation Table), which has It is the file system of choice for
three varieties: FAT-12 (used only for Windows-XP.
floppy disks), FAT-16, and FAT-32. The Unix. All versions of Unix (Linux,
number 12, 16, or 32 in the name FreeBSD, Solaris, etc.) use minor vari-
refers to the number of bits used in ations on the same file-system design
the unique identification number of which, surprisingly, has no name
each file. For example, FAT-12 uses only (apart from UFS, which simply stands
12 bits, which can produce only 212 for Unix File System). Its most recog-
or 4096 different numbers, so, under nizable and namable feature is the I-
the FAT-12 system (and therefore on Node, a small data structure used to
any PC floppy disk), it is not possi- represent an individual file. I-Nodes are
ble to have more than 4096 files. The probably also the most troublesome
same logic imposes a limit of 65 536 feature, since they must be created
files on any FAT-16 system. This restric- in advance when a disk is formatted;
tion, together with the maximum of if the number of I-Nodes needed was
11 characters in a file name and 4 GB underestimated and they are all used
maximum disk size, has rendered FAT- up, it is impossible to create any new
16 obsolete. FAT-32, introduced in 1996, files. Unix file systems generally allow
relieves some of these limits, support- for very long file names; and always
ing disks of up to 2048 GB with indi- provide individual access permissions
vidual files of up to 4 GB and the now- on files and directories.
standard long file names. However, FAT ISO9660 is the file system used for
systems provide hardly any support data on compact disks (CDs). It has
for recovery after disk crashes, and two forms: Level-1, which allows only
even the unexpected loss of electrical 11-character file names; and Level-
power can result in enormous loss of 2, which allows up to 32-character
data. FAT systems also suffer from Clus- file names. Both permit folders to
tering, a technique that significantly be nested only eight deep. Joliet is
raises the maximum supportable disk a commonly used set of extensions
size at the expense of also raising the that greatly relieves these restrictions.
minimum possible file size. A normal Micro-UDF is similar in spirit to
FAT-32 system on a large disk will ISO9660, and is used for data DVDs.
require every file, even if it contains UDF (the Universal Disk Format) is
only a single byte of data, to occupy another newer extension to ISO9660,
at least 32 768 bytes of disk space. FAT- which is used only for optical disks
32 is generally available only under (CDs, DVDs, etc.)
Windows. HFS (Hierarchical File System) and HFS-
NTFS (the Windows-NT File System) is Plus were the file systems used by
a Microsoft product introduced for Apple Macintosh computers before
Windows-NT, and supported by all MacOS-X, which is based on Unix
modern Windows variants. It provides (the BSD version). They support long
faster access to data in large files, file names, file access permission set-
automatic encryption, logging of ting, and the more complex file
file accesses at various levels, indi- structures that Macintosh systems
vidualized file-access protection were famous for, but have other
settings, and good recovery after limitations similar to those of FAT
disk, computer, or software failures. systems.

147
Firewall

Business value proposition Overview


The systems administrator must under- A Firewall may be a separate hardware
stand the underlying structural limits, unit installed directly between a computer
strengths, and weaknesses of the file sys- and its internet connection, it may be an
tem in use as part of an organization’s sys- integral part of another item of network
tems architecture. A file system is an inte- hardware such as a switch or broadband
gral part of an operating system, and the modem, or it may simply be a piece of soft-
whole is a very sophisticated piece of soft- ware that works with the operating system
ware that continues to evolve; it is essential that has already been installed.
for administrators to keep abreast of the lat- Whatever form it takes, the job of a fire-
est technology applicable to their systems. wall is to monitor and police network activ-
Operating systems are increasingly ity. Not spying on the content of trans-
designed to provide robust file systems missions, and never recording them, but
that can survive hardware failures and rather making judgments based on the
operator errors, and can even be config- source, destination, and protocol of individ-
ured to provide warnings of impending ual messages.
problems (such as running out of storage If a particular computer is threatened
space, or problems with the disk hard- by frequent break-in attempts, and it does
ware). Unfortunately, system problems and not provide any public network services, a
instabilities tend to occur more frequently simple solution is to instruct the firewall
at the edges of a system’s performance to block all incoming connection attempts.
limits, and, when a system is operating Communications initiated from the inside
under extreme stress, failure can occur will be unaffected, but those initiated from
without any detectable precursors. It is outside (as most break-in attempts are) will
an operator’s responsibility to manage the never be seen.
system proactively. If a threatened system does provide some
public network services, perhaps acting as
References a web server or email server, the firewall
r A. Silberschatz, P. Galvin, and G. Gagne
can be instructed to block all ports except
(2004). Operating System Concepts (New those directly associated with web and
York, John Wiley and Sons). email services (i.e., ports 25, 80, 109, 110,
r B. Carrier (2005). File System Forensic
143, 993, and 995) from receiving connec-
Analysis (Boston, MA, Pearson). tion attempts. Of course, it is then the sys-
tem administrator’s duty to ensure that the
Associated terminology: Operating applications responding to those ports are
system, File server. secure and trustworthy.
Firewalls can also be used to prevent out-
going connections, so, for instance, chil-
Firewall dren or employees could be prevented from
browsing the web, accessing off-site email
Foundation concepts: Network, Virus, Spyware, Inter- servers, or playing with instant-messaging
net protocol. systems. If a system has been infected by a
Definition: A device that confidentially monitors all virus, it is possible that break-in-attempting
applications attempting to establish internet com- connections can be initiated from inside
munications, and stops those that meet predefined the system; some firewall products can
standards of suspicious behavior or that fail to satisfy be programmed to report unusual connec-
predefined standards of acceptability. tions, and ask the user ‘‘forbid or permit?”

148
Firewall

on a case-by-case basis for anything unex- ine network traffic and resource requests,
pected. assess the threat level, and take the appro-
The primary purpose of firewalls is, of priate action. The application of firewall
course, to restrict communications into technology is the responsibility of the net-
or out of a controlled area. If used care- work manager and the chief security officer
fully, they can prevent attacks from get- (CSO) (in a small organization the CSO may
ting in and secrets, corporate or personal, well be the network manager). An appro-
from getting out. Firewalls can be a heavy- priate firewall technology and vendor need
handed solution; employees often bemoan to be selected and put in place. In the
the fact that they can not use FTP and selection of a firewall, the network man-
therefore can’t download the latest version ager needs amongst other things to con-
of some application, but that is just the fire- sider the reliability of the device, its costs
wall doing its job especially well. It is more relative to performance, whether a hard-
often firewall-management policy that is at ware or software device is appropriate, the
fault rather than the firewalls themselves. scalability of the selected system, vendor or
Of course, a firewall is just another third-party support for the device, the his-
human construct, as fallible as any other. torical record of the vendor, and the total
If a firewall system itself has security vul- cost of ownership.
nerabilities through which it can be taken The implementation of a firewall device
over, then all is lost. Single-purpose fire- needs to take into account the sensitivity
wall hardware units are the least likely to requirements of the network to which it is
have such vulnerabilities, but are by no attached and the requirements of its users.
means to be considered immune. Software Should all traffic be prevented from enter-
firewalls are certainly the most vulnera- ing the system or just select traffic? Insen-
ble, since they are merely software appli- sitivity will allow too much traffic and the
cations themselves, and are just as likely potential for breaches in security; if the
to be replaced or modified by a success- sensitivity levels are too high users will
fully attacking virus. For any system hold- be unhappy because ‘‘normal” emails may
ing critical data, the selection of a firewall be rejected. These aspects of deployment
is a critical stage in setting up security, and need to be contained within a firewall-
requires diligent research: some providers management policy document.
have expensive products and a long history
of satisfied customers; some providers are Summary of positive issues
very cheap, but have never been heard of Firewall technology helps to prevent
before. A small saving at setup may have a unwanted data traffic from entering a
very large cost later. A vulnerable firewall network. The technology is mature, well
means an open computer system. understood, and supported by vendors, con-
sultants, and network specialists. Firewalls
Business value proposition can be used to prevent access to specific
The key to a secure IT operation is to ensure resources.
that unwelcome traffic stays outside the
walls of the organization. If there is any Summary of potentially negative issues
possible doubt about the security of a com- Firewalls are not immune from attack
puter system (and with closed-source sys- themselves, and software firewalls in many
tems how can there not be a doubt?), then instances may themselves be targets of hos-
the provision of a reliable firewall system tile intent. Poor firewall-management poli-
is central to safe operation. Firewalls exam- cies can result in restricted access to the

149
Flash memory

network both for internal and for external memory) that satisfies requirements (4) and
users. (5). It is a relatively new technology, and
advances toward the first three requireme-
Reference
r J. Wack, K. Cutler and J. Pole (2002). nts are being made, but there is still a long
way to go before flash memory can replace
Guidelines on Firewalls and Firewall Policy:
either RAM or disks in general-purpose
Recommendations of the National Institute
computers. At the time of writing the best
of Standards and Technology, US
statistics for commercially available units
Department of Commerce Special
are very approximately as follows:
Publication 800--41 (Gaithersburg, MD,
US Department of Commerce).
Criterion Units RAM Flash Disk
Associated terminology: Proxy, Security.
1
Cost per GB $100 $100 2 c/
Capacity GB 2 4 500
Speed ns per byte 0.5 50 250
Flash memory
Non-volatile -- No Yes Yes
Robust -- Yes Yes No
Foundation concepts: Storage, Memory, Disk.
Definition: A form of stable memory maintaining its
contents indefinitely without any power consump- (In this table, each technology is given
tion, and having the robustness of a solid-state device, its best possible circumstances. For exam-
with no moving parts. ple, flash-memory cards can attain trans-
fer rates of 20 MB per second, which gives
Overview 50 ns per byte, even though a single byte
In an ideal world, the memory used can not be accessed in 50 ns; similarly, disk
in a computer would have five essential drives can provide one byte in an average of
characteristics: it would be 250 ns only as part of a large transfer.)
(1) of low cost, Currently, flash memory provides a use-
(2) of high capacity, ful portable medium. Data may be trans-
(3) of high speed, ferred to it very quickly and conveniently,
(4) non-volatile (meaning that it should and it is very light and hard to break,
not forget its contents when the power so it may be transported with great ease.
supply is turned off), and Usually, flash memory is in the form of
(5) robust (both solid state, having no small cards or other portable devices. The
moving parts that will inevitably suffer cards must be connected to a computer
from wear and fail, and having no through a special USB device; the popular
delicate components). (and mutually incompatible) forms are Com-
pact flash (two different forms), Secure digi-
Naturally, the real world is far from perfect, tal, and Memory stick (two forms). Other flash
and no form of memory yet devised satis- memory devices have a built-in USB plug,
fies all five of the requirements. Tradition- so they need no extra support.
ally, the problem has been solved by build- There is one further disadvantage to cur-
ing computers with two forms of mem- rent flash memory technologies. Although
ory, one (usually called RAM) that satisfies they are completely solid state, the pro-
requirements (3) and (5), and the other (usu- cedure required to erase and write data
ally in the form of a disk drive) that satisfies is slightly destructive, and flash memory
requirements (1), (2), and (4). devices can not be expected to survive more
Flash memory is a form of EEPROM (electro- than one million cycles (fewer for many
nically erasable programmable read-only kinds). Reading data has no destructive

150
Formal methods

effect, only erasing and writing. This limi- flash memory stores considerably less data.
tation is of no significance to current uses, There is no built-in security associated with
but, if flash memory is ever to be used the devices.
to replace RAM, for which a million read--
write cycles occur in less than a second, this Reference
r P. Cappelletti, C. Golla, P. Olivo, and E.
will be another hurdle to be overcome.
Zanoni (1999). Flash Memories (Boston,
Business value proposition MA, Kluwer Academic).
Flash memory has many desirable prop-
erties for a memory device. It is capable
of holding a meaningful amount of data,
as opposed to the floppy disk which has Formal methods
become almost redundant due to its small
Foundation concept: Software development lifecycle.
storage capability, which forces users to
Definition: Mathematically rigorous software devel-
employ compression or other techniques
opment methodologies.
to save even modest files. Flash memory
has relatively fast data transfer rates, and
most useful of all is that it is portable and Overview
can employ the convenience of a USB con- Software development is sometimes viewed
nection. These characteristics make it ideal as a science, sometimes as an art, but
for the business traveler or individual who mostly is considered to lie in an ill-defined
wishes to carry their data with them on a variable position somewhere between the
trip. two. The scientific method is based on
Security concerns also need to be a continuous cycle of observation, the-
addressed by flash memory users; the orization, and experimentation. Observa-
devices have no inherent security built tion corresponds to investigating what is
into them and thus users need to ensure needed in new software, or observing the
that the data stored is encrypted as neces- results of tests of existing software. Theo-
sary (e.g., to meet HIPAA requirements). Of rization corresponds to working out what
course, the same security concerns apply those observations mean, and thinking
equally to other portable media. The small of a way to implement or improve the
size of the memory devices aids their porta- implementation of a solution. Experimen-
bility and can enhance the data security tation corresponds to running or testing
because users can wear the device on a cord the implementation, and seeing what hap-
around their necks and thus reduce the pens in that test brings the cycle back to
potential for data theft; a laptop is a much observation.
more obvious target for a thief than is a The artistic method is based on experi-
hidden flash memory device. ence and the intuitive expertise that comes
with it: the experience of having imple-
Summary of positive issues mented many software projects before,
Flash memory is convenient, portable, easy and knowing which techniques work well
to use, possesses a relatively high transmis- under which circumstances, and the aca-
sion rate, can hold more data than a floppy- demic experience of having learned the
disk device, and is solid state in nature. ‘‘right” way to do things and knowing how
to apply that knowledge. Neither method
Summary of potentially negative issues can be successful alone; successful software
The cost of flash memory is higher, per unit developers slowly discover their preferred
of capacity, than that of disk memory, and place on the art--science spectrum.

151
Formal methods

Neither method is particularly suited to matics might not be as practically useful as


getting programs right. The artistic method we had imagined.
is used because we don’t know any better; Modern mathematics is the result of
programming hasn’t really been ‘‘worked thousands of years of careful study, and is
out,” and never can be. There is no ‘‘correct” a tailor-made methodology for investigat-
method for producing a program to solve ing, developing, and manipulating knowl-
a problem, so humanistic techniques rely- edge in artificial, abstract, and completely
ing on flashes of inspiration, experience, known systems. In other words, mathemat-
and even superstition are inevitable. The ics is the ideal system for software develop-
scientific method sounds ideal, but it is ment. Formal methods are techniques that
designed for real scientific research, inves- allow all stages of software development
tigating the real natural world, where abso- to be represented in purely mathematical
lute knowledge is not possible, and the best terms, which can be manipulated to pro-
we can do is invent theories that explain all duce a program in ways that are guaran-
experimental results, and keep on testing teed to be correct.
those theories with new experiments. The first stage in formal software devel-
Programming is not part of the natu- opment is to produce a Formal specification
ral world. It is a totally artificial activity, of the problem to be solved. This means
and absolute knowledge is possible. Peo- that a mathematical model for the problem
ple invented computers; we know exactly domain is constructed, and logical rules
what they do under all circumstances. describing inputs to a problem and exactly
People invented programming languages; how a correct solution is related to them
we know exactly what effect every cor- are defined. All further work is based on
rect program will have under all circum- this model and these rules; they provide
stances. People write programs; so we the definition of correctness for the pro-
can work out exactly what they will do gram to be produced. This is a very diffi-
under all circumstances. There is nowhere cult stage, requiring fairly intensive train-
where experimentation can reveal any- ing in abstract algebra and mathematical
thing that we could not already have been logic. It is made more difficult by the fact
certain of; the scientific method is totally that the customer must agree that the for-
inappropriate. mal specification does indeed describe the
Mathematics is the language of artificial, problem that they want solved, and it is
abstract, completely known systems. Why is most unlikely that the customer will be
two plus two equal to four? Simply because able to understand the specification at all.
the laws of mathematics make it so. If one The next, and final, stage is to trans-
had a box with two apples in it, then added form the formal specification into a pro-
two more apples to the box, one would gram. The formal specification is by defi-
strongly expect to have a box with four nition correct; the only thing that stops it
apples in it. That is not why two plus two from being a solution in itself is the fact
is four; in reality it is not much more than that it will not naturally be in the form
a coincidence. We have a very strong belief of an efficiently computable algorithm. For
that the laws of mathematics do model the example, part of the specification might
way the world works, but mathematics is demand ‘‘the smallest positive number that
in no way dependent upon that belief. If it satisfies [some condition]”; that would be a
turned out that our box actually had five correct description of the result, but does
apples in it, that would not mean that two not say how to find that result. It must
plus two is not four, or that mathematics be transformed into an explicit and exe-
is wrong; it would only mean that mathe- cutable form such as ‘‘take each number 0,

152
Formal methods

1, 2, and so on without limit, testing each has been a lot of research into formal
in turn to see whether it satisfies [some methods, and some variably useful tools
condition]; when one is found, stop the have been produced to aid the process.
search because you’ve got the answer.” This The most promising of these are Z (pro-
describes the same result, but this time also nounced ‘‘zed”), VDM (the Vienna Devel-
says how to find it. opment Method), The Hoare Logic, the
The well-understood method of mathe- ACL2 Theorem Prover, the HOL Theorem
matical proof supports exactly that kind Prover, CSP (Communicating Sequential
of transformation. It repeatedly applies Processes), and Extended ML, although
small transformations that are known to there are many more. Some non-traditional
preserve correctness to an initial state- programming language paradigms, partic-
ment. If the initial statement was correct, ularly pure functional programming and
the result will also be correct. The for- logic programming, transform specifica-
mal specification is transformed into an tions that would normally seem to be
executable program by applying a care- completely abstract and unexecutable into
fully chosen sequence of simple changes directly executable programs.
that have already been proven to preserve
correctness. This is also a very difficult Business value proposition
stage. There are software development tools The need to create programs quickly and
that will apply the chosen transformations correctly is at the very center of computing.
to the guaranteed correct specifications, The promise of perfect programs, flawlessly
removing the drudgery and the possibil- working and optimized, first time and
ity of transcription errors, but at each step every time, is the ideal sought by program-
there will be an enormous number of possi- mers, IT-project managers, and everyone
ble transformations that could be applied, connected to an IT organization, includ-
and it is absolutely impossible for a soft- ing the end users. However, this utopia
ware tool to know which is the ‘‘right” is rarely, if ever, reached by anyone, even
one. those organizations certified as Level 5 on
Even formal methods rely completely on the Software Capability Maturity Model.
human intelligence. The difference is that Typically software is developed in a very
any program produced by formal meth- ad hoc manner. For example, some code
ods is absolutely guaranteed to be correct. may be taken from a previous project,
What is not guaranteed is that any effec- edited, tested, and then added to. The pro-
tively executable program will be produced cesses may follow a rough informal pro-
at all. Heuristic searches and other tech- cess model, e.g., a waterfall model, or a
niques of artificial intelligence can help to spiral model, or be termed a ‘‘prototyping”
automate the search for the right transfor- process model and couched in JAD termi-
mations, but it is logically impossible for nology. Fundamentally, many professional
there to be any method that will always and trained computer scientists acknowl-
find the answer. Even human programmers edge that, if houses, bridges, and high-rise
can’t guarantee to be able to program a buildings were constructed to the same
solution to every problem they might be standards as the majority of software, they
given. would either fall down during construction
The potential gains from making for- or be condemned as unsafe and unfit for
mal methods easier to use are enormous. occupation.
The ability to produce guaranteed per- The rise of formal methods as a disci-
fectly correct programs just sometimes pline grew from the recognition that the
would be worth a lot. As a result there profession needs to move away from the

153
Formal methods

position where anyone can claim to be a that professional engineers, chartered sur-
programmer and proceed to sell bad code veyors, and others do, with a requirement
to the public, toward becoming an engi- for continuous training and acknowledge-
neering profession. Formal methods aim to ment of liability for errors, the current ad
provide the professional programmer with hoc situation of random development will
a set of tools and methodologies that, when continue.
used correctly, will result in systems that
match their specifications, as agreed upon Summary of positive issues
by the software engineer, the user, and, Formal methods are the subject of much
when appropriate, the corporate manage- current research and great academic
ment. interest; successful use of formal methods
The use of formal methods in IT organi- would provide great benefits to an orga-
zations is very limited, and there are sev- nization. The Software Capability Maturity
eral reasons for this. The first is that the Model is available to measure an orga-
training to become a true software engi- nization’s progress from the informal to
neer is long, technically challenging, and the repeatable formal style of develop-
requires intellectual rigor. Second, poten- ment. A wide range of formal methods
tial software engineers are put off from exists, including functional programming
studying this topic because the demand and proof systems. Formal methods enable
from industry is low and the rewards con- the user and the software engineer to agree
sequently limited. Third, companies do not on a specification of the product that will
recruit software engineers because many be delivered. Formal methods permit proof
IT organizations do not understand what that the specification and the delivered
true software engineering offers them. A code are the same.
mistaken view is that adhering to a pro-
cess model such as MIL-STD-2167A, using Summary of potentially negative issues
the waterfall method, or using RAD meth- Formal methods are very much an evolv-
ods equates to software engineering. This ing technology. There is a large academic
is simply wrong, weak development meth- literature base, but few usable software
ods that do not encapsulate mathematical development tools that truly adhere to
design principles are not formal methods formal methods. Formal methods require
of software engineering. Fourth, IT organi- intense specialist training. Formal specifica-
zations are typically not willing to change tions are difficult for users and customers
their development model to this higher to read and hence be able to comment
standard, since it demands a higher caliber upon or give informed assent to. Adoption
of human resources, is perceived as taking of formal methods requires different devel-
longer to develop code, and is unintelligible opment processes from those of non-formal
to non-technically trained developers, ven- styles of development.
dors, and users.
Some organizations have attempted to References
use formal methods, with varying degrees r W. Gibbs (1994). ‘‘Software’s chronic
of success. Formal methods are particularly crisis,” Scientific American, September.
suitable to problems for which failure may r A. Harry (1996). Formal Methods Fact File,
threaten human life, e.g., in the space pro- VDM and Z (New York, John Wiley and
gram, the nuclear industry, and airlines. Sons).
However, until the computer industry as r C. Jones (1990). Systematic Software
a whole demands that its workers possess Development Using VDM (Englewood
true professional qualifications in the way Cliffs, NJ, Prentice-Hall).

154
Fortran

tations of modern compiled languages will


Fortran
include these optimizations and more. Tra-
Foundation concepts: Programming language, ditional Fortran is also very easy for an
Compiler. experienced programmer to learn, so there
Definition: A primitive programming language. is little risk of this stockpile of legacy code
becoming unusable. Fortran is still used by
older scientists who learned to program in
Overview Fortran and nothing else, and are unwill-
Fortran was one of the first programming ing to learn a newer language when their
languages ever created, and is easily the old favorite still suits their purposes as well
record holder for longevity, being still in as ever.
use today. The original design and imple- Over the years, Fortran has evolved
mentation was begun by John Backus of through a number of standards. The one
IBM in 1954, but it was many years before known as Fortran IV, defined in 1961, is
the language was usable, and years more true to the original vision of the language,
before it became available to the computer- whilst fixing all of the ‘‘teething troubles”
using public. Compilers were unknown at of the language, and is generally consid-
the time, so the implementation of Fortran ered to be the Real Fortran. WatFor was a
required the invention of a whole new soft- famous implementation of Fortran IV from
ware technology, and gave birth to a new Waterloo University in Ontario, which was
area of research. The name ‘‘Fortran” is a very popular because it gave comprehensive
contraction of ‘‘Formula Translator”; in the and meaningful diagnostic error messages.
1950s programming was mostly a matter Fortran 66, Fortran 77, and Fortran 90 were
of translating mathematical formulæ into later major revisions (each named after the
a form that could be directly executed by year of their design), the last two espe-
computer, so the name is apt. cially trying to bring Fortran into the mod-
To call Fortran a very primitive program- ern world by introducing the possibility
ming language is a true statement, but not of structured program design and object
a fair criticism. When it was designed, it orientation. However, these changes went
was far in advance of anything else. Pro- against the fundamental nature of the lan-
gramming was in its infancy, and the tech- guage, and many programmers took the
niques that are now well known as the reasonable attitude that it would be just as
basis of successful programming had not easy to learn a fully modern programming
yet been thought of. Fortran would be emi- language as it would be to learn the new
nently unsuitable for any new software versions.
development -- it does not even support Until recently, the correct rendering was
structured programming -- but it served its ‘‘FORTRAN,” but now, especially since the
purpose admirably. Fortran-90 standard, it is treated as a nor-
There is still a considerable amount of mal proper noun, and written ‘‘Fortran.”
legacy code, programs that were written
long ago in Fortran, and for one reason Business value proposition
or another never got recoded in a newer The use of Fortran programming within
language. For a long time, Fortran was organizations has decreased significantly
the most efficient programming language since the 1980s as new generations of
for scientific calculations. It incorporated programmers coming out of universities
many optimizations that were not consid- and colleges have been taught through
ered suitable for a general-purpose pro- newer programming languages. Fortran
gramming language, but good implemen- was widely taught in the 1960s and 1970s

155
Fourth generation

as an engineering programming language programming and offer programmers bet-


and thus many professional engineers, sci- ter access to operating systems functions.
entists, and academics of that generation Owing to their essentially unstructured
consider this language to be their preferred form, traditional Fortran programs are
medium for programming solutions based very difficult to debug and maintain. For-
upon mathematical calculations. The pop- tran can be difficult for organizations to
ularity of this language during the 1970s support. Many legacy Fortran programs
and 1980s also grew upon the availability exist.
of library programs and packages provid-
References
ing easy access to statistical and numeri- r R. Barlow and A. Barnett (1998).
cal analysis routines, numerical control sys-
Computing for Scientists: Principles of
tems, and computer-aided-design systems.
Programming with Fortran 90 and C++
Subsequently many of these libraries or
(New York, John Wiley and Sons).
routines have been upgraded and may now r D. McCracken (1972). Guide to Fortran IV
be used with more modern languages. As
Programming (New York, John Wiley and
a consequence many of the programmers
Sons).
who run Fortran programs are running sys-
tems that are weakly supported at best and Associated terminology: C and C++, Java,
with legacy versions of the packages. While Cobol.
the statistical abilities of the old libraries
and packages may be correct, it is increas-
ingly becoming more difficult to connect Fourth generation
and embed these systems to other mod-
ern systems. Similarly, IT organizations will Foundation concept: Hardware.
find it more and more resource-intensive Definition:Firstgenerationcomputerswerebuiltwith
to support the few Fortran programmers vacuum-tube technology, second generation with
who wish to continue developing scientific transistors, third generation with integrated circuits,
solutions in that manner. and fourth generation with large-scale integration.

Summary of positive issues Overview


Fortran is easy to program, but hard to The terms first generation, second gener-
debug. It has been in existence for over 40 ation, etc. define the technological basis
years and there is a very substantial litera- upon which a computer system is built. The
ture behind the language. Fortran formed first automatic computing devices, starting
the basis for several generations of pro- with Babbage’s Difference Engine in 1822,
grammers to learn to program scientific and culminating with Turing’s Bombes
applications, and many software libraries in 1940, were mechanical or electro-
and packages were written to leverage For- mechanical in nature -- they contained
tran programs. Fortran can still be useful at least some physical moving parts --
for small, quickly written programs that and were not programmable in the mod-
have a short lifespan and are produced by ern sense, and can not really be considered
experienced programmers who do not wish computers as the term is normally under-
to learn a new programming language. stood.
Moving parts create major problems.
Summary of potentially negative issues They are inherently slow and unreliable:
Fortran has been superseded by newer pro- overcoming friction requires considerable
gramming languages that facilitate struc- power consumption, and it is physically
tured programming and object-orientated impossible for moving parts not to be

156
Fourth generation

subject to continuous wear, which gradu- transistor is a very robust (compared with a
ally destroys the device. vacuum tube, almost indestructible), small,
Fast reliable computing could only be cheap, low-power device built from semi-
built on a technology that allows machines conductor materials. Electricity can flow
without moving parts to be built. This through semiconductors just as electrons
seems to be a contradiction in terms, can flow through a vacuum, and applying
but the only thing that moves in a CPU a voltage correctly to a third connection
is electricity: electrons flowing through a can disrupt or enhance a current flowing
crystalline lattice of atoms. Other parts of a between two primary connections, allow-
computer (disk, keyboard, mouse, CD drive) ing one signal to control another just as
still have mechanical components, and are with a triode. Although the physics is very
the only parts of a modern computer (apart different, a transistor is conceptually simi-
from software) that are ever likely to fail. lar to a triode. Second generation comput-
The ability to use one electrical signal to ers used transistors in their logic circuitry
control or switch another electrical compo- where the first used vacuum tubes. This
nent without any mechanical intervention greatly improved reliability, price, and com-
was realized with the invention of the tri- puting power. A first generation computer
ode in 1906. Triodes and related devices, with perhaps 1000 vacuum tubes might
known collectively as vacuum tubes in be expected to run for a couple of hours
the United States and valves in the UK, before a tube burns out and has to be repla-
are extremely delicate and expensive. They ced. Significantly increasing the number
consist of a fragile glass tube containing, of tubes would significantly decrease the
in a vacuum, a small heating element mean time between failures. A second gene-
that heats an electrode (the cathode) to ration computer with many more transis-
around 4000 ◦ F. Electrons are boiled from tors would usually operate reliably until
the surface, and fly through the vacuum to some auxiliary mechanical component
another electrode (the anode), completing failed. The second generation lasted until
an electrical circuit. A voltage applied to a about 1970.
third mesh electrode (the grid) interposed A transistor is essentially a tiny piece of
between the anode and cathode suppresses silicon subtly modified (‘‘doped”) in strate-
the flow of current, and thus allows one gic places to make different kinds of semi-
purely electrical signal to control another. conductor material. It is possible to take a
With their thin glass, hard vacuum, and larger piece of silicon, make a more com-
high temperatures, vacuum tubes have a plex pattern of modifications to it, and con-
very short operational lifetime, but they struct multiple transistors on a single solid
can operate many thousands of times faster component. Individual transistors have to
than any mechanical device. They were in be individually mounted and soldered onto
common use for decades, but it was not circuit boards along with wire connectors
until the 1930s that anyone realized that and other parts to make a useful electronic
they could be applied to digital signals component. An integrated circuit is a single
to perform numerical computations. This chip of silicon with many transistors and
was the first generation: electronic com- connectors etched onto it. One small sili-
puters with logic circuitry built from vac- con chip can replace a whole circuit board
uum tubes. The first generation of comput- and be manufactured in large quantities
ers started with Colossus in 1943 and ENIAC completely automatically. Third-generation
in 1946, and continued until the late 1950s. computers use integrated circuits for all
Vacuum tubes made computers possi- of their major components. This makes
ble, but transistors made them practical. A another significant reduction in size, power

157
FTP (file transfer protocol)

A triode, an equivalent transistor, and two integrated circuits, one SSI, containing about 25 transistors,
and one VLSI, containing about 25 000 000 transistors. The scale on the left is in inches.

consumption, cost, and fallibility. The sec-


in the context of a museum. The latest gen-
ond generation made computers practical;
eration simply represents the latest tech-
the third made them commonplace.
nology and quickly renders previous gen-
The fourth generation was a less sig-
erations obsolete. Of course, it might not
nificant step than any of the first three,
be wise to jump to the next generation as
and was really just a continuation of the
soon as it is announced (or claimed), but,
third. Fourth generation computers apply
once new technology has become estab-
the idea of putting whole circuits on a chip
lished and is known to be reliable, the
to put whole CPUs on a single chip. The
old effectively ceases to exist, leaving busi-
fourth generation began between 1971 and
nesses with no options to worry about.
1972 when Intel introduced the 4004 and
8008, the first microprocessors. Although
great improvements to integrated circuits
have been made, the same basic technolog- FTP (file transfer protocol)
ical idea is still in use, and we are really still
in the third generation. The integrated cir- Foundation concepts: Network, File.
cuits characteristic of the third generation Definition: A network data-transfer application.
are referred to as SSI (small-scale integra-
tion), and those of the fourth as LSI (large- Overview
scale integration) and VLSI (very-large-scale FTP is a simple but effective and fairly effi-
integration). cient method for transferring files of any
Generations beyond the fourth do not kind between networked computers. It is a
really exist, and the term is used by various typical client--server system.
authors to refer to hardware based on arti- Although FTP is almost universal, and
ficial intelligence, neural networks, quan- is very simple to use, it does require
tum computing, or just about anything. some foresight. One of the two comput-
ers involved in the data transfer, either the
receiver or the transmitter, must have an
Business value proposition FTP server installed on it. FTP clients are
There is no possible advantage to using a ubiquitous -- every operating system pro-
first or second generation computer except vides at least a basic one -- and they require

158
Functional programming

no special setup, but FTP servers are not availability on computers makes this app-
always provided for free, and need to be roach useful for data transfer anywhere a
installed and configured correctly. Installa- network connection is available.
tion and configuration is just a few min-
utes’ work for the right personnel, but must Summary of positive issues
be done correctly. A mis-configured FTP FTP is a cheap, efficient, and widely avail-
server can leave every piece of data on the able method for data transfer.
computer open to the entire internet.
In normal use, a user wishing to transfer Summary of potentially negative issues
data will start up the FTP client software It requires care in configuration. Security
and tell it to connect to the desired server issues in several areas need to be addressed,
system. The user provides a user-name and including the use of ports used to commu-
password, and, if they are accepted by the nicate through, the firewall configuration,
server, files and folders may be transferred and the data files that are allowed to be
in either direction, often with a single click accessed.
or drag-and-drop operation. Server admin-
istrators may control exactly which files References
r W. Stevens (1994). TCP/IP Illustrated (New
and folders any given user or group of
users may access, and what kind of accesses York, Addison-Wesley).
r http://www.ietf.org/rfc/rfc0959.txt?
(read, modify, delete) they may make.
FTP also has an anonymous mode, in number=959.
which certain files are made accessible to Associated terminology: File server,
all, without the need for a user-name or Server--client, Electronic data interchange,
password. This provides a very convenient X.12.
means for making files publicly available,
but is also a possible security flaw: acci-
dentally allowing anonymous FTP circum- Functional programming
vents file protections. Most web browsers
understand FTP, and provide a very intu- Foundation concepts: Programming language, For-
itive interface for downloading files. For mal methods.
large amounts of data, FTP is superior to Definition: A style of programming in which compu-
HTTP, the default web protocol. tations are expressed as pure mathematical functions
FTP can be problematical for firewalls, that can be directly executed.
since every FTP transfer requires the use of
an arbitrarily selected internet port, so sim- Overview
ply leaving ‘‘the FTP port” unblocked is not Functional programming is a widely mis-
enough. Modern firewalls are constructed understood concept. The description is
with complete knowledge of FTP and other simple -- pure mathematical functions are
commonly used protocols, and are usually used to express computations (programs) --
able to handle the situation, albeit with but the true significance of ‘‘pure math-
some complexity. ematical” is often missed. A function in
the mathematical sense is not what most
Business value proposition programmers think of as a function. A
FTP is a simple, efficient mechanism for mathematical function is a fixed mapping
data transfer between two computers con- from each of a set of possible inputs to a
nected over a network. The approach allows corresponding output. The same function
large amounts of data to be transferred applied to the same inputs must always
easily and effectively. FTP’s almost universal deliver the same results, regardless of

159
Functional programming

circumstances. In fact, the idea of ‘‘circum- major areas of computer science research.
stances” does not mean anything in this Popular implementations that either are
setting; mathematics does not vary accord- purely functional or at least support mathe-
ing to context. matically functional programming include
This very basic notion of requiring that LispKit, KRC (Kent Recursive Calculator),
the same function applied to the same Miranda, Haskell, AFL (A Functional Lan-
inputs always produces the same results guage), FP, Hope, and Clean, but there are
is essential to all of mathematical reason- many more. A theoretical construct known
ing. The standards of proof simply do not as Monads is a topic of current research,
work if you can write the same thing twice attempting to make the benefits of func-
but have it represent a different value each tional programming available to more tra-
time. If this absolute constancy is applied ditionally oriented programmers.
to the design of programming languages, it
becomes possible to produce perfect proofs
Business value proposition
of a program’s correctness. New techniques
Functional programming has a vast poten-
of program development become possible,
tial for the development of mathematically
in which a statement of the problem to be
correct programs. Functional programming
solved is simply transformed into a solution
constructs also help programmers to cre-
to that problem.
ate solutions in such a way that they are
The difficulty is that standard program-
not constrained by arcane rules of proce-
ming languages do not behave at all like
dural programming. Unfortunately, many
this. The vast majority of programmers
programmers are unaware of this branch of
would consider it impossible to write a pro-
computer science and hence the use of this
gram to perform any useful task if it is
style of programming in the ‘‘real world”
not allowed to change anything. True func-
has been limited.
tional programming does not permit vari-
It should be noted that the amount of
ables, loops, ‘‘print statements,” or any of
‘‘lines of code” required to perform an oper-
a multitude of features normally thought
ation in a functional language is frequently
of as essential to programming. In order
very small compared with that for a pro-
to take advantage of the vast benefits that
cedural language, and the functions them-
functional programming has to offer, pro-
selves can also be easy to read with some
grammers have to almost relearn their
training. The style of programming has the
trade right from the beginning. That is not
potential for providing cleaner, smaller (in
an easy, or a cheap, thing to do.
terms of amount of code), partially self-
There are some popular programming
documenting programs, which reduces the
languages that are often called functional,
overhead on the developer. However, when
when they really are not. Lisp, the work-
programmer productivity is measured in
horse of artificial intelligence, is very fre-
terms of lines of code delivered, a very dis-
quently thought of as a functional lan-
torted picture is given.
guage because in Lisp everything looks like
a function. Appearances are irrelevant; Lisp
is not a mathematically functional lan- Summary of positive issues
guage, and hence does not provide the ben- Functional programming is mathemati-
efits of one. That is not to say that it is a cally pure and is open to undisputable
bad language; Lisp is very useful, it is just proof. A significant academic literature has
not functional. developed on the topic. Many free func-
Many real functional languages are avail- tional programming languages are avail-
able, and it has occasionally been one of the able. The programs themselves can be

160
Fuzzy logic

transformed and reasoned about mathe- does not mean that it is probably tall. It is
matically. because the idea of tallness is not precisely
defined, it is a fuzzy concept. The 0.85 rep-
Summary of potentially negative issues resents the judgment that it is quite tall.
Functional programming needs an aware- By most standards the Empire State Build-
ness and knowledge of pure mathematics. ing is tall, as when compared with people,
The approach is not widely understood or other skyscrapers, or red ants, but it is not
known amongst programmers. The inter- tall by all standards: when compared with
preters are often relatively slow at execut- mountains or thunderclouds, it is distinctly
ing the code. short.
The use of fuzzy logic allows automatic
References decision-making processes to progress
r S. Peyton-Jones (1997). Implementing
when an exact algorithm can not be
Functional Programming (Englewood found, and it is generally considered to be
Cliffs, NJ, Prentice-Hall). a technique of artificial intelligence and
r J. Darlington (1982). Functional
expert systems working in the domain
Programming and Its Applications of human descriptions. For example, a
(Cambridge, Cambridge University search for ‘‘big orange cats” would be hard
Press). to code with an exact algorithm because
r R. Bird (1998). Introduction to Functional
both ‘‘big” and ‘‘orange” are fuzzy human-
Programming (Boston, MA, Pearson). oriented descriptions. It would be pointless
r P. Hudak (2000). The Haskell School of
to ask the user for further details, exactly
Expression (Cambridge, Cambridge specifying how big in terms of inches or
University Press). pounds, and it would be absurd to attempt
to quantify how orange the cat must be.
Associated terminology: Logic
A fuzzy logic system would allow ‘‘big” to
programming, Formal methods, Software
be determined as a fuzzy combination of
development.
‘‘long” and ‘‘heavy,” accepting candidates
that are quite long and quite heavy,
together with those that are very long but
Fuzzy logic only moderately heavy, and those that are
very heavy but only moderately long, thus
Foundation concept: Knowledge-based systems. encapsulating all reasonable meanings of
Definition: A system of logical reasoning that handles ‘‘big.” It could then similarly filter all can-
imprecise knowledge or information. didate cats, having reduced requirements
for the degree of orangeness for those
Overview that are exceptionally big, and reduced
Fuzzy logic is a reasoning system that allows requirements for the degree of bigness for
truth values other than ‘‘yes” and ‘‘no.” The those that are startlingly orange.
validity of a statement is expressed as a Fuzzy logic is very difficult to work with
number somewhere between 0.0 (for com- effectively. The assignment of levels of tall-
pletely false) and 1.0 (for completely true). ness, bigness, and orangeness on the basis
For example, the truth value of ‘‘two plus of measurable quantities is entirely arbi-
two is four” would probably be exactly 1.0, trary and a matter for the system designer
whereas the truth value of ‘‘the Empire to decide. This makes success almost impos-
State Building is tall” might be 0.85. This sible because all users will have different
is not because we are uncertain about the ideas of how these terms should be defined.
height of the Empire State Building, and it Then there is the problem of how to

161
Fuzzy logic

combine fuzzy values: if a cat is big to the through which ‘‘fuzzy” problems could be
tune of 0.74, and orange to the tune of 0.9, considered. While probability theory can be
how should those numbers be combined to applied to a variety of problems, it is not
give a single measure of how ‘‘big orange” suited to all, and it requires that the prob-
it is? Again, there are no clear answers; it is lem solver has sufficient historical data to
for the system designer to decide, and for make a determination. Fuzzy logic allows
the users to inevitably disagree. software engineers to create programs that
For these reasons, fuzzy logic has not work for a variety of situations based upon
been a very commonly used tool. At first imprecise knowledge of inputs, the state of
sight it seems to be an obvious and good the system, and even the desired results.
idea, but it is very resistant to encoding in Fuzzy logic technologies have been incor-
the precise language of computer logic. porated into a wide variety of processes
Fuzzy logic was introduced by Lofti and products, including those found in
Zadeh of the University of California in automobiles, e.g., cruise-control systems,
Berkeley in 1965. He and others have devel- traction-control systems, anti-lock systems,
oped a properly mathematically rigorous and cornering systems; household goods
basis for fuzzy logic, which is known as such as toasters, microwave ovens, and
possibility theory. The notation and con- refrigerators; and entertainment products
cepts of possibility theory can easily be such as video cameras for automatic expo-
confused with probability theory, but it sure and focus control.
is a completely different thing. Probabil-
ity theory is entirely uncontroversial and Summary of positive issues
admits no imprecision, only uncertainty. Fuzzy logic allows programmers to use
Fuzzy logic and standard probability the- truth systems that are non-binary in
ory may be applied to similar computa- nature. It allows determination of out-
tional situations, but have very different comes when probability theory is not appli-
semantics. Saying that the probability of cable. Fuzzy logic is a well-researched aca-
X being tall is 0.85 means that we don’t demic discipline and can be applied to a
really know whether X is tall or not; it wide variety of problems and processes.
probably is, but we can’t be sure. Saying
that the fuzzy measure of X’s tallness is
0.85 means that we are absolutely certain
Summary of potentially negative issues
Fuzzy logic mathematics can become com-
that it is quite tall. Probability deals with
plex and it requires some special training.
uncertainty over well-defined attributes;
Fuzzy logic is difficult to work with effec-
fuzzy logic deals with certainties about ill-
tively. The combination of fuzzy values in
defined attributes. Exactly how connectives
logic is difficult to determine in many sit-
like ‘‘and” and ‘‘or” combine probability val-
uations.
ues is well known; how they combine fuzzy
logic values is inherently unknowable. Reference
r G. Klir and B. Yuan (1995). Fuzzy Sets and
Fuzzy Logic (Boston, MA, Pearson).
Business value proposition
Fuzzy logic was originally developed in Associated terminology: Artificial
the 1960s by Lofti Zadeh as a mechanism intelligence.

162
Global positioning system

mine orientation by comparing their rela-


Global positioning system
tive positions, but would need to be some
Definition: Also known as NavStar GPS, an electronic tens of feet apart to do so.
device that may rapidly and accurately determine its Because GPS can be (and is) used to guide
own geographical location by receiving signals from a weaponry, the US military has retained the
network of orbiting satellites. ability to deliberately degrade the trans-
mitted signal (make it less accurate), or
Overview turn it off completely, during emergen-
There are a few dozen special-purpose satel- cies. Naturally, military receivers are able
lites, launched and maintained by the to overcome the degradation, and give
US military, in constant Earth orbit. They positions much more accurately than do
contain extremely accurate clocks, pro- models available to civilians. Even during
grammed knowledge of their own orbits normal operation, military receivers pro-
(which is constantly corrected by signals duce a much more accurate fix than do
from ground stations), and high-frequency civilian receivers.
radio transmitters. These satellites trans-
mit a constant stream of data, which Business value proposition
describes the exact time and their position, The United States-based NavStar GPS and
and can easily be picked up by terrestrial the European Galileo GPS, which is sched-
receivers. uled to be available from 2008, are techno-
A GPS receiver listens to the stream of logies used by civilian organizations in a
data from a number of these satellites, and variety of ways. Typical uses include auto-
from it, using relatively simple geometry, mobile and maritime guidance systems,
can calculate its own exact position (lon- vehicle-tracking systems (some United
gitude, latitude, and elevation) to an accu- States-based car-rental companies prevent
racy of just a few feet. These receivers can their vehicles from being restarted when
be made very small (around one square driven into Mexico), surveying systems,
inch) and moderately cheaply (tens of dol- and automobile safety systems (crash
lars). Their primary use is as an aid to navi- locations can be automatically reported
gation (in aeroplanes, in smart guided mis- to emergency services through GPS and
siles, and in hand units for use by people), transponder technologies).
but they may also be embedded into larger
systems. By comparing a sequence of posi-
tions, a GPS unit may also determine its Summary of positive issues
own speed and direction of movement, and GPS systems are inexpensive, accurate, and
the relevant portion of a digitized map may can locate a position globally. From 2008
be automatically displayed. two versions of the system will be avail-
Coverage is global, but does require an able.
almost unobstructed ‘‘view” of the sky. Sig-
nals do not penetrate anything but the Summary of potentially negative issues
flimsiest of structures, and can even be The GPS could be downgraded in times of
obscured by trees overhead. GPS receivers political tension. As a satellite-based system
built into vehicles need to be in a rela- it is subject to solar radiation and other
tively open position. A single GPS receiver solar threats. The GPS works only outdoors
can determine position, but not orienta- where there is a clear ‘‘line of sight” with
tion (i.e., it can not act as a compass); the satellite; even tunnels and trees affect
two receivers working together can deter- coverage.

163
Groupware

Reference CAD for an engine component), but also


r A. El-Rabbany (2002). Introduction to GPS: needs to provide project management sup-
The Global Positioning System (New York, port in the form of email or instant messag-
Artech House). ing, calendar management functions, and
so on.
Another popular form of groupware is
Groupware multi-user computer gaming, in which
users share a virtual memory space to expe-
Foundation concept: Decision support system. rience the synchronous interaction asso-
Definition: Groupware is a software platform that ciated with the game. Although this is
enables a team to undertake a joint task and provides an entirely recreational use, the concepts
support. of multi-user interactive games have been
transported into the business world in the
Overview form of virtual meetings.
Groupware provides a software platform
that allows groups to collaborate on a Business value proposition
task or project. It is sometimes termed Groupware has the potential to help orga-
Computer-supported cooperative work (CSCW) nizations utilize their workforce in a more
or Computer-mediated communications. Group- productive manner in that team mem-
ware is a mechanism that Johansen cate- bers may be at different locations and
gorized as being composed of two param- work in shifts, but can still work together
eters. The first is location: are the team on a project. Groupware enables team
members located in the same place or members to work in asynchronous mode,
not? The second is temporal: are the team which, when supported by knowledge-
members working on a task simultaneously management software, can produce high
or not? For example, video conferencing degrees of efficiency.
is a simultaneous process wherein mem-
bers participating are at different locations, Summary of positive issues
whereas text messaging via cellular phones Groupware is available and supported by
is asynchronous for participants at differ- consultants and vendors, and can be used
ent locations. to promote greater flexibility within the
Groupware may be used to support syn- organization’s structure. Groupware can be
chronous tasks, for example the production used to monitor activities of team members
of an engine component for an automobile; and track human-resource expenditure.
in such a situation a team may be located
in a single central design studio with Summary of potentially negative issues
each team member working on a separate Companies have been hesitant to imple-
task. One team member could be work- ment groupware since it requires organi-
ing on the interface between the compo- zational and process changes. When group
nents, while another team member could members operate remotely through group-
be developing the packaging. In this sit- ware, their activities may need human
uation product-versioning and document- monitoring to ensure that performance
management software is also a useful com- levels are maintained.
ponent to support the team and task
development. References
Groupware not only supports the process r T. Hall (2006). ‘‘Intelligence community
task (such as the virtual modeling of the collaboration” (Washington, DC,

164
Groupware

Solutions Office for Advanced r R. Johansen, J. Charles, R. Mittman, and


Analytical Tools, Central Intelligence P. Saffo (1988). Groupware: Computer
Agency), available online only, from Support for Business Teams (New York,
http://collaboration.mitre.org/prail/ Free Press).
IC Collaboration Baseline Study
Associated terminology: Knowledge
Final Report/toc.htm.
management.

165
Hacker

already exist in the installation’s software


Hacker
infrastructure (operating system and appli-
Foundation concept: Security. cations) that were included by careless
Definition: developers.
On the basis that ‘‘it takes a thief to catch
1. Anundisciplined,unprofessionalprogrammer. a thief,” ex-hackers (or supposedly ‘‘ex,” at
2. Someone who attempts to break into other least) can be valuable members of a security
computer systems. team. Who would be better able to guard
against break-ins than the very person who
Overview had previously devoted his life to perform-
1. Originally, a hacker was a programmer ing them?
who worked without any clear plan, but
would just type something in and ‘‘hack it Business value proposition
around” until it seemed to work. A hacker There is very little business value asso-
would often be very well informed about ciated with the first definition of the
the particular programming language and term, because systems that were gener-
the technical details of the operating sys- ated by hackers do not typically conform
tem they were using, and would know a to corporate programming methodologies
wide variety of surprising tricks, but would or standards. This makes them difficult to
usually not be very well educated in the sci- read, interpret, and alter; and efforts to
ence and art of programming. do so expend large amounts of resources.
However, hackers always considered These systems are typically problematic
themselves to be the ultimate experts, and legacy in nature, although companies
and their facility with interesting tricks that still employ programmers with this
and their propensity for constructing soft- type of mentality toward programming will
ware incomprehensible to others tended continue to generate poor-quality systems
to reinforce this view, and ‘‘hacker” grad- and code until best-practice standards and
ually became a mostly positive term. It is approaches to software design are enforced.
from these roots that the second definition A traditional hacker may be able to produce
evolved. a mock-up of software somewhat quickly,
2. In modern usage, ‘‘hacker” has come but totally unreliable software is of very
to mean only one thing: a person who little value.
attempts to break into other computer sys- The modern form of hackers who break
tems. The motivation may be anything into systems are a corporate liability. The
from basic pride or vanity to theft, sabo- employment of a hacker unbeknownst to
tage, espionage, or revenge. The results may the organization can lead to severe prob-
be anything from a simple web-site post- lems for the company because intellectual
ing to prove that a break-in was success- property may be stolen and released to the
ful, to complete destruction of corporate public, the security and integrity of the cor-
data and modifications to essential records poration may very well be compromised,
or access to confidential data that is never and the company may be used as a base for
detected. The methods range from sim- illegal activities. Companies must under-
ply guessing at passwords (a surprisingly take background vetting of all systems per-
successful technique), through befriending sonnel prior to the assignment of a security
careless or disgruntled employees, to trans- clearance.
mitting targeted viruses, Trojan horses, Organizations need to put in place secu-
and worms. A very common means of rity precautions to prevent all unautho-
illicit entry is to exploit insecurities that rized systems access. This includes access

166
Hardware, software, firmware, application, program

attempts from hackers both internal and


Hardware, software, firmware,
external to the organization. The employ-
ment of a chief security officer is critical in
application, program
large organizations, and the use of skilled Definitions and overview
specialists to undertake frequent security Hardware is the physical part of a
assessments is vital in order to maintain computer: the circuit boards, the silicon
objectivity. chips, the wires, the boxes, the disks, etc.
Hardware is the parts that can be seen or
Summary of positive issues touched.
Hacking as a style of programming is less Software is the more ephemeral compo-
common in organizations employing strict nents that exist only as electrical impulses
programming standards and methodolo- in conductors, magnetic fields or optical
gies. distortions in storage media, or electri-
The unauthorized hacking into systems cal charges in memory capacitors. Soft-
can be defended against through the ware is the programs and applications
employment of security measures, such as that run on a computer and make it use-
mandating well-chosen passwords, instal- ful. The term software is also frequently
ling a reliable firewall, closing all accounts used to include the passive data stored on
immediately an employee leaves employ- computers: images, databases, email mes-
ment at the firm, and installing the latest sages, and so on.
security patches from vendors. Firmware is the very-low-level software,
often safely encapsulated in immutable
ROM (read-only-memory), which is so inti-
Summary of potentially negative issues mately connected with the operation of the
Hacking as a style of programming is waste- hardware that it almost seems to be part of
ful of resources, difficult to maintain, and the hardware itself. The best known exam-
potentially dangerous because the system ple is the BIOS: very basic software respon-
produced might not perform as required sible for starting up a computer when it is
under all circumstances. first turned on, before the operating system
The issue of hackers who break into has started up.
corporate systems to access unauthorized A program is an active piece of software,
information is extremely problematic, and the encoding of an algorithm or computa-
can be the cause of significant losses and tional procedure in a form suitable for exe-
liabilities. Hackers internal to the company cution by the computer. The occasionally
can reveal intellectual property and cause heard term ‘‘software program” is redun-
liability issues for the company. dant. In American English, the word ‘‘pro-
References gram” is spelled the same way in all of
r S. Harris, M. Lester, A. Harper, C. Eagle, its senses. In British English, the spelling
and J. Ness (2004). Gray Hat Hacking: The ‘‘program” is now always used when ‘‘soft-
Ethical Hacker’s Handbook (New York, ware” is meant, the spelling ‘‘programme”
McGraw-Hill). is reserved for the other senses, e.g., ‘‘pro-
r K. Mitnick and W. Simon (2005). The Art gramme of entertainment.”
of Intrusion (New York, John Wiley and Application is simply an alternative
Sons). word for program. Often, ‘‘application” is
reserved for major software products, but
Associated terminology: Virus, Trojan the terms are generally taken to be synony-
horse, Cracker. mous.

167
Health Insurance Portability and Accountability Act of 1996 (HIPAA)

References the organizations that interact with those


r D. Groth (2003). A+ Complete (Hoboken, providers. The act impacts information
NJ, Sybex--John Wiley and Sons). systems in a variety of areas, includ-
r R. Thompson and B. Thompson (2003). ing electronic-transaction standards and
PC Hardware in a Nutshell (Sebastopol, information-security standards.
CA, O’Reilly Press). The HIPAA has mandated standards for
‘‘administrative and financial health care
transactions” based upon ANSI ASC X.12N,
Health Insurance Portability and Version 4010, except for retail pharmacy
Accountability Act of 1996 (HIPAA) transactions, which will use the National
Council for Prescription Drug Programs
Foundation concept: Security. (NCPDP) Telecommunications Standards
Definition: The Health Insurance Portability and Format Version 5.1. The standards are mon-
Accountability Act of 1996 (HIPAA) protects the pri- itored by six designated standards-mainte-
vacyofpersonalmedicalrecords.Itaimstoimprovethe nance organizations (DSMOs), namely ASC
portability and continuity of health-insurance cover- Committee X.12, The Dental Content Com-
age and to provide a framework for obtaining, manag- mittee, Health Level Seven, the NCPDP, the
ing, and disseminating medical or personal informa- National Uniform Billing Committee, and
tion, termed electronic protected health information the National Uniform Claim Committee.
(EPHI). The HIPAA details the obligations of enti-
ties operating under the Final Rule (45 CFR
Overview Parts 160, 162, and 164) of the act including
The HIPAA legislation consists of five titles: administrative safeguards on
Title I covers ‘‘Health care access, porta-
bility and renewability”; Title II covers (1) security-management processes;
‘‘Preventing health care fraud and abuse; (2) assignment of security responsibility;
administrative simplification; medical lia- (3) workforce security;
bility reform”; Title III covers ‘‘Tax related (4) information-access management;
health provisions”; Title IV covers ‘‘Applica- (5) security awareness and training;
tion and enforcement of group health plan (6) the implementation of
requirements”; and Title V covers ‘‘Revenue security-incident procedures;
offsets.” (7) contingency-planning processes;
The US Department of Health and (8) evaluation policies; and
Human Services issued a special rule (9) business-associate contracts.
known as the ‘‘Standards for Privacy of
Individually Identifiable Health Informa- Physical safeguards are also proposed,
tion” (‘‘Privacy Rule”) to help healthcare including
providers and businesses implement and
manage the HIPAA. The privacy rule focuses (1) assigned security responsibilities;
upon the protection of personal health- (2) media controls,
care information and covers the standards (3) physical-access controls;
acceptable for the electronic exchange, pri- (4) workstation-use policies;
vacy, and security of health information. (5) workstation-security location; and
(6) security-awareness training.
Business value proposition
The HIPAA provides a comprehensive The following technical safeguards have
framework for healthcare providers and been proposed:

168
Host, host name, hosting

(1) access controls; Overview


(2) authorization control; Host: In technical descriptions of internet
(3) data and entity authentication; protocols and related documents, the word
(4) integrity controls; Host is used to refer to any kind of computer
(5) message authentication; or other device that can perform the same
(6) encryption; tasks, such as ‘‘smart” telephones and other
(7) alarms; appliances, intelligent routers, and so on.
(8) audit trails; and This is simply to provide a single cover-all
(9) event reporting. term and avoid clumsy phrasing; it has no
other significance.
Summary of positive issues Host name: The name or address of
The act provides a comprehensive frame- an internet-accessible computer, in user-
work for healthcare organizations. The friendly form. Often called a Web address.
Department of Health and Social Security A host name consists of a Domain name
has set compliance dates for each part of (q.v.) with optional prefixes to specify an
the act. individual host. For example, ‘‘www.corpo-
ration.com” and ‘‘sales.eastern.company.
co.uk” are both host names, not domain
Summary of potentially negative issues names; the domain names are ‘‘corpo-
Compliance with the HIPAA standards ration.com” and ‘‘company.co.uk,” respec-
requires significant expenditure of human tively.
and capital resources. Hosting: When an organization or indi-
vidual wishes to have a web presence, the
References
r Public Law 104--191, August 21, 1996, associated costs can be significant. If only
light access is expected, a simple DSL, ISDN,
Health Insurance Portability and
or cable-modem service from a telecommu-
Accountability Act of 1996.
r D. Solove and M. Rotenberg (2003). nications or cable provider, connected to a
desktop computer in the office or home,
Information Privacy Law (New York,
may be sufficient. If exceptionally light
Apsen).
r M. Rotenberg (2003). Privacy Law access is expected, the expense of main-
taining an ‘‘always-on” internet connection
Sourcebook 2003 (Washington, DC,
with the static IP address required for reli-
Electronic Privacy Information Center).
able public access may make even this too
Associated terminology: Security, expensive. If even moderately heavy access
Encryption, Law cross-reference. is expected, such a simple system will prob-
ably not suffice. In both of these cases, a
Hosting service may be the solution.
A hosting service is a company that
Host, host name, hosting already has a physical location with a fast
internet connection, uninterruptible power
Definitions: supplies, some physical security, computers
Host: Any computer. ready for use, and support staff to keep
Host name: A unique identifier for a networked them running. The services they provide
computer. range from a small share of space and band-
Hosting: A third-party service providing computer width on an existing web server to support
equipment, software, maintenance, and network a personal web site, all the way up to 24-
access. hour maintenance on a group of servers

169
Human–computer interaction

dedicated to providing a range of network Overview


services for a single customer. Human--computer interaction (HCI) has been
a part of computing in a variety of ways
Business value proposition since computer applications transitioned
A web-hosting service is an attractive propo- from being batch-mode transaction process-
sition for companies that wish to relieve ing and became interactive. HCI as a disci-
themselves of the burden of managing pline has two main dimensions: Human fac-
their own web server. The ability to out- tors, which considers the problems humans
source this function allows organizations encounter with information presentation
to focus upon their core web activities (e.g., does a pilot prefer dials or digital read-
such as managing the site content, devel- outs for their aircraft instruments); and
oping the applications, and managing Ergonomics, which considers the issues sur-
the customer-service relationships, logis- rounding the design of objects for human
tics, and other aspects of their business. use (e.g., keyboard design to reduce the risk
Web hosting services can also be used as of repetitive-stress injury).
backup service providers should a disaster The study and development of HCI accel-
occur at the corporate site. erated in the 1970s when researchers at
Xerox PARC devised the first Windows-
based GUI (graphical user interface) and
Summary of positive issues invented the mouse as an interface device.
Web hosting provides a relatively low-cost, Other researchers pioneered interface devi-
highly reliable solution to managing web ces such as the tracking ball, the joystick,
servers and an alternative means of pro- touch-screen interfaces, and the light pen.
viding web service should in-house systems These mechanisms allowed greater flexibil-
fail. ity in the way that interfaces could be con-
structed and the means by which users
Summary of potentially negative issues could work with their systems.
Naturally, prospective customers of a host- A chief proponent of HCI development
ing service should carefully consider the has been the military, whose systems have
guaranteed minimum levels of service and become progressively more interactive.
security specified in contracts, and be fully Military simulations require the use of
aware of the costs associated with higher- virtual-reality models and flight simulators
than-anticipated bandwidth use. that allow participants to practice within
a controlled environment. In the simula-
Reference tors, as on their aircraft, pilots wear a
r D. Kaye (2001). Strategies for Web Hosting
‘‘heads-up” display on which the flight and
and Managed Services (New York, John weapons data is projected into their field of
Wiley and Sons). view; this form of HCI is aimed at improv-
Associated terminology: Web services, ISP. ing the reaction times and performance of
the pilots.
The computing environment used by mil-
lions of people every day is frequently just
Human–computer interaction taken for granted, but there are many
devices and mechanisms through which
Definition: Human–computer interaction is the study, users can and do interact with computers,
design, and evaluation of interactive computing envi- again usually without thought, and this
ronments as used by human subjects in order to ease of use is a direct result of HCI research.
accomplish a process or task. For example, many automobile controls are

170
Hypertext, HTML

actually computer interfaces, including the Overview


‘‘drive-by-wire” accelerator and braking sys- Plain-text files are simply sequences of char-
tems, the hands-free (Bluetooth) mobile- acters and line-breaks exactly as they could
phone operation, the navigation system, have been produced on an old-fashioned
heating and entertainment systems, and manual typewriter. When computers were
the steering-wheel ‘‘paddles” for gear shift- strictly technical tools and computer
ing. interconnectivity was low-bandwidth if it
existed at all, all documents were either
Business value proposition plain text or in some application-specific
The value of HCI research is to be found format. Plain text was perfectly adequate
in a variety of areas, including the safer for the intended purposes, but is of little
and more accurate use of computer systems use in the world of desktop publishing and
(e.g., automotive, military). Correct assess- graphic-conscious web pages.
ment of HCI issues can result in a more pro- When computer typesetting and desk-
ductive workplace where computer users top publishing first appeared, macro lan-
are not taking time off for computer-use- guages became popular. The premise is that
related medical problems such as back text is typed exactly as normal, but, when-
pain, carpal tunnel syndrome, and eye ever some special action is required, spe-
strain. cial symbolic tags are inserted into the text.
When the author is working on the text,
Summary of positive issues these tags are nothing special; they just
HCI is an area of active research in aca- appear as strange symbols in the text, and
demic and commercial organizations. HCI can be edited along with everything else.
has an extensive literature. HCI can lead to When the document is specially processed
more productive workforces. for printing, the tags are not displayed at
all; instead, they control various aspects of
Summary of potentially negative issues the document’s appearance.
The incorporation of HCI research into a One of the most successful applications
standard human--machine interaction may of this kind was the TeX/LaTeX system from
require significant expenditure of resour- Donald Knuth and Leslie Lamport. It was
ces, but there is a potential for a high return observed that the character ‘‘ \ ,” which
on this investment. appears on all computer keyboards, is never
used in normal writing, so it was taken as
Reference a special flag symbol. In the LaTeX system,
r J. Preece, Y. Rogers, H. Sharp, D. Benyon, text is taken exactly as it is typed, until a
S. Holland, and T. Carey (1994). ‘‘ \ ” sign is encountered, at which point the
Human--Computer Interaction: Concepts and following text component is taken as a spe-
Design (New York, Addison-Wesley). cial typesetting command. For example, a
LaTeX document may contain this text:
Associated terminology: Virtual reality.

the shop-keeper insisted it was a


\ it heffalump, \ rm
Hypertext, HTML and cost \ pounds 7.50; this was
less than \ frac{3}{4} of the
Foundation concept: World Wide Web. price I had
Definition: Text enhanced with additional features seen elsewhere, but the coloration
such as automatic formatting, images, and links to was
other documents rather pass \ ’{e}.

171
Hypertext, HTML

It is quite easy to read as it appears, to adapt to a great variety of conditions.


which is very helpful when editing, espe- A user may have reduced the web-browser
cially with the knowledge that ‘‘ \ it” is window to a long thin stripe to keep it out
a command to start using the italic type- of the way. HTML should still be able to fill
face, ‘‘ \ rm” is a command to return to the it with text and keep it readable, whereas
normal roman typeface, ‘‘ \ pounds” repre- a more exacting language like LaTeX would
sents the British currency symbol, and so insist on an exact layout for a specific page
on. After processing, the above text would size and would not leave browsers free to
be rendered on a printer as adapt.
This great flexibility in HTML has led
the shop-keeper insisted it was a heffalump, to a great deal of dissatisfaction amongst
and cost £7.50; this was less than 3/4 of the
the authors of web content, who often
price I had seen elsewhere, but the col-
have a very fixed idea of exactly how their
oration was rather passé.
web pages should appear, with carefully
Hypertext systems are really just an planned positions for graphical elements
extension of this basic style. By far the most and perfectly selected artistic fonts. It is
common in current use is HTML, the Hyper- often possible to wrest complete control
Text Mark-up Language, the standard for from the browser by using newer HTML
which is maintained by the W3C organi- extensions and such features as CSS (cas-
zation. HTML is the universal language for cading style sheets) and JavaScript con-
web pages, and, although various manu- trols. The result can be most unfortu-
facturers’ browsers make significant varia- nate, producing web pages that can be
tions on the details of the standard, basic viewed only at one particular size; there are
HTML is universally understood by all web many commercial sites that are completely
browsers. unprintable because they have been over-
Although HTML has a slightly different engineered for the traditional short-but-
intent from LaTeX, it is very similar in wide monitor configuration. HTML should
appearance, and gives the same advantages: be used to provide general layout guidance,
authors may work with simple plain-text not exact typesetting instructions. Hyper-
files but produce very sophisticated layouts. text is not necessarily tied to HTML, but,
The example above would appear thus: simply because there are currently no com-
peting systems with such universal support
the shop-keeper insisted it was from web browsers, and such universal and
a <em>heffalump,</em> and cost free accessibility to authors, HTML does cur-
&pound;7.50; this was less than
rently define hypertext.
&frac34; of the price I had seen
The most essential feature of hypertext
elsewhere, but the coloration was
rather pass&eacute;. is not its ability to specify the appearance
of text with markup elements, not even its
The markup tags in HTML are intended ability to include full graphics integrally
to indicate the purpose of the text, rather in the display of a document; these fea-
than to exactly control its appearance. A tures merely reproduce what is available
clear example is the <em> tag used above; in books. The definitive feature of hyper-
this indicates that the affected text is to text is the ability to actively link documents
be emphasized in some way. Most web together. Whilst reading one document, a
browsers will provide emphasis by using person may decide to follow a reference or
italics, but that is not a requirement. In look up an unfamiliar word or phrase. With
retaining some flexibility in how docu- a simple click of the mouse, the appropri-
ments are to be displayed, HTML is able ate document appears on the screen along

172
Hypertext, HTML

with the original. The viewer may read the ing Group. HTML continues to evolve, with
new document, then return to the original, a working draft of the Extensible HyperText
or read the two together, or follow up on Markup Language (XHTML) version 2.0 hav-
another reference in a limitless chain. The ing been released on May 27, 2005. XHTML
immediate and concurrent access to linked (HTML written in XML) advances certain
documents without the requirement for aspects of earlier versions of HTML and is
any searching is what puts the ‘‘hyper” into backwards compatible, such that all older
‘‘hypertext.” HTML documents continue to display cor-
The support for connected documents, rectly on newer browsers.
or Hyperlinks, provided by HTML is surpris-
ingly primitive. The author of a document Summary of positive issues
must decide exactly which other docu- HTML is an open-source, non-proprietary
ments should be linked with any given sec- system. Basic HTML is easy to program, flex-
tion of his or her composition, and explic- ible, universally used in hypertext devel-
itly add markup tags giving the location of opment, and understood by all browsers.
the document as a URL (universal resource HTML continues to evolve through the
locator, q.v.). There is no mechanism for activities of W3C.org. HTML provides a
specifying a general look-up rule for all strong basis for web-site development and
words in a document, nor is there any tools are available to help developers
mechanism for ensuring that the linked improve their productivity.
document has not changed, and is still
relevant. What we have today is barely Summary of potentially negative issues
more sophisticated than the Memex sys- HTML has become by default the standard
tem invented by Vannevar Bush (but never vehicle for hypertext presentation. Organi-
built) in 1932; he envisioned a library of zations need to deploy resources to under-
documents on microfilm physically linked stand it fully and understand its limita-
together with fine threads, and a special tions. This may be in the form of training
machine to display those documents and and/or being active in the W3C.org HTML
follow threads when the user pulls a lever. Working Group. The flexibility of HTML can
The World Wide Web as we know it today be problematic to developers wishing to
is for the most part a widely distributed col- develop a web page with fixed dimensions.
lection of hypertext documents. Most are HTML is the dominant web development
written in HTML, some still in plain text, language and will remain so for the fore-
and some use proprietary systems (the most seeable future.
popular of which is Acrobat), for which
References
viewers are available for free, but com- r C. Musciano and B. Kennedy (2002).
position requires commercially available
HTML and XHTML: The Definitive Guide
software. Surprisingly, there are no LaTeX
(Sebastopol, CA, O’Reilly Press).
viewers available that can be integrated r P. Gralla (2004). How The Internet Works
with the standard web browsers.
(Indianapolis, IN, Que).
r L. Lamport (1986). LaTeX User’s Guide &
Business value proposition Reference Manual (New York,
HTML is universal in its use as the base for Addison-Wesley).
hypertext web documents. HTML is a non- r V. Bush (1945). ‘‘As We May Think,”
proprietary format developed through the Atlantic Monthly, July.
W3C.org consortium and the HTML Work- r http://www.w3.org/TR/REC-html40/.

173
ICANN (Internet Corporation for Assigned Names and Numbers)

ICANN (Internet Corporation for Business value proposition


ICANN is the ultimate authority in the allo-
Assigned Names and Numbers)
cation of domain names and IP addresses,
Foundation concept: Domain name. and in the creation of top-level domains.
Definition: The non-profit corporation delegated There are currently seven top-level domain
responsibility for the allocation of IP addresses and names aside from the national domains;
other internet-related identifiers by the US Depart- however, there have been proposals to cre-
ment of Commerce. ate new top-level domain names, includ-
ing ‘‘.asia,” ‘‘.cat,” ‘‘.jobs,” ‘‘.mail,” ‘‘.mobi,”
‘‘.post,” ‘‘.tel,” ‘‘.travel,” and ‘‘.xxx.” ICANN
Overview has requested public comments on these
IP addresses and domain names are criti- proposals.
cal to the operation of the internet, and at Most organizations will have no direct
the top level must be managed by a rec- dealings with ICANN, since domain name
ognized authority. The US Department of registration and IP address allocation is
Commerce somehow became the interna- usually handled by intermediaries such as
tional authority for such things, and del- ISPs and hosting services.
egates that authority to the Internet Cor-
poration for Assigned Names and Numbers Summary of positive issues
(ICANN). It is a non-profit corporation, with ICANN controls the top-level domain
an international board, and has subsumed names. Registration of domain names is
the former responsibilities of the Internet well organized and they can be acquired
Assigned Numbers Authority (IANA). through ICANN-authorized companies such
ICANN usually allocates IP addresses in as ISPs.
large groups to regional authorities, who
then allocate them individually or in smal- Summary of potentially negative issues
ler groups to ISPs and other organizations. The top-level domain names have been
ICANN is also responsible for the overall fixed and expansion is slow to occur. Many
management of DNS, the Domain Name domain names have already been taken.
Service, and creates the top-level domains Some domain name categories are less
(such as ‘‘.com” and ‘‘.org”) and the national attractive to corporations attempting to cre-
domains (such as ‘‘.uk” and ‘‘.nu”). ate their brand than others.
ICANN is often blamed, completely
unfairly, for the fact that IP addresses References
r http://www.icann.org.
are running out, and organizations in the
r ICANN, 4676 Admiralty Way, Suite 330,
developing world might not be getting the
share they would like. They are also some- Marina del Rey, CA 90292, USA.
times blamed, perhaps less unfairly, for Associated terminology: Host, Domain
the fact that prices for registration are name, Internet, World Wide Web.
much higher than they once were, and this
is seen as unreasonably favoring the old
established organizations over new ones. IEEE (Institute of Electrical and
Most organizations have no direct dealings Electronic Engineers)
with ICANN, since domain-name registra-
tion and IP address allocation is usually Definition: A professional organization for individ-
handled by intermediaries such as ISPs and uals involved in all areas of electronics, particularly
hosting services. computers.

174
Index

Overview reviewed articles. The IEEE also hosts con-


The IEEE, the Institute of Electrical and ferences and seminars on a wide range of
Electronic Engineers, is a major interna- computer-related topics, and has a long his-
tional society, probably the largest profes- tory of involvement in the development of
sional organization in the world, with over standards and protocols. The society has
300 000 members. It is concerned with all a membership progression from Associate,
aspects of electrical systems, and the IEEE Member, Senior Member, to Fellow. The
Computer Society is a major subgroup ded- grade of Member requires that the person
icated to all areas of computing: hardware, has ‘‘satisfied IEEE-specified educational
software, standards, and people. requirements and/or has demonstrated pro-
The IEEE is a major publisher of peer- fessional competence in IEEE-designated
reviewed research, and produces a large fields of interest,” whereas Senior Members
number of technical journals. It is also a need to have demonstrated at least ten
major influence on standards and proto- years of ‘‘professional maturity.” The career-
cols used throughout the computing world, development model is a useful mecha-
most famously producing the IEEE floating- nism for IT organizations to employ to
point format and all of the major standards ensure that their workforce is continu-
for network communications. The society ing to develop professionally and being
takes a leading role in advanced education involved in the development of a profes-
and accreditation of academic programs in sional body.
electrical, electronic, and computer engi-
neering.
Summary of positive issues
The society also promulgates a code of
The IEEE provides the IT community with
ethics, but, since computing is an almost
a wide range of resources and a career-
entirely unregulated profession, member-
development path for its membership. The
ship in professional organizations is totally
IEEE is a recognized body for the creation
voluntary, and codes of ethics are almost
of technology standards and protocols.
entirely unenforceable. The lack of regu-
lation is proving to be a major source of
danger and financial loss in this computer- Summary of potentially negative issues
dependent world, and, although there will Membership by IT workers is voluntary and
be strong objections from established prac- adherence to its code of ethics is therefore
titioners, it seems inevitable that profes- also voluntary.
sional regulation will eventually become
References
mandatory, as it is in all other engineering r www.ieee.org.
professions. r IEEE Corporate Office, 3 Park Avenue,
17th Floor, New York, NY 10016, USA.
Business value proposition Associated terminology: ACM, BCS, W3C.
The IEEE provides a wide range of ser-
vices for IT professionals, academics, and
other practitioners. These include the high-
quality IEEE publications, which range Index
from the very theoretical peer-reviewed
IEEE Transactions series to the more popu- Foundation concept: Database.
lar IEEE magazine format (e.g., IEEE Com- Definition: An organization imposed upon data that
puter), all of which provide leading peer- makes future searches faster or more efficient.

175
Index

Overview for example, a telephone directory may be


Data is of no use unless it can be found. If indexed both on name and on number, so
a major city’s telephone directory imposed that reverse lookups (look up a number
no useful organization upon the data to find the subscriber’s name) are just as
it contained, perhaps listing subscribers fast.
in the order in which they originally When creating a database, the designer
signed up for service, it would contain must be aware of the kinds of queries
exactly the same data as a conventional and other accesses that are likely to
telephone directory, but would be entirely be performed. With this information, an
useless. The few million entries in a informed decision can be made on what
telephone directory pale into insignifi- indexes should be created. Any complete
cance when compared with the number of DBMS (database management system) will
records in a major commercial database. automatically create whatever indexes are
Any significant collection of data relies requested. Usually, indexes are set up when
upon an effective indexing scheme for its that database’s tables are defined; creating
usefulness. an index after the data has been added
Indexing data may involve rearranging can be a computationally intense process.
that data; the familiar example of the tele- The method described above, namely that
phone directory illustrates this. The data of creating an alphabetical list of head-
may be physically organized in alphabeti- words and page numbers, is very popular
cal order (or some other logical arrange- in database implementations, and is often
ment), and this certainly allows effective known as ISAM (indexed sequential access
fast access. However, maintaining a special method). Other popular forms of database
ordering on a dynamic data set can be pro- index are the B-Tree (which provides much
hibitively expensive in terms of time and faster access, but is much more demanding
effort. If the telephone directory had to in terms of storage requirements), and the
be recomposed and reprinted every time Hash table (which is even faster, but does not
a new subscriber joined the system, the support alphabetical browsing).
result would be extremely wasteful.
Most practical indexing schemes involve Business value proposition
leaving the original data exactly where it Indexing is an essential part of data orga-
is in a totally arbitrary order, and creat- nization, and it takes some technical train-
ing a new Index file that contains only the ing in database management to make the
parts of the data necessary for rapid access. right decisions about which indexes should
For example, if a very large encyclopædia be created. Every index requires significant
allocated exactly one page per entry, but storage space and adds to the time required
those pages were totally unordered, a very to perform an insertion or deletion, so hav-
effective index could be created by sim- ing too many indexes can be a costly mis-
ply listing the head-words and correspond- take, but having too few indexes can result
ing page numbers. The index itself would in a query taking literally millions of times
be in alphabetical order, but the pages it longer than it needs to.
refers to would not. The result would work While the DBMS software performs the
just as well, but would require significantly actual tasks of indexing, it is the role of
less effort to maintain. Ordering the index the database administrator (DBA) to make
instead of the actual data makes it possible a determination on the trade-offs between
for the same data to appear to be sorted system performance and resource requi-
in two incompatible orders concurrently; rements. The DBA may then create and

176
Information lifecycle management (ILM)

implement an appropriate data structure


Information lifecycle management
in an attempt to meet the requirements.
DBMS software is not the only com-
(ILM)
mon application of indexing. Commercial Foundation concept: CIO.
web-search engines that perform keyword Definition: Information lifecycle management is the
searches on billions of web pages in a process of managing data from its point of creation to
fraction of a second could not do so if when it is purposely and permanently deleted.
the searches were actually performed when
requested. Instead, an index is created by Overview
pre-searching for all possible words, and The ILM concept covers the storage and
saving the results, so that each simple user management of data during its entire life-
query may be answered immediately by cycle. Corporate data takes several forms,
simply retrieving already-known answers. including live transaction data that resides,
Many operating systems for personal com- for example, on an ERP system; data that
puters provide an indexing service that, has been through an extraction, trans-
once it has been set up, allows rapid search- formation, and loading (ETL) process and
ing for any file containing a particular word resides in a data warehouse; data that is
or phrase. in backup mode to enable transactional or
data warehouse data to be replaced; and
Summary of positive issues data that is archived or held in long-term
Indexing is a well-known method for struc- storage (10+ years) to satisfy legal compli-
turing and organizing data with a large ance requirements or other needs.
theoretical, academic, and practitioner lit- ILM stresses not only the security of the
erature. DBMS systems contain tried and data but also the managerial issues that
tested indexing algorithms that are capa- surround the lifecycle, assessing the value
ble of automatically indexing data very effi- of the data (its age, whether the data is in
ciently. live usage, whether the data is warehoused,
required for compliance, or merely backed
Summary of potentially negative issues up for completeness), and using the appro-
Poorly designed indexes can lead to a priate technology to store the data at mini-
lower overall performance in some circum- mal cost over its entire lifecycle. This dual-
stances than other methods for file struc- ity of view is necessary because corporate
turing. Index file structures in some imple- data ownership continues to grow, as do
mentations may require more direct-access the requirements to keep data for longer.
memory space to store their indexes. In order to reduce overheads, manage-
rial and technical, CIOs deploy a variety of
References technologies. These include Storage resource
r C. Date (2000). An Introduction to Database
management (SRM) systems that manage the
Systems (New York, Addison-Wesley). data across the organization and Hierar-
r R. Elmasri and S. Navathe (1994).
chical storage management (HSM) tools that
Fundamentals of Database Systems (New manage the data upon different types of
York, Addison-Wesley). storage device and facilitate the archiving
r T. Cormen, C. Leiserson, R. Rivest, and
of data (e.g., from a high-performance disk
C. Stein (2001). Introduction to Algorithms storage unit to a tape library). As data archi-
(Cambridge, MA, MIT Press). ving has grown in importance with the
Associated terminology: Database, changes in regulatory compliance require-
Database administrator. ments, the value and urgency of moving

177
Information Technology Infrastructure Library (ITIL)

data from one type of storage medium to Many vendors support the ILM philosophy
another has changed. Accordingly, HSM has through the provision of tools and services.
evolved into what are termed Automated
data migration (ADM) tools that intelligently Summary of potentially negative issues
move data from one type of data stor- The development of an ILM initiative can
age to another. These tools are often built be resource intensive, potentially requiring
upon a Storage area network (SAN), a high- a company to re-engineer its systems to
performance independent sub-network of ensure alignment between the IT strategy
a larger corporate network that specifi- and the ILM strategy. The risks associated
cally connects dedicated storage devices with corporate data security, location, and
together, allowing high volumes of data ownership over the entire lifecycle need to
to be stored, accessed, and moved around be assessed.
a network as required, without impacting
Reference
the existing corporate network. r R. Maier, T. Hädrich and R. Peinl (2005).
Enterprise Knowledge Infrastructures
Business value proposition (Berlin, Springer-Verlag).
ILM aims to provide complete control and
Associated terminology: ERP, Data
management of the data lifecycle. The use
warehouse, ETL.
of an ILM philosophy that is backed with
SRM and ADM tools allows organizations to
achieve several goals. Firstly, they create a
Information Technology
system that balances data accessibility with
value and need. Secondly, the ILM philos-
Infrastructure Library (ITIL)
ophy helps ensure regulatory compliance Foundation concepts: CIO, MIS.
with regard to data storage. Thirdly, the Definition: A collection of best practices for the provi-
systems allow efficient data storage across sion of IT service management (ITSM).
each data storage level. Fourthly, the sys-
tems allow much of the overhead to be Overview
automated, allowing the systems manager The Information Technology Infrastructure
to focus upon other issues. Fifthly, the ILM Library† is a collection of IT service manage-
philosophy and systems may be designed ment best practices that aim to provide the
to map onto an organization’s disaster- leaders of IT organizations with a top-down
recovery and contingency plans. business-driven approach to the manage-
ILM is supported by a variety of tools and ment of IT processes, people, and technol-
vendors, including Storage service providers ogy. The library was originally developed in
(SSP), which act as providers of storage ser- the 1980s by the UK government’s Central
vices, including offering a wide range of Computer and Telecommunications Agency
services for outsourcing of data storage. (CCTA). Since then it has been refined
These include storing data, providing and extended by the UK’s Office of Gov-
backup services, and remote data-conti- ernment Commerce (OGC), which subse-
nuity services via SANs, VPNs, or internet quently published a Version 2 of the ITIL
connections. (1998--2004) and has commenced work on
Version 3. A British Standard (BS 15000-
1:2002) based upon ITIL has also been deve-
Summary of positive issues
loped; this standard has subsequently been
ILM enables a data-value approach to data
storage to be taken. ILM assists organi- †
ITIL is a registered trade mark of OGC -- The Office of
zations with regulatory data compliance. Government Commerce.

178
Information Technology Infrastructure Library (ITIL)

developed into an international standard: processes associated with more advanced


ISO 20000. and continued support are also defined.
The ITIL model is broken down into eight These may include routing the problem to
areas: more advanced technical-support person-
nel, notifying the appropriate levels of IT
r service support, and business-process owner of the prob-
r service delivery, lem, and identifying alternatives should
r planning to implement service the problem not be resolved within an
management, acceptable predefined time limit, so that
r security management, service can be restored to the agreed levels.
r ICT (information and communications The components are managed in con-
technology) infrastructure junction with a database, known as the
management, Configuration management database (CMDB),
r application management, that maintains a ‘‘map” of the IT infras-
r software asset management, and tructure within the organization and a log
r business perspective. of the problems or ‘‘incidents” associated
with aspects of the infrastructure. These
Each of the eight components focuses upon incidents can then be analyzed and prob-
a specific aspect of IT operations and breaks lem areas identified and resolved, further
that aspect down into services. For each ser- improving the quality of the IT service.
vice, best practices are provided at three
levels: the strategic, the tactical, and the Business value proposition
operational. The ITIL has been used in the UK and
The methodology is hierarchically based the European Union since its introduction
and allows the selective implementation in the 1980s and has continued to grow
of only those aspects relevant to the in popularity in other parts of the world.
organization. For example, within the Acceptance and use of ITIL in the United
service-support component there are sub- States has been driven by the need to have
components including help-desk manage- stronger IT governance in order to accom-
ment, incident management, problem modate regulatory compliance legislation
management, configuration management, such as Sarbanes--Oxley and HIPAA.
change management, and release manage- The ITIL allows organizations to stan-
ment, some or all of which may be adopted dardize the processes associated with their
and implemented by the IT organization. use of IT. Central to the approach is the pro-
While the ITIL does not specify the actual vision of standardized definitions for many
steps to be performed in order to exe- of the terms used in the IT organization,
cute the best practice, the operational guid- which allows all members of the organiza-
ance is sufficient that IT professionals can tion to have a common basis for discussions
be confident in their implementations. In pertaining to people, processes, and the IT
the case of help-desk management, the systems. This common basis of understand-
incoming calls from computer users can be ing enables all parties to have a more accu-
divided into appropriate categories, such rate understanding of the status and role of
as laptop, desktop, network, and main- technology within the organization, from
frame, and routed to specialists in each of which more developed business cases and
these areas. The service-level requirements solutions can in turn be created.
for each area can then be defined (e.g., By standardizing on best processes and
to resolve each problem in 30 minutes). easing communications between technolo-
Should the problem not be resolved, the gists and the users in the business units,

179
Instant messaging (IM)

ITIL is aimed at reducing the costs of ature on ITIL and a growing body of con-
IT deployment and maintenance while sultants and practitioners to support the
raising service levels. The standards-based implementation of ITIL. Certification of ITIL
approach allows consistent guidance to practitioners is available, this being sanc-
be provided by the IT organization, and, tioned by professional IT bodies such as
through certification-based training, it cre- the British Computer Society and the UK
ates a consistent knowledge level within Government’s Office of Government Com-
the workforce, also leveraging the ability merce. The ITIL approach to service man-
to recruit certified individuals to support agement is scalable and structured, and
ITIL-based procedures. The standards-based allows a standardized and more efficient
approach to IT service delivery can also be and effective implementation of IT prac-
applied to contract vendors delivering out- tices. It can be combined with other IT gov-
sourced services, in that a contactor can be ernance frameworks, including CobiT and
required to be ITIL-certified. This provides ISO 17999.
a higher basis for IT service quality and
for competitive bidding on service provi- Summary of potentially negative issues
sion due to the processes being fully under- ITIL provides guidance on how to achieve
stood by multiple vendors. best practice IT service management, but is
ITIL certification is sanctioned by the ITIL not specific on the fine details of implemen-
Certification Management Board (ICMB) tion. The implementation process can be
and is composed of the OGC, the ITSMF slow and resource-intensive, and the asso-
(IT Service Management Forum) and the ciated process changes can be disruptive in
two examination institutes EXIN and ISEB. the short term.
The ICM has created three levels of certi-
fication: foundation, practitioner, and man- References
r http://www.bs15000.org.uk/.
ager. Foundation certification recognizes
r http://www.itsmf.com/.
that the holder is familiar with best prac-
r http://www.ogc.gov.uk/.
tices in ITSM, practitioner certification rec-
r B. Worthen (2005). ‘‘ITIL power,” in CIO
ognizes that the holder understands the
theory of ITSM and how to apply that the- Magazine, September 1, 2005.
ory in practice, and the manager certifica- Associated terminology: ISO/IEC 17999,
tion recognizes that the holder can manage British Computer Society, Sarbanes--Oxley.
ITIL solutions across a range of service-
management areas.
The ITIL is frequently used in combina-
tion with ISO/IEC 17999 and the COSO/ Instant messaging (IM)
CobiT frameworks. Typically, ITIL is used as
the backbone framework for delivery and Foundation concepts: Email, Client–server.
support processes in conjunction with ISO Definition: An interactive internet communications
17999, around which security controls are method, akin to a typed telephone call.
created, while COSO/CobiT is used on con-
trols and metrics that pertain to financial Overview
systems and Sarbanes--Oxley compliance. Whereas email is strictly non-interactive, a
message is typed and sent, and the recipi-
Summary of positive issues ent opens and reads it when they feel like
ITIL is a widely used methodology for it, Instant messaging (IM) makes the basic
implementing best practices in IT service concept of email interactive. Communi-
management. There is a substantial liter- cating by email is the internet version of

180
Instant messaging (IM)

sending telegrams or letters; communica- the provider is not spying on communica-


tion by IM is the internet equivalent of a tions, the vicissitudes of operating systems
typed telephone call. Both (or all) partici- can result in messages being accidentally
pants in communication must be actively preserved on servers, and these are subject
using the IM system at the same time; as to discovery in legal proceedings.
one types or inserts graphical elements, Just as the popularity of email has been
what they type is immediately shown to a boon to e-commerce at the same time as
the other(s). All participants may, and often proving a great destroyer of office produc-
do, type at the same time, producing a very tivity, IM is also a double-edged sword, but
confused ‘‘conversation.” this time the positive aspects can be hard
Instant messaging systems work in one to detect. Unlike email, IM requires the
of three modes. In Character mode, every key- full attention of both parties for the dura-
press is sent to every participant as soon as tion of the conversation, the same as with
it occurs; in Line mode, nothing is sent until telephone calls, except that IM takes far
a whole line has been typed, and the enter more time than a telephone call because
key has been pressed, giving typers the most users are not very good typists. The
opportunity to correct mistakes before they extreme ease and absence of direct finan-
are seen; and in Message mode, participants cial cost associated with IM means that
compose whole sentences or paragraphs or people are far more likely to initiate com-
more, and it is all sent at once when they munications for simple matters that could
click a ‘‘send” button or key. Message mode have been resolved with just a few seconds’
allows users to see what they are writing thought. The immediacy of IM makes seri-
and perhaps think better of it before send- ous thought most unlikely.
ing, so it always results in more rational This does have one possibly positive
and productive conversations, makes more aspect: because IM frequently involves very
efficient use of network bandwidth, and little thought, it is sometimes possible
is easier to implement correctly. Character to substitute an artificial intelligence (AI)
mode is generally the most popular because application for a real human correspon-
there are no frustrating delays while some- dent. The famous Eliza experiments con-
one is typing, and it gives the illusion of a ducted by J. Weizenbaum starting in 1965
real face-to-face communication, although were simple (by modern standards) pro-
it is exceptionally inefficient. Line mode grams that simulated real people in con-
seems to be the most common choice for versation. They work by picking out what
IM service providers. seem to be key words in what a person
Although instant messaging has become types, and reinserting them, transformed,
the poster child of the Peer-to-peer move- into standard responses selected from a
ment in network communications, many predefined list. The original Eliza simu-
commercial systems in fact use a Client- lated a psychotherapist, and was astonish-
-server model: when two participants are ingly successful, eliciting embarrassingly
communicating, what they type does not frank admissions from people who should
go directly from one computer to the other; have known better. Modern Elizas can often
instead it goes first to a central server, answer the majority of customer-service
and thence to the intended recipient. This enquiries through the same techniques.
makes implementation easier, and allows Organizations must take great care if con-
the service provider to maintain complete sidering this option, since an Eliza-like
control over all aspects of its use. It also system, or any other AI technique yet dis-
means that communications might not be covered, will be unable to answer any com-
as private as users would imagine; even if plex questions that could not have been

181
Internet

answered by simply reading the manual, operate. Instant messaging provides a low-
and may completely alienate customers. cost communication option.

Business value proposition Summary of potentially negative issues


Instant messaging has become a ubiquitous Abuse of IM by workers can be detrimen-
aspect of communications in today’s soci- tal to corporate performance. The messages
ety. It is easy to use and accessible wher- may need to be stored for regulatory com-
ever there is an internet connection. It has pliance. Instant messaging systems provide
taken its place alongside the conventional yet another unsecured channel into a com-
telephone call and provides a convenient puter system; although it is currently not a
mechanism for communicating when the real problem, it is quite likely that this will
telephone might not be appropriate, e.g., prove to be an exploitable vulnerability in
when a person is at work and a long ‘‘chat” the future.
to a friend on the telephone could be
noticed and contravenes corporate policy. References
r R. Khan and B. Blair (2004). Information
The IM systems can, however, be disabled
by network managers and this would be Nation: Seven Keys to Information
likely to raise corporate productivity. It is Management Compliance (New York, Aim
unlikely that IM would be advantageous International).
r P. Gralla (2004). How the Internet Works
inside a corporation when a telephone con-
versation is more appropriate. However, (Indianapolis, IN, Que).
r J. Weizenbam (1966). ‘‘ELIZA,”
when there is a need to communicate over
long distances and where the telephone Communications of the A.C.M., January.
charges are high, IM can provide an advan- Associated terminology: Email, Voice over
tageous alternative. However, Voice over IP IP, Peer to peer, Client--server.
(internet telephony) can offer the commu-
nicating parties the same advantages as IM
but with true voice communication. Internet
In some industries, e.g., the brokerage
Foundation concept: Network.
business, the use of any communication
Definitions:
tools requires that the communication be
recorded. This prevents disputes over a Internet: the entire worldwide community of
potential ‘‘buy” or ‘‘sell” order between a interconnected computer networks.
broker and a client. Instant messaging is Internet communications: communications
not immune to this requirement; however, between two systems that are not part of the same
technically this is a more challenging task local-area network.
and requires that procedures to capture Private internet: a group of local-area networks
and store the interactions be put in place. sharing a private IP-address space, and unable to
An organization with moderate IT capa- communicate directly with the internet as a whole.
bility should be able to produce its own
proprietary IM system that conforms to Overviews
any requirements and restrictions they may Private internet: one of the ways to over-
have. come the global shortage of IP addresses
is for an organization to allocate private
Summary of positive issues IP addresses to its own systems. Private IP
Instant messaging systems are easy to addresses are those beginning with ‘‘10.”,
install on a computer and require only an ‘‘172.16.” to ‘‘172.31.”, and ‘‘192.168.”. Pri-
internet connection in order for them to vate IP addresses are valid for use on an

182
Internet

internal network if it is configured cor- by ICANN, but that is not usually a cost-
rectly, but are not valid for use on the effective solution.
whole internet. Computers arranged in this Credit for inventing the internet has
way are unable to communicate with the been awarded to many people. Three dif-
outside world except through Proxy servers ferent groups, Paul Baran of the RAND
(see Network Address Translation and Proxy corporation, Leonard Kleinrock of MIT,
for details), and are said to form a Private and Donald Davies and Roger Scantle-
internet. bury of NPL (National Physical Laboratory,
Internet communications: communica- UK) seem to have developed the idea of
tions within one Local-area network (LAN) Packet-switched networks (see Packet switch-
are referred to as intranet communications. ing) independently and at about the same
Communications between systems that are time, around 1962. In 1968, the Advanced
not on the same LAN are referred to as Research Projects Agency (ARPA) of the US
internet communications. Internet com- government funded the first internet-like
munications can not be based on aspects computer network, which was known as
of LAN protocols, such as ethernet hard- ARPANET, and came on line in 1969, con-
ware addresses (MAC addresses), and must necting UCLA (University of California, Los
use higher-level protocols (such as IP, the Angeles) and SRI (Stanford Research Insti-
internet protocol) instead. This merely is a tute), soon joined by UCSB (University of
technical distinction. California, Santa Barbara) and the Univer-
The internet: The internet is a cover- sity of Utah. ARPANET gradually expanded
all term for all of the computers, both over the years, linking with other national
servers and clients, and other network- networks, until it grew to be what is today
enabled devices, together with the connec- called the internet.
tions between them, that form the global The TCP/IP protocol suite was invented
digital communications network. by Vinton Cerf of Stanford and Bob Kahn
The internet is tied together by IP, the of ARPA (by then renamed DARPA, the D is
Internet protocol, a system that defines how for Defense) starting in 1973, and in 1974
individual devices are addressed, and how the word ‘‘internet” was used for the first
transmissions are routed throughout the time in their published paper describing
network from one device to another. It pro- TCP. In 1983, the University of Wisconsin
vides a uniform format for transmissions introduced the Domain-name system, which
that is capable of carrying any form of dig- completed the basic infrastructure of the
ital data over any kind of connection, with modern internet. The system that is con-
reasonable efficiency. sidered by many to give life to the modern
Every device must have an IP address allo- internet, the ‘‘World Wide Web,” was not
cated to it before it can participate in the invented until 1990 (see World Wide Web for
internet. IP addresses may be static (per- details).
manently allocated to a particular device), The Backbone of the internet consists of
or dynamic (allocated on demand and a large number of major switching cen-
released after a short period). Blocks of ters each connected to a moderate number
IP addresses are allocated to organizations of others. Many smaller switching centers
by an international organization known as may be connected to each of the major
ICANN (Internet Corporation for Assigned ones, and many smaller yet to each of
Names and Numbers), and those organi- them, down to the level of the LAN. Each
zations then allocate individual addresses individual computer generally has one
within their blocks as they see fit. Individ- communications path available to one of
ual IP addresses can be allocated directly the major switching centers, but the major

183
Internet

switching centers are very highly intercon- architecture, and cost of these systems is
nected, and may communicate with one classified.
another over a wide variety of paths, choos-
ing whichever is best at any given moment. Business value proposition
This is what gives the internet its strength: The internet has enabled individuals, cor-
any switching center could fail, and only porations, and entities within corpora-
those individual computers relying on it tions to connect together in many new
would become disconnected; the internet and productive ways. These include Online
as a whole would be undamaged, since communities in which virtual communities
there will always be a variety of paths that unite members to solve problems, discuss
avoid any out-of-service node. One of the issues, and foster commerce. The inter-
original purposes behind packet-switched net has spawned a whole new mechanism
networks was to provide a military com- for the development and delivery of busi-
mand structure that would still be opera- ness models and commerce including Col-
tional after a disastrous nuclear attack. laborative commerce, Business-to-business (B2B)
Since the World Wide Web converted commerce, and business-to-consumer (B2C)
the internet from a governmental and aca- commerce.
demic research tool into a wildly popular The technology has grown to include
public utility, and email became a mass new mechanisms for connecting to the
communications (and miscommunication) internet and protocols such as WAP allow
medium, the original purpose of the inter- mobile devices using Micro-browsers to
net has become compromised. This has receive and send information over wireless
led to the creation of the Internet2 con- internet connections. The technologies that
sortium, devoted to the creation of an run the internet such as TCP/IP have allowed
advanced internet, with the intent of fos- companies to base their intranet and inter-
tering new technologies in both hardware net architectures upon one consistent set of
and software that will improve the speed, protocols, and this relieves them of the bur-
reliability, safety, and bandwidth of the dens associated with managing multiple
internet. protocols and their interactions. Languages
As the original internet became deregu- such as XML have revolutionized the way
lated, the US military decided to create its data is transmitted, releasing organizations
own secure ‘‘internet” known as SIPERNET from outdated fixed formats or proprietary
(Secure IP Routing Network). It is intended data formats.
to link all five branches of the military in
one network and possess greater connec- Summary of positive issues
tivity for mobile forces, connecting ship The internet provides an open-standard-
to shore, and field troops to operational based architecture for inter-computer com-
bases and to headquarters through satel- munication. The internet is available uni-
lite and other mechanisms. A second mil- versally via landlines or wireless devices
itary internet was created for the Navy (including satellite phones). The internet
and Marines and is known as NIPERNET protocols have simplified network manage-
(Navy Internet Protocol Router Network). A ment for CIOs and individuals.
third military internet is the JWICS net-
work (Joint Worldwide Intelligence Com- Summary of potentially negative issues
munications System); through it the US The internet has morphed into a commer-
military securely transmits data classified cial, academic, and public-domain space
as top-secret to specific designated recip- that is becoming more congested, and
ients. Naturally, much of the technology, its users are often subject to those with

184
Internet protocol (IP)

malicious intent, e.g., viruses, worms, computer on the internet to ‘‘know” exactly
Trojan horses, phishing, and the use of net- how to communicate with all the others.
work sniffers for those wishing to attempt To make communications practical, a lay-
internet fraud. ered system has been developed, and the
Internet protocol (IP) is one of its fundamen-
References tal components.
r K. Hafner (1998). Where Wizards Stay Up
At the lowest level, every computer has
Late: The Origins of the Internet (New York, the ability to communicate along the kind
Simon and Schuster). of connection that it has, and no other. Typ-
r D. Groth (2003). A+ Complete (Hoboken,
ically, a small-to-medium office will have a
NJ, Sybex--John Wiley and Sons). group of computers all connected together
r P. Gralla (2004). How the Internet Works
on a Local area network (LAN) that con-
(Indianapolis, IN, Que). sists of cheap copper wires and small elec-
tronic components called Switches and Hubs.
Associated terminology: Virtual private
This collection of wires and hubs connects
networks, World Wide Web, Internet
together only the local group: it does not
protocol.
extend to any great distance, and the major
arteries of the internet certainly use much
more sophisticated technology. Since all
Internet protocol (IP) the computers in the local group have the
same kind of connection, they could eas-
Foundation concepts: Internet, Network, Protocol. ily communicate with each other, but not
Definition: The internet is the worldwide commu- with anything outside the group.
nity of computers interconnected by telephone lines, Every local group of computers has one
cable, satellite, etc. The internet protocol is the set of special member, sometimes a normal com-
establishedrulesandproceduresthatmakestheinter- puter with two network connections, some-
net work, by establishing a universal communications times a special-purpose piece of hardware.
language and addressing scheme This device with two network connections
forms a Bridge or Gateway between one LAN
Overview and another, or between a LAN and a larger
The internet exists as a large heterogeneous internet artery. This one device must, of
collection of interacting technologies, and course, be capable of communicating on
is likely to continue to do so for the fore- both of the networks it is connected to,
seeable future. When computers are physi- and they may be using completely differ-
cally close together, a simple, cheap copper ent technologies. The purpose of the IP is
wire connection between them gives excel- to provide a single uniform ‘‘language” that
lent results. When computers are widely allows all of the other computers to make
separated and their owners do not have use of this gateway. A computer on one side
large budgets (home users particularly), of the gateway might be capable only of
a connection through the public utilities communicating along cheap copper wires;
(e.g., telephone lines) is often the only the computers on the other side may (for
viable solution. When major businesses are example) be capable of communicating by a
spread over geographically large areas, no high-bandwidth satellite link, perhaps even
physical connection is viable, so wireless using technology that didn’t exist when the
(satellite) connections are the method of other computers were made. When a ‘‘cop-
choice. All of these different physical con- per wire” computer needs to communicate
nection media work in different ways, and with a ‘‘satellite” computer, they can not
it would not be at all practical for every use their own built-in methods, because

185
Internet protocol (IP)

they are totally incompatible. Instead, they There is some logic to address assign-
use the IP. ment. A large company with many thou-
The IP strictly defines message formats sands of computers may be given all the
that are independent of the technology IP addresses that begin with (for example)
being used. A copper-wire-connected com- 127.35 (known as a Class B address) and
puter composes a message in the IP for- allowed to allocate them as they see fit.
mat, then sends it to the gateway of its A smaller company may be given a Class
own LAN over the copper wires that con- C address, which would give them con-
nect it. The gateway computer receives the trol over all the addresses beginning with
message, and, because the IP is universal (for example) 127.35.101. An individual com-
and invariant, it can understand the mes- puter owner may be given a single full
sage, and see that it needs to be sent to address for their own personal use.
one of the computers on the satellite side. The current IP addressing scheme com-
It is capable of communicating both along bines four small numbers (as in the exam-
copper wire and by satellite, so it simply ple above) to make one numeric address.
‘‘re-wraps” the IP message in the specialized Because each of the small numbers has
satellite message format, and sends it on its a restricted range, the total number of
way. possible IP addresses is something under
Typically, a message transmitted over 4 000 000 000. That seemed an absurdly
the internet will make a few dozen such large number at the time, but is much less
Hops through gateways between LANs and than the current population of the world. If
internet arteries before it reaches its the average number of computers per per-
destination. The IP completely defines the son exceeds about 23 , there just won’t be
universal message format that allows this enough addresses to go round. Since many
to happen, and also provides a universal things that are not really computers are
addressing scheme. now internet-connected, the world has
It is essential that computers on the already nearly run out of IP addresses. The
internet should have fixed, simple addres- next generation of addresses, combining
ses that uniquely identify them, just as 16 small numbers to make one numeric
homes and businesses must have known address, is known as IP6, and is already
addresses to receive their mail. being implemented. This will provide an
The internet protocol uses simple large enormous number of different addresses
numbers as addresses. Usually, these nume- (over 300 000 000 000 000 000 000 000 000 000
ric addresses are seen as four smallish num- 000 000 000).
bers separated by dots (e.g., 127.35.101.98), ATM (Asynchronous transfer mode) is ano-
but this is just a notational convenience. ther routing system, which was originally
It is really just one big number split into intended as an alternative or even replace-
four parts to make it easier (for humans) to ment for IP. Under ATM, data is divided
handle without error. Every computer on up into much smaller packets, all with a
the internet at any time has one of these fixed size (48 bytes of data plus 5 bytes
numbers assigned to it and no other. The of header information), called Cells. Before
IP tells the Gateway computers how to cor- data is communicated between two sites,
rectly forward any message; given the IP a Virtual circuit is set up, which specifies
address of the destination, it is easy for the communications path to be taken; then
any gateway to work out which one of its all data uses that virtual circuit, following
neighbors should receive the message in the same path from source to destination.
order for it to eventually reach its intended These two changes allow greater control
recipient. over the overall transmission rate, and this

186
ISO/IEC 17799

can significantly reduce Jitter (signals break- way, without any specialist or non-standard
ing up and becoming ‘‘choppy”) in audio tools or software.
and video signals. However, improvements
in WAN bandwidth made IP work much Summary of potentially negative issues
more smoothly, and ATM is a very complex The IP was designed at a time when nobody
protocol, so it did not achieve its propo- anticipated the vast number of computers
nents’ goals, and IP still reigns. that would eventually be interconnected.
As a result, its universal addressing scheme
Business value proposition is rapidly running out of addresses to use.
Access to the internet, through the IP, is IP4 (the current version) will have to be
almost a sine qua non for any modern busi- replaced by the already designed and imple-
ness. The use of internet protocols allows mented IP6, or some other alternative.
a clear and simple technology strategy to Although both IP4 and IP6 are well known
be adopted by the whole organization. The and stable, the potential for disastrous
acknowledgement of IP as the standard pro- upheaval as one system is replaced by the
tocol for messaging within and between other is as great as the potential for dis-
organizations allows the internal IT organi- aster was with the Y2K bug. It will proba-
zation to focus its protocol monitoring on bly turn out to be a problem-free transition
just one set of technology protocols. The IP (like Y2K turned out to be), but nothing is
will continue to evolve and advance in the guaranteed until it is all over.
future, and a corporate IT group needs to
References
monitor these changes. IP working groups r http://www.ietf.org/.
have been set up by major software and r W. R. Stevens (1994). TCP/IP Illustrated
technology vendors such as The Internet
(New York, Addison-Wesley).
Engineering Task Force to discuss and mold r D. Groth (2003). A+ Complete (Hoboken,
the shape that the future IPs will take. The
NJ, Sybex--John Wiley and Sons).
IP is fundamental to all devices and soft- r P. Gralla (2004). How the Internet Works
ware that operate over the internet; thus
(Indianapolis, IN, Que).
all vendors will need to comply with the
standard as it evolves, and there are no seri- Associated terminology: Internet, TCP,
ously competing protocols in this area. In DHCP, Network.
the future, a major area of concern for busi-
ness users will be in the transition from
IP4 to IP6, ensuring that devices connect
correctly and that all of the software is ISO/IEC 17799
upgraded.
Foundation concept: Security.
Definition: ISO/IEC 17799 is an international standard
Summary of positive issues intended to provide guidance for IT professionals in
The IP is what makes the internet work. establishing a set of security processes and policies for
All of the positive and negative issues for their organization.
the internet itself are essentially issues for
the IP. Overview
Specific to the protocol is the fact that it The ISO/IEC standard 17799 has its ori-
is very simple and universally established. gins in the British Standard BS7799 origi-
This means that any two computers that nally developed by the UK’s Department of
can be connected together can be expected Trade and Industry. The standard is really
to be able to communicate in a meaningful a code of practice that aims to provide

187
ISO/IEC 17799

a comprehensive set of guidelines that IT policy groups, the delegation of informa-


organizations can follow in order to pro- tion security ownership, and the policies
vide a high degree of information security. required to enable third-party access to cor-
The code of practice itself is comprised porate information systems.
of ten sub-sections that cover The development of policies pertaining
to the operational aspects of the information
(1) Security policy, systems within a corporation is central to
(2) System access control, the code of practice and covers an extensive
(3) Asset classification and control, array of issues, including capacity manage-
(4) Personnel security management, ment, provision of security steps to prevent
(5) Organizational security, malicious attacks upon the network, estab-
(6) Computer and operations lishment of incident logs, and the establish-
management, ment of physical and system-based security
(7) Physical and environmental security, measures to protect intellectual property.
(8) Systems development and The establishment of practices to pro-
maintenance, tect the physical aspects of the informa-
(9) Business continuity management, and tion systems is also important, and the
(10) Compliance management. code discusses areas of concern that need
to be addressed, including the protection of
The issues surrounding security policy are systems using high-security areas, network
centered upon providing information and technologies, physical protection mecha-
data to the organization relating to infor- nisms, and enforcement of equipment use
mation security that can be used to cre- policies.
ate an overall security policy. The creation The guidelines also pertain to systems
of the post of chief security officer may development and maintenance, and stress the
follow the initial assessment of a corpo- need to build into code high degrees of
ration’s information security performance security as well as mechanisms for data val-
level. idation and verification. The code also cov-
The ability to control systems access is vital ers contingency planning and the mecha-
to corporations wishing to be compliant nisms behind compliance. The adherence
with Sarbanes--Oxley and other US regula- to these guidelines can be certified by orga-
tions, which require organizations to docu- nizations such as the British Standards
ment and enforce strict access control. Institute (BSI).
The third factor, asset classification and con- The ISO/IEC code of practice is a very
trol, requires organizations to identify all high level set of guidelines and requires
corporate information assets, and define sec- extensive ‘‘filling in” through other tech-
urity levels for those assets. The code also nical standards and policies such as those
presents guidelines pertaining to person- proposed and developed by the US National
nel policies and security that include issues Institute of Standards and Technology
such as performing background checks (NIST) and published in the Special Publi-
on employees and enforcing non-disclosure cation 800 series. While no code of prac-
agreements, as well as documenting and tice, systems policies, systems, or people
investigating systems errors. for that matter can be considered to have
The code of practice details a set of guide- a zero security risk, the guidelines when
lines relating to the organizational security combined with practical steps can provide
and access to information, including such the first steps toward developing a secure
issues as the establishment of management IT organization.

188
ISP (Internet service provider)

Business value proposition Definition: An organization that provides internet


The need for information security across access to its customers.
the organization is of paramount impor-
tance to all organizations and the ISO/IEC Overview
17799 code of practice provides a frame- Internet access is an expensive commod-
work covering all aspects of information ity. The equipment, staffing, technical, and
security, including the physical issues, soft- administrative requirements for establish-
ware issues, human-resources issues, and ing and maintaining a node on the inter-
the establishment of internal policies and net Backbone are onerous to say the least.
processes. The code of practice describes For a large corporation, it makes financial
the macro issues that need to be addressed sense to set up an in-house access node, but
individually with physical and practical for others the cost is prohibitive. For indi-
processes such that a complete security pol- viduals, the cost of internet access would
icy is enacted and enforced. be completely unaffordable without some
way to share the burden.
Summary of positive issues An Internet service provider (ISP) is simply
The ISO/IEC 17799 provides a code of con- an organization that has invested in its own
duct that helps organizations develop a set internet access node, and sells to others
of security policies and procedures. Compa- the rights to use it. Of course, some means
nies that adhere to the code can be audited of connection between the customers’ com-
and certified by independent entities. The puters and their site must be provided.
code is comprehensive, covering both phys- In some cases customers make a dial-up
ical and intangible aspects of security. The connection using a modem and the pub-
code can be further developed and sup- lic telephone system. In others some high-
ported by combining it with more technical bandwidth or broadband infrastructure
standards and security process models. is used, such as existing cable-television
cables, enhanced telephone lines (such as
Summary of potentially negative issues DSL), or specially installed lines (such as
The standard is very high level and does not
ISDN and T1).
detail specific technologies. The standards
An ISP may simply provide internet
are not legally mandated and adherence is
access and nothing more, but more often it
voluntary.
will provide secondary value-added services
References such as email servers and data backup facil-
r International Standard ISO/IEC ities. Charges may be based on a monthly
17799:2005 Code of Practice for flat fee, a bandwidth usage fee (‘‘per mega-
Information Security Management, byte,” etc.), or a combination of the two.
www.iso.org. ISPs sometimes also provide Hosting services,
r US Department of Commerce (2005). but that is a different but related line of
An Introduction to Computer Security, NIST business.
SP 800--12 (Gaithersburg, MD, US
Department of Commerce). Business value proposition
ISPs provide a mechanism for companies
and individuals to access the internet with-
ISP (Internet service provider) out having to set up their own internet
access node, which is prohibitively expen-
Foundation concepts: Internet, Broadband. sive for all but the largest companies.
Not to be confused with: Hosting. The selection of an ISP is an important

189
ISP (Internet service provider)

consideration for a business. Unfortunately, Other issues that can influence the selec-
in some markets there is a limited choice, tion of an ISP include the scalability of
by regulation, because of limited backbone the ISP’s capacity, since a rapidly growing
access due to incumbents with embed- company may rapidly outgrow a small ISP.
ded ownership, or simply through the What additional services are offered by the
dynamics of supply and demand. When ISP ISP (e.g., email server, firewalls, security)
options are available the cheapest company and how long the ISP has been operational
might not always be the best; similarly are also reasonable considerations. It may
the most expensive might not offer a truly be worthwhile for a company locating in an
superior product and thus metrics beyond emergent market to ask for references from
simply the financial need to be employed. the ISP. However, in mature markets the
Frequently companies allow users a grace costs of switching from one ISP to another
trial period during which they can connect are usually relatively low.
for free. During this period system connec-
tivity tests can be made, such as running
a ‘‘speed-tester” application to ensure that Summary of positive issues
the speed of the connection is as adver- ISPs offer an easy and relatively inexpen-
tised. sive mechanism for connecting to the inter-
Sometimes ISPs advertise the speed of net. There are many ISPs in developed
their systems in asymmetric terms such economies. Costs of switching from one ISP
that downloads are one speed while input to another are low. Many ISPs offer many
from the user is another; if bi-directional additional services beyond connection ser-
speed is important then the true speed of vices.
the system connection need to be estab-
lished. A second issue is the nature of the Summary of potentially negative issues
connection in terms of down-times. Some Some markets have limited ISP access. In
companies offering high-speed connections some markets, ISPs are owned and oper-
such as DSL offer a dial-up backup option ated by governments that regulate and
when DSL is unavailable. The nature and monitor the traffic that can be passed
quality of customer service also needs to through them. The speed of the connection
be established -- if you’re having problems offered by some ISPs may fluctuate depend-
with a connection, will the ISP be there ing upon load and conditions on the
24 × 7 to assist? The contract with the ISP network.
needs also to be examined to ensure that
the ISP adheres to the level of privacy Reference
you expect. While legislation exists in the r T. Casey (2002). ISP Liability Survival
United States, European Union, and many Guide: Strategies for Managing Copyright,
other countries, it is by no means universal Spam, Cache, and Privacy Regulations
and, should corporate secrets or even what (New York, John Wiley and Sons).
could be considered ‘‘free speech” be trans-
mitted through VoIP, strong encryption is Associated terminology: Hosting,
recommended. Internet, Broadband, Network, Modem.

190
Java

utilities, and allow secure execution, guar-


Java
anteeing that programs from untrusted
Foundation concept: Programming language. sources can not do any harm. Surpris-
Definition: A popular object-oriented programming ingly, it succeeded admirably in nearly all
language, derived from C++ but with many signifi- of these objectives, with the result that
cant differences, some of them improvements. Java is probably the second most popu-
Not to be confused with: JavaScript. lar general-purpose programming language
today (some surveys already put it first), and
Overview is still gaining ground.
For many years, C++ was the only pro- Java programs are not compiled in the
gramming language available for large- traditional sense of being translated into
scale programming projects that properly code that can be directly executed on a
supported modern programming method- particular real computer. Instead Java uses
ologies, apart from the very unpopular a Virtual machine (the JVM) designed to
Ada. For all its popularity, which con- provide optimal support for Java-specific
tinues unabated, C++ has many faults, features; Java programs are compiled to
which can not possibly be rectified with- code executable on the virtual machine,
out changing it into a fundamentally dif- called Bytecode, and a JVM emulator applica-
ferent kind of language. C++ has many tion is responsible for executing the code.
safety features, which would go a long As a result, Java programs usually run
way toward making programs more reli- rather more slowly than those written in
able, but none are enforced; programmers more traditional languages, but this differ-
can and do override them at will. C++ ence is almost canceled out by the use of JIT
carries with it the unfortunate legacy of (Just-in-time) compilers, which convert the
older languages (most of the old C language bytecode to normal machine code immedi-
survives as part of C++), and it almost ately before execution.
encourages an unstructured overcompli- The use of a virtual machine to sup-
cated style of programming. The language port execution of programs has two major
itself is exceptionally complex (the offi- advantages. One is that Java programs will
cial standard fills 774 pages), and very few run identically on all computers, requiring
expert programmers are aware of all of its only that the JVM has actually been imple-
rules. mented for that platform; this is an advan-
Java was developed by a small team tage that no other general-purpose pro-
at Sun Microsystems, and released to the gramming language can boast. The other is
public in 1995. It was intended to do that security may be completely controlled.
many good things: remove or replace the A normal executable program, once it has
worst features of C++, be a relatively been allowed to run, has complete control
simple language that can be completely of the computer and can do anything from
known and understood, provide safety fea- functioning correctly to erasing the disk
tures that programmers can not choose drive and sending obscene emails. A vir-
to ignore, fully embody the object-oriented tual machine retains control, and allows
design methodology, provide automatic the user to control exactly which opera-
cross-platform compatibility for all pro- tions are permissible. Java programs can
grams, provide a uniform simplified access be run in a totally secure mode in which
to ‘‘windowy” graphical user interfaces, they are not allowed to access any disk files
be ‘‘internet-friendly,” provide an extensive or make any internet connections, or they
thoroughly tested library of programming can be run in trusted mode, allowing them

191
Java

to do anything, or they can be run at any marked the end of the old mainframe-era
intermediate security setting. languages of Cobol and Fortran. Simulta-
The guaranteed cross-platform compati- neously it moved programmers toward a
bility and control of security make Java more structured and controllable method-
programs very web-friendly. A Java pro- ology of program design.
gram may be embedded into a web page Java also provided an open-source alter-
and automatically run when that page is native to traditional programming environ-
accessed. This allows much greater func- ments associated with C and C++ with
tionality than a normal passive web page which programmers had wrestled for years.
could provide. The program itself is run One of Java’s strengths is that it facili-
through the JVM, so it will work equally tates the creation of powerful and flexible
well on all platforms, and the user retains code for internet-based applications. The
complete control of all security settings, strength of Java for CIOs is based upon the
so they can happily let the program run, portability of the system: the code is com-
secure in the knowledge that it can’t do piled down to a form that is capable of
anything they wouldn’t want it to do. being run on nearly any device (including
Smallish Java programs designed to be some Java-enabled telephones). This style
accessed from web pages in this way are of development alleviated the platform-
called Applets, meaning ‘‘little applica- specific problems that in the past were so
tions,” with the ‘‘-let” suffix as in ‘‘piglet.” problematic for IT organizations. For many
The only serious widely upheld com- organizations, a lack of cross-platform
plaints about Java were that its libraries operability was taken as sufficient reason
were produced perhaps too quickly, and for for staying with an outdated software--
many years the standard libraries remai- hardware platform combination for far too
ned very ‘‘buggy” while new features were long.
being cranked out at high speed. The major A key to the language’s success is the
remaining complaint is that Java very ability of programmers to learn it quickly,
heavy-handedly enforces some extremely especially if they had previous experience
arguable policy decisions: it reports as with C++. Additionally, programmers like
errors things that are common and accep- the availability of software libraries which
ted programming practice, and certainly keeps them from having to ‘‘reinvent the
not wrong; some of the essential libraries wheel” and enables them to focus upon
make decisions that put speed of execu- more specific and important tasks. This
tion above flexibility and clean design. clearly raises productivity and systems reli-
These libraries can not in any practical ability, since many of the library functions
way be replaced by the programmer, and have been improved over the years as pro-
force some very unfortunate implemen- grammers find and document bugs and
tation practices on experienced expert problems.
programmers.
Summary of positive issues
Business value proposition Java is free; the compiler, libraries, and
In 1995 the release of Java coincided with virtual machine for a wide variety of
the deregulation of the internet and these platforms may be downloaded from Sun’s
two events radically changed the way that web site. There are many online resources
organizations thought about their software to help Java programmers, and very many
development. Java and the need for pro- books, including technical references,
grams to address the internet in essence introductions, and academic texts, that

192
JavaScript

explore all aspects of the Java program- Definition:Averybasicprogramminglanguage,based


ming experience. on Java, but much restricted, used to add simple
dynamic content to web pages.
Summary of potentially negative issues
Overview
The official documentation of the required
JavaScript is a simple scripting language
libraries is not always of the highest qual-
barely based on Java. It is rarely if ever
ity. For example, the documentation for
used for any serious programming tasks; its
JDK (Java Development Kit) 1.4.1 and others
main use is for small fragments of execu-
‘‘explains” the ‘‘get preferred size” method
table code (scripts) that are to appear inside
of the component class thus: ‘‘Gets the pre-
normal web (HTML) pages, and make things
ferred size of this component. Returns: a
happen under certain circumstances. The
dimension object indicating this compo-
actions triggered by JavaScript scripts are
nent’s preferred size.” That is not documen-
usually annoying graphical devices that
tation, it is simply repetition. That is not an
many viewers would be happier without,
atypical example; Java programmers must
such as images being replaced by others at
have reliable independent sources of infor-
set periods, to display a succession of adver-
mation. The need to scan multiple sites for
tisements in one place, or cursors changing
clues about how to work around bugs in
shape, or captions changing in front of the
the libraries was for a long time a notable
viewer’s eyes.
drain on productivity.
When used with care, JavaScript can add
The use of Java by some web sites means
positive value to a web page, and can sim-
that a Java virtual machine is required to
plify web development somewhat. It can be
be installed on client systems. Access to
used to create tasteful, unobtrusive graphi-
such a machine requires a visit to the Sun
cal layouts, and to take proper control of
Microsystems web site, since some other
‘‘pop-ups.” It is simply the fact that the
operating systems vendors who have pop-
most basic abilities of JavaScript are so easy
ular browsers do not support the Java ini-
to learn that attracts frequent over-use.
tiative and it can not be downloaded from
JavaScript, being an active element on a
their sites.
web page, is very often a vector for virus
or spyware delivery. If not limited, sim-
References
r J. Gosling, K. Arnold, and D. Holmes ply viewing a page that contains a script
is enough to allow that script to install
(2005). Java Programming Language, 4th
illicit software on a computer and arrange
edn. (Boston, MA, Addison-Wesley
for it to be effectively unstoppable. Web
Professional).
r http://java.sun.com/. surfers must take active control of web
browser security settings, and ensure that
JavaScript is either disabled, or at least for-
Associated terminology: C++, bidden from accessing disk and making its
Object-oriented. own network connections.

Business value proposition


JavaScript JavaScript was created to enable pro-
grammers with little training or exper-
Foundation concepts: Dynamic web pages, Program- tise to rapidly create small dynamic
ming languages. web pages. The applications are typically
Not to be confused with: Java. intended to be graphical in nature and

193
Joint application design (JAD)

add informational or decorative value to Overview


the viewing of a web page. Unfortunately, Joint application design (JAD) originated
the very ease with which these systems with Chuck Morris of IBM Raleigh and
can be created and their ability to deliver Tony Crawford of IBM Toronto in the late
web content via the internet have allowed 1970s and was developed by IBM Canada
malevolent users to create scripts contain- in the early 1980s. It uses a collaborative
ing viruses, spyware, and harmful mistakes. style of system specification development
Hence many users prohibit the acceptance under which the parties involved come
or viewing of emails and other incom- together in structured workshops or ses-
ing data streams that contain JavaScript sions to determine a system’s specification.
elements, and that clearly reduces their Many variants of the original JAD meth-
impact. Network managers and users must ods have been developed by researchers
be made aware of the risks associated with and consulting organizations, but all share
unrestricted scripts, and a security policy a common focus upon the conceptualiza-
should be enforced to diminish their threat tion, design, and specification stages of
potential. the software development lifecycle. The JAD
method breaks the development down into
Summary of positive issues a series of phases.
JavaScript is easy to learn and allows quick The first stage is the Conceptualization
development of basic dynamic web pages. stage (also termed the Project definition stage
and the Requirements gathering phase), in
Summary of potentially negative issues which the project members are identified
JavaScript can be a very inefficient and inef- and then educated on the JAD process. The
fective programming language. It should JAD session then attempts to define the sys-
not be considered a general-purpose pro- tem’s boundaries, the basic processes, and
gramming language and is typically useful high-level data flows. A document is pro-
only for small applications. Hackers have duced that then acts as the requirements
frequently embedded viruses and other document for stage two. Stage two is the
malevolent applications within JavaScript Research phase in which stakeholders (e.g.,
elements in web pages. Consequently function owners, managers, and informa-
many computer users have learned to dis- tion-systems personnel familiar with the
able JavaScript in order to protect their domain area) are interviewed, with the aim
computers; this has the consequence of of determining the project’s parameters.
making the pages unviewable. Having developed a weak specification, the
third stage can be undertaken: in this Prepa-
Reference
r D. Flanagan (2001). JavaScript: The ration stage, the structure documents and
questions to be asked of the team in the
Definitive Guide (Sebastopol, CA, O’Reilly
fourth stage, the JAD session itself, are cre-
Press).
ated. In the fifth stage the specification
Associated terminology: Hypertext. document is prepared.
The JAD session itself (stage four) fol-
lows a detailed agenda and the workshop
Joint application design (JAD) participants implement a predefined set of
procedures that must be closely followed.
Foundation concept: Software-development The participants include but are not lim-
lifecycle. ited to a Facilitator, who is knowledgeable
Definition: Joint application design is a collaborative in the JAD methodology and acts to drive
methodologyforthespecificationofsoftwaresystems. the discussion and ensure that procedure

194
Joint application design (JAD)

is followed in the sessions; Subject matter approach is well known and a wide vari-
experts, who understand the business func- ety of resources is available to support the
tion being addressed by the system; End development approach.
users, who ultimately will use the system;
Developers, software and hardware special- Summary of positive issues
ists who will create the system; a Scribe JAD is a well-known and well-supported
who records all decisions and compiles the method for systems development. It focuses
documentation; and a Tie breaker, who is on the problem and brings all stakeholders
the final arbiter and usually a member of together. JAD sessions help to develop sys-
senior management. Additionally, Observers tems that are supported by good documen-
are usually present but do not partici- tation.
pate in active discussion. The sessions are
Summary of potentially negative issues
intended to be peer-to-peer meetings where
The JAD approach works well for small
participants can express themselves freely
applications, but can incur a significant
regardless of rank.
resource overhead for the developer in
development of larger applications.
Business value proposition
The JAD methodology has been a popu- References
lar approach to systems development, espe- r J. August (1991). Joint Application Design:
cially among consultants. The approach The Group Session Approach to System
can be used for development of new sys- Design (New York, Yourdon Press).
tems and of new procedures that will r J. Wood and D. Silver (1989). Joint
link to existing systems, and as a mecha- Application Design (New York, John Wiley
nism to perform systems maintenance. The and Sons).

195
Knowledge-based systems

Once the knowledge base has been cre-


Knowledge-based systems
ated, it is necessary in the third stage to
Also known as: Expert systems. build an Inference mechanism that is able
Foundation concept: Artificial intelligence. to manipulate the knowledge relating to a
Definition: Knowledge-based systems are programs particular problem under consideration in
that aim to perform at the same level as a human a way that may eventually solve the prob-
expert over a limited knowledge domain. lem. This typically occurs through a struc-
tured dialog between the system and the
user, composed of questions such as ‘‘Are
Overview the lights dim?” The inference engine takes
Systems that could perform at levels requir- the input and uses the knowledge repre-
ing high degrees of intelligence were orig- sented in the knowledge base to arrive at
inally proposed as an aspect of artificial a solution to the problem. The final stage
intelligence (AI) in the 1940s by Alan Tur- in development is the Verification stage,
ing. As general and steady progress was in which the knowledge engineer ensures
made toward this aim during the period that the system works correctly. This is
1950--1980 the concept of a program that a difficult problem because many systems
attempts to replicate the level of perfor- are highly complex and use multiple forms
mance of a human expert over a limited of knowledge representation together with
domain, e.g., a system that helps diagnose specialized reasoning techniques and may
faults in an automobile engine at the same employ fuzzy logic and involve problem
level as a trained mechanic, became pop- areas in which the implementor is not
ularly termed an Expert or Knowledge-based expert.
system. Early systems were created through the
Knowledge-based systems are composed use of AI programming languages such
of a knowledge base, an inference engine, as LISP and Prolog, and during the 1980s
a human--computer interface, and, usually, special computers known as LISP Machines
an explanation mechanism. Various com- were developed by Symbolics and Xerox
plex methodologies have been produced by to support the development of AI and
the research community to assist the Know- knowledge-based systems. Although cus-
ledge engineer in developing their systems, tomized systems were developed through
but the basic process consists of four stages. programming in LISP and Prolog, this was
First, the knowledge engineer elicits knowl- seen as a slow and difficult process and it
edge from a domain expert. This process was eventually realized that the inference
may involve interviews, reviews of manu- engine and interface components could
als, and the examination of case studies. be standardized in significant ways. This
The result of the elicitation process is led to the creation of expert-system Shells,
knowledge in the form of rules, heuristics, which incorporate a general purpose infer-
facts, etc., which in stage two is refined and ence engine that works for any knowl-
encoded in a form known as the Knowl- edge base so long as it is encoded in one
edge representation that may be manipula- specific representation. Knowledge engi-
ted by the inference engine. One popular neers then only have to tune the ques-
knowledge representation is the use of sim- tions that the interface will ask of the user
ple ‘‘if . . . then . . .” rules (e.g., IF light and the responses that it will produce, and
brightness = dim THEN check battery, IF the system is complete. This led to some-
light brightness = dim AND battery = what inflexible systems, but it simplified
charged THEN check alternator). the process considerably and became the

196
Knowledge-based systems

development method of choice for many Summary of potentially negative issues


organizations whether or not they lacked Knowledge-based systems rely upon knowl-
the resources to create customized systems. edge, and the knowledge elicitation process
Successful expert systems are used in a can be long and difficult. It is important
wide range of domains, including the diag- that developers employ rigorous develop-
nosis of system faults and preparation of ment methodologies and techniques. An
schedules for workers, and to assist medi- expert system’s ability to explain its conclu-
cal staff. sions can be weak, and this justifiably inhi-
bits users from trusting results in critical
Business value proposition situations. The knowledge base of a system
Knowledge-based systems are useful pro- may, depending upon the domain, need to
grams that can be used to encapsulate cor- be updated and maintained, incurring a
porate knowledge in a form that may be recurring overhead for the developers.
consulted on a variety of corporate prob- The maintenance aspects of the lifecycle
lems by any user. The knowledge in a also need to include a verification phase to
knowledge base acts as a repository that ensure that no conflicts or redundancies
is helpful to companies that have valuable exist in relation to the specification. A
human experts but whose knowledge can potential problem with knowledge-based
not be duplicated in any other way; the systems is that, since they are computer-
human experts may retire, be overloaded, based, over-reliance on their output with-
or otherwise indisposed, but their knowl- out question can be misleading and at
edge remains available for consultation. times dangerous, especially at the outer
Knowledge-based systems can be used to edges of the system’s knowledge where true
relieve the human expert of routine tasks expertise is needed. A poorly constructed
and allow them to focus upon truly expert system may be fragile and might not
activities. present the user with an assessment of the
validity of the solution proposed. People
Summary of positive issues have a tendency to believe what a computer
The literature on knowledge-based systems tells them: a poorly constructed knowledge-
is well developed, and significant work has based system that provides wrong answers
been performed on design methodologies can be much more harmful than a simi-
that facilitate their specification and cre- larly inept human who gives exactly the
ation. Significant research has also been same wrong answers.
performed on the verification and valida- It is essential to keep in mind that gen-
tion of knowledge-based systems, which is uine expertise is not merely a matter of
an important aspect of any systems design, memorized facts and procedures. A true
especially in the case of systems used expert in most domains is able to solve,
in situations with strong constraints and or at least invent reasonable approaches
requirements (such as the closedown rou- to, problems of kinds that they have never
tines for nuclear power stations). The devel- seen before. This is an aspect of true intel-
opment of knowledge-based systems has ligence, and is outside the scope of any
been eased by the availability of program- knowledge-based system.
ming environments that support the cre-
ation of the knowledge base through tool References
sets, and provide pre-built inference mecha- r R. Plant and R. Gamble (2003).
nisms that operate on a specific knowledge ‘‘Methodologies for the development of
representation. knowledge-based systems 1982--2002,”

197
Knowledge engineer

Knowledge Engineering Review, Volume 18, structed, provide expert levels of perfor-
Issue 1. mance that can be duplicated at multiple
r S. Murrell and R. Plant (1997). ‘‘A survey locations and applied to multiple prob-
of tools for validation and verification lems. The systems can also act as corporate
1985--1995,” Decision Support Systems, knowledge repositories.
Volume 21, No. 4.
Associated terminology: Artificial
Summary of positive issues
The knowledge-based systems community
intelligence, Machine learning, Knowledge
has a mature literature and application
engineer, Logic programming.
tool sets have been developed to support
the creation of such systems. Faster and
more accessible hardware and software sys-
Knowledge engineer tems have facilitated the wider develop-
Foundation concept: Knowledge-based systems. ment and use of knowledge-based systems.
Definition: A knowledge engineer is an information
systems professional who works in the area of arti- Summary of potentially negative issues
ficial intelligence, eliciting knowledge from domain The acceptance of knowledge-based tech-
experts and representing that knowledge in a com- niques remains limited due to the special-
puter system. ized systems requirements and the high
cost of development. Knowledge engineer-
Overview ing is an area of active research and the the-
Knowledge engineers perform the task of oretical basis of knowledge-based systems
extracting information from domain expe- continues to develop as an aspect of artifi-
rts. This process is known as Knowledge elic- cial intelligence.
itation. A variety of elicitation techniques References
are used, including interviews, document r A. Gonzalez and D. Dankel (1993). The
reviews, and structured techniques such as Engineering of Knowledge-Based Systems:
the Repertory grid (a method for extract- Theory and Practice (Englewood Cliffs, NJ,
ing knowledge in which experts compare Prentice-Hall).
the similarities and differences between r B. Gaines, and M. Shaw (1993).
two sets of task-related parameters within ‘‘Knowledge acquisition tools based on
a grid structure). Having elicited the knowl- personal construct psychology,”
edge, the knowledge engineer then per- Knowledge Engineering Review, Volume 8,
forms a careful analysis and structures that No. 1.
information so that it can be represented r G. Kelly (1955). Psychology of Personal
effectively in a data structure on a com- Constructs (New York, Norton).
puter. These data structures are known as
Knowledge representations and include Produc- Associated terminology: Artificial
tion rules, frames, and Semantic networks. The intelligence.
knowledge engineer is also responsible for
maintaining the system and verifying that
the system performs correctly.
Knowledge management
Business value proposition
Knowledge engineers perform the valuable Definition: Knowledge management is the process
task of compiling scarce knowledge and of creating, locating, encoding, and utilizing corpo-
placing that knowledge into a computer rate knowledge to support and enhance corporate
system. The resultant systems, if well con- processes.

198
Knowledge management

Overview development of a knowledge portal. One


The idea of Knowledge management (KM) was such task is operating a call center, where
proposed in the late 1990s as a mecha- an information system or knowledge portal
nism through which organizations could can be developed to support the call-center
capture their tacit and explicit intellectual operator. The portal would support them
property. Tacit knowledge is that informa- by providing scripts, easily accessible solu-
tion or knowledge that is not explicitly doc- tions to the most frequently asked ques-
umented, primarily residing in the heads of tions, information resources that callers
the workforce. Explicit knowledge resides may request, hot links to transfer the cus-
in some physical medium within the com- tomer to another level of support, and a
pany, such as files, process manuals, and knowledge base that contains solutions to
charts. The KM initiative was promoted in documented problems. The system would
part by the corporate business process re- also be self-monitoring, maintaining logs
engineering efforts that were also taking of calls so that improved solutions could
place during the late 1990s, which saw be created in the future.
many firms downsize their workforce only Workgroup collaboration involves provid-
to realize that the workers who were leav- ing support to project teams whose mem-
ing were actually in possession of valuable bers may be located at different physical
assets: corporate knowledge, insight and sites. The system would provide the ability
‘‘know how.” To counter this, many orga- to share documents and also allow mem-
nizations created a senior-level position bers to share knowledge bases that support
of chief knowledge officer (CKO), whose the processes and task being undertaken.
responsibility was to create and utilize a Human capital management KM involves
knowledge base to enable and enhance cor- developing a corporate resource that iden-
porate performance. tifies experts on a specific topic; these
The CKOs quickly found that the pro- experts can then be drawn upon by other
cess of capturing corporate knowledge was individuals as resources to help move a
a difficult and resource-intensive activity, project forward.
and the majority of initiatives were sig- Document management uses Electronic
nificantly reduced in scope. This resulted document management (EDM) software to cap-
in the companies focusing upon key initia- ture, link, and provide version control for
tives, amongst which were process-focused corporate data. Specialist EDM software
KM, workgroup collaboration, document overcomes the problem of corporate data
management, human capital management often being in many diverse data forms and
knowledge bases, and corporate knowledge file types. The software allows any autho-
portals. rized member of the organization to search
Process-focused KM is the creation of data that has been structured by the sys-
resources around the performance of a tem. These are sometimes referred to as
process. This may require knowledge engi- ‘‘enterprise knowledge portals.”
neers to perform knowledge elicitation
upon domain specialists to determine how Business value proposition
they perform their task. This is very similar The use of KM systems allows corpora-
in nature to the tasks associated with build- tions to capture and improve their process
ing a knowledge-based system or expert knowledge, share knowledge, and sup-
system, except that in KM the result may port knowledge-intensive tasks across work-
be a set of process manuals rather than groups. They also allow the structuring
a working knowledge-based system. Some of archival corporate knowledge in an
process tasks are very amenable to the electronic-document system.

199
Knowledge management

Summary of positive issues period of time. Knowledge elicitation and


Knowledge management software systems encoding in a suitable representation
have been developed to support a variety can be an extremely difficult and time-
of initiatives, ranging from specialist pro- consuming task, as the knowledge engi-
cess tools (such as supporting call-center neering community can attest.
operation) to EDM systems that can capture
and manage a wide array of types of cor- Reference
porate knowledge. The effective use of KM r P. Drucker, D. Garvin, L. Dorothy, S.
can increase and speed up access to rele- Susan, and J. Brown (1998). Harvard
vant information by employees, ultimately Business Review on Knowledge Management
reducing the costs associated with perform- (Boston, MA, Harvard Business School
ing a task. Press).
Summary of potentially negative issues
The KM initiatives are expensive and can Associated terminology: Data-flow
require significant support over a long diagram, Knowledge-based system.

200
Law cross-reference

as an adjective (as in ‘‘internet computing”)


LAN/WAN/subnet/internet/intranet
to indicate communication between net-
Foundation concepts: Network, IP (internet protocol). works, as opposed to Intranet, which means
Definitions: communications within one network.
LAN: Local-area network. References
WAN: Wide-area network. r P. Gralla (2004). How the Internet Works
Subnet: The group of computers served by a (Indianapolis, IN, Que).
particular gateway. r D. Groth (2003). A+ Complete (Hoboken,
Internet: The worldwide network of interconnected NJ, Sybex--John Wiley and Sons).
computers.
Intranet: An organization’s internal network of Associated terminology: Virtual private
computers. network, Host, Web services.

Overview
A Subnet is a subset of a larger network con- Law cross-reference
sisting of a number of computers and other
network access devices that have related IP In the United States, national or federal
(internet protocol) addresses, and are all law relevant to computer systems is gen-
connected to the larger network through erally limited to protecting privacy, intel-
the same network bridging device, or Gate- lectual property rights, and computer sys-
way. The existence of subnets is an arti- tems owned by governments or financial
fact of the way the internet protocol works; institutions (although the Computer Fraud
the broadcast network messages that com- and Abuse Act does provide some protec-
prise ARP (the Address Resolution Protocol, tion to corporations). Other matters usually
which allows IP addresses to be associated fall within the domain of state law, which
with particular hardware) will not be trans- of course means that there are more than
mitted beyond a local subnet. However, 50 different versions currently in effect.
the organization of computers into smaller- Those responsible for or dependent upon
than-necessary subnets can be an aid to computer systems should always bear in
network administration. mind that self-protection in the form of
A Local-area network (LAN) is similar in security and vigilance is the only real pro-
concept to a subnet, but less rigidly defi- tection. No matter how strong and fiercely
ned. A LAN is an organizational device that enforced national and state laws may be,
may consist of part of a subnet or a num- they provide no protection against abuses
ber of combined subnets. It is simply the originating from other nations. The idea of
portion of a network that is devoted to one enforceable global laws against spam, child
organization or sub-organization, and gen- pornography, fraud, and software piracy is
erally has some uniform management. completely out of the question for the fore-
A Wide-area network (WAN) is even less seeable future; there will always be law-
firmly defined. A WAN can simply be a less jurisdictions in which computer crim-
group of connected or related LANs, most inals may hide. So long as the law only
commonly the collection of LANs belong- forbids the origination of illegal materials,
ing to a single organization, or it could it will provide very little protection. Only
be any grouping up to and including the forbidding internet service and telecommu-
whole internet. nications companies from relaying illegal
The term Internet is generally used to material would have a real effect, and with
refer to the whole world of connected com- current technology that would be excep-
puters, the World Wide Web. It is also used tionally difficult.

201
Law cross-reference

United States federal law Enacts security and privacy


This section provides no new information; requirements for digital medical
it is a cross-reference to the articles con- systems.
cerned with individual computer-related Privacy Act, 1974,
laws. References of the form ‘‘XX USC YY” Public Law 93--579; 5 USC 522.
are to section YY of title XX of the United Regulates the collection of personal
States Code (USC), where the most relevant data by the federal government.
parts of the law in question may be found. Privacy Protection Act, 1980,
The USC is arranged under 50 titles indicat- 42 USC 2000.
ing the general subject matter of the laws Protects confidentiality of journalists’
included; some laws fit under a variety of sources before publication.
titles.∗ Unlawful Access to Stored
Communications Act,
Cable Communications Policy Act, 1984, 18 USC 2701.
Public Law 98--549; 47 USC 521--551. Guarantees privacy and
Protecting personal information confidentiality for electronic
collected by cable-service providers. communications.
CAN-SPAM Act, 2003, discussed in Spam Voluntary Disclosure of Customer
article, Communications or Records Act,
Public Law 108--187; 15 USC 7701. 18 USC 2702.
Enacts requirements on commercially Guarantees privacy and
motivated email. confidentiality for electronic
Children’s Online Privacy Protection Act, communications and data stored by
1998, computing services.
15 USC 1301; 15 USC 6501--5606. Wiretap Act,
Restricting online data collection 18 USC 2511.
from minors. Forbids tapping of telephones and
Computer Fraud and Abuse Act, 1986, other communications.
Public Law 99--474; 18 USC 1030.
Forbids unauthorized access to UK law
certain computer systems, the release UK Computer Misuse Act, 1990.
of harmful viruses, and trafficking Forbids unauthorized access to
in passwords. computers and the data held on
Computer Security Act, 1987, them.
Public Law 100--235. UK Data Protection Act, 1998.
Requires the National Bureau of Protecting personal information and
Standards to produce computer- restricting its transfer.
security standards.
Digital Millennium Copyright Act, 1998,
US state law
The varieties of state law make it impossi-
Public Law 105--304; 17 USC 1201--1205.
ble to provide any general coverage. We pro-
Extends and strengthens copyright for
vide here references to two of the most sig-
digital materials.
nificant computer-related laws for the state
Health Insurance Portability and
of Florida, simply as an illustration.
Accountability Act, 1996,
Public Law 104--191; 42 USC 1320. 815.06, Offenses against computer users.

Please note that the content of these and other law-related
articles represents the views of the authors who are not
This act makes accessing any computer
legal experts, and the articles should not be read as without authorization, disrupting computer
presenting definitive legal advice. services to authorized users, damaging

202
Legacy system

computer systems, and introducing any port the present and future processes of a
‘‘computer contaminant” (viruses, etc.) into company presents an opportunity to assess
a computer system a third-degree felony (5 the point at which systems will need to
years’ imprisonment, $5000 fine), but if the be replaced or upgraded. Legacy systems
illegal action results in damage valued at can incur high maintenance and support
$5000 or over, furthers fraud, or interrupts
costs and increased costs for the orga-
a public service, it becomes a second-degree
nization as a whole if the technologies
felony (15 years’ imprisonment, $10 000 fine).
employed cannot support changing process
817.568, Criminal use of personal identi- needs.
fication information.
This act makes the fraudulent use of
another’s personal identifying information Summary of positive issues
without consent a third-degree felony, but There are few positive aspects to systems
that increases to a first-degree felony (30 architectures that do not support changing
years’ imprisonment, $10 000 fine) if the processes other than the fact that those sys-
fraud is valued at $50 000 or more, or if it tems may be well known and understood.
involves 20 or more victims. For example, a Cobol system that supports
the purchasing functions may, in a best-
Legacy system case scenario, be completely specified, doc-
umented, tested, and maintained. However,
Foundation concepts: Software, Hardware, Architec- maintaining such a purchasing system for
ture. a very large organization can be extremely
Definition: A legacy system is a systems architecture difficult and a resource drain upon the
that does not support the functional requirements of organization. The ability to run such a sys-
an organization. tem with a deep knowledge and resource
base facilitates an orderly transition to a
Overview more flexible environment such as an ERP
The term Legacy system comes from the fact system.
that a great quantity of program code is
handed down to programmers from previ-
ous generations of programmers; hence it Summary of potentially negative issues
is a legacy of the past. The term is gener- Legacy systems might not support the
ally used to indicate that a system is old, architectural or process requirements of an
but this is an over-generalization. A system organization. At best they will be expen-
needs to be considered in terms of its abil- sive to maintain and potentially deny the
ity to support the current and future pro- company access to the required processes
cesses of an organization; an inability to necessary to support operations. At worst
support changing process requirements is the system could fail or not adhere to legal
now taken as the definition of a legacy sys- requirements (such as HIPAA compliance),
tem. The total architecture needs to be con- and has the potential to put a company out
sidered in respect to the alignment of pro- of business.
cess requirements and requires an assess-
ment of the components of the system’s Reference
r W. Ulrich (2002). Legacy Systems:
architecture: the software, the hardware,
and the network components. Transformation Strategies (Englewood
Cliffs, NJ, Prentice-Hall).
Business value proposition
Assessment of a corporate system’s archi- Associated terminology: HIPAA, ERP,
tecture with respect to its ability to sup- Cobol.

203
Logic programming

very basic facts will establish the partici-


Logic programming
pants in the problem domain. We have two
Foundation concept: Programming language. mice, a cat, and a dog:
Definition: A style of programming based not on giv-
is-a (mouse, jerry).
ing detailed commands, but on describing the rela- is-a (mouse, jerrys-nephew).
tionship between inputs and correct outputs through is-a (cat, tom).
logic formulæ. is-a (dog, spike).

Overview The syntax may seem strange, but that is


The traditional Procedural or Imperative style in the nature of programming languages
of programming, practiced by almost all in general. The meaning is probably clear.
programmers in the commercial environ- Next, some more general facts about how
ment, is based on the notion that the pro- the world works are stated. One essential
grammer should first work out how to solve motivation for this world is that cats eat
a problem, working out an exact Algorithm mice, or, in the terminology of logic pro-
for it, and should then Code that algorithm gramming, if X is a cat and Y is a mouse,
as a complex of detailed instructions that then X eats Y:
the computer will blindly obey. This model
eats (X, Y):- is-a (cat, X),
of programming is so ingrained in pro-
is-a (mouse, Y).
grammers that it is often thought of as
the definition of programming, so that no Again, the syntax is strange, but if ‘‘:-”
other way is even conceivable. is read as ‘‘if,” and the non-parenthetical
Logic programming is another way. It is comma is read as ‘‘and,” it should make
well known to experienced workers in the sense. Other motivating facts are that dogs
areas of artificial intelligence, but rela- chase cats, everything chases anything that
tively unheard-of and almost completely it would like to eat, and one thing likes
misunderstood by the vast majority of pro- another if that other chases away a third
grammers. Prolog is by far the most popular thing that is trying to eat the first:
programming language for logic program-
ming, although others do exist, and Prolog chases (X, Y):- is-a (dog, X),
and logic programming are often seen as is-a (cat, Y).
chases (X, Y):- eats (X, Y).
one and the same thing. In Prolog, no algo-
likes (X, Y):- chases (Y, Z),
rithms are designed, and no instructions
eats (Z, X).
are given. Instead, some facts are stated,
both facts about data and facts about the and that, perhaps surprisingly, is a com-
relationship between inputs and correct plete, albeit simple, Prolog program.
outputs. The Prolog system uses these The program does not tell the computer
facts, according to the ancient established how to do anything; it certainly does not
laws of logic, to transform questions into embody an algorithm. The Prolog interpreter
answers. The programmer must, of course, (the run-time system) knows how to select
take great care to ensure that facts are pre- a sequence of facts to apply in a given situ-
sented correctly, in a way that will allow ation, and how to search for a path to a
the Prolog system to perform the trans- solution. It employs two essential techni-
formations successfully; programming ques that are fundamental to logic prog-
with Prolog requires significant specialized ramming: Unification, which allows it to
expertise. work out which rules are applicable, and
A small example may help to understand replace variables like ‘‘X” and ‘‘Y” with
the logic programming style. First, some solid objects like ‘‘tom” and ‘‘jerry”; and

204
Logic programming

Backtracking, which allows it to try out one ?- likes (X, jerry),


possible path in a computation, but aban- (eats (tom, X);
don it if it does not work out, and try a chases (X, spike)).
different path instead. no solutions.
A user runs the program by asking ques- Of course, nobody likes jerry. This is only
tions. A question may be a simple inquiry a very basic example, but illustrates the
about a fact, such as ‘‘is tom a cat” or ‘‘does style of logic programming. Logic program-
spike eat tom,” and the system will reply ming in general, and Prolog in particu-
with ‘‘yes” or ‘‘no,” thus: lar, opens up a whole new field to simple
?- is-a (cat, tom).
programming and rapid prototyping. Every-
yes. thing that can be done in a ‘‘normal” pro-
?- eats (spike, tom). gramming language can be done in Prolog,
no. and conversely everything that can be done
in Prolog could be done with a traditional
A user may ask a more general question programming language, but in most cases
by asking about a fact with a variable in it. it would be a very complex and error-prone
The system finds all possible valuations of endeavor.
the variable that would result in the fact
as typed being true. We may ask ‘‘who is a Business value proposition
mouse” and ‘‘who eats jerry”: Logic programming systems have been used
extensively in the areas of artificial intelli-
?- is-a(mouse, X).
X = jerry;
gence (AI) known as Knowledge engineering,
X = jerrys-nephew; Intelligent tutoring systems, and Natural lan-
no more solutions. guage processing, since this approach to pro-
?- eats (X, jerry). gramming allows domain knowledge to be
X = tom; represented easily and facilitates the cre-
no more solutions. ation of rule sets that can operate upon
that knowledge base.
The computations work just as well in While logic programming languages
the reverse direction, a feature that is never such as Prolog have been in existence for
available in traditional programming lan- over 30 years, their adoption has not been
guages. For example, ‘‘whom does spike widespread. Logic programming is almost
chase”: exclusively limited to AI applications and
?- chases (spike, X). use as a prototyping mechanism. However,
X = tom; the technology continues to evolve and the
no more solutions. suitability of the approach to large projects
has been enhanced by newer versions of
Questions may even have multiple vari- Prolog with GUIs that are more aligned
ables, as in ‘‘tell me everything you know to ‘‘office”-type users, and by the provision
about who likes whom”: of development environments to assist pro-
?- likes(X, Y). grammers in rapidly developing systems.
X = jerry, Y = spike;
X = jerrys-nephew, Y = spike; Summary of positive issues
no more solutions. Logic programs in some domains are easy
to develop and do not suffer from many
Questions may also involve multiple of the problems associated with procedural
parts, as in ‘‘find somebody who likes jerry programming. Logic programs are easy to
and either is eaten by tom or chases spike”: maintain and modify. There are many free

205
Logic programming

open-source implementations of logic lan- tions have poor input and output facilities.
guages available, and there are many tools Very few programmers are trained in the
and development environments that sup- logic style of programming. Logic programs
port this approach to programming. The can be difficult to interface with other sys-
documentation associated with logic pro- tems, and logic programs are usually much
gramming is extensive. slower than equivalent programs written in
the traditional style.
Summary of potentially negative issues
Logic programs that solve traditional com- Reference
mercial problems can be difficult to write r W. Clocksin and C. Mellish (1984).
and database handling can also be diffi- Programming in Prolog (Heidelberg,
cult due to the fact that many implementa- Springer-Verlag).

206
Machine learning

ization is context dependent and thus vari-


Machine learning
able in nature. One researcher examining
Foundation concept: Artificial intelligence. this problem is Douglas Lenat, who cre-
Definition: Machine learning is the ability of programs ated the Cyc (pronounced as in ‘‘psychol-
to make inferences and expand their understanding ogy”) project in 1984 in order to understand
of the domain in which they operate. and collect common-sense knowledge (for
example, we all know that we pour tea from
Overview a tea pot into a cup in order to make a cup
The term Machine learning is usually asso- of tea, but that depends upon the knowl-
ciated with automated learning or Self- edge that gravity pulls the water down-
programming capabilities rather than the ward and that concave objects like cups are
production of solutions from static knowl- capable of holding fluids). Much of what is
edge bases. Mechanisms for learning under considered to be common-sense is in fact
current research include learning by a very complex web of uncounted essen-
knowledge acquisition, learning by exam- tial facts that are usually learned through
ples, neural networks, case-based reasoning many years of continual experience. Cap-
(CBR), genetic algorithms, and learning by turing all of this implicit knowledge in
discovery. computer-usable form is an exceptionally
Knowledge acquisition is a vital com- difficult task that has not been completed.
ponent of machine learning and is the
mechanism through which a system gath- Business value proposition
ers knowledge, which then has to be rep- Machine learning seemed to hold a sig-
resented internally to provide the basis for nificant degree of promise for several
inference. Major challenges to researchers decades, but has only recently started to
into machine learning lie in each part of deliver techniques that help to solve busi-
this process: how to capture knowledge ness problems. These techniques are gen-
of different types (from pictures, speech, erally task-specific, for example, biomet-
text, diagrams), how to represent each of ric identification techniques that include
these knowledge types, how to integrate facial recognition or signature analysis
them all, and, of course, how to infer use- through pattern recognition, and auto-
ful new knowledge from old. In light of mated telephone-based customer relation-
the multitude of different types of knowl- ship systems using machine learning to
edge that systems need to acquire and improve their understanding of speech and
be able to infer from, many techniques enhance their ability to perform natural
have been developed to assist in this pro- language processing.
cess, including computational learning the- Neural network technologies have been
ory, explanation-based learning, delayed used widely in diverse tasks such as stock
reinforcement learning, temporal differ- market analysis, image recognition, and
ence learning, and using version spaces for extrapolation from examples. Neural net-
learning. works are mathematical programs that are
One of the key problems still facing trained with a set of known inputs and out-
researchers is that of common sense reaso- puts, and adapt themselves to produce the
ning; whereas a child having once touched correct sample output whenever presented
a hot-plate can extrapolate that anything with one of the sample inputs. For some
glowing red is potentially dangerous, this types of problem, they are also able to pro-
ability to generalize an experience is very vide correct solutions for inputs that did
difficult for programmers to build into not appear in the training set. Neural net-
computer systems, because every general- works are used in systems such as character

207
Maintenance

recognition, automomobile warranty anal- Overview


ysis, and detection of credit-card fraud. The activity commonly known as Mainte-
nance is fundamentally the set of processes
Summary of positive issues performed in order to ensure that a system
Machine learning has a long academic performs to specification and will continue
history and some of the techniques have to do so. This term is broadly used to cover
recently become sufficiently mature to be a variety of cases.
embedded in consumer and military prod-
ucts. Machine learning tools resulting from (i) Proactive maintenance ensures that the
academic projects exist and are also avail- quality of a system’s performance will
able from commercial vendors. be at or above the minimum service
level requirements. This type of
Summary of potentially negative issues maintenance may involve both
Many techniques associated with machine hardware and software processes, e.g.,
learning are still difficult to implement adding an extra server to handle a
in large-scale industrial-strength systems; forecast increase in peak system
they are better suited to less wide-ranging demand by users.
problems with limited domains. Some tech- (ii) Reactive maintenance occurs when a
niques such as neural networks are inca- systems failure or emergency
pable of explanation and therefore the threatens the functionality of the
integrity of the solution produced may be system. Again this can be hardware-
difficult to verify. or software-related or a combination
References of both, e.g., a database hard drive
r T. Mitchell (1997). Machine Learning (New crashes and the data needs to be
York, McGraw-Hill). recovered.
r C. Matuszek, M. Witbrock, R. Kahlert, J. (iii) The traditional use of the term
Cabral, D. Schneider, P. Shah, and D. maintenance in relation to software
Lenat (2005). ‘‘Searching for common has been to describe the work a
sense: populating Cyc from the Web,” programmer does to make changes or
in Proceedings, The Twentieth National additions to existing software, e.g., to
Conference on Artificial Intelligence and the change the layout of data on a screen
Seventeenth Innovative Applications of or to add a new section to a generated
Artificial Intelligence Conference, report.
Pittsburgh, Pennsylvania, ed. M. M.
Veloso and S. Kambhampati
(Cambridge, MA, AAAI Press/MIT Press), Business value proposition
pp. 1430--1435. The maintenance function is one that all
organizations have to undertake in one
Associated terminology: Neural networks, form or another. Proactive maintenance
Knowledge-based systems. generally provides a positive return on
investment but requires that organizations
establish and develop metrics, and are
Maintenance proactive in using them. Without a set of
baseline performance metrics, the benefits
Definition: Maintenance is the work performed to of proactive maintenance can be difficult
ensure that systems continue to perform to specifi- to assess. Metrics are also established to
cation. determine the complexity of the code being

208
Memory

developed, since this will then help the IT the expense of setting up a quality assur-
organization not only to determine the cost ance program that establishes baselines
and efforts required to perform mainte- and metrics. Companies can then look at
nance but also to establish better processes processes by which to migrate their systems
and development methods to reduce code to environments that require less, easier, or
complexity in the future, and hence make cheaper maintenance.
the code more maintainable. Maintenance programming is not gener-
The establishment of a software quality ally considered a popular career path for
assurance program helps to ensure that systems professionals, who would rather be
proactive maintenance prevents the need creating new systems than modifying old
for reactive maintenance. Reactive main- ones. Some forms of maintenance program-
tenance is expensive and generally prob- ming are inherently dangerous: some sys-
lematic because programmers and systems tems that have been in existence for over
analysts are under pressure to perform a 30 years will have been modified exten-
‘‘quick fix” in order to re-establish services, sively during that period, and there might
and might not be allowed sufficient time to be no surviving overall structure, nobody
determine the true underlying cause of the who understands the existing system, and
problem. Endlessly fixing the symptoms of certainly nobody who understands how all
an undiscovered software or hardware fault the previous quick fixes might interact. A
can be very expensive. mistake in the maintenance process may
With the move toward configurable sys- lead to complete systems collapse.
tems such as ERP systems, there has been a
Reference
decline in the amount of traditional main- r The Journal of Software Maintenance:
tenance activities, such as modifying legacy
Research and Practice (New York, John
Cobol programs, and this has freed pro-
Wiley and Sons).
grammers to work on more critical core
systems development. Associated terminology: Cobol, ERP,
Software maintenance will never disap- Software metrics.
pear, since systems are always being mod-
ified or upgraded due to changes in the
regulatory or technical frameworks within Memory
which they operate.
Foundation concepts: Central processing unit,
Storage.
Summary of positive issues
Definition: The physical components of a computer
Maintenance is a well-researched area of
system that support long- or short-term retention of
software engineering. Consultants and con-
information.
tract programmers are available to support
a wide range of maintenance requirements,
Overview
including the maintenance of legacy sys-
The need for memory in a computer system
tems that are written in languages long
is absolute. Memory is needed in order to
thought to be extinct.
keep the sequence of commands that the
computer is currently obeying (i.e., the
Summary of potentially negative issues ‘‘code” for the program or application
In order to prevent maintenance from itself) and the data that it is working on.
becoming a never-ending and very expen- All other programs that are available to the
sive activity, it is necessary to undertake computer must also be stored so that they

209
Memory

can be executed on demand; similarly, all etc.), and records just a few lines of inform-
data that those applications could access ation for each interaction, the total amount
must be kept available. of information that must be kept is already
In human terms, the distinction between 100 times the size of the Bible. A large,
intellectual effort (processing or comput- active corporation could easily generate
ing) and remembering (using memory) is trillions of pieces of information.
difficult to see: we have a single organ, Current technology is capable of satis-
the brain, that performs both tasks. Indeed, fying all four of those requirements, but
according to current theories in neuro- not at the same time. The main conflict is
science, there is no difference between the between speed and capacity. Small amou-
two tasks. In a computer system, there is a nts of incredibly fast memory are afford-
sharp distinction that must be understood. able and reliable. Vast amounts of relatively
If a computer is capable of performing slow memory are affordable and reliable.
2 000 000 000 operations per second (which Vast amounts of incredibly fast memory are
is not an unrealistic proposition), that not just unaffordable, but also technologi-
means that it has a Central processing unit cally infeasible.
(CPU) that is capable of obeying 2 000 000 To satisfy all of the requirements, a com-
000 instructions per second. Those instruc- puter system has a complex memory struc-
tions have to come from somewhere: the ture, consisting of many layers of different
CPU does not make its own decisions technology working together to cover all of
about what to do. Those instructions are the requirements, as listed below.
the program or application that is being
run, which was previously created by a CPU registers: a few dozen bytes of
human programmer, and is kept in the memory, physically part of the CPU,
computer’s memory. The computer’s mem- and working at the same speed as it,
ory component must work at incredible holding just the immediate details of
speeds: 2 000 000 000 times per second, the one step of the computation.
CPU will effectively ask it ‘‘what shall I do Cache: around a megabyte of
next?,” and expect to be given the next exceptionally fast memory used to
instruction to obey. hold temporarily just the portion of
In addition to this requirement for excep- an application that is currently
tional speed, there are other conflicting running and the portion of the data
demands. Absolute reliability is essential: a that it is actively processing; this is
single wrong step in a program invalidates copied from slower memory as and
everything it is doing, and an error rate of when it is needed. Part of the cache
just one in a billion would result in two (called ‘‘L1”) is usually part of the CPU
failures every second. chip, and another part (called ‘‘L2”) is
Long-term retention of data is another usually an extra chip soldered onto
essential. Corporate records and databases the motherboard.
must be kept safely for an unlimited Main memory: hundreds, or even
amount of time. Computers must have thousands, of megabytes of fast
memory that continues to work even memory. This usually holds the entire
when the computer itself is turned off or running application and all data it is
damaged. likely to make use of in the near
The fourth requirement is for capacity. future. Applications are copied from
If a company just has 25 000 customers, slower memory into main memory
has an average of only 100 interactions with when execution begins. Main
each customer (sales, purchases, enquiries, memory, cache, and CPU registers are

210
Memory and disk size

volatile, meaning that their contents Business value proposition


are lost when the computer is turned Memory is a critical component and signif-
off. Main memory (sometimes called icantly impacts the performance and relia-
RAM) is usually in the form of one to bility of the whole computer system. Gen-
four replaceable components (DIMMs, erally, the more memory at each of the
SIMMs, etc.), which are plugged into levels, the higher the performance of the
the motherboard. system up to a limit set by the ability of
Disk: hundreds, or even thousands, of the processor to perform its tasks. Gener-
gigabytes of slow memory. Disks are ally, the performance of a computer system
used only for long-term storage, or can be significantly improved by the addi-
when applications need more data tion of memory above and beyond the min-
than can fit into main memory. Disks imal configuration requirements of the sys-
(often called hard disks) are tem. A significant proportion of untrace-
non-volatile, meaning that they retain able problems will just ‘‘disappear” if more
their information even when the memory is made available.
power is turned off.
Backing storage (for example, DVDs and Summary of positive issues
CDs): very slow memory, but of The decreasing cost associated with mem-
unlimited size, since the media ory at all levels has been one of the driv-
(individual disks) can be swapped ing forces behind the production of faster
while an application is running. and more robust systems. Memory devices
are available in many types and configura-
Main memory may be literally a million tions, internal to the computer, removable,
times faster than disk, but is also around redundant backup, and memory sticks to
a thousand times more expensive, and it name just a few. This provides a high
is also volatile. So a real-world computer degree of flexibility in the design of systems
system can not reasonably be made faster and architectures.
by using entirely main memory instead of
Associated terminology: Memory and
disks, or cheaper by using entirely disk
disk size, Disk.
instead of main memory.
Disks may be around 100 times faster
than DVDs, but, whereas DVDs can be Memory and disk size
exchanged in a matter of seconds, chang-
ing a hard disk is a fairly major and slightly Overview
risky operation, to be performed only when A computer’s memory capacity, just like
the computer is off. that of a human being, is finite. If an appli-
Portable memory devices based on Flash- cation is required to ‘‘remember” more
memory technology provide another alterna- things than its memory can hold, it will
tive. These small devices are easily portable, simply not work. Unlike human mem-
and can be plugged into almost any com- ory, the size of a computer’s memory is
puter, through its USB port, without any both exactly quantifiable and changeable.
preparation. They can currently store up If more memory is needed, it can be bought
to a few gigabytes (although that figure and plugged into the motherboard in a
is still increasing), but, due to their slow matter of a few minutes (subject to a max-
speed, they can be considered only as data- imum limit imposed by the computer’s
transport or backup media, and are cer- design).
tainly not an alternative to ‘‘main” memory The total size of a computer’s memory is
or RAM. measured in terms of how many pieces of

211
Memory and disk size

data it can record (this corresponds to the almost invariably a mistake. The metric or
length of the memory), and how big each of S.I. prefix kilo (as in kilogram or kilometer)
those pieces of data can be (corresponding is always written as a small ‘‘k,” so there
to its width). A computer that can store a should be no confusion.
million ten-digit numbers has more capac- From a design point of view, the idea
ity than does one that can store two million of a computer with 1000, 1 000 000 or
three-digit numbers. 1 000 000 000 bytes of memory is absurd,
Since the beginning of the computer era, whereas 1024, 1 048 576, and 1 073 741 824
the width of memory, the size of an indi- bytes would be perfectly reasonable sizes. In
vidual data item, was always called a Word. computer technology 1024 is a ‘‘nice round
Different kinds of computer would have number,” just as 1000 is a nice round num-
different word sizes (e.g., ICL in the 1970s ber to humans. Memory capacity is always
frequently used 24-bit words, giving 7-digit some multiple of 1024 bytes, and, with 1024
data items; DEC/PDP in the same period fre- being so close to 1000, the letter ‘‘K,” being
quently used 36-bit words, giving 11-digit so close to the metric prefix ‘‘k,” was an
numbers). Thus memory size was always obvious choice. 12K is an easy number to
quoted in terms of how many words there remember; 12 288 is not. Originally, when
are and how big a word is (e.g., 49 152 24- only technically proficient people ever dis-
bit words for an ICL-1902). The word Byte cussed memory capacity, there was never
was always used to refer to some portion any confusion.
smaller than a whole word of memory. The other prefixes, M, G, and T, have
Since the supremacy of the desktop com- a less clear standing. They were ‘‘bor-
puter the terminology has changed. In com- rowed” directly from the metric system,
mon usage today, the word byte exclusively where M for ‘‘mega” means 1 000 000, G for
means exactly eight bits (enough to store ‘‘giga” means 1 000 000 000 and T for ‘‘tera”
a two-to-three-digit number) and word, if means 1 000 000 000 000. When somebody
it is used at all, just means two bytes. talks about ‘‘one megabyte” or 64 MB of
The amount of memory in a computer is memory, what do they mean? Although
always given simply as the total number ‘‘mega” very clearly means ‘‘one million,”
of bytes, with no reference being made to no computer ever had one million bytes of
the width. This simplifies comparisons, but memory. What is meant in these cases is
hides useful data: a computer with 64-bit- 1 048 576 (i.e., 1024 × 1024), which is a stan-
wide memory will generally be much faster dard unit of memory size.
than a similar computer with 8-bit-wide
A ‘‘megabyte,” 1 MB, is most commonly
memory.
1 048 576 bytes.
Since a computer usually has a very large
A ‘‘gigabyte,” 1 GB, is most commonly
amount of memory, special abbreviations
1 073 741 824 bytes.
are used for large multipliers; these are K,
A ‘‘terabyte,” 1 TB, is most commonly
M, G, and occasionally T. There is a great
1 099 511 627 776 bytes.
deal of confusion and a certain amount of
deceit involved in the use of these abbrevi- Some manufacturers take advantage of
ations. this unfortunate naming to exaggerate the
The abbreviation ‘‘K” is pronounced sim- capacity of a disk drive by using pre-
ply as ‘‘kay”; it is not ‘‘kilo.” In memory fixes from the metric system in a con-
capacities, a capital ‘‘K” is always used, and text in which the computer-technology pre-
the multiplier that it represents is 1024. A fixes are obviously expected. For example,
computer with 4 KB (‘‘four kay bytes”) has if a disk has a capacity of 3 221 225 472
4096 bytes of memory. The word kilobyte is bytes, an honest label could say ‘‘3 GB,”

212
Middleware

‘‘3072 MB,” or even ‘‘3 221 225 472 bytes.” A these machines were not capable of com-
deceptive label might say ‘‘3.22 GB.” municating without additional software to
The difference seems too small to worry act as an intermediary.
about: 3.22 GB is only just over 7% more Another form of middleware is known
than 3 GB; but the problem can be signif- as a Front-end processor. This is software, or
icant. $10 is just one tenth of one per- even special-purpose hardware, that inter-
cent more than $9.99, but the psychological faces a system with the outside world, per-
effect of the price difference is well known. haps reformatting or condensing inputs,
When confronted with a choice between a or perhaps converting output into a more
3.22 GB disk drive from an unheard-of man- human-friendly form. Front-end processors
ufacturer and a 3 GB disk drive from a rep- are a variety of middleware that makes a
utable one, many purchasers opt for the single system accessible, rather than inter-
3.22 GB drive because it seems to be bet- connecting two systems.
ter value for money, even though it is in Middleware requires specific information
fact exactly the same in terms of capacity, about the protocols used and the formats
and may well be significantly worse value of the data being received from each sys-
in terms of reliability. tem. The development of middleware has
traditionally been a complex task, further
Business value proposition complicated by the need to have differ-
Read the label carefully. A difference of a ent middleware solutions for each partner
few percent in disk capacity will have no computer that might ever be connected.
significant effect, but a difference in relia- Then, when a data format, operating sys-
bility is the difference between successful tem, or communications protocol changed,
productivity and ruinous disaster. the middleware would also have to be
changed; if many different computers were
involved, each change would represent a
major expense.
Middleware Many of the problems associated with
enabling computers and applications to
Foundation concept: Software. communicate that arose during the pre-
Definition: Middleware is code that is used to enable internet era have subsequently been eased
two or more other applications that would otherwise or resolved. It should be noted that the
not be able to communicate to do so. term ‘‘middleware” was not commonly
used until the late 1980s, when distributed
Overview corporate systems architectures started to
The term Middleware is used to denote soft- evolve. Prior to this the term ‘‘systems inte-
ware that enables two other software sys- gration” tended to be used.
tems to communicate and pass data from A primary enabler of easier communi-
one to the other. The need for middleware cation has been the adoption of standard
became pressing in the 1970s, when orga- internet protocols such as TCP/IP, which
nizations commonly ran different applica- provide well-defined languages and media
tions on different machines made by dif- through which computers can exchange
ferent manufacturers, that each had its data. The development and adoption of
own proprietary operating system, data- XML, which defines a uniform format for
base, and network-communication pro- all kinds of data, has freed developers from
tocols (a heterogeneous environment). having to write customized data-conversion
Standardized, system-independent network software for every entity with which their
protocols had not yet been developed, so system communicates.

213
MIS (management information systems) department

The problem of inter-operability has also create middleware in line with a standard
been addressed by the Object Management methodology and specification system.
Group (OMG), whose members focus upon
producing and maintaining specifications Summary of potentially negative issues
for Inter-operable enterprise applications. Their The development of middleware can be a
best-known product is the Common Object difficult and resource-intensive operation,
Request Broker Architecture (CORBA) speci- and the standards, particularly CORBA,
fication, which provides a specification may be very complex.
(through an interface definition language,
IDL) that allows developers to define inter- Reference
r P. Bernstein (1996). ‘‘Middleware: a
faces for their programs using a standard
language that can be understood and used model for distributed system services,”
by any IT organization. Communications of the A.C.M., Volume 39,
No. 2.
Business value proposition Associated terminology: ERP.
Middleware has always been used to enable
computer systems to communicate with
each other. The term Systems integrator is MIS (management information
used for consultants who build middleware systems) department
to interconnect disparate systems. An exam-
ple of this is when an organization pursues Definition: The management information systems
a ‘‘best-of-breed” strategy for an ERP imple- department is the functional entity within a corpo-
mentation. In such a project, a company ration that is responsible for information systems
may buy an HRM system from one company and the alignment of those systems with corporate
and a financial system from another. In strategy.
order for the systems to communicate effec-
tively with each other, the system would Overview
potentially require some middleware to be In the context of a business, the MIS depart-
written. This would require, on the part ment or ‘‘IT organization” is the functional
of the system integrators, specific infor- area tasked with providing the business
mation about the packages, including the enterprise with the technology resources
protocols used and the data requirements necessary to perform its functions effec-
of each. The process of integrating mod- tively. The MIS department, through the
ern systems is simplified when vendors office of the chief information officer (CIO),
adopt open standards and protocols such as provides advice on technological strategy
TCP/IP and XML. It is further aided by ven- to the board of directors and senior exec-
dors’ adoption of the CORBA interface defi- utives of the company, while maintain-
nition language, or a similar technology, so ing current systems, developing systems
that a system’s interface requirements can to meet future needs, ensuring regula-
be clearly understood. tory compliance, developing disaster con-
tingency plans, ensuring security, and sup-
Summary of positive issues porting and enabling the business units
Middleware techniques are well known and divisions.
and supported within the IT community. In the context of academic study, the MIS
The Object Management Group has created department is usually located within the
the CORBA specification, which provides a School of Business and focuses upon provid-
specification (the interface definition lan- ing business-related information systems
guage, IDL) that allows organizations to skills to its students. This differentiates

214
Modem

MIS departments from computer science form the very core of the business’ com-
departments, which focus upon the theo- petitive strengths. The MIS department is
retical aspects of computing and program- also tasked with ensuring systems integrity
ming systems, and computer engineering and security, maintaining and upgrad-
departments, which tend to focus upon the ing current systems, assessing outsourcing
technological aspects of systems. options, working with software vendors,
developing systems directly with customers
Business value proposition and suppliers, complying with regulatory
Corporate MIS departments emerged in the agencies, developing disaster contingency
mid 1980s, evolving from Data processing plans, and supporting and enabling the
departments. The role of MIS departments business units and divisions.
has been transformed in many ways and The study of MIS occurs in the MIS
can be considered to reflect what Venkatra- departments of Business Schools where
man terms a Value center rather than a Cost scholars undertake research into the mul-
center as they were previously considered titude of activities associated with corpo-
to be. rate systems. There are over fifty scholarly
A value-centered MIS organization is journals that publish the research of the
based upon four concepts. Firstly, rather academic community and several profes-
than being a cost center the MIS depart- sional organizations and societies that pro-
ment provides a technological environ- mote scholarship, education, and standards
ment that not only supports corporate pro- in the professional IT community.
cesses but also enables those processes to
References
achieve high operational efficiency. Sec- r N. Venkatraman (1997). ‘‘Beyond
ondly, the MIS department also acts as a
outsourcing: managing IT resources as
service center that provides ‘‘drivers” of
a value center,” Sloan Management
competitive advantage to the organization.
Review, Spring.
A typical example is to think of the help r MIS Quarterly.
desk as a cost center, but in the value- r Communications of the A.C.M.
centered approach it is the contribution of r Information Systems Research.
the help desk to the overall performance r Journal of MIS.
of the enterprise that is considered, not its r Information & Management.
absolute cost. The third dimension of the r European Journal of Information Systems.
value-centered MIS department is its r Journal of Information Technology.
ability to act as an investment center that
attempts to derive new high-yielding busi- Associated terminology: Chief
ness opportunities from existing IT resou- information officer, Enterprise resource
rces and emerging technologies. The fourth planning.
aspect of the value-centered approach is to
consider the MIS department as a profit
Modem
center that provides services and products
to external markets at a profit. Foundation concepts: Binary, Bit, Analog, Bandwidth.
The MIS department is tasked with many Definition: A device that converts digital data signals
responsibilities in the modern organiza- into audio form (and vice versa), enabling transmis-
tion, ranging from providing advice on sion on telephone lines.
technological strategy to the CEO and
Sarbanes--Oxley compliance information to Overview
the board of directors, to the development The word ‘‘modem” is a contraction of
of mission-critical software systems that Modulator--demodulator. Modulation is the act

215
Modem

of modifying one signal, the Carrier, so that stood but most basic facts of information
it carries another signal with it; Demodula- theory. If the carrier signal is cut off at
tion is the reverse: extracting a signal that 5 kHz (quite a high note really, in musi-
was carried on the back of another. cal terms), then no clever inventions will
With a modem, the carrier is an audio get more than 5000 bits of information
signal, an audible whistle, and the signal through per second. High-speed modems
carried is any piece of digital data. All data rely on the fact that data and information
may be encoded as a simple sequence of are not quite the same thing: the data that
1s and 0s: binary. A sequence of 1s and 0s people want to transmit usually has a high
can easily be used to control an audio sig- degree of redundancy, and can easily be
nal by varying its pitch. A 1 is represented compressed into something smaller: 50 000
by one note, and a 0 by another. The elec- bits of personal data could be quite easy to
tronic circuitry required to perform this compress into 5000 bits of real information.
kind of modulation is extremely simple The use of dial-up service over a normal
and cheap. Demodulation, namely detect- telephone line requires a modem at both
ing which note is being received, and con- ends. Broadband connections (DSL, cable
verting it back to a 1 or 0 as appropriate, is modem, etc.) use devices that are called
similarly simple and cheap. modems, and they do modulate and
In essence, that is all a modem is. Digital demodulate signals, but they are not using
data is converted into a series of musical a standard unmodified telephone line; so
notes and back again. The reason for this they are not subject to the same bandwidth
is, of course, the desire to make computers limitations. Direct network connections
communicate over long distances. Setting (ethernet, fiber optics, etc.) use devices that
up a large-scale digital network is a very are never called modems, even though they
expensive and time-consuming task. The do perform the same tasks of modulation
public telephone network was already in and demodulation outside the realm of
existence, and reached almost everywhere, audio signals.
so it provided an obvious solution. Tele-
phone networks can transmit only audi- Business value proposition
ble signals; purely digital data will sim- Modem technology has been an integral
ply not get through, so the modem is the part of computer networks for decades, and
essential intermediary, converting digital today it provides high degrees of reliability
data into a form that can be transmitted at low cost. Modems range from external
on a standard unmodified public telephone devices through which any computer may
network. be connected to an external network to the
Public telephone networks apply band- built-in devices typical of laptops, and pro-
width limitation to all transmitted signals. vide a series of options for systems man-
The highest frequencies are cut off, which agers to select from. Wireless and other
is why the letters S and F can be hard to network connections use a device that is
distinguish (the difference is in the details technically a modem, but almost never so
of the high-pitched hiss), and dog whistles called.
are completely cut off. The reason is sim- While traditional modems are in gen-
ple economics: with bandwidth limitation, eral terms slower than broadband modems,
many signals can be multiplexed to share they are still sometimes useful as a mecha-
the same long-distance lines. nism for interactive situations that require
If the carrier signal has a limited band- a consistently low degree of latency in the
width, then so does the carried signal. This response. A common example occurs when
is one of the most commonly misunder- connecting to a command-line system such

216
Motherboard

as Unix system through a Telnet client. Reference


A broadband connection may go through r M. Banks (2000). The Modem Reference: The
many dozens of routers, and produce a sig- Complete Guide to PC Communications
nificant delay in feedback. A delay of half (New York, Cyberage Books).
a second when downloading web pages or
Associated terminology: Bandwidth,
transferring files is barely detectable, but
Cable, and connectors, Voice over IP.
for an inexpert typist a delay of half a
second between pressing a key and see-
ing the letter appear is disastrous. Dial-up
modems use circuit-switched connections, Motherboard
providing a direct router-free path, and
feedback delays are virtually nonexistent. Definition: An essential circuit board on which CPU,
The lower bandwidth is irrelevant for sim- memory, and required support chips are mounted,
ple human-speed character-mode commu- which also provides connectors for auxiliary or
nications. Modems also provide an addi- optional components, and forms the backbone of a
tional but small level of security: dial out minicomputer or microcomputer.
can be disabled, and caller identification
may be used to ensure that only authorized Overview
accesses are possible. Even though the fourth generation was her-
Modems also provide systems managers alded as whole computers on a single chip,
with a mechanism for remotely dialing into no general-purpose mainstream computer
a system to perform corrective or mainte- actually exists on a single chip. It is not
nance activities should they be required. possible to take (for example) a Pentium
Telephone systems are not directly powered 4 chip, connect it to a keyboard, monitor,
by the public electricity supply, and usu- and power supply, and expect it to work.
ally remain operational during blackouts, Of course, one can expect it to work, but it
so a systems administrator with a battery- won’t.
powered laptop can still connect to central CPUs still require a large amount of
systems during emergencies. external support circuitry just to achieve
the basic required functionality of a com-
puter. Required components such as mem-
Summary of positive issues
ory, disk controllers, clock generators,
Modems are reliable devices with relatively
etc. need intricate carefully designed con-
low costs associated with their purchase
nections between themselves and the CPU.
and maintenance; they provide a valuable
The motherboard is simply a large printed
secondary means of access to systems in
circuit board with all the support compo-
emergencies, or when low bandwidth is
nents soldered in and empty sockets ready
acceptable and fast latency is required.
to receive the CPU and memory chips. The
CPU and memory are not fixed parts of
Summary of potentially negative issues the motherboard because there are too
Modems are limited in speed by the many buyer-selectable options. There are
bandwidth provided to them by the pub- many types and speeds of CPU, and many
lic telecommunications companies which sizes and speeds of memory that could be
carry their signals. Some characteristics used with one uniformly designed mother-
typical of broadband connections may be board.
detrimental to certain activities, such as Additionally, a motherboard has a num-
connecting to remote computers over a ber of vacant connectors ready to receive
Telnet connection. extra ‘‘optional” components. Originally,

217
Motherboard

A motherboard with CPU, video, and memory cards.

disks, graphical displays, printer ports, and


and CPU have already been made by the
network connections were considered to
manufacturer. Mainframe computers gen-
be optional components. They were pur-
erally do not have motherboards, but access
chased as separate items and plugged into
to a mainframe computer will almost cer-
the motherboard’s sockets. As those compo-
tainly be through a normal personal com-
nents gradually became regarded as essen-
puter acting as a workstation.
tial and the potential for miniaturization
For those who decide to save money
increased, it became more and more com-
by purchasing computer components and
mon to provide them as fixed parts of the
assembling them in-house, selection of the
motherboard. It is now quite commonplace
right motherboard is absolutely critical,
for a personal computer or workstation not
probably more important than selection
to require any additional circuit boards
of the right CPU. There is an incredi-
beyond the motherboard. This makes com-
ble number of motherboard manufactur-
puters cheaper to purchase, but does have
ers throughout the world, and some seem
disadvantages: once a motherboard compo-
to offer incredibly good deals. The correct
nent fails or becomes outdated, it can not
design of a motherboard is a very difficult
be replaced unless the whole motherboard
procedure, and very small design flaws will
is replaced.
render a computer totally unreliable. It is
very difficult to make an objective assess-
Business value proposition ment of the quality of a motherboard with-
For many, motherboards are not a consid- out using it extensively for a long period.
eration. Off-the-shelf ready-to-run comput- Most motherboards never get print reviews,
ers are the most common purchase, and and it is often not possible to tell whether
all decisions concerning the motherboard online reviews are written by real users and

218
Multicast

disinterested third parties, or were planted communications involve a direct trans-


or financed by manufacturers or retailers. mission from one sender to one specific
It is also essential to realize that CPUs receiver, and this is how nearly all inter-
come in different, incompatible families. net traffic is delivered. Broadcasting requires
An AMD Athlon (for example) will not fit a common carrier medium, often called an
in a motherboard designed for Intel Pen- Ether; a transmission is put out ‘‘on the air”
tiums or Apple G4s. Motherboards also have by one transmitter, and everything on the
a predetermined range of supported CPU network receives it. Many, perhaps most,
speeds, so, although a 3000 MHz Pentium 4 receiving stations will choose to ignore the
may work in a motherboard designed for broadcast, but it is still available to all.
1000 MHz Pentium 4s, it probably will not Broadcasting is how radio and television
work at full speed. Memory components signals are delivered, and does not exist over
also come in different families, only one the internet. The third mode, Multicasting,
of which will be compatible with any given is used when the effect of a broadcast is
motherboard. This all means that upgrad- desired, and no common carrier medium
ing a CPU often requires upgrading the exists.
motherboard and memory too. There is no common carrier medium
spanning the internet. This simply means
Summary of positive issues that broadcasting is impossible: when one
A uniform basis for computer construction system transmits a message, it can be
allows the CPU, memory, and other com- received only by other stations on the same
ponents to be freely chosen and config- local-area network (LAN); other systems
ured in a ‘‘mix-and-match” fashion. Dam- have no connection to it. Long-distance
aged major components may be replaced transmissions (anything beyond the LAN)
with compatible new components without require repeated retransmission over indi-
having to replace the whole computer. vidual network branches until the desired
receiver can be reached.
Summary of potentially negative issues If a system is required to send the
Different families and speeds of CPU and same message to multiple receivers, it must
memory create expensive compatibility either send multiple exact copies of that
problems, often making expected upgrades message, one to each recipient, or use mul-
impossible. Essential components built into ticasting. Sending multiple copies is clearly
a motherboard can not be replaced if they a very inefficient use of resources, espe-
fail. cially network bandwidth, but it is still a
common implementation choice, because
Associated terminology: CPU, Memory. multicasting is a very difficult technology.
True network multicasting comes in two
varieties. The first is extremely easy and
Multicast efficient to implement. If all recipients
are on the same LAN, there is no prob-
Foundation concepts: Network, Internet protocol. lem. Broadcasting does work over LANs,
Definition: A network transmission that is intended so a message can be sent once, using the
to have more than one recipient. standard Layer 2 Protocol, typically Ether-
net. The other variety is very complex. IP
Overview multicast, the only kind that applies to the
There are three different modes of commu- existing internet infrastructure, requires
nication amongst the members of an inter- that systems adopt a special shared ‘‘mul-
acting network community. Point-to-point ticast IP address”; the transmitter sends a

219
Multicast

message to that address, and all routers and ing, multicasting within a LAN is a useful
gateways must know to forward messages mechanism for distributing data. Multicast-
with multicast addresses in multiple direc- ing provides an efficient and effective tech-
tions. This requires a lot of cross-network nique for corporate audio and video con-
organization; most internet routers do not ferencing, and for online training over the
support multicasting. True multicasting same internal network.
can be relied upon only in a controlled
network environment, perhaps a corporate Summary of positive issues
network in which all routers are subject to Multicasting is a useful, efficient, and effec-
the same corporate policy. Without such a tive technique for communicating amongst
controlled and friendly environment, out- a group of users on a LAN.
of-LAN multicasting is a very complex and
inefficient technology.
Summary of potentially negative issues
Multicasting beyond the confines of a LAN
Business value proposition
is not at all well supported. It is still a devel-
Multicasting is a possible technique for dis-
oping research topic.
tributing a communication amongst the
members of an interacting network com- Associated terminology: LAN (local-area
munity. While the technology for internet network), OSI seven-layer model, Ethernet,
multicasting beyond a LAN is still develop- Internet protocol.

220
Natural language processing (NLP)

ing question would be matched with a set


Natural language processing (NLP)
of input--output templates such as (I’m feel-
Foundation concept: Artificial intelligence. ing X)--(How long have you been feeling
Definition: The development and use of techniques to X?). When a match was found for an input
understand natural language input and to formulate template, the corresponding output tem-
an intelligent response. plate would be printed. Together with a few
simple rules for transforming inputs, such
Overview as You→I, My→Your, etc., and some stock
Since the earliest days of computing, the responses for when nothing matches such
idea of interacting with a machine through as (Do carry on), that was the whole system.
speech has been very appealing. To do so Eliza is exceptionally simple; it does not
requires solving two problems: firstly, a sys- understand what the user is saying in any
tem must be able to take a captured audio meaningful way, and it does not seem pos-
signal and recognize the words within it; sible that such a system could be of any use,
and secondly, the system then has to be but Eliza-like systems proved very provoca-
able to understand what the whole thing tive when they were developed and many
means. Researchers have been working people were initially fooled by the interac-
upon both of these problems and have tion. The simplistic pattern-matching tech-
achieved a certain degree of success in niques used by the system place a serious
each. It is not uncommon for business users limitation on how far the system’s interac-
to interact with airline reservation systems, tions can go. No real attempt is made by the
for example, through an automated voice- system to understand the dialog, since the
recognition system. Many of the more suc- system progresses from one pattern-based
cessful systems bypass most of the first response to another, without any memory
stage by accepting only typed input, not of what was said before.
audible speech. More recently, very sophisticated systems
The origins of Natural language processing have been designed, which do seriously
(NLP) are almost as old as computing itself. analyze user inputs, construct an internal
In the 1950s and 1960s researchers began representation of what they mean, and con-
to develop a variety of techniques aimed at duct directed meaningful conversations.
understanding the ways computers could Such NLP systems are a valuable inter-
be used to provide NLP. One such system, face tool for Expert systems and automated
Eliza, was created by Joseph Weizenbaum, customer response systems. The computer
whose system pretended to be a psycholo- processing of human languages is an excep-
gist with whom a dialog could be enacted; tionally difficult task, as is the representa-
a typical interaction with Eliza might start tion of meaning and the understanding of
like this: a train of thought; no system has yet come
close to the levels of performance expected
Eliza: How are you? of any competent human.
Human: I’m feeling ill. Another early research area was in
Eliza: How long have you been feel- Machine translation, which attempts to auto-
ing ill? matically translate one natural human
Human: Just a few days. language into another. It is immedi-
Eliza: Do carry on. ately clear that simplistic dictionary-based
word-for-word translation does not work,
And the dialog would continue. The secret because different languages have differ-
to the system was that it used pattern ent grammatical structures and vastly dif-
matching, in that the response to the open- ferent idioms. The only way to translate

221
Natural language processing (NLP)

properly is to perform full NLP to under- One aspect of NLP technology that has
stand the grammatical structure of the been deployed in a variety of ways, with
input, and then reverse the process to con- a variety of levels of success, is that of
vert the structure back to speech. Auto- voice recognition. Voice recognition is an
matic translation is still an incompletely extension of NLP in which spoken words
solved problem. in audio format are taken as input, rather
The ability to capture context was stud- than typed text. This adds a whole new
ied by Noam Chomsky, who formulated layer of difficulty to an already intractable
a new theory for the basis of linguis- problem: that of recognizing words, one
tic research and computational NLP sys- of the many tasks that humans handle
tems. His theory is based upon the use with ease but for which no reliable com-
of Generative grammars, constructs used to putational method has yet been discov-
describe how a sentence is formed: e.g., a ered. This is not to say that the technolo-
[sentence] may be decomposed into a [noun gies associated with voice recognition have
phrase] followed by a [verb phrase], the not been commercially exploited; indeed,
[noun phrase] and the [verb phrase] may specialist systems do exist. For example,
then be broken down into other constructs word-processing systems that recognize the
such as [determiner], [noun], [verb], [article], user’s voice commands have been created,
[adjective], [adverb], etc. These constructs although considerable training of the sys-
may then be used to create formal gram- tem is frequently required when the sys-
mars through which an input stream of tem is to be used in domains with specialist
words may be parsed as a first step toward vocabulary requirements.
extracting their meaning. Call centers use a technology known
There have been many types of grammars as Interactive voice response (IVR) to auto-
used within natural language research, mate aspects of their customer service oper-
including Phrase-structured grammars, Trans- ations. These systems work in a dialog
formational grammars, Case grammars, and mode and, to be successful, are required
Augmented transition networks, and parsing is to be very flexible in their voice recogni-
a central task upon which NLP is based. tion component due to the potential range
Typically, commercial systems such as of voices, accents, and dialects with which
those used in customer-relationship man- they have to interact. So far, they can
agement are built to cover dialogs in very successfully handle only the simplest of
limited areas, and the limited vocabu- situations.
lary helps simplify the processes of gram- NLP is also used in systems that automate
mar construction, parsing, and response the reading of customer emails, attempt-
generation. ing to interpret their desires and respond-
ing appropriately. An area of growth for
Business value proposition NLP is that of computer-based training, for
Public perception of NLP has been very pos- example for instruction in a foreign lan-
itive since the days of seeing Robby the guage, mathematics problems, or corporate
Robot from Forbidden Planet and the tinny- processes.
voiced computer of Star Trek responding
helpfully and in perfect natural language Summary of positive issues
to every question posed by the crew. Even The academic literature on NLP and voice
though reality still does not come close to recognition systems is extensive. Many com-
those ideals, and shows no signs of doing so mercially supported systems are available
in the foreseeable future, the general pub- for a wide variety of tasks. NLP systems
lic continues to believe otherwise. are best deployed on automating routine

222
Network

tasks, rather than ones that are potentially Overview


customer-sensitive in nature, and, used as A network is simply a group of comput-
such, can provide a positive return on ers and the equipment that connects them.
investment. It may range in scale from a home setup
consisting of a desktop and a laptop that
Summary of potentially negative issues are occasionally synchronized, to the entire
Voice recognition and NLP systems have not internet.
yet reached the level at which they can The purpose of a network is, of course, to
interact naturally with human users, either enable communications, but there is great
in writing or vocally. Complex data min- variety in what can be achieved by those
ing or open-ended questions are generally communications. A network may be con-
not handled well, and levels of understand- figured to make all workstations appear
ing are adequate only in narrowly specific identical, so that a worker may go to any
domains. The training overhead associated computer at any time, and find exactly the
with voice recognition systems can be high. same files and applications, in exactly the
Interactive voice response systems can have same states. At the other extreme, a net-
a negative impact upon customers’ percep- work may be configured so that every com-
tions of an organization’s level of service puter remains totally independent, and the
and commitment. only network traffic permitted is simple
messages under explicit user control (such
References as emails and ‘‘chat” sessions).
r A. Turing (1950). ‘‘Computing Most networks fall somewhere between
machinery and intelligence,” Mind, the two extremes, allowing individual
No. 59. users to control what is public and what
r N. Chomsky (2002). Syntactic Structures is private on their own computers. Arbi-
(Berlin, Mouton de Gruyter). trary parts of a file system may be made
r J. Weizenbaum (1966). ‘‘A computer to appear to be shared across a network
program for the study of natural (perhaps using NFS, the Network file system,
language communication between man or shared folders), may be made available
and machine,” Communications of the only to specific access requests (by setting
A.C.M., Volume 9, No. 1. up an FTP or HTTP/web server), or may be
r A. Barr and E. Feigenbaum (1981). The shared on a more cooperative basis using
Handbook of A.I., Volume 1 (London, one of the many peer-to-peer file-sharing
Pitman). services.
r R. Plant and S. Murrell (2005). ‘‘A Networks are generally not uniform orga-
natural language help system shell nizations, but rather consist of a number
through functional programming,” of different structures at different levels.
Knowledge Based Systems, 18, 19--35. The entire internet is a single network, but
is also a large number of interconnected
Associated terminology: Machine sub-networks. The usual arrangement is for
learning, Data mining. each organization to have its own LAN
(local-area network) consisting of the com-
puters it uses, interconnected with only the
bandwidth (transmission capacity) required
Network for their own use. Each computer on a
LAN can communicate directly only with
Definition: Any system of interconnected communi- other computers on the same LAN. Except
cating computers. in the few cases where internet access is

223
Network

not desired, each LAN will also incorpo- cheap and easy to build. This is typical of
rate a bridging device (sometimes a com- Thinwire ethernet.
puter with two network cards, sometimes A Star topology comprises a central hub
a special-purpose device such as a Switch, a or Switch into which a cable from each com-
Bridge, or a Router). The bridging device com- puter runs, so that a computer sends a
municates both on the LAN and on the next message to the hub, which then relays the
level of network above. Any computer on message to either all the computers in its
the LAN wishing to communicate outside network, or, in the case of a switch, only
the LAN must have all communications for- the correct one. This configuration is easy
warded by the bridging device. This simpli- to build and is resilient unless the hub has
fies network management (only one device a problem. Most modern LANs have a star
needs to know how to deal with the inter- topology.
net as a whole), and provides a useful secu- A Ring network, sometimes called a Token
rity bottleneck. ring, is a network of machines connected
The bridging devices for a number of in a circle. Messages are sent in one direc-
LANs may be similarly connected together tion around the ring, from one computer to
on a larger-scale network, possibly built the next, until a machine recognizes itself
with higher-bandwidth components since as the intended recipient. If the message
it must carry all inter-LAN and internet makes its way all around the ring, then
traffic, and this larger network has its own the original sender recognizes it and the
bridging device that connects it to the next message is recognized as undeliverable. To
level of network. Ultimately, there will be a control proper sharing of network band-
very-high-capacity connection to the entire width, a dummy message called the Token
internet backbone. LANs are most commonly is passed around the ring when there is no
built on ethernet technology, giving band- real message to be transmitted; a computer
widths from 10 to 1000 million bits per sec- may initiate the passing of a new mes-
ond. Higher-level connections use a much sage only when it receives that token. Ring
wider variety of technology to support the networks are not very robust, since a sin-
much higher capacities needed, includ- gle machine failure will prevent the whole
ing fiber optics, microwaves, and satellite network from functioning.
communications. In a pure physical Mesh topology every
The way in which computers are con- machine is connected to every other, offer-
nected in a network is sometimes referred ing redundancy and fault tolerance; how-
to as the ‘‘topology” of that network. These ever, this is impractical in reality and hence
topologies take their names from the shape Incomplete-mesh topologies are usually built
of the network and include Bus, Star, Ring, instead, in which only a selected few of
and Mesh. the many possible connections are actually
A Bus topology consists of a single cable made. The backbone of the internet has an
along which computers are connected in incomplete-mesh topology.
a long row. A computer communicates by There are several hybrid models that have
putting on the cable a message that is read been created and these are typically Star bus
by all the computers, but acted upon only and Star ring. A star bus uses a bus back-
by the one with the correct address. This bone to link hubs together and run star
topology is effective only when there are configurations from those hubs. A star ring
only a few computers, since only one com- uses a hub as a means of transmitting the
puter can put a message on the cable at message or token around a virtual star con-
any one time. The bus topology is, however, figuration linked into the hub.

224
Network

Business value proposition needed, transmission speeds, response time


The networks associated with businesses requirements, and the number and type
have evolved considerably since the 1970s, of network devices that will be connected
when the majority were based upon main- to it. There is a wide array of hardware,
frame technologies. During this period, software, and systems solutions available
mainframes were connected via their int- and business cases can be developed for
ranet to ‘‘dumb” terminals, and externally varying options, including the total cost
to other companies and remote terminals of ownership involved over the life of the
via modems. During the 1980s, organiza- project, and the scalability of the configura-
tions supplemented their mainframes with tion and the length of time a configuration
personal computers that had their own would take to provide a positive return on
LANs and were also interconnected to the investment.
corporate mainframe and other resources. The decision to create a network also
The 1990s saw the mainframe transition involves the construction of a business case
from being at the center of the network in which total cost of ownership is projec-
to being used for specialist high-volume ted upon the basis of a given usage require-
transaction processing requirements, the ments scenario. This may include issues
mainframe being replaced at the center such as the hardware costs associated with
of the network by server architectures. each configuration option, the support cost
Modern network architectures are typi- of proprietary versus open-source software,
cally composed of many types of comput- the use of thin client versus fat client, and
ing resource, which may include a main- the use of a value-added network versus the
frame, individual workstations and servers, use of the internet for messaging.
industrial computers (robots), sensor-based
technologies (RFID), and remote computing Summary of positive issues
devices (PDAs, laptop computers) connect- Network technologies, including the hard-
ing via the internet. ware, software, operating system, and pro-
In any situation the network topology tocols involved, are mature and supported
or design needs to be considered in rela- by vendors, contractors, and developers.
tion to the goals for that network and, by Networks may be created in a wide variety
extension, the strategic goals of the organi- of configurations from completely homoge-
zation. These requirements may be to sup- neous to completely heterogeneous, and at
port a global ERP system, a super-computer a wide variety of cost levels depending on
performing design work, a LAN of work- requirements.
stations supporting a call center, or for a
company to support all of its computing Summary of potentially negative issues
devices on an extended network. IT organi- Network-related technologies are contin-
zations and network managers use a vari- ually evolving and an upgrade strategy
ety of metrics and benchmarks to deter- needs to be established as part of IT plan-
mine the performance and effectiveness ning. Older technologies and protocols do
of their networks, for example end-to-end become deprecated and sometimes eventu-
response time for a user request to be ser- ally unsupported.
viced by a web server may be vital to an
e-commerce company. References
The creation of a network must take into r L. Peterson and B. Davie (2003).
account the volume of data that is to be Computer Networks: A Systems Approach
passed over it, the bandwidth that will be (San Francisco, CA, Morgan Kaufmann).

225
Network-address translation (NAT)

r W. R. Stevens (1994). TCP/IP Illustrated desire is to provide internet access, but it


(New York, Addison-Wesley). is a very positive aspect from the security
point of view. Systems that can’t be seen
Associated terminology: Client--server,
can’t be attacked, and employees can not
LAN, Internet protocol, TCP/IP, ERP.
spend office time ‘‘surfing the web.”
Network address translation (NAT) allows
computers on an internal network with pri-
Network-address translation (NAT) vate addresses to have controlled access to
the outside world. A NAT server must be
Foundation concepts: Internet protocol, DHCP, Proxy. set up as a member of the internal net-
Definition: Using an intermediate proxy server to work, but with a real IP address. The NAT
change the IP addresses in incoming or outgoing server acts as a bridge between the inter-
transmissions. nal network and the internet as a whole.
It accepts all communications intended for
Overview the outside world, and changes the asso-
The problem of an organization having too ciated addressing information, substitut-
few IP addresses to cover all of its network- ing its own real IP address for the real
enabled devices is only partly solved by sender’s private address, and retransmits
DHCP (Dynamic Host Control Protocol, q.v.). those modified communications on the
DHCP requires that enough IP addresses are external network. When a reply is eventu-
available to cover the number of devices ally received, the reverse procedure is per-
actually connected to the internet at any formed, and the response is retransmitted
given time, sharing them out on an on- on the internal network to the original
demand basis. As the supply of IP addresses sender.
is rapidly diminishing while the number of NAT may be used exactly as described
internet devices is rapidly increasing, DHCP above, to allow a whole network of literally
alone is not enough. millions of computers to access the inter-
There are ranges of IP addresses that net with just one shared real IP address. A
have been set aside for special purposes, similar arrangement allows the NAT server
and are guaranteed never to be allocated to be pre-programmed to redirect incom-
as the true IP address of any computer. ing connections to the appropriate private
These are the 16 777 216 IP addresses that address, so that even publicly accessible
begin with ‘‘10.,” the 1 048 576 that begin servers will be supported.
with ‘‘172.16.” to ‘‘172.31.,” and the 65 536 NAT servers may be configured to mon-
that begin with ‘‘192.168.” Any organiza- itor network traffic and refuse to forward
tion is free to use any of that vast number certain kinds of communications. NAT may
of free addresses, but for internal purposes be used to prevent all but a select few exter-
only. Any data transmitted over the inter- nal sites from being accessed or from hav-
net using one of those addresses is guar- ing access to the internal network, so it can
anteed not to be delivered, but an internal act as a powerful security tool.
network may be configured to handle them
in any way that is desired. Business value proposition
Use of these Private addresses has a conse- NAT servers allow organizations to
quence that may be seen as both positive heighten the security of their network by
and negative: no computer from outside using IP addresses that are not resolvable
the private network will be able to commu- externally (they are not part of the nor-
nicate with these systems (or vice versa, of mal address system used on the internet),
course). This is a very negative aspect if the ensuring that the organization’s computers

226
Network devices

can not be found or sent messages through Overview


the internet. A network device is anything that is con-
NAT servers can also act as security nected to a network and able to communi-
guards, taking a message from a user inter- cate with it. The best known of these are,
nal to the organization and then sending of course, individual computers with net-
it out to an external site, reversing the pro- work adapters (internal network cards or
cess should an incoming message be deliv- devices added to a free port), but most net-
ered. However, the NAT can be configured works depend upon other kinds of net-
so that only authorized addresses can be work device: special-purpose hardware with
visited or received from. This would, for some processing power, but completely devo-
example, help to prevent the plans of a new ted to one network-support task. Nearly all
‘‘secret weapon” being sent to anyone other of these tasks could be handled by a nor-
than those working on the project and mal computer, but it is usually more cost-
holding the clearance ‘‘Top Secret,” if those effective and reliable to use task-specific
plans could be recognized in outgoing hardware.
transmissions. Similarly, because the client A Hub is the simplest kind of network
computer has no resolvable IP address, it interconnection. It connects together two
would be much harder for external spies or more network cables and makes them
to break into the system. behave logically as though they were a sin-
gle cable. Every signal that appears on any
Summary of positive issues one cable is repeated exactly and almost
NAT provides a safe and secure partial simultaneously on all other cables. Hubs
mechanism for protecting corporate com- provide ‘‘signal cleaning,” which allows net-
puters and networks; it can also prevent work cables to cover more distance than
users from reaching external sites. they normally could without excessive sig-
nal degradation, but do nothing to reduce
the levels of network traffic that build up
Summary of potentially negative issues
when too many devices are connected. Net-
Some internet protocols (notably FTP, the
work hubs are almost obsolete, having been
file transfer protocol) do not work well
replaced by Switches, although USB hubs,
through NAT; the consequences of using
which are conceptually similar but not a
NAT must be carefully considered.
piece of network hardware, are current
Reference technology.
r T. Baumgartner and M. Phillips (2004). A switch is an improved or ‘‘smart” hub.
Implementation of a Network Address Switches also connect together two or more
Translation Mechanism over IPv6 network cables, but apply some intelligence
(Washington, DC, Storming Media). to their processing. A signal that is ‘‘heard”
by a switch on one of its cables will be
Associated terminology: Internet, Host, repeated on other cables only if it is rel-
Network. evant to the other devices connected to
them. Unlike hubs, switches require some
setup and occasional management; they
Network devices must be provided with some data on the
structure of the network in order to do
Foundation concepts: Network, IP (internet protocol), their jobs.
LAN. Switches are valuable in reducing net-
Definition: Any device, hardware or software, that can work traffic and increasing the speed of
communicate with a network. communication. The amount of traffic on

227
Network devices

a network increases rapidly as the num- SMTP (email-sending), and IMAP (email-
ber of computers connected increases, until receiving) services are all provided together.
it reaches a point at which the network Similarly, a computer may be both a server
is saturated and performance degrades and a client (for different applications) at
sharply. A switch breaks a network into the same time. When the term ‘‘server”
smaller sub-networks with much less cross- is used to refer to an actual computer, it
traffic. means either a computer that runs some
A Bridge is a particular form of switch server software, or a computer that is sup-
that usually has just two connections, and posedly more powerful than the average
is generally used to connect a Local area workstation and therefore would be more
network (LAN) to a larger grouping. Instead suitable for running heavily used server
of being pre-programmed with information software.
on network structure, a bridge will often
‘‘learn” the network for itself by simply Business value proposition
keeping note of the kind of traffic that New network devices continue to be devel-
occurs on either side. oped and planned: these include routers
A Router is similar to a switch, but that can ‘‘understand” the type of traffic
works at a different level. Switches gener- that they direct and facilitate faster, more
ally decide whether or not to repeat traf- efficient data routing; elevators (lifts) that
fic on the basis of physical identification can transmit their status back to the man-
of the traffic’s destination. A router typi- ufacturer, allowing proactive maintenance
cally looks at the IP address, and may use planning; and refrigerators with bar-code
much more general rules (such as ‘‘all readers that can order groceries over the
IP addresses beginning with ‘111.222’ are internet from preferred vendors.
found along cable A”). Many devices sold The development of network devices
as switches for domestic or office use also that communicate over wireless networks
perform some router functions. also continues to evolve, involving devices
A Gateway is a kind of router that very embedded in cellular telephones, automo-
frequently has the form of a computer biles, PDAs, and laptop computers.
running special software, rather than a
single-purpose smaller piece of hardware. A
gateway is the bottleneck connection bet- Summary of positive issues
ween a LAN and the network as a whole. There is a very large set of options for
The terms ‘‘server” and ‘‘client” properly network developers to build their networks
refer not to an item of hardware, but to from, connecting ‘‘intelligent” devices, such
a particular implementation style for net- as PDAs and cellular phones, together thro-
worked software. A Server is a software app- ugh routers, switches, and gateways.
lication that provides some network ser-
vice; once started, it ‘‘listens” for requests Summary of potentially negative issues
coming from other systems, processes As network device use grows, bandwidth
them, and sends back responses. Servers are demands will continue to grow and tech-
passive systems that wait until some other nology upgrades will continue to be
system specifically requests that they act. required.
A Client is a software application that
makes use of a server in that manner. Reference
Any particular computer may be running r L. Peterson and B. Davie (2003).
a large number of different servers con- Computer Networks: A Systems Approach
currently. Commonly, FTP, HTTP (i.e., web), (San Francisco, CA, Morgan Kaufmann).

228
Neural network

This view of neural networks as a path


Neural network
to artificial intelligence is not universally
Foundation concepts: Artificial intelligence, Cell accepted. A powerful counter-argument is
computing. that duplicating Nature hardly ever pro-
Definition: Artificial devices, namely electronic or soft- duces a viable technological solution. In
ware simulations, that copy the function of brain cells, the early days of flight, many inventors
connected together in large numbers in the hope of produced machines that attempted to fly
duplicating the function of a brain. by flapping their wings, on the grounds
that that is what birds do, and birds fly,
so it must work. It never did; aeroplanes
Overview do not flap their wings, and nor do they
The original goal of artificial intelligence, have feathers or beaks. Lawn-mowers do
and still one of the future hopes for gen- not attempt to emulate goats. Road vehicles
eral computing research, is to construct an successfully use wheels, but nobody has yet
artificial device, a computer or robot, that made a successful legged vehicle. The argu-
behaves intelligently. That is, something ment is that of course successfully dupli-
that can think for itself, making decisions, cating a natural brain would produce some-
having original ideas, and perhaps even thing that has all the qualities of a natural
having feelings and personality (although brain, but that does not mean that it will
the last two might not be so desirable). be an effective solution or an achievable
There were great hopes in the early days of goal.
computing, but they were rapidly dashed In addition, there is the undeniable
when the problem was discovered to be observation that new-born babies do not
much more than just a matter of having make very good workers. The natural brain
enough computing power. takes many, many years of intensive train-
One proposed solution seemed to be fool- ing to produce a usefully intelligent being.
proof: find out how human brains work, A successful duplication of the human
and make machines that do exactly the brain will obviously have exactly the same
same thing. If we have an electronic ver- requirements, plus, of course, exactly the
sion of a human brain, how could it possi- same side effects: personality, desires, and
bly fail to have the same intelligence? It is fallibility. How can anyone expect to avoid
now known in great detail just what a brain all of those human traits whilst still devel-
cell (or Neuron) does; they are extremely oping the same intelligence? This raises
simple in operation, and it is very easy to the further question of why we would
implement very exact copies both electron- need to create artificial intelligent beings:
ically and in software simulations. A Neural when the world’s population is increas-
network or Neural net is exactly that: artifi- ing by about 1 000 000 every day, why put
cial brain cells connected together as they all that effort into making more artificial
might be in a real brain. Unfortunately, ones? The typical answer, that artificially
the human brain contains approximately intelligent systems could be very useful in
100 000 000 000 neurons, each with an aver- environments where humans do not wish
age of 1000 connections (or Synapses) to to be sent or where the environment is
other neurons. It is the enormous number hazardous such as in one-way deep-space
of neurons and synapses, and the dynamic flights, becomes cruel and inhumane if
nature of synapses, that produces intelli- intelligence can not be isolated from the
gence. It will be a long time before human other aspects of being human.
technology is capable of building anything Clearly there is a need to create systems
on that scale. that can behave intelligently but do not

229
Neural network

need to display cognitive behavior, such as Business value proposition


a biometric security system which needs The much hyped use of neural networks
to fulfill only a few clearly statable goals as a mechanism for achieving true artifi-
and need not replicate the full behavior cial intelligence has clearly not succeeded.
or abilities of a human security guard. On However, the nature of the techniques
reducing the scale of the goals, neural net- makes them suitable for many smaller,
works can be used successfully, and are an well-defined application domains. Applica-
effective solution to some kinds of prob- tions have been developed and deployed
lems. Instead of having many billions of in the areas of healthcare, finance, and bio-
neurons to create a whole brain, just a mod- metrics, amongst many others.
erate number is used, to perform just one In healthcare, neural network technol-
small task. With a reasonable number of ogy has been used to help physicians in a
artificial neurons connected together, some variety of areas, including clinical diagno-
receiving stimulation from input sensors, sis and image and signal analysis, as well as
some sending, having their states read by emergent areas such as drug development.
output indicators, and all widely intercon- Specific applications include assisting in
nected, they can, through a process of train- the screening tests for cervical cancer, the
ing, learn to recognize complex patterns detection of acute myocardial infarction,
in data. This process requires no program- the prediction of metastases in breast can-
ming in the traditional sense; neural net- cer patients, the prediction of coronary-
works are trained by repeatedly showing artery disease, and testing for cirrhosis of
them sample inputs together with correct the liver.
solutions. The trainer does not even need Diverse neural network applications have
to have any knowledge of how correct solu- been created in many business areas,
tions can be produced from the inputs, including stock market analysis, futures
since, if there is a pattern of the right and commodities trading, risk analysis, and
sort, a neural network will gradually learn predicting bond price movements. How-
it, and self-configure to produce repeatable ever, many of these systems are proprietary
results and interpolate for input patterns to their developers and open details of
never presented in the training data. their methods are thus scarce. Other busi-
These limited scope neural nets, altho- ness applications include the use of neu-
ugh nowhere near to being an artificial ral networks for detection of credit card
brain, are capable of learning, and do pro- fraud, and for focused direct marketing
duce useful results in some situations. campaigns.
Neural net based systems have been used in The application of neural network sys-
scoring credit applications, optical charac- tems to security problems has included sys-
ter and handwriting recognition, and many tems being developed for facial, voice, and
areas of data mining. A neural net can signature recognition.
never explain its results; inputs that do
not exactly match any that appeared dur- Summary of positive issues
ing training may be correctly interpolated, Neural network technology has a well-
but may also produce totally unexpected developed theoretical basis. The technology
glitches: inexplicable wrong answers. has been applied to systems in a variety
Rarely, the term ‘‘neural network” is used of domains, including healthcare, business,
to describe any network of very simple pro- and security systems. The technology is sup-
cessors cooperating to perform one task, ported by a variety of tools, platforms, and
almost as a synonym for ‘‘cell computing.” software from research labs, universities,
This is not standard usage. and commercial vendors. The use of tools

230
Normalization

greatly simplifies the development of appli- easy to see that 3/7 and 3/7 are the same,
cations and the training of neural nets. but noticing that 27/63 and 3/7 are equal
requires some intellectual effort. Normal-
Summary of potentially negative issues ization provides a single standard represen-
Neural network technologies are applica- tation for all possible values. If all data is
ble to small specific problems rather than represented by its normal form, checking
to systems within large ill-defined problem for equality is as simple as noticing that
areas. The training of a neural network can two things look the same.
take a considerable amount of resources In computing, it can save a significant
and time. A complete understanding of amount of processing time if all data is
neural network technologies requires train- stored in a normal form. No loss of expres-
ing in mathematics and computer science. sivity is incurred, since output filters may
The use of neural networks requires that be applied to ensure that data is printed in
the user understand that the system is not whatever form is desired.
perfect and dependence upon the results For data items more complex than frac-
implies an understanding of the risk level tions, the design of a normal form and
associated with that system; neural net- the procedures for converting data into
works do not have the ability to explain it can be a complex task, but the bene-
their behavior. fits are correspondingly greater. The most
important commercial aspect is in the
References design of Multi-table relational databases. As
r J. Freeman and D. Skapura (1992). Neural an example, consider a small business that
Networks (New York, Addison-Wesley). accepts orders from a number of regular
r J. A. Anderson (1995). Introduction to customers. It would be perfectly sensible
Neural Networks (Cambridge, MA, MIT for them to arrange their database in (at
Press). least) two tables. One table would record
full information on all of their customers
Associated terminology: Machine
(name, address, tax identification, account
learning.
balance, etc.), and another table would
record all the orders (customer name, item
ordered, date, quantity, etc.). Although the
Normalization division into multiple tables is a sound
decision, the given design of those two
Foundation concept: Database. tables is not.
Definition: The conversion or reformatting of data into What happens when a customer changes
a standard, rationalized, uniform representation that their name, an option commonly taken
still provides the same information. by people and corporations? Naturally the
database will need to be updated. Chang-
Overview ing the customer table is a simple oper-
Fractions all have multiple representations: ation, but the order table will also need
1/2, 2/4, 5/10, and 48/96 are all really the to be changed to keep it consistent. Every
same thing. In general use, this does not single record of an order made by the cus-
present a problem, and sometimes has a tomer will need to be updated, and that is
benefit: the notion of ‘‘five out of ten” may a long procedure. While the update is in
be more accurately expressive in some cir- progress, the whole database will be in an
cumstances. However, when a comparison inconsistent state and unsafe for use. The
is required, having many representations design is also unsafe; if the customer name
for the same thing does cause trouble. It is for an order is mistyped, and that error

231
Normalization

goes undetected, there will be no way for require but also allows these activities to
an automatic database system to associate be carried out with precision. The high
that order with the originating customer. technical standards associated with nor-
A set of rules introduced by E. F. Codd in malized systems also help to ensure that
1972, known as Boyce--Codd Normal Form the systems work to specification. This may
(BCNF), provides a normal form for rela- require more effort on the part of the tech-
tional database design. If the tables of a nology team, but ensures that the system
database satisfy the BCNF rules, then the will not have unexpected problems later
problems indicated above, and many oth- and require troubleshooting, which is an
ers, will not occur and the database will expensive and unpredictable activity.
generally make more efficient use of disk
space, provide faster response times, and be Summary of positive issues
more robust. The rules for database normalization have
Normal forms exist for many other been developed and used for over 30
domains. Disjunctive normal form is used years and they provide a formal basis for
for logical formulæ; it allows equivalence database development. In other areas, nor-
to be determined by simple visual inspec- malization of data may be expected to
tion, common operations to be performed increase efficiency and reliability.
quickly, and direct conversion into hard-
ware designs. Backus normal form is used Summary of potentially negative issues
to specify the syntax of programming lan- Normalization can require very specialized
guages, and permits automatic parser gen- expertise or an increased training effort on
eration. Church’s normal form for Lambda- the part of the organization and technolo-
expressions is an essential tool of theoretical gist.
computer science.
Reference
r E. Codd (1972). ‘‘Further normalization
Business value proposition
of the data base relational model,” in
Normalization enables technologists to
Data Base Systems, ed. R. Rustin
develop database systems that are efficient
(Englewood Cliffs, NJ, Prentice-Hall).
and effective in their structure. This not
only facilitates the development, mainte- Associated terminology: UML, Data-flow
nance, and growth of the database as needs diagrams, Entity-relationship diagram.

232
Object-oriented

level operations, without the programmer


Object-oriented
ever having to consider the details of the
Foundation concepts: Programming language, Algo- data again. This resulted in much more reli-
rithm. able program development methods, and
Definition: A programming style that focuses on data is the foundation of Formal methods. There
objects and the operations performed on them as a was great resistance to the adoption of this
whole,ratherthanconcentratingpurelyonexecutable programming style in the established soft-
code and treating data as something that passively ware industry, with ADTs being derided as
flows through it. a mere academic toy, and claims that ‘‘real
programmers use Fortran,” which does not
Overview support ADTs. However, it did eventually
Object-oriented is one of the most popular catch on. Object-oriented programming is
bandwagons for computing professionals nothing more than the practical embodi-
and organizations to jump upon. In real- ment of ADTs.
ity it is a simple logical evolution in pro- Object-oriented programming was first
gramming methodologies that has gradu- presented to the world as an integrated
ally been solidifying since the 1960s. It is usable product in 1972, with the language
certainly a Good Thing, but not necessarily Smalltalk, designed by Alan Kay and Adele
the shiny bright new technology it is usu- Goldberg of Xerox. It became a commer-
ally thought to be. cially usable technique with the intro-
In the older, non-object-oriented style of duction of C++, which took the object-
programming, programmers concentrate oriented style to heart, although it still
on designing programs and algorithms treated it as completely optional. Java was
that will process data. The data is viewed the first major programming language to
as a passive entity that flows through the make object orientation a compulsory part
program, and is worked on by it. Starting of the design process, and that is undoubt-
in a small way with Cobol in 1961, and edly a major factor in its success.
fully taking off with the far more rational The notion of object orientation as a sim-
Algol-68 in 1968, programming languages ple clarifying design technique has been
provided ways of encapsulating large and largely lost on the world, mainly because
complex data items, even whole databases, C++ added so much additional baggage to
as single manipulable objects in a program. it. Some of those additions are undeniably
Programming was still viewed as producing useful, some less so. C++ is the definition
instructions that say what to do with the of object-oriented programming in many
data, but the data could be handled as a programmers’ eyes, so object-oriented pro-
single well-defined object, not as an amor- gramming is often seen as a big, complex,
phous collection of binary digits. and arcane system, which it does not need
These developments gave rise to the idea to be.
of Abstract data types (ADTs) as a tool used The usual extensions to object orienta-
in the early stages of software development. tion can be seen as major improvements in
With ADTs, all of the different kinds of data their own right, but they do have a strong
that a program will have to deal with are complicating tendency that is not strictly
fully and formally defined, and an exact necessary. Inheritance is the technique of
specification is provided for all of the valid programming that allows a whole new data
operations on those data types. Parts of type to be defined by simply specifying how
programs are written to implement those it varies from an existing data type. This
operations, and then the whole program increases code reusability, and in the long
may be defined in terms of those higher- term, reliability. Polymorphism and Virtual

233
Object-oriented

methods are dependent upon inheritance, oriented programming and the language
and make it simpler to define data objects C++ as merely different names for the
that contain within them diverse kinds of same thing, and C++ is a language of
other data objects. This also increases code almost impenetrable complexity. Object-
reusability and reduces testing and main- oriented programming can be clean, clear,
tenance turn-around times. Unfortunately, simple, and understandable, as the orig-
inheritance and polymorphism, although inal object-oriented language, Smalltalk,
simple concepts, bring with them a host showed.
of attendant concerns that seriously com- The advent of the Java programming
plicate programming languages: what hap- language from Sun Microsystems in the
pens when a new data type is defined mid 1990s gave programmers a somewhat
in terms of two incompatible older data more manageable environment in which to
types? What happens when inherited code develop their applications, since it prohib-
operates on data types that had not been ited many of the unsafe constructs freely
defined when it was written? There are available in C++. The availability and use
many other very technical questions that of Java has helped to reduce the complex-
have to be resolved for any system that ity associated with the object approach and
incorporates these ideas. has encouraged programmers to produce
Java solved many of the arcane questions systems in a manner more closely associ-
by simply forbidding the conditions that ated with their original intent, that of facil-
lead to them being relevant. Adherents of itating correct, clearly structured, reusable
C++ claimed that this made the language software.
far inferior to their favorite, but that claim
has not been supported by the evidence
of years of practice. Object-oriented pro- Summary of positive issues
gramming can be a very simple technique, Object-oriented programming allows for
and, when it is, it is of great benefit to all the development of readable, reliable and
concerned: programmers, their managers, reusable code, if it is used properly. The
and their customers. It is part of the logical object-oriented style has an extensive lit-
progression of software technology, in the erature and is supported by many major
same direction as Structured programming, software vendors. Java is becoming more
which also met with great resistance from popular as a language and supports the
industry for a very long time, but is now a object-oriented style of programming.
sine qua non of software design.
Summary of potentially negative issues
Business value proposition
The use of C++ and the poor use of Java
Object-oriented programming was inten-
have in many cases caused complicated,
ded to be a methodology through which
impenetrable programs to be written, espe-
programmers could produce readable, reli-
cially in those instances when program-
able, and reusable code, and was populari-
mers are allowed to believe that complica-
zed by the C++ programming language.
tion is acceptable.
However, while the theory upon which it
was based is sound, the practical use of References
the programming style has been clouded r M. Weisfeld (2003). The Object Oriented
to varying degrees by misunderstandings Thought Process (Indianapolis, IN, Sams).
and unbroken old programming habits. A r J. Rumbaugh, M. Blaha, W. Premerlani,
primary problem is that many see object- F. Eddy, and W. Lorensen (1991). Object

234
OLAP (online analytical processing)

Oriented Modeling and Design (Englewood There are differing degrees of sophistica-
Cliffs, NJ, Prentice-Hall). tion available in the OLAP applications and
r A. Eliëns (1995). Principles of Object the mechanisms through which the data
Oriented Software Development (New York, is stored. The term OLAP was originally
Addison-Wesley). used by Dr. Ted Codd in 1993 and he used
r A. Goldberg and D. Robson (1989). his relational data model (he also created
Smalltalk-80, 1st edn. (Boston, MA, the term Relational database) as the basis of
Addison-Wesley Professional). OLAP. This is still used in some instances
and is known as ROLAP. Subsequently to
Associated terminology: C++, Java,
the relational model, the use of multidi-
Programming language, Algorithm.
mensional databases evolved and the terms
Data cube, Star model, and Snowflake model
were coined to describe the data structures
OLAP (online analytical processing)
underlying the OLAP implementation. The
Foundation concept: Database. OLAP model has been extended to be web-
Definition: Online analytical processing is a set of ana- based or web-enabled (termed WOLAP) and
lytical tools that enable users to examine the relation- this allows users to access the OLAP via a
ships that exist within a structured data set, usually network connection. A fourth level of OLAP
held in the form of a data warehouse. is the use of spreadsheets, which, through
their in-built data-manipulation tool sets
Overview (e.g., pivot tables) and their graphical tools,
OLAP is a collection of analysis techniques can be useful for handling smaller data
that are applied to data sets in order to sets held in relational and non-relational
determine relationships between the ele- structures.
ments of that set. A well-known example is
that a supermarket used OLAP to look for Business value proposition
correlations in sales, and it identified that OLAP systems are beneficial to corporations
customers coming into their stores in the wishing to analyze their data sets. The
evenings just to purchase diapers typically historical accumulation of corporate data
also purchased beer at the same time, leav- provides a potentially rich source of infor-
ing with two items rather than one item. mation for companies. The data is typi-
Knowledge of this unusual combination of cally divided into live transactional data
products purchased allowed the store own- and historical data, namely data that has
ers to achieve better stock modeling and been extracted from the live data set, trans-
demand forecasts. formed into a new data structure, and then
Typically OLAP is used to manipulate and loaded into a separate database (which can
examine financial or sales data models (e.g., be a data warehouse). The OLAP systems
looking at sales at a variety of geographical then manipulate the data to allow for the
levels: sales by country, region, district, or identification of trends and examination
zone) and uses statistical and other tools to of subsets of data, and present the data
identify trends and anomalies. To help in in a variety of formats. OLAP systems have
the analysis, mathematical and statistical several advantages, including the ability
tools are built into the OLAP packages and to add to revenues through better ana-
accessed through a variety of graphical user lysis of sales and marketing, improved
interfaces that allow data to be presented customer satisfaction through enhanced
as graphs, pie charts, linearity charts, and product quality and service, provision of
so on. better, more accurate reporting, and the

235
One-way hash

provision of more timely information resul-


One-way hash
ting in faster decision-making across the
organization. Foundation concepts: Encryption, Security.
The decision to purchase an OLAP sys- Definition: Irreversible encryption, reducing a mes-
tem must be made both through the busi- sage to a small “digest” from which the original can
ness case and with consideration of the not be recovered.
technical issues surrounding the technol-
ogy choices. Issues such as the scale of Overview
the data set that the OLAP can manipu- The idea of irreversible encryption at first
late and the timescales for the manipula- sight seems foolish. What is the point of
tions based upon the underlying platform encrypting something in such a way that
and database server must be considered. it can not be decrypted under any circum-
The OLAP’s compatibility with the database stances, even with the key and unlimited
upon which it operates needs to be taken computing resources? If the data were so
into account, as must the nature of the secret that it must never be seen again, it
GUIs through which the data is viewed. The could much more easily have been secreted
ability to access the data remotely through by simply erasing it.
the internet also needs to be considered, The true purpose of a one-way hash func-
together with the security issues surround- tion is to provide a kind of Fingerprint or
ing any remote access of corporate data Signature for a digital document: some code
warehouses. that can positively identify a document or
data set without being able to reveal it.
Summary of positive issues Take for example the vital message
OLAPs provide tool sets for the manipula- ‘‘HELLO.” Assign each letter a number in a
tion of data sets. OLAPs are available for uniform way (perhaps the simple scheme
a variety of database structures. Various A = 1, B = 2, C = 3, etc.), so the message
vendors provide OLAPs, including specialty becomes 8--5--12--12--15. Then add 1 to the
OLAP tool sets, OLAPs that are integrated as first number, 2 to the second, 3 to the
an aspect of an ERP, and spreadsheet OLAPs third, and so on throughout the message,
for smaller data sets. producing 9--7--15--16--20. Then multiply all
the numbers together: 9 × 7 × 15 × 16 ×
Summary of potentially negative issues 20 = 302 400. Finally, take just the middle
OLAPs require correct ‘‘clean” data in order two digits of the result, giving the number
to provide high-quality outputs. The estab- 24. This is not a secure one-way hash by any
lishment of an OLAP and its maintenance means, but it is sufficient to illustrate the
typically requires a high degree of technical point.
knowledge and specific training. The number 24 provides a digest of the
message. The same message will always pro-
References
r E. Thomsen (2002). OLAP Solutions: duce the same number, but essential infor-
mation was discarded, so it is not possible
Building Multidimensional Information
to work back from the number 24 to dis-
Systems (New York, John Wiley and
cover what the original message was. There
Sons).
r N. Pendse (2004). ‘‘Drilling into OLAP are only 100 different possible results from
this oversimplified operation, but there are
benefits,” DM Review Magazine, March.
certainly more than 100 possible origi-
Associated terminology: Data warehouse, nal messages, so reversing the process is a
ERP, ETL. logical impossibility.

236
Online communities

Just as important is the observation that, Summary of potentially negative issues


if the original message were changed, the There is a concern that many of the algo-
final result would change too. With only rithms that utilize only a small num-
two digits, there are only 100 possible ber of bits may have been broken, which
results, and it would not take long to find would mean that it is possible to produce
a second message that also results in 24. a fake document that produces the same
Real one-way-hash functions produce much hash value as the genuine one. In 2004 it
larger results (MD5 is a prime example, was reported that some popular algorithms
with 30-digit results, and SHA produces 48- including MD4, MD5, and SHA-0 had been
digit results), so finding a second message broken, but the stronger algorithm SHA-1
that produces the same result as any given remains safe, although it is to be phased
first message is practically impossible. out from government use by 2010.
That is the power of a one-way hash. If
you are not concerned with keeping data References
r B. Schneier (1996). Applied Cryptography
secret, but simply want to be sure that
it has not been modified, simply apply a (New York, John Wiley and Sons).
r NTIS brief comments on recent
one-way hash function to it, and keep the
result safe (but not necessarily secret). cryptanalytic attacks on secure hashing
Every time the data is accessed, simply functions and the continued secuity
recalculate the one-way hash and make provided by SHA-1 (http://crsc.nist.gov/
sure that the value is the same. When hash standards comments.pdf).
r NTIS (2002). Secure Hash Standard, 2002,
used in conjunction with a Public-key (q.v.)
encryption system, one-way-hash functions August 1, Federal Information
provide a secure Digital signature mecha- Processing Standards Publication 180--2
nism. (Washington, DC, NTIS)
(http://csrc.nist.gov/publications/fips/
fips180 2/fips180 2.pdf).
Business value proposition
One-way hashing is a technique by which
companies or individuals may ‘‘fingerprint”
their documents or data to ensure authen- Online communities
ticity and prevent changes. The method is
Foundation concept: World Wide Web.
used by some software vendors to allow
Definition: A group of entities that utilizes the inter-
their customers to ensure that software
net technologies to interact to effect a common goal,
applications they receive are the original
practice, or interest.
authorized versions and no third party has
tampered with them en route, perhaps by
Overview
adding some spyware to them.
The term Online community relates to the
concept of an online environment that is
Summary of positive issues composed of members who have a common
One-way hash functions are inexpensive goal or interest. Examples of online com-
to implement and easy to use, and many munities can be found in many areas: Com-
algorithms are openly available. Strong munities of interest, e.g., a financial web site
algorithms provide an almost unbreakable through which participants interact and
mechanism for ensuring the authenticity discuss investments; Communities of relation-
of a document. The algorithms are widely ship, e.g., a forum for cancer-related issues;
published and relatively easy to implement. Communities of fantasy, e.g., communities

237
Open source software

interested in games such as Dungeons and running race, or an event such as a plan-
Dragons or Multi-User Dungeons; and Com- ning meeting). At the other end of the
munities of transaction, e.g., online business- technology spectrum are communities
to-consumer (B2C) and business-to-business built around large-scale technology plat-
(B2B) communities such as Covisint, a pro- forms that enable data synchronization,
curement exchange focused upon the auto- exchange, and procurement between the
motive industry. members.
Online communities use a variety of
levels of technology to execute their inter- Summary of potentially negative issues
actions. These include ‘‘list servers” that Unregulated communities can lead to poor
distribute emails; chat rooms; web sites and unreliable data and information. The
through which emails and messages are overhead required to monitor and maintain
posted; electronic procurement software communities can be high. The costs associ-
systems through which members buy, sell, ated with data synchronization exchanges
and interact; and highly secure, highly reli- and communities can also be high. Compet-
able data-synchronization exchanges thro- itive pressures to join procurement portals
ugh which members distribute the data can force companies to join communities
pertaining to their products and pricing. and be subject to their rules and norms.
The control over access to communities can
be unregulated (as with many communities Reference
of interest, for which activity and access are r R. Plant (2004). ‘‘Online communities,”
self-regulated by the membership), open Technology & Society, No. 26.
but regulated by a moderator, or fully reg-
ulated by a controlling body. Associated terminology: e-Commerce/
e-Business
Business value proposition
An online community can be a power-
ful binding force that can be harnessed Open source software
to drive commerce. Commercial ventures
such as eBay and Amazon started out in Definition: Open source software is any software
the form of online ‘‘communities” within whose source code is freely available, and is typically
which members regulated and monitored not subject to fees or royalties.
trades and recommended books to each
other. List servers can be sponsored or Overview
augmented by the provision of advertising Since the origins of computing there has
and product placement. Electronic procure- always been a culture of ‘‘free software”
ment systems provide a focused industry- available within the computing commu-
or product-specific location for participants nity. The free software has included soft-
to concentrate their efforts, reducing the ware from corporations and individual pro-
need to build their own individual systems. grammers who have placed tools, games,
applications, and ‘‘fixes” (patches) on their
Summary of positive issues systems for users to download via FTP or
There is a variety of mechanisms for pro- the internet. The Free Software Foundation
viding a central location through which (FSF) has been a strong advocate of free soft-
participants can interact. The range of ware since it was established in 1985; its
technologies allows low-cost communities members are ‘‘dedicated to promoting com-
to be created quickly and spontaneously puter users’ rights to use, study, copy, mod-
(e.g., for a specific event such as a marathon ify, and redistribute computer programs.”

238
Open source software

The FSF ‘‘promotes the development and ‘‘open source” and recognized as such (they
use of free software” (http://www.fsf.org/) offer certification signified through a trade-
and is particularly well known for GNU, marked ‘‘OSI Certified” logo) the software
its collection of Unix utilities. Additionally, must satisfy ten criteria:
the Open Source Initiative is an organiza-
tion that provides more structure to the (1) Free redistribution,
provision of ‘‘free” code. (2) Source-code availability,
Open source code is typified by Linux, a (3) Modification to and subsequent
variant of the Unix operating system that redistribution of the code must be
was created mostly by Linus Torvalds, who permitted,
worked on the system in the early 1990s (4) Integrity of the author’s source code,
and finalized the first version in 1994. He (5) No discrimination against any
made this freely available via the internet persons or groups,
to other programmers, who subsequently (6) No discrimination against any fields
added and contributed to the development of endeavor,
of the system, which they made into a (7) Distribution of license,
POSIX-compliant system (POSIX is a set of (8) License must not be specific to a
requirements for a standardized core of product,
Unix). Linux has subsequently been dis- (9) License must not restrict other
tributed under the GNU General Public software, and
License from the Free Software Foundation. (10) License must be technology-neutral.
This allows its source code to be freely dis-
tributed and made available to the general Interestingly, in the early days of com-
public subject to certain conditions (see puting, a large proportion of all software
http://www.linux.org/info/gnu.html). was free. Software would generally only
Open source became a well-known term work on one kind of computer, and, given
as a consequence of the browser wars that the extraordinarily high cost of computers
occurred at the height of the internet rev- themselves, manufacturers would usually
olution in the late 1990s. The battle of the throw in the software as part of the pack-
web browsers gained a high profile dur- age, not having to worry about their com-
ing this period because it is a concern petitors’ customers being able to benefit
central to all internet users and develop- from it. Failure to provide and maintain an
ers. The battle culminated in 1998 when operating system at no extra charge would
the Netscape Communications Corporation have made computers virtually unsellable.
decided to make the client-side source This all changed with the advent of cheap
code of their well-known eponymous sys- personal computers at a price that could
tem available via the internet for free. not possibly support adequate software
Following the Linux and Netscape initia- development, especially when desktop com-
tives the open source community became puters became relatively uniform, so one
more formalized under the auspices of manufacturer’s free software would equally
the Open Source Initiative (OSI) (http:// benefit their competitors.
www.opensource.org/), which was founded
in 1998 as a non-profit corporation ‘‘dedica- Business value proposition
ted to managing and promoting” what they Open source systems allow organizations
have termed the ‘‘Open Source Definition,” to have access to free software for which
which is itself derived from the Debian the source code is available. The sup-
‘‘Free Software Guidelines.” This definition port and quality of the code can vary
states that, for software to be considered considerably from program to program and

239
Operating system

organizations need to be aware of the risks The existence of free software can put
associated with using such systems. How- excessive pressure on traditional software
ever, it is also the case that some com- companies, which may well have been pro-
mercial vendors, rather than funding the ducing a far superior product, but may be
continued support of an aging software unable to stay in business in the face of
package for a small (but sometimes dedi- zero-cost ‘‘competition.” There are enough
cated) group of users, will place the source users who value cheapness over quality and
code in the public domain. This allows enough uninformed corporate buyers who
users and interested parties to work on the never have to use the free software that
code, upgrade it, and maintain it. This has they recommend to make this a serious
the effect of keeping the product’s users concern in many areas.
happy (they no longer have to pay a license It is not totally clear whether the free-
fee) and having the software supported for software movement is really beneficial over-
free by a user group, and as a consequence all. There are situations in which reli-
the vendor’s brand is not weakened by hav- able software is absolutely critical, and
ing disenchanted ex-customers. the typical free-software distribution agree-
The open-source movement has devel- ment about providing no warranties and
oped and maintains products in a very wide accepting no responsibility is inadequate.
range of product areas, including operat- If skilled programmers are expected to give
ing systems, word processing, spreadsheets, away the fruits of their labors for free, they
databases, and internet browsers. This has will not be able to devote themselves pro-
allowed some companies to use exclusively fessionally to the project.
free-source software in their selection of
products and thus is highly favored in orga- References
r Free Software Foundation, 51 Franklin
nizations and countries that can not afford
proprietary vendor-managed software. Street, Fifth Floor, Boston, MA 02110,
USA.
r http://www.fsf.org/.
Summary of positive issues r http://www.opensource.org/.
The Open Source Initiative attempts to
bring professionalism and standardization Associated terminology: Linux (see Unix).
to the domain of open-source systems.
Open-source software is of low cost and fre-
quently under continuous development by
user groups. Open source allows users to Operating system
read, modify, and develop the source code
of an application. Foundation concept: Software.
Definition: Essential, permanently running software
Summary of potentially negative issues that keeps a computer operating.
Open-source code can vary in quality and
might not ever have been rigorously tested. Overview
The code might not have any documen- In the old days, running a computer
tation or specifications associated with it. required a full-time staff. Programs had to
Companies using open-source code need be ‘‘fed in” manually from decks of cards
to ensure that the code is actually legally or magnetic tapes, then the input data
open-source code rather than just copied had to be made available in a similar way,
code that has been released under another and everything had to be tended carefully
name. until the final results could be extracted,

240
Operating system

and another program could be started in Operating systems today often present
its place. two alternative faces to users. The tradi-
As more convenient forms of storage, tional interface involves typed commands
principally disks, became available, the and responses, and is familiar to many in
requirement for human operators became the form of DOS (Disk Operating System)
an intolerable bottleneck, wasting far too or the ‘‘Dos shell.” By modern standards,
much of the computer’s still very expen- DOS is a very basic job-control language,
sive time. The solution was to make the capable of running only one application
computer operate itself. For each program at a time, and arranging for its own
run, a set of operating instructions would execution to be continued only if nothing
be written, exactly stating which program goes wrong, after an application has fin-
should be run, which files contain the data, ished. ‘‘Industrial-strength” operating sys-
what to do if anything goes wrong, and tems usually have a similar typed com-
what to do with the results. These instruc- mand interface, known as a Shell, Monitor,
tions were written in a form that could be or Command-line interface, that is gener-
understood and obeyed directly by the com- ally preferred by technical users. The now-
puter, usually called a Job-control language. discontinued TOPS and VMS, together with
The job-control instructions could even be all varieties of Unix, follow this pattern.
stored in a disk file, completely automat- The alternate interface is the now univer-
ing the whole process, relegating human sal Windowing environment, or GUI (graphi-
intervention to disaster recovery and paper cal user interface), preferred, or, more pre-
loading. cisely, demanded, by non-technical users.
In order for the computer to be able Usually this windowing environment is just
to understand the job-control instructions, a secondary interface running on top of
some programming is required. Software a traditional primary one, although the
must read and interpret the instructions, popular Windows operating systems have
and carry out the actions indicated. This abandoned this model and reversed the
software was the earliest form of operat- situation.
ing system. When a computer is first started Since these early beginnings, the scope,
up, some special action is required to load complexity, and error-proneness of oper-
and start execution of the operating sys- ating systems have all grown enormously.
tem’s software. From then on, it is self- Typical operating systems provide a large
maintaining; the operating system never library of standard program components
relinquishes control until the computer is to simplify the software-implementation
turned off. Of course, all software can (and process. Multiprogramming allowed more
usually does) have design flaws; an error than one application to be runnable con-
in the operating system can have a wide currently, so that, when one is temporarily
range of effects, from minor annoyances to unable to continue, perhaps waiting for a
introducing subtle and undetected errors slow input device, another could run in its
into the results of programs and destroying place. Multiprogramming grew into time-
data. Some kinds of error will simply cause sharing, the system under which multiple
the operating system to stop running, at applications do in fact run concurrently,
which point the computer becomes unus- with control switching between them per-
able (‘‘freezing up”) until it is restarted; this haps 100 times per second. Virtual mem-
is known as a System crash, and is some- ory, which allows more applications to run
times visible as the infamous ‘‘blue screen than would actually fit in the computer’s
of death.” memory, is another system that requires

241
Operating system

constant operating-system supervision. Net- system across all these devices benefits the
work access requires oversight to ensure organization by potentially reducing costs,
that incoming messages are delivered to through a reduction in maintenance, train-
the right application. The ability to inter- ing, and license expenses.
pret the movement and clicks of a mouse A second strategy organizations may
and to provide a visual ‘‘desktop” requires wish to pursue is the ‘‘best-of-breed” strat-
much more computational effort than egy, whereby one operating system is
is needed with a traditional job-control selected to work on the server cluster,
language. another on the desktops, and yet others
The full, efficient, and effective use of on the laptops, PDAs, and cellular tele-
an operating system requires a significant phones. Alternatively, a low-cost strategy
degree of training. Many science-fiction may be appropriate, implying that free-
books and films depict operating systems ware and open-source systems can be cho-
as intelligent agents (e.g., the HAL-9000 on sen with less regard to their applicability
the spaceship in the film 2001: A Space or usability.
Odyssey and the computer on the Star Trek A major concern in operating system
spaceship Enterprise). While it is possible selection is the level of support associ-
to have natural-language voice interfaces, ated with a particular system and ven-
the ‘‘cognizant” systems in these films are dor. The metrics associated with support
still a long way off; their study is an level include the cost of the license and
ongoing aspect of research into artificial the associated maintenance levels (if any),
intelligence. the cost associated with service packs
(upgrades and fixes), user groups and user
Business value proposition community, conferences, technical papers,
The selection of an operating system and the level of support from application
depends upon the use to which it is to be vendors.
put, the knowledge level and skill of the It is also important to consider the rela-
operator, and the resources available to sup- tionship between the operating system and
port that system. There exists a range of the applications that are to be supported
operating systems to meet a user’s require- by that system. This is important in the
ments depending upon the configuration overall design of the systems architec-
of these variables. One aspect of operat- ture (the total combination of applications,
ing system selection that may override the operating systems, networks, etc.) because
others is the maintenance of consistency not all operating systems can run all appli-
across the organization. Hence an organi- cations or database systems. Equally, some
zation may choose to select a GUI-based applications such as ERP systems may run
operating system that a majority of the in a thin client mode, making the operat-
users will find easier to understand than ing system on the client almost irrelevant
a command-line interface. Versions of that so long as the client can run a browser.
system may be used across a range of plat- A further issue that is important in many
forms supporting a variety of applications. purchasing decisions is the lifespan of the
At one end of the spectrum one version operating system, and the nature of any
may support the corporation’s ERP system, superseding operating system. The nature
while clients may run another version of of the technology upon which an operating
the operating system, portable laptops a system works will be one factor that deter-
third, and hand-held devices such as PDAs mines whether the operating system needs
and cellular telephones yet another ver- to be changed to stay in alignment with
sion. The use of one ‘‘family” of operating the technology in use (e.g., CPU, memory

242
Optical character recognition (OCR)

systems, network requirements, monitor mal user groups and are considered legacy
and peripheral devices). A second factor is by their original manufacturer. Some oper-
the amount of support (and cost) that a ating systems are not backward-compatible
vendor provides to transition to the next and thus do not support the applications
version. A third factor is the ability of that ran on previous versions.
new versions of the operating system to be
Reference
backward-compatible so that all the previ- r A. Silberschatz, P. Galvin, and G. Gagne
ous applications will continue to run.
(2004). Operating System Concepts (New
Simpler items of machinery, such as
York, John Wiley and Sons).
automobiles and washing machines, are
often computer-controlled, and have soft- Associated terminology: Unix, Compiler.
ware that controls their functions (e.g., the
engine management system, traction con-
trol, braking system, etc.). However, that
software is not usually considered an oper- Optical character recognition (OCR)
ating system, since it is really just a piece
of Firmware, and does not control the oper- Definition: Processing a digital scanned image of
ations of other software. In some cases, printed or written words, to isolate and identify the
onboard computers do interact with each individual letters and the words they form.
other, and do run a number of different
applications depending on changing cir- Overview
cumstances, and many manufacturers use It seems like a very simple problem. Given a
a special reduced-size version of a normal good digital image of some printed words,
computer operating system to act as the work out what they say. The shapes of char-
basis of the onboard controller. acters are well known, and printed words
provide a great deal of consistency. Natu-
Summary of positive issues rally, understanding handwriting could be
There are several types of operating system more difficult, but understanding printed
interfaces available, including command- characters seems to be a simple matter of
line-style and GUI-style systems. Applica- seeing which of 26 already known patterns
tions vendors can provide data regarding each letter matches.
operating systems requirements for their Optical character recognition (OCR) is in fact
system. Operating systems are available in a an exceptionally difficult task. Even with
variety of configurations to support a vari- the enormous commercial gains to be had
ety of applications and platforms. Many from being able to process printed doc-
operating systems have both formal and uments automatically, no system has yet
informal support groups. been perfected. The problems are numer-
ous. When scanned with enough resolu-
Summary of potentially negative issues tion to give sufficient detail, no two printed
The types and variety of operating systems characters ever look exactly the same. Varia-
have been consolidating since the 1980s, tions in lighting, texture of paper and ink,
resulting in three distinct primary camps: slight imperfections in the image, imper-
the Unix-like, the Windows family, and the fect alignment, and a host of other prob-
others. Some operating systems lack suffi- lems mean that looking for an exact match
cient support from vendors, have limited with predetermined letter shapes will
functional options, contain bugs, and have never succeed. Added to that is the prob-
poor security. Some operating systems are lem that different fonts, type sizes, and
not supported at all except through infor- styles (bold, italic, etc.) provide a potentially

243
Optical storage

infinite variability to how any letter may inal by eye to ensure correctness. Predomi-
look. nantly, OCR technology is effective in appli-
There are many commercially available cation areas that use special optical char-
software packages that do a creditable job acter sets such as those found on checks.
of OCR, even down to properly setting Another application of OCR is in comput-
paragraphs and diagrams in popular word- ing devices that accept stylized handwrit-
processors’ file formats. None of them is ten input, but these systems also require
totally reliable, or even comes close to the extensive training of users if they are to
degree of accuracy that a capable and care- work effectively.
ful human reader could achieve. For the
collection of essential data, for which accu- Summary of positive issues
racy is of some importance, it is always OCR can enable large quantities of doc-
necessary to have a human operator com- umentation to be entered into a system
pare the original printed document with (with varying degrees of accuracy). OCR sys-
the results produced by OCR software. This tems that use specially designed characters
does not by any means mean that OCR and that are read in special-purpose readers
is pointless. Human operators can com- are effective in automating processes (such
pare the two versions much more quickly as mail sorting). Some computing devices
than they could possibly type the entire allow hand-written input through the use
document, so OCR is a great time saver; of special styli and screens.
it simply means that a totally automatic
process can not be expected to produce Summary of potentially negative issues
error-free digital transcriptions of printed OCR of documents does not provide 100%
documents. reliability of duplication when scanning
The similar problem of handwriting ‘‘real-world” documents that were not
recognition is many times more difficult. specifically created to be processed elec-
Either users must write in a specially desi- tronically.
gned highly stylized way (such as the stra-
Reference
nge alphabet familiar to users of the Apple r H. Bunke and P. Wang (1997). Handbook
Newton), or the system must be trained:
of Character Recognition and Document
hand-held through a selection of samples to
Image Analysis (Singapore, World
learn a new writer’s style. Even then, reli-
Scientific Publishing Company).
ability is far lower than for printed-word
OCR. Associated terminology: Natural language
processing, Neural networks, Machine
Business value proposition learning.
The potential for OCR is vast, since the abil-
ity to capture electronic versions of whole
collections of documents would move orga-
Optical storage
nizations closer to the envisioned paper-
less office. However, the technology has not Foundation concepts: Storage, Disk, Backup.
yet reached the point at which this can Definition: Data-storage systems based on reflection
be effortlessly and reliably achieved. OCR and refraction of light, rather than on the traditional
software is available to scan documents magnetic properties of materials.
and store the results in a digital form, but
the reliability of these systems is less than Overview
total, and it is necessary to cross check the Primary storage in computer systems, that
electronic document against the paper orig- known as main memory or RAM, is now

244
Optical storage

exclusively built from solid-state electro- occurs more quickly in cheaply made CDs:
nics; data is stored as static electrical when the glue holding the two layers of
charges. Secondary storage, usually known plastic tightly together starts to fail, air can
as disk, is almost exclusively built from get to the aluminum layer, and that is the
spinning disks on whose surfaces data is beginning of the end. If CDs are to be used
recorded as microscopic magnetic fields. for archival or backup purposes, the life-
Optical storage provides another alternative, time of the chosen media must be carefully
in which data is stored as minute deflec- investigated.
tions or deliberate defects in a reflective or DVDs (Digital video discs or Digital versa-
refractive medium (often just a thin layer tile discs) provided a large increase in capac-
of shiny aluminum). Data is read from opti- ity, and hardware for writing data to DVD
cal media by shining an accurately focused blanks is now as widely available as that for
laser light onto it, and measuring how the CDs. A DVD uses the same basic technology
reflection of that light is affected. as a CD, in the same way; improvements in
Laser disks, or Disco-vision, were introduced technology between the introduction of the
in 1969, but were fundamentally analog two simply allow for more to be done in the
systems, and could not have been used same space. DVDs use smaller bumps, read
effectively for digital data storage. The first by a shorter-wavelength laser (visible red
optical storage devices suitable for com- instead of infrared) from a thinner sheet of
puter systems were the now universally aluminum. The standard capacity DVD can
familiar Compact discs or CDs. Although CDs store 4.7 GB of data, although more expen-
were originally used exclusively for audio sive dual-layer versions with a capacity of
recordings, they store their contents in a 8.5 GB are available. Dual-layer is not the
completely digital form, and are ideally same thing as double-sided; double-sided
suited for all kinds of digital data. A CD DVDs are not generally used in computers
consists of two layers of transparent plas- because of the need to turn them over or
tic with a very thin layer of metal, usu- have two laser pickups.
ally aluminum, sandwiched between them. The DVD market is confused by a profu-
Microscopic bumps are printed onto the sion of standards. The most basic is DVD-
aluminum during manufacture, or may be ROM: this is a read-only format; data must
‘‘burned” in by laser light in a CD recorder. be written during manufacture, and can
When a CD is played, another laser shines not be modified later. DVD--R and DVD+R
onto the surface, and the microscopic are two slightly different write-once for-
bumps cause detectable changes in its mats: data may be written onto a blank
reflection. disc once; after that it can not be modi-
CDs have a usual capacity of 650 MB, fied. These are the cheapest and most use-
but that may be slightly increased in some ful for archival backups and software dis-
models. Data may be read from a CD at tribution. Most hardware can read both
rates of about 10 MB per second, and CD- the --R and the +R versions, but some recor-
recording hardware may fill a CD in about ders can write onto only one format.
5 minutes at best. Although CDs do pro- Domestic DVD video players are also often
vide good long-term data storage, they do capable of playing only videos recorded
not last for ever. Scratches on the plastic in their preferred format. DVD--RW and
surface may often be repaired completely, DVD+RW are re-writable formats; the discs
or even ignored successfully, but deterio- may be erased and re-recorded many
ration of the aluminum layer can not be (but not unlimited) times. The difference
repaired. This deterioration is much accel- between --RW and +RW is the same as the
erated by storage in poor conditions, and minor incompatibility between −R and +R.

245
Optical storage

DVD--RAM and DVD+RAM are extensions of rent research. If successful, it could result
the RW idea; instead of having to erase the in a rapid increase in available storage
whole disc before writing new data onto capacities, and greater reliability since it
it, single blocks of data may be erased and might not require any moving parts. No
re-written individually, so a DVD±RAM disc such products are commercially available,
behaves like a normal but slow hard disk. or promised for the near future; holo-
The DL suffix indicates ‘‘dual layer”; dual- graphic memory currently lingers in the
layer blanks are much more expensive than world of fiction.
single-layer ones, and require special hard-
ware for writing. Currently, only DVD--R DL Business value proposition
and DVD+R DL are available. Optical storage offers organizations and
Two further improvements to optical individuals the ability to store and archive
disc technology have recently become avail- data in a semi-permanent medium at rel-
able. Both take the same basic technology atively low cost. For several decades tapes
even further, using a violet laser to detect and ‘‘floppy” disks have been used to
even smaller, more closely packed bumps archive data; however, these media have the
in the reflective layer. One is HDDVD (High- same inherent problem in that the data is
density DVD) with a standard capacity of encapsulated in a magnetic coating that
15 GB, but a 30 GB dual-layer version; the can degrade over time. The same is true
other is called Blu-ray (because of the laser for audio and video cassette tapes, which
color), and has a standard capacity of 25 GB, frequently give very-low-quality play-back
but also dual- and quadruple-layer versions after a few years. While computer tapes are
for capacities of 50 GB and 100 GB. usually stored in a temperature-controlled
Magneto-optical discs, which were popu- environment and used infrequently, they
lar for a period, are not very widely seen suffer the same inherent limitations (in
now. In a magneto-optical system, data is some systems the tapes need to be replaced
stored as microscopic magnetic fields on as frequently as every ten backup cycles).
the surface of a rotating disk, but a tightly CDs and DVDs solve this problem by not
focused laser light is also applied when using magnetic materials that can degrade,
data is written. This has the dual effect of but instead their data is burnt onto the
allowing smaller magnetic fields to be writ- disc, although it must not be forgotten
ten (thus increasing capacity) and render- that these media do still degrade in their
ing those fields optically detectable. When own ways. While CD and DVD technology
data is read back, only a beam of polar- has been evolving, its major limitation is
ized light is needed, no magnetic pickups; the amount of data that can be written
the light interacts with the magnetic field onto a disk. ‘‘Juke-box” technology for CDs
when it is reflected, and the data is read has been developed, allowing many CDs to
from the reflected light beam. Magneto- be created without human intervention or
optical discs are much more complex and delays.
expensive than traditional magnetic discs, The blu-ray optical technology has the
and, once the price/capacity ratio for the potential to expand the options open to
latter had fallen significantly, the former network users and managers since the data
fell out of favor. storage capacity is much higher than for
Holographic storage, in which data is older optical storage devices, and the speed
stored throughout a three-dimensional vol- at which data may be saved is also much
ume instead of merely on a two-dimen- increased. Blu-ray technologies have been
sional surface, is the subject of some cur- designed to accommodate the needs of

246
OSI seven-layer model

the high-definition television communities; the kinds of interactions permitted, all the
however, the technologies will inevitably be way to providing a convenient end-user
adapted to computer-related uses. interface. Such complex systems can be
implemented reliably only if they are divi-
Summary of positive issues ded into more manageable subsystems that
Optical storage technologies offer individ- may be developed independently. For most
ual users an easy way to store relatively new development, the ISO’s OSI (Open Sys-
high amounts of data on almost non- tems Interconnection) seven-layer model is
volatile media. New technologies such as used.
blu-ray have been developed to provide
higher capacity storage at higher data 1. The ‘‘physical layer” is concerned with
transfer rates. the kinds of cables and connectors used,
and the nature of the electrical (or
Summary of potentially negative issues other) signals transmitted.
Optical storage technologies have been lim- 2. The ‘‘data-link layer” describes the for-
ited in size, and over a long period of time mat of data packets sent on the first
the medium itself is subject to decay if not layer, and communications when a
well maintained. The variety of acronyms direct link between two systems exists.
for CD and DVD media is large and confus- 3. The ‘‘network layer” controls long-
ing and care needs to be taken when select- distance communication, when a direct
ing a disc type for use. Holographic storage link does not exist, and data packets
is currently a research area but no products have to take a number of ‘‘hops” to
have yet been demonstrated for commercial arrive.
use. 4. The ‘‘transport layer” specifies how data
flows between applications, and handles
Reference delivery receipts and retransmissions if
r F. Yu and S. Jutamulia (1996). Optical
necessary.
Storage and Retrieval (New York, Marcel 5. The ‘‘session layer” specifies the
Dekker). sequences of messages that must
be sent and received in order to
achieve particular goals, the language
OSI seven-layer model of request and response between
connected applications.
Foundation concepts: Network, Protocol. 6. The ‘‘presentation layer” describes the
Definition: A commonly used abstraction of network format and representation of the indi-
software architecture. vidual messages, and any data that they
carry.
Overview 7. The ‘‘application layer” is the layer of
Network applications can be very complex interest to users, which provides the
pieces of software, because there are so purpose behind using the other layers,
many different levels of control that they namely the application that controls
need to exert: from correctly controlling communications and provides, uses, or
the transmission of signals on the network displays the transmitted data.
hardware, through routing data over poten-
tially long distances and complex network Developers can expect the first four lay-
topologies, ensuring that all signals are ers to be provided as part of any working
received correctly as transmitted, defining computer system. If a totally new kind of

247
Outsourcing

application is under development, the last The TCP/IP protocol stack happily existed
three layers must all be created, but, if an with four levels long before the OSI model
interface to an existing kind of system is was introduced. Insisting on seeing every
being developed (perhaps a new email ser- system as consisting of seven layers can pro-
vice or a new web browser), then layers 5 duce a distorted and confusing view.
and 6 will already exist, and only the final
References
layer is needed. r W. R. Stevens (1994). TCP/IP Illustrated
At each layer of implementation, the
(New York, Addison-Wesley).
developer may rely upon the previous lay- r International Organization for
ers already existing and working correctly.
Standardization (ISO) (1994). ISO/IEC
Each new layer simply adds a new level of
7498: Open Systems Interconnection, The
functionality, making full use of the previ-
Basic Model (Geneva, ISO).
ous layer. For example, IP (the internet pro-
tocol) is a layer-3 protocol, it is responsible Associated terminology: Ethernet,
for delivering data from any computer to Internet protocol, TCP/IP, Client--server.
any other anywhere on the network. It uses
a layer-2 protocol (such as ethernet) which
is already capable of delivering data over a Outsourcing
local area network (LAN), so all it has to be
concerned with is the forwarding required Foundation concepts: Business process re-
when two computers are not on the same engineering.
LAN. Definition: Outsourcing is the use by a company of a
third-party provider to perform a function or imple-
Business value proposition ment a process on its behalf.
Protocol stacking based on the OSI seven-
layer model is of great benefit to network Overview
software developers. The clean division of Outsourcing has been used in business for
network tasks into a number of layers decades. As organizations began to under-
allows the developer to concentrate on just stand that they did not have to perform
the one relevant aspect of design, knowing all the functions of business themselves,
that surrounding layers are isolated and known as being Vertically integrated, they
well specified. Within businesses that are moved toward an outsourcing model, con-
not actively involved in technology develop- tracting with other vendors to supply com-
ment there is less need for the MIS staff to ponents and provide services. For example,
be concerned with all the minutiæ of the an automobile manufacturer may choose
OSI model, since the majority of the soft- not to make brakes at all, but to have a
ware in use has been designed to be simply company that specializes in brakes produce
‘‘plugged in” to a network, with all issues and supply them under contract.
related to the protocols already resolved. The history of outsourcing in the context
of information technology (IT) has a slightly
Summary of positive issues different past and rationale from those of
The OSI seven-layer model provides a robust outsourcing in the manufacturing sector.
theoretical model for all systems developers In the 1960s not all companies could afford
to construct systems around. computers, and they utilized ‘‘computing
service bureaus” to undertake their data
Summary of negative issues processing. During the 1970s and 1980s,
The OSI seven-layer model is frequently as computing became more pervasive, com-
seen as the way the internet works. It is not. panies began to change their views on

248
Outsourcing

software development. Rather than writing be prohibitive, move IT assets and staff off
all their own programs, it became much the balance sheet, or alleviate the effects of
easier to purchase packages, in effect out- staffing shortages.
sourcing their development effort.
Outsourcing remained a useful option Summary of positive issues
for many organizations that wished to The ability to source technology skills
supplement their own internal IT organi- around the planet has made a wide variety
zation; in particular, the outsourcing of IT of options available to organizations. While
training for employees became a popular costs are frequently considered to be a pri-
option. However, a seismic change occurred mary driver of outsourcing, other issues
in 1989 when Kathy Hudson, the CIO of (such as the ability to provide a call cen-
Eastman Kodak, decided to outsource the ter that is located in a suitable time zone)
company’s entire IT operation. This strate- also drive the decision to locate IT services
gic decision changed the way that IT in distant locations. Strategic outsourcing
outsourcing was viewed by organizations; enables organizations to use IT to achieve
instead of considering it an expense, they a variety of strategic ends, including cost
saw the opportunity to ‘‘sell” their IT oper- advantages, human capital advantages, and
ations and then pay for a service oper- customer-service provision.
ated by another company according to a
contract. Summary of potentially negative issues
In the twenty-first century, IT outsourc- Outsourcing can be associated with a vari-
ing has evolved significantly from the total ety of positive attributes, but many of them
outsourcing experienced by Kodak. While may be positive in the short term and
very large total outsourcing contracts still problematic in the long term. Outsourc-
occur, companies now undertake selective ing may provide a cash bonus from the
process sourcing, which, due to advances in sale of the IT function, but, if the out-
technologies, can potentially be performed sourced functions are not performed as
at any place on the planet. well as expected, it may be very difficult
to bring the IT function back in-house (also
known as Insourcing). While different loca-
Business value proposition tions have different costs associated with
The use of process and technology out- them, they also frequently have differing
sourcing has become a popular mechanism levels of security provision and legal regu-
by which organizations can achieve a strate- lations to work within, which may prove
gic goal. The goal may be to allow the com- detrimental should a problem occur.
pany to focus on strategic IT initiatives,
Reference
achieve a higher level of operational perfor- r S. Cullen and L. Willcocks (2004).
mance than could be developed in-house,
Intelligent IT Outsourcing (Burlington, MA,
achieve a level of performance at lower cost
Butterworth-Heinemann).
than could be achieved in-house, complete
a short-term specialist project when the Associated terminology: Application
setup costs for in-house development would service provider, Hosting, ISP.

249
Packet switching and circuit switching

is added to each packet, identifying the


Packet switching and circuit
contents of the packet, its sequence num-
switching ber (its ‘‘chunk number,” which is used to
Foundation concept: Networks. ensure that the original data is reassem-
bled in the correct order), and its desti-
Overview nation. The labeled packet is transmitted
The standard model for a voice telephone to a switching station. The switching sta-
system is a circuit switching network. When tion receives the entire packet, inspects
two users or end-stations wish to commu- its label, and retransmits the packet to
nicate, a direct connection must be made another switching station, which is (at least
between them, and dedicated to that call hoped to be) closer to the ultimate desti-
for its duration. The connection may be nation. The packet makes its way, hop by
through copper wires, fiber-optic cables, or hop, across the network, until it reaches
even radio or microwave transmissions. It a switching station that has a direct link
may pass though any number of switch- to the ultimate destination, whither it is
ing stations, staffed by human operators delivered.
with plug-boards, or electronically oper- Each station on the network is responsi-
ated. It is even possible, using frequency ble only for sending a packet across one sin-
shifting, for many conversations to be car- gle link. The individual packets of a trans-
ried by the same wire at the same time. The mission may take different paths across the
key concept is that a complete connection network, and may even arrive out of order.
must be made, then the conversation can If one link is broken, packets may simply
happen, and then the connection may be take an alternate route, or be held in stor-
dismantled. age until the link is repaired. This gives
Circuit switching allows very simple tele- packet-switching networks a robustness not
phony equipment built on nineteenth- available with circuit switching. The divi-
century technology to operate effectively sion of data into small packets, and the
(the first telephone exchange was working loss of the requirement that transmission
in 1877), but it does not allow very efficient by the originator and reception at the ulti-
use of the high-cost infrastructure (long- mate destination should be simultaneous,
distance cables and transmitters). allows the most efficient possible use of
The alternative is a Packet-switching net- network bandwidth.
work, and requires that all communica- Carefully designed protocols are an abso-
tions are in digital form. In a packet- lute requirement, to ensure that packets
switched network, every end-station has are delivered to the right destination and
a permanent connection to at least one reassembled into an exact duplicate of the
switching station; each switching station original data, to ensure that any transmis-
has a number of permanent connections to sion errors are detectable if not correctable,
other switching stations. The connectivity and to ensure that the original sender
does not change during use: connections is able to infer that data has not been
are not made before data is transmitted or received when there is a larger problem.
broken thereafter; there is a path (perhaps The best known and probably most used
a very long path, through dozens of switch- of these is the TCP/IP/ethernet protocol
ing stations) from any end-station to any family.
other. Frame relay is a term often confused with
Digital data is split into manageably packet switching. It is certainly a simi-
sized chunks, known as packets, normally lar concept, but by no means a synonym;
a few thousand bits long. A digital label the confusion results from the fact that

250
Parallel processing

the data packets transmitted by packet- Associated terminology: IP (internet


switching networks, especially ethernet protocol), Voice over IP.
and IP packets, are often called Frames.
Frame relay is just one of the many possi-
ble implementations of packet switching; it Parallel processing
was originally developed as a simpler alter-
native to the X.25 standard, to allow WAN Foundation concept: Algorithm.
(wide-area network) connections to utilize Definition: The use of more than one computer or CPU
ISDN lines, and is now commonly used over concurrentlytosolveasingleproblemortorunasingle
T1 and T3 lines. Frame relay is expected application.
to be gradually replaced by networks using
ATM (asynchronous-transfer mode). Overview
It would seem obvious that a number of
Business value proposition computers working together on a single
The development of packet switching has problem will produce faster results than
enabled a robust mechanism through a single computer could. That is the basis
which data can be sent reliably from one of parallel processing. As with most things
point to another. Packet-switching technol- that seem obvious, it is usually not that
ogy forms the basis of the internet and is simple. There are two major considerations
composed of a wide variety of standards that have great influence on the effective-
and protocols such as TCP/IP and ethernet. ness of parallel processing.
The open nature of most of the packet- One is the communications overhead.
switching standards and the fact that there Most problems can not be devolved into
are organized working groups actively sup- smaller sub-problems that are totally inde-
porting their ongoing development have pendent. Take for example the elementary-
led to their widespread adoption. school problem of long division. Two peo-
ple working together to divide a big
number by a smaller one will take just
Summary of positive issues as long as a single person would, because
The technology associated with packet- there is no opportunity for them to work
switching networks is reliable and mature. concurrently on different parts of the prob-
The technology is extremely well supported lem. On the other hand, when compiling
and almost universally adopted. Packet- statistics on a large number of data items,
switching technologies are used as the two people can produce a significant speed-
basis of the internet and include TCP/IP and up: they can divide the data items into
ethernet. two smaller sets and independently calcu-
late the statistics for their own halves. Even
Summary of potentially negative issues here, they will not be able to double their
The technology continues to evolve and speed, because some extra calculations will
requires organizations and individuals to be required to combine the statistics for the
deploy resources in order to maintain the two half-piles into overall results. Finally,
compatibility of their systems with current there are problems for which parallel pro-
industry standards. cessing does produce optimal results: two
people can eat a cake twice as fast as a sin-
Reference gle person could; no communications are
r L. Peterson and B. Davie (2003). required at all.
Computer Networks: A Systems Approach In the general case, the amount of time
(San Francisco, CA, Morgan Kaufmann). each processor must spend communicating

251
Parallel processing

with other processors can be expected to records each. Each person would sort their
grow directly with the number of proces- own half, but, because of the quadratic
sors involved. Viewing simple graphs show- nature of the sorting method used, each
ing (1) how processing time reduces with person would take only a quarter of a year
increasing number of processors sharing to sort their share of the records. They
the load, (2) how communications time would then have to merge the two sorted
grows, and (3) the sum of processing time piles into a single sorted pile, but that oper-
and communications time illustrates the ation is an order of magnitude faster than
point clearly: in cases like this, there is the sorting. So two people could sort the
an optimal number of processors to use, records in just over one quarter of the time
beyond which adding more actually pro- taken by one person. Ten people could do it
duces a slower final result. in just over one hundredth of the time. The
The second essential consideration is the only reason why 250 000 people working on
algorithmic Complexity of the problem at the problem couldn’t do it in a few seconds
hand. If the algorithm chosen is signifi- is the communications overhead: merging
cantly worse than linear in its time com- 250 000 piles of four sorted records each
plexity, then parallel processing is very would take quite some time.
likely to be beneficial. If it is better than lin- So, there are cases in which parallel pro-
ear, the benefits are not so clear, and deeper cessing pays off handsomely, cases in which
analysis will be required. For example, con- it has a negative effect, and, of course,
sider the problem of sorting a large num- everything between those two extremes.
ber of database records into alphabetical Parallel processing is not a simple matter,
order. Naive sorting algorithms, the kinds and needs thorough investigation for any
of method that a person naturally uses, given problem domain. Fortunately, this
are quadratic in time. That means that the is a well-researched area, and there is a
time taken to solve a problem grows with vast literature available to those who know
the square of the number of data items what to look for.
involved: twice as much data means four There are also choices to be made regard-
times as long to process; ten times as much ing the style of parallel processing to be
data means a hundred times as long to pro- used. Should there be a number of inde-
cess. Now suppose that it would take one pendent computers connected by a net-
person a whole year to sort a collection of work, or a single specially designed com-
one million records into order (that is not puter with multiple CPUs? The former,
a totally unreasonable figure, if the records known as Distributed computing, or a Loosely
are small and easy to handle). Two people coupled system, allows systems to be built
working on the problem would divide the from off-the-shelf components, but has far
records up into two halves, with 500 000 slower communications; it is the kind of

252
Parallel processing

system that is often used when a prob- to some useful network-coordinated back-
lem is devolved into a number of quite ground task in their idle periods. For exam-
large sub-problems, such as with queries on ple, the SETI@home (Search for Extraterres-
very large databases. The latter, known as a trial Intelligence at Home) research group
Tightly coupled system, is a much more pow- is looking for a signal from outer space, and
erful strategy, but very expensive. Very few relies on vast numbers of volunteers via the
vendors offer large multi-CPU computers; internet to load a program that acts like
it may even require customized computer a screensaver running in the background
design. only when the computer is idle, to perform
Another possibility is to have a single signal processing. It is estimated that they
powerful computer act like a larger num- have many millions of users on their net-
ber of smaller computers by dividing up its work, contributing many teraflops of com-
time amongst a number of separate tasks, puting power and trillions of floating-point
rapidly switching its attention from one to operations a second to solve the problem. In
another, maintaining the illusion that all 1999, a similar cooperative effort involving
are being processed concurrently. This is about 100 000 personal computers world-
called Multi-processing, and is a technique wide succeeded in cracking a secret mes-
used by all general-purpose operating sys- sage encrypted with 56-bit DES in less
tems to support multiple applications run- than 24 hours. The combined idle time of
ning at the same time. Multi-threading is a large corporation’s network of comput-
an emerging technique that allows a single ers is a formidable resource waiting to be
CPU to run more than one program at the tapped.
same time without having to switch atten- The use of tightly coupled systems is cur-
tion rapidly from one process to another. rently limited to either very simple small
This is related to the concept of a Thread, systems or highly complex problems requir-
which is a stream of execution within a sin- ing extensive processing-time and data-
gle application, allowing one application to throughput capabilities. One such system
work on multiple problems without resort- is IBM’s massively parallel Blue Gene/L com-
ing to complex programming techniques. puter, which has been designed to scale to
65 536 dual-processor nodes with a peak
Business value proposition performance of 360 teraflops (a teraflop
Parallel processing is an area of comput- is 1 000 000 000 000 arithmetical operations
ing that holds great potential for a vari- per second). This system is being used to
ety of large and complex problems. It is investigate, amongst other things, the prob-
unlikely that a simple office application lems of protein folding in the biological
such as word processing would benefit sciences.
substantially, if at all, from implementa- The adoption of loosely coupled and
tion on a parallel-processing architecture widely distributed processing in the style of
at current levels of technology, but signifi- SETI@home requires special software to be
cant improvements are likely for computa- deployed, and it is clearly not suitable for
tionally intensive tasks in the foreseeable applications in which security or response
future. time is a major overriding concern, but
The use of a distributed, loosely coupled its applicability to problems in the public
form of parallel processing is attractive for domain and for the public good is clear.
those with significant computational prob- The adoption of tightly coupled processing
lems. Many corporate PCs spend as many is constrained to a certain class of major
as 17 hours a day doing nothing (24 hours problems and is also constrained from a
a day on weekends), and could be assigned user’s perspective by the limited number of

253
Password

super-computers that are available and the Overview


expense in utilizing their capacity. The idea of a password is very simple: some
word, phrase, or jumble of letters and dig-
Summary of positive issues its known only to one person, which they
Parallel processing greatly improves the use to verify their identity when accessing a
speed of processing for certain classes of secure computer system. The security prob-
problems. Several approaches to parallel lems associated with passwords are less
processing are available, including loosely well known.
and tightly coupled processing. The use The first problem is with transmitting
of the internet and volunteered comput- the password. The user must type their
ing resources can allow enormous com- secret password and send it to the secure
puting power to be unleashed. Very large, computer system so that that system may
tightly coupled parallel processing systems perform the identity check itself. If the
are available for complex problems. There user is connected through a modem over
is a vast literature on parallel processing. a telephone line (dial-up connection), there
could be a ‘‘tap” anywhere on the line,
Summary of potentially negative issues allowing a third party to see everything
Parallel processing is a complex computing that is transmitted in either direction, and
environment. Tightly coupled, massively easily record passwords. If a network con-
parallel computers are very expensive, diffi- nection is used (ethernet, fiber-optic, DSL,
cult to use, and limited in availability. Par- cable modem, etc.), a similar device called
allel processing has limited application to a ‘‘packet sniffer” could also be attached
the majority of commercial programming at any point, and could intercept all trans-
problems under current technology. missions (to be precise, a packet sniffer is
a software application that may be run on
References any computer; it is that computer that taps
r B. Wilkinson and M. Allen (2004). into the network connection).
Parallel Programming: Techniques and The second problem is with password
Applications Using Networked Workstations management on the secure computer itself.
and Parallel Computers (Englewood Cliffs, In order to check that the password entered
NJ, Prentice-Hall). by the user is correct, the secure system
r I. Foster (1995). Designing and Building must have some kind of record of every
Parallel Programs (New York, Addison- user’s password. If security at that instal-
Wesley). lation is not perfect, if they fall victim to
r A. Gara M. A. Blumrich, D. Chen et al., any form of ‘‘hacking,” all passwords could
‘‘Overview of the Blue Gene/L system be revealed. If a password gives access to
architecture,” IBM Journal of Research and any kind of financial control (such as order-
Development 49, 195--213. ing transfers of funds), how could any user
prove that they did not authorize a par-
Associated terminology: Algorithm,
ticular transfer if their password has been
Complexity.
compromised?
The third problem is with the choice
of password. Many people are required to
Password remember a large number of passwords for
the various systems they use, and have
Foundation concept: Security. to pick something memorable, for obvious
Definition: A secret identification code used to verify reasons. Some users are truly careless and
identity. choose passwords like ‘‘password,” ‘‘abc,”

254
Password

or their own names. The reason this is a could not be accessed by normal users.
problem is that guessable passwords really The computer operating system’s own file-
can be guessed. Someone trying to break protection services were used to ensure
into a system can simply try a lot of com- that only the system administrator and the
mon password choices for a large num- password-verifying software could access
ber of accounts, and can expect occasional the password file. In modern times, when
successes. ‘‘Cracking kits” are readily and new security flaws seem to be discovered
freely available for download: these are every day, this is clearly not sufficient.
software packages that automatically scan One successful virus, worm, or Trojan-horse
through whole dictionaries, trying out an attack could reveal everything. The solu-
endless stream of possible passwords at tion is that passwords are not stored at
very high speed. No real word or simple pat- all. When a user sets or changes their pass-
tern is safe. Systems that require their users word, it is encrypted using a One-way hash
to pick four-digit PINs (personal identifica- (a kind of encryption scheme that can not
tion numbers) as passwords could almost be reversed: it is impossible to recover the
be considered pre-compromised. original message from the encrypted ver-
There are solutions for each of these sion, even if the key is known). Only the
three problem areas. The first is a software encrypted version is stored. Each time the
solution called SSL (secure-sockets layer). If user logs in, the password they enter is also
both computer systems have SSL installed encrypted in the same way, and the result
and elect to use it, every byte transmitted is compared with the previously stored ver-
is automatically encrypted using a strong sion. This means that the password file
crypto-system. Security is enhanced by the could be published in a newspaper with-
fact that a new, very long encryption key out any passwords being revealed. Until
is automatically generated for each ses- recently, it was standard for all Unix sys-
sion, so long-term eavesdropping provides tems to keep their password files in a uni-
no additional leverage. Incredibly, it is pos- versally readable location.
sible for two systems to securely choose Unfortunately, increased computing
a new encryption key so that they both power has made that system less than
know what it is, but even an interloper who perfect. Since the one-way-hash encryption
had complete access to everything trans- system is well known, it is possible for
mitted (even knowing both parties’ public a programmer who has access to an
and private keys) could not know it (the encrypted password to create a program
best known of these methods is known as that simply tries out all possible passwords,
Diffie--Hellman). SSL is available as a stan- encrypting them all, and waiting for a
dard part of nearly all modern operating match. For systems that allow passwords to
systems, and costs nothing to use beyond a be of any length, it is not physically possi-
small requirement for extra computation. ble to try out all possible passwords because
It simply requires that the server is set there will be an infinite number of them,
up to use it. When a web browser warns but most users, if left to their own devices,
that information is being sent insecurely, it will pick a simple memorable password
usually means that the server is not using of five or six letters, and that is an open
SSL. invitation to automatic cracking software.
The second problem can not be solved Cracking kits commonly run through all
perfectly, but may be significantly reduced words in a very large dictionary of many
by simple standard software methods. It languages, try all short combinations of
used to be thought enough to store pass- letters even if they don’t make words, and
words in a specially protected file that even try adding digits in various places.

255
Password

Cracking software can be left running in password file has been released, because
the background for many days until it the cracking software does not then need
happens upon a match. to go through the official log-in procedure
The only way to make passwords at all for each attempt.
safe is to use a combination of one-way Another good security measure is to
encryption, plus putting the password file require that the password verification pro-
in an inaccessible place, plus keeping every cedure makes a note of each log-in attempt
possible security measure in place. Even that fails because of an incorrect password.
then, it is essential to prevent users from If each user who successfully logs in is
picking easily guessable passwords (any informed of the number of recent failures,
software system can be constructed to auto- they will remember if they did not mistype
matically reject passwords that are not long their own password, and at least know
enough, or that appear in some standard that a break-in attempt has been made.
dictionary, or even ones that don’t con- It is also common for systems to disable
tain a sufficiently complex mixture of non- accounts temporarily, making them inac-
letters). However, if password memoriza- cessible even with the correct password,
tion is rendered too difficult, then users if there have been some (typically three)
must resort to writing down their pass- attempts at access with incorrect passwords
words in supposedly secret places, and all within a moderate period (often ten min-
hopes of security are then lost. If sys- utes to an hour). This reduces the number
tems designers let users make a totally free of attempts that an automatic cracking sys-
choice of passwords (subject to their being tem can make to such a degree that it can
long and complicated enough of course), not be expected to succeed at all.
then users will be able to use the same
password for each system, and if only one Business value proposition
password has to be remembered, it can be The processes surrounding the assignment
quite long and complex. Systems that reject and management of passwords in an orga-
passwords for seeming offensive or politi- nization are primarily the responsibility
cally incorrect provide no benefit (nobody of the chief security officer; in smaller
but the user in question ever sees the offen- organizations this task usually falls to
sive password), and serve to reduce overall the network administrator. The approaches
security. taken to secure systems access must be
Another effective solution is to deliber- based upon best-practice principles from
ately slow down the password verification the assignment of passwords to the stor-
process. It should normally be possible to age of passwords as one-way hashes. It must
verify a password with less than a microsec- be remembered that the value of the infor-
ond of computing, but if the verification mation stored by the system and password
process includes deliberate delays, so that it ‘‘protected” is proportional to the amount
takes a whole second or two, then the crack- of effort that someone dedicated enough
ing kits that work by repeatedly attempt- will undertake to break the password and
ing to log in with different passwords will access the system. Even though one-way-
be slowed down to such an extent that hash algorithms are typically unbreakable,
they could never scan through all the pos- reports since 2004 have shown that small
sibilities. Genuine users will not be disad- bit-length algorithms are capable of being
vantaged, because logging in happens only broken and thus, if the data is valuable
once per session. This solution does not enough, even more secure hash algorithms
help in situations in which the encrypted need to be used.

256
Patent

It is therefore imperative that all organi- processes to ensure higher levels of secu-
zations and individuals follow the security rity. Security policies need to be established
procedures completely and revise those pro- and independently verified. Policies need
cedures on a regular basis. It is advanta- to reflect the real security risks that a vio-
geous to have an independent security con- lation may cause. Many users choose the
sultant examine the procedures and report same password for all of the systems that
directly to a senior member of the organi- they access. While this may simplify their
zation, such as the CIO. This avoids prob- lives, it does increase the potential harm if
lems such as network administrators pro- ever that password is compromised. How-
viding a log-in account under a false name ever, that is unlikely to be as dangerous as
so that, in the event that they get fired, forcing users to write down their passwords
they can return to the company’s network because they have too many to remember.
and perform illicit activities. The security
Reference
scan can cover all access points, remote-log- r R. Smith (2001). Authentication: From
in practices, encryption methods to prevent
Passwords to Public Keys (New York,
the use of packet sniffers, the possibility
Addison-Wesley).
of keyboard-monitoring programs that cap-
ture key-strokes including passwords, and Associated terminology: Encryption,
spyware technologies that monitor users to Hacking, Cracking.
capture password activity. Even the possibil-
ity of rogue hidden cameras in buildings
being used to capture pictures of keyboard Patent
entries needs to be considered.
Definition: A patent is a legally enforceable property
Summary of positive issues right granted to an inventor, allowing the inventor for
By following best-practice principles, pass- a limited time to prevent or regulate others making,
word security may be incorporated into using, offering for sale, or selling the invention. This
an overall technology-security plan. Techno- right is assigned in exchange for public disclosure of
logies exist to help ensure that passwords the invention when the patent is granted.
selected conform to a pattern that is not
simple to break. Most operating systems Overview
may be configured to ensure that pass- The establishment of a patent requires a
words are changed on a regular basis government agency to acknowledge the
and that multiple log-in attempts are not validity of the patent and assign the rights
allowed, thus reducing the risk of robotic of that patent to an individual or other
programs running through a dictionary. legal entity. Most countries have their own
Large-bit-length one-way-hash algorithms patent system and intellectual-property (IP)
may be used to store passwords centrally in laws and a large agency to administer
a secure manner. Passwords may be multi- them. The US Patent and Trademark Office,
ple levels deep and strengthened by the use the European Patent Office, the UK Patent
of a physical access device such as a finger- Office, the Japanese Patent Office, and the
print or a security swipe card. State Intellectual Property Office of the Peo-
ple’s Republic of China are among the most
Summary of potentially negative issues influential. In addition there is The World
No system that uses passwords alone can Intellectual Property Office (WIPO) which
be considered totally secure and it is neces- was formed in 1883 to enforce the Paris
sary to use supplemental technologies and Convention, a treaty that allows patents

257
Peer to peer

granted in one nation to be enforced in


Peer to peer
others. The WIPO covers IP rights in the
areas of inventions, trademarks, and indus- Foundation concepts: Network, Internet.
trial designs. Definition: The connection of two computing devices
The various bodies administering IP or applications of the same status to communicate
rights each have different categories and over a network.
rules according to which a patent can be
granted. For example, in the United States Overview
a patent can be granted as the result of a Peer to peer describes communications
successful application in three categories between two computers or applications
(Utility Patents, Design Patents, and Plant over a network, in which both have the
Patents), whereas the UK has just a sin- same status. Neither acts as a server, in the
gle category. The various patent offices also sense of client--server connections passively
have different rules regarding what is eligi- waiting for and servicing requests from a
ble for patenting; for example, in the UK client.
an invention is not patentable if it is a dis- Peer-to-peer communications are gener-
covery, a scientific theory or mathemati- ally more complex than client--server con-
cal method, an esthetic creation (literary, nections. It is understood that a server
dramatic, or artistic work), or a scheme or must have a known, relatively static IP
method for performing a mental act, play- address in order to be accessed, whereas
ing a game, or doing business; and the clients can be very mobile. In peer-to-peer
presentation of information or a computer communications, both correspondents are
program is not patentable. expected to behave as clients. The modal-
The rules for patentability do change ity of two peers finding each other, so that
from time to time. It is only relatively communications may commence, is one
recently that software methods became of the major problems in any peer-to-peer
patentable in the United States, and it is system.
widely believed that that change was not Peer-to-peer communication does have its
entirely intentional. While the USPTO has advantages, not the least of which is secu-
stated that software as a general class is rity. In a normal client--server system, all
not patentable, it adopted its ‘‘Final Com- communications, whether they be emails,
puter Related Examination Guidelines” in interactive ‘‘chats,” or exchanged files, go
1996 in order to clarify what aspect of through a server, and remain on that server
software is patentable. The guidelines help for some period. There is a risk that a
patent examiners to determine the rela- third party may also be able to access pri-
tionship between the software and associ- vate information before or after the trans-
ated ‘‘processes” or ‘‘machines,” which are fer is complete, and, when total privacy
patentable classes. Patents in the United is required, it is possible and undesirable
States and the UK allow the patent owner that the server will retain logs of which
(the ‘‘proprietor”) 20 years of protection. clients communicated, where, and when.
Peer-to-peer communications (beside using
References the internet as a common carrier) are
r http://www.wipo.int. direct, and data does not rest on any server.
r http://www.uspto.gov. The risk of a server keeping logs, or failing
r http://www.european-patent-office.org. to erase sensitive messages after transfer,
r http://www.patent.gov.uk/. does not exist if there is no server.
r http://www.jpo.go.jp/. The public perception of peer-to-peer
r http://www.sipo.gov.cn. systems may be unreasonably negative

258
Person month

because of press interest in file-sharing software, non-secure data or other informa-


activities that violate commercial copy- tion. Peer-to-peer technologies are not lim-
right, such as dissemination of pirated ited to file sharing, but may also allow users
music and videos. There is nothing inher- to perform tasks remotely, or to perform
ently underhand about a peer-to-peer complex computations using a large collec-
system; it is the positive aspect of enhanced tion of connected computers.
privacy that makes these systems (the best
known of which are undoubtedly Napster Summary of positive issues
and Kazaa) attractive to those engaged in Peer-to-peer systems can assure high deg-
piracy. Traditional client--server systems, rees of security because there is no third
particularly FTP and HTTP, would provide a party or server through which the informa-
much more efficient backbone for copying tion is routed.
music and videos, but the chances of being
detected are much greater, just as they are Summary of potentially negative issues
for legitimate private communications. The use of peer-to-peer file-swapping tech-
nology is open to abuse and illegal activ-
Business value proposition ities. The technology can be complex to
Peer-to-peer computing became a high- develop and maintain, and even more dif-
profile technology when Napster developed ficult to monitor.
a method for file sharing across the inter-
net. Napster worked by maintaining a mas- Associated terminology: Client--server,
ter list of resources (files) and the host com- Protocol, Parallel processing.
puter (location) of those resources, which
enabled users to easily locate the resources
they wanted. A user could then form a Person month
direct peer-to-peer connection with the
host computer and download the file. Since Definition: A unit of overall effort used in the determi-
the resources being swapped were typically nation and description of project-development times.
copyrighted music files, a US court judged
the activity illegal and ordered the Napster Overview
index to be shut down. A Person month (PM), occasionally still
The Napster model was superseded by referred to as a Man month, is the amount
a slightly different model typified by of work performed by one person working
Gnutella. The Gnutella system does not for one standard working month. A PM is
have a central server or contain a master in fact the integral of the number of people
resource index, but rather a user makes working over time: it could be one person
a request to another Gnutella user for a working for a month, 152 people all work-
file. If the second user has the file, this is ing for one hour, or any other equivalent
acknowledged and a peer-to-peer link is cre- combination.
ated such that the file can be downloaded. The PM is used in project management
If the second user does not have the file, to estimate and record the effort expected
it then propagates the request to all of the to be required, or actually expended, on
other Gnutella systems that it is aware of. a project. Traditional software project-
This cascading effect continues until the management models, originating with the
request times itself out or the file is found. COnstructive COst MOdel (COCOMO) proposed
Both of these models can usefully be by Barry Boehm in 1981, equate cost with
used for legal file sharing between groups effort. In COCOMO a PM is considered to be
of users; for example, sharing open-source 152 hours of working time, which averages

259
Phishing

approximately six hours a day and takes r B. Boehm (2000). Software Cost Estimation
into account holidays, average sick days, with COCOMO-II (Englewood Cliffs, NJ,
and weekends. Prentice-Hall).
Associated terminology: Software
Business value proposition development.
A standard work period helps project man-
agers to plan development schedules. The
European Union has an upper limit to the
number of hours an employee can work per Phishing
month and this number can also act as a
Foundation concept: Security.
basis for PM calculations.
Definition: Attempting to trick others into revealing
personal and financial information through fraudu-
Summary of positive issues lent web sites or electronic mail.
PMs standardize the process of software
development and scheduling. Overview
Phishing has become a very widespread
problem for users of the internet. A crim-
Summary of potentially negative issues
inal organization or individual sends out
The use of a standard number in project
emails that are an exact copy of legiti-
scheduling and management to represent
mate emails from a well-known company
a group of programmers is dangerous. As
that many random recipients are likely to
with any project development, there is a
have accounts with. The emails look exactly
standard distribution of programmer skills
like official emails from that company, but
in any project team. The use of a single
direct the recipient to a web site for some
number that equates to programmer effort
vital purpose. Often the email will warn
also might not take into account the vari-
that an account is about to be involuntarily
ances across projects and the project teams
closed, or a large bill is about to be referred
that work together on different aspects of
to a collection agency, and this is the cus-
a development.
tomer’s last opportunity to do something
It is vitally important to take into
about it.
account the fact that, when more than one
The web site that the customer is referred
person works on a project, a significant
to will have an address (URL) that seems
amount of time and effort is expended on
to be right for the company, and the con-
communications. Forty people working for
tent will be carefully set up to duplicate
one hour can not achieve anything close to
the company’s real web sites. Of course, it
the same productivity as one person work-
will be completely controlled by the crimi-
ing for forty hours. This is a fundamental
nal ‘‘phishers.” Victims may be made to feel
flaw in simplistic use of PMs to represent
more comfortable by the fact that they have
an amount of effort.
to log in using their pre-established user-
References name and password, but, of course, the site
r F. Brooks (1995). The Mythical Man-Month: is not checking your identity, it is simply
Essays on Software Engineering (New York, recording the user-names and passwords
Addison-Wesley). that are entered. That might be all that
r B. Boehm (1981). Software Engineering happens; stealing a customer’s user-name
Economics (Englewood Cliffs, NJ, and password is often all that is required
Prentice-Hall). for great financial gains, in which case the

260
Phishing

customer will quite probably be redirected comprehensible to most readers, but they
to the company’s real web site with a mes- contain a list of the names of some of the
sage saying that the password entered was email servers involved in transmitting the
incorrect. Otherwise, the illicit site may message. Look through them quickly, just
continue to ask for an ever-widening vari- the endings, which will normally show two-
ety of personal details, account numbers, letter country codes for any foreign server.
and anything else the customer can be per- If an email that purports to come from
suaded to divulge. a company that you frequently deal with
An alternate means of attack is to set up went through a server in an unusual or
web sites with names very similar to those remote location it is unlikely to be legiti-
of major corporations, but with slightly mate. For example, if you are in the United
misspelled names, and just wait for cus- States it would be suspicious if the com-
tomers to fall upon those sites accidentally. pany you deal with to buy fruit in Wash-
The key to safety from phishing attacks ington State is routing a message through
is to realize that email is an exception- an Asian or West African site.
ally unsafe medium, and that, with stan- Of course, phishing does frequently orig-
dard email clients, it is impossible to know inate from within one’s home country, and
where an email message really came from. the only way to be really safe is very sim-
The return address and all other informa- ple. Don’t click on links in email messages
tion in an email header are exceptionally unless you positively know the sender.
easy to fake. For this reason no reputable The word ‘‘phishing” is just a respelling
company would expect customers to trust of ‘‘fishing,” which is really what it is,
email for anything vital. Legally required fishing for information. The change of ‘‘f”
notices can not be delivered by email, as to ‘‘ph” is just a strange modern fad of
there is no possible proof of delivery. Any no significance. The US Identity Theft and
time an email message asks the recipient to Assumption Deterrence Act of 1998 makes
visit a web site, for any reason, it must be most phishing criminal, but is of little help
viewed with grave suspicion, and remem- when the perpetrators are untraceable or
ber that visiting a web site can be achieved overseas.
by as little as simply clicking on a link
embedded in the message.
If an email recipient is for some reason Business value proposition
tempted to visit a web site as instructed, As a customer service it is vital that orga-
they should first re-read the email care- nizations educate and inform their cus-
fully. A vast quantity of phishing emails tomers about the phishing phenomenon.
now come from overseas locations, and the Organizations involved in internet com-
use of language is often poor. Incorrect merce must at least give the appearance
grammar and punctuation are common of taking a proactive position. Failing to
give-aways, but, of course, correct usage of respond when customers report suspicious
language is no evidence of legitimacy. As emails can destroy customer confidence,
another check, look at the email ‘‘internet and it takes very little effort to invite cus-
headers” (with the ‘‘outlook” mail client, tomers to report suspicions, and in turn
select from the menu ‘‘view” → ‘‘options,” warn them of any known form of attack, or
and they appear at the bottom of the to inform them that you would never send
dialog; other mail readers provide the same an email directing them to follow a link to
information but with a different access a web site. Such a stance, and, if possible,
sequence). The ‘‘internet headers” are not taking legal action against the offenders,

261
Port

helps to preserve a brand and maintain cus- patibility, then, when it is working and has
tomer confidence in the organization. been well tested, and its value has been
assessed, it is Ported to another system.
Summary of positive issues For well-written code, and when the two
There is no positive business value from systems are in some way similar (perhaps
phishing unless your business is actively two varieties of Unix), the porting pro-
criminal. cess can be almost trivial and does not
require any great expertise in program-
Summary of potentially negative issues ming or even full understanding of the soft-
In individual terms phishing can be ware itself. For applications that make use
extremely damaging for any victim of infor- of system-dependent features, in contrast,
mation, identity, or data theft. Phishing porting can be a major project.
can weaken customers’ confidence in an 2: Every computer has a number of con-
organization that does not have a policy nectors at the back into which other things,
to deal with phishing violations against its such as keyboards, mice, printers, network
customers. cables, and USB devices are plugged. These
References connectors are known as Ports.
r M. Whitman and H. Mattord (2004). Network ports, TCP/IP ports, and inter-
Principles of Information Security (Boston, net ports: The same word is also used to
MA, Course Technology). describe a Virtual socket, most commonly for
r L. Peterson and B. Davie (2003). network connections. Even though a com-
Computer Networks: A Systems Approach puter normally has just a single network
(San Francisco, CA, Morgan Kaufmann). port, with a single network cable plugged
into it, it can maintain a large number of
Associated terminology: Cracking, simultaneous connections with other com-
Hacking, Law cross-reference. puters. This is made possible by the operat-
ing system, which creates tens of thousands
of virtual ports or Network sockets.
Port The underlying internet protocols TCP/IP
and UDP/IP allow messages to be given two-
Definitions:
part destination addresses. The well-known
1. Reconstruction of software designed for one IP address (q.v.) specifies which other com-
kind of computer to work on another. puter the message should ultimately be for-
2. A communications connection point. warded to; the second part, the Port num-
ber, is used once the message arrives to
Overviews decide which of the many concurrently
1: Most software is originally designed running applications the message should
to work on one particular kind of com- be delivered to. Network-enabled applica-
puter under one particular kind of oper- tions obtain a port number from the oper-
ating system. Multi-platform software is a ating system on start-up, and use that port
highly commendable target, and certainly number as part of their unique address
is achievable, but it is unusual for software for receiving messages. Although a com-
that was designed to operate under one puter may have up to 65 500 network ports,
kind of system to work perfectly on another they are all just virtual constructs accessed
without a significant outlay of additional through the one physical network port at
development effort. It is very common for the back of the computer.
software to be developed without any spe- Serial ports: nine or fifteen individual
cial consideration for cross-platform com- connectors in two parallel rows. Serial ports

262
Port

are almost obsolete and usually go unused sometimes called Full USB) is now used
on modern computers. Serial ports were only on older or low-end equipment. USB
used to connect to external dial-up modems 2.0 (also called High-Speed USB) is fully
or very old mice. Communications through backward-compatible, so any USB 1.1 device
a serial port occur as a series of single bits may be plugged into a USB 2.0 socket. USB
in a long stream; since only a single bit 2.0 is rapidly replacing all of the other
can be sent at once, serial port communica- ports normally found on the back of a
tions are usually quite slow. The speed and computer (keyboard, mouse, serial, parallel,
other details of bit transmission are not and SCSI). At such high signal rates, it is
fixed, and must be set on a per-device basis. essential that good solid connections are
Some serial devices Auto sense the settings, made through properly constructed and
and are almost plug-and-play, whereas oth- undamaged cables; the only slight disad-
ers require an exact configuration to be vantage of USB 2.0 is that the cables tend
pre-set, which causes trouble when the to cost a little more.
instructions are inevitably lost. SCSI ports: SCSI (Small Computer Sys-
Parallel ports: 25 individual connectors tems Interface) became an ANSI standard
in two parallel rows. Until recently, parallel in 1981, and rapidly became the connec-
ports were the only low-cost means of con- tion of choice for the high-speed con-
necting external devices to a single com- nections needed for external disk drives,
puter with a fast transfer rate. Parallel ports CD-ROM burners, high-resolution scanners,
were used to connect printers and scanners, and other devices. Its original transfer
but have now largely been replaced by USB. speed of 40 000 000 bits per second was
PS/2 ports: small round sockets with six much faster than anything else available
individual connectors and a small rectan- for PCs or any other small-to-medium-sized
gular tongue, normally used for connecting computers. SCSI now has many versions,
a keyboard and mouse, but they are in the which vary in their levels of compatibility:
process of being replaced by USB. PS/2 ports Fast, Wide, Ultra, Ultra-Wide, Ultra-2, Ultra-
are usually color-coded, mauve for the key- 2-Wide, Ultra-3, and Ultra-320. The fastest of
board and green for the mouse. these are still faster than USB 2.0 (Ultra-2-
USB ports: small rectangular connectors Wide can transfer data at 640 000 000 bits
with a central tongue. USB (Universal Serial per second), but SCSI is far more expensive,
Bus) is a relatively new technology that always requiring costly shielded multi-core
adapts the old idea of serial ports to an cables, and usually requiring an additional
age of ‘‘smart” devices. A USB device has interface card in the computer.
some minimal processing power of its own,
and can accurately identify itself when Business value proposition
plugged into a system and recognize com- The term port has two major meanings, as
mands intended for it rather than for other described above. The first refers to trans-
devices. This means that a moderate num- plantation of software from one system
ber of USB devices may be connected to a to another. This may be highly desirable,
single port (through a Hub). Improved tech- and much depends upon the degree to
nology means that USB devices can com- which the original designer utilized func-
municate at speeds that were impossible tions specific to a particular computer or
for serial devices. There are two standards operating system. If the design is highly
in current use: USB 1.1, which allows com- specific in its use of a unique hardware
munication at 12 000 000 bits per second; resource, then portability will be limited
and USB 2.0, which allows communica- or may require significant effort. If the
tions at 40 times that speed. USB 1.1 (also system uses only standard, non-customized

263
Power protection

features then the system’s potential for References


portability is greatly enhanced. The man- r D. Groth (2003). A+ Complete (Hoboken,
agement of a software development team NJ, Sybex--John Wiley and Sons).
must enact policies requiring the use of r P. Gralla (2004). How the Internet Works
standard resources if portability of the final (Indianapolis, IN, Que).
product is desired. r L. Peterson and B. Davie (2003).
The second use of the term refers to Computer Networks: A Systems Approach
a computer’s physical connectors, typically (San Francisco, CA, Morgan Kaufmann).
located at the back of desktop computers,
Associated terminology: Network devices,
into which external devices are plugged
Cables and connectors, TCP/IP.
(although it is now common for USB con-
nectors to appear at the front of machines
due to their frequent use and the difficulty
in plugging devices into a small connector Power protection
slot at the back hidden from view). Physi-
cal connections are made for the internet, (UPS, Battery backup, Surge protection, Line condi-
and for devices such as printers, mice, scan- tioning, Power conditioning)
ners, and external hard drives. Typically, Definition: Devices designed to prevent damage to
the trend is toward standardization of con- computers and other electronic equipment caused by
nectors, and USB is now very much the or borne upon electrical power-supply lines.
favorite. As USB becomes the de facto stan-
dard, devices that connect through serial Overview
ports will become obsolete as computer Since the introduction of solid-state elec-
manufacturers drop support for this type tronics, the intrinsic reliability of the main
of device. Virtual ports, such as those used parts of a computer has become excep-
for internet connections, need to be consid- tionally high; CPUs fail only in fabricated
ered as part of the security assessment of excuses. Unfortunately, there are still exter-
the network, since ports connected to inse- nal sources of failure: any component with
cure applications allow hackers to illegally a moving part will eventually fail. One of
access a system and cause problems. the most significant causes of failure is the
power supply: even a computer with 100%
reliability will not work without electricity.
Summary of positive issues
There are four common kinds of power-
Standardization of ports on computers
supply problem that lead to computer fail-
allows connections between pieces of
ure, they are as follows.
equipment to be made more easily. The use
Power failure. Someone accidentally
of virtual ports allows a large number of
pulls the plug out, or a circuit breaker
connections to be made by devices with-
trips, or a power line needs to be repaired.
out the need for a large number of physical
For whatever reason, electrical power fails
ports.
to reach the computer. It is very rare for
this to cause any physical damage. What-
Summary of potentially negative issues ever operation the computer was perform-
The move to standardization of ports (USB) ing will be unsuccessful, but on restoration
will reduce support available for other, of power a computer is normally expected
older port types such as serial and paral- to remain in good working order. However,
lel ports. Virtual network ports need to be damage to data can be extensive. Applica-
secured to prevent intrusion from hackers. tions must be written with special care to

264
Power protection

ensure that data integrity is maintained the right voltage at all times. Particularly in
during a power failure. industrial areas, the turning on and off of
An Uninterruptible power supply (UPS) is large inductive loads (essentially anything
usually a rechargeable battery unit: it is with a big electric motor) can result in volt-
plugged into the normal mains electric- age fluctuations, both over and under the
ity supply, and the computer is plugged nominal value. Modern computers tolerate
into it. Normally, the UPS remains passive, a wide range of supply voltages, but there
but, when the mains supply fails, the bat- are always limits. A UPS constantly moni-
tery automatically takes over and temporar- tors the supply voltage, and should auto-
ily provides power to the computer. The matically switch to battery power not just
changeover is so fast that the computer is for total failures, but also for the duration
totally unaffected. The capacity of any bat- of excessive voltage swings.
tery is severely limited, so a UPS will keep Incorrect wiring. It is not unheard-of for
a computer working for no longer than a the electrical system in a building to be
few minutes. This is enough for most power incorrectly installed: the live and neutral
failures, and in critical installations should conductors can be crossed over, the ground
be enough time to start up an emergency conductor can be left unconnected, etc.
generator. These errors can lead to safety problems
UPS hardware ranges from small low-cost themselves, and can prevent other safety
devices that provide a few minutes’ pro- equipment from functioning. A surge pro-
tection to a small computer, to industrial- tector can not divert excess voltage to
strength units with their own electrical ground if the ground wire is disconnected.
generators that automatically start before Inexpensive wiring checkers are widely
the batteries become discharged, and could available, when plugged into the electri-
keep a whole installation running for days cal supply they indicate via red and green
without an external power source. lights whether the wiring is correct.
Surges. In areas with very primitive pub- Electro-magnetic interference (EMI).
lic utility infrastructure, electrical supply Whenever wires pass through a varying
lines are often above ground, exposed to magnetic field, stray currents are induced
the weather. If lightning strikes the dis- in it. All electrical equipment creates
tribution line, there can be an enormous varying magnetic fields, so the modern
surge of power with the potential to destroy office is something of a farm for induced
any connected electrical equipment. Lesser currents in wires. The introduction of
events (such as transformers switching) can high-frequency signals into electronic
cause lesser, but still damaging, surges. A equipment can have unfortunate effects,
Surge protector is a small unit connected in so power-protection equipment often
the same manner as a UPS. The most com- includes EMI filters. Many power cables
mon design uses a simple semiconductor have EMI filters built in; they are small
device, a Varistor, which shunts the excess thick tubes that appear to be clamped to
voltage to the electrical ground. The varis- the wire.
tors can be destroyed by a severe surge, leav-
ing the computer unprotected from subse- Business value proposition
quent surges: the indicator light should be UPS devices are used to prevent unexpected
checked frequently. A UPS usually incorpo- systems outages that can lead to the loss
rates a surge protector. of data, hardware damage, and business
Under- and over-voltage. It is very diffi- downtime. The devices are universally avail-
cult for a public utility to maintain exactly able, in a variety of specifications, and are

265
Privacy Act of 1974

easy to install, with relatively little main- Overview


tenance required. UPS devices allow sys- The Computer Matching and Privacy Act
tem administrators to bring a system down was first enacted in 1974 and provided a
in a controlled manner if the duration of legislative framework through which the
a power failure is expected to exceed the information and data collected on individ-
UPS’s battery life. uals (US citizens and legal aliens) could be
stored and compiled. The act has periodi-
cally been amended to keep pace with tech-
Summary of positive issues
nology. Its key provisions are to
UPS technology is understood and well
supported by vendors and manufacturers
(1) prevent unauthorized matching of
worldwide. The lifespan and quality of
disparate federal databases or the
units vary, but they can be purchased
matching of federal databases with
with 10-year warranties. UPS systems are
non-federal databases;
available in a wide variety of configu-
(2) ensure accountability by maintaining
rations (e.g., dual-voltage outputs). Many
records of who accessed which data
manufacturers provide extended-run bat-
and when;
tery options and ‘‘hot swapping”: exchang-
(3) ensure that only pertinent data is
ing batteries while the UPS is running a
maintained on individuals by federal
live system. Some vendors provide compli-
agencies;
mentary software to monitor the system,
(4) require that the name and location of
provide alerts, and shut down or restart
the ‘‘store” or database be published
equipment automatically, and to assist
in the federal register;
with maintenance and provide self-test
(5) maintain data that is used for making
logs.
‘‘determinations about individuals”;
(6) maintain accurate records, which
Summary of potentially negative issues includes allowing individuals to gain
There are different battery technologies access to their data and amend the
available; older ones such as nickel-- data when necessary;
cadmium have a ‘‘memory effect” that pre- (7) ‘‘maintain no record describing how
vents full charging after a partial discharge. an individual exercises rights
They need to be completely discharged in guaranteed by the First Amendment
order to regain their full operational capac- unless expressly authorized by statute
ity. Battery devices can only be a tempo- or by the individual about whom the
rary solution to a power outage and must record is maintained or unless
be supplemented with generator capabil- pertinent to and within the scope of
ity if longer-term secondary power-supply an authorized law enforcement
requirements are needed. activity”;
(8) enforce code-of-conduct rules for
technology workers maintaining and
developing the systems; and
Privacy Act of 1974 (9) enforce security measures to prevent
unauthorized access to information.
Definition: The US Computer Matching and Privacy Act
is a “code of fair information practices” that attempts These and other aspects of the act are over-
to regulate the collection, maintenance, use, and dis- seen and coordinated through the data-
semination of personal information by federal execu- integrity boards that are mandatory in
tive branch agencies. each federal agency.

266
Programming language

Business value proposition Business value proposition


The act provides a framework through Registered journalists are not required to
which US federal agencies must operate. divulge sources of data or information
The act also has provision for contrac- pertaining to work in progress to law-
tors working with US federal agencies and enforcement agents, thus allowing them to
details their obligations. examine potentially ‘‘sensitive” topic areas.
Encryption technology may be used to pre-
Summary of positive issues vent unauthorized access to data and work
The act protects personal information and in progress.
data at the federal level.

References Summary of positive issues


r D. Solove and M. Rotenberg (2003). Encryption mechanisms can protect data
Information Privacy Law (New York, during all phases of a journalistic inves-
Apsen). tigation. Data storage and data transmis-
r M. Rotenberg (2003). Privacy Law sion are vulnerable to unauthorized access,
Sourcebook 2003 (Washington, DC, e.g., internally, at the internet service
Electronic Privacy Information Center). provider, at the source, and at the destina-
r 5 USC §522a (2000). tion. Encryption technologies based upon a
wide range of techniques and with various
Associated terminology: Database, Law degrees of security are available.
cross-reference.

Summary of potentially negative issues


Privacy Protection Act of 1980 Encryption adds a layer of complexity
and overhead to the technology used by
Foundation concept: Security. the members of the journalistic project.
Definition: The US Privacy Protection Act of 1980 The accidental loss of encryption keys
protects the “work in progress” and “sources of infor- can result in sensitive information being
mation” used by journalists from access by law irrecoverably lost.
enforcement prior to public dissemination.
References
r D. Solove and M. Rotenberg (2003).
Overview
The US Privacy Protection Act of 1980 Information Privacy Law (New York,
(‘‘PPA”) (Title 42 United States Code Section Apsen).
r M. Rotenberg (2003). Privacy Law
42, 2000aa) enacts a provision based upon
the First Amendment of the US Constitu- Sourcebook 2003 (Washington, DC,
tion protecting ‘‘work in progress” and the Electronic Privacy Information Center).
‘‘sources of information” used by journal- Associated terminology: Encryption, ISP.
ists from access by law enforcement prior to
public dissemination. Government officers
and employees are forbidden from search-
ing for or seizing work materials that are Programming language
intended for publication in a newspaper or
book or for public broadcast. There are, of Foundation concept: Algorithm.
course, exceptions when danger to life or Definition: A compromise language, accessible to
limb is imminent, or when the author is computers and specially trained humans, that allows
believed to be involved in a related crime. instructions to be written clearly and unambiguously.

267
Programming language

Overview it automatically into the incomprehensi-


Computers and human beings do not work ble numeric codes that the computer can
in the same ways. Contrary to popular obey. It would be as though the computer
belief, computers are vastly ignorant and were programming itself. A human oper-
not in the least bit intelligent. The only ator could type “SET A=2; MULTIPLY A
thing they are good at is following instruc- 3; ADD A 5; PRINT A”, and the com-
tions, and even then those instructions puter would automatically convert it to ‘‘B8
must be written in exactly the right way. 02 00 00 00 6B C0 03 83 C0 05 . . . ,”
Even when it comes to calculations, the and obey those instructions. This was called
things that everyone knows computers are Automatic programming, and was originally
best at, computers are clueless. A computer viewed as an aspect of Artificial intelligence
would be incapable of working out a sim- (AI).
ple calculation like 2 × 3 + 5 unless some- The first program that was able to take a
one gave it exact step-by-step instructions. human-friendly set of instructions and con-
The level of detail required in the instruc- vert it automatically into the computer’s
tions given to computers can be quite mind- own internal language was an enormous
numbing. For example, to tell an Intel Pen- undertaking, consuming the efforts of the
tium how to work out 2 × 3 + 5, the instruc- established experts in programming (A. M.
tions are like this: ‘‘Put the number 2 into Turing, R. A. Brooker, and others) for a
A, then multiply whatever is in A by 3, great deal of time. The result was the
then add 5 to whatever is in A,” and those first Compiler and the first Programming lan-
instructions must be encoded in a spe- guage (known as Autocode). The benefits
cial computer-oriented form (called Machine were immediate and obvious. Programmers
code) like ‘‘B8 02 00 00 00 6B C0 03 83 C0 could spend most of their time on the truly
05” before they can be obeyed. All that just intellectual process of working out how to
addresses the simple calculation; it says solve a problem, since writing the solution
nothing of the far more complex opera- method in a computer-accessible form took
tions required to display the result. only a fraction of the effort. Also, many
In the very early days of computing, errors became easily detectable; if someone
that is exactly what programming meant. enters B9 instead of B8, it could easily go
First work out how to solve a problem, unnoticed, since both are valid, but if some-
then work out how to represent that solu- one enters MYLTIPLY instead of MULTIPLY,
tion method as a sequence of exceptionally the error is so obvious that the computer
small detailed steps, then encode it into the itself could detect it and issue a warning.
strange numeric form that the computer Since then, the development of program-
expects, then try to get that (now unread- ming languages and their compilers has led
able) program into the computer without to ever more sophisticated systems. A pro-
making any mistakes. Programming was a gramming language strikes a compromise
very difficult labor-intensive task that only between the computer’s internal numeric
a very few highly trained experts were capa- codes at one end of the spectrum and
ble of. straightforward human-oriented or math-
Early on, an inspirational breakthrough ematical notation at the other. The con-
occurred, and it was realized that, with tinuing development of programming lan-
an exceptional outlay of effort and inge- guages generally moves the point of com-
nuity, it would be possible to write a very promise ever closer to the human end
complex program that tells the computer of the spectrum, automating more and
how to take something written in a mod- more of the simpler (or less intelligence-
erately human-friendly form, and convert demanding) parts of the process. Modern

268
Programming language

languages allow the user to enter an a while it seemed to be gaining ground,


instruction as simple and clear as ‘‘Print and many government contracts required
2 × 3 + 5,” and have the computer work out the use of Ada, but it has now fallen far
everything else for itself. behind. The current languages C++ and
Ideally, the development of program- Java are just the latest generation of the
ming languages would be a linear process, Algol family.
with there being at every stage one lan- Currently, the leading language for pro-
guage which is the best we are currently fessional and technical programming is
capable of creating, and which every pro- C++, with Java a close second. Variants of
grammer uses. There has never been a situ- Basic are very popular with non-technical
ation like that. Designers of programming programmers because some programs can
languages have their own personalities and be constructed without training in pro-
motivations, sometimes take a lot of pride gramming, and Visual Basic provides simpli-
in their creations, and don’t always see fied interfaces to GUI (Graphical user inter-
what is best for their users until it is too face) components and common office appli-
late. There has always been a wide vari- cations. JavaScript, Python, and other script-
ety of languages in use at any one time, ing languages enjoy high popularity due to
with a few (but never just one) recognized the ease of performing quick experiments
as the leaders of the pack. Languages cer- and embedding small programs inside doc-
tainly do improve with time, but there is uments.
a lot of inertia, since it takes significant Legacy languages, particularly Fortran
effort for a programmer to mentally retool. and Cobol, survive from the late 1950s
Even if one programming language were and early 1960s. They survive because of
clearly superior to all others, the chance the large amount of useful code that
of it being universally adopted would be was written when they were the standard
about the same as the chance of the entire languages; little new development takes
world deciding to speak English exclusively. place now. Cobol was for a very long time
Furthermore, there will always be the need the standard language for business applica-
for future improvements: language devel- tions, as Fortran was for scientific program-
opment is often driven by new discoveries ming. Folklore still bizarrely insists that
regarding what can be done. Fortran is more efficient for scientific cal-
There have been well-meaning attempts culations.
to create the universal programming lan- For special purposes, various targeted
guage. Starting in 1958, an international languages exist. Every computer has its own
committee defined the International Pro- Assembly language, which is used for small
gramming Language IPL, which in 1960 parts of new operating-system and com-
became Algol. This was a very successful lan- piler development. AI research makes sig-
guage, and was the basis of a large propor- nificant use of Logic-programming languages,
tion of university teaching for the next 20 such as Prolog, and so-called functional lan-
years. As new discoveries were made, Algol guages like Lisp. There are true functional
gave rise to a whole family of languages, languages, which have a pure mathemat-
including the very complex but seminal ical semantics and are useful in research
Algol-68, and, in a backlash, the very pop- into programming languages and software
ular but now equally extinct Pascal. During engineering.
the 1980s and 1990s, under the direction The standard paradigm for programming
of the US Department of Defense, another is called the Imperative or Procedural style.
attempt at a universal language derived In the imperative style, a program tells
from Algol, called Ada, was developed; for the computer explicitly what to do as a

269
Programming language

sequence of ordered steps. Without spe- the de facto standard language for scien-
cial training, programmers tend to think tific programming and Cobol became the
that this is the only way programming de facto standard for commercial systems
could be. An alternative is the Functional development. Various attempts were made
style, in which a program is a set of to produce a language that would be good
pure mathematical functions that specify for everything; PL/I is the most notable,
the correct mapping from inputs to out- combining features of Algol, Fortran, and
puts without saying what order the indi- Cobol into one big confusing whole. Pro-
vidual steps of the computation must fol- gramming was also performed in a vari-
low. In the pure functional style every- ety of other languages that ranged from
thing involved in a program is strictly con- assembly language to system-command lan-
stant, but many languages follow the form guages such as JCL (Job Control Language),
of functional programming without follow- which was used by systems programmers
ing the rules. Lisp is an example of this: to ensure that programs were executed cor-
programs are thought of as functions, but rectly, and that they had access to the cor-
in the mathematical sense they are not, rect resources of the system.
and Lisp does not provide the full ben- The weight of maintaining and develop-
efits of true functional programming. In ing the huge amount of Fortran and Cobol
the logic-programming style, programs are code, which was typically poorly designed,
simply logical formulæ specifying the cor- structured, and developed, was one of the
rect relationship between questions and factors that led to their decline. In the
answers, not stating how that relationship 1980s and 1990s programmers were influ-
is to be realized. The non-procedural styles enced by the need to develop programs that
require great expertise and deep theoret- were more in alignment with the needs
ical understanding on the part of a pro- of the end user rather than the technolo-
grammer, but, once mastered, they provide gist. They also needed to develop code that
major advantages, including rapid proto- was able to operate on a range of systems
typing and the possibility of Proving a pro- that included PCs and workstations, and
gram correct. The Object-oriented style is a would work correctly on various operating
view of how data and computations should systems. Managerial considerations were
be encapsulated in a program, and may also becoming a key influence on which
equally be applied to all three of the major programming languages lived and which
paradigms. The two popular languages, died. Project managers were keen to ensure
C++ and Java, are both imperative and that any code developed was reusable and
object-oriented. structured, and could be maintained by
any programmer versed in that language.
Business value proposition The programming languages also needed
Programming languages come in all styles to facilitate inter-program communication,
and flavors. They can generally be classi- since many systems were becoming either
fied in terms of their ages and ‘‘eras.” For embedded inside other systems, or con-
example, the mainframe era of the 1960s nected as part of a heterogeneous program-
and 1970s was dominated by Algol-type ming environment. This led to the emer-
languages and the two major commercial gence and adoption of several key current
languages Fortran and Cobol. The Algol languages, chief amongst them C, C++,
style facilitated the development of creative Java, and Visual Basic.
programming techniques and led directly While the C programming language
to the modern languages Pascal, Modula, gives programmers the power to perform
BCPL, C, C++, and Java. Fortran became assembly-level types of operations within

270
Programming language

higher-level programs and to access all to become difficult to maintain, and will
aspects of the computer’s hardware, it is become a heavy legacy environment within
too easy for inexperienced or untrained the US Department of Defense, since very
programmers to abuse or accidentally mis- few programmers are being trained in
use the language. This was corrected to Ada, and it is not an easy language to
a large degree with C++, which retained pick up.
the power of C but provided a more struc- In the 1980s and 1990s there was also a
tured programming environment. C++ resurgence of interest in AI programming
has proved to be a popular programming systems and languages, with Lisp and Pro-
language; however, it is a very complex lan- log becoming popular for the creation of
guage to learn and to extract the full poten- knowledge-based systems. The prominence
tial from. One answer to this was the cre- of these systems has declined since then,
ation of Java, which offers many of the fea- however, but many of these systems are
tures of C++ but simplifies the program- still in operation. Some KBS programs have
mer’s workload. Visual Basic provides non- been translated into C or used through
technical developers with a mechanism ‘‘shell” environments. Again, these legacy
through which they can develop simple systems pose major maintenance problems
programs, and is considered especially use- for project managers because the support
ful for creating systems with graphical user environments that were available at their
interfaces of the type familiar to office and creation have disappeared or not been
commercial users. updated. Lisp and Prolog require special
In the 1980s there was a major move- training, since they do not work in the
ment toward developing one standard pro- manner familiar to most programmers.
gramming language that would be good The use of programming within many
for all purposes. The movement was led organizations to create applications has
by the US government, and resulted in the been on the decline since the 1990s as CIOs
programming language Ada. In 1987 The US began to understand the value of using
Department of Defense mandated the use commercially produced applications such
of Ada for nearly all projects under its con- as enterprise resource planning (ERP) sys-
trol, with directive 3405.2 ‘‘Computer Pro- tems to replace many home-developed pro-
gramming Language Policy,” which stated grams. Interestingly, a leading ERP vendor,
that ‘‘The Ada programming language shall SAP A.G., chose to write their system in
be the single, common, computer pro- their proprietary language ABAP/4, which
gramming language for Defense computer not only discourages customization of an
resources used in intelligence systems, for SAP ERP system (typically a good thing) but
the command and control of military also ensures that SAP A.G. has complete
forces, or as an integral part of a weapon control over the development of the ABAP/4
system.” For various reasons, the Ada move- language and as a consequence their ERP
ment was unsuccessful, and in 1997 the application.
policy was reversed. A new memorandum The role of programming languages in
from the Assistant Secretary of Defense the future is likely to continue along a
stated ‘‘I have directed my staff . . . to elim- development path that increasingly allows
inate the mandatory requirement for use programmers to create systems that are
of the Ada programming language in favor able to interact easily, use the features
of an engineering approach to selection of of the web, provide high-quality graphical
the language to be used.” As a consequence interfaces, and operate equally well upon a
of this there exists ten years’ worth of range of operating systems and hardware
Ada code in critical systems that are likely platforms.

271
Protocol

Summary of positive issues Overview


A very wide range of programming lan- Protocols cover a very wide range of sit-
guages is available, including assembler, uations, from remembering to bow and
functional, procedural, logical, and object- say ‘‘Your Majesty” when speaking to the
oriented programming. Modern program- Queen, through the Kyoto Protocol which
ming languages facilitate structured pro- seeks to govern how nations interact in con-
gramming, reusability, and portability. trolling and trading carbon dioxide emis-
There is a huge literature associated with sions, HTTP, the language of communi-
programming, systems design, compiler cations between web browsers and web
design, and software project management. servers, and IEEE 802.3, which specifies how
messages are communicated on an ether-
net network, to X3.131, which defines the
Summary of potentially negative issues operations on a SCSI bus. A protocol is sim-
Poorly designed and written programs can
ply an agreed set of rules that govern how
cause disaster, and programming correctly
communications are to be performed, no
requires a significant degree of training.
more and no less.
Some programming languages encourage
Since modern computing is dominated
poor programming unless great care is
by communications, protocols abound. If
taken in all stages of design and implemen-
an acronym ends in the letter ‘P’, there is
tation. Incorrect choice of programming
a good chance that it is a protocol name
language can lead to expensive failures or
(FTP, HTTP, ICMP, SMTP, etc.). Some are
poorly performing systems. The wide vari-
backed by the authority of major standards
ety of programming languages available
organizations, such as the ISO, ANSI, and
can be a negative aspect: nobody is expert
BSI; some are proprietary systems belong-
in, or even knowledgeable about, all of
ing to individual corporations; some are ad-
them.
hoc creations designed by a programmer to
References meet an immediate need.
r R. Wexelblat (1981). History of When a protocol exists for a given sys-
Programming Languages, Volume 1 (New tem, it is normally considered to be essen-
York, Elsevier). tial to follow that protocol exactly. Failure
r T. Bergin, R. Gibson, and P. Gordon to do so usually results in failure of the sys-
(1986). History of Programming Languages, tem. However, it is not unknown for prod-
Volume 2 (New York, Addison-Wesley). ucts to be made deliberately to diverge from
r Memo: ‘‘Command, Control, the governing protocol, in an attempt to
Communications and Intelligence,” undermine a protocol seen as unfavorable
Assistant Secretary of Defense, 6000 to the developer’s interests, to add a unique
Defense, Pentagon, Washington D.C. ‘‘finger print” to the product so that unau-
20301--6000 April 29, 1997. thorized use of intellectual property may
be detected, or in an attempt to add ‘‘out-
Associated terminology: Compiler, C++, of-band” communications undetectable by
Java, Fortran, Cobol, Logic programming. compliant components. Whatever the rea-
son, the result of variation is often
failure.

Protocol
Business value proposition
Definition: A set of rules governing the interactions The ever-increasing ability for comput-
between a number of systems or components. ing devices to connect together, to run

272
Proxy

programs, and to enable applications


Proxy
to interact with each other is directly
attributable to the adoption of standard Foundation concept: Network.
protocols by manufacturers and vendors. Definition: A system on a network that “stands in” for
Protocols have, since the 1980s, moved another, performing some or all of its normal tasks.
away from proprietary corporate ownership
toward being open and accessible, usually
with documentation available for down- Overview
load over the internet. In general terms, a Proxy is a system that
Many protocols are managed by for- intercepts communications intended for
mal bodies such as ANSI, ISO, and W3C, another system and deals with them itself.
which foster development through working Proxy servers perform a useful function
groups. These groups issue calls for com- when a local-area network (LAN) is pro-
ment from their membership, and steadily tected by a Firewall. The firewall may be
develop existing protocols, keeping them configured to prevent any access to com-
up to date. puters inside the LAN from any outside,
It is important for all organizations, espe- perhaps as a security measure. When some
cially those heavily involved with devel- controlled access is desirable, reconfigur-
opment of leading-edge technology, to be ing the firewall might not be an accept-
aware of the protocol formulation process able solution; instead, a proxy server may
and to play an active part in the work- be used. The firewall may be pre-configured
ing groups that develop them. This ensures to accept any communications from the
that they are not caught by surprise by designated proxy, and the proxy may be
any new protocols or changes to existing easily configured to accept all communica-
protocols. tions destined for the protected network,
and filter them. Those that are considered
innocent are forwarded; others are simply
Summary of positive issues ignored. A strong firewall together with
The development of open protocols is accel- a flexible (but secure) proxy server allows
erating and the move toward standardiza- complete control over access to and from
tion on commonly used protocols that are the internet.
non-proprietary will reduce administrative Another use for proxies is to provide
and cost overheads associated with main- a fast-access Cache of frequently requested
taining software compliance. material. For example, if there are certain
web documents that are very frequently
needed by users in a LAN, copies of those
Summary of potentially negative issues
documents can be kept on a local proxy
The fact that protocols change frequently
server. All web accesses can be intercepted
requires that organizations be aware of the
by that server, and inspected. If an access
changes and their implications. The fact
is requesting one of the already-held doc-
that a protocol exists does not imply that
uments, the copy may be sent back imme-
it is good; many poorly designed or out-of-
diately in response without requiring any
date protocols exist and some are perpet-
further traffic on the external network. If
uated for reasons of self-interest by their
the requested document is not already held
designers rather than in the true interest
on the proxy, the request would be retrans-
of their adopters.
mitted to the real server. The proxy may
Associated terminology: Internet protocol, be configured to monitor the response and
TCP/IP, ANSI, ISO. keep a copy of the new document so that

273
Public key–private key

future accesses to the same may be handled the corporate network, and the potential
locally. for using third-party providers for specialist
This second kind of proxy is also used services.
more controversially to filter internet com-
munications. Proxies may intercept web- Summary of positive issues
page accesses, file transfers, and even Proxy servers are strong mechanisms for
email messages, and scan for content that protecting corporate networks. The tech-
is deemed somehow inappropriate. Appli- nology is mature and widely understood by
cations of this range from corporations network managers. Proxy servers may act
attempting to prevent leaks of industrial as a mechanism for servicing web requests
secrets, or to prevent employees from wast- quickly without having to route every
ing time with web browsing, to shielding request internally and thereby increase net-
minors from explicit sexual material, or work traffic. Third-party providers are avail-
even to preventing customers from seeing able to offer specialist services such as
rivals’ materials. spam filtering.
An even more controversial use of prox-
ies involves modifying material that is sent Summary of potentially negative issues
or received without the author’s consent or Proxy servers that are used to filter spam
even knowledge. If a user can access the and other traffic need to be maintained
internet only through a proxy, then that and monitored to keep them current and
proxy has complete control over all of their effective.
communications. One application of this is
References
adding Banner advertisements to email mes- r P. Gralla (2004). How the Internet Works
sages, but far more sinister uses are easy to
(Indianapolis, IN, Que).
imagine. r L. Peterson and B. Davie (2003).
Computer Networks: A Systems Approach
Business value proposition
(San Francisco, CA, Morgan Kaufmann).
Proxy servers provide a mechanism
through which a network may be pro- Associated terminology: Internet
tected from unwanted external messages. protocol, DHCP, Firewall, Network-address
They also add value by providing a mecha- translation.
nism for servicing requests without having
to place unnecessary extra loads upon the
corporate network. A proxy server, whether Public key–private key
acting as a filter or providing redirection
services, does not need to be physically loca- Foundation concept: Security.
ted at the business, but may instead reside Definition: An encryption system using two keys, one
at the ISP telecommunications company’s for encryption and a different one for decryption; one
site, reducing the bandwidth requirements key is made public, whereas the other is kept secret.
on the corporate internet connection. The
provision and maintenance of proxy servers Overview
may in many cases be sensibly outsour- Public-key/private-key systems (often just
ced to a web-hosting company. called Public key) are based on asymmet-
While there are costs associated with the ric Encryption algorithms that require two
setup, maintenance, and operation of proxy separate keys (or pass-codes) for each com-
servers, the cost is offset by the decrease munication. Data that was encrypted using
in internal network traffic that results, the one key can be decrypted only by using the
reduction in malicious data traffic entering other key, and vice versa; it is not possible

274
Public key–private key

to deduce one key from the other. RSA Total secrecy: a message from person A
(named after Rivest, Shamir, and Adleman, to person B may be encrypted twice, once
the system’s inventors) is the best-known using person A’s private key, then again
and most respected of such systems. using person B’s public key. The result is
For each individual participating in a that only person B can read the message
public-key/private-key system, one matched because only they have the required pri-
key-pair is created. The keys are usually very vate key, and person B can be confident
large numbers, hundreds of digits long, so that it really did come unmodifed from
that a Brute-force attack, systematically try- person A.
ing every possible key to decrypt an inter- Private records: naturally, any partic-
cepted message, is impossible. One of those ipant may encrypt their own personal
keys is given in absolute secrecy to the secrets using their own public key, and be
individual; it is known as their Private key. secure in the knowledge that only they can
The private key must never be revealed or decrypt them.
left unsecured under any circumstances. Key exchange: The asymmetric encryp-
The other key, known as their Public key, tion systems needed for public-key systems
is published as widely as possible, and are more thoroughly trusted, but much
becomes part of the individual’s public digi- more computationally demanding than the
tal identity. Public keys must be well known usual symmetric systems, so the transmis-
and easily accessible. This arrangement per- sion of large amounts of data can be very
mits a wide variety of secure communica- difficult. A valuable technique for large
tions. secure transmissions is to make up a totally
Digital signature: if you wish to make a new key for each transmission, and send
statement in such a way that everyone read- that key only using the slow but very secure
ing it can have absolute confidence that public-key system. Once the recipient has
what they are reading is exactly what you received and decrypted this one-time key
wrote, simply encrypt it using your own (or Session key), it is used to encrypt the
private key. Everybody in the world has confidential data under a fast but perhaps
access to your public key, so they can eas- less trustworthy symmetric encryption sys-
ily decrypt it and read the message; the tem, and never used again. Many cryptan-
fact that your public key did successfully alytic attacks rely on having a collection
decrypt the message is proof that your pri- of intercepts all encrypted with the same
vate key must have been used to encrypt it. key, so the ability to use a new key for each
Only you have your private key, so you must transmission, and be sure that that key can
have sent the message. Successful encryp- not be intercepted, provides a significant
tion methods are carefully designed so that increase in security.
the tiniest change to an encrypted message
will result in absolute gibberish when the Business value proposition
message is decrypted, so illicit modifica- The use of public-key/private-key technolo-
tions are impossible. gies has a wide range of applications. Pri-
For your eyes only: if you wish to send a mary amongst them is the encryption of
message and be sure that only person X can data to be communicated between two
read it, simply encrypt it using person X’s or more parties and of data to be held
freely available public key. The encrypted within a company or by an individual.
message may be sent over a completely Emerging uses include the incorporation of
open and insecure channel; it doesn’t mat- encryption into a variety of processes such
ter who intercepts it, since only person X as voting and election mechanisms, smart
has the private key required to decrypt it. cards, digital time stamps, authentication

275
Public key–private key

methods, and watermarking of intellectual known as Chosen plaintext. If a code breaker


property (e.g., music). is able to design a specific message, and per-
suade somebody to encrypt it with their
Summary of positive issues private key, and is able to intercept the
A variety of public-key/private-key encryp- result (so that they have in their possession
tion mechanisms have been openly pub- both the encrypted and the unencrypted
lished and are widely available. These version of a carefully chosen text), it can
mechanisms enable data to be encoded and result in the key being deduced. The sim-
decoded with relative ease in an automated ple but essential rule is ‘‘Never encrypt a
manner. The mechanism can also be used stranger’s message.” When encrypting for
to encode data for use within an organiza- the purposes of digital signature, always
tion for security properties. The fact that make a cosmetic change, perhaps just to
an encryption mechanism has been widely the spacing of the document.
published is a strength, not a weakness:
References
many cryptanalysts, not all of them emplo- r B. Schneier (1996). Applied Cryptography
yed by secret government agencies, will
(New York, John Wiley and Sons).
have tried to crack it; keeping success secret r K. Schmen (2003). Cryptography and
for a long time is almost inconceivable.
Public Key Infrastructure on the Internet
(New York, John Wiley and Sons).
Summary of potentially negative issues
Many public-key/private-key encryption sys- Associated terminology: Encryption,
tems are vulnerable to a cracking technique Password, One-way hash.

276
Quantum computing

with 30 gates, over 1 000 000 000 different


Quantum computing
states could be achieved simultaneously;
Foundation Concept: Bit. the numbers are unlimited. The ability to
Definition: The use of quantum mechanics, the laws perform any number of computations in
governing sub-microscopic particles, in the construc- the time currently taken by one computa-
tion of computing and communication devices. tion would have an incredible effect in all
areas.
Overview The second feature is easier to picture.
Quantum computing is firmly in the domain Two particles can be entangled, which
of science fiction, and can be expected to means that they are always tied together
stay there for quite some while. Very basic in the same state, regardless of how far
experiments confirming that the underly- apart they are or what objects may separate
ing ideas are at least viable have been per- them. If they are thought of as switches,
formed successfully, but the construction then turning one of them on instantly
of usable computing or communications results in the other one turning on (and
devices is at best very distant. We are cur- vice versa). This provides a method for
rently at a stage akin to having discovered transmitting logic signals within a quan-
that electricity does actually exist; people tum computer, but is also believed by
can start thinking about the possibility of some to provide a possible mechanism for
building computers, but the possibility is faster-than-light communications: two peo-
far off, and it may never happen. ple vast distances apart could communi-
Quantum mechanics governs the behav- cate by repeatedly switching the state of a
ior of very small particles, single atoms and shared pair of entangled particles. It would
smaller, and shows that they do not behave be impossible for any interloper to ‘‘tap
as experience of the larger world would into” such communications. This possibil-
lead us to expect. The two key concepts that ity remains highly controversial.
might turn out to be practically useful are
Superposition and Entanglement. Business value proposition
In conventional computing, a switch or Quantum computing does not yet exist in
logic element is either on or off, 1 or 0. any practical form. If it ever does, it will
Ten logic elements give ten things that can completely change the face of computing,
be on or off in any combination, result- and probably all other human endeavors,
ing in 2 × 2 × 2 × 2 × 2 × 2 × 2 × 2 × 2 × in ways we can barely imagine.
2 = 1024 different states. A search through
1024 possibilities is performed by stepping Summary of positive issues
10 such elements through all possible com- Quantum computing could potentially pro-
binations of states in turn. A single logic vide an unlimited increase in the power
element at the quantum scale does not and speed of any computing device, and
have to be either on or off, it can, by virtue perhaps even permit instantaneous very-
of the principle of superposition, be both low-power secure communications.
on and off at the same time. Ten logic ele-
ments can all be both on and off at the Summary of potentially negative issues
same time, which means that in combi- It will not exist for the foreseeable future.
nation they can be in 1024 states at the
same time. Thus 1024 different computa- Reference
tions could be performed simultaneously r M. Hirvensalo (2001). Quantum
with one set of circuitry. Of course, the Computing (New York, Springer-Verlag
number 10 was chosen just as an example: New York, Inc.).

277
RAID (redundant array of independent disks)

known as Striping, and may be extended to


RAID (redundant array of
more than two disk drives for even greater
independent disks) speed-ups.
Foundation concepts: Disk, Storage. RAID, a redundant array of independent
Definition: A system for making a number of disk (or sometimes inexpensive) disks, is a stan-
drives act as a single one, for improved access speed, dardized system based on these two ideas.
reliability, or both. There are seven different forms of RAID,
known as Levels 0--6, which use different
Overview forms of mirroring, striping, and a combi-
Of all the critical components of a com- nation of the two.
puter system, the disk drive is by far the RAID level 0 uses only striping, with any
most unreliable, and often the worst bot- number of disk drives configured to behave
tleneck for speed. as a single one. It provides a great improve-
A very simple system for gaining ment in speed, but a loss in reliability,
almost perfect reliability would be to since a failure in any one of the disk drives
have two identical disk drives instead of results in the failure of the whole system,
one, and store every piece of data on and, with two disk drives, failures are twice
both drives simultaneously. Naturally, both as common as with one.
drives could fail, but the chance of both RAID level 1 uses only mirroring, usually
failing at exactly the same time is beyond with exactly two drives configured to act as
remote. As soon as one drive starts to one. It behaves exactly as described above,
exhibit problems it can be replaced with providing exceptional reliability with no
a new one, primed with a fresh copy of all cost in terms of performance.
the data. This system is known as Mirroring; RAID levels 2 and 3 use striping and
it gives almost perfect reliability, and has a reduced form of mirroring. The data
no other impact on performance. Having is spread across multiple drives, together
three identical disks at all times allows with a Digest of that data (known as Par-
replacements to be made as needed with- ity bits, Hamming codes, and Error-correction
out interrupting system operation. codes). This gives enough information to cor-
A relatively simple system for improv- rect small errors, and reliably detect larger
ing access speed can also be implemented ones.
with multiple disk drives. One of the lim- RAID levels 4--6 use ever more complex
itations on disk speed is the capacity of systems of striping, mirroring, and error
a single Track: only one track of data can correction, to achieve different balances of
be read in a single operation without a performance, reliability, and cost. The orig-
delay caused by switching to another track inal idea of RAID came from research car-
(see Disk for further details). Tracks typi- ried out at the University of California in
cally contain only 50 000--100 000 bytes of Berkeley in 1988. It quickly became very
data, so reading a typical file can involve popular, and is now an industry standard;
a great many track-switching delays. If the relatively inexpensive PC motherboards are
contents of a file are carefully distributed widely available with built-in hardware
between two disk drives, so that perhaps support for RAID arrays.
odd-numbered tracks are stored on one and
even-numbered tracks on the other, access Business value proposition
speed can be doubled, with data being read RAID technology allows data to be stored
from one drive while the other is waiting with a variety of options to improve data
for a track switch. This simple scheme is storage and recovery. Each RAID level has

278
Rapid application development (RAD)

its strengths and weaknesses that are deter- References


mined by the level of fault tolerance, r D. Patterson, G. Gibson, and R. Katz
the input and output data rates that can (1988). ‘‘A case for redundant arrays of
be employed, the reliability levels, the inexpensive disks,” in SIGMOD 1998,
technical performance levels of hardware Proceedings ACM SIGMOD International
required, and the cost. A cost-based busi- Conference on Management of Data, ed.
ness case must be made to determine L. M. Haas and A. Tiwary (Seattle, WA,
the technology level required for a given ACM Press), pp. 109--116.
organization. r D. Vadala (2002). Managing RAID on
Linux (Sebastopol, CA, O’Reilly Press).
Summary of positive issues
Associated terminology: Backup,
The technology associated with the basic
Information-lifecycle management.
open industry-standard RAID levels is
mature and, as disk storage costs decrease,
the deployment of larger RAID systems
Rapid application development (RAD)
becomes more cost effective in terms of
dollars per megabyte. RAID continues to Foundation concepts: Software development life-
be developed and some companies have cycle.
created proprietary RAID levels to address Definition: Rapid application development is a soft-
specific needs they perceive (or wish to ware development methodology designed to improve
create) in the market. The hardware and the overall speed of system implementation.
software associated with RAID systems are
highly reliable and very large data stor- Overview
age is available. Each RAID level has its The Rapid application development (RAD) app-
own positive dimension: RAID 0 is simple roach to software development has its roots
and easy to implement; RAID 1 has 100% in iterative prototyping and computer-
data redundancy, protecting against a disk aided software engineering (CASE), both of
failure; RAID 3 and 4 facilitate very high which were popular approaches to software
data-transfer rates and have a very efficient development in the 1980s. The approach
mechanism for limited error correction; was first developed as an iterative software
RAID 5 has the ability to write data at a development methodology by Scott Shultz
very high data rate; and RAID 6 provides at DuPont in 1984, and was refined and
extremely good error recovery, but incurs popularized by technology author James
a substantial penalty when data is written Martin in 1991.
onto the disks. The RAD approach is to develop software
in pieces, using software tools to speed the
Summary of potentially negative issues development of prototypes. It uses teams
Each RAID level has its own limitations: typically composed of four to six members
RAID 0 is not actually a ‘‘redundant” mech- that include managers, users, and develop-
anism since it is not fault-tolerant; failure ers, and the aim is to have a team develop
of any one disk can result in significant a piece of the total system in 4--6 months.
data loss. RAID 1 is very inefficient in terms A significant degree of support is afforded
of its use of storage capacity. RAID 5 can to the teams through software tools and
be difficult to rebuild in the event of data development environments. The project
failure. The higher RAID levels can be very methodology underlying RAD is a standard
complex to implement, but that is only a phased model, with an orientation phase,
concern for a very few developers. a training phase, business modeling, data

279
Reliability

modeling, process modeling, code genera- to each segment, may result in a loss of
tion, testing, and, finally, production. Ide- overall consistency and deviation from the
ally the RAD team members will have original specification.
undertaken significant prior training and
Reference
will be ready to go, allowing rapid com- r J. Kerr and R. Hunter (1994). Inside RAD
mencement and execution of a project.
(New York, McGraw-Hill).
Business value proposition Associated terminology: Joint application
The RAD methodology is well known development, CAD/CAM.
amongst developers and versions of it
are incorporated into the development
approaches used by many consulting orga-
nizations. The approach is well supported Reliability
and tools are available to developers. It
relies upon iterative prototyping and this Foundation concept: Software development lifecycle.
enables the development team to deliver
prototype solutions to users more quickly Overview
than would be possible using stage-based People normally think of ‘‘one in a million”
methods (such as the waterfall method), as being the archetypal statement of long
which present solutions only at the end of odds, and a failure rate of one in a mil-
the development cycle. lion describes a very reliable thing. How-
ever, a modern computer system can exe-
Summary of positive issues cute 2 000 000 000 operations or more per
RAD has the potential to deliver software second, and an error rate of one in a mil-
solutions faster than stage-based models lion means that it will go wrong 2000 times
do. It allows the development of compo- every second. That would be totally useless.
nents in parallel and simplifies the amend- The reliability required of even cheap toy
ment of components should the specifica- computer systems is higher even than that
tion change during development. RAD has required of pacemakers and rocket-ships,
significant support in the form of soft- not because of any critical need, but sim-
ware tools and consultants. RAD works well ply because the speed of operation magni-
with other techniques for structured sys- fies the probability of error.
tems development such as Joint application The reliability of electronic hardware is
design (JAD). usually expressed in terms of the Mean time
between failures (MTBF). If a large number
Summary of potentially negative issues of identical devices were all turned on at
The RAD approach requires that the teams the same time, and used continuously until
be trained in the RAD techniques and phi- they were all broken, the MTBF would be
losophy. The approach works best when the average time that each one survived.
the scope of a project is constrained and The arithmetic of failure rates is not nec-
good use can be made of tools and Appli- essarily intuitive. For example, consider a
cation program interfaces. Owing to the frag- system consisting of three components, A,
mented nature of the development, with B, and C. A has an MTBF of 1000 hours, B
the project development being divided has an MTBF of 1000 hours, and C has an
amongst many groups, optimization issues MTBF of 2000 hours. What is the overall
need to be carefully considered if speed MTBF of the system? The solution is to con-
constraints are important. The fragmented sider a long period of time, such as 100 000
aspect of the development, with user input hours. In 100 000 hours, we would expect

280
Reliability

component A to fail 100 times, B to fail 100 Sadly, most manufacturers can only guess
times, and C to fail 50 times, giving a total at the number of faults contained in their
of 250 failures in 100 000 hours. This trans- software.
lates back to an MTBF of 400 hours for the The fact that automated space probes
whole system. manage to reach their destinations in
Unfortunately the situation is not always the outer solar system after traveling for
so straightforward. Component failures are many years, and successfully beam back
not usually evenly distributed. Consider a photographs of previously undiscovered
simple battery as a counter-example: a new moons, is a testament to the fact that peo-
battery put into an electric clock usually ple can make reliable software if they really
works for about a year; perhaps the MTBF try hard enough. One might find oneself
for clocks with new batteries is 10 months. asking why software for home and com-
If you have 100 new clocks, and put a new mercial use is so unreliable. Sadly, in most
battery in each, all at the same time, you cases the answer is simply that it can be.
will not find the clocks failing at an aver- Unlike all other engineering professions,
age of 10 per month. Many months will there is absolutely no professional regula-
pass without a single clock failing, then, tion for software engineers; expensive soft-
at around the 10-month mark, they will all ware is unguaranteed, and even comes with
fail at times very close together. The MTBF user agreements explicitly stating that it
for the system of 100 clocks will be only might not work and that the manufac-
slightly less than the MTBF for one of its turer will not take responsibility. It costs
component clocks. It is this non-linearity more to produce a good product than a bad
of the failure distribution that makes sys- one, people continue to purchase software
tems consisting of literally millions of com- from manufacturers of other software that
ponents (such as computers) able to work they already know to be unreliable, and
at all. well-trained programmers cost more than
The reliability of software is of a some- untrained ones. Human nature and market
what different nature. Unlike physical com- forces ensure that general production soft-
ponents, software that was constructed per- ware will remain unreliable for as long as
fectly does not break after extended use it is allowed to.
and suddenly fail. No matter how heav-
ily it is used, there is no wear and tear Business value proposition
on software. If software ever fails, that The ability to create reliable software is
means that it was faulty right from the a function of the development methodol-
very beginning, and the fault simply had ogy utilized to create that software. The
not previously been observed. Expressions range of software development styles and
of the mean time between failures for soft- methodologies is wide, ranging from the
ware are somewhat deceptive; what they very informal styles in which the system
really express is how long the manufac- requirements are informally encoded, if at
turer expects you to be able to use the all, in informal language, through to the
software before becoming aware of a fault use of mathematical equations to specify
that is already there. Manufacturers can required behavior, a type of development
only guess at how users will use their soft- known as Formal methods.
ware, so the mean times are no better than The discipline of software development
guesses themselves. The only real measure has not yet evolved to the point at which
of the reliability of software is the num- the programmers, analysts, and other tech-
ber of faults it already contains. If it has nologists involved in the development
no faults, it is reliable; if it doesn’t, it isn’t. processes have all become ‘‘professional

281
RFID (Radio-frequency identity) tags

engineers” in the sense that architects or orously developed but probably would not
civil engineers have in their professions. wish to incur the cost and make the effort
Professional bodies such as the ACM, IEEE, required. Thus the management issue on
and BCS have developed professional codes software reliability comes down to ques-
and membership levels such as Member tions of balancing the total cost of own-
or Fellow, and further accreditation lev- ership of the software over its life, includ-
els such as Chartered Engineer and Char- ing liability issues, against return on invest-
tered Scientist have also been designated, ment for a project.
but there is no requirement for practition-
ers to be members of any professional body, Summary of positive issues
nor is there any enforcement of minimum The reliability of software and hardware
standards of competence. within a computer system can be managed
Programmers and other computer tech- by developing software through the adop-
nologists do not need any accreditation, tion of appropriate levels of formality in
and any responsibility for product qual- the development process using the SEI’s
ity rests solely on the management team. developmental guidelines.
Managers can demand that everyone in
their development team should be mem-
bers of a professional organization such Summary of potentially negative issues
as the IEEE and have Chartered Engineer Without strong management and process
status in order to work on a project, and controls in place, the development of soft-
they can further stipulate that the team ware is subject to large degrees of variance
use a formal methodology such as Z or in terms of quality, methodological style,
VDM and that the specifications be refined and programming technique, leading to a
mathematically until they are encoded in corresponding variability in reliability.
a program. Managers rarely make such Reference
demands, since the increase in costs would r D. Ince (1991). Software Quality and
be significant. Reliability: Tools and Methods (London,
An organization that is procuring soft- Chapman and Hall).
ware may also specify that the develop-
ment team and the software vendor as Associated terminology: Formal methods,
a whole must meet Software Engineering RAID, Y2K problem.
Institute (SEI) Level 5 Engineering certifi-
cation, the highest level of certification,
which requires that code is produced under RFID (Radio-frequency identity) tags
certifiably high levels of process control.
The business value proposition for this style Definition: Radio-frequency identity tags are small,
of development is that the systems will be relatively low-cost label devices that can be used to
of very high quality and extremely well doc- automatically identify and track objects.
umented, and that future changes to the
system can be made with a complete under- Overview
standing of the consequences. The prob- The idea of automatically identifying an
lem is again that this style of development object from a distance gained prominence
is extremely expensive and technologists and became practical during the Second
who can perform at this level are extremely World War when aircraft carried transpon-
rare. A trade-off is inevitable; organizations ders tuned to a certain radio frequency,
would like to have all their systems rig- and, upon receipt of a particular signal,

282
RFID (Radio-frequency identity) tags

they transmitted back an individual iden- similar way to the active tag system, but
tifier to state that the aircraft was friendly. the passive system, having no energy source
Subsequent to the war the concept of a of its own, needs to convert the incom-
reflected-signal device was discussed in the ing radio waves into energy and use that
writings of Harry Stockman, but the tech- energy to transmit back data.
nology available at the time was not ade- The amount of data that can be held on
quate to implement his theories. passive RFID tags is growing and products
Since the 1940s many people have with 2000 bytes are available (the Fujitsu
worked on remote-identification technolo- MB89R116 has 2048 bytes of total memory,
gies and the concept has been refined 48 bytes of which are used by systems pro-
considerably. The first patent was not grams), with a ‘‘data-retention” period of 10
awarded until 1973, when Mario Cardullo years under normal conditions.
was awarded a US patent for an RFID There are current moves toward the cre-
device (the patent expired in 1990). Tech- ation of protocols and standards for RFID
nical implementation problems remained tags. Primary amongst these is the effort
for some considerable time and it was not by EPCglobal Inc., a joint venture of GS1
until the late 1980s that the technology was (formerly the EAN), GS1-US (formerly the
able to be put to practical use as a means UCC), and a group of universities, whose
of allowing drivers of cars to pay high- focus is to ‘‘establish and support the EPC-
way tolls without stopping at a toll booth. global Network as the global standard for
The tags used in cars are known as active real-time, automatic identification of infor-
tags because they have their own internal mation in the supply chain of any com-
power source (usually a battery) and this pany, anywhere in the world.” In 2004 EPC-
enables them to transmit their data fur- global approved a specification for an RFID
ther and in a wider field than is possi- air-interface protocol, known as the ‘‘Class-
ble with so-called passive tags that have no 1 Generation-2 UHF RFID specification” or
internal power supply and can only reflect just as ‘‘RFID Gen 2.” Other standards are
or interact with an externally generated being developed by ISO/IEC and the Euro-
signal. pean Telecommunications Standards Insti-
Active tags hold significant advantages tute. These protocols address problems in
over passive tags in that the internal power electromagnetic interference, sensitivity to
source enables signal quality to be higher nearby metallic objects, range of data trans-
whilst providing a greater range of reliable mission, and the use of multiple readers in
operation. However, they are also much close proximity.
more expensive to manufacture and may
require occasional human intervention, for Business value proposition
example to change the batteries. RFID tags are mechanisms for automati-
While active tags are useful, their cost cally detecting and identifying an object
and need for maintenance prohibit them from a distance. The concept bears some
from being used in mass commerce, espe- similarity to that of bar codes, but RFID
cially in situations where they would be tags can be read through an opaque cover
discarded after a brief period of use. Thus, without being in direct line of sight,
passive tags have been the subject of inten- and can have the information upon them
sive research and development, and, as the changed remotely. RFID tags may be read
cost of their production has dropped, more unobtrusively at a distance, so they could
commercial opportunities have been real- eventually lead to systems allowing shop-
ized. A passive tag system works in a very pers just to wheel their purchases past a

283
RFID (Radio-frequency identity) tags

checkout, still in their shopping cart, and lifts) and on packing cases, as well as actu-
get an accurate accounting. They are also ally embedded with items.
an effective theft-prevention device, since Civil liberties organizations and politi-
tags may be hidden inside products, with cal groups have raised concerns about the
readers placed at all exits. use of RFID. Central amongst these con-
RFID technologies have been used in a cerns is the ownership of personal infor-
variety of ways. Active tags that contain mation, and the ability of RFID readers to
their own power source are frequently used track and profile individuals. The United
by motorists who wish to pass quickly States and the European Union are inves-
through toll booths without stopping to tigating the possibility of embedding RFID
pay with cash. The tag’s identification num- tags into their citizens’ passports and driv-
ber is collected by a reader on the toll ing licenses.
booth, and the toll is automatically debited
from an established account or charged to Summary of positive issues
a credit card. RFID technologies allow rapid remote iden-
The decreasing cost associated with the tification of labeled objects. The data in
passive tags has led to their use in a grow- the device can be changed as required. Pas-
ing number of areas, for example the orga- sive tags do not require an internal power
nizers of marathons frequently have the source. Tags are easy to place on objects.
athletes attach RFID tags to their shoes Protocols and standards are being devel-
so that their ‘‘real times” can be recorded oped. RFID tags can withstand a large range
(in large events such as the Boston and of ambient conditions. Unlike with bar-
London Marathons some runners do not code systems, the readers do not have to
actually pass over the start line until sev- be in line-of-sight proximity.
eral minutes after the starter’s gun has
been fired, due to the number of run- Summary of potentially negative issues
ners ahead of them in the crowd). Busi- RFID tags contain a relatively small amount
ness applications using RFID are also grow- of data. Costs are relatively high compared
ing, including embedding them in credit with older solutions. It is difficult to ensure
cards so that the card can simply be waved that a tag has been read, and especially
in front of a RFID reader. Casinos embed that all tags have been read when process-
RFID tags inside their chips to prevent the ing collections of labeled objects, and it
use of counterfeit chips in their establish- requires extra receipt acknowledgements
ments, placing readers at the gambling to be performed. Standards and protocols
tables. are still evolving. There are controversial
Large-scale RFID use is dependent upon social issues associated with the use of
the dramatic decrease in their cost, and RFID in the public domain. It is difficult to
this is dependent upon large orders being ensure that tags are deactivated after they
placed with their manufacturers. Progress have served their legitimate purpose.
in this area is strong: in 2002 the Gillette
Company ordered 500 million RFID tags, References
and Wal-Mart, the world’s largest retailer, r H. Stockman (1948). ‘‘Communication
has moved toward 100% RFID compliance by means of reflected power,”
by vendors. Both of these initiatives indi- Proceedings of the I.R.E., pp. 1196--1204.
cate the strength of corporate confidence r J. Landt (2005). ‘‘The history of RFID,”
in this technology. Companies are placing IEEE Potentials, Volume 24, Issue 4,
RFID tags on pallets (with readers on fork- pp. 8--11.

284
Robotics

r EPCglobal Inc. (2005). ‘‘EPC puter. Unimate moved castings to a plat-


radio-frequency identity protocols form and then performed spot welds upon
Class-1 Generation-2 UHF RFID car bodies.
conformance requirements,” Since the 1960s computer scientists and
http://www.epcglobalinc.org. artificial intelligence (AI) researchers have
r D. Ewalt (2003). ‘‘Gillette orders 500 worked to improve the flexibility, auton-
million RFID tags,” Information Week, omy, and dexterity of robotic systems.
January 6. Robotic arms have evolved to be multi-
jointed with hands and fingers enjoying
Associated terminology: Bar code,
seven degrees of freedom, and having tac-
Business process re-engineering.
tile sensors on the fingers that can be
used to regulate the pressure used to grasp
items.
Robotics Robotics research at laboratories such as
the MIT AI lab has created robotic systems
Foundation concept: Artificial intelligence. that utilize a variety of technologies to
Definition: Robotics is the branch of computer engi- make them more human, exhibiting the
neering covering the development of computer- human characteristics of adaptability, cre-
operated devices that perform useful physical tasks. ativity, and initiative. These are amongst
the characteristics required to differentiate
Overview ‘‘intelligent” robotic systems from robotic
The idea of a robotic servant that can per- systems that can merely perform tasks as
form tasks for its human owner has a defined by their programmers.
long and involved history. Humans have At the MIT AI Leg Lab researchers have
been attempting to create robots since the created robots that can run and perform
eighteenth century when inventors cre- somersaults, demonstrating agility and bal-
ated ‘‘automata” based upon clockwork sys- ance. They have also created robots that
tems and metal-cam ‘‘memory” devices, can find their way around the corridors
very much in the way that a musical box or and rooms of the laboratory, demonstrat-
a player-piano worked. However, the term ing their visual and spatial reasoning capa-
‘‘robot” did not come into use until the bilities. Some of the robots in the labo-
beginning of the twentieth century and it ratory can communicate with the people
is generally acknowledged that the Czech whom they meet in the building and pro-
playwright Karel Čapek coined the term, vide assistance to visitors, demonstrating
derived from the Czech word for serf or their skills in understanding natural lan-
servant. guage, speech recognition, and reasoning.
It was not until the 1960s that working Not all robots are human-like in design,
robots started to appear. While these early and researchers such as Professor Rod-
machines were called robots, they were far ney Brookes at MIT have created smaller
from the science-fiction vision of a robot robots designed to work in hostile environ-
portrayed as a fully operational, thinking, ments, such as those found on the planet
and super-human, if emotionally stunted Mars. Brookes’s robots, such as Boadicea and
being (the primary example seems to be Genghis, provided the basis of many of the
Robby the Robot from the 1956 film Forbid- technologies used in robots such as the
den Planet). In 1961 General Motors deployed NASA Mars Exploration Rovers Spirit and
the Unimate robot, which was essentially a Opportunity, which touched down on Mars
4000 lb hydraulic arm attached to a com- in 2004.

285
RSS (Really simple syndication)

At the outset of the twenty-first century, mic issues such as operational costs incur-
robots may not be able to make our beds, red by the consumption of electricity and
cook our breakfast, and drive the children maintenance, and the expected useful life-
to school, but robotic help systems for the time before new technology renders an
home are available from several vendors, expensive robotic workforce obsolete.
including Honda’s ASIMO P3, which can The military use robotic systems to oper-
navigate around a home and supposedly ate in high-risk environments such as
perform basic household tasks such as examining potential bombs or hazardous
fetching the newspaper from the front gar- materials. They have spent considerable
den. However, those who expect robots to amounts of money attempting to develop
provide any real independent assistance in automated fighting machines to replace or
the near future are doomed to disappoint- assist their human soldiers.
ment.
Much has also been written about a new Summary of positive issues
class of robotic systems termed ‘‘nanotech- Robotic systems have been researched since
nology,” that is technology whose size is the 1950s and a large literature has devel-
measured in nanometers or one-billionths oped. Industrial robotic systems are widely
of a meter and hence are built at the molec- available for manufacturing. Robots can
ular level. While the theory holds great possess the capabilities of walking, com-
promise, the reality of nanotechnology municating, and vision processing. Robots
robots being used to perform heart oper- are in use in hostile and inaccessible envi-
ations or other procedures in the human ronments such as volcanic craters and
body remains a considerable way off. Mars.

Business value proposition Summary of potentially negative issues


The vision of genetically based humanoid Intelligent robotic systems are typically
androids, such as those depicted in the programmed to work in limited domains,
film Blade Runner, capable of working in and their level of intelligence is debatable.
the mines on Mars is at present still only Industrial robots need to have their pro-
science fiction, and is virtually certain to cesses defined completely for them.
remain so. The use of robots within cor-
porations is currently limited to industrial References
r Computer History Museum, 1401 N.
robots that work on and around assembly
lines. Manufacturers such as FANUC, ABB, Shoreline Blvd, Mountain View, CA
Panasonic, Motoman, Kuka, and Nachi pro- 94043, USA.
r http://marsrovers.jpl.nasa.gov/overview/.
vide a wide range of equipment for cus-
r D. Kortenkamp, R. Bonasso, and R.
tomers such as automobile companies.
The technology associated with indus- Murphy (1998). AI-based Mobile Robots:
trial robots is primarily focused upon pro- Case Studies of Successful Robot Systems
viding precision performance in processes (Cambridge, MA, MIT Press).
such as welding. The adoption of a robotic Associated terminology: Neural networks,
system also requires a company to build Machine learning.
a business case that examines operational
issues such as the system’s ability to reduce
work-cycle times, operate in small work RSS (Really simple syndication)
areas (so that more robots can operate
together on the assembly line), and econo- Foundation concepts: Internet, XML.

286
RSS (Really simple syndication)

Overview ders whose sites may simply not be noticed.


Really simple syndication, RSS, was originally An aggregation service collects together
introduced in a simple form by Netscape feeds from a variety of sources in one well-
in 1999, and has since undergone a series known place, where potential subscribers
of modifications and improvements. As a may browse at leisure and see whatever
result of this, several different versions are attracts their interest.
in use today, and the acronym is actually The business model associated with RSS
defined three ways, as RDF site summary, feeds is based on allowing users to sign up
Rich site summary, and Really simple syn- for information updates in specific areas
dication. The last is the most used meaning. of interest and then pushing information
Syndication is a process traditionally out to those users. This builds a strong
used in the newspaper industry when con- relationship between the provider and the
tent is used by a variety of outlets. For subscriber since the subscriber has to go
example, a well-known columnist may have via the hyperlink to the story’s originat-
their work printed in many different news- ing site to read the full article. This is
papers at the same time, and nearly all pop- advantageous in that it allows providers to
ular comic strips are published this way. understand and manage their content bet-
Web syndication is a mechanism through ter, focus the content upon specific audi-
which an individual may subscribe to con- ences, and, of course, focus advertising and
tent providers, and automatically receive other revenue-generating aspects of their
regular updates from them. RSS is used not sites where applicable.
only for news stories, but also for blog shar- RSS technology has drawn a lot of atten-
ing, and even pictures and music. RSS read- tion because it overcomes the limitations of
ers collect content and place it in a folder email as a distribution channel, on which
for viewing, displaying an icon to notify the similar content is likely to be blocked as
user that new items are available. Gener- spam or just ignored. Users no longer find
ally, the whole item is not copied to the signing up to email distribution lists an
folder, but rather a summary together with attractive choice due to the widespread
a hyperlink to the full story. abuse of email and the bland nature of the
text delivery. RSS is also attractive to busi-
nesses since it leverages information that
Business value proposition they have already provided. The system may
RSS provides any entity that deems its be automated, and requires relatively low
information or opinion to be worthy of resource overheads to manage.
reading by a subscriber with a semi-
standardized mechanism for subscription
management and distribution. RSS viewers Summary of positive issues
(client receivers) are universally available, RSS is an easy-to-use mechanism for syndi-
so there is no need to distribute special- cating information. RSS readers are widely
ist software to subscribers. Content may be available. Some browsers have RSS readers
made available in one of two ways, either built into them, and there are many ven-
by setting up an individual or corporate dors supplying commercial tools and solu-
RSS server, or by registering with an Aggre- tions. Open-source and free RSS readers and
gator. Providing an RSS service is suitable tools are also available. RSS leverages exist-
for a large well-branded company whose ing information for low cost, and provides
feeds will be sought out by subscribers, but organizations with the ability to have a bet-
unsuitable for relatively unknown provi- ter understanding of their customers and

287
RSS (Really simple syndication)

leverage that relationship. RSS technology Reference


is relatively straightforward. r B. Hammersley (2003). Content
Syndication with RSS (Sebastopol, CA,
Summary of potentially negative issues O’Reilly Press).
RSS technology is still evolving and several
technology groups are attempting to pro- Associated terminology: Web services,
vide direction for future standardization. Client--server.

288
Sarbanes–Oxley Act of 2002 (SOX)

key from a CIO’s perspective, including the


Sarbanes–Oxley Act of 2002 (SOX)
following.
Definition: The Sarbanes–Oxley Act of 2002 was Title II -- Auditor Independence, Section
enacted to “protect investors by improving the accu- 201: Services Outside the Scope of Practice of
racy and reliability of corporate disclosures made pur- Auditors ‘‘(g) Prohibited Activities . . . (2)
suant to the securities laws, and for other purposes” financial information systems design and
(Sarbanes–Oxley Act of 2002, Report 107–610). implementation.” Termed ‘‘non audit ser-
vices,” accounting firms are banned from
Overview providing systems consulting, and as a con-
The Sarbanes--Oxley Act of 2002 was a sequence most accounting firms have spun
response to the financial reporting and dis- off their consulting practices, which now
closure problems associated with compa- have to maintain independence from the
nies such as Enron, whose 2001 collapse process of implementing controls and per-
was the largest bankruptcy in US history. forming audits.
This large and complex act pertains to cor- Title III -- Corporate Responsibility, Sec-
porate governance practices in public com- tion 302: Corporate Responsibility for Financial
panies and contains eleven titles: Reports ‘‘(a) . . . that the principal executive
officer or officers, or persons performing
1. Public company accounting oversight similar functions, certify in each annual or
board quarterly report filed or submitted under
2. Auditor independence either such section of such Act that (1)
3. Corporate responsibility the signing officer has reviewed the report;
4. Enhanced financial disclosures (2) based upon the officer’s knowledge, the
5. Analyst conflicts of interest report does not contain any untrue state-
6. Commission resources and authority ment of a material fact or omit to state a
7. Studies and reports material fact necessary in order to make
8. Corporate and criminal fraud the statements made, in light of the cir-
accountability cumstances under which such statements
9. White-collar crime penalty were made, not misleading . . . (4) the sign-
enhancements ing officers (A) are responsible for estab-
10. Corporate tax returns lishing and maintaining internal controls;
11. Corporate fraud and accountability (B) have designed such internal controls
to ensure that material information relat-
While the act is wide-reaching in scope ing to the issuer and its consolidated sub-
and focuses on corporate and executive sidiaries is made known to such officers
accountability for financial data, the main- by others within those entities, particu-
tenance of internal control structures, and larly during the period in which the peri-
the role of accounting firms in the audit odic reports are being prepared; (C) have
process, for the CIO or IT professional evaluated the effectiveness of the issuer’s
it does not contain any specific systems internal controls as of a date within 90
requirements and in fact never even men- days prior to the report.” The effect of this
tions the word computer in its 66 pages. aspect of the act is to enforce strict con-
However, it is clear that technology and trols upon access to data and reporting
information systems will be central to cor- systems. The mechanisms through which
porate compliance with the act. this can be done are varied, and the use
The eleven titles of the act contain 69 of modern ERP-type systems or packages as
sections, several of which are regarded as opposed to legacy systems typically eases

289
Sarbanes–Oxley Act of 2002 (SOX)

the compliance officer’s task because many retention of relevant records such as work
of the security barriers have been con- papers, documents that form the basis of
structed into the system by the vendor. an audit or review, memoranda, correspon-
Title IV -- Enhanced Financial Disclo- dence, communications, other documents,
sures, Section 404. Management Assessment and records (including electronic records)
of Internal Controls, ‘‘(1) State the respon- which are created, sent, or received in
sibility of management for establishing connection with an audit or review and
and maintaining an adequate internal con- contain conclusions, opinions, analyses or
trol structure and procedures for finan- financial data relating to such an audit or
cial reporting,” together with Section 409, review which is conducted by any accoun-
‘‘(1) Real Time Issuer Disclosures -- Each tant who conducts an audit . . .”
issuer reporting under section 13(a) or The retention of data and information by
15(d) shall disclose to the public on a the public company and the auditors is a
rapid and current basis such additional major aspect of Information-lifecycle manage-
information concerning material changes ment (ILM), in which the frequency of use
in the financial condition or operations of the data is related to the type and cost
of the issuer, in plain English, which of the storage medium utilized for stor-
may include trend and qualitative infor- ing that data. Included in this data set is
mation and graphic presentations, as the the email correspondence of the organi-
Commission determines, by rule, is nec- zation, the archiving of which needs spe-
essary or useful for the protection of cial concern since the volume of emails
investors and in the public interest.” The may be considerable, thus requiring spe-
development of policies and procedures to cialist storage services. Additionally, special
maintain internal controls typically falls techniques will be needed to capture other
under secure identity management (SIM), forms of electronic communication such as
whereby all security points for data and instant messaging and text messaging.
personnel are assessed and the required In order to comply with these and the
level of security is implemented. The pri- other sections of the act, CIOs have utilized
mary vehicle through which Section 409 (1) a series of frameworks. Primary amongst
is approached is the posting of information these are COSO, CobiT, Trust Services, and
upon a web site accessible via the internet. the ISO 17799 security standard.
Additionally, document-management sys- The Committee of Sponsoring Organiza-
tems are also valuable resources to provide tions of the Treadway Commission (COSO)
facilities such as versioning, archiving, and is an organization founded in 1985 to
managing heterogeneous file types (audio, examine the ‘‘causal factors that can lead
video, email, text, web documents, etc.). to fraudulent financial reporting” and to
Title VIII -- Corporate and Criminal Fraud ‘‘develop recommendations for public com-
Accountability, Section 802. Criminal Penal- panies and their auditors.” The COSO
ties for Altering Documents, ‘‘(a) (1) Any framework was developed in the 1990s to
accountant who conducts an audit of an manage the SEC’s demands for internal
issuer of securities . . . shall maintain all audit controls and has been adopted by
audit or review work papers for a period of SOX-compliance managers to meet the act’s
5 years from the end of the fiscal period in demands. The framework is composed of
which the audit or review was concluded. five dimensions,
(2) The SEC shall promulgate, within 180
days after adequate notice and an oppor- 1. Control environment
tunity for comment, such rules and regu- 2. Risk assessment
lations, as are reasonably necessary to the 3. Control activities

290
Sarbanes–Oxley Act of 2002 (SOX)

4. Information and communication Trust services are a set of ‘‘professional


5. Monitoring assurance and advisory services based upon
a common framework to assess and address
While these dimensions are very useful, the risks and opportunities of IT” devel-
they are not specific step-by-step instruc- oped through the American Institute of
tions that can be implemented. This is Certified Public Accountants. The founda-
by design, since they were intended to be tion for trust services encompasses four
guidelines for a wide variety of companies sets of principles and criteria: policies, com-
and situations; there can never be a specific munications, procedures, and monitoring,
set of ‘‘one-size-fits-all” instructions on how together with five sets of principles and
to be compliant. The COSO framework is criteria developed for policy implementers.
typically used with other frameworks such These consist of attributes and controls in
as those provided by CobiT, or by trust ser- the areas of security, availability, process-
vices that are more specific in nature and ing integrity, online privacy, and confiden-
help organizations focus upon technology- tiality.
related issues.
The Control Objectives for Information Business value proposition
and Related Technology (CobiT) framework The business value proposition of the SOX
was developed by The Information Sys- act is to provide evidence for the share-
tems Audit and Control Foundation to holders, employees, suppliers, vendors, and
assist organizations to manage compliance others associated with an organization
through the provision of good practices that the financial control structures are
that work toward balancing risk, control, in accordance with strict audited account-
and technical issues. ing standards, and that a financial controls
The control of access to financial sys- structure is indeed in place.
tems and financial data is at the very
center of the SOX act and the control Summary of potentially positive issues
responsibilities are typically divided into The SOX act helps investors in public com-
three areas: General control issues such panies to have confidence that the finan-
as data management, disaster recovery, cial aspects of the company will adhere to
and the physical and logical security of the stringent requirements of the act, and
resources; Company-level controls, includ- that, should an organization not adhere
ing such issues as implementing codes of to the act, then the executives and board
conduct, policies to prevent fraud, and of directors will be held accountable. The
operational governance policies (these con- act makes organizations consider their
trols establish the ‘‘control environment” strengths and weaknesses in terms of their
within the organization); and Application control structures and financial-reporting
controls which are controls actually embed- systems.
ded in the organization’s processes in order
to ensure their correct use (these include Summary of potentially negative issues
such controls as authorization approvals SOX is frequently discussed in terms of
and approval routings). These three aspects the issues surrounding Y2K compliance,
of control are then considered and devel- except that failure to meet Y2K compli-
oped further in terms of the risk criteria ance did not potentially result in the CEO,
with each process and task, the monitoring CFO, and others in the organization going
of the controls, and the communications to jail. SOX has been criticized for being
that surround the management of control extremely expensive and hard to imple-
functions. ment, and for requiring large amounts of

291
Scalability

corporate resources and commitment. This perfectly. If that merchant expands, and
has been especially hard on many smaller needs to store 24 000 cubic feet of merchan-
public companies and has resulted in com- dise, simply building another 3000 square
panies delisting from the public markets foot warehouse will solve the problem
and returning to being privately held enti- equally well. If they eventually need to store
ties. It has also been suggested that, as a 120 000 cubic feet, the process of building
cost of business, it has acted as a barrier a new 3000 square foot warehouse, applied
to entry for companies wishing to locate in ten times, will yield a working solution.
the United States or use the US public stock The idea of building warehouses to store
markets to raise capital. merchandise is a Scalable solution.
If a publisher decides to produce a
References small encyclopedia with 4000 entries, they
r The IT Governance Institute (2004). IT
may decide to print it as a single large
Control Objectives for Sarbanes--Oxley book, with 1000 pages printed on very
(Rolling Meadows, IL, The IT lightweight paper, in the style of a large
Governance Institute), www.itgi.org. dictionary. If the encyclopedia proves pop-
r The IT Governance Institute (2000).
ular, and a second, extended edition with
CobiT Framework, 3rd edn. (Rolling 5000 entries is planned, they could simply
Meadows, IL, The IT Governance expand the book to 1250 pages and use a
Institute), www.itgi.org. slightly stronger spine. However, the same
r The IT Governance Institute (2000).
expansion process can not be applied too
CobiT Executive Summary, 3rd edn. many times. It is almost impossible to pro-
(Rolling Meadows, IL, The IT duce a 10 000-page book: the binding falls
Governance Institute), www.itigi.org. apart, and it is too heavy to lift. This is not a
r Committee of Sponsoring Organizations
scalable process. Eventually a jump to new
of the Treadway Commission (2004). technology is required.
Integrated Control Framework, For book production, the jump to new
Volumes I & II, www.iacpa.org. technology is simple: just publish the book
Associated terminology: Internet, ILM, in more than one volume. The analog for
Email, Instant messaging, ISO/IEC 17799. computer systems, both hardware and soft-
ware, is usually not so simple.
A single desktop PC, with a suitably
robust operating system, when acting as an
Scalability enterprise web server can support a surpris-
ingly heavy load. A thousand gigabytes of
Foundation concept: Software development lifecycle. data and a million hits per day are realistic
Definition: The ability to expand a system without expectations. However, there are limits. The
making significant changes. Getting more of the same standard IDE/EIDE disk system supports a
results by using more of the same solution. total of four devices, and even wide SCSI
is limited to 15, so disk storage can not be
Overview expanded without limit. A T1 internet con-
Scalability is a very simple concept. If a nection provides 193 000 bytes per second
merchant needs to store 12 000 cubic feet of bandwidth: as usage increases, that pro-
of non-perishable merchandise, they might vides an absolute upper limit to capacity.
decide to build a 3000 square foot one- Simply using two computers instead of one
storey warehouse. With merchandise stac- does not provide a viable solution. How are
ked 8 feet high, covering 50% of the floor- accesses to be distributed between the two
space (to allow access), the size matches computers? How can one ensure that the

292
Secure

two computers have consistent versions of operational 24 hours a day, 7 days a week,
the data at all times? more sophisticated design is needed. It
In a similar vein, the simple techniques takes many years for a programmer to gain
of programming used by beginning pro- sufficient expertise in such areas.
grammers are also very often not scalable.
The commonly understood methods for Summary of positive issues
searching data sets for example are lin- There exists an extensive body of knowl-
ear with time, which means that, as the edge pertaining to the problems of scaling
amount of data to be searched grows, the software and hardware systems. Commer-
time required to search it grows in direct cial systems that will scale to significant
proportion. This sounds like an obvious levels are available for common problems
observation, but is in reality a critical prob- and tasks.
lem that can be fatal to ill-conceived soft-
ware projects. Under light load, testing
Summary of potentially negative issues
with 1 MB of data might reveal that the
Failure to understand the scale to which
average search takes one tenth of a second,
a task may grow can result in solutions
and that sounds quite acceptable. But when
that fall short, cause system failure, and be
the system goes live with perhaps 10 GB
detrimental to the business. It is expensive
of data (10 000 times as much), that one
to train IT personnel to the levels required
tenth of a second grows to one thousand
in order for them to plan, design, build,
seconds, or a quarter of an hour, reducing
and maintain systems that scale signifi-
the system to a maximum of 96 agoniz-
cantly.
ingly slow accesses per day. (See Complex-
ity, Algorithm, and Data structure for more Reference
depth.) r K. Hwang (1992). Advanced Computer
Ensuring scalability requires intimate Architecture: Parallelism, Scalability,
knowledge of algorithm analysis, data- Programmability (New York,
structure design, and network protocols, McGraw-Hill).
and requires familiarity with the vast cor-
pus of human experience that 60 years of Associated terminology: Reliability,
intense study has created. Of course, an Computability.
intelligent programmer could in principle
rediscover all of this knowledge for himself
or herself, just as an intelligent blacksmith Secure
could in principle independently reinvent
the Apollo lunar-landing module. Foundation concept: Encryption.
Definition: Secure web pages and secure transactions
Business value proposition are transmitted over the network with an encryption
Lack of scalability in software solutions is method that ensures complete confidentiality even
the hallmark of the untrained program- if the entire interaction, including key exchange, is
mer. It is often possible to learn a little pro- intercepted.
gramming by reading a book or two, and
then produce small-scale solutions, per- Overview
haps using Visual Basic or JavaScript, that When reviewing personal records, or mak-
stand up to simple test scenarios. When ing an online purchase with a credit card,
software has to support literally millions or when privacy is a concern for any
of accesses every day, with perhaps hun- other reason, the internet can be an alarm-
dreds happening concurrently, and remain ing communications medium. There are so

293
Secure

many opportunities for communications to credit-card transactions that was intro-


be intercepted anywhere along what may duced jointly by Visa and Mastercard,
be a very tortuous path between browser which works by the same basic methods.
and server. Physical security, preventing Secure communications do offer some
transmissions from being intercepted, is protection against identity theft, but are
impossible, so Encryption must be used to no protection if identity theft has already
ensure that intercepted communications occurred. If a criminal can successfully pre-
will be unreadable. tend to be another person, having stolen
The Secure-sockets layer (SSL) and other their password, they will be able to engage
similar technologies provide a universally in secure transactions under that identity.
standard means of encryption that gives
complete protection regardless of how Business value proposition
badly compromised the communications Security is one of the fundamental aspects
network may be. Traditional encryption of computer and network design, and the
methods always had one weak spot: the technologies that support the desired secu-
sender and receiver must agree on a key rity level need to be designed into the
to be used for encryption and decryption. system at the outset whenever possible.
If the sender and receiver are never phys- The design of the security component is
ically in the same location, the key must the responsibility not only of the network
be transmitted, and could easily be inter- administrator and the chief security offi-
cepted, resulting in all communications cer but also of the CIO, the CEO, and the
being readable. If the same key is used board of directors. For example, should a
more than once, it provides eavesdroppers security breach occur or if customers don’t
with a much greater chance of decoding see the user interface that they under-
messages by analytic means. stand and are familiar with (e.g., the pad-
Newer developments, owing much to lock that signifies the secure-sockets-layer-
the Diffie--Hellman technique, completely based connection), it is less likely that cus-
remove this problem, with a cleverly tomers will use the site. However, the level
designed protocol that allows sender and of security needs to be based upon the
receiver to agree upon a secret key that circumstances and basis of the business
can not be determined even by an eaves- and transactions occurring: a major bank
dropper who hears the entire conversation. will be clearly required to have a higher
SSL and other systems use these secure key- level of secure technology than a small
generation methods to securely create a trader, and a purely informational site
totally new encryption key for each com- with no financial transactions may need
munications session, so there is also no risk none.
associated with the reuse of old keys.
A Secure web page is one that is trans- Summary of positive issues
mitted from the web server to the user’s Established and well-understood technolo-
browser using this technology, so that gies exist to ensure the security of the data
its content is guaranteed to be readable transmitted on the internet between com-
only by the intended recipient. A Secure panies or individuals. Technologies exist to
transaction is an entire online ‘‘conversa- provide security at a variety of levels.
tion” between a web browser and the
server, every stage of which is transmit- Summary of potentially negative issues
ted securely. SET (Secure Electronic Trans- Even though technologies exist, security
action) is a particular protocol for secure can be maintained only if the technologies

294
Security

are implemented and used correctly. No secure areas needs to be considered. For
security system can be expected to be per- example, if a trader at an investment bank
fect, but processes and measures must be left their computer logged in to the trading
in place to identify immediately any break- system, a cleaner with legitimate access to
down in security that has occurred, as the area at night could gain access to that
well as reporting any unsuccessful security system, send rogue emails, place trades, or
breaches so that weak spots may be identi- acquire proprietary information about an
fied before it is too late. Security should not IPO pricing strategy. Similarly, all techni-
only be reactive to incidents, but should cal correspondence should be shredded or
also be a continuous background presence. disposed of in a secure manner, because
It is important to assess carefully the true even seemingly irrelevant information can
value of security policies in action, since be useful to someone with bad intent, e.g.,
some seemingly valid policies can have a the CEO’s internal email address or the
negative effect; for example, the common CFO’s system ID could be a starting point
requirement to change passwords every for a hacker’s attack. The physical security
month results in users having meaningless of the computers in the organization needs
unmemorable passwords that they need to be examined since it is not unknown
to write down, and a written password is for people to load deliberate software bugs
much easier to find. that copy log-in data or corporate infor-
mation and then transmit it secretly to
Reference the attacker (or an accomplice) via the
r K. Day (2003). Inside the Security Mind:
internet.
Making the Tough Decisions (Englewood Technology clearly can play a large
Cliffs, NJ, Prentice-Hall). part in securing an organization’s systems.
Associated terminology: Security, For example, the use of encryption, thin
Internet, Protocol. clients, and password-protection systems,
requiring the operating system to time-out
and disconnect users after short periods of
inactivity, having servers disable worksta-
Security
tions and clients on the network after nor-
Foundation concept: Secure. mal work hours, and using biometrics to
Definition: An information system’s security pertains log in to systems can all play a part in secur-
to the task of understanding those aspects of an orga- ing the systems from intrusion.
nization’s information system and processes that may Security concerns also emanate from
be vulnerable to unauthorized intrusion, abuse, or external entities that attempt to intrude
attack. upon the network and its contents. Attacks
come in many forms, including attempts to
Overview hack into the network, or crack the secu-
Security of information systems is primarily rity codes associated with a network. Alter-
concerned with access, physical as well as natively, networks may come under attack
through the network. Good security com- from Viruses, Worms, Trojan horses, and
mences with an appraisal of access rights Denial-of-service attacks. Defenses against
and then enacting physical and technolog- each of these can be mounted, includ-
ical solutions to guarantee that security is ing the creation of software and hardware
achieved. barriers such as Anti-virus software, Fire-
Physical security is frequently over- walls, Proxy servers, and the use of networks
looked, but all access to secure and non- that are not capable of connecting to the

295
Security

internet unless Network address translation would potentially have the negative effect
is used. Other plagues that are usually of reducing customer confidence in their
not fatal but are detrimental to corporate system and defenses as well as making the
systems’ performance include Spam, Pop-up company a potentially attractive and chal-
ads, and Phishing. These can be addressed lenging target for other attacks. Laws are
through the use of spam filters and of also not ‘‘practically” helpful in the event
software that removes spyware and pre- that an attack successfully crashes the cor-
vents pop-ups from being executed. Phish- porate network or when private data has
ing requires corporations to be proactive been stolen; the damage has already been
and to help customers affected by such done.
attacks, and to have e-Commerce policies that
minimize their effectiveness. Business value proposition
While computers and networks can Security is an aspect of a corporate IT orga-
seem secure, frequently security measures nization that is usually regarded as a sunk
are undermined by individual weaknesses, operational cost; while the prevention of
such as leaving a laptop computer with intrusion or attack is valuable, it does not
no password protection in a car that is actually help raise the top line of the busi-
stolen, downloading data to unauthorized ness (revenue) unless a company uses it as
portable computers that are taken out of a selling point. In the last century banks
secure premises, using wireless systems used to have large safes visible to customers
that beam the signal through office or entering the bank and this encouraged
hotel walls, and using laptop computers on their customers to think that their deposits
airplanes where corporate information or would be safe. In the modern technological
client information can be surreptitiously organization it is becoming more and more
viewed. Even cell phones and cell phone a corporate imperative to ensure the secu-
conversations need to be considered as part rity of their systems and, more importantly,
of the security umbrella of an organiza- the security and privacy of their customers’
tion and solutions found for these areas of data.
concern. Many companies have created a position
Corporate IT security concerns are of Chief Security Officer (CSO) responsible
addressed by many laws, including the for all security concerns, including physi-
CAN-SPAM Act of 2003, which enacts cal and intangible assets. This is an impor-
requirements on commercially motivated tant position, especially in a public com-
email; HIPAA (the Health Insurance Porta- pany that has to adhere to corporate gover-
bility and Accountability Act, 1996) that nance legislation such as Sarbanes--Oxley.
legislates the security requirements asso- CEOs, CIOs, and CSOs all need to be aware
ciated with digital medical systems; the of technology-related security issues and
UK Computer Misuse Act of 1990, which these may influence decisions regarding
forbids unauthorized access to computers corporate strategy. For example, it is vital
and the data held on them; and the US to understand the legal consequences of
Identity Theft and Assumption Deterrence outsourcing information-based services to
Act of 1998, which makes most phish- countries with weak IT laws and protection.
ing criminal. These and other laws come One approach that has been used success-
with strong penalties to deter potential fully has been to adhere to ISO 17799:2000,
offenders, but the nature of the internet which is a code of practice for information-
makes many offenders difficult to track security management and offers guidelines
down and many companies do not wish to on the development of policies and good
prosecute every small attack because this practice in this area.

296
Server

The need for security within an orga-


Server
nization’s IT systems is a critical factor
in maintaining stakeholder (customer, ven- Foundation concept: Network.
dor, shareholder, and governmental) confi- Definition: Servers provide the resources to a network;
dence in the organization. However, many these resources are accessed by end users through
organizations are unable or can not afford client devices.
to staff a full-time CSO position and have
instead used the services of vendors who Overview
specialize in security issues to perform A Server may be any device that provides
security assessments. computational, data-processing, storage, or
other services to other Client systems. The
usual model is that a server has a long-
Summary of positive issues
term passive network presence, providing
Many of the security concerns for organiza-
services to client systems only on request,
tions are well known and procedures and
whereas a client is an independent compo-
policies are available to address them. The
nent, actively requesting services when it
laws of many countries cover the majority
(or its user) sees fit.
of abuses relating to systems security and
A server can be as simple as a personal
specify significant penalties for offenders.
computer acting as a file system or web-
Organizations can appoint a CSO to head
page host for a small group of clients,
the security initiative. There is an ISO code
or as complicated as a cluster of proces-
of practice for information security man-
sors hosting complex software applications
agement to guide organizational policies.
such as an ERP or distributed database for
There are many vendors and consultants
thousands of clients over a global network.
worldwide that provide security services,
Servers provide centralized computing ser-
tools and resources.
vices that may be secure, scalable, and
regulated. The centralization allows net-
Summary of potentially negative issues work administrators to better protect their
Security problems continue to surface in organization’s systems from unauthorized
all areas associated with corporate (and access, attack, and failure than if an orga-
personal) computing environments: hard- nization had a wide variety of resources
ware, software, the physical infrastructure, spread over the business units. The scala-
and human resources are all vulnerable. bility of the system comes from the mod-
Thus the target for CSOs continues to ular design of the typical server system, to
move, requiring continuous investment in which computing power, in terms of the
resources to combat these problems. The number of processors, can be added as the
legal framework acts as a deterrent but can demand for system resources grows.
not prevent criminal activity. The ISO code Operating systems suitable for support-
of practice for information security man- ing large-scale servers, such as Unix (and
agement is a set of guidelines, not a man- Linux, a variety thereof) and Microsoft
ual of solutions. server products, allow the user demand
to be balanced across multiple servers
Reference
r K. Day (2003). Inside the Security Mind: if demand increases, and thus help to
optimize performance levels. Specialized
Making the Tough Decisions (Englewood
servers are also generally incorporated into
Cliffs, NJ, Prentice-Hall).
the overall design of a corporate server-
Associated terminology: Encryption, based architecture; these, for example, act
Internet, Protocol. to provide data files (file servers), web pages

297
SIIA (the Software and Information Industry Association)

(internet servers), and applications (appli- and that decisions affecting the return on
cation servers), and to relieve the load investment are correctly considered.
on the main server which is interacting
with its clients and determining how best Summary of positive issues
to service requests (the main server in a Servers facilitate strong centralized control
Microsoft-based system is termed the pri- of an organization’s computing resources.
mary domain controller). Servers provide a flexible, scalable, secure
architecture through which to deploy com-
Business value proposition puting resources. There is a choice of oper-
The server concept allows corporations to ating systems upon which servers can run.
develop flexible, scalable computer systems Servers can run major applications such as
that run across networks. The associated ERP systems or act as file servers for a com-
client--server technologies allow greater pany’s web-page service.
flexibility in terms of configuration, the
variety of applications that can run on the
platform, and control of user access com- Summary of potentially negative issues
pared with traditional mainframe dumb- Large server architectures require substan-
terminal configurations. Server technology tial expertise to run effectively. Depen-
may be configured to support a wide array dence upon a central system can leave the
of applications, including specialist servers users vulnerable should that system fail.
such as web servers, that address the com- References
putational needs of particular groups of r J. Chellis, C. Perkins, and M. Strebe
users. (2000). MCSE: Networking Essentials, Study
Server technology allows companies to Guide (Alameda, CA, Sybex Press).
create and administer a strong security pol- r P. Gralla (2004). How the Internet Works
icy based on the centralization of data and (Indianapolis, IN, Que).
applications in a small number of server r D. Groth (2003). A+ Complete (Hoboken,
locations. This allows the CIO, the network NJ, Sybex--John Wiley and Sons).
administrator, and the Chief Security Offi-
cer to focus resources on the problem of Associated terminology: Client, Client--
system security. Security policies may be server, Web services, Peer to peer.
implemented more efficiently on a server-
based architecture than when corporate
IT operations are distributed amongst the
business units with no central control. SIIA (the Software and Information
Server architecture also allows the IT Industry Association)
organization to standardize and regulate
its software systems rather than allowing Definition: The Software and Information Industry
individuals and business units to make Association is a trade association representing mem-
their own software choices. Server technol- bers of the software and digital-content industry.
ogy is mature enough that good cost-of-
ownership decisions can be made centrally Overview
by the IT organization. A key aspect of this The Software and Information Industry
calculation is centered upon the software Association (SIIA) is a trade association
license fees, which may be based on the for the software and digital-content indus-
number of users and the processing power try. The association provides a variety of
of the system. It is therefore important services, including representing its mem-
for the CIO to ensure license productivity bers’ interests at the governmental level,

298
Software development lifecycle

protecting the intellectual property of bers with educational resources. The SIIA is
members, promoting business develop- proactive on anti-piracy issues and provides
ment, and providing educational services certification for IT professionals pertaining
at a corporate level and as a public to piracy issues.
service.
The association is divided into four Summary of potentially negative issues
divisions: Software, Content, Education, The requirement to join an umbrella orga-
and Financial Information Services, nization such as the SIIA is optional and
through which a common policy formu- it can not be seen as the only voice in the
lation is developed and put forth in the technology industry.
form of white papers, briefings, and other
Reference
media. Each division develops the agendas r SIIA Main Office, 1090 Vermont Ave
that are important to its members, and,
NW, Sixth Floor, Washington, D.C.
through working groups and studies,
20005, USA; http://www.siia.net/.
develops policy in each area. For example,
the Software division has working groups Associated terminology: ACM, BCS, IEEE,
that consider initiatives such as web ser- W3C.
vices, open-source systems, and software
as a service.
The SIIA is also active in the development Software development lifecycle
and enforcement of anti-piracy policy, and
has created an anti-piracy department to Foundation concept: Software.
focus upon this area (the division was for- Definition: The software development lifecycle is the
merly known as the Software Publishers series of processes through which a concept is trans-
Association). The SIIA also offers a course formed into a program.
that leads to the Certified Software Man-
ager (CSM) qualification, which is aimed at Overview
network managers and technology profes- The Software development lifecycle (SDLC) cov-
sionals. ers the transformation of a concept into a
verified and validated software product. Ver-
Business value proposition ification is the process of ensuring that the
The SIIA provides a forum through which system created matches its specification;
common subject matter can be discussed Validation is the process of ensuring that
and policy positions ultimately developed. the specification satisfies the customer’s
These policy positions can then be put requirements.
before legislators by the SIIA to help pro- An SDLC can be designed to consist of
vide advice and education. The body also any combination of development methods
provides industry intelligence to mem- as long as the resulting code matches the
bers through briefings, reports, and confer- specification to a degree of rigor that sat-
ences. isfies the customer. The weakest form of
lifecycle is one in which somebody verbally
Summary of positive issues proposes an idea for a program and then
The SIIA has a large membership with a just starts to code directly on the computer,
wide range of interests. Membership fees running the resultant code and testing it
are on a sliding scale depending upon the until they are satisfied. This method is
revenues of the company joining. The SIIA clearly not scalable and makes proper veri-
is influential in providing representation at fication impossible since there is no speci-
the governmental level and provides mem- fication to measure the system against.

299
Software development lifecycle

At the other end of the development ware Engineering Institute at Carnegie


spectrum are Formal methods, which can be Mellon University proposed in 1995 the
considered a development lifecycle based Capability Maturity Model for Software
upon mathematical methods, first for spec- (CMM), which in 2000 became the CMMI
ifying the program to be developed, and (Capability Maturity Model Integration).
then again in applying mathematical-proof The CMMI defines parameters and practice
rules to refine the program from the levels in a set of models (software, sys-
abstract specification to the concrete code tems, integrated products, process develop-
level. ment, and supplier sourcing) that combine
Between these two extremes lies a wide to determine which practice level an orga-
range of lifecycle models, including rapid nization has achieved.
prototyping (also known as the Evolution-
ary development model) whereby a proto-
type is constructed to represent the func- Business value proposition
tionality of the program that is ultimately In order for organizations and individ-
to be developed. This prototype undergoes uals to develop efficient, effective pro-
cyclical refinement until it matches the grams, each and every time, it is necessary
user’s requirements, at which point the for them to adopt the appropriate SDLC
system should then be re-written using methodology for each project undertaken.
more efficient and structured code. (This The use of a lifecycle model enables orga-
step, however, is frequently not under- nizations to establish internal design stan-
taken.) dards for their programming. This helps
Many models based upon the ‘‘Water- them to be more efficient at measuring
fall” model have been developed. This their productivity, to be consistent in their
approach is based on the unlimited repe- training and recruitment policies, and to
tition of five basic stages. The stages are develop metrics programs in accordance
requirements definition, system and soft- with their processes.
ware design, implementation and unit test- There is a wide variety of SDLC method-
ing, integration and system testing, and ologies available to developers, ranging
operation and maintenance. In 1988 Barry from the formal to the informal. In
Boehm advanced a Spiral model, which between the informal and the formal are
has four repeated phases: determine objec- Structured methods such as Jackson structured
tives, alternatives and constraints, evaluate design, which have been popularized and
alternatives and identify and resolve risks, modified by consultants and by industry
develop and verify the next-level product, practices (such as JAD and RAD). Some
and plan next phase. The model can be SDLC models are the result of the require-
visualized as a ‘‘nautilus shell,” such that ments of a particular industry or regu-
the developer starts in the middle with a latory body, and contain design criteria
review, determines the risk and builds a necessary for a specific type of software
prototype, considers the prototype in opera- (such as MIL-STD-498, a merger of DOD-STD-
tional issues, and then plans the next stage. 2167A and DOD-STD-7935A for military soft-
This set of process operations is repeated, ware applications).
gradually moving development toward a The software industry has a vast array
production-quality system with each itera- of tools, vendors, and consulting firms
tion of the model. that support and offer advice on method-
To assist developers in understanding ologies or parts of methodologies and
what level of formality is associated with the management techniques that surround
the methodology they are using, the Soft- them.

300
Software metrics

Summary of positive issues ity associated with a software program and,


There is a wide array of well-documented by counting the unique number of oper-
and -researched methodologies that range ators µ1 and number of unique operands
from the formal to the informal. Consult- µ2 , together with the total number of
ing, tools, and support are available for occurrences of operators N1 and the total
many methodologies. number of occurrences of operands N2 in a
program, constructed a formula intended
Summary of potentially negative issues to measure the intellectual effort required
Choosing the wrong method can be expen- to produce the program, measured in units
sive, time-consuming, and potentially catas- of ‘‘mental discriminations,” of which a
trophic, depending upon the project. The competent person is capable of 5--20 per
professional skill levels required for each second, according to psychologist John
vary considerably; formal methods, for Stroud:
example, while being the most rigor- Effort = µ1 N 2 (N 1 + N 2 ) log2 (µ1 + µ2 )/(2µ2 )
ous also require extensive rigorous formal
Another metric that has frequently been
training on the part of the software engi-
used is the basic count of the number
neer and of any user who has to be able to
of lines of code in a program and is an
read the mathematical specification.
example of a direct metric that does not
References depend upon any other measurement. This
r www.sei.cmu.edu. metric has been incorporated into many
r I. Sommerville (2004). Software other models, including Boehm’s Construc-
Engineering (New York, Addison-Wesley). tive Cost Model, known as COCOMO, a
r R. Plant and R. Gamble (2003). model that can be used to estimate the
‘‘Methodologies for the development of potential costs associated with a software
knowledge-based systems 1982--2002,” project through the equation
Knowledge Engineering Review, Volume 18,
Effort = a(KDSI)b
No. 1.
where effort is measured in person months,
Associated terminology: Structured a and b are constants determined by
design, Formal methods, JAD, RAD, the mode of development, and KDSI
Waterfall method. stands for thousands of deliverable source
instructions. There are three modes of
development: organic (the team has worked
Software metrics together before, the domain and system
are familiar), semi-detached (the team has
Foundation concept: Software development lifecycle. mixed skill levels, and some familiarity
Definition: A well-defined measurement of some with the domain and the system), and
aspect of software or software production processes. embedded (complex heterogeneous systems
development). The COCOMO metric is an
Overview example of an indirect metric that involves
Software metrics have been used in rela- one or more other metrics.
tion to software development and software Both of these metrics have enjoyed wide
engineering since the 1960s with varying popularity, regardless of the fact that they
degrees of success and relevance. The early can produce wildly different results when
metrics research of Stroud attempted to applied to the same sample programs.
measure the skill levels of programmers, They are intended to work with increas-
whereas Halstead considered the complex- ing degrees of accuracy as the developer

301
Software package

moves from the highest levels of design Summary of positive issues


to more concrete implementation designs. There is a large and established metrics
However, they also share the disadvantage literature. There have been many tools
of being able to produce final effort estima- developed to automate the metrics data-
tions only once the software has actually collection process. Algorithm analysis is a
been created and the actual total number form of metric that is based upon sound
of lines of code in the system is known. Nei- mathematical principles and helps develop-
ther can directly predict the effort required ers understand the complexity of the code
for a new project with any accuracy, which they are developing in a formal manner.
would be a much more valuable metric.
It is also problematic that the code devel- Summary of potentially negative issues
oped and the circumstances in which it is There is considerable debate over the worth
developed are never the same twice, mak- of many software metrics such as Hal-
ing effort forecasts difficult to produce. stead’s, and many metrics are no longer
There are metrics that attempt to predict used in commercial systems development
the size, complexity, or effort required to because they do not map onto object-
create a program on the basis of its require- oriented design and development methods.
ments, so that they can be used to estimate Metrics need to be adjusted as the systems
the costs for a new project. These include they monitor change (e.g., measurements
Albrecht’s Function Point Analysis, DeMarco’s of attributes of Cobol programs can not
Bang Metric, and Putnam’s SLIM model, but be equated with the same measurements
no single model has been shown to be applied to C++ programs); failure to adapt
especially accurate in predicting software- the metrics will result in wildly incorrect
development costs or associated effort, so results.
models need to be used with a degree of
caution. Reference
r N. Fenton (1991). Software Metrics
Business value proposition (London, Chapman and Hall).
Metrics have been used in many ways
Associated terminology: Benchmark,
within computing; these include the
Algorithm.
benchmarking of systems’ performance,
developing reliability models of systems’
performance, quality-of-service metrics,
and measurement of algorithm complex- Software package
ity. Owing to the weak statistical base
and mathematical underpinning of many Definition: A software package is an application or a
metrics, care needs to be taken in the suiteofapplicationsanddatapurchasedfromavendor.
selection of any metric utilized. There is a
wide variety of opinion on the validity and Overview
usefulness of the various software metrics: The term Software package is typically asso-
Halstead’s system has been described as ciated with an application written by a ven-
‘‘convoluted” and ‘‘unreliable” (‘‘Software dor independent from the company pur-
metrics: good, bad, and missing,” C. Jones, chasing that software. This can range from
in Computer, September 1994), and as a custom macro for a spreadsheet applica-
‘‘among the strongest measures of main- tion to a full ERP system.
tainability” (‘‘HP-MAS: a tool for software Commercial software packages are
maintainability,” P. Oman, University of generally written by vendors for sale to a
Idaho Test Laboratory, 1991). wide range of customers (for example, a

302
Software package

word-processing application such as panies from having to write systems for


Microsoft Word is a software package for many activities, including human-resource
a mass market). While many packages are management (HRM), customer-relationship
available to users only in an executable management (CRM), manufacturing, and
form (binary code), some vendors provide supply-chain management (SCM).
open access to the source code of the While ERP systems are powerful and pro-
software. Such packages are known as vide many features, and even though ven-
‘‘open-source” systems, allowing users to dors attempt to write industry-specific ver-
modify the code to meet their own needs. sions (healthcare, retail, automotive, etc.),
Vendors who provide open-source code a package can never be all things to all
access place restrictions on what may people. Independent software vendors fill
legally be done with the code, controlling, the void, producing software modules for
for example, whether that code may itself specialty functions. One such example is
be sold or whether access to it may be the need for wine producers to have a soft-
provided only for free to other users of the ware package to manage the production of
package. wine; the problem of blending and mix-
Commercial software vendors who pro- ing wines to achieve a consistent blend for
vide only the executable code usually allow the vineyard is a niche task, outside of the
programmers to access their code through mainstream products of the ERP vendors.
what are known as Application program inter- To solve this problem, they would turn to
faces (APIs). These are vendor-defined con- an independent developer, or typically to a
ventions, routines, protocols, and tools for company that is certified by the ERP vendor,
accessing a particular aspect of the system, for a solution. Sometimes, these third-party
but not the whole system. The APIs are vendors have ‘‘canned” solutions that may
essential for developers who need to move be quickly modified to meet a customer’s
the package between operating systems or need in a particular area, or else they may
interface it with another application, for have to write a completely new application.
example. A third category of software pack- In making a determination of whether to
age is known as ‘‘freeware” and this is asso- build or buy a system, the question CIOs,
ciated with software that is open to use by business owners, and IT managers need
anyone, potentially with no restrictions on to address is whether or not the system
use but usually with the provision that it they are considering is core to their busi-
is not repackaged and resold. A final cate- ness. If it is, then they need to determine
gory or term associated with software pack- whether they should buy a software pack-
ages is ‘‘shelfware,” which refers to soft- age, buy and modify a package, or build the
ware that is purchased but never used and system themselves. They also need to con-
resides on the shelf of the user. sider the operation and maintenance of the
system, and whether this will need to be
Business value proposition outsourced. If the system is not core,
Software packages free individuals and the software should reflect best practices,
organizations from developing their own be easily modifiable, and adhere to indus-
code or ‘‘reinventing the wheel” as it is try standards; typically such a solution can
commonly termed. Clearly it is in very few be found in a pre-existing package.
people’s interest to write their own word- Other key issues are the following: will
processing system or spreadsheet applica- the package run on the infrastructure that
tion. At the other end of the spectrum, ERP the company has in place or will it require
vendors provide very large systems with a new infrastructure (hardware, additional
high degree of functionality, and save com- software, and net ware); will the vendor

303
Spam

be in existence during the entire expected to be examined from a process and tech-
lifespan of the product; and what levels nical perspective prior to purchase. This
of support will they be able to offer in includes such testing as stress testing to
regard to upgrades, maintenance, and tech- determine the limits of an application to
nical support? Other issues involve the abil- handle boundary problems, e.g., a sales-
ity of the software to support the com- entry system that is capable of handling
pany’s existing and future communications downloads of 1000 point-of-sale data entries
protocols (e.g., X.12, XML), and the cost per hour into its database might not be
per user to run the software. All of these scalable and hence would be unsuitable
issues and many others need to be consid- for a hypermarket dealing with 10 000
ered during the due-diligence period of ven- entries per hour. The purchase of an appli-
dor selection and should ultimately be cap- cation in a software package will come
tured in a legal contract. with a contract; however, all possible future
Organizations need to determine their issues can not be predicted in the con-
ultimate optimal business process require- tract. For example, in 1999 a contract
ments (business process re-engineering) for an accounting system would not have
prior to selecting the package, rather predicted the need to include provision
than letting the package completely dom- for Sarbanes--Oxley; hence the question of
inate the selection process. While there who pays for this upgrade would not have
might not be a complete one-to-one match been addressed, even a provision to ensure
between the organization’s desired best- that the vendor make available, for a fee,
practice model and the process model upgrades to cover such process require-
provided by the system, this sequence of ments might not have been included in the
actions will allow the selection of a package contract.
that comes closest to the ideal, and adjust-
ments may be subsequently made at the Reference
r N. Hollander (2000). A Guide to Software
system and organizational levels.
Package Evaluation and Selection: The R2 ISC
Method (New York, American
Summary of positive issues Management Association).
A wide variety of software packages is
available, covering requirements for mass Associated terminology: ERP, X.12, XML.
and niche markets. These include open
source, freeware, and commercial software
to which there is no code access; each type
of vendor and code offers different solution Spam
options based upon the process and sourc-
Foundation concepts: Email, Internet.
ing needs of the organizations.
Definition: Unsolicited, unwanted, bulk commercial
email.
Summary of potentially negative issues
While the range of vendors and solutions Overview
is wide, there is also a wide variety of Advertising is big business; direct mailing
solutions on the market. Some software to consumers or potential consumers is
is full of errors and bugs, some known, very popular, as is demonstrated by the
some unknown; and unfortunately price is quantity of junk mail delivered daily. Bulk
not an indicator of quality as one would mailing of small cards costs 24c/ per item
expect in, say, an automobile purchase. through the US Postal Service, and the
The quality of a software package needs cost is 39c/ per item for anything larger

304
Spam

than a postcard.† In addition to the cost of subject line for pre-chosen key sequences
postage, there are also printing, assembly, such as ‘‘Lotto Number Predictor” and the
and addressing costs, but still unsolicited trade names of certain little blue pills.
direct-mail advertising abounds. Email is Deliberate misspellings or false headings
free. Hence spam. easily defeat most systems.
The advantages to the advertiser are Spam exists only because sending email
manifold. Complex typesetting and high- is free. The decision that the cost of inter-
quality images may be used without cost; net access for most subscribers should
emails may be sent automatically to entire be charged as a monthly flat rate, with
easily purchasable mailing lists without no usage-dependent costs, was arrived at
any human processing; and, most impor- mostly by default, and certainly favors the
tantly of all, email is unregulated. In com- spam industry. If the origination of data-
parison with mail fraud, which is a well- bearing transmissions actually cost some-
understood crime, whose victims have lit- thing, some fractional amount, it would
tle difficulty finding out how to report it, not impact the average subscriber: the total
victims of electronic mail spamming have monthly bill would on average remain the
a more difficult time seeking redress, since same, it would just be calculated differ-
the spammers work hard to circumvent the ently. Only those who transmit inordinate
regulations and remain untraceable. amounts would find significant increases
Although email messages seem to have in costs. There is no fundamental right to
return addresses, they do not really mean free long-distance telephone calls; it may
anything. Email protocols do not require a seem surprising that internet traffic should
valid ‘‘from” address, and make no attempt be treated so differently, to the detriment
to verify those provided. Once an email has of all users. The TCP/IP protocol family as
reached the recipient, it is virtually impos- it currently exists has no mechanism for
sible to determine where it really came accounting and charging, but far greater
from, so unscrupulous retailers feel free design challenges have been met before.
to say anything they want. Illegal and con-
trolled substances, defective or nonexistent Business value proposition
merchandise, stolen goods, everything is on The spam debate has positives and neg-
offer. atives depending upon whether you are
Spam-prevention software generally a company that uses email as a direct-
attempts to flag unwanted email after it marketing channel or are an unhappy
has been delivered, telling the recipient recipient. Email is a cheap mechanism
that they probably won’t want to waste for pushing a direct-mail campaign out to
time reading it, but not preventing it potential consumers. Recipients, however,
from arriving. The effectiveness of even find spam annoying and defending against
this technique is severely limited by the spam to be a waste of corporate and indi-
fact that spam can be detected only by vidual resources.
reading and understanding the content of
a message. The natural-language problem Summary of positive issues
simply has not been solved: computers can Spamming has led to the development of
not understand free-form human commu- several laws to protect the consumer. The
nication. Current techniques are barely Controlling the Assault of Non-Solicited
more sophisticated than scanning the Pornography and Marketing (CAN-SPAM)
Act of 2003 was developed to regulate com-

USPS pricing, obtained from www.usps.gov, September merce by imposing limitations and penal-
2006. ties on the transmission of unsolicited

305
Spam

commercial electronic mail via the inter- prohibits deceptive subject headings and
net. The act covers and attempts to prevent requires the inclusion of a return address
spamming and the avoidance of account- or comparable mechanism in commercial
ability by spammers. Its more important electronic mail. The act also has a provision
sections forbid the following actions: to establish a ‘‘do not email” registry that
would allow recipients who object to receiv-
r the unauthorized use of a protected ing spam email to register, and requires
computer to trasmit multiple that the sender must stop sending the
commercial emails; email within ten days of the objection, but
r the use of a protected computer to in a Federal Trade Commission report to
relay or retransmit multiple messages the US Congress they stated that this could
with the intent to mislead the be counter-productive since spammers, pos-
recipients; sibly outside US jurisdiction, might use
r the use of materially false header the list as a source of email addresses,
information in multiple commercial and further that such a list ‘‘would raise
emails, and the distribution of such serious security, privacy and enforcement
emails; difficulties.”
r registering five or more email accounts
or two domain names using a false
identity with the intent to transmit Summary of potentially negative issues
multiple commercial messages; The CAN-SPAM legislation has been crit-
r the acquisition of email addresses by icized for having several gray areas and
‘‘harvesting” web sites without the loopholes through which spammers can
consent of the site’s operator; attempt to circumvent the law. Spammers
r the false registration of an internet continue to use a variety of mechanisms to
domain for the purpose of setting up place temporary or untraceable addresses
a web site to capture email addresses; in their email headers and even place just
and one line of ‘‘news” in the text to ensure
r sending commercial messages to that the message contains ‘‘newsworthy”
randomly generated email addresses. content as the act demands. An additional
problem is that a large volume of spam
To protect the public and allow the spam originates from nations beyond the juris-
filters to operate more effectively, the CAN- diction of the CAN-SPAM law. The United
SPAM act stipulates requirements for the States has not yet established a ‘‘National
transmission of messages. The act requires Do Not Email Registry.”
that the header information be technically References
accurate: it must include a correct orig- r J. Zdziarski (2005). Ending Spam: Bayesian
inating email address, domain name, or Content Filtering and the Art of Statistical
IP address. The act also requires that the Language Classification (San Francisco,
‘‘from” line in the email accurately iden- CA, No Starch Press).
tify the person sending the email, and r T. Muris, M. Thompson, O. Swindle, T.
the ‘‘header information” accurately repre- Leary, and P. Harbour (2004). National Do
sent the computer used for transmission of Not Email Registry: A Report to Congress
the email and not hide the fact that the (Washington, DC, Federal Trade
email has been relayed through a series Commission).
of machines. The header must also place
a warning label on commercial email con- Associated terminology: Email, Internet,
taining sexually oriented materials. The act Internet protocol, Law cross-reference.

306
Spoofing

once computer A has confused computer C


Spoofing
with computer B in its address-resolution
Foundation concept: Security. tables, all transmissions intended for B will
Definition: A form of attack (attempted break-in or be sent directly to C instead. Computer C
interception of data) in which one system pretends to simply invites the user to log in exactly as
be another. computer B (the real server) would, records
the user-name and password, and that’s
Overview that.
Spoofing comes in many forms, the sim- Spoofing can also occur at the DNS
plest of which, IP-spoofing, gives a good (Domain name service) level. When a user is
example of the whole spectrum. IP spoof- first making a connection to www.what-
ing relies on the fact that computers on ever.com, their computer will issue a DNS
a network generally know each other by enquiry to find out the corresponding IP
their IP addresses, but communications on address. An attacker need only intercept
a Local area network (LAN) use a totally dif- the enquiry and provide their own false
ferent identification method, MAC addresses, response, and the user will be tricked into
instead. If a user (on computer A) wishes connecting to the wrong server again.
to make use of a password-protected server There are many other levels on which
(on computer B) that has an IP address spoofing can occur, since so many of the
of 192.168.123.9, that user’s computer will commonly used internet protocols are com-
use the address resolution protocol (ARP) to pletely insecure. The best defense against
find the equivalent MAC address. Unfor- spoofing is physical security, ensuring that
tunately ARP is not a secure protocol. It no unauthorized systems can possibly be
works in a very simple way: the request- connected to the LAN. Of course, it is
ing computer broadcasts a ‘‘Who has IP incumbent upon the administrators of
192.168.123.9?” query to every computer on other levels of the network, and ISPs, to pro-
the LAN. If everything is working properly, vide the same security for their equipment.
just one computer will respond with ‘‘I am The adoption of secure encryption-based
192.168.123.9, and my MAC is 11-22-11-44- protocols instead of the traditional inse-
33--11.” The requesting computer notes that cure ones would make spoofing virtu-
reply, and thenceforth sends all communi- ally impossible, but replacing the standard
cations over the LAN directly to 11-22-11-44- internet protocols either requires agree-
33--11. ment and cooperation from the entire
ARP is a very simple, efficient, and effec- internet community, or results in a locally
tive protocol, when everybody is on the secure network that can not communicate
same side. All it takes is for an attacker to with the rest of the insecure internet.
momentarily make the real secure server
inaccessible (perhaps just by unplugging
its network cable for a fraction of a sec- Business value proposition
ond, but there are many more sophisticated The use of spoofing techniques to fool a
techniques), and transmit its own dishon- secure source into giving access rights or
est response to the ARP query, presenting other privileges to an unauthorized com-
its own MAC address as the translation for puter is a well-known and easily tried
192.168.123.9. ARP is so trusting that the technique used by hackers to gain valu-
invading system can transmit its dishon- able information, should a network be left
est response at any time, before or after undefended. Therefore it is vital that the
the query is made, and the false answer CIO act through the company’s chief secu-
will be noted and used by all. Of course, rity officer and network administrator to

307
Spyware

ensure that all adequate protections are in tions’ budgets; however, some companies
place. simply lack the funds to develop sophis-
Defending against spoofing is best ticated market research and advertising
achieved through physical security mea- campaigns. A system that would provide
sures, ensuring that no unauthorized com- almost unlimited opportunities for market
puter can access the network, and this pol- research and direct-to-consumer advertis-
icy should be enforced at all other net- ing would be very tempting indeed. That
work levels, including the ISP. This requires is exactly what Spyware is.
that organizations use ISPs and service Simply visiting a web site once is enough
providers that take spoofing seriously and to cause software from that site to be
actively enforce their policies. Concern in downloaded, installed on your own com-
this area may lead to the decision to puter, and executed, if your web browser
use only private networks and to adopt does not have sufficient security settings
secure encryption-based protocols rather selected. Software can similarly be embed-
than relying upon the insecure ones. ded in emails and automatically installed if
the email client application is not secure.
Summary of positive issues Very commonly, spyware takes the form of
Most spoofing techniques are well known a Trojan horse, some piece of free software
and may be defended against through the that is claimed to perform some useful task
use of private networks, encryption, and is made available for download through the
other techniques. internet. Many users will happily download
it, install it, and try it out. Maybe it does
Summary of potentially negative issues perform its advertised task, and maybe it
New spoofing techniques are always being doesn’t, but, if it is a spyware delivery vec-
developed and resources need to be tor, it will continue to run after it appears
deployed to keep track of threat levels to have stopped, and will have added itself
and address any newly revealed security to the system’s start-up list, so that it will
weaknesses. automatically run every time the computer
Reference is started.
r B. Schneier (2004). ‘‘Customers, Once it is running, spyware has the unob-
passwords, and web sites,” IEEE Security structed use of the entire computer. It
and Privacy, Volume 2, No. 4. can be designed to search out certain files
(address books, password files, etc.), or to
Associated terminology: Phishing, Virus, look for any file containing certain kinds
Trojan horse, Worm. of data (account numbers, PINs, etc.) and
send the contents to the spyware’s author.
It can monitor the keyboard and record
Spyware everything that is typed; it can record every
application that is run, and every file that
Foundation concept: Security. is accessed; it can record every web page
Definition:Illicitandusuallyillegalsoftwarethatspies that is viewed, and every email that is sent
on a computer’s stored data and its users’ activities, or received. Everything the computer does
reporting them through an internet connection or an can be recorded and transmitted. It could
autodialing modem to a third party. even potentially turn on a web-cam and
capture real-time pictures of the user in
Overview their home or office.
Market research and advertising consume Traditionally, the term spyware is limited
an enormous proportion of many corpora- to software that passively spies, as detailed

308
Spyware

above. Software is not limited by the term prevents proper system functions. Much
used to describe it, and spyware will com- depends upon the type of spyware that
monly play a more active role. Spyware is resident upon a system. Typically spy-
can be written to redirect web searches, so ware attaches itself to a computer through
that the operators can insert their own pre- an access to a web site or through an
ferred sites instead of the results provided email message, and then collects informa-
by the genuine search engine. It can redi- tion on the computer or its user’s activi-
rect simple web accesses so that the opera- ties. This information is then transmitted
tors’ advertisements are shown instead of over an internet connection to a remote
the real pages. The most common non- user. Clearly, in the worst instances spy-
passive act is to ‘‘pop up” unsolicited ware can run undetected for long periods
advertisements. of time, transmitting industrial secrets to
In many cases, spyware is illegal, but, in a third party, or set a web-cam in action,
the United States, legislation to protect pri- monitoring an individual’s activities in the
vate parties (both individuals and corpo- ‘‘privacy” of their home. However, not all
rations) is very weak at the federal level spyware is put on a system from an external
(government computers and those of finan- source; spyware can be placed on a system
cial institutions are protected, of course), by an employer to monitor employees’ com-
and left to state legislatures. State laws are pliance with company policies, or to check
not at all uniform, and, since most use against access to unauthorized sites. Par-
of spyware is an interstate attack, state ents may wish to place spyware on their
laws can be difficult to enforce. Many spy- children’s computer for similar reasons.
ware operators manage to convince them- The best way to avoid spyware is by not
selves that their actions are neither legally being connected to the internet; however,
nor morally wrong, and are even willing this is not always practical and the next
to believe that their victims are really stage is to install a Firewall and software
beneficiaries. to filter out problem files. The comput-
Spyware is mainly a plague on users of ers themselves can also have their security
personal computer systems. This is partly levels raised to forbid the installation or
because spyware requires some effort to running of web-resident software. Finally,
create, so that effort is directed toward the there are many commercial vendors who
systems that most people have. It is also provide anti-spyware software (some for
partly because of the large numbers of secu- free) and these act to sweep all aspects of
rity flaws discovered in popular operating a computer’s memory, flagging and then
systems and third-party software made for removing undesired spyware and adware.
them. The only way to maintain privacy is Spyware itself evolves continually and thus
to avoid spyware, and the only way to avoid it is necessary for users to update their soft-
spyware is to run a secure computer system, ware’s library of possible spyware sources
in particular never running software that and categories.
didn’t originate from a completely trusted
source. As with viruses, there are a few anti- Summary of positive issues
spyware applications available, and in this Spyware can be used to actively monitor
case some of the best are also free. any computer. Companies may wish to
monitor the online activity of an employee,
Business value proposition whereas a parent may wish to monitor
For corporations and individuals, spyware their child’s online behavior. Anti-spyware
is generally an annoyance that is ignored software is available. Certain categories
until it builds to a critical level and of spyware activity are illegal in some

309
Storage

jurisdictions. In the UK, all software no capacity for any sustained activity; hun-
installed on a computer without the dreds of millions of times per second, it
owner’s consent is illegal. In the US many must access its memory to find out what
states have similar laws, but there is no it is supposed to be doing next. Without
federal equivalent. memory, it would grind to a halt in a few
nanoseconds.
Summary of potentially negative issues Computer storage is usually divided into
Spyware can cause serious damage to an two distinct categories, Primary storage and
organization by stealing intellectual prop- Secondary storage. Primary storage, often just
erty, passwords, or other corporate data. called Memory, is the kind that is essential
Lesser damage is caused by its occupation for the nanosecond-to-nanosecond opera-
of valuable network bandwidth and slow- tion of the processor. Every instruction to
ing of infected computer systems, but this be obeyed by the computer, and every piece
can become critical if left untreated for too of data to be used, must be present in pri-
long. Spyware can be created to monitor mary storage. This means, of course, that
the personal behavior of a computer user, a computer must have as much primary
for example their online shopping activity storage as possible, but the importance of
and their web-surfing habits, or even to size is nothing when compared with the
turn on computer equipment such as importance of speed and reliability. The
web-cams. Legal recourse against spyware processing speed of a computer is com-
is limited in many jurisdictions to certain pletely bound by the speed (or Access time)
categories of software and intrusion. Spy- of its memory.
ware may be impossible to detect without There is an essential balance to be struck.
the aid of special software. The faster memory is, the more it costs,
and the more memory costs, the less of it
Reference a computer can have. Additionally, faster
r E. Schultz (2003). ‘‘Pandora’s box: memory consumes more electrical power,
spyware, adware, autoexecution, and and results in more heat that must be dis-
NGSCB,” Computers and Security, sipated from the circuits, which is a major
Volume 22, No. 5. problem in modern computer design. The
generally used solution is a complex com-
Associated terminology: Advertising,
promise. A typical modern computer has a
Virus, Trojan horse, Phishing.
large amount of reasonably fast memory
known as Main memory or ‘‘RAM,” usually
in the hundreds of millions of bytes, plus a
Storage much smaller amount of exceptionally fast
memory, known as the Cache. Whole appli-
Definition: Memory, disks, and other media used to cations and data sets are loaded into main
retain data and applications both in the short and in memory as usual, but, as an application
the long term. runs, the portions of that data that are in
immediate use are copied into the cache.
Overview This is a very complicated dynamic process
Memory or storage is one of the sine quibus performed by the CPU, but it is responsible
non of any computer system. Unlike with for significant improvements in processor
humans, it is not possible for a computer speed at minimal cost.
with poor memory retention to get by. The Another economic factor is the unsur-
central processing unit (CPU) of a computer, prising fact that it is much cheaper to make
its brain in anthropomorphic terms, has Dynamic RAM, memory that can retain

310
Storage

information for just a few thousandths of a be programmed with its permanent con-
second, than it is to make Static RAM, which tent at the factory. PROM (Programmable
retains information for unlimited periods. ROM) comes from the factory blank, and
The difference is so great that all modern can be programmed (once only) using spe-
computers use almost exclusively dynamic cial equipment. Another variety, EPROM
RAM, and are equipped with extra circuitry (Erasable programmable ROM) can actually be
required to Refresh, or remind every mem- erased and reprogrammed (erasure is either
ory cell of what it is supposed to be stor- by exposure to ultraviolet light, or elec-
ing, hundreds of times per second automat- tronically), but erasure is a very slow pro-
ically. So much research and development cess, and EPROM does not provide a viable
has been devoted to improving the perfor- form of general-purpose non-volatile mem-
mance of dynamic RAM that the overall ory. ROM is used primarily for control-
effect is a benefit. Static RAM has been rel- ling embedded processors and for storing
egated to a very small unit that runs on the permanent system-start-up procedures
battery power to keep essential system set- (BIOS) in a modern computer.
tings when a computer is turned off. The term ‘‘RAM,” which is now used to
Another unfortunate economic aspect mean general-purpose memory, is in fact
is that the only affordable memory is an acronym for Random-access memory. Ran-
Volatile. That means that it works only dom access means that the various items
when the power is on. Up until the mid stored in the memory may be accessed in
1970s, when all computer memory was any order at any time, rather than Serially.
expensive, the main memory technology in A good analog is the difference between
use was magnetic: vast arrays of minute watching a film on a DVD and on a video
magnetizable doughnuts, called Cores, that tape: DVDs are random access, you can skip
each stored one single bit in a magnetic to any point at any time; video tapes are
field. These cores were Non-volatile; a com- serial, the only way to get to a point half
puter could be turned off for a week, and way through the film is to wind through
still retain everything. Unfortunately, core half of the tape. There is no serial-memory
memory takes up a lot of space, is much technology in current use, and the mean-
slower than semiconductor memory, and is ing of RAM has changed slightly. Techni-
vastly more expensive. The only part of core cally speaking, ROM is a form of RAM; its
memory still surviving is the name; many contents may be accessed in random order.
still call main memory ‘‘the core,” and the In modern usage, however, RAM means
extensive status report produced when an strictly normal general-purpose memory.
application crashes is called a Core dump. Because non-volatile RAM is prohibitively
One cheap form of non-volatile memory expensive, and even the usual volatile RAM
is available; it is known as ROM, or Read- is by no means cheap, computers also need
only memory. As the name suggests, ROM secondary storage. In the economic trade-
provides only half of the functionality nor- offs of computer design, primary storage
mally required of computer memory: its is chosen to be as fast as possible with-
content can be read (i.e., looked at), but out becoming too small to use; secondary
can not be written (i.e., modified), so it storage is chosen to be as large as possi-
is suitable only for data that absolutely ble without becoming too slow to use. Sec-
never changes. Even essential basic appli- ondary storage currently uses totally differ-
cations and operating systems can not rea- ent technology, almost invariably the Hard-
sonably be stored in ROM because it would disk drive. Disk drives use magnetic fields to
make upgrades impossible. ROM comes store data, and so are non-volatile, retain-
in many varieties. The basic kind must ing data for many years without use. The

311
Storage

magnetic fields are stored in concentric Technologists can, however, place their
circular tracks on a rapidly spinning disk software into a hardware memory device
(hence the name). Typically, random access that is non-volatile in nature if desired.
to an item of data stored on disk takes of Various devices are available that allow
the order of a million times longer than developers to burn their software into a
random access to a data item in primary PROM, test the program in the PROM, and
storage, but capacities tend to be approxi- move the PROM onto production memory
mately one thousand times higher for sim- devices.
ilar prices. The significantly greater capac- Secondary storage is available in a wide
ities and non-volatility of disks make the variety of configurations and needs to
combination of semiconductor RAM for pri- be considered as part of a company’s
mary storage and disks for secondary stor- Information lifecycle (ILC) and the manage-
age a universal choice. ment of the company’s data. There are
many companies offering data-storage ser-
Business value proposition vices that can be used as outsourcing
The selection and use of storage in a corpo- partners.
rate setting requires that a business case be
developed for the whole system, taking into
Summary of positive issues
account the processes that are going to run
A wide range of devices and specifications
upon it. In a system for a person whose only
is available for both primary and secondary
requirement is to use a word processor,
storage. The technology is well understood
memory speed and size will typically not
and supported by vendors. Software and
be the bottleneck, but rather the person’s
hardware tools are available to help users
typing speed. However, in the case of a file
to analyze and optimize their memory. New
server that is used by corporate users, an
forms of memory are continually evolving
assessment of memory requirements must
(e.g., USB flash-memory devices). Primary
consider issues of performance, loads, and
storage is available in a wide variety of
network demands. Adding lots of mem-
configurations.
ory to a corporate server that has a low-
performance processor could still greatly
increase performance. Summary of potentially negative issues
RAM is available in a wide variety of The wide variety of memory types requires
types, each of which is available with a that the correct memory be selected for
wide array of performance options. These each situation. Adding memory to a com-
choices facilitate the acquisition of a spe- puter system might not make the system
cific memory component for a specific more effective or faster since the bottle-
purpose (e.g., cache for main memory), neck may lie elsewhere in the system. Some
and allow systems to have their memory memory specifications and types lose their
upgraded, potentially meeting new require- manufacturer’s support as they age and are
ments without the user having to purchase superseded by newer specifications. While
a completely new system. primary memory can last for ever, barring
Commercial and personal computer physical harm, secondary memory has a
users usually never change the ROM config- shorter lifespan because it contains mov-
urations of their systems, since these typi- ing parts, and an assessment of the con-
cally store system-specific applications, but dition of the memory needs to be under-
newer systems use Flash ROM, which does taken regularly, so that the appropriate
allow the BIOS to be updated without any upgrades, maintenance, and adjustments
special equipment needs. can be made.

312
Structured design methodologies

Associated terminology: Memory, Disk, It should be noted that there is no


Optical storage, Backup, Memory and disk one universal diagrammatic structure or
size. notation; many different notations have
been developed since the 1970s, some as
open standards such as Wirth’s Step-Wise
Structured design methodologies Refinement, and some associated with a par-
ticular company or designer such as Jack-
Foundation concept: Software development lifecycle. son Structured Programming and the Yourdon
Definition: Structured design methodologies are a set Design Method. All of the methods do tend
of techniques available to systems analysts and pro- to have some shared properties, and involve
grammers to assist them in producing system designs structures that are primarily based upon
and programs with higher levels of validation and ver- the hierarchical decomposition of func-
ification than would otherwise be attainable. tional modules into smaller modules with
specifications of the data passing between
Overview them. The modules are usually drawn or
Structured design methods started to marked in such a way as to denote that
appear in the late 1960s, when the num- a sequence of activates occurs in that
ber of systems under development started module (step one, then step two, then
to increase dramatically and the quality step three), that a selection process occurs
of those systems, in terms of the code (perform process one or perform process
matching the specification, was generally two), or that an iterative process occurs
poor. In addressing this, researchers such (perform process one until condition two
as Dijkstra, Wirth, Baker, Knuth, Stevens, occurs).
Myers, and Constantine focused upon devel- Various methodologies treat the con-
oping structuring techniques to help devel- structs with different rules to avoid conflict
opers and implementers produce better sys- in the code (e.g., Jackson Structured Design
tems. Some of the researchers concentrated does not permit sequence, selection, and
on design issues and some upon techni- iteration to be at the same level in a tree; a
cal and implementation issues, and eventu- sequence of activities that themselves can
ally, by the 1980s, a series of methodologies be decomposed into further activities is
and techniques for the creation of complex permitted, but no one node may be directly
high-quality systems had been developed. decomposed into a mixture of the differ-
The term Structured design was originally ent kinds of sub-nodes). Such rules are
used by Stevens, Myers, and Constantine in intended to ensure that the final leaf nodes
their book Reliable Software through Compos- in the design tree represent program state-
ite Design, which was published in 1974 and ments or activities that ultimately can be
in essence an extension of what IBM had translated into program code with good
termed the Hierarchical input process output design characteristics.
documentation technique or HIPO for short. Structured design was originally aimed
The structured design techniques use box- at the development of transaction-
and-line notation in order to present the processing systems in languages such as
design diagrammatically, with boxes of cer- Cobol, but has evolved in various directions
tain shapes representing a type of object since the 1970s and led to methodologies
such as a Module or a Predefined module, the such as Entity-relationship diagrams (ERDs)
lines linking boxes together in a tree struc- and Data-flow diagramming (DFD). These and
ture with a notational marker associated other structured methods have also been
with each link between the modules stat- built into the software lifecycle models
ing what data is passed between modules. such as the Waterfall model, which is used

313
Structured design methodologies

to develop a concept into a program, and supported by consultants and software tool
encourages the use of structured methods vendors.
at each stage. The structured approach is
also employed in commercial methodolo- Summary of positive issues
gies such as RAD. The structured approach The structured approach to systems devel-
continues to be used in various forms opment is well established, with a long his-
and hybrids and has evolved into methods tory and extensive literature. A wide vari-
such as UML to accommodate the current ety of software tools is available to sup-
Object-oriented paradigm. Structured models port structured methods and the associ-
are generally not considered rigorous ated management techniques that accom-
enough to develop mission-critical systems pany this approach to systems design.
that may, for example, involve potential
risk to human life; for such systems a
more mathematically rigorous approach Summary of potentially negative issues
to design, known as Formal methods, was Structured design was developed in the
developed. 1970s to aid in the design of transaction
processing systems in Cobol. Early method-
Business value proposition ologies may be less suitable to the devel-
The use of structured design techniques opment of object-oriented program designs
leads to higher-quality, more maintain- and care needs to be taken in method-
able, and better-documented programs and ology selection. Structured methods tend
systems than would be the case if no uni- to be considered ‘‘less rigorous” in nature
fying design approach were taken. The since they don’t involve the mathematical
adoption of structured methods allows formalism of formal methods; as a conse-
companies to use a consistent approach to quence they may be considered unsuitable
developing their systems, allowing consis- for systems requiring absolute precision in
tency in training and more efficient plan- the development process.
ning and scheduling, recruitment, and References
management. The structured approach r E. Yourdon (1979). Classics in Software
allows for the adoption of commercial Engineering (New York, Yourdon Press).
tools and packages to support this style r G. Myers (1974). Reliable Software through
of development as well as the adoption Composite Design (New York, Mason and
of management practices such as RAD to Lipscomb Publishers).
speed up development, and cost model-
ing techniques such as COCOMO II to help Associated terminology: Formal methods,
determine system costs and effort require- UML, RAD, Object-oriented, Cobol, ERD,
ments. The methods are well known and DFD, Waterfall model.

314
TCP/IP (Transport Control Protocol for the internet protocol)

T-channel), and using the letter E instead of


T-Carrier
T. So an E-3 carrier, for example, would have
Foundation concepts: Network, Bandwidth. 33% more bandwidth than a T-3 carrier.
Definition: T-Carrier is the generic name for digital
high-speed data-transmission lines more specifically Business value proposition
known in the United States as T-1, T-2, etc. Organizations can purchase capacity from
common carriers in terms of a leased T-x
Overview carrier line. For example, a company can
The basis of the T-Carrier system is the dig- lease a whole T-1 line, or, if a T-1 line
ital Channel, designated DS0 or T-0. A single provides more bandwidth than required, a
DS0 channel has a capacity or Bandwidth of smaller number of DS0 channels (known as
64 kbps (64 000 bits per second), which is a Fractional T-1 subscription) may be avail-
the amount of carrying capacity that Bell able at a lower price.
Labs originally defined as necessary to carry It is important to note that the speed of
a digitized voice signal over their network. downloads or uploads through a network
Originally, the name T-1 was used for a digi- depends not only upon the bandwidth pro-
tal transmission channel with a bandwidth vided by the ISP and common-carrier con-
of 1544 kbps that carried 24 separate DS0 nection, but also upon the nature of the
channels together with one 8 kbps chan- network that it is connected to. Thus, for
nel for system-related information. Now the speed-sensitive environments, it is worth
technology may be slightly different, and T- testing the actual speed of transfers.
1 has become a less specific name for any
1544 kbps digital carrier. Similar names (T-
2, T-3, etc.) are used for higher-bandwidth Summary of positive issues
carriers: ISPs and common carriers provide a vari-
ety of bandwidth capacities. The European
a T-1 line is equivalent to 24 channels or standards are compatible with the North
1.544 Mbps, American and Japanese standards.
a T-2 line is equivalent to 96 channels or
6.312 Mbps,
a T-3 line is equivalent to 672 channels Summary of potentially negative issues
or 44.74 Mbps, Organizations need to check the band-
a T-4 line is equivalent to 4032 channels width that they are actually achieving
or 274.2 Mbps, and given the nature of the network through
a T-5 line is equivalent to 5760 channels which they are connected.
or 400.3 Mbps. Reference
r The International Telecommunications
although it is not common to hear of any-
thing other than T-1 and T-3. In Japan, Union (http://www.itu.int).
exactly the same system is used, except Associated terminology: ISP.
with the letter J instead of T. Originally T-
lines were made of copper wire, and the
designation FT was used for fiber optics,
but now the distinction is often ignored. TCP/IP (Transport Control Protocol for
In Europe, a slightly different system is the internet protocol)
used, still based on 64 kbps DS0 channels
(called E-0), but in multiples of 32 rather Foundation concepts: Network, Protocol.
than 24 (giving a bandwidth one third Definition: The preferred protocol for reliable network
higher than that of the corresponding communication.

315
TCP/IP (Transport Control Protocol for the internet protocol)

Overview TCP is quite a complex protocol, and


is not always the most appropriate choice
TCP/IP is often thought of as a single for all applications. The need to transmit
thing, but TCP and IP are separate proto- acknowledgements for data, and acknowl-
cols with separate purposes. IP (Internet pro- edgements for acknowledgements, puts a
tocol, q.v.) is the logical mechanism used significant load on network capacity. In
to deliver data over a network; it provides cases in which responsiveness is more
an addressing and adaptive-routing system important (such as network telephony) or
that allows one computer to communicate reliability is not required (such as a net-
with another without knowing where it is work time service that transmits the time
or how to find it. IP does not provide any of day every minute: it would not matter if
reliability of service: most transmissions a few transmissions went astray), UDP (User
arrive at their destination, but not all. IP Datagram Protocol) is a popular and effi-
does not depend upon any other particular cient alternative.
protocol, it often uses ethernet, but does The division of responsibility between IP
not have to; in particular, it is not bound (routing) and TCP (delivery) is part of one
to TCP in any way. of the most important principles of systems
TCP (Transport Control Protocol) is one of design: Divide and conquer. Two separate sys-
the higher-level systems that are built on tems, individually testable and with their
top of IP to provide additional functional- own non-overlapping responsibilities, will
ity. TCP does require IP, but not vice versa. be much easier to develop reliably than
TCP relies upon IP to attempt to deliver will a single system trying to perform both
data correctly, but adds its own procedures tasks. For network programming, divide
for ensuring reliability. and conquer is usually taken much further,
A networked application that requires with systems divided into many layers. The
reliable communications, such as an email most popular view is the OSI (Open Sys-
server, could use IP to send all of its data. tems Interconnection) Seven-layer model (see
In that case, it would have to be carefully OSI for details), but the more traditional
designed to take account of the fact that four-layer model is often preferred. The
some transmissions will not arrive, some layers are as follows.
portions of the data may arrive in the
wrong order, and some may even arrive Layer 1, ‘‘link”: this describes the phys-
twice. The client and server would have ical connections (cables, connectors,
to exchange acknowledgements for every etc.) and how data is sent directly
piece of data, establish time-outs so that between two network participants that
lost data will eventually be re-sent, and pro- have a direct physical link. Example:
vide data identification so that duplicated ethernet.
and disordered data may be corrected. Layer 2, ‘‘network”: this describes how
TCP provides all of that functionality data moves around a network as a
automatically. The same application using whole, providing addressing and rout-
TCP plus IP instead of IP alone would have ing procedures. Example: IP.
no extra work to do beyond simply send- Layer 3, ‘‘transport”: this describes the
ing the required data. TCP provides appli- flow of data between one application
cations with thoroughly tried and tested and another, providing reliability to
procedures for all the essentials of net- whatever degree is desired. Examples:
work communications, relieving develop- TCP and UDP. Whereas IP is concerned
ers of one of the most difficult and error- with delivering data to the right com-
prone tasks. puter, TCP and UDP are concerned

316
Telnet

with directing it to the right applica- may require a higher network bandwidth
tion running on that computer. capability than less convenient alternatives
Layer 4, ‘‘application”: this is the ‘‘use- would need.
ful” layer, performing the high-level
Reference
tasks that are actually wanted, such r W. R. Stevens (1994). TCP/IP Illustrated
as delivering email or remotely con-
(New York, Addison-Wesley).
trolling equipment. The other layers
are really just services, whose sole pur- Associated terminology: Internet protocol.
poses are to make this fourth layer
possible.
Telnet
Business value proposition
TCP in conjunction with IP is the most Foundation concepts: Client–server, Operating
widely known protocol for general network system.
communications. However, it should not be Definition: A network application that gives remote
forgotten that there are alternatives. When access to any software that provides a command-line
deciding upon which protocol to choose, interface and is itself network enabled.
one must develop a business case that
examines the future requirements of the Overview
systems under consideration. This can be Before graphical user interfaces and win-
done through the creation and evaluation dowing systems became affordable and
of a set of performance indicators (such as popular, the regular way to control a com-
reliability) and then systems metrics (such puter was by typing verbal commands on
as performance requirements) determine a keyboard and reading a typed response.
which protocol would be the most suitable. This kind of interaction is still familiar
The business case will also consider the pro- to anyone who uses Unix or ‘‘the DOS
tocol requirements of other systems that prompt,” and is often still the preferred
the network would interact with, includ- mode of operation for programmers and
ing those beyond the corporate boundary. the technically minded. There is only so
Many organizations will also have technical much information that can be conveyed
staff capable of creating a proprietary pro- by clicking and dragging icons, and, when
tocol that exactly fits their specific needs; more is needed, the use of complex dialogs
a proprietary layer-3 (transport) protocol to provide the extra information can be
may certainly be a viable solution, since very tedious and inefficient.
it would ride upon IP, the lingua franca of The use of a Command-line interface,
the internet; a proprietary layer-2 (network) through which users type commands that
protocol is probably impractical, because it may be as simple or as complex as is
would not be recognized by existing net- desired, can provide a great increase in effi-
work equipment. ciency. The only cost is that users must
learn the language of the interface. Dialog
windows do at least ask for specific infor-
Summary of positive issues mation in specific places, so a user need
TCP is a well-known, highly reliable, and only understand the questions in order
established protocol. to use them properly. When faced with a
command-line prompt, there is generally
Summary of potentially negative issues no information on what the system will
TCP is a complex protocol, which places accept, and a user can not reasonably guess
a higher demand on networks and thus at the syntax and hope for the best. Use

317
Telnet

of a command-line interface requires some Of course, password protection will pre-


preparation and even expertise. vent unauthorized access if implemented
Some operating systems have a correctly, but typing passwords over net-
command-line interface as their pri- work connections can be very unsafe. When
mary means of control; Unix (and its the legitimate user types their password, it
variants, Linux, BSD, etc.) is a prime exam- must be transmitted to the server for veri-
ple. Control of the system and interaction fication. Anyone with a tap on the network
with applications are both conducted (e.g., a Packet sniffer) can capture transmis-
through typed commands (although win- sions and learn the password. As a conse-
dowing systems are of course available). quence, Telnet should be used only over
Other operating systems (for example a secure connection, protected with SSL
windows) are primarily controlled through (Secure-sockets layer), SSH (Secure shell) or a
a graphical user interface, but individual functionally equivalent encryption scheme.
applications may still provide their own Secure Telnet clients are widely available.
command-line interfaces for convenience. Telnet is also known as the Hacker’s
Telnet is a simple network application Friend, since it allows human users to com-
that provides remote access to any sys- municate directly with automated servers
tem or application that provides a network- devoted to some network task. For exam-
enabled command-line interface. Network ple, it is very easy for someone familiar
enabled is a key requirement. The system with the SMTP protocol (the language of
or application must be designed not just outgoing email servers) to connect directly
to accept commands typed on a keyboard, to a server and send email messages with
but also to act as a Server, willing to accept fraudulent return addresses and origina-
TCP/IP connections over the internet, and tion dates. This is, of course, really a prob-
treat transmissions received over that con- lem with unsecured servers rather than
nection as though they had been typed at with Telnet itself; if Telnet did not exist,
the local keyboard. This is not at all dif- someone would create it.
ficult for an application to do, but is still
essential. If the operating system itself pro- Business value proposition
vides a command-line interface, then that The value of Telnet is that it provides a
interface is normally inherited by all appli- simple command-line means of accessing
cations that run, so no extra programming a remote system via a network connection.
is required. This enables users to connect to remote ser-
The Telnet application acts as a Client, vices as desired over the internet. A posi-
connecting to any server as directed by tive aspect of the Telnet approach to sys-
the user, then simply acting as a commu- tems access is that network managers can
nications intermediary. Anything typed at remotely log into their system and perform
the keyboard is transmitted directly to the any system function as desired. Thus, for
server’s command-line interface; any out- example, if an oil company has a Unix sys-
put produced by the server is displayed tem on an oil rig, the network manager
immediately for the user to see. Telnet can restart processes and perform mainte-
provides a simple, cheap, and uniform nance from his or her office and does not
interface that allows systems and applica- have to board a helicopter and fly to the
tions to be controlled from any networked middle of the ocean, which would prob-
location. ably be the case for a systems manager
Pure Telnet is a very insecure system. trying to fix a non-‘‘reboot” problem via
If legitimate users can connect to a sys- an operating system without this access
tem using Telnet, then so can anyone else. feature.

318
Thirty-two bit

Summary of positive issues −128 to +127: more or less two-digit


Telnet provides a tried, trusted, and con- numbers in human terms. A 16-bit num-
venient method for accessing remote com- ber is restricted to the range −32 768 to
puters over a network. Applications to add +32 767: roughly four human digits. A 32-
varying levels of security to the system are bit number corresponds to approximately
readily available. nine human digits, and a 64-bit number
has up to 19 digits in human terms.
Summary of potentially negative issues An eight-bit computer has the hardware
Pure Telnet is a vulnerable access method required to work on eight-bit numbers
and needs to be reinforced with methods directly. It can still work on larger num-
such as SSL to encrypt the messages. Tel- bers, but has to perform a lot of extra work
net connects to line-command operating to do so. In general, a 32-bit computer can
systems and applications and these require perform a calculation on a nine-digit num-
instruction and training of users since they ber in the same time as a corresponding
are not as intuitive as the drag, drop, point, 8-bit computer could perform the same cal-
and click operating systems. culation on a two-digit number. More bits
means more processing power.
References The first microprocessors were 4-bit, and
r W. R. Stevens (1994). TCP/IP Illustrated never found their way into general-purpose
(New York, Addison-Wesley). computers. The first desktop computers, in
r L. Peterson and B. Davie (2003). the 1970s, were 8-bit. The first commercial
Computer Networks: A Systems Approach PCs (the IBM PC AT and XT) were 16-bit.
(San Francisco, CA, Morgan Kaufmann). Currently, 32-bit processors are in the vast
majority of computers (the Pentium and all
Associated terminology: Protocol, of its relatives are 32-bit), but 64-bit proces-
Internet. sors are gradually gaining ground.
The number of bits that a processor has
can be measured in a number of ways. The
Thirty-two bit most common is the width of the Arith-
metic logic unit (ALU), which determines the
Overview magnitude of the numeric values that can
This is closely related to the discussion of be processed in a single operation. Another
memory and disk size (q.v.). To a computer, common measure is the width of the Data
a Bit is the same as a Digit is to a human. If bus, which determines how much data can
you had a desk calculator that could work be retrieved from memory in a single oper-
only with three-digit numbers, it would be ation, and yet another is the width of the
very inefficient to use: it could calculate Address bus, which determines how much
5 × 7 perfectly well, but if you needed to memory can be directly accessed. These
know 1215 × 49 there would be trouble. The three measures all have a strong tendency
calculator would still be better than noth- to increase as technology progresses, but
ing, you could use it to calculate 12 × 49 they do not increase uniformly. Most desk-
and 15 × 49, and then combine the results, top computers today have a 32-bit integer
but it would be a lot of work. A calcu- ALU, an 80-bit floating-point ALU, a 64-bit
lator that could handle up to eight-digit data bus, and a 36-bit address bus. The
numbers would make everything so much exact meaning of the phrase ‘‘thirty-two
easier. bit” must be determined from its context.
The same is true with computers. An Software is specific to a particular kind of
eight-bit number is restricted to the range processor. Most programs may be designed

319
Thirty-two bit

to be Platform-independent, which means which is far more price-driven, has evolved


that they could potentially be run on any from computers that happily used 4 bits
computer, but the application itself needs (although not so happily for their pro-
to be recompiled (reconstructed) for each grammers), to computers that used 8 bits,
kind of processor, and system-critical soft- then 16 bits, followed by 32 bits and
ware (most especially the operating system) 64 bits. A larger bit size allows computers
is often designed to take advantage of spe- to quickly access larger amounts of mem-
cific features of specific processors. Soft- ory, produce more realistic-looking graph-
ware created for a 16-bit processor may still ics, and perform computations involving
work on a 32-bit processor, but it will not be larger or more accurate numbers with
able to take full advantage of the new pro- greater efficiency.
cessor’s capabilities. System software cre- The consequence for organizations
ated for a 16-bit processor often does not relates to the procurement decisions sur-
work at all on a 32-bit processor. rounding both hardware and the software.
Thus, there is 16-bit, 32-bit, and 64-bit As the hardware of computers changes,
software. The number of bits is not really software vendors produce new packages
an aspect of the functionality of the appli- or new versions of their existing software,
cation, but an indicator of the kind of pro- which might not be compatible with older
cessor it works (best) on. As hardware is hardware. Frequently, using the most
upgraded, it is often necessary to upgrade up-to-date software requires purchasing
software too. Conversely, sometimes a soft- the most up-to-date hardware, and this
ware upgrade may unexpectedly require a may be a significant expense. When legal
hardware upgrade. The best-known exam- requirements change, the high cost of an
ple is probably from the history of Win- upgrade may be unavoidable.
dows. Windows 3.11 was a very popular 16- This lifecycle forces users to upgrade
bit operating system. It had been designed their systems more frequently than they
to be compatible with earlier 16-bit proces- may have expected, and users who may be
sors, and could not properly support 32-bit happy with their existing systems have to
software or make effective use of the 32-bit endure the expense and upheaval of an
microprocessors (such as the Intel 80486) unwanted upgrade. The upgrade require-
on which it most commonly ran. It was ments are compounded by the fact that an
replaced primarily by Windows 95, which is upgrade to one aspect of the system may
a 32-bit operating system, and could make require an upgrade to the whole system to
much better use of the processor’s power, ensure that components continue to inter-
but a lot of software designed for 16-bit operate with each other. Clearly, there is
systems became inoperable when it was also a danger in being first mover to a new
installed. platform because systems bugs and incom-
patibility problems might not have been
Business value proposition corrected or even discovered, so any move
An aspect of technology that continues to a new platform needs to be carefully
to evolve is the number of bits a com- considered.
puter uses internally to perform its compu-
tations and other operations. Mainframe- Summary of positive issues
style computers, for which keeping the Larger bit sizes enable computers to per-
price low was not such a vital design deci- form faster computations, data transfer,
sion, traditionally had large bit sizes, such and processing, to access more memory,
as 36 or 60. Microprocessor technology, and to produce better graphical displays.

320
Transaction processing system (TPS)

Summary of potentially negative issues stand-alone have become ‘‘modules” in


The transition from one generation of com- these larger systems. They still perform
puter to another may require software ven- their transaction-based functions but are
dors to re-write or adapt their system to fully integrated with the other modules
meet the new hardware architecture. This of the systems. However, it should be
action frequently requires users to upgrade noted that extremely high-volume transac-
to the new hardware platform and the new tion processing of such items as credit card
version of the software at additional cost. bills, for which millions of items of data
need to be passed through the same pro-
Associated terminology: Bit, Bus. gram, is still performed as a dedicated pro-
cess on mainframes designed for TP at high
processing speeds.
Transaction processing system (TPS)
In database systems, a transaction is
Foundation concept: Database. defined to be a set of operations that must
Definition: A transaction processing system is an be completed as a whole or not at all. For
information system whose principal intent is to pro- example, when funds are transferred from
cesscompletetransactionsinvolving“goods,serviceor one account to another, the debiting of one
money” (www.tpc.org). For database systems, a trans- account and the crediting of the other have
action is a set of operations that must be performed as to be performed as separate operations on
one, in order to ensure consistency. the database, but must be treated as a sin-
gle transaction, so that, if there is a sys-
Overview tem failure, it will be certain that either
The term Transaction processing (TP) can be the transaction happened or it didn’t: there
traced back to the early Data-processing can be no half-way result. The design of effi-
(DP) era of information systems (approxi- cient, effective, and guaranteed Commit and
mately 1969--1979) when the primary activ- Rollback operations that ensure this behav-
ity that occurred in the Data center was ior in transactions is a major challenge in
the repetitive processing of data through the implementation of database systems.
programs that resided upon a mainframe One organization that attempts to bench-
computer operating in batch mode. The mark the performance abilities of compu-
organizational view was that the TP sys- ter systems on standard tasks is the Tran-
tem needed to work at maximum operating saction Processing Performance Council
capacity in order to achieve an economic (www.tpc.org), whose aim is ‘‘to define
payback on the large investments that the transaction processing and database bench-
mainframe environments required. marks and to disseminate objective, veri-
Since the 1980s, as systems became more fiable TPC performance data to the indus-
integrated and moved from batch pro- try.” This organization publishes the results
cessing to online processing, transaction- of its benchmarking studies for con-
processing systems came to be known as sumption by the computing industry and
Online transaction-processing (OLTP) systems. consumers.
The early OLTP systems differed from their
batch predecessors in that they allowed Business value proposition
online (interactive) processing and retrieval Transaction processing systems facilitate
of data as well as reporting capabilities. optimal resource utilization and the effi-
OLTP systems have since become embed- cient processing of large blocks of data (e.g.,
ded within larger systems such as ERP sys- a payroll module of an ERP can be run at
tems, and the functions that were once night when system usage is low).

321
Trojan horse

Summary of positive issues admired it, and retired to bed thinking


Transaction processing systems provide a the war to be over. The Greek soldiers
cost-effective mechanism for the process- climbed out of the horse, killed them all,
ing of large amounts of functionally sim- and opened the city gates to their waiting
ilar transactions. Vendors and outsourcers comrades.
are available to support TP. The story has been well known since
ancient times, and as a result the term
Summary of potentially negative issues Trojan horse has been used for anything
Transaction processing systems, depending disguised as a gift in order for it to be
upon the scale of data passing through the taken into a normally secure location with
system, may require significant dedicated evil intent. The horse was actually Greek of
resources. course, but the term Trojan horse is always
used. The same story also gave rise to the
References slightly insensitive phrase ‘‘Beware Greeks
r P. Bernstein and E. Newcomer (1997). bearing gifts.”
Principles of Transaction Processing (San The concept of a Trojan horse has proved
Francisco, CA, Morgan Kaufmann). irresistible to the modern cyber-criminal,
r Transaction Processing Performance and is a very common virus delivery
Council, Presidio of San Francisco, method. Trojan horses are extremely easy
Building 572B, Ruger St., San Francisco, to create, since they do not require any clev-
CA 94129-0920, USA. erly designed vector (delivering agent) or
trigger, and people happily load them onto
Associated terminology: Batch processing, their own computers and run them.
ERP. To create a Trojan horse, a program is
written to perform some evil, and usually
criminal, act. It could be anything from
Trojan horse deleting files, reformatting disks, or creat-
ing false data files to secretly reading con-
Foundation concept: Security. fidential data and sending it back to the
Definition: A piece of software, disguised as some- creator, or using the modem to dial expen-
thing desirable, but which is actually maliciously sive toll calls. The program is then openly
destructive. posted on a web site, emailed to unsuspect-
ing recipients, or packaged with genuine
Overview software on a CD, but with a false and
According to Homer (the Greek, not the inviting description. Commonly, the Trojan
Simpson), some time around 3000 years horse is described as a ‘‘cracked” version of
ago, the Greeks laid siege to the city of some expensive and popular software pack-
Troy (now known as Hissarlik) in west- age, an MP3 recording of a popular song,
ern Turkey, in order to retrieve Helen, or compromising photographs of a popular
the abducted/escaped wife of the King of person.
Sparta. The siege seemed to be heading Anti-virus software can detect Trojan
for failure, until the Greeks devised a new horses, if they are mere copies of already-
plan. They built a giant wooden horse, caught and analyzed versions, and they are
which the Trojans would believe to be a scanned before they are run. However, any
tribute from a humbled enemy. The horse programmer, at almost any skill level, can
was secretly filled with Greek soldiers and easily create a totally new Trojan horse. The
left outside the city gates. The Trojans fell key to avoiding Trojan horses is skepticism:
for the trick, brought the horse inside, do not execute software that didn’t come

322
Trojan horse

from a trusted source, and be very care- Summary of potentially negative issues
ful about whom you trust. It is very tempt- Vendors of anti-virus software need to con-
ing to solve your budget woes by download- tinually modify their software to detect
ing free versions of software, but consider new and variant Trojan horses, which is
this: if a stranger walked up to you in the a never-ending and nearly impossible task.
street and gave you a free pie, would you Opening a malicious Trojan horse can
eat it? lead to serious or even catastrophic conse-
quences.
Business value proposition
References
There is no legitimate business value to a r M. Egan and T. Mather (2004). The
Trojan horse.
Executive Guide to Information Security:
Threats, Challenges, and Solutions (New
Summary of positive issues
York, Addison-Wesley).
The software device known as a Trojan r P. Szor (2005). The Art of Computer Virus
horse is known and understood, so vendors
Research and Defense (New York,
of anti-virus software can work to ensure
Addison-Wesley).
that their software identifies and removes
the offending program. Associated terminology: Virus, Spyware.

323
UML (Unified Modeling Language)

a wide variety of modeling techniques


UML (Unified Modeling Language)
and diagramming methods, enabling a
Foundation concept: Software development lifecycle. wide variety of systems and system types
Definition: UML is a standards based modeling lan- to be developed. The UML approach has
guage for specifying, visualizing, constructing, and been adopted as the standard development
documenting software systems. methodology by many large vendors and
as a consequence there is a wide avail-
Overview ability of consulting, software, and toolsets
UML is a mechanism for specifying soft- available. UML aids communication within
ware systems. It provides developers with project teams and the management of large
a relatively intuitive methodology for the projects. The wide acceptance and use of
design of systems and programs. UML emer- UML has increased access to human capital.
ged as the leading object-oriented model-
Summary of positive issues
ing language in the mid 1990s, and com-
UML is a widely used, standardized
bines many popular and common concepts
methodology for systems development. The
found in earlier methods into its model-
methodology is supported by a wide array
ing notations. UML provides a development
of software tools and user groups. UML is
structure applicable to any systems prob-
amenable to the development of a wide
lem, and supports the system modeler from
array of system types and environments.
specification to implementation.
UML incorporates a large set of graphical Summary of potentially negative issues
models, including Class diagrams, Use-case UML can be conceptually difficult for pro-
diagrams, Component diagrams, Deployment grammers and developers not familiar with
diagrams, State-chart diagrams, Activity dia- the object-oriented style of systems develop-
grams, Sequence diagrams, and Collaboration ment. UML is not a formal mathematical
diagrams. These are used to represent dif- model of development.
ferent viewpoints of an overall design: the
Static view, the Case view, the State-machine References
r G. Booch, J. Rumbaugh, and I. Jacobson
view, the Interaction view, and the Model-
management view. Each view provides the (2005). Unified Modeling Language User
modeler with a different perspective of the Guide (New York, Addison-Wesley).
r M. Fowler and K. Scott (1999). UML
system. For example the static view pro-
vides details of the basic concepts from Distilled: A Brief Guide to the Standard
which the system is composed, whereas the Object Modeling Language (New York,
activity view provides a visualization of the Addison-Wesley).
r J. Rumbaugh, I. Jacobson, and G. Booch
process flows through a model such as an
activity diagram. (1999). The Unified Modeling Language
UML is supported by a variety of soft- Reference Manual (New York, Addison-
ware tools both from open-source providers Wesley).
and from commercial organizations. These Associated terminology: Object-oriented,
tools range from assisting in the drawing Formal methods.
and creation of UML diagrams to code-
generation systems.
Unix
Business value proposition
UML provides developers with a standard- Foundation concept: Operating system.
ized approach to the development of sys- Definition: A popular operating system for computers
tem designs. The approach incorporates large and small.

324
Unix

Overview (from Berkeley Software Distribution), and


Linux (based on Andrew Tanenbaum’s light-
It used to be that every computer manu- weight system Minix, but independently
facturer also produced their own operating redeveloped by Linus Torvalds). Sun Micro-
system. They were totally tuned to one par- systems’ Solaris and SunOS are also versions
ticular kind of computer, and gave carefully of Unix, as now is MacOS. It is unclear
tailored control over all aspects of its opera- whether these new systems can really be
tions. As computers became cheaper, their called Unix in a legal sense, since they are
purchase price could no longer cover the distinct products, but in every other sense
high costs of such complex software devel- they certainly are.
opment, and now the practice is very rare The command-line interfaces to Unix (the
indeed. It seems as though every computer shells) are characterized by exceptionally
there is runs either Unix or Windows, and arcane syntax and a complete lack of any
things really are almost exactly what they overall design. For example, to see the con-
seem. There are a few exceptions, but only tents of a text file, the command is ‘‘cat”; to
a few. Even the Macintosh operating system see the file with all non-text characters ren-
is now a graphical shell running on top of dered visible, it is ‘‘cat -e.” The basic com-
a standard Unix. mand for file backup is ‘‘tar”; to obtain a
Unix started its life in the early 1970s as listing of the files that were backed up in a
an AT&T Bell Labs project led by Kenneth particular archive one has to remember the
Thompson and Dennis Ritchie. It was origi- incantation ‘‘tar xvf.” The terse syntax can
nally a very simple and basic operating sys- be very dangerous: the command to delete
tem, providing only the absolutely essen- all files whose names begin with the four
tial features, relying on third-party appli- letters ‘‘temp,” a common and reasonable
cations to perform most of its tasks. It task, is ‘‘rm temp∗,” but, if it is accidentally
has since grown considerably, but still fol- typed as ‘‘rm temp∗” (just an extra space),
lows the same principle of design: a Ker- every file is deleted, regardless of its name,
nel covering the essentials, a Command-line without warning. Even experienced users
interpreter, or Shell, with very little intrin- spend a lot of time referring to reference
sic functionality, and a large collection of material in an annoyed manner. For non-
independent applications and components technical users, one of the many Graphical
to do the work. user interfaces (GUIs), or Windowing environ-
For many years, Unix was a strictly pro- ments, is almost essential.
prietary piece of software, with use per- The fact that there is no one source
mitted only under license from AT&T. The responsible for the overall design of Unix-
source code was kept secret, making many like systems, and essential utilities are cre-
aspects of system design very difficult, ated by arbitrary third parties, means that
and rendering the system unsuitable as a there is no uniformity to the Unix expe-
basis for academic study. Over the years, rience. Different applications will behave
several complete ‘‘clean-room” redevelop- in different ways, with different paradigms
ments were undertaken, so that there of interaction, and different command con-
now exist Unix-like operating systems that ventions. For a long time, reliability was
were created by people who had never a major issue: if software had been con-
seen the AT&T code, and such systems tributed for free by random programmers,
are not derived from any earlier systems. what level of reliability could be expected?
The result is free and open implementa- Nowadays, however, some of the varieties of
tions not requiring any licensing. The most Unix are amongst the most reliable oper-
popular free versions of Unix are FreeBSD ating systems available; the commercial

325
URL (Uniform resource locator)

pressure of millions of users surrounded by organizations and even more individuals


thousands of hackers has resulted in great have adopted free versions of Unix and
improvements. reduced their systems costs.
In order to bring together all the dif-
ferent versions of Unix, and at least pro- Summary of positive issues
vide some common core of functionality, so Unix is a well-known operating system that
that cross-version software could be devel- can be used in line-command mode or, if
oped with reasonable ease, the IEEE devel- desired, through a GUI. Free versions of
oped the POSIX standard (Portable Oper- Unix are widely available.
ating System Interface for uniX, now ISO
9945). The ‘‘I” in Posix represents not the Summary of potentially negative issues
user interface, but the programmer’s inter- Unix is extremely hard for untrained users
face. A ‘‘Posix-compliant” system will pro- to understand and operate. It has very pow-
vide a standard set of library functions with erful instructions that can perform signif-
uniform functionality for software develop- icant systems operations (e.g., remove all
ers to make use of. data from a disk) and is unforgiving of mis-
takes (there is rarely an undo option). The
Business value proposition quality of third-party components can not
The choice between Unix and another oper- always be guaranteed.
ating system depends in part upon the skill There is a highly controversial dispute
levels of the user base, the applications to between the owners of the original AT&T
be run on the system, and the technology Unix code and the developers of other ver-
to be used in conjunction with the operat- sions of Unix as to the ownership of source
ing system. code and claims of infringements of intel-
Unfortunately, Unix remains a dangerous lectual property rights, although varieties
operating system in the wrong hands. Its of Unix produced by ‘‘clean-room” develop-
users are expected to know what they are ment and containing no AT&T code are
doing, and it is very unforgiving when they common.
don’t. For a computer owner with some rea-
References
sonable technical knowledge, or an orga- r G. Glass (1993). Unix for Programmers and
nization with a properly trained technical
Users (Englewood Cliffs, NJ,
staff, Unix is often the operating system of
Prentice-Hall).
choice. For those who merely need to run r http://www.freebsd.org.
visual applications (word processors, spread r http://www.linux.org.
sheets, web and email browsers), a well-
designed and controlled windowing envi- Associated terminology: Open-source
ronment hiding the arcane Unix shells may software.
be much safer.
Many technical systems administrators
do in fact like Unix for the very reason that URL (Uniform resource locator)
typed command interfaces are available
for manipulating specific aspects of the sys- Foundation concept: World Wide Web.
tem, rather than having to wade through Definition: A uniform and well-defined form for
layers of pull-down menus and screens. The addressing internet resources.
open-source nature of many versions of
Unix also appeals to certain user groups, Overview
allowing them to construct customized Anything that is available over the inter-
variants of the operating system. Many net, whether it be a web page, a graphical

326
URL (Uniform resource locator)

image, an e-commerce form, a pirated A URN, or Uniform resource name, pro-


movie, or a virus, is generically termed vides a more abstract way of identifying
a Resource. In order to access any such a resource. URNs give names to resources,
resource, information on how to access and those names are supposedly guaran-
it (which server, which directory on that teed to always uniquely identify the same
server, which file in that directory, and resource regardless of how or where it is
which protocol to use) must be provided. stored. For example, a URN might be some-
In order to package that necessary infor- thing like ‘‘urn:isbn:0-201-63346--9”; this
mation in one consistent form that may be indicates the book that has ISBN 0-201-
understood by all network applications, the 63346--9 (ISBNs, International standard book
Uniform resource locator (URL) was invented numbers, are an already-established unique
(by Tim Berners-Lee, then a contractor for identifier for all books). A URN is in a sense
CERN, in 1989). the exact opposite of a URL -- it uniquely
URLs are the familiar ‘‘web-site addres- identifies a resource, but gives no clue as
ses” that users enter into web browsers, and to how to find it. Sadly, URNs are at present
that are embedded in hyperlinks in hyper- not a very practical proposition, since there
text documents. A URL consists of four is no established way to allocate or resolve
main parts, and has this general form: them.
protocol://hostname/resourcelocation?parameters The term URI, Uniform resource indicator,
All but the host-name are optional. The is a general term covering all URLs and all
protocol part specifies the communica- URNs. If anything requests a URI, in practi-
tions language to be used with the server cal terms it is really asking for a URL.
to access the resource; it is most com- Originally, the ‘‘U” in URL, URI, and
monly ‘‘http,” which is the language of web URN was intended to stand for ‘‘universal”
servers. The host-name indicates the inter- rather than ‘‘uniform”; the change has no
net address of the server to be used; it is particular significance.
typically of the form ‘‘www.company.com,”
but may be a numeric IP address. The Business value proposition
resource location specifies exactly which of URLs are a mechanism for accessing a
the many resources on that server is to be resource on the internet and provide an
accessed, and generally has the form of a organization’s customers with a means of
file name; if it is left out, then that server’s locating them. Organizations need to be
default ‘‘home page” is usually requested. careful about changing the URLs on their
The parameters are occasionally used to websites, since customers who have ‘‘book-
provide extra information specific to an marked” part of a site for future reference
individual request. have really just automatically recorded the
URLs give instructions on how to find a URL, and may be disappointed and frus-
particular resource by saying exactly where trated to receive a ‘‘404 Error” instead of the
to look for it. This is a very practical page with ‘‘Discounted Sports Equipment.”
scheme, but is sometimes problematical. Businesses should also take care to en-
Often web sites are rearranged, and resour- sure that all their URLs are carefully man-
ces change locations. Using an old URL aged and that old ‘‘legacy” URLs from the
could easily result in the wrong resource early days of their web presence are main-
being accessed. Even more frequently, old tained, redirected, or properly removed
URLs point to locations that no longer con- from servers no longer considered opera-
tain anything. URLs do not include expir- tional for public viewing. These legacy sites
ation dates or any other means for verifying can be found by using a search engine and
that the correct resource has been found. scanning for all public access materials.

327
URL (Uniform resource locator)

Summary of positive issues strong perception of unreliability and lack


URLs provide a universal access mechanism of devotion to customer service.
for locating resources on the World Wide
Web, and are an inevitable part of any web References
r P. Gralla (2004). How the Internet Works
presence.
(Indianapolis, IN, Que).
Summary of potentially negative issues r D. Gourley and B. Totty (2002). HTTP, the
URNs (as opposed to URLs) are not cur- Definitive Guide (Sebastopol, CA, O’Reilly
rently a practical proposition. The main- Press).
tenance of web sites can be more com- r http://www.w3c.org.
plex than expected, since changes that are
not carefully managed result in customers’ Associated terminology: Internet, Host,
bookmarks becoming invalid; this creates a Hypertext.

328
Video

tions and entities. They provide an organi-


Value added network (VAN)
zation with an alternative to running and
Foundation concept: Network. maintaining its own networks.
Definition: A communication network provided by
a third-party vendor that also provides guaran- Summary of potentially negative issues
tees of service quality, security, and reliable data VANs are connected over a common carrier
transmission. network. While these networks can pro-
vide high bandwidth and services, internet-
Overview based network services provide a popu-
Value added networks (VANs) are provided by lar alternative. Internet-based networks are
third parties such as the common carri- known as virtual private networks (VPNs) and
ers (telecommunications providers) which can be created by corporations or through
lease secure telecommunication connec- third-party VPN providers.
tions to subscribers. VANs provide a vari-
ety of services together with the connec- Reference
r L. Peterson and B. Davie (2003).
tion, including support of a variety of
EDI standards (such as ANSI X.12, EDIFACT, Computer Networks: A Systems Approach
and XML), data security standards (for data (San Francisco, CA, Morgan Kaufmann).
encryption, a Secure sockets layer, etc.), mul- Associated terminology: Virtual private
tiple levels of password authorization, and network, T-Carrier, Encryption, Security.
communication protocols such as X.25 and
TCP/IP.

Business value proposition Video


VANs provide secure networks capable Foundation concept: Digital.
of high-bandwidth data transmission that
range from T1 at 1.544 Mbps to OC48 at Overview
2.4 Gbps, together with service-quality guar- Video captured by a camera compresses
antees and data security services. The deci- a three-dimensional scene into a one-
sion to purchase may be based upon the dimensional stream of data. Simply looking
technical need together with the financial at a scene with visible light reduces it to
aspects of the lease: subscribers pay a base a two-dimensional projection; the camera
fee plus per-service costs. The analysis of scans this projection in a series of horizon-
total cost of ownership has to be based tal scan lines from top to bottom, and each
upon the use of and need for the services. line is reduced to a series of dots running
VANs offer a variety of services and charges from left to right, so the whole scene is sim-
are based upon their usage: the amount ply a series of dots with varying colors and
of data transmitted, fees for connections intensities. A moving picture is created by
to other VANs, mailbox charges, help-desk repeatedly scanning the same scene, mak-
service charges, data-archiving fees, service- ing a longer series of colored dots. Con-
plan fees, and user and technical training verting the color and intensity of each dot
fees. into a numeric form gives a complete digi-
tal representation of one view of a moving
Summary of positive issues scene.
VANs provide a secure, high-quality, profes- Natural colors correspond to different
sionally managed and maintained mecha- wavelengths of visible light, and there
nism for connecting with other organiza- is a huge range of possible wavelengths.

329
Video

Fortunately, Nature took a short-cut when most real scenes, so compression can come
designing the human eye, and the color- to the rescue. Compression ratios of 30 : 1
sensing cells in the retina react to stimuli may be achieved with generally acceptable
only in three broad wavelength ranges. This results, although rapidly changing or sub-
is why there are three primary colors: it has tly detailed scenes do suffer degradation.
nothing to do with the physics of light; it The most used format for digital video is
just arises from the structure of the human now MPEG (Motion Picture Experts Group),
eye. To accurately represent the color and which is currently available in four ver-
intensity of any point of light from the sions. (The popular audio format MP3 is an
human point of view, only three numbers abbreviation for MPEG-3.) MPEG is the stan-
are required, to represent the intensities dard used for domestic DVD recordings,
in each of the three bands: red, green, most digital satellite television signals, and
and blue. Thus the storage space required most streaming video systems. MPEG is
for any video signal may be computed as closely related to the JPEG system for
3 × the number of frames scanned per still images; it views the image as a series
second × the number of horizontal rows of 8 × 8-pixel blocks, which are individu-
per scan × the number of resolved dots per ally compressed. When too high a compres-
line × the length of the video sequence in sion ratio is demanded, these little squares
seconds × the amount of space required become easily visible. MPEG also groups
to store a single number. To make a pic- together small sequences of scanned
ture smooth enough to please the aver- images, taking the first whole (but com-
age viewer, each number requires only one pressed) as a Key frame, then recording only
byte, about 500 lines and 500 dots per how subsequent frames in the group dif-
line, and 30 scans per second are required fer from the first. When the scene changes
as a minimum. This results in a total of slowly, this produces great improvements
22 megabytes per second, or 120 gigabytes in compression without noticeable loss of
for a short movie; a very large quantity quality; when the scene changes quickly, it
of data indeed, and an impossible band- produces highly visible, strangely colored
width requirement for almost any network rectangular ‘‘MPEG artifacts.”
connection. This belies the deliberate myth of ‘‘dig-
The figures of 500 rows and 500 dots ital quality.” A digital signal is always an
are somewhat arbitrary; standard television approximation to an original analog sig-
signals are approximately 500 × 300 and nal, and a compressed one more so. Digital
‘‘DVD quality” is, in the United States at video does have two advantages: it is in a
least, 720 × 480. The resolution perceivable conveniently computer-manipulable form,
by the eye is much higher than that, but and perfect copies can be made without
cameras could easily be designed with a the ‘‘generation loss” familiar to those who
higher resolution if needed. The figure of have tried to copy VHS tapes. Digital video,
30 scans per second is based soundly on like digital audio, has no innate lead over
biology: the eye can not react to changes a corresponding analog signal in terms of
quite that fast, so a higher scan rate quality; it is simply cheaper and more con-
would produce no improvement for human venient for providers.
viewers; a scan rate as low as 20 per second Other video formats abound. The AVI
would produce a very irritating flickering (Audio/Video Interleave) format introduced
effect. by Microsoft is popular in video production
Twenty-two megabytes per second is an because it is easier to process and tends
almost impossible standard to meet. For- to produce better quality than MPEG, but
tunately, there is a lot of redundancy in as a consequence produces much larger

330
Virtual machine

files, so it is not popular for distribution. internet and short video clips sent to cel-
Apple’s Quicktime ‘‘MOV” format is also pop- lular telephones are certainly just the first
ular, and frequently used for short down- steps.
loadable movie clips.
Summary of positive issues
Business value proposition There are several well-known and estab-
There has been explosive growth in the lished standards associated with video.
use of digital video since low-cost video Video technologies and internet technolo-
technologies became available during the gies are continually improving, allowing
1990s. The newly affordable technologies higher-resolution and larger video files to
include moderate-quality cameras costing be transferred. Digital video delivered over
only a few tens of dollars but requiring a network is a cost-effective mechanism
direct connection to a computer, video for the delivery of short or low-bandwidth
cameras with their own built-in record- transmissions, which is particularly valu-
ing mechanisms, and, of course, digital able for training and educational purposes.
versions of traditional VCRs using com-
puter disks instead of tapes. A significant Summary of potentially negative issues
increase in affordable network bandwidth Digital video requires a large amount of
and the availability of affordable CD- and storage capacity and bandwidth. Video
DVD-burning hardware were also major compression when taken too far signifi-
influences. cantly reduces the quality of images.
The proliferation of video devices was
paralleled by the development of technolo- Reference
gies and software to allow the duplication, r D. Austerberry (2002). Technology of Video
storage, and manipulation of the video and Audio Streaming (Woburn, MA, Focal
data. These include software for the pro- Press).
duction of online videos capable of being
downloaded through the internet or multi- Associated terminology: Compression,
cast through Local area networks, which has Audio, Digital, Bandwidth.
led to a large increase in the use of, if not
in the quality of, on-demand instructional
and training video. Virtual machine
The potential uses of video technologies
are vast, and the delivery of video over Definition: A software application that perfectly
the internet has become an area of com- imitates a real computer.
mercial and academic interest. The long-
awaited delivery of video-on-demand televi- Overview
sion is now possible over high-bandwidth It is possible to design a program that pro-
networks. Wide availability of high-speed vides a perfect imitation of the hardware
broadband internet connections is needed of a real computer. By duplicating, as a
both to create a positive economic demand software simulation, all the operations of a
model and to facilitate the delivery of video given piece of hardware, a purely software
signals with sufficient image resolution. application can arrange to produce exactly
Computer systems may be connected to the same effect. Naturally, a software sim-
large projection systems to provide a home ulation will be slower than the real thing,
or corporate in-house theater effect. There and generally less efficient in other ways,
is great potential for future expansion in but in all important respects a piece of digi-
this direction; the delivery of radio over the tal hardware that is fully understood can be

331
Virtual machine

perfectly rendered in software. The result is only harm the simulation, and the simu-
called a Virtual machine. lation can be restarted in a fraction of a
Virtual machines provide four valu- second.
able services, relating to experimentation, Software construction and portability:
hardware--software compatibility, system this is currently the most popular use of
security, and software construction and virtual machines. Sometimes, the basic way
portability. that modern computers work makes cer-
Experimentation: when the purchase of tain software solutions very difficult to
a new kind of computer system is being implement, and designers conclude that it
considered, buyers can get a strong idea of would be much easier if they could design
what the new system is like to use without a whole new kind of computer just for this
having to buy or borrow a demonstration one program. Virtual machines make this
model if they can run a software simula- into a viable solution. The new computer,
tion of it on their existing computer sys- idealized for the solution of this one prob-
tem. When a new computer system is first lem, and perhaps totally infeasible for real
designed, enormous savings can be realized construction, is designed and implemented
by first constructing a virtual machine to as a software simulation. The difficult prob-
test the design and iron out the major bugs lem is then solved with a program writ-
before constructing any new hardware. ten for that new computer, and the com-
Hardware–software compatibility: most puter simulation packaged together with
software runs exclusively on one par- that program is the product. This is an
ticular hardware platform: PC software especially valuable technique when new
does not run on Macintosh hardware. programming languages are designed and
Although some software manufacturers do first implemented. Once a product has been
produce versions for different platforms, developed by this method, it is much eas-
this is by no means universal. A virtual ier to produce versions for other hardware
machine allows one computer to behave platforms: only the virtual machine needs
like another, and therefore run software to be Ported to another kind of computer;
designed for that other system. For exam- the program that it supports, the major
ple, a Macintosh computer can run a vir- part of the product, remains unchanged.
tual machine that imitates a PC, and This is part of the philosophy of Java. All
that virtual PC can then run any PC Java programs run on a virtual machine
software. known as JVM, the Java Virtual Machine.
System security: virtual machines are All Java programs are automatically cross-
not limited to imitating another kind of platform because an equivalent JVM has
computer. It is quite possible, and common, already been written for all major hardware
for a PC with Windows to run a virtual platforms.
machine that imitates a PC with Windows.
The reason for this is that untrusted soft- Business value proposition
ware (perhaps it is suspected of being a Tro- Virtual machines are software programs
jan horse or bearing a virus, or perhaps it that simulate the hardware of a real or the-
just doesn’t work very well and keeps crash- oretical computer. The best known use of
ing the computer) can be run on the vir- virtual machines is in the Java program-
tual machine; it will behave exactly as it ming area, where Java programs run on
would on the real computer, but will be a Java Virtual Machine (JVM) and as such
incapable of doing any harm, because its are highly portable, since the JVM can be
environment is simulated. It can’t erase the re-written for any hardware platform. This
real disks or crash the real computer; it can prevents the problem associated with many

332
Virtual memory

programming languages, in that a specific Summary of potentially negative issues


version of the language will only run on a Training as a computer scientist is required
single machine and locating or writing a if one is to create a useful virtual machine.
new compiler for the language so that it The creation of a virtual machine requires
can be used with a new computer is dif- the expenditure of focused specialized
ficult, expensive, and time-consuming. The resources and may take some time to com-
virtual machine concept is one of the rea- plete for a challenging simulation prob-
sons behind the popularity of Java and its lem. Developing a virtual machine may be
development environment. a lengthy exercise and should not gener-
A second popular use of the virtual ally be considered a quick fix for a critical
machine concept is to allow one machine problem. Virtual machines are less efficient
to simulate another machine, such as a PC than the real hardware system that they
being simulated on a Macintosh, and con- simulate.
sequently PC applications can run on the
Macintosh through the simulation. Reference
r J. Smith and R. Fair (2005). Virtual
Another use that is gaining popularity
amongst network managers and chief secu- Machines: Versatile Platforms for Systems
rity officers is the ability to use virtual and Processes, Morgan Kaufmann Series
machines as a further layer of defense in Computer Architecture and Design
against intruders and malicious software (San Francisco, CA, Morgan Kaufmann).
attacks. A virtual machine can be made to Associated terminology: Java.
simulate the machine upon which it is in
fact running and therefore, should the soft-
ware turn out to be malicious, it would Virtual memory
only be attacking the simulation and not
be able to do any real harm. Foundation concepts: Memory, Disk.
Virtual machines may be used in orga- Definition: A system that makes a computer appear to
nizations as mechanisms to determine have more memory available than it really has.
how a software package would work upon
another hardware platform. For example, Overview
an ERP system module may represent A Virtual memory system usually has two
a large investment and a company may components: Memory mapping, which allows
wish to assess whether the system would many applications to run concurrently,
perform as desired upon a new platform with all of them able to make use of all of
prior to its purchase. This would allow per- the computer’s memory without conflict;
formance issues to be considered and aid and Paging, which allows applications to
the organization in performing capacity make use of more memory than is actually
planning. installed in the computer. Virtual memory
is an essential component of any general-
purpose operating system.
Summary of positive issues Memory mapping is a service provided by
The theoretical aspects of virtual machine special-purpose hardware, which requires
development are well known amongst com- special support from the operating sys-
puter scientists. Virtual machines exist to tem software. A computer’s memory is
support a variety of languages and of plat- viewed as a collection of literally millions
forms that aid the portability of applica- of individual Locations, each capable of stor-
tions. Virtual machines can be used as an ing a single simple item of data such
extra security layer. as a small number or an alphanumeric

333
Virtual memory

character. Each of these locations has a rently stored in immediate-access memory


fixed numeric address, and all accesses to to be Paged out, and stored on disk instead,
memory must specify the numeric address thus freeing up that area of memory to be
of the location to be accessed. For example, used for something else. Of course, when
a database application may store the cur- an application needs to access data that has
rent count of the number of records in the been paged out, it must be slightly delayed
database in location number 301 440, and while that data is Paged in, and brought
might store the list of employees’ social- back into memory from disk. Since disk
security numbers in the region of memory drives work approximately 1 000 000 times
between locations 1 241 780 and 5 241 779; more slowly than normal memory, the sys-
a word-processing application might store tem must take great care in deciding which
the name of the file it is working on areas of memory can reasonably be paged
between locations 301 300 and 301 526, and out.
the contents of that document starting at Fortunately, most applications on a per-
location 2 000 000. If those two applications sonal computer are almost completely User-
are ever to be run concurrently, their mem- bound. This means that they spend the vast
ory usages would conflict. majority of their time just waiting for a
Memory mapping hardware performs on- human user to press a key or click the
the-fly location translations for each run- mouse. Modern computers are so fast that a
ning application. Every time a memory whole application can sometimes be paged
location is accessed, the memory map- out, or suspended to disk, while waiting for
ping hardware looks up the location to user input, and restored so quickly when
be accessed in a mapping table, finds the that input is provided that the user is
alternate location that that access should totally unaware of any delay. With a well-
be translated to, and causes that address designed paging algorithm, a modern oper-
to be used instead. The operating system ating system can easily over-subscribe its
sets up a mapping table for each appli- use of memory by a factor of ten or more,
cation as it is started, and ensures that allowing a computer with 128 MB of mem-
all applications’ memory accesses will be ory to behave like a computer with 1280 MB
translated to unique locations, avoiding of memory.
all conflicts. A typical modern processor Paging is so called because memory is
may make 100 000 000 or more memory divided into a number of small regions,
accesses per second, so the careful design usually of 512 or 4096 bytes each and called
of memory mapping hardware to work Pages. For efficiency, it is not individual
at these speeds is critical to system per- memory locations that are paged out, but
formance. Without memory mapping, it whole pages at once. Paging is not essen-
would be not quite impossible but very dif- tial for any operating system, and, when
ficult and inefficient to have more than one computationally intensive applications, or
application running concurrently through Real-time applications (those that must be
Time-sharing. able to respond to emergent conditions
Paging is a service provided by operat- immediately), are running, it may be desir-
ing system software, which requires special able to turn paging off, either for indi-
support from the memory mapping hard- vidual applications or for the whole sys-
ware. Paging allows a computer system to tem. Not all operating systems provide fine
act as though it has much more memory control of paging, and that is an impor-
available than is actually installed. It does tant aspect to consider when choosing a
this by selecting certain things that are cur- system.

334
Virtual organization

Business value proposition Virtual organization


The technical aspects of virtual memory are
fixed by the hardware and operating system Foundation concepts: e-Commerce, e-Business.
manufacturers and thus there is little that Definition: A virtual organization is an organizational
the end user can do in this regard. Usu- structure under which value-chain activities may be
ally only the most minor of adjustments located anywhere, either owned by the organization
are possible. If a system does not offer vir- or sourced from a third-party provider, and connected
tual memory capability, that may be a seri- together through information technology.
ous impediment to its candidacy. For some
applications, such as small embedded sys- Overview
tems, virtual memory is not appropriate, The concept of a virtual organization arises
but that assessment should be made only from the notion that all aspects of a tradi-
by technical experts. tional company’s value chain can be either
outsourced or provided by a third party
and managed through information tech-
Summary of positive issues nology. The goal of creating a virtual orga-
Virtual memory provides systems with nization is to enable flexible production to
the ability to service multiple applications occur, resulting in what Davidow and Mal-
and users concurrently. Memory mapping one have termed the Virtual product.
allows more flexible use of memory and as The term Virtual in both of these con-
a consequence optimizes memory resour- texts is derived from the use of the word
ces. Paging allows computers to extend the in computer engineering (virtual memory,
capability of their memory by bringing virtual reality, virtual terminal, etc.), where
in and out of the page file the resources it denotes some form of nonexistence or
required by applications as the operating abstraction from reality. A virtual organiza-
system switches between them. tion is one in which some or all of the core
processes have been outsourced to a third-
party service provider. A simple example
Summary of potentially negative issues would be a company that receives orders for
Memory management is a highly technical some product from the public through the
aspect of computer engineering and should internet, buying the product directly from
not be interfered with by non-specialists. their suppliers only when full payment
Paging is also a complex aspect of the has been received, using a delivery service
design of a memory management system, such as FedEx or UPS to send the prod-
so any adjustments to the paging mecha- uct on to the customer. Even the process-
nism such as turning it off need careful ing of payments may be outsourced, allow-
consideration and technical advice is rec- ing the company to provide a vast range
ommended. of products by coordinating and manag-
References ing processes while performing hardly any
r D. Groth (2003). A+ Complete (Alameda, processes themselves.
CA, Sybex Press). Similarly, a company selling virtual prod-
r M. Gorman (2004). Understanding the ucts may advertise a wide range of highly
Linux Virtual Memory Manager customizable products, such as computer
(Englewood Cliffs, NJ, Prentice-Hall). systems with almost infinitely variable
options selected by the customer. Such a
Associated terminology: Operating company may have absolutely no stock in
system. hand, simply ordering the required parts

335
Virtual private network (VPN)

once an order has been received and full ing or provision of service to align with
payment made, then assembling them and demand or market fluctuations.
shipping the product out. This strategy
may provide great savings in inventory and Summary of potentially negative issues
warehousing costs. The concept of virtual organization incurs
In practice, pure virtual organizations a high overhead in terms of command and
exist only in the service sector (e.g., consult- control, information technology, and rela-
ing), while in the manufacturing sector a tionship management. Partners may find it
close approximation can be found in orga- difficult to provide services to the virtual
nizations that utilize techniques associated organization with the flexibility desired.
with mass customization, as advocated by Virtual organizations are dependent upon
researchers such as Joseph Pine. The term their partners to perform core and non-core
Agile has also been associated with discus- functions successfully.
sions surrounding the concept of virtual
References
organization. Steven Goldman and other r W. Davidow and M. Malone (1992). The
pioneers in this area have promoted the
Virtual Corporation (New York, Harper
agile organizational model, in which prod-
Business).
ucts and production resources are designed r J. Pine (1993). Mass Customization: The
to allow continuous product evolution as
New Frontier in Business Competition
market demands change.
(Boston, MA, Harvard Business School
Press).
Business value proposition r S. Goldman, R. Nagel, and K. Preiss
The concept of a virtual organization is in
(1995). Agile Competitors and Virtual
operation to some extent in the majority
Organizations: Strategies for Enriching the
of organizations that use a third-party ser-
Customer (New York, Van Nostrand
vice provider for any activity and manage
Reinhold).
that relationship through technology. The
pure virtual organization concept in which Associated terminology: Business process
every function is outsourced has not been re-engineering.
embraced by many organizations, but as
the ability of an organization to monitor
its outsourced functions through sophisti- Virtual private network (VPN)
cated systems increases the adoption of this
model will increase. The concept of virtual Foundation concept: Network.
organization allows organizations to con- Definition: A network that has no independent phys-
centrate on deriving the best value from ical existence, but exists only as a set of connections
each service provider, reducing the oper- over a larger real network, usually the internet.
ating expenses through technology, and
using the managerial resources to execute Overview
the business plan effectively. If an organization has offices in London,
Sydney, Grytviken, and New York, the con-
Summary of positive issues struction of a private network connect-
The concept of a virtual organization ing the four would be prohibitively expen-
allows greater flexibility and partnership sive. A worldwide network, the internet,
with ‘‘world-class” service providers. The already exists and provides public con-
application of leading technologies allows nections between any two points. A Vir-
greater agility in the company, rapidly tual private network (VPN) uses the existing
adjusting the production of manufactur- infrastructure of the internet, or any other

336
Virtual private network (VPN)

physical network, as the foundation of a ever funding they see fit to maintenance


new free network. A VPN could almost be and redundancy. When the network is vir-
viewed as a parasite on a real network; it tual, and exists only over public internet
exists only as an abstraction, an alternate connections, no such control is possible.
view of parts of the real network. The main connections of the internet,
In a Client--server-based VPN, a server is set called its Backbone, are extremely reliable,
up at some designated location to represent but the end connections provided by local
the private network. All other Nodes (sites telecommunications companies and ISPs
on the VPN) run special-purpose client soft- can be problematical.
ware, which makes a standard internet con-
nection to the server, and maintains that Business value proposition
connection for as long as the VPN is active. VPNs provide an alternative mechanism for
All messages that are supposed to be trans- organizations that wish to take a higher
mitted on the private network are actually level of control of their network, extend-
transmitted to the server, and the server ing and managing their network beyond
retransmits them to the appropriate des- their corporate boundaries. VPN technolo-
tination site. This provides the same func- gies have evolved considerably since the
tionality as a real private network, but with 1960s and 1970s when the options were
slightly less efficiency (all messages must to use a dial-up modem or lease a line
make two trips instead of one), and virtu- from the telephone company. The ‘‘leased-
ally no cost. line problem,” however, remains the same
In a Peer-to-peer-based VPN, all sites on the today as it was in 1960, since the costs
private network make and maintain direct associated with this approach are still the
internet connections to some or all of the biggest barrier to their use, and the com-
other sites. Messages that are intended to binatoric explosion of ‘‘peer-to-peer” con-
be sent on the private network are instead nections makes the scalability of this solu-
sent to the same destination over the exist- tion unsustainable for all but the largest of
ing network. When a VPN has more than organizations.
a small number of nodes, it is not practi- The business case for a VPN is based upon
cal for each to maintain a connection to all security, speed requirements, and the qual-
of the others. In such cases, a connection- ity of service provided by the network. The
less protocol such as UDP is more likely to use of an internet-based solution with off-
be used than TCP; the end results are the the-shelf hardware and software solutions
same with suitable software support. can create a highly effective solution appli-
There are two advantages that a true pri- cable and acceptable to the majority of
vate network would have. One is security: organizations wishing to use a VPN. These
if an organization is completely in control technologies include firewalls (hardware- or
of its network hardware, it can physically software-based), the Secure sockets layer (SSL),
shut out spies and unwelcome communi- and other encryption protocols. The use of
cations from external sources. The same widely available technologies allows a VPN
level of security can easily be provided to scale well, while maintaining a techno-
with the proper use of strong encryption logical basis that is generally compatible
and firewalls to exclude unauthenticated with the remainder of the corporate net-
transmissions. work architectures.
The second advantage is control over reli-
ability. The organization owning the net- Summary of positive issues
work hardware can decide exactly how Leased-line VPNs provide secure, reliable
important reliability is, and allocate what- networks that allow the quality of service

337
Virtual reality

to be controlled by the membership of sound. During the subsequent 40 years,


the network. Internet-based VPNs provide researchers such as Ivan Sutherland devel-
lower-cost solutions for companies that can oped the fundamental theory of computer
work within the parameters of the internet graphics and 3D imaging which under-
(speed variances based upon the network’s lie virtual reality. During the 1960s and
bandwidth and levels of use by others, etc.). 1970s virtual reality received a great deal
VPNs can be built out of well-known, well- of attention in military circles as a means
understood, and fully documented proto- for creating aircraft and battlefield sim-
cols and technologies. ulators and other military training envi-
ronments. During the 1990s the advent
Summary of potentially negative issues of more powerful and flexible computing
Leased-line VPNs can be expensive, depend- equipment allowed the practical implemen-
ing upon the speed and other requirements tation of creative projects such as CAVE,
of the members. Public internet-based VPNs a room-sized high-resolution 3D video and
are less secure than leased-line systems audio environment developed at the Uni-
and issues such as speed might not be versity of Illinois at Chicago in 1991, and
guaranteed by the ISP. The creation of a Jack, a 3D interactive environment for con-
VPN requires a significant amount of tech- trolling detailed models of human figures,
nology training and an understanding of which was developed at the University of
networks. Pennsylvania.
Reference
r R. Yuan and W. Strayer (2001). Virtual Business value proposition
Private Networks: Technologies and Solutions Since the 1950s the use of virtual real-
(New York, Addison-Wesley). ity has moved from the research labora-
tory into the workplace, factory, and home.
Associated terminology: Internet,
The development of products such as auto-
Bandwidth, Security, Internet protocol,
mobiles and airplanes would be very dif-
TCP.
ficult without the use of such technolo-
gies. The technology is used as a model-
ing aid in computer-aided styling, factory
Virtual reality design whereby models such as Jack are
used to examine work practices, including
Definition: Virtual reality is a computer-generated
the physical stresses placed upon workers,
immersive three-dimensional simulation of an envi-
and sophisticated simulators are used for
ronment.
the training of pilots. Virtual reality allows
manufacturers to examine their product
Overview
ideas more thoroughly, more quickly, and
Virtual reality has its roots in the work
in more depth than is possible with other
of Morton Heilig, who in 1957 invented
forms of modeling.
the Sensorama Machine (US patent 3 050 870
granted in 1962) that attempted to cre-
ate a three-dimensional (3D) cinematic Summary of positive issues
experience including smell, stereo sound, Virtual reality technology is commercially
and even seats that vibrated. In 1960 he available and software provides platforms
invented the first head-mounted display for a variety of industries and processes,
(HMD), known as the Telesphere Mask (US including medical systems, space applica-
patent 2 955 156 granted in 1960), which tions, manufacturing, aviation, and mili-
provided 3D vision combined with stereo tary applications.

338
Virus

Summary of potentially negative issues traced. In 1988, a virus known simply as


Virtual reality can be resource-intensive ‘‘The Internet Worm” virtually shut down
and may require specialized hardware for the entire internet within a few hours of
high-end performance applications. Dis- its release.
tributed virtual reality has the added com- Originally the term ‘‘virus” was reserved
plication of potentially requiring a high- specifically for illicit software that became
bandwidth network between the entities attached to some already-existing appli-
interacting in the simulation. cation or executable code, which would
become active only when that infected
Reference piece of software was run. Conversely, a
r W. Sherman and A. Craig (2003).
‘‘worm” is self-contained, and does not
Understanding Virtual Reality: Interface, need to infect anything else. Worms are typ-
Application, and Design, Morgan ically delivered and immediately activated
Kaufmann Series in Computer Graphics through some security flaw in the operat-
(San Francisco, CA, Morgan Kaufmann). ing system. When activated, they immedi-
ately attempt to retransmit themselves by
Associated terminology: Artificial
the same means to other computers in the
intelligence, Video.
vicinity. Recently the distinction between
‘‘virus” and ‘‘worm” has become blurred,
and ‘‘virus” is generally taken to cover all
Virus such things.
Delivery: any computer with any kind
Foundation concept: Security. of internet connection is likely to be vul-
Definition: A virus is an illicit piece of software, delib- nerable to viruses. Even if the computer is
erately installed on a computer by an unauthorized used only for browsing the web, it will still
user without the owner’s consent. The usual purpose have many hidden internet connections. An
of a virus is to cause damage to, or destruction of, files. instant of carelessness by the programmers
who created the operating system can leave
Overview a security hole that will eventually be dis-
Computer viruses are now universally covered and exploited, allowing viruses to
known, but generally very poorly under- arrive through ‘‘back doors” that the com-
stood. The original viruses were self- puter owner didn’t know existed. Viruses
replicating programs. Once a virus had may also arrive hidden inside web pages
been delivered to a computer, it would that are being browsed, or in email attach-
make copies of itself to thwart attempts at ments. Any time two computers are con-
removal, and then send copies of itself to nected, even indirectly, there is a potential
other computers (using email contact lists, for a virus to spread from one to the other.
shared document services, and many other Any time files are shared (using floppy
means). The name ‘‘virus” is used because disks, CDs, or DVDs, for example), those
of this strong resemblance to the behav- files could be delivering a virus.
ior of a biological virus. The term is now Activation: contrary to popular belief,
used much more loosely, generally mean- simply having a virus on a computer does
ing any software deliberately run without absolutely no harm at all. A virus is a
authorization. program, a piece of software; it can not
A successful virus could be quietly sent to do anything unless it is executed. With
just a few vulnerable computers, and then older, poorly designed computer systems,
spread itself throughout the world with vir- programs could easily be run automatically
tually no chance of its origin ever being without the owner’s consent. If an email

339
Virus

or a web page contained a piece of soft- its scope. It could erase every byte of data
ware, just opening the email or browsing in the whole system; it could transmit all
the page would be enough to start that soft- of your personal data over an internet con-
ware running. With more recent operat- nection to anywhere in the world; it could
ing systems, user-selectable safety settings make your modem dial a very expensive 1-
exist. The user can decide that no software -900 number; or it could just pop up a mes-
ever runs automatically, or perhaps only sage saying ‘‘Got you!”
software from trusted sources. The user Quite frequently, when a virus gets into
must be aware of these settings, and delib- a computer, it will secretly take over that
erately activate them, but they do provide computer, allowing the virus creator to
some amount of protection. send commands and mount other attacks,
Unfortunately, a virus delivered through or store illegal copies of copyrighted mate-
a ‘‘back door” can still often be activated rial on a computer without the owner ever
automatically. There is a sadly common knowing. Most of the seriously damaging
software fault called a ‘‘buffer over-run.” In attacks are mounted indirectly from com-
simple terms, a software application with puters that have been taken over by a virus,
an internet connection has to set aside so that the originator of the attack is virtu-
some memory for incoming messages to be ally untraceable. This could have unpleas-
assembled in. If the incoming message is ant legal consequences for the innocent
longer than it should be, and the program- computer owner.
mer left out all the checks, an incoming It is not impossible for a virus to cause
message can over-run the boundaries of the physical harm to a computer, but it is very
assembly area and become part of the appli- unlikely, and extremely rare. Just as with a
cation itself. The checks to prevent buffer real biological virus, if it kills its victim,
over-runs are exceptionally simple, and the it has no opportunity to spread itself to
problem of buffer over-runs is well known. others.
For a commercial programmer to allow a Creation: virus creation can be a highly
buffer over-run is an act of carelessness bor- skilled and intellectually demanding job,
dering on gross incompetence, but, sadly, but, as with most criminal enterprises,
buffer over-runs are still probably the most the miscreants are usually completely
common means of entry for a virus. unskilled users of widely available ‘‘virus
Purpose: the general expectation is that kits.” Finding a security flaw in a com-
a virus is intended to erase or modify monly used application, and then work-
essential data, or even destroy computers, ing out a way to exploit it, is a very dif-
to cause great mischief and financial loss. ficult task. Unfortunately, it rarely has to
While this is definitely true in some cases, be done. Responsible software companies
the most common virus is simply an ‘‘ego are always testing their own products, and,
trip.” Somebody wants to cause just enough when they find a flaw, the only responsible
annoyance to be noticed, and earn them action is to release a ‘‘patch” (a small pro-
bragging rights amongst the community of gram that automatically repairs the flaw)
‘‘hackers,” as they like to call themselves. to their customers. Once such a patch has
The fact that great harm is not usually been released, it is much easier for a virus
intended should be of little comfort. Virus designer to work out where the flaw is,
software is often very poorly designed, and then release a tailor-made virus, knowing
can cause a lot of unintended damage. Most that most computer users will not install
importantly of all, a virus can be truly mali- the patch in time to be safe.
cious. Once an illicit program is running on By far the largest group of viruses con-
your computer, there are very few limits to sists of those created by people who either

340
Virus

blindly follow a script for virus creation, or is to expect to be the victim of a disas-
take an existing virus and insert their own trous attack and take appropriate steps in
names into it. Thus all such viruses tend advance. These include encrypting all sen-
to do the same thing, and often don’t work sitive data stored on a network-accessible
(either doing nothing, or being far more computer, and ensuring that all valuable
destructive than was intended). data is effectively backed up: copied onto
some removable media (CD, DVD, tape,
Business value proposition etc.), and physically removed from the com-
Even a virus that does no tangible harm puter and the site. The only data that can
has a negative impact on the bottom line. not be destroyed by a virus is data that is
It takes time and resources to remove a not on a computer.
virus from an infected computer, and they Most importantly, corporations and indi-
must always be removed, because seeming viduals should always install security
harmless is no guarantee of being harm- patches as soon as they are released by the
less. A virus consumes resources, tapping software company that created the operat-
the computer’s processing power and mem- ing system and other applications in use
ory, and uses up valuable communications upon the organization’s computers. It is
bandwidth. Deliberately destructive viruses also important to make sure that the patch
are another matter altogether and can par- being installed is really from the actual
alyze a company through its computing original software company. (Another form
resources. of attack, known as the ‘‘Trojan horse,” is
The only way to be completely safe from to send out free software that claims to be
viruses is either not to have a computer, something useful, perhaps a security patch
or to have a computer but no modem, and even, but is in fact just a virus in different
no internet connection, and never to accept packaging.)
floppies, CDs, or DVDs from anyone. That Special-purpose ‘‘anti-virus” software can
is clearly not a practical approach. To be be effective, but is not the silver bullet that
relatively safe from viruses, there are a few it is often believed (or hyped) to be. Compa-
simple precautions that organizations and nies that create anti-virus software monitor
individuals can follow. the viruses that are flying about the inter-
net. Whenever they detect a new one, they
r Do not buy software from companies work out a way to detect and disable it,
that have a history of exploitable and then send out a (usually free) update
software. to all of their customers. Except for a few
r Turn on all the user-selectable security rare cases, anti-virus software can protect
levels. only against a virus that has already been
r Disable JavaScript in web pages and caught and analyzed. If a company or an
other active software applications. individual is the victim of a new virus (and
r Do not open an email attachment from most victims do fall in the first few days
somebody you do not know unless it after a virus release), the anti-virus soft-
has been scanned by an anti-virus ware will probably not be able to do any-
software package. thing until the ‘‘antidote” for the new virus
r Do not access floppies or CDs that have has been released.
not been scanned for viruses. It should be noted that personal comput-
r Use a reliable Firewall. ing devices that are not ‘‘company issue,”
such as home PCs, need to be considered as
No matter how good your security is, the part of the company’s security assessment,
only responsible attitude for corporations since executives who take home sensitive

341
Visual Basic

or vital corporate data may be vulnera- Associated terminology: Email,


ble in their less secure home-computing Encryption, Internet, Trojan horse,
environments. Spyware, Law cross-reference.

Summary of positive issues


Theoretically, it is possible for there to Visual Basic
be a good virus. A skilled programmer
Foundation concept: Programming language.
could use all the techniques of the virus
Definition: A simplified programming language pop-
creator to produce software that dissem-
ular for low-volume data-transfer tasks and creation
inates itself about the internet, ‘‘attack-
of Windows user interfaces.
ing” every computer it finds, by repairing
all known security flaws and killing any
Overview
active viruses. Such a product is unlikely to
The name Basic is an acronym for Begin-
exist, since nobody would make any money
ners’ All-purpose Symbolic Instruction Code,
from its creation, and there is a signifi-
which was created at Dartmouth College
cant legal liability if the creator makes a
in 1963. It was designed to make program-
mistake, and the ‘‘good” virus does some
ming possible for those who do not have
harm. There is also a significant legal lia-
the time to learn it. Owing to the altruis-
bility if the good virus works perfectly,
tic efforts of its designers to make it widely
because it would be illegal in most juris-
and freely available, Basic was already well
dictions. Practically, there are no positive
known when the microcomputer revolu-
issues other than to say that governments
tion began. Because it is so small and sim-
are actively creating and developing legisla-
ple to implement, it became not just the
tion to prevent, or at least discourage, the
language of choice, but the only language
malicious use of viruses or other software
available for users of many small com-
technology.
puters. A more ‘‘serious” language would
not have been able to fit on computers
Summary of potentially negative issues with such limited memory capacity. By the
The potential negative impact of viruses 1980s, Basic was probably the most widely
upon companies, individuals, and govern- known of all programming languages.
ments can not be understated and anti- The most common tasks for non-
virus measures need to be in place prior technical commercial programmers (that
to an attack occurring. It has been esti- is, the vast majority of all programmers
mated that viruses and their relatives have today) are the transfer of data between two
a total annual cost (for repair, removal, and applications or between an application and
incurred losses) of over $13 billion in the a human user. This kind of programming
United States alone, although there are no does not require great efficiency or the abil-
truly reliable statistics. ity to express advanced algorithms; it is
relatively straightforward data processing
References and user-interface control. Data produced
r M. Egan and T. Mather (2004). The by one application may need to be refor-
Executive Guide to Information Security: matted before it can be used by another;
Threats, Challenges, and Solutions (New human users now expect to be able to pro-
York, Addison-Wesley). vide input to applications through conve-
r P. Szor (2005). The Art of Computer Virus nient and well-laid-out dialogs and win-
Research and Defense (New York, dows, and also expect to see the results in
Addison-Wesley). the same way.

342
Voice over IP (VoIP)

Any operating system that expects to opers from having to create every program
gain popular acclaim is going to need to for every user. This is especially important
provide some way for people to be able to because many Visual Basic programs are
carry out this kind of programming easily. used to develop reports and are applica-
There are not enough technically trained tions that are used once only or for a short
software engineers in the world to meet the period of time.
demands created by the ubiquity of com-
puters, so a non-technical programming Summary of positive issues
language is needed. Since one of Microsoft’s Visual Basic is easy to learn and widely
earliest products was a popular version of available. It was designed to work in a
Basic for microcomputers, it is hardly sur- Microsoft environment, but applications
prising that they chose Basic as the basis can be ported over to other operating
for this language. Visual Basic is a highly systems through special software. Visual
extended version of Basic, designed par- Basic programs can be developed by end
ticularly to work with the Windows oper- users, which frees technical programmers
ating systems, performing data process- and other resources to be used on other
ing and interface tasks for non-technical projects.
programmers.
Edsger Dijkstra, one of the universally
Summary of potentially negative issues
recognized fathers of computer science,
Visual Basic is not suitable for large, com-
said of Basic (Basic in general; Visual Basic
plex, or technically sophisticated projects.
did not exist at the time) ‘‘It is practically
Visual Basic is primarily designed to run
impossible to teach good programming to stu-
on Microsoft operating systems and extra
dents that have had a prior exposure to BASIC: as
resources need to be deployed to run these
potential programmers they are mentally muti-
applications in other environments.
lated beyond hope of regeneration.” He was
partly joking. When used for its intended Reference
purpose, data transfer and implementation r F. Balena (2002). Programming Microsoft
of graphical user interfaces, Visual Basic Visual Basic.NET (Core Reference)
is an immensely successful tool, and it is (Redmond, CA, Wintellect).
hard to argue with success. The only prob-
lems arise when dialects of Basic are used Associated terminology: Fortran, Cobol,
beyond their scope. C++, Java.

Business value proposition


Visual Basic provides end users and non- Voice over IP (VoIP)
technical programmers with a tool for
developing their own programs quickly Foundation concepts: Audio, Digital.
and easily. The language does not require Definition: Speech and other low-bandwidth audio,
advanced training and can be quickly digitally encoded and transmitted over a network
assimilated and used. The language can by computers substituting for normal telephone
also be used to develop prototypes quickly; equipment.
these are non-production versions of soft-
ware that simulate the interfaces and basic Overview
functionality of a system that may poten- Capturing sounds and converting them, in
tially be developed. real time, into digital signals that could
The adoption of Visual Basic frees the IT at any time be converted back into audi-
organization’s trained and expensive devel- ble form and replayed is a very simple task.

343
Voice over IP (VoIP)

For most people with an internet connec- Internet telephony is not the panacea it
tion, data may be transmitted to any point may seem to be. Unlike the Circuit-switched
in the world quickly, simply, and without regular telephone network, internet traf-
any additional cost. Putting those two facts fic is Packet-switched. This means that vari-
together results in Internet telephony or Voice ous parts of a single transmission may be
over IP. transmitted by different routes and at dif-
A computer with an internet connection, ferent speeds, and thus might not reach
a microphone, and a loudspeaker effec- the receiver at the ideal rate. When this
tively gives free long-distance and interna- happens, the result is either an extremely
tional telephone calls to any other simi- annoying lag in reception, or the sound
larly equipped computer. Because human becomes ‘‘choppy” or stuttering, which can
speech remains completely intelligible render speech incomprehensible. Another
under severe bandwidth limitations (3 kHz minor point for consideration is safety:
is quite good for a regular telephone call), the regular telephone network has its own
there is no requirement for a fast internet power source; after natural disasters or dur-
connection; a bi-directional 3000 bytes per ing bad weather there may be power out-
second or 24 000 baud is enough for normal ages, but often the telephones still work.
telephone quality (dial-up modems usually Computers need an operating power sup-
provide the equivalent of 56 000 baud). ply, so they can not provide telephone ser-
With even a standard speed local area vice during power outages.
network, the bandwidth may be increased
to give speech reproduction with much
higher fidelity, and still carry a great many Business value proposition
concurrent conversations, making intra- The use of VoIP technologies is growing
and inter-network conference calls a sim- as they become more advanced, broadband
ple and cheap reality. With the data being reaches high degrees of penetration, and
already in digital form, making a perfect interface technologies become easier and
recording of the entire conference is a triv- more convenient to use.
ial feature to add. The basic calculation of the total cost
In recent years, ‘‘web phones” have of ownership is based upon the need to
emerged. These are small devices looking have a physical connection, a computing
just like a traditional telephone, but con- device, and a ‘‘telephone” interface. For
taining a microprocessor and network con- a large organization in which the major-
nectivity hardware. These devices may be ity of the work force uses a computer in
plugged directly into a DSL, cable modem, performing everyday business tasks, the
or other connection, and provide internet- addition of VoIP carries a low overhead
based telephone service without the need when only straightforward telephone con-
for a computer. nections are required. However, more com-
VoIP, an acronym for voice over internet plex requirements such as call waiting,
protocol, is a standard protocol that runs voice-mail boxes, call forwarding, and tele-
on top of IP (internet protocol, the main conferencing make additional demands on
transport for internet traffic) to implement network bandwidth and storage capacity.
internet telephony. This is a rapidly grow- The major impediment to the adoption of
ing market, and new protocols are appear- VoIP for small organizations is the cost of
ing very frequently. A currently popular the equipment relative to other options
direction is to combine internet telephony such as cellular telephones and land
with Peer-to-peer technology. lines.

344
Voice over IP (VoIP)

The network administrator and other Summary of positive issues


corporate officers need to address several VoIP technology continues to advance in
other issues that may affect their network quality and ease of use. The running cost
through the use of VoIP, including the once the physical equipment and network
impact VoIP would have in terms of band- service have been purchased is low.
width requirements. A network with lit-
tle spare capacity may experience severe Summary of potentially negative issues
problems when extra VoIP requirements VoIP works only when the host device is
are placed upon it. Quality of service is powered up. Sound quality may not be at
affected when an echo is present on the an acceptable level for some users. VoIP
‘‘circuit”; this can normally be solved by requires capital outlay to provide a hard-
adopting acoustic or digital echo-canceling ware and software platform for operation.
technology. Overall, it is important to mea-
Reference
sure the quality of the speech over the r T. Wallingford (2005). Switching to VoIP
network and several vendors provide tools
(Sebastopol, CA, O’Reilly Press).
for speech performance measurement, typ-
ically using the G.711 standard as a Associated terminology: Compression,
benchmark. Internet protocol, Audio, Digital.

345
W3C (the World Wide Web Consortium)

r MIT, 32 Vassar Street, Room 32-G515,


W3C (the World Wide Web
Cambridge, MA 02139, USA.
Consortium) r ERCIM, 2004 Rue des Lucioles, BP 93,

Foundation concepts: Internet, World Wide Web. 06902 Sophia-Antipolis Cedex, France.
Definition: In their own words, “The World Wide Associated terminology: Hypertext, URL,
Web Consortium (W3C) is an international consor- XML.
tium where member organizations, a full-time staff,
and the public work together to develop web stan-
dards” (http://www.w3.org/Consortium/).
Walk-through review
Overview
The W3C was founded in 1994 by Tim Foundation concept: Software development lifecycle.
Berners-Lee, the inventor of the World Definition: A walk-through review is a peer-based
Wide Web, and MIT in collaboration with review of a software development process or product.
CERN, DARPA, and the European Commis-
sion. The organization aims to develop Overview
inter-operable technologies and provide Walk-through reviews are performed on a
specifications, guidelines, software, and software component or on an aspect of the
tools to help the web to reach its full poten- software development process by a ‘‘peer”
tial. A primary aim of the consortium is group. Group members can not be from the
to develop non-proprietary standards and team that developed the product or process
promote inter-operability. The consortium and are removed from the evaluation pro-
is international in scope, has offices in cess. The aim of the review is to examine
fourteen locations around the world, and the process or product from multiple per-
draws its membership from around the spectives. An example review team might
globe (http://www.w3.org/). be composed of engineers representing dif-
The W3C is behind many of the technolo- ferent technical disciplines providing dif-
gies that are in common global use, such ferent perspectives: a software designer, a
as HTML, SOAP/XMLP, URLs, and XML. software interface designer, and a hard-
ware engineer. Alternatively, a review team
may be formed by end users, customers,
Business value proposition and systems analysts, performing a review
The W3C is an organization that provides along business performance lines. The goal
leadership in the area of web technologies. of walk-throughs is to assess the product
Individuals and corporations can become or process and, through a moderator, pro-
members and, through working commit- vide critical positive feedback rather than
tees, help to develop the formulations of criticize the process manager, developer, or
future web-based technologies. software designer.
The review process is meant to be infor-
mal and not confrontational in nature.
Summary of positive issues
While there are many different ways of per-
W3C provides a forum through which web-
forming the review process, the approach
based technologies are developed. Organi-
typically involves the developer of the soft-
zations are able to join the consortium and
ware presenting the specification and the
work with the committees to innovate and
system to the peer-review team. The team
create new technologies.
then acts as surrogate customers and traces
References the system requirements through the devel-
r http://www.w3.org/. opment to the system that customers or

346
Waterfall model

end users would receive. The team also issues and buffers the developers and their
aims to be objective in determining the egos from direct criticism.
usability aspects of the system, including
References
the human factors, the interface expe- r E. Yourdon (1989). Structured
rience, and other aspects of the soft-
Walkthroughs (New York, Yourdon Press).
ware as presented to the user. The third r M. Deutsch and R. Willis (1988). Software
aspect of the team’s role is to ensure that
Quality Engineering: A Total Technical and
the developers followed the methodology
Management Approach, Prentice-Hall
and standards as required by the vendor
Series in Software Engineering
and the customer. Having documented the
(Englewood Cliffs, NJ, Prentice-Hall).
outcomes of these review categories, the
results are presented back to the devel- Associated terminology: Software
opment team together with an evaluation development.
of whether the system needs to undergo
another walk-through review before being
released for use. Waterfall model
Business value proposition Foundation concept: Software development lifecycle.
The objective of walk-through reviews is Definition: The waterfall model is a software develop-
to provide intellectual input to a process ment lifecycle model.
or product quickly and effectively. Walk-
through review teams can be drawn from a Overview
variety of skill bases and the mix changed The Waterfall model developed by Royce in
according to the needs of the problem. 1970 was the first software development
The team may be put together rapidly, the model to address all phases of the software
assessment process executed in a timely lifecycle, capturing each phase before ‘‘cas-
manner, and the feedback quickly pre- cading” its results into the next. The model
sented to the process owners, completing is usually drawn as a series of descending
the review cycle. steps, drawn as boxes, with the results of a
higher box flowing into the next lower one,
Summary of positive issues suggesting its name ‘‘the waterfall model.”
Walk-through reviews provide a rapid val- While there is no ‘‘official” number of
idation and verification tool for managers phases in a waterfall model, a basic ver-
and organizations. The teams performing sion begins with a Requirements definition
the review can be put together on an ‘‘as- phase, in which the goals of the system to
needed” basis. Review cycle times can be be developed are considered by the techni-
very brief and reviews may be performed cal staff, the ultimate users, and the man-
by different teams for different aspects of agement from whose budget the system is
the lifecycle. being paid for. The notation and the level
of formality used to capture the require-
Summary of potentially negative issues ments are selected according to the stan-
If reviews are too informal the results dards enforced by the organization. The sec-
may be inconsistent. The team structure ond phase is Systems and software design, in
and skill sets might not be appropriate which the hardware and software process
for the task, and thus managerial care is requirements are determined and consid-
required in team selection. The review pro- ered in relation to the overall IT architec-
cess should be semi-formal and requires a ture that the system is to be incorporated
moderator who understands the technical within. The software design aspect leads to

347
Web services

the specification and design of each indi- a high-level ranking on the Software Engi-
vidual software component, and again the neering Institute’s (SEI) Software Capability
degree of formality of specification is deter- Maturity Model (CMM). The waterfall model
mined by the organization and its poli- in its non-iterative version does not reveal
cies. Stage three is Implementation and unit problems until the end of the lifecycle, the
testing, in which the programs are coded point at which the cost to fix the error is
and tested as individual entities. Stage four highest. The iterative model does facilitate
is Integration and system testing, in which earlier detection of errors but requires an
the individual programs are tested as com- update revision of the documents and pro-
plete systems. The final phase is the Opera- cesses at each level, a task that is frequently
tional and maintenance phase, generally the underperformed or not performed at all.
longest phase of any software development,
involving modifying the code, continued References
r http://www.sei.cmu.edu/sei-home.html.
testing, and integration with other soft- r I. Sommerville (2004). Software
ware programs.
The waterfall model was originally con- Engineering (New York, Addison-Wesley).
sidered a one-pass development, with each Associated terminology: Formal methods.
stage being undertaken once. However, it
was soon realized that, when an error was
found or a modification was required in Web services
any stage, it was important to return to
earlier stages and amend the design there, Foundation concept: World Wide Web.
so that all stages would reflect the same Definition
model of the system. Hence the water- 1. Any service provided by a web server and
fall model is sometimes referred to as the accessed through a web browser.
Iterative waterfall model. 2. A particular protocol or group of protocols
used by applications to exchange data over a
Business value proposition network.
The waterfall model has a long history and
many organizations have incorporated a Overview
version of it into their application devel- This is one of those troublesome phrases
opment model. Many other techniques, that for many years had a perfectly obvious
such as cost modeling systems and project and straightforward meaning (a ‘‘web ser-
management systems, have been built vice” is any service provided via the World
around the waterfall model and provide Wide Web), but was later kidnapped and
an established methodology for software used for a different meaning by a small but
development. demanding subset of society. Since the two
meanings are closely related, both being
Summary of positive issues kinds of internet software technology, but
The model is an established approach very different, it can be hard to tell which
to software development; tools and tech- meaning any given speaker intends.
niques associated with the waterfall model 1: Any service provided through the
have been developed and there is a wide World Wide Web, in other words anything
user base. that can be accessed over the internet
through a web browser, is a Web service. This
Summary of potentially negative issues covers a great multitude of things, from
The approach generally lacks formality and simple static web pages, through informa-
is not a methodology that would achieve tion services, digital music and movies, to

348
Wireless Application Protocol (WAP)

e-commerce and complex interactive envi- applications have been embraced by many
ronments. organizations and government entities,
2: When applications need to interact and provide a flexible approach to systems
or share data across the internet without access. For example, The US Navy--Marine
human assistance, some common language Internet initiative, commenced in 2000,
and protocol of communication must be aims to allow over 300 000 users access to
provided. Web services refers to the entire all of the US Navy’s estimated 80 000 appli-
collection of data languages and communi- cations over the internet.
cations protocols used by a group of appli- The technologies that underlie web
cations for this purpose. Most versions of services also facilitated the creation of
web services insist that data is represented business-to-business e-commerce and data
using the XML language. Various protocols exchanges, including the provision of data
are available, including the existing stan- synchronization engines and data pools.
dards FTP and HTTP, and the special-purpose
SOAP (Simple Object Access Protocol) and Summary of positive issues
CORBA (Common Object Request Broker Web technologies continue to mature and
Architecture). Additional systems such as their development allows greater connec-
WSDL (Web Services Description Language) tivity between entities. Web services allow
and UDDI (Universal Description Discovery access to remote applications.
and Integration) are available to allow sys-
tems and applications to identify them- Summary of potentially negative issues
selves and others, and to communicate how The terminology associated with web ser-
a particular web service is to be used. vices is used ambiguously. Web services can
be technically demanding to implement.
Business value proposition
Reference
The development of internet and web- r H. Deitel, P. Deitel, B. DuWaldt, and L.
based technologies has resulted in the
Trees (2002). Web Services: A Technical
development of many different models
Introduction (Englewood Cliffs, NJ,
through which businesses can interact and
Prentice-Hall).
communicate.
The most basic form of ‘‘web service” is Associated terminology: Data pool,
typically any interaction with a remote web Internet, Protocol.
page, such as viewing its contents, using
the web site to order a product, accessing
a database, or other function. The sophisti- Wireless Application Protocol (WAP)
cation of this form of web service leads to
the creation of e-commerce business mod- Foundation concept: Wireless network.
els through which commerce is transacted. Definition: The Wireless Application Protocol is an
A second form of web service, some- open-specification communications protocol for pro-
times referred to as a ‘‘web-enabled applica- viding information services on wireless devices.
tion,” is just that, an application accessible Not to be confused with: WAP – wireless access point.
through the use of internet technologies
and through which data and/or services Overview
can be accessed. For example, an employee The Wireless Application Protocol (WAP) evol-
who travels extensively on company busi- ved out of a need for wireless providers to
ness may remotely connect through the create a common platform through which
internet with their company’s systems and their devices could inter-operate. WAP 1.0,
use a variety of applications. Web-enabled which was developed under the auspices

349
Wireless Application Protocol (WAP)

of the WAP Forum, was released in 1998 closely related to HTML, the standard lan-
and was subsequently developed into the guage for web pages, and is compatible
backwards-compatible WAP 2.0, released in with the XML standard, but is optimized
January 2002. for wireless use. Instead of using web
WAP provides an environment in which pages, WML treats data for display as a
web applications can run on hand-held Deck of Cards, each card containing a small
devices using Micro-browsers (browsers desi- amount of data, with the user expected
gned to operate within the limited envi- to navigate through the deck at their own
ronment of a very small computer). As pace. Many wireless devices have severely
such, WAP is based on a six-layer proto- limited display regions (such as the small
col stack similar to the OSI standard Seven- screen on a cellular phone), and could
layer model, and has been developed to not be used to view normal web pages.
support internet protocols including TCP A common practice is to provide a WAP
and HTTP. Central to these layers is the gateway, acting as a proxy between servers
Wireless Application Environment Layer and viewers, converting standard HTML to
which provides the application-developer a reduced-size WML document for conve-
interface through which developers can nient viewing.
write Extensible Hypertext Markup Langu-
age Mobile Profile (XHTMLMP) documents Business value proposition
that may be displayed on any device with The WAP protocol was a major break-
a WAP micro-browser. WAP’s WAE (Wire- through in mobile computing since it uni-
less Application Environment) also facili- fied vendors under a single set of proto-
tates the development of more advanced cols and avoided the frequent problem of a
‘‘presentations” through other standards ‘‘land-grab” mentality under which differ-
including WML (Wireless Markup Lan- ent proprietary protocols battle it out for
guage), WML Script, and WBMP (Wireless market dominance. WAP provides a stable
Bitmap image). protocol environment through which deve-
The basis of WAP operations is the lopers can create web presentations aimed
standard client--server model, in which a at devices with micro-browsers, including
request for service is sent over the network PDAs, cell phones, and two-way radios.
to the appropriate server identified by its WAP must be assessed from a security
IP address or URL. In this case, communi- perspective, in that, when the system’s full
cations are wireless-oriented; the request capabilities for strong encryption (similar
commences as a wireless signal, and is to the Secure sockets layer) are used together
responded to by the web server and sent with a secure server gateway at the recipi-
back across the network. ent’s side of a request, the security of the
The WAP system is capable of running data is high. However, there is a possibil-
on top of any operating system, includ- ity of using a weakened encryption system
ing those specially written for hand-held and disabling the security system. For orga-
devices with small amounts of memory nizations or even individuals using WAP-
and limited processing capabilities (these enabled devices to transfer sensitive data,
operating systems are often optimized, careful consideration of end-to-end security
slimmed-down versions of the full operat- needs to be made by a professional systems
ing system), as well as being optimized to manager.
maximize the relatively low data transfer
rates associated with wireless systems. Summary of positive issues
Closely associated with WAP is WML (the WAP is an open standard supported by
Wireless Markup Language). WML is very the WAP Forum. The standard is universal

350
Wireless network

in hand-held devices that support micro- Metal wires are not perfect; they get in
browsers. WAP allows developers to create the way, trip people up, and get broken and
presentations specifically for micro-brow- tangled. If the public is to be given access
sers. WAP supports a form of Digital certifi- to a network, connectors and sockets have
cates and other security enhancements. to be provided, and the public will quickly
break them or fill them with foreign sub-
Summary of potentially negative issues stances. Wires also prevent mobility; it is
WAP offers a development and presenta- hard to move around freely while using a
tion environment that is limited by pro- computer that is connected by a cord to the
cessing hardware and memory, and by the wall.
data transfer rates of the wireless system. When mobility or public access is requi-
WAP has some security concerns associated red, or under other circumstances when
with the WTLS (Wireless Transport Layer wires may be undesirable, a wireless net-
Security) protocol which encrypts the data work is of course the solution. On a wireless
transmitted, and governs the security at network, every device must be fitted with
the WAP gateway where the server connects a transceiver (a transmitter and receiver in
to the network. one) that converts between the native elec-
tronic signals of the computing device and
References
r www.wapforum.org. the radio-frequency transmissions used by
r C. Arehart and N. Chidambaram (eds.) the wireless network. This additional equip-
ment adds some small expense, but at least
(2000). Professional WAP (Chicago, IL,
there is a single universally adopted stan-
Wrox Press).
dard, or at least a family of mutually com-
Associated terminology: Network, LAN, patible standards (in the IEEE 803 suite), so
Protocol, HTML, XML, Security, Digital wireless network adapters are as standard
certificate. as the ethernet adapters that they replace.
Unless a wireless network is to be totally
isolated and not part of the internet, there
Wireless network must also be a Base station or Wireless access
point, to provide a gateway.
Foundation concepts: Network, LAN. Wireless networks are not well suited
Definition: A local area network in which the connec- to all circumstances. Wireless signals are
tions between computers use not wires, but radio- not easily contained or directed. When two
frequency broadcasts. networks are close together, great confu-
sion can be caused for transceivers in the
Overview area of overlap. If a base station is not
Electrical signals naturally flow along properly secured, unauthorized users can
metal wires; they require no special trans- ‘‘hitch a free ride” on the network, using
mitters or receivers, and are exceptionally an internet connection that someone else
cheap and reliable. They are also highly has paid for. Most wireless network base
directional: a signal inserted at one end stations are easy to secure against unau-
of a wire flows straight to the other end, thorized use, but it does require some user
and does not leak out in unwanted direc- configuration, which is most often skipped.
tions. This makes metal wires seem like the If any network traffic is not encrypted,
ideal medium for inter-computer data con- it can easily be intercepted by an unde-
nections, and indeed most network connec- tectable interloper. Inside buildings, walls
tions are in the form of normal electrical and equipment, and even moving people,
cables. can interfere with network signals, and

351
World Wide Web

cause unpredictable weak spots where Summary of positive issues


there is no network access. Wireless networks provide computer users
and computing devices with flexibility in
Business value proposition terms of their operational locations. The
Wireless networks provide organizations standards are well defined both at the IEEE
with the ability to maintain or even level and at public telephone network level.
increase the connectivity between com- Many wireless networks in public spaces
puters whilst reducing the complexity are provided as a ‘‘free” service and some
involved. Wireless network systems allow municipalities are providing or contemplat-
a variety of system solutions to be created ing the provision of wireless networks to
within an organization, including so-called their communities. RFID and other tech-
wireless ‘‘hot spots” where a transceiver nologies use wireless network technologies.
allows wireless connections to be made by
any authorized device. While wireless net- Summary of potentially negative issues
works enable employees to work anywhere Wireless network systems are slower than
there is a hot spot, many organizations are cable-based systems. The cost of providing
structured so that employees work in fixed wireless networks needs a solid business
locations or can access their data anywhere case to be built around it. For some orga-
within an organization by logging onto a nizations such as coffee shops, it may be a
fixed cable-connected networked computer. positive in attracting customers but a large
While the benefits can easily be deter- negative if the customers don’t buy coffee
mined for an institution such as a univer- and just linger for free network connec-
sity where the ‘‘customers” (the students) tions in a nice atmosphere. The use of an
are working from a variety of locations in insecure wireless system is detrimental to
a day, or a hotel where the guests require any organization’s networks and intellec-
internet access as a service within ‘‘pub- tual properties.
lic areas,” the cost justifications in devel-
oping a business case for a corporate move Reference
to wireless networks may be more difficult r K. Pahlavan and P. Krishnamurthy
in the case of a traditional static work- (2001). Principles of Wireless Networks: A
force. The case for equipping a company’s Unified Approach (Englewood Cliffs, NJ,
mobile workforce with wireless technology Prentice-Hall).
can be based upon productivity gains, but
must be balanced with the risk of security Associated terminology: Network, Local
breaches. area network, Wireless Application
Wireless network use goes beyond the Protocol.
simple connection of laptops to corporate
networks, and includes the development
of devices such as RFID tags that allow World Wide Web
the monitoring of devices or items within
a wireless coverage zone. Beyond the IEEE Foundation concepts: Internet, Hypertext.
803 standards-based technologies there are
other mechanisms through which wireless Overview
communications can occur. One such tech- Although the term ‘‘World Wide Web” is
nology is the wireless cellular network that often taken as a synonym for ‘‘the inter-
supports the 3G computing devices embed- net,” there is a subtle shade of differ-
ded in telephones. ence in meanings. ‘‘The internet” really

352
World Wide Web

describes the worldwide collection of inter- servers) use to communicate requests and
connected computers, the hardware that their corresponding documents.
connects them, and the fundamental soft- It is HTML and HTTP that make the seam-
ware and protocols that make communica- less browsing of the web, or surfing of
tions between them possible. The ‘‘World the net, possible. HTML and HTTP together
Wide Web” refers to the web of intercon- allow the carefully designed graphical lay-
nected documents, data, and applications outs with an almost infinite variety of
that spans the internet. fonts, text sizes, backgrounds, colors, ima-
The internet existed long before the ges, and other graphical elements, and the
World Wide Web that we know today (see online forms and their automatic process-
Internet for details). Since the early days of ing on which e-commerce is founded. Of
ARPANET, it has been possible for anyone course, many other applications provide
with internet connectivity to put any col- the same mixture of text and graphical
lection of data online, so that anyone else elements, and many provide better control
with the same connectivity may download over the appearance of a page, but the fun-
and access them at will (‘‘A file transfer damental thing that makes the web more
protocol” was published by Abhay Bhushan than just a collection of files accessible
of MIT in 1971). However, those data files through the internet is the ability to follow
were stand-alone documents: each could be a reference to another online document
viewed in any convenient way, and they anywhere in the world instantly and auto-
may contain references to other online doc- matically. That is the difference between
uments, but those references would be sim- text and hypertext; it is also the difference
ple inert text as might be printed in a between the internet and the World Wide
book. Any reader wishing to view a ref- Web.
erenced document would have to find it, In order to publish documents on the
manually download it, and separately view web, only one thing is required: a Web
it; not necessarily a difficult process, but server. A web server is usually thought of
not the smooth operation we are familiar as a networked computer with this desig-
with today. nated task, but it is actually just a simple
The World Wide Web did not exist until piece of software running on a networked
the introduction of a practical Hypertext computer. A single computer may in fact
system by Tim Berners-Lee of CERN (Conseil support a large number of different web
Conseil Européen pour la Recherche Nucléaire or servers. Web server software compatible
the European Council for Nuclear Research) with all popular computers and operating
in 1990. This included HTML (the HyperText systems is freely available for download,
Markup Language) and HTTP (the Hyper- but there are also commercial versions
Text Transfer Protocol). HTML (see Hypertext available that may provide more support.
for details) is a simple language that Once a web server has been installed and
allows the creation of documents that mix started on a networked computer, it is con-
text, graphics, and other media, which figured to make certain groups of files avail-
includes Hyperlinks: references to other able for browsing, and it simply waits for
online documents that can be retrieved requests composed in the HTTP language.
and displayed automatically, using the now Each request received is analyzed to dis-
familiar point-and-click technique. HTTP cover which actual file or application is
is the even simpler language that docu- wanted, and, if that access is permitted,
ment display software (web browsers) and the results are transmitted back. The selec-
the remote document storage systems (web tion of files made available by a single web

353
World Wide Web

server is usually referred to as a Web site, site. A reference to a file, document, or


although this term is not strictly defined. application on a web site is encoded as a
For a web server to be accessible, its URL, which includes both a domain name
IP address must be made known. The and additional file designations (see URL for
most common way is through DNS, the details). URLs are the basic address desig-
Domain name service. This is a system that nators for the World Wide Web, just as IP
associates humanly understandable and addresses are for the internet.
memorable names (such as ‘‘YourCompany- Recently, two newer models for web-
Name.com”) with the totally unmemorable based information dissemination have
and unfriendly numeric IP addresses (such become popular, the Blog and the Wiki. A
as ‘‘192.168.31.108”) actually used in inter- blog (web log) is little more than an online
net communications (see Domain name for public diary: a person writes thoughts,
details). Given just a domain name or an opinions, poems, or whatever else comes to
IP address, web clients or browser applica- mind, perhaps adding pictures and audio
tions know exactly how to contact the indi- components, to a continuously evolving
cated web server, and request an appropri- personal web site in the hope that others
ate starting document. will read it. Bloggers usually arrange to
Normally, each web site has its own have links to their sites placed on blog-
Home page, a specially designed hypertext directory sites so that others will find
document (usually stored in a file called them. Most blogs are like most other self-
‘‘index.html”) that serves as an introduc- published work, of no interest to anyone
tion and guide to the whole site. When a but the author, but a few of the more tal-
web browser is given only a domain name ented authors do gain quite a following.
or IP address, it asks the associated web A wiki (derived from a Hawaiian word)
server for its home page file, and displays is a public web site at which anyone in
that to the user. The home page would nor- the internet community may add content
mally be expected to contain hyperlinks or even edit existing content. Occasion-
to other documents on that site or oth- ally there may be a designated modera-
ers, enabling visitors to find whatever they tor who checks all changes and additions,
want. but most commonly there is no oversight.
The term Web browser usually refers to The best known of all wiki sites, Wikipedia
the software application (Mosaic, Netscape, (http://en.wikipedia.org/wiki/Main Page), is
Internet Explorer, etc.) through which a growing into a large and sometimes useful
human user accesses web content, alth- free online encyclopedia. Anyone may con-
ough sometimes the term is applied instead tribute articles, but, since there are a great
to the human user. It is the browser applica- many frequent users, untruthful or mali-
tion’s responsibility to take web addresses cious additions are usually caught quite
entered by the user and convert them into quickly, although vandalism of its content
proper HTTP requests sent to the appropri- is, of course, a recurring problem. Wiki
ate server, to display the response appropri- sites may be used as general meeting places
ately, and monitor mouse clicks, converting for people to share opinions, as bulletin
them into additional HTTP requests when boards for posting notices, or as the center
those clicks occur over hyperlinks. of a collaborative online workforce.
Domain names refer to whole web sites.
Web addresses entered by users, or encap- Business value proposition
sulated in hyperlinks, may instead refer to The World Wide Web is a medium for
individual documents on a particular web the distribution of information. It allows

354
World Wide Web

individuals, organizations, and governme- under peak demands would be valuable in


nts to access data and to make their own order to avoid crashes at true peaks in sys-
data accessible very cheaply. The devel- tems demand such as Christmas shopping
opment of an externally facing corporate periods, and for stock trading at key IPOs.
web site is usually centered on one of seve- There is a very large array of tools and tech-
ral business models, including business- nologies available to assist web masters in
to-customer (B2C) commerce, business-to- the development, testing, and deployment
business (B2B) commerce, business-to-gov- of web sites.
ernment (B2G) commerce, government-to- The choice of browser to access and read
citizen (G2C) commerce, online-community web sites has evolved since the ‘‘browser
models, and blogs. It is also possible wars” of the 1990s. While there are domi-
to use web technologies to provide web nant vendors in some computer categories,
sites for internal corporate use only; these there are many emergent browsers for sev-
would house such things as internal cor- eral classes of computer. Free open-source
porate training materials, data on human browsers are available for desktop systems
resources, and documents relating to tech- and many vendors offer micro-browsers for
nology and research. This provides a low- hand-held devices such as cell phones.
cost model for the dissemination of infor- The advent of blogs has required many
mation and helps companies to control and corporations to develop a formal policy
monitor access to resources. on employees’ online behavior, and strictly
The development of externally facing cor- enforce it. Policy usually includes the obli-
porate web sites requires careful considera- gations of the blogger toward the com-
tion not only of the design of the hyper- pany, and requires that blogs relating to
text but also in terms of the intent of any company activity or information be
the web site as related to the corporate authorized by the company’s media rela-
strategy. This requires achieving a balance tions department, and include the author’s
amongst the market demands that the site real name and their official role in the com-
aims to address, the service level to be pro- pany. The policy document should stress
vided by the site, and the brand implica- the need to maintain corporate intellectual
tions of the site and the technology that property and that blogs should not bring
supports it. A key aspect of the brand is the company into disrepute by including
the selection of the domain name, since a personal opinions on sensitive subject mat-
memorable name such as ‘‘YourCompany- ters, such as on an IPO during its quiet
Name.com” is clearly easier for customers period. Blogs can be positive and many exe-
to remember than an IP addresses such as cutives and business personalities who blog
‘‘192.168.31.108.” Companies need to deter- have a large following. The blog can act as a
mine whether they wish to use the same vehicle for expressing intellectual content
name as their physical brand name or and building an online community, ulti-
whether the online brand is going to have mately adding value back to the company.
a separate identity. The development of wikis has proved a
The deployment of web sites on a cor- useful mechanism for collaborative efforts.
porate web server also requires an assess- Within corporations, wiki web sites allow
ment of the impact of that sales channel for a wide variety of collaborative work
upon the corporate value chain, as well as spaces, such as the joint development of
an assessment of the scalability and robust- documents, product plans, and specifica-
ness of the technology behind the site. For tions. The relatively low cost of maintain-
example, a stress test to assess the system ing wiki sites combined with their easy

355
WYSIWYG

access through a standard browser makes produced on an output device such as a printer or
this form of collaboration tool a medium through a browser then the document production
through which remote or distributed work- system is a WYSIWYG system.
ers may contribute to a project in an asyn-
chronous manner. Overview
Word processors and other content-
Summary of positive issues creation software may be classified as
The World Wide Web provides a popular either WYSIWYG (pronounced whiz-ee-wig)
mechanism for disseminating information or non-WYSIWYG. Word processors in the
across the internet. A web site can be inter- 1970s and early 1980s such as WordStar
nal to an organization, whereby content generally operated in one of two modes:
may be restricted to a workforce or group of Edit mode and View mode. When in edit
workers. Externally facing web sites can be mode, strange annotations would appear
used to provide information or to sell prod- in the text (like the HTML markup tags of
ucts. The World Wide Web is based upon today, but far more arcane); users could
a set of open standards and protocols that edit the text, but it would often be hard
continue to evolve to meet new challenges. to visualize how it would appear when
The technologies and services associated printed. In view mode, the document
with web development are widely available would appear in an approximation of its
and have a wide range of capabilities; many final printed or display form, with roughly
tools are open-source and free to download. correct text layout, and perhaps even the
Domain names can be acquired to reflect right fonts, but it would not be possible
the branding statement of the organization to make any modifications to the docu-
or individual. ment without switching back to edit mode.
These were non-WYSIWYG systems. Under a
Summary of potentially negative issues WYSIWYG system there is only one mode of
The technology underpinning the World operation: editing is performed while view-
Wide Web continues to evolve and requires ing the document in its production form,
that the corporate or individual web and the document is shown in as close
master maintains a working knowledge to its final form as the display hardware
of advances. Popular domain names may permits.
already have been acquired by others. Not The first WYSIWYG word processor,
all browsers have the same capabilities: BRAVO, was created at Xerox PARC in 1974
micro-browsers on hand-held devices do and this style of word processing has con-
not have the same capabilities as desktop tinued to grow in use, having been popu-
computer-based browsers. larized by Microsoft’s Word program and
Adobe’s Frame-Maker. However, development
Associated terminology: Internet, Server-- of markup-language-based document pre-
client, Hypertext, Internet protocol. paration systems continues and systems
such as LaTeX facilitate the production of
high-quality documents and publications,
WYSIWYG especially when specialized notations and
symbolic forms are required. Often a doc-
Definition: An acronym for “What You See Is What ument processing system will provide two
You Get,” describing the visible form of a document interfaces for working in HTML and simi-
while it is being developed. If the visible form the doc- lar markup languages, allowing the user to
umenttakeswhileitisbeingcreatedisthesameasone directly edit the markup tags, or to work in

356
WYSIWYG

WYSIWYG mode, indirectly modifying the Summary of positive issues


document’s final form. WYSIWYG systems are easy to use, intuitive,
and may require less training than non-
Business value proposition WYSIWYG systems. WYSIWYG word pro-
The use of WYSIWYG environments is cessing systems have achieved widespread
intuitive; they frequently require less user adoption.
training and can speed up the development
of documents and projects. Non-WYSIWYG
document preparation systems and envi- Summary of potentially negative issues
ronments can provide superior control over The WYSIWYG systems are more expen-
the documents produced, and may even be sive, may reduce direct control over the
easier to use when specialized document documents being produced, and may lack
formats are to be developed. Environments the resources available to non-WYSIWYG
that facilitate markup languages such as developers.
HTML can also reduce the complexity of
development. Associated terminology: Hypertext.

357
X.12

a Mapping document, which specifies exactly


X.12
which segments, data items, and fields are
Definition: An ANSI protocol that defines a set of to be used, is agreed upon and shared by
standards for documents involved in electronic data the two parties.
interchange.
Business value proposition
Overview The X.12 standard is widely used and has
The American National Standards Institute undergone consistent revision and modern-
(ANSI) chartered an Accredited Standards ization since its creation in 1979. The stan-
Committee (ASC) in 1979 to develop a dard covers a very wide range of docu-
set of standards for the documents that ments, with approximately 300 transaction
are exchanged through Electronic data inter- sets having been developed.
change (EDI). The standard has periodically The X.12 standard continues to evolve
been revised by the ASC, with version 2 to meet the demands of practitioners. Ver-
being established in 1986, version 3 in sion 4010 onwards is year-2000 compli-
1992, version 4 in 1997, and version 5 in ant, and addresses new regulatory require-
2004. The standard defines Transaction sets ments such as the 1996 Health Insurance
to capture the information that would be Portability and Accountability Act (HIPAA),
found in a paper document. For example, which in the United States requires that all
the X.12 transaction set 850 represents an healthcare providers be X.12-compliant by
outbound purchase order, 860 a purchase August 1, 2003.
order change document, 855 a purchase The use of the X.12 standard for EDI
order acknowledgement document, set 856 provides companies with the opportunity
represents packing slips, and 810 repre- to reduce transaction costs, speed data
sents invoices. exchange, increase reliability, and simplify
The X.12 standard breaks a transaction procedures.
set into logical groups of data called Seg-
ments. For example, the 850 purchase order Summary of positive issues
set could contain the following segments: The standard is mature, having undergone
a transaction set header, an initial seg- more than 25 years of development. It has
ment, a reference number, a date/time ref- been adopted by a wide variety of indus-
erence, carrier details, name, product iden- tries and by more than 300 000 companies
tifier, physical details of the item, marking/ worldwide. The standard has continued to
packing/loading, destination quantity, and evolve and is a required federal standard in
transaction totals. All segments for each the United States for healthcare providers.
transaction set are extensive because they
have to be usable by many different organi- Summary of potentially negative issues
zations, and it is usual for only subsets The three major EDI standards, namely the
of the defined segments to be utilized by United Nations Guidelines for Trade Data
each company. The segments are themsel- Interchange (UNTDI), the United Nations
ves composed of data items; thus the item rules for Electronic Data Interchange for
‘‘marking/packing/loading” could contain Administration, Commerce and Transport
the fields item description type, packag- (EDIFACT), and ANSI X.12, use transaction
ing characteristic code, association quali- sets that are not compatible and need
fier code, and packaging description code. middleware systems to translate the data
In order for the two companies that are from one format to another. The ASC com-
exchanging the data to understand which menced work in 2001 to make X.12 and
codes are to be used in each transaction, EDIFACT share common transaction sets

358
XML

and move toward a global standard, since sion over the network. If the called ter-
X.12 is widely used in the United States minal accepts the call, a full duplex (bi-
and EDIFACT is widely used in the Euro- directional) transfer of data can occur. The
pean Union. The ASC X.12 Standards orga- communication session can be terminated
nization also began an initiative in 2001 to by either party at any time.
harmonize X.12 with XML-based messaging
standards, but XML and X.12 remain sepa- Business value proposition
rate approaches to document messaging. The X.25 protocol was initially developed in
Reference 1976 to facilitate the creation of wide area
r http://www.x12.org/x12org/index.cfm. networks over public telecommunications
equipment. The technology is mature, hav-
Associated terminology: XML. ing been revised in 1980, 1984, and 1988,
resulting in a debugged and stable envi-
ronment that provides high-quality data
X.25 connectivity. X.25 has been implemented
widely and products with X.25 certification
Foundation concepts: Packet switching, Network. are widely available.
Definition: X.25 is an ITU-T protocol that defines a
set of standards for connecting user and network
Summary of positive issues
devices.
The technology is mature, stable, and pro-
vides high-quality connections. Products
Overview
adhering to the standard are widely avail-
The X.25 standard was defined by the Inter-
able. X.25 applies to dial-up connection
national Telecommunications Union (ITU),
methods.
an organization that operates under the
auspices of the United Nations to develop
and coordinate global telecommunications Summary of potentially negative issues
standards. The ITU-T (the final ‘‘T” signify- The X.25 protocol incurs a turn-around
ing the telecommunications sector group) delay on messaging between terminal
uses a naming convention of a letter and devices. This can be detrimental to heavy
a number separated by a period. The letter bi-directional communication over an X.25
‘‘X” signifies a group of related standards in network. Line speeds of 64 kbps are too slow
a common area, namely data networks and for many applications that require higher
open systems communications. The num- levels of bandwidth to perform their tasks
ber ‘‘25” signifies that the standard covers usefully.
the area of packet switching.
References
The X.25 standard applies to the first r http://www.itu.int/home/.
three layers (network, data, and physical) r M. Clark (2001). Networks and
of the OSI Seven-layer model and is typically
Telcommunications (New York, John Wiley
used on the networks of telecommunica-
and Sons).
tions companies (generally known as com-
mon carriers), and enables user devices and
network devices to communicate with each
other regardless of their type. XML
Communication over an X.25 network
begins when a data terminal such as a per- Definition: The Extensible Markup Language (XML) is
sonal computer calls another data termi- a standard language for describing and encapsulating
nal to request a joint communication ses- data.

359
XML

Overview i.e., the element types, their attributes, and


The concept of an Extensible Markup Lan- so on, are defined by an XML schema (itself
guage (XML) took shape in November 1996, written in XML) that is transmitted to the
evolving from the Standard Generalized partner system in a data exchange prior
Markup Language (SGML) which became to the data transmission itself. However,
ISO 8879. The major concept underlying another group of companies may use a
markup languages is that they contain tags completely different set of tags to denote
or markers that signify some form of cod- purchase orders to suit their particular
ing. In the example needs. The flexibility of XML allows easy
adaptation to individual requirements, but
<bold> word </bold> this may require additional software to
the tags <bold> and </bold> signify that translate between different XML schema
whatever comes between them should be designs.
printed in bold; mechanisms of this nature
were employed by many early word pro- Business value proposition
cessing systems. This concept is employed XML was developed under the auspices of
by a variety of markup languages, most the World Wide Web consortium (W3C.org)
famously HTML, which is used to mark and the W3C XML working groups con-
up data so that the format of the data tinue to develop the language and its exten-
can be controlled when displayed by a web sions, providing recommendations, work-
browser. While the tags in HTML are pre- ing documents, and documentation. XML
defined so that all systems use the same is platform-independent and widely sup-
tag sets and their meanings are universally ported, with open-source and proprietary
known, the tags in XML are not predefined. parsers available, and XML technologies
This allows (and requires) developers to cre- have been embraced by a wide array of
ate their own data structures unique to software vendors from specialist XML ven-
a task’s requirements. This is useful when dors to ERP vendors such as SAP. The tech-
organizations wish to exchange data and nology has continued to grow in use, and
choose not to follow the strictly defined reg- a registry through which companies can
imen of X.12. For example, the X.12 trans- automatically discover and connect with
action set 850 for purchase orders could each other via XML, known as the Universal
be replaced by a purpose-designed purchase Description, Discovery and Integration (UDDI)
order schema in XML: initiative, has been created.

<?xml version=“1.0”?>
<purchase order> Summary of positive issues
<description> The XML approach to data description
<order type> transfer is flexible, cross-platform, and
“Urgent Shipment”</order type> widely utilized by software vendors. The
<Destination Address> language is supported by W3C and contin-
“Head Office”</Destination Address> ues to evolve, supporting encryption, com-
<Item Description> municating with web services, and facilitat-
“Large Widget”</Item Description>
ing end-to-end transactions.
</description>
</purchase order>
Summary of potentially negative issues
The names and placements of the tags are XML may not be used for the transmis-
purely at the discretion of the individual sion of healthcare data in the United States,
designer. The structural details of the tags, because HIPAA mandates the use of X.12N

360
XML

(The N signifies ‘‘insurance” documentation sockets layer (SSL) approach is frequently


under the X.12 EDI protocol) for dental, pro- adopted for securing XML data transfers.
fessional, and institutional transactions,
Reference
and the National Council for Prescription r http://www.w3c.org.
Drug Programs (NCPDP) standards for phar-
macy transactions. Security when using Associated terminology: X.12, Hypertext,
XML has to be considered and a Secure Protocol.

361
Y2K problem

systems represent dates and times by the


Y2K problem
number of seconds elapsed since midnight
Definition: The problem caused by using two digits to on January 1, 1970, but frequently use a
represent the year in recorded dates (e.g., 97 to mean numeric format that can count only up to
1997),andbyextensionanyproblemcausedbyusinga 2 147 483 647. Two billion is a large num-
data-limitedrepresentationbasedoninvalidassump- ber, but two billion seconds is less than 70
tions about the possible range of values. years; many Unix systems have their own
version of the Y2K problem scheduled for
Overview some time in January 2038.
The ‘‘year-2000” problem, or Y2K, is very Those who do not learn from history are
well known. Many years ago, high-capacity doomed to repeat it. George Santayana
long-term storage was expensive, and by
modern standards not very high-capacity We learn from history that we learn nothing
from history. George Bernard Shaw
at all. Every commercial programmer knew
the need to compress data so that more
effective use could be made of the limited Business value proposition
capacity available. In the mid 1960s, pro- The Y2K data problem should act as a
grammers working on commercial appli- reminder to all IT managers and program-
cations never imagined that their prod- mers that the data format they choose at
ucts would still be in use 35 years later, the outset of a project needs careful con-
and storing the ‘‘19” that appeared without sideration since data structures created by
fail at the beginning of every year number their systems may be around for a very long
seemed like an absurd waste. time.
In all respects but one, they were right.
The Y2K problem was not caused by incom- Summary of positive issues
petence or laziness amongst programmers. The vendors of software applications today
The applications they wrote 35 years ago aim to build scalable solutions and provide
are generally not still in use. Unfortunately, tools to help organizations migrate their
the data sets that those applications pro- systems from one vendor to another, allow-
duced, replete with their missing 19s, out- ing companies to move from outdated soft-
lived those applications and have remained ware using proprietary vendors’ formats.
in continuous use, growing rapidly, ever
since.
Summary of potentially negative issues
Fortunately, the world’s programmers
The Unix problem and other ‘‘time bombs”
successfully updated all (with a very few
are still in existence, some known and
exceptions) applications and data sets in
some still unknown at this time. Software
time, and the financial ruin and end of the
from vendors that either have gone out
world as we know it predicted by the pop-
of business or no longer support their
ular press was averted.
software may contain undetected problems
Unfortunately, very similar problems still
that could surface at any time, at which
exist in many guises. The use of seven dig-
point there may be no ‘‘quick fix.” Or worse,
its for telephone numbers within large geo-
the problem could be catastrophic for the
graphical areas was a variant of the Y2K
organization.
problem: who would ever have expected
the nearly ten million telephone numbers Associated terminology: Reliability,
in each area code not to be enough? Unix Database.

362
Zip

(3) A Zip operation is a simple data process-


Zip
ing step, in which two or more equal-
Definitions: length lists of data records are com-
bined into a single list of the same
1: A popular file compression system. length by merging or concatenating the
2: A medium-capacity removable disk format. items from corresponding positions.
3: A data processing operation.

Overviews Business value proposition


(1) A Zip file is a single file that may be Zip files are a very convenient and uni-
copied, transmitted, or archived as a versally accepted format for archiving data
single unit, which may contain within for backup and for packaging collections of
it an unlimited number of other files, data and applications for electronic trans-
structured in folders and sub-folders. mission.
Depending on their content, these Zip-like disks filled an important niche
files may be very heavily compressed; in the business and personal computing
this makes zip files a popular form world, but have been left behind by tech-
of attachment for emails. Using stan- nological evolution.
dard software, the contents of a zip
file may be viewed and modified in
Summary of potentially negative issues
the familiar ‘‘drag-and-drop” windows
The Zip file system works through the
style, and available application program
use of proprietary software owned by the
interfaces allow applications to directly
PKware corporation, or a competing version
access sub-files from a zip file.
owned by WinZip International LLC. Until
(2) Zip disk is the trade name for the most
recently the two versions were completely
popular variety of removable disks, and
compatible, but with the addition of strong
has almost become a generic name for
encryption that ceased to be the case. In
that kind of product. They are small
2003, the PKware corporation filed for a US
devices in protective casings, similar in
patent on the zip specification, which it
concept to the floppy disk. There are
had previously published openly since its
various varieties of zip disks with dif-
invention in 1986.
ferent capacity, speed, and price con-
Zip disks are also proprietary; the intel-
figurations. For a period, zip disks
lectual property rights are owned by the
were a very popular medium for data
Iomega Corporation.
backup and transport, but they have
now largely been eclipsed by USB flash Associated terminology: Compression,
memory devices. Disk, Storage, Database.

363
Index

10 base 2, 100 base T, 139 assembler, 20


16 bit, 32 bit, 64 bit, 320 assigned name, internet, see ICANN, 174
asynchronous transfer mode, 186
abuse of computers, 182, 287 ATM (Asynchronous Transfer Mode), 186
ACM (Association for Computing Machinery), 3 attack
Ada, programming language, 271 brute force, 275
address, network, 102 denial of service, 99
administrator, database, 90 password file, 255, 256
ADSL (Asymmetric Digital Subscriber Line), 69 spoofing, 307
adware, 4 Trojan horse, 322
agent, artificially intelligent, 5 virus, 339
agile organization, 336 audio, 21, 23
Algol, programming language, 270 audit, data quality, 96
algorithm, 9 authority, certification, 105
complexity, speed, 73 autocode, programming language, 268
efficiency, 124 AVI, video format, 330
Amdahl’s law, 56
analysis, of web site, 59 B-tree, database index, 176
anonymous FTP, 159 B2B (Business to Business), 123
ANSI C, programming language, 47 B2C (Business to Customer), 123
anti-Virus, 11 backbone, internet, 189
Applet, Java, 4 backup, 24
application, 167 bandwidth, 26, 27
architecture bar code, 29
client--server, 62 Basic, programming language, 342,
hardware, network, software, 16 343
legacy, 203 batch processing, 30
two-tier, three-tier, 14 BCS (British Computer Society), 31
archive, 24 benchmark, 32
ARPANET, see Internet, 333 binary, 33
artificial intelligence, 17 biometric, 34
agents, 5 bit, 26, 35, 41
fuzzy logic, 196 16, 32, and 64 bit, 319
instant messaging, 181 block, 66
knowledge-based systems, 196 blog, 354
logic programming, 205 blu-ray, optical storage, 246
machine learning, 207 blue screen of death, 241
natural language processing, 221 bluetooth, 36
neural network, 229 bot, 5
robotics, 285 bridge, network device, 185, 228
ASCII, 19 British Computer Society, 31
ASP broadband, 37
(Active Server Pages), 121 broadcast, network, 201
(Application Service Provider), 15 brute force attack, 275

365
Index

BS-15000, 178 CMM (Capability Maturity Model), 154


bug, 28, 40 CobiT, 291
bus, 40 Cobol, programming language, 67,
business 270
continuity, 41 COCOMO (Constructive Cost Model), 259
decision support system, 98 coding, see programming, 121
intelligence, 42 collision, network, 138
process, 44 Colossus, 157
byte, 212 commerce, electronic, 227
communicating sequential processes, 55
C, programming language, 46 communications
C#, programming language, 47 bandwidth, 27
C++, programming language, 47 interception and disclosure, 110
CA (Certification Authority), 105 middleware, 213
Cable Communications Policy Act 1984, 48 peer to peer, 258
cable modem, 37 community, on-line, 237
cache, 60, 210 compact disc, 245
CAD (Computer Aided Design), 51 compact flash, 150
CAM (Computer Aided Manufacturing), 51 compiler, 69
CAN-SPAM Act 2003, 296 compression, 73
capability maturity model, 154 audio, 22
capacity of disk, memory, 211 data, 75
capacity planning, 52 Zip, 363
cascading style sheets, 172 computability, 76
cash, digital, 103, 104 computer, 77
CCPA (Cable Communications Policy analog, 10
Act 1984), 48 CPU, 85
CD (Compact Disc), 147, 245 quantum, 277
cell computing, 54 computer aided design, CAD, 51
central processing unit, CPU, 85 computer aided manufacturing, CAM, 51
certificate, digital, 105 computer-based training, 79
certification authority, 105 computer society
CGI (Common Gateway Interface), 120 ACM, 3
chaos theory, 56 BCS, 31
chief information officer, CIO, 59 IEEE, 174
responsibilities of, see also security conceptual data model, 137
bandwidth assessment, 60 connectivity standard, 82
business continuity planning, 42 constructive cost model, 259
capacity planning, 52 continuity, business-service provider, 41
data backup, 42 cookie, 83
database administrator, 90 COPPA (Children’s Online Privacy Protection Act 1998),
distributed database, 115 59
information lifecycle management, 177 copyright, Digital Millennium Copyright Act
management information systems, 214 1998, 107
Sarbanes--Oxley Act 2002, 289 CORBA, 214
system architecture, 62 core, storage, 310
chief security officer, CSO, 296 COTS (Commercial Off The Shelf), 66
circuit switching, 250 CPU (Central Processing Unit), 85
CISC (Complex Instruction Set Computer), 85 cracking, 86, 87
click-stream tracking, 60 crash
clicks and bricks, 123 computer system, 40, 100
client disk, 114
application server, 14 cryptanalysis, 131
network access application, 61 CSCW (Computer Supported Cooperative Work), 164
networked computer, 353 CSO (Chief Security Officer), 296
clone, 64 CSS (Cascading Style Sheets), 172
cluster, 65 cybersquatting, 118

366
Index

data disclosure and interception of communications,


cluster, 66 110
compression, 63 disk, 111
cube, 98, 235 cache, 50
definition language, DDL, 88 compact, 245
encryption standard, DES, 131 DVD, 25, 245
flow diagram, 91 failure, 24
interchange, 126 file system, 146
mining, 93 optical, 147
pool, 68, 94 RAID, 278
Protection Act, 95 size, 211
quality audit, 96 storage, 24
storage, 310 striping, 278
synchronization hub, 68 zip, 363
warehouse, 97 diskless, 78, 145
database, 166 distributed
administrator, 90, 176 database, 115, 116
data interchange, 126 denial of service attack, 100
data mining, 93 processing, 115, 253
data pool, 94 divide and conquer, 316
data quality audit, 96 DNS (Domain Name Service), 117
data warehouse, 97 domain name, 117
distributed, 115 domain name service, 117
enterprise resource planning, 135 DoS (Denial of Service Attack), 100
entity relationship diagram, 137 DPA (Data Protection Act 1998), 95
extracting, transforming, and loading, 139 DSA (Digital Signature Algorithm), 109
index, 176 DSL (Digital Subscriber Line), 37
normalization, 231 DSS
on-line analytical processing, 235 (Decision Support System), 98
query language, 89 (Digital Signature Standard), 109
transaction, 321 DVD, 25, 245
DBMS (Database Management System), 88 DVD-RW, +RW, etc., 25
death, blue screen of, 40 Dvorak keyboard, 119
debugging, 21, 38 Dynamic Host Control Protocol, 101
decimal, 33 dynamic web content, 121, 172
decision support system, 98, 99
denial of service attack, 99 e-business, 123
DES (Data Encryption Standard), 131 e-commerce, 123, 349
development EDI (Electronic Data Interchange), 126, 127
end user, 133 EEPROM, 150
software, 299 efficiency, of processing method, 124
devices, network, 227 EIP (Enterprise Information Portal), 134
DFD (Data Flow Diagram), 91 EISA (Extended Industry Standard Architecture), 41
DHCP (Dynamic Host Control Protocol), 101 electro-magnetic interference, 265
Dhrystone, benchmark, 32 electronic
Diffie--Hellman, 255, 294 business, commerce, 123
digest, message, 236 mail, 306
digital, 102 electronic data interchange, 126
cash, 103 electronic signature, 109
certificate, 105 Eliza, 221
compression, 22 email, 118
signature, 107 spam, 304
video, 330 embedded system, 78
wallet, 110 EMI (Electro-magnetic Interference), 265
Digital Millennium Copyright Act, 107 encoding
digital signal processing, 22 cryptography, 241
digital signature, 107 digital, 74

367
Index

encryption, 131 generation, first to fourth, 157


one-way hash, 236 generation loss, 76
public key--private key, 274 gigabit
end user development, 133 bandwidth, 27
entanglement, 277 ethernet, 139
enterprise information portal, 134 gigabyte, 211
enterprise resource planning, 135 Global Positioning System, 163
entity relationship diagram, 137 GNU, 239
ERD (Entity Relationship Diagram), 137 GPS (Global Positioning System), 163
ergonomics, 170 grid computing, 54
ERP (Enterprise Resource Planning), 135 groupware, 164
estimation, software metrics, 301
ethernet, 254 hacker, 166
ETL (Extracting, Transforming, and Loading), 139 hardware, 24, 167
expert systems, 161 hash, one-way, 236
Export restrictions on cryptographic hash table, database index, 176
systems, 243 Haskell, programming language, 160
Extensible Markup Language, 359 HCI (Human--Computer Interaction), 170
extracting, transforming, and loading, 139 head crash, disk, 112
Health Insurance Portability and Accountability
FAT, file system, 147 Act 1996, 168
fat client, 64 HFS (file system), 147
FDDI (Fiber Distributed Data Interface), 143 HIPAA (Health Insurance Portability and Accountability
feed, RSS, 286 Act), 168
fiber optic, 143 Hoare logic, 153
file server, 144, 145 holographic storage, 246
file sharing, 259 home page, 354
file system, 146 host, 169
File Transfer Protocol, 158 host name, 169, 327
fingerprint, 35, 108 hosting, 169, 189
firewall, 148 HTML (HyperText Markup Language), 353
firmware, 167 Dynamic Content, 353
first generation, 157 HTTP (HyperText Transfer Protocol), 272, 353, 354
flash memory, 150 hub, 185, 227
flop, performance measure, 32 human--computer interaction, 170
floppy disk, 113 hyper link, 173
forecasting hypertext, 171, 353
capacity requirements, 54
weather, 55, 57 IANA (Internet Assigned Numbers Authority), 174
formal methods, 40, 151 IC (Integrated Circuit), 157
formatting, disk, 114 ICANN (Internet Corporation for Assigned Names and
Fortran, programming language, 155, 156 Numbers), 174, 183
four-layer model, 316 IEEE (Institute of Electrical and Electronic Engineers), 174
fourth generation, 156 ILM (Information Lifecycle Management), 177
fractals, 57 IM (Instant Messaging), 180
frame relay, 250 IMAP (E-mail Protocol), 128
fraud, Computer Fraud and Abuse Act 1986, 80 imperative programming style, 204
Free Software Foundation, 238 index, database, 176
freeBSD, 325 information lifecycle management, 177
front end, 213 information technology infrastructure library, 178
FSF (Free Software Foundation), 238 Institute of Electrical and Electronic Engineers, 174
FTP (File Transfer Protocol), 158 instant messaging, 180
functional programming, 159, 160 integrated circuit, 157
fuzzy logic, 161 intellectual property, 108
intelligence
gateway, network device, 185, 228 artificial, 17
GB (Gigabyte), 211 business, 42

368
Index

interactive voice response, 222 kilobit, 27


interception of communications, 110 kilobyte, 212
internet, 182, 183, 184 knowledge
address, 101, 117, 183, 186 based systems, 196, 197
advertising, 5 discovery, 93
hosting, 170 engineer, 198, 199
multicast, 219 management, 198
protocol, 185
internet service provider, 189 LAN (Local Area Network), 183, 201
telephony, 344 language
Internet 2, 184 Ada, 271
intranet, 201 Algol, 269
IOS (Inter-Organizational Systems), 127 Assembler, 20
IP autocode, 268
intellectual property, 166 Basic, 342
internet protocol, 185 C, 46
IP address, 101, 117, 183 C#, 47
ISA (Industry Standard Architecture), 37 C++, 84
ISAM (Indexed Sequential Access Method), 176 Cobol, 20, 269
ISDN (Integrated Services Digital Network), 37 computer, 77
island of automation, 44 Fortran, 155
ISO-1989 (Cobol Programming Language), 67 functional, 159
ISO-8859 (Extended ASCII), 19 Haskell, 160
ISO-9660 (file system), 147 Java, 191
ISO-9899 (C programming language), 47 JavaScript, 193
ISO-9945 (Posix), 326 Lisp, 160
ISO-14882 (C++ programming language), 47 Miranda, 160
ISO-17799 (security), 290, 296 Occam, 55
ISO-20000 (IT infrastructure library), 179 Pascal, 270
ISP (Internet Service Provider), 189 PL/I, 270
ITIL (Information Technology Infrastructure Prolog, 204
Library), 178 Visual Basic, 342
IVR (Interactive Voice Response), 222 Latex, 171
law
Jackson structured programming, 313 Amdahl’s, 55
JAD (Joint Application Design), 352 Cable Communications Policy Act 1984, 48
Java, 191 CAN-SPAM Act 2003, 296
virtual machine, 332 Children’s Online Privacy Protection Act 1998, 59
J2EE application server, 15 Computer Fraud and Abuse Act 1986, 150
JavaScript, 193, 194 Computer Misuse Act 1990, 81
Jaz disk, 113 Computer Security Act 1987, 82
job control language, 241 Cyber-Security Enhancement Act 2002, 111
joint application design, 194 Data Protection Act 1998, 95
JSP Digital Millennium Copyright Act 1998, 107
( Jackson Structured Programming), 313 Digital Signature and Electronic Authentication
( Java Server Pages), 121 Act 1998, 109
Jurassic Park, 56 Electronic Signature Directive 1998, 109
JVM (Java Virtual Machine), 193, 332 Electronic Signatures Act 2000, 109
Health Insurance Portability and Accountability
KB, memory size measurement, 212 Act 1996, 168
KBS (Knowlegde-Based Systems), 271 Moore’s, 86
DDD (Knowledge Discovery in Databases), 93 Privacy Act 1974, 266
kernel, 325 Privacy in Electronic Commerce, EU directive, 141
key Privacy Protection Act 1980, 267
session, 275 Regulation of Investigatory Powers Act 2000, 133
exchange, 275 Sarbanes--Oxley Act 2002, 289
keyboard, 119 Uniform Electronic TransActions Act 1999, 109

369
Index

law (cont.) multi-threading, 253


Unlawful Access to Stored Communications multicast, 219
Act, 110 multiprogramming, 241
Voluntary Disclosure of Communications, 110
Wiretap Act, 111 NAT (Network Address Translation), 226
learning, machine, 207 natural language processing, 221
legacy system, 203 network, 102, 223
lifecycle, software development -, 299 address, 102
line conditioning, 264 address translation, 226
Linux, 239 architecture, 17
Lisp, programming language, 160 bandwidth, 6
Lisp machine, 196 circuit switched, 344
Listserv, 129 devices, 227
load balancing, 65 DHCP, 101
local area network, 183, 201 domain name, 117
logic ethernet, 138
fuzzy, 161 file server, 144
programming, 204 file transfer protcol, 158
logical data model, 137 firewall, 148
internet, 182
machine intranet, 201
code, 20 IP, 185
intelligence, 17, 18 LAN, 201
language, 20 multicast, 219
learning, 207 neural, 207, 229
mail, electronic, 306 OSI seven-layer model, 247
mainframe, 49 packet switched, 344
maintenance, 208 peer to peer, 258
man-month, 259 port, 262
management information systems, 7, 214 private, 337
master data registry, 94 proxy, 273
MB (Megabyte), 210 server, 297
MD5 (Message Digest Algorithm), 237 subnet, 201
measurement of computing power, 32 TCP/IP, 315
megabit, 27 value added, 329
megabyte, 210 virtual, 264
megaflop, performance measurement, 32 voice, 22
magnetic WAN, 201
disk, 111 web services, 348
tape, 25 wireless, 351
magneto-optical disk, 246 Wireless Application Protocol, 349
memory, 209 World Wide Web, 353
flash, 150 X.25, 359
size, 212 neural net, 229, 230
stick, 211 NFS (Network File Server), 145
storage, 310 NLP (Natural Language Processing), 221
virtual, 333 normal form, 232
message, instant, 180 normalization, 231
metrics, software, 301 NTFS (file system), 147
middleware, 213
Miranda, programming language, 160 object oriented, 233
mirroring, disk, 278 C++, 47
MIS (Management Information Systems), 214 Java, 191
modem, 215 UML, 324
Moore’s law, 86 Occam, programming language, 55
motherboard, 217 OCR (Optical Character Recognition), 243
MPEG (Motion Picture Experts Group), 330 OLAP (On-Line Analytical Processing), 235

370
Index

OLTP (On-Line Transaction Processing), 321 Institute of Electrical and Electronic


on-line analytical processing, 235 Engineers, 174
on-line community, 237 program, 209
one-way hash, 236 programming
open source, 239 algorithm, 204
operating system, 114, 240 application development methods, 12
optical storage, 244 application generator, 14
optical character recognition, 243 choice of method, 185
OSI compiler, 155
(Open Source Initiative), 239 formal methods, 314
(Open Systems Interconnection), 247 logic programming, 204
outsourcing, 248 programming langauge, 268
Ada, 269
package, software, 302 Algol, 269
packet, 250 assembler, 20
packet sniffer, 254 autocode, 268
packet switching, 250 Basic, 269
page, home, 354 C, 46
palmtop, 77 C#, 46
parallel port, 263 C++, 46
parallel processing, 251, 253 Cobol, 67, 209
partition Fortran, 20, 155
database, 117 functional, 159
disk, 114 Haskell, 160
Pascal, programming language, 270 Java, 191
password, 86, 256 JavaScript, 193
patent, 257 Lisp, 270
PCI (Peripheral Component Interconnect), 41 logic, 186
PDA (Portable Digital Assistant), 77 Miranda, 160
peer to peer, 258, 337 object oriented, 233
performance Occam, 55
of computational method, 14, 321 Pascal, 270
of process, 72 PL/I, 270
person month, 259 Prolog, 204
phishing, 260 Visual Basic, 269, 342
physical data model, 137 Prolog, 196
piconet, 36 protection, see security
PL/I, programming language, 270 Protocol, 272
POP (Post Office Protocol), 128 connectivity standard, 82
port, 262 DHCP, 101
portal, enterprise information, 134 FTP, 158
Posix, 326 HTTP, 259, 272, 327
power protection, 264 IMAP, 128
primary domain controller, 298 IP, 185
privacy (see also law and security), 141 POP, 129
Privacy Act 1974, 202 SMTP, 128
Privacy Protection Act 1980, 202 TCP/IP, 315
private key, 274 WAP, 349
procedural programming style, 160 X.12, 358
process, re-engineering, 45 proxy, 273
process modeling, 92 public key, 105
processing efficiency, 145
productivity, of programmer, 160 quantization effect, 21
professional association quantum computing, 277
Association for Computing Machinery, 3 quarantine, 11
Association for Information Systems, 7 query language, 89
British Computer Society, 31 QWERTY, keyboard, 119

371
Index

RAD (Rapid Application Development), 12 data warehouse, 97


radio frequency identity tag, 282 database, 14
RAID (Redundant Array of Inexpensive Disks), 26, 278 farm, 65
RAM (Random Access Memory), 245, 310 file, 144
rapid application development, 279 network address translation, 226
really simple syndication, 286 networked application, 316
redundant array of inexpensive disks, 26 proxy, 183
relational database, 89 services
reliability, 280 application service provider, 15
RFID (Radio Frequency Identity Tag), 282 business continuity service provider, 41
RISC (Reduced Instruction Set Computer), 85 hosting, 189
robotics, 285 web, 630
ROM (Read-Only Memory), 311 session key, 275
router, 224 SET (Secure Electronic Transaction), 294
RSS (Really Simple Syndication), 286 SETI at home, 253
seven-layer model, 316
Sarbanes-Oxley Act 2002, 289 SHA (Secure Hash Algorithm), 256
scalability, 292 shell
scripting, 121, 193, 269 operating system, 241
SCSI (Small Computer Systems Interface), 263 secure, 318
SEAL (Digital Signature and Electronic Authentication SHTML (Server Parsed HTML), 120
Act 1998), 109 signature, digital, 107, 275
search, database, 135 SIIA (Software and Information Industry Association),
second generation, 157 298
secure, 293 SIPERNET, 184
digital, 150 small computer systems interface, 263
electronic transaction, 294 Smalltalk, programming language, 234
shell, 318 SMTP (Simple Mail Transfer Protocol), 128
sockets layer, 294, 350 software, 304
security, 295 capability maturity model, 300
Children’s Online Privacy Protection Act 1998, 59 compiler, 20
Computer Security Act 1987, 202 development (see also programming), 13, 55
cracking, 86 data flow diagram, 313
denial of service attack, 99 formal methods, 40
digital cash, 103 joint application design, 280
digital certificate, 351 object oriented, 137
digital signature, 237 rapid application development, 279
digital wallet, 110 structured, 67
encryption, 307 UML, 324
encryption keys, 86 walk-through review, 346
firewall, 42 waterfall model, 347
hacking, 167 maintenance, 209
ISO-17799, 290 metrics, 301
one-way hash, 107 middleware, 213
password, 257 package, 302
phishing, 261 publishers association, 299
scripting, 121 reliability, 282
secure, 293 scalability, 293
spoofing, 307 sound, 10
spyware, 99 spam, 304
Trojan horse, 99 special interest group
virus, 339 Association for Computing Machinery, 3
seek time, disk, 112 Association for Information Systems, 7
serial port, 262 speed
server, 297 of computational method, 14, 133
application server, 14 of data access, disk, 112
computer, 228 of process, 254

372
Index

spoofing, 307 Unicode, 19


spyware, 308 Unified Modeling Language, 93
advertising, 4 uniform product code, 29
Computer Misuse Act 1990, 202 uniform resource locator, 326
SQL (Structured Query Language), 89 uninterruptible power supply, 265
squatting, domain name, 118 Unix, 324, 326
SSH (Secure Shell), 318 UPC (Uniform Product Code), 29
SSL (Secure Sockets Layer), 255, 318 UPS (Uninterruptible Power Supply), 265
standard template library, 47 URL (Uniform Resource Locator), 326
star topology, network, 224 USB
step-wise refinement, 313 key drive, 275
STL (Standard Template Library), 47 port, 114, 263
storage, 111
capacity, 77 vacuum tube, 157
disk, 114 validation and verification, 188, 347
flash memory, 150 value added network, 593
memory, 310 VAN (Value added network), 329
size, 292 video, 329
optical, 244 virtual
RAID, 278 Java virtual machine, 332
service provider, 178 machine, 331
striping, disk, 278 memory, 333
structured design, 300 organization, 335
style sheet, 172 private network, 336
subnet, 201 reality, 338
superposition, 277 virus, 11, 339
surge protection, 264 Visual Basic, 121, 269, 342
switch, 93, 224 VLSI (Very Large Scale Integration), 158
VM
T-1 line, 315 (Virtual Machine), 191
T-carrier, 315 (Virtual Memory), 333
tape, magnetic, 25 Voice over IP, 182
TB (terabyte), 212 VoIP (Voice over IP), 343
TCP/IP, 315 VPN (Virtual Private Network), 608
telephony, internet, 344 VR (Virtual Reality), 338
Telnet, 317
terabyte, 116 W3C (World Wide Web Consortium), 346
text mining, 93 walk-through review, 346
thin client, 63, 114 wallet, digital, 110
third generation, 157 WAN (Wide Area Network), 251
thread, 253 WAP (Wireless Application Protocol), 349
time sharing, 334 waterfall model, 347
token ring network, 224 Watfor Fortran, 155
topology, network, 225 web, 355
TPS (Transaction Processing System), 321 address, 169
training, computer-based, 79 hosting, 15, 170
transaction, 358, 360 hypertext, 353
transaction set, x.12, 358 ICANN, 174
transistor, 157 internet, 185
transport control protocol, 315 page, 51, 350
transputer, 55 phone, 344
Trojan horse, 322 services, 348
Turing test, 17 site analysis, 112
URL, 327
UDP (User Datagram Protocol), 316 webcast, 23
UFS (Unix File System), 147 wide area network, 201
UML (Unified Modeling Language), 324 wiki, 354

373
Index

Winbench, benchmark, 32 X.12, 127, 358


Windows operating system, 146 X.25, 359
winstone, benchmark, 32 XML (Extensible Markup Language), 359
wire tap, 111
Wireless Application Protocol, 349 Y2K (Year 2000 Problem), 362
wireless network, 36, 351 year 2000 problem, 362
WML (Wireless Markup Language), 350 Yourdon design method, 313
word, unit of memory, 20
World Wide Web, 348 Z, 153
World Wide Web Consortium, 360 zip, 43, 363
WYSIWYG (What You See Is What You Get), 356 zombie computer, 100

374

S-ar putea să vă placă și