936 lines
50 KiB
Plaintext
936 lines
50 KiB
Plaintext
|
||
|
||
Computer underground Digest Wed Oct 21 1993 Volume 5 : Issue 82
|
||
ISSN 1004-042X
|
||
|
||
Editors: Jim Thomas and Gordon Meyer (TK0JUT2@NIU.BITNET)
|
||
Archivist: Brendan Kehoe
|
||
Shadow-Archivists: Dan Carosone / Paul Southworth
|
||
Ralph Sims / Jyrki Kuoppala
|
||
Ian Dickinson
|
||
Copy Ediort: Etaoin Shrdlu, III
|
||
|
||
CONTENTS, #5.82 (Oct 21 1993)
|
||
File 1--Fair Info Practices with Comp. Supported Coop Work
|
||
File 2--LA Times does cyphertech; odds & ends
|
||
File 3--IGC Wins Social Responsibility Award
|
||
File 4--Full Description of Proposed "Hacker" Documentary"
|
||
|
||
Cu-Digest is a weekly electronic journal/newsletter. Subscriptions are
|
||
available at no cost electronically from tk0jut2@mvs.cso.niu.edu. The
|
||
editors may be contacted by voice (815-753-0303), fax (815-753-6302)
|
||
or U.S. mail at: Jim Thomas, Department of Sociology, NIU, DeKalb, IL
|
||
60115.
|
||
|
||
Issues of CuD can also be found in the Usenet comp.society.cu-digest
|
||
news group; on CompuServe in DL0 and DL4 of the IBMBBS SIG, DL1 of
|
||
LAWSIG, and DL1 of TELECOM; on GEnie in the PF*NPC RT
|
||
libraries and in the VIRUS/SECURITY library; from America Online in
|
||
the PC Telecom forum under "computing newsletters;"
|
||
On Delphi in the General Discussion database of the Internet SIG;
|
||
on the PC-EXEC BBS at (414) 789-4210; and on: Rune Stone BBS (IIRG
|
||
WHQ) (203) 832-8441 NUP:Conspiracy; RIPCO BBS (312) 528-5020
|
||
CuD is also available via Fidonet File Request from 1:11/70; unlisted
|
||
nodes and points welcome.
|
||
EUROPE: from the ComNet in LUXEMBOURG BBS (++352) 466893;
|
||
In ITALY: Bits against the Empire BBS: +39-461-980493
|
||
|
||
ANONYMOUS FTP SITES:
|
||
AUSTRALIA: ftp.ee.mu.oz.au (128.250.77.2) in /pub/text/CuD.
|
||
EUROPE: nic.funet.fi in pub/doc/cud. (Finland)
|
||
UNITED STATES:
|
||
aql.gatech.edu (128.61.10.53) in /pub/eff/cud
|
||
etext.archive.umich.edu (141.211.164.18) in /pub/CuD/cud
|
||
ftp.eff.org (192.88.144.4) in /pub/cud
|
||
halcyon.com( 202.135.191.2) in /pub/mirror/cud
|
||
ftp.warwick.ac.uk in pub/cud (United Kingdom)
|
||
|
||
COMPUTER UNDERGROUND DIGEST is an open forum dedicated to sharing
|
||
information among computerists and to the presentation and debate of
|
||
diverse views. CuD material may be reprinted for non-profit as long
|
||
as the source is cited. Authors hold a presumptive copyright, and
|
||
they should be contacted for reprint permission. It is assumed that
|
||
non-personal mail to the moderators may be reprinted unless otherwise
|
||
specified. Readers are encouraged to submit reasoned articles
|
||
relating to computer culture and communication. Articles are
|
||
preferred to short responses. Please avoid quoting previous posts
|
||
unless absolutely necessary.
|
||
|
||
DISCLAIMER: The views represented herein do not necessarily represent
|
||
the views of the moderators. Digest contributors assume all
|
||
responsibility for ensuring that articles submitted do not
|
||
violate copyright protections.
|
||
|
||
----------------------------------------------------------------------
|
||
|
||
Subject: File 1--Fair Info Practices with Comp. Supported Coop Work
|
||
Date: Wed, 20 Oct 1993 09:54:21 -0700
|
||
From: Rob Kling <kling@ICS.UCI.EDU>
|
||
|
||
Fair Information Practices with Computer Supported Cooperative Work
|
||
|
||
Rob Kling
|
||
|
||
Department of Information & Computer Science
|
||
and
|
||
Center for Research on Information Technology and Organizations
|
||
University of California at Irvine,
|
||
Irvine, CA 92717, USA
|
||
kling@ics.uci.edu
|
||
|
||
May 12, 1993 (v. 3.2)
|
||
|
||
Based on a paper which appears in SIGOIS Bulletin, July 1993
|
||
|
||
+++++++++++++
|
||
The term "CSCW" was publicly launched in the early 1980s. Like other
|
||
important computing terms, such as artificial intelligence, it was coined
|
||
as a galvanizing catch-phrase, and given substance through a lively stream
|
||
of research. Interest quickly formed around the research programs, and
|
||
conferences identified with the term advanced prototype systems, studies of
|
||
their use, key theories, and debates about them. CSCW offers special
|
||
excitement: new concepts and possibilities in computer support for work.
|
||
|
||
CSCW refers to both special products (groupware), and to a social movement
|
||
by computer scientists who want to provide better computer support for
|
||
people, primarily professionals, to enhance the ease of collaborating.
|
||
Researchers disagree about the definition of CSCW, but the current
|
||
definitions focus on technology. I see CSCW as a conjunction of certain
|
||
kinds of technologies, certain kinds of users (usually small self-directed
|
||
professional teams), and a worldview which emphasizes convivial work
|
||
relations. These three elements, taken together, differentiate CSCW from
|
||
other related forms of computerization, such as information systems and
|
||
office automation which differ as much in their typical users and the
|
||
worldview describing the role of technology in work, as on the technology
|
||
itself (Kling, 1991). CSCW is the product of a particular computer-based
|
||
social movement rather than simply a family of technologies (Kling and
|
||
Iacono, 1990).
|
||
|
||
The common technologies that are central to CSCW often record fine grained
|
||
aspects of people activities in workplaces, such as typed messages, notes,
|
||
personal calendar entries, and videotapes of personal activity. Electronic
|
||
mail is the most popular of the CSCW technologies (Bullen and Bennett,
|
||
1991) and is a useful vehicle for examining some of the privacy issues in
|
||
CSCW. Many electronic mail messages contain personal communications which
|
||
include opinions and information which many senders would prefer not to be
|
||
public information. However, most electronic mail system users I have
|
||
spoken to are ignorant of the conditions under which their transmissions
|
||
will be maintained as private communications by their own organizations.
|
||
(They often assume that their electronic communications will be treated as
|
||
private by their organizations. Others are extremely sensitive to the
|
||
possible lack of privacy/security of email transmissions.)
|
||
|
||
Discussions of computerization and privacy are highly developed with
|
||
respect to personal record systems which contain information about banking,
|
||
credit, health, police, schooling, employment, insurance, etc. (Kling and
|
||
Dunlop, 1991:Section V). Definitions of personal privacy have been examined
|
||
in extensive literature about personal privacy and record-keeping systems.
|
||
Analysts have been careful to distinguish security issues (e.g., lock and
|
||
keys for authorized access) from privacy issues -- those which involve
|
||
people's control over personal information. There has also been significant
|
||
discussion of the interplay between privacy and other competing social
|
||
values. The privacy issues in CSCW both have important similarities and
|
||
differences when compared with the issues of personal record systems. We
|
||
can gain helpful insights by building on this body of sustain thinking
|
||
about privacy and record systems to advance our understanding of privacy
|
||
issues in CSCW.
|
||
|
||
Another related and helpful set of inquiries examines the surveillance of
|
||
workers in measuring activities related to quality of service and
|
||
individual productivity (Attewell, 1991; Kling and Dunlop, 1993). Some of
|
||
the most intensive fine grained electronic monitoring involves listening to
|
||
the phone calls of service workers such as reservationists, and
|
||
fine-grained productivity counts, such as the number of transactions that a
|
||
worker completes in a small time period. While all managers have ways of
|
||
assessing their subordinates' performance, clerks are most subject to these
|
||
fine grained forms of electronic surveillance. The CSCW community has
|
||
focussed on professionals as the key groups to use groupware and meeting
|
||
support systems. Consequently, electronic monitoring has seemed to be
|
||
implausible.
|
||
|
||
The computing community is beginning to be collectively aware of the
|
||
possible privacy issues in CSCW applications. Professionals who use CSCW
|
||
can lose privacy under quite different conditions than clerks who have
|
||
little control over the use of electronic performance monitoring systems.
|
||
And personal communications, like electronic mail or systems like gIBIS
|
||
which supports debates, record personally sensitive information under very
|
||
different conditions than do information systems for regulatory control
|
||
such as systems of motor vehicle, health and tax records.
|
||
|
||
The use of email raises interesting privacy issues. In the case of email,
|
||
privacy issues arise when people lose control over the dissemination of
|
||
their mail messages. When should managers be allowed to read the email of
|
||
their subordinates? One can readily conjure instances where managers would
|
||
seek access to email files. These can range from curiosity (such as when a
|
||
manager wonders about subordinates' gossip, and requests messages which
|
||
include his name in the message body), through situations in which a legal
|
||
agency subpoenas mail files as part of a formal investigation. A
|
||
different, but related set of issues can occur when a manager seeks mail
|
||
profiles: lists of people who send more than N messages a day, lists of
|
||
people who read a specific bulletin board or the membership of a specific
|
||
mailing list.
|
||
|
||
CSCW systems differ in many ways that pertain to informational control. For
|
||
example, systems such as email and conferencing systems retain electronic
|
||
information which can be reused indefinitely with little control by the
|
||
people who were writing with the system. One can imagine cases in which
|
||
managers may wish to review transcripts of key meetings held by computer
|
||
conferencing to learn the bases of specific decisions, who took various
|
||
positions on controversial issues, or to gain insight into their
|
||
subordinate's interactional styles. Other systems, such as voice and video
|
||
links, are often designed not to store information. But they can raise
|
||
questions about who is tuning in, and the extent to which participants are
|
||
aware that their communication systems is "on." In the literature about
|
||
computerization and privacy, similar questions have been closely examined
|
||
-- regulating the duration of records storage, the conditions under which
|
||
people should be informed that a third party is seeking their records, and
|
||
conditions under which individuals may have administrative or legal
|
||
standing in blocking access to their records (See Dunlop and Kling, 1991,
|
||
Section V).
|
||
|
||
One of the peculiarities of CSCW in contrast with traditional record
|
||
keeping systems is the nature of the social settings in which systems are
|
||
being developed and explored. Most personal record systems are developed in
|
||
relatively traditional control-oriented organizations. In contrast, most
|
||
CSCW applications have been developed in academic and industrial research
|
||
labs. These settings are protective of freedom of speech and thought and
|
||
less authoritarian than many organizations which ultimately use CSCW
|
||
applications. In fact, relatively few CSCW applications, other than email
|
||
and Lotus Notes, are used by the thousands of people in traditional
|
||
organizations (Bullen and Bennett, 1991). Further, CSCW systems are
|
||
primarily designed to be used by professionals rather than technicians and
|
||
clerks. Professionals generally have more autonomy than clerks, who are
|
||
most subject to computerized monitoring (Attewell, 1991). As a consequence,
|
||
many CSCW developers don't face problems of personal privacy that may be
|
||
more commonplace when prototype systems are commercialized and widely used.
|
||
|
||
These contrasts between R&D with CSCW and the likely contexts of
|
||
application should not impede us from working hard to understand the
|
||
privacy issues of these new technologies. CSCW applications are able to
|
||
record more fine grained information about peoples' thoughts, feelings, and
|
||
social relationships than traditional record keeping systems. They can be
|
||
relatively unobtrusive. The subject may be unaware of any scrutiny. In R&D
|
||
labs, we often have norms of reciprocity in social behavior: monitoring can
|
||
be reciprocal. However, in certain organizations, monitoring may follow a
|
||
formal hierarchy of social relations. For example, supervisors can monitor
|
||
the phone conversations of travel reservationists and telephone operators,
|
||
but the operators cannot monitor their supervisors. The primary
|
||
(publicized) appropriations of "private email" have been in military
|
||
organizations, NASA, and commercial firms like Epson, rather than in
|
||
university and industrial laboratories.
|
||
|
||
CSCW creates a new electronic frontier in which people's rights and
|
||
obligations about access and control over personally sensitive information
|
||
have not been systematically articulated. I believe that we need to better
|
||
understand the nature of information practices with regard to different
|
||
CSCW applications that balance fairness to individuals and to their
|
||
organizations.
|
||
|
||
It is remarkable how vague the information practices regulating the use of
|
||
the few commonplace CSCW applications are. Yet we are designing and
|
||
building the information infrastructures for recording significant amounts
|
||
of information about people thoughts and feelings which are essentially
|
||
private and not for arbitrary circulation, without the guidelines to
|
||
safeguard them. People who use computer and telecommunications applications
|
||
need to have a basic understanding about which information is being
|
||
recorded, how long it is retained (even if they "delete" information from
|
||
their local files, who can access information about them, and when they can
|
||
have some control over restricting access to their information.
|
||
|
||
In the late 1970s the U.S. Privacy Protection Study Commission developed a
|
||
set of recommendations for Fair Information Practices pertinent to personal
|
||
record keeping systems (PPSC, 1977:17-19). A concern of Commission members
|
||
was to maximize the extent to which record systems would be managed so that
|
||
people would not be unfairly affected by decisions which relied upon
|
||
records which were inaccurate, incomplete, irrelevant or not timely.
|
||
Commission members believed that record keeping systems in different
|
||
institutional settings should be regulated by different laws. For example,
|
||
people should have more control over the disclosure of their current
|
||
financial records than over the disclosure of their current police records.
|
||
On the other hand, the Commission proposed that each institutional arena
|
||
should be governed with an explicit set of Fair Information Practices. In a
|
||
similar way, different families of CSCW applications or different
|
||
institutional settings may be most appropriately organized with different
|
||
Fair Information Practices. In the case of CSCW applications, fairness may
|
||
have different meanings than in the case of decisions based upon personal
|
||
records systems.
|
||
|
||
We need fearless and vigorous exploratory research to shed clear light on
|
||
these issues. This rather modest position contrasts strongly with that
|
||
taken by Andy Hopper of Olivetti, one of the panelists at this plenary
|
||
session on CSCW'92. He was enthusiastic about the use of "active badges"
|
||
(Want, Hopper, Falcao, and Gibbons, 1992) and insisted on discussing only
|
||
their virtues. He argued that one can imagine many scenarios in which
|
||
people are harmed by some uses of a particular technology, but that
|
||
discussing such scenarios is usually pointless. Hopper's 1992 co-authored
|
||
article about active badges examines some of the privacy threats their use
|
||
can foster. But on the plenary panel he was critical of people who asked
|
||
serious questions about the risks, as well as the benefits of new CSCW
|
||
technologies. In this way, he took a position similar to that taken by
|
||
spokespeople of many industries, including such as automobiles, who have
|
||
delayed serious inquiries and regulatory protections for environmental and
|
||
safety risks by insisting on unambiguous evidence of harm before
|
||
investigating plausible problems.
|
||
|
||
The active badge systems which Hopper described seem to be regulated by
|
||
Fair Information Practices in his own research laboratory (e.g., no long
|
||
term storage of data about people's locations, reciprocity of use,
|
||
discretion in use). These sorts of Fair Information Practices may be
|
||
required to help insure that active badges are a convenient technology
|
||
which do not degrade people's working lives. Other kinds of information
|
||
practices, such as those in which location monitoring is non-reciprocal,
|
||
and non-discretionary may help transform some workplaces into electronic
|
||
cages. Hopper and his colleagues briefly mention such possibilities in
|
||
their 1992 ACM TOIS article about active badges. And their article deserves
|
||
some applause for at least identifying some of the pertinent privacy
|
||
problems which active badges facilitate. However they are very careful to
|
||
characterize fine grained aspects of the technological architecture of
|
||
active badges, while they are far from being comparably careful in
|
||
identifying the workplace information practices which can make active
|
||
badges either primarily a convenience or primarily invasive. I believe that
|
||
CSCW researchers should be paying careful attention to social practices as
|
||
well as to technologies. Richard Harper's (1992) ethnographic study of the
|
||
use of active badges in two research labs illustrates the kind of nuanced
|
||
analyses which we need, although Harper also glosses the particular
|
||
information practices which accompanied the use of active badges in the two
|
||
labs.
|
||
|
||
Unfortunately, delays in understanding some risks of emerging technologies
|
||
have led the public to underestimate the initial magnitude of problems, and
|
||
to make collective choices which proved difficult alter. Our design of
|
||
metropolitan areas making individually operated cars a virtual necessity is
|
||
an example. In the early stages of use, the risks of a new family of
|
||
technologies are often hard to discern (See Dunlop and Kling, 1991, Part
|
||
VI). When major problems develop to the point that they are undeniable,
|
||
amelioration may also be difficult.
|
||
|
||
I characterized CSCW, in part, as a social movement (Kling and Iacono,
|
||
1990). Most of us who study, develop, or write about CSCW enthusiastically,
|
||
(and sometimes evangelistically) encourage the widespread use of these new
|
||
technologies. However, as responsible computer scientists, we should temper
|
||
our enthusiasms with appropriate professional responsibility. CSCW
|
||
applications open important organizational opportunities, but also opens
|
||
privacy issues which we don't understand very well.
|
||
|
||
The new ACM Ethical Code (ACM, 1993) also has several provisions which bear
|
||
on privacy issues in CSCW. These include provisions which require ACM
|
||
members to respect the privacy of others (Section 1.7), to improve public
|
||
understanding of computing and its consequences (Section 2.7), and to
|
||
design and build information systems which enhance the quality of working
|
||
life (Section 3.2). The ACM's code is rather general and does not give much
|
||
specific guidance to practitioners. The CSCW research community is well
|
||
positioned to conduct the kinds of research into the social practices for
|
||
using these technologies which could shape meaningful professional
|
||
guidelines for their use in diverse organizations. Will we take a
|
||
leadership role in helping to keep CSCW safe for users and their
|
||
organizations?
|
||
|
||
=================================
|
||
Note: I appreciate discussions with Jonathan Allen, Paul Forester, Beki
|
||
Grinter, and Jonathan Grudin which helped clarify some of my key points.
|
||
|
||
|
||
REFERENCES
|
||
|
||
1. Association of Computing Machinery. 1993. "ACM Code of Ethics and
|
||
Professional Conduct." Communications of the ACM. 36(2)(Feb.):99-103.
|
||
|
||
2. Attewell, Paul. "Big Brother and the Sweatshop: Computer
|
||
Surveillance in the Automated Office" in Dunlop and Kling 1991.
|
||
|
||
3. Bullen, Christine and John Bennett. 1991. Groupware in Practice: An
|
||
Interpretation of Work Experience" in Dunlop and Kling 1991.
|
||
|
||
4. Dunlop, Charles and Rob Kling (Ed). 1991. Computerization and
|
||
Controversy: Value Conflicts and Social Choices. Boston: Academic
|
||
Press.
|
||
|
||
5. Harper, Richard H.R. "Looking at Ourselves: An Examination of the
|
||
Social Organization of Two Research Laboratories" Proc. CSCW '92:
|
||
330-337.
|
||
|
||
6. Kling, Rob. 1991. "Cooperation, Coordination and Control in
|
||
Computer-Supported Work." Communications of the ACM
|
||
34(12)(December):83-88.
|
||
|
||
7. Kling, Rob and Charles Dunlop. 1993. "Controversies About
|
||
Computerization and the Character of White Collar Worklife." The
|
||
Information Society. 9(1) (Jan-Feb:1-29.
|
||
|
||
8. Kling, Rob and Suzanne Iacono. 1990. "Computerization Movements"
|
||
Chapter 19, pp 213-236 Computers, Ethics and Society, David Ermann,
|
||
Mary Williams & Claudio Guitierrez (ed.) New York, Oxford University
|
||
Press.
|
||
|
||
9. Privacy Protection Study Commission. 1977. Personal Privacy in an
|
||
Information Society, U.S. Government Printing Office, Washington D.C.
|
||
(briefly excerpted in Dunlop and Kling, 1991.)
|
||
|
||
10.Want, Roy, Andy Hopper, Veronica Falcao and Jonathan Gibbons. 1992.
|
||
"The Active Badge Location System" ACM Transactions on Information
|
||
Systems. 10(1)(January): 91-102.
|
||
|
||
------------------------------
|
||
|
||
Date: 05 Oct 93 03:09:50 EDT
|
||
From: Urnst Kouch <70743.1711@COMPUSERVE.COM>
|
||
Subject: File 2--LA Times does Cyphertech; odds & ends
|
||
|
||
(MODERATORS' NOTE: Urnst Kouch is editor of Cyrpt Newsletter, a 'Zine
|
||
specializing in techno-political commentary, satire, and virus
|
||
information)).
|
||
|
||
CuD readers might want to look for the October 3 and 4 issues of The
|
||
L.A. Times. In a two-part series, the paper's "Column One" was devoted
|
||
to privacy/cryptography issues.
|
||
|
||
"Demanding the Ability to Snoop: Afraid new technology may foil
|
||
eavesdropping efforts, U.S. officials want phone and computer users to
|
||
adopt the same privacy code. The government would hold the only key"
|
||
was the title and subhead of Robert Lee Hotz's 60+ inch piece.
|
||
|
||
Hotz focused on the Clipper/Skipjack end of the story, in part,
|
||
because Mykotronx, Inc., the manufacturer of the chip for the National
|
||
Security Agency, is based in Torrance, Los Angeles County. The
|
||
newspiece did not delve into any of the recent events surrounding
|
||
Pretty Good Privacy and Phil Zimmerman. Pretty Good Privacy was
|
||
referred to as "one of the best codes . . . free and [it] can be
|
||
downloaded from computer network libraries around the world"; the
|
||
people who make up the citizen-supported cryptography movement as
|
||
"ragtag computerzoids."
|
||
|
||
The L.A. Times series also included statistics documenting the steady
|
||
rise in court-ordered wiretapping from 1985 to 1992 and the almost
|
||
100% increase in phones monitored by pen registers - which record
|
||
outgoing numbers - from 1,682 (1987) to 3,145 in 1992. These numbers
|
||
do not include monitoring by such as the NSA and said so.
|
||
|
||
Whitford Diffie earned a boxed-out quote, too. "Recent years have seen
|
||
technological developments that diminish the privacy available to the
|
||
individual. Cameras watch us in the stores, X-ray machines search us
|
||
at the airport, magnetometers look to see that we are not stealing
|
||
from the merchants, and databases record our actions and
|
||
transactions."
|
||
|
||
The October 3 installment wrapped up with this succint bit from
|
||
Diffie: "Cryptography is perhaps alone in its promise to give us more
|
||
privacy rather than less."
|
||
|
||
Moving on from The L.A. Times, readers could find interesting the
|
||
following hodgepodge of facts, which taken together, lend some
|
||
historical perspective to the continuing conflict between privately
|
||
developed cryptography and the government.
|
||
|
||
For example, in reference to the Clipper chip, take the old story of
|
||
Carl Nicolai and the Phasorphone.
|
||
|
||
In 1977 Nicolai had applied for a patent for the Phasorphone telephone
|
||
scrambler, which he figured he could sell for $100 - easily within the
|
||
reach of John Q. Public. For that, the NSA slapped a secrecy order on
|
||
him in 1978. Nicolai subsequently popped a nut, took his plight to
|
||
the media, and charged in Science magazine that "it appears part of a
|
||
general plan by the NSA to limit the freedom of the American people .
|
||
. . They've been bugging people's telephones for years and now
|
||
someone comes along with a device that makes this a little harder to
|
||
do and they oppose this under the guise of national security."
|
||
|
||
The media went berserk on the issue and the NSA's Bobby Ray Inman
|
||
revoked the Phasorphone secrecy order. If the cypherpunks have a
|
||
spiritual Godfather, or need a likeness to put on a T-shirt, Carl
|
||
Nicolai and his Phasorphone could certainly be candidates.
|
||
|
||
About the same time, Dr. George Davida of the University of Wisconsin
|
||
was also served with a NSA secrecy order, in response to a patent
|
||
application on a ciphering device which incorporated some advanced
|
||
mathematical techniques.
|
||
|
||
Werner Raum, chancellor of the University of Wisconsin's Milwaukee
|
||
campus, promptly denounced the NSA for messing with faculty academic
|
||
freedom. The Agency backed off.
|
||
|
||
Both setbacks only made the NSA more determined to exert ultimate
|
||
control over cryptography. In an interview in Science magazine the
|
||
same year, Bobby Inman stated that he would like to see the NSA
|
||
receive the same authority over cryptology that the Department of
|
||
Energy reserved for research which could be applied to atomic weapons,
|
||
according to James Bamford's "The Puzzle Palace." "Such authority
|
||
would grant to NSA absolute 'born classified' control over all
|
||
research in any way related to cryptology," reads his book.
|
||
|
||
Readers have also seen the acronym ITAR - for International Traffic in
|
||
Arms Regulation - used a lot in reference to the government's interest
|
||
in controlling private cryptography. ITAR springs from the Arms
|
||
Export Control Act of 1976, in which "The President is authorized to
|
||
designate those items which shall be considered as defense articles
|
||
and defense services." ITAR contains the U.S. Munitions List, the
|
||
Commodity Control List and the Nuclear Referral List which cover,
|
||
respectively, munitions, industrial and nuclear-related items.
|
||
|
||
Cryptographic technology falls into the Munitions List which is
|
||
administered by the Department of State, in consultation with the
|
||
Department of Defense. In this case, the NSA controls most of the
|
||
decision making.
|
||
|
||
The Arms Export Control Act (AECA) exists _primarily_ to restrict the
|
||
acquisition of biological organisms, missile technology, chemical
|
||
weapons and any items of use in production of nuclear bombs to
|
||
embargoed nations or countries thought inimical to the interests of
|
||
the United States. (Examples: South Africa, North Korea, Libya, Iran,
|
||
Iraq, etc.)
|
||
|
||
That the AECA is used as a tool to control the development of private
|
||
cryptography in the US is secondary to its original aim, but is a
|
||
logical consequence of four considerations which the ITAR lists as
|
||
determinators of whether a technological development is a defense
|
||
item. These are:
|
||
|
||
1. Whether the item is "inherently military in nature."
|
||
|
||
2. Whether the item "has a predominantly military application."
|
||
|
||
3. Whether an item has military and civil uses "does not in and of
|
||
itself determine" whether it is a defense item.
|
||
|
||
4. "Intended use . . . is also not relevant," for the item's
|
||
classification.
|
||
|
||
If you're brain hasn't seized yet - often, this is what the government
|
||
counts on - you may have the gut feeling that the determinators are
|
||
sufficiently strong and vague to allow for the inclusion of just about
|
||
anything in the U.S. Munitions List or related lists of lists. That
|
||
would be about right.
|
||
|
||
Which is basically what Grady Ward has been yelling about, only he
|
||
doesn't kill you with jargon, bureaucrat-ese or Orwell-speak, God
|
||
bless him.
|
||
|
||
[Yes, you too can be an armchair expert on the topic using acronyms,
|
||
insider terms, secret handshakes and obscure facts and references to
|
||
go toe-to-toe with the best in this controversy. Just take advantage
|
||
of this little reading list:
|
||
|
||
1. Bamford, James. 1982. "The Puzzle Palace: Inside The National
|
||
Security Agency, America's Most Secret Intelligence Organization"
|
||
Penguin Books.
|
||
|
||
Nota Bene: The NSA really hated James Bamford, so much so that it
|
||
attempted to classify _him_, all 150,000 published copies of "The
|
||
Puzzle Palace," his notes and all materials he had gained under the
|
||
Freedom of Information Act. Of this, NSA director Lincoln D. Faurer
|
||
said, "Just because information has been published doesn't mean it
|
||
shouldn't be classified."
|
||
|
||
2. Foerstal, Herbert N. 1993. "Secret Science: Federal Control of
|
||
American Science and Technology" Praeger Publishers.
|
||
|
||
3. "Encyclopedia of the US Military", edited by William M. Arkin,
|
||
Joshua M. Handler, Julia A. Morrissey and Jacquelyn M. Walsh. 1990.
|
||
Harper & Row/Ballinger.
|
||
|
||
4. "The US and Multilateral Export Control Regimes," in "Finding
|
||
Common Ground" 1991. National Academy of Sciences, National Academy
|
||
Press.
|
||
|
||
------------------------------
|
||
|
||
Date: Tue, 5 Oct 1993 21:02:30 EDT
|
||
From: Nikki Draper <draper@EUPHRATES.STANFORD.EDU>
|
||
Subject: File 3--IGC Wins Social Responsibility Award
|
||
|
||
BAY AREA COMPUTER NETWORK ORGANIZATION
|
||
WINS PRIZE FOR SOCIAL RESPONSIBILITY
|
||
|
||
Palo Alto, Calif., September 15, 1993 - Computer Professionals for
|
||
Social Responsibility (CPSR), the national public interest
|
||
organization based in Palo Alto, announced today that the Institute
|
||
for Global Communications (IGC) has been named the winner of the 1993
|
||
Norbert Wiener Award for Social and Professional Responsibility.
|
||
Beginning in 1986, CPSR has presented this award each year to a
|
||
distinguished individual who, through personal example, demonstrated a
|
||
deep commitment to the socially responsible use of computing
|
||
technology. In 1992, the CPSR Board expanded the nominations to
|
||
include organizations. IGC is the first organizational recipient of
|
||
this prestigious award.
|
||
|
||
"The award is particularly appropriate this year because of the
|
||
enormous interest in computer networks generated by the debate over
|
||
the proposed National Information Infrastructure (NII)," said Stanford
|
||
professor and CPSR Board president Eric Roberts. "IGC has worked
|
||
diligently to use network technology to empower previously
|
||
disenfranchised individuals and groups working for progressive change.
|
||
CPSR has a strong commitment to making sure that everyone has access
|
||
to the resources and empowerment that networks provide. IGC has been
|
||
providing such access ever since it was founded in 1986."
|
||
|
||
"We're honored to be recognized by CPSR and to be the Norbert Wiener
|
||
Award recipient," says Geoff Sears, IGC's Executive Director. "Of
|
||
course, this award honors not just IGC, but the efforts and
|
||
accomplishments of all our network members, our entire network
|
||
community."
|
||
|
||
Sears will accept the Wiener award at CPSR's annual meeting banquet in
|
||
Seattle, Washington, on Saturday, October 16th.
|
||
|
||
This year's annual meeting is a two-day conference entitled
|
||
"Envisioning the Future: A National Forum on the National Information
|
||
Infrastructure (NII)" that will bring together local, regional, and
|
||
national decision makers to take a critical look at the social
|
||
implications of the NII. The keynote speaker will be Bruce McConnell,
|
||
Chief of Information Policy at the Office of Information and
|
||
Regulatory Affairs in the Office of Management and Budget (OMB), who
|
||
will present his views on the major NII issues now facing the
|
||
administration. Other highlights of the meeting include Kit Galloway
|
||
of Electronic Cafe International in Santa Monica, California, as the
|
||
featured speaker at the banquet. Using videotapes and a live
|
||
demonstration with CPSR chapters, Kit will present an innovative
|
||
approach to electronic communication and discuss how the Electronic
|
||
Cafe concept has been used.
|
||
|
||
The Institute for Global Communications is a nonprofit computer
|
||
networking organization dedicated to providing low-cost worldwide
|
||
communication and information exchange pertaining to environmental
|
||
preservation, human rights, sustainable development, peace, and social
|
||
justice issues. IGC operates the PeaceNet, EcoNet, ConflictNet, and
|
||
LaborNet computer networks. With a combined membership of 10,000
|
||
individuals and organizations ranging in size and scope from United
|
||
Nations Commissions to local elementary schools, IGC members
|
||
contribute to more than 1200 conferences covering virtually every
|
||
environmental and human rights topic.
|
||
|
||
The Wiener Award was established in 1987 in memory of Norbert Wiener,
|
||
the originator of the field of cybernetics and a pioneer in looking at
|
||
the social and political consequences of computing. Author of the
|
||
book, The Human Use of Human Beings, Wiener began pointing out the
|
||
dangers of nuclear war and the role of scientists in developing more
|
||
powerful weapons shortly after Hiroshima.
|
||
|
||
Past recipients of the Wiener Award have been: Dave Parnas, 1987, in
|
||
recognition of his courageous actions opposing the Strategic Defense
|
||
Initiative; Joe Weizenbaum, 1988, for his pioneering work emphasizing
|
||
the social context of computer science; Daniel McCracken, 1989, for
|
||
his work organizing computer scientists against the Anti Ballistic
|
||
Missiles deployment during the 1960s; Kristen Nygaard of Norway, 1990,
|
||
for his work in participatory design; Severo Ornstein and Laura Gould,
|
||
1991, in recognition of their tireless energy guiding CPSR through
|
||
its early years; and Barbara Simons, 1992, for her work on human
|
||
rights, military funding, and the U.C. Berkeley reentry program for
|
||
women and minorities.
|
||
|
||
Founded in 1981, CPSR is a national, nonprofit, public-interest
|
||
organization of computer scientists and other professionals concerned
|
||
with the impact of computer technology on society. With offices in
|
||
Palo Alto, California, and Washington, D.C., CPSR challenges the
|
||
assumption that technology alone can solve political and social
|
||
problems.
|
||
|
||
For more information about CPSR, the annual meeting, or the awards
|
||
banquet, call 415-322-3778 or send email to <cpsr@cpsr.org>.
|
||
|
||
For more information about IGC, contact Sarah Hutchison, 415-442-0220
|
||
x117, or send email to <sarah@igc.apc.org>.
|
||
|
||
------------------------------
|
||
|
||
Date: Sat, 16 Oct 93 17:44:16 PDT
|
||
From: annaliza@NETCOM.COM(Annaliza T. Orquamada)
|
||
Subject: File 4--Full Description of Proposed "Hacker" Documentary"
|
||
|
||
((MODERATORS' NOTE: In CuD 5.82, we ran a short description of a
|
||
proposed documentary film on "Hackers," which intends to be an
|
||
antidote to conventional media depictions of the topic. We asked for
|
||
a more lengthy description of the project and received the following
|
||
summary. We combined two files after a long day of teaching, and hope
|
||
we have not omitted or re-edited inappropriately. Any errors or
|
||
omissions are the result of our editing, and not necessarily gaps in
|
||
the original posts.
|
||
|
||
We have long-argued that conventional media depictions of "hacking"
|
||
are flawed. The more we learn about the proposed documentary, the more
|
||
encouraged we are that there exist film makers with both the talent
|
||
and the knowledge to produce antidotes to Forbes Magazines "Hackers in
|
||
the Hood," Geraldo's "Mad Hacker's Tea-party," and Datelines' modem
|
||
hysteria, to name just a few of the more egregious examples of media
|
||
madness. Annaliza's group may or may not tell the "hacker story" in a
|
||
way that will please everybody, but we remain impressed with her
|
||
meticulous research and her open-mindedness. She is about to begin a
|
||
cross-country jaunt to interview/film those willing to talk with her,
|
||
so if you have a story to tell, think about letting her know)).
|
||
|
||
=====================================
|
||
|
||
TREATMENT FOR DOCUMENTARY: UNAUTHORIZED ACCESS ONLY
|
||
16, October, 1993
|
||
annaliza@netcom.com
|
||
|
||
Lately the media have widely publicized the on-going dilemmas of
|
||
computer security experts whose job it is to stop systems crackers
|
||
(what the media have labelled as hackers) from breaking into secure
|
||
systems. There have been accounts of teenagers being sentenced for
|
||
stealing information, running up phone bills of thousands of dollars
|
||
and even espionage.
|
||
|
||
What is the real threat? Who are these people who break into computer
|
||
systems? Why do they do it?
|
||
|
||
Since the computer was first put on line and hooked up to a phone,
|
||
there has always been a risk to security. Breaking into computers is
|
||
viewed by many hackers as a mental game of chess. Often computer
|
||
professionals tolerate such break-ins as nothing more than inquisitive
|
||
minds trying to see if they can outwit the security experts. Most
|
||
hackers, when caught show no remorse. In fact, they rarely view
|
||
themselves as criminals. They even hold conventions in various global
|
||
locations, often inviting their prosecutors to join them. so why is
|
||
hacking such a threat? How does it affect the computer community?
|
||
Who are these hackers and what are their objectives? Is there any
|
||
positive side to hacking?
|
||
|
||
The focus of this documentary will be to follow the hackers and see
|
||
what motivates them. It will be to show how they feel about the
|
||
underground computer community, and their own place within it. What
|
||
are their stories and their explanations? Do they have a political
|
||
agenda, or are they just joyriding through computer systems? How do
|
||
they feel about the media and its sensationalized attitude towards
|
||
computer cracking and the "outlaw cyberpunk"? What do they think is
|
||
the future of the computer underground?
|
||
|
||
The hacker scene is fractionalized. There are many types of hackers.
|
||
Some work in solitude, others in groups. Some use cellular, others
|
||
are interested in programming. Some hackers obtain passwords and
|
||
codes through the underground or by "social engineering" company
|
||
employees or by using electronic scanners to listen in on phone
|
||
conversations. Some hackers know computer systems so well that they
|
||
don't need passwords but can log on to the computer directly by using
|
||
various security holes.
|
||
|
||
In most countries hacking is now illegal, so everyone who does hack
|
||
risks major penalties, even prison. Some groups have a political
|
||
agenda, or at least some unwritten moral code concerning the right to
|
||
information. There are various interests in the hacker scene
|
||
depending on the individual.
|
||
|
||
Some use hacking for personal gain. Kevin Poulsen, a hacker from Los
|
||
Angeles, used his knowledge of the phone system to block phone lines
|
||
to a radio station to win a new porsche (Littman, 1993).
|
||
|
||
Some hackers are into military systems. One case in particular was
|
||
comprised of a group of hackers in Germany who sold computer software
|
||
programs to the KGB. Though the software given to the Russians was
|
||
freely available in the West, the group faced espionage charges. The
|
||
hackers who sold the software displeased many in the W. German Hacker
|
||
Underground who believed it to be morally wrong to hack for monetary
|
||
gain. The project itself was allegedly started to bring the Soviet's
|
||
military computer software standard to a grade matching the Americans.
|
||
It was called "Project Equalizer" (Hafner and Markoff, 1991; Stoll,
|
||
1989).
|
||
|
||
The documentary will aim to find out more about what the political
|
||
premise of the hackers is presently and what its role will be in the
|
||
future. Are hackers using their skills for political reasons? Will
|
||
individual hackers play a major role in influencing the radical left
|
||
or the radical right in the future? Are hackers being used as
|
||
government or corporate spies? How do the hackers feel about computer
|
||
politics? How do hacker politics vary according to the nationalities
|
||
of the hackers themselves?
|
||
|
||
To date, the media have concentrated on systems crackers as the
|
||
entirety of the hacker community. Even though the community is
|
||
fractionalized, each sections interacts with the other. The
|
||
documentary will explore other parts of the underground.
|
||
|
||
Mark Ludwig, author of "The Little Black Book of Computer Viruses",
|
||
recently unleashed one of his latest virus programs at Def Con 1, a
|
||
hacker convention that was held in Las Vegas in July of 1993. The
|
||
virus infects the computer hard drive encrypting everything
|
||
automatically. The only way to recover the data is to know the secret
|
||
password. This sent a buzz through the conference. The ramifications
|
||
being that any information stored on the hackers hard drive would be
|
||
impossible to retrieve should the Secret Service come bursting through
|
||
the door simply by rebooting the computer.
|
||
|
||
Some hackers see themselves as artists. These hackers are always
|
||
offended when one confuses them with systems crackers. They see
|
||
themselves as more of an intellectual elite and are very condescending
|
||
towards systems crackers. One such hacker was able to penetrate a
|
||
NASA satellite probe. When the satellite was launched into space a
|
||
peace sign appeared on it's monitor.
|
||
|
||
The hacking community is growing. Every year conventions are held in
|
||
the United States, Germany, France and Holland, as well as through out
|
||
the world. SummerCon, HoHoCon, Def Con, and The Hacking at the End of
|
||
the Universe Conference are some of the best known. In August of
|
||
1993, The Hacking at the End of the Universe Conference was reported
|
||
as having over 600 attendees. This particular global conference, put
|
||
on by Hactic, was held outside of Amerstam in Holland. The speakers
|
||
ranged from hackers to security experts to Police Agents. The press
|
||
was everywhere. A spread even appeared in Newsweek Magazine (July 26,
|
||
1993: 58). Though most Cons are places for exchanging information,
|
||
meeting electronic friends, and generally having a good time,
|
||
sometimes there are problems. Last year at PumpCon arrests were made.
|
||
At Def Con, Gail Thackeray, a woman who spends much of her time
|
||
prosecuting hackers, started her speech by saying she wasn't there to
|
||
bust anyone. Another speaker, Dark Druid, was unable to talk about
|
||
his planned topic because his persecutor happened to be sitting in the
|
||
audience.
|
||
|
||
More and more hackers are breaking headlines in the news. The AT&T
|
||
crash of 1990, (though caused by a wrongly written line of code in a
|
||
the switching software program), led to speculation among some media
|
||
stories and law enforcement officials that hackers might have been
|
||
responsible.
|
||
|
||
So why are hackers such a threat??? What does a hacker do that could
|
||
affect the average person?? One of the objectives of the documentary
|
||
will be to explore the technology available to the hacker.
|
||
|
||
Hackers are experts on the phone systems, they have to be in order to
|
||
hack systems without being traced. The really good hackers are able
|
||
to dial into the phone systems and trick the phone computers into
|
||
believing that they are part of the system, or even that they are the
|
||
controller of the system. So how do the hackers do it? Where do they
|
||
obtain their information? How do they get onto systems? How do they
|
||
get out without being traced? What can they do with their hacking
|
||
abilities?
|
||
|
||
Kevin Poulsen, in the instance of the KIIS FM radio contest was able
|
||
to use his knowledge of the phone system to take control of the phone
|
||
lines and wait until 119 calls had been placed. On the 120st he
|
||
simply blocked all of the incoming lines to make sure that only his
|
||
call got through.
|
||
|
||
A prank by another hacker involved taking control of the phone system
|
||
and then using it to reroute the calls of a certain probation officer.
|
||
When someone called up the probation officers's office, the caller
|
||
would be connected to a phone sex service (Sterling, 1992: 98-99).
|
||
|
||
Some European hackers broke into South African computer systems during
|
||
the boycott against the Apartheid system. The hackers deleted files
|
||
in South Africa to disrupt the political system and also were able to
|
||
monitor which companies were breaking the boycott by monitoring
|
||
computer systems.
|
||
|
||
A serious case that was to initiate Operation Sundevil and lead to
|
||
many arrests was to involve a document called E-911. This document
|
||
(though later found to be obtainable through legal channels for about
|
||
$13.95) was obtained by a hacker on one of his jaunts through the
|
||
phone system computers. The document was kept by the hacker as a
|
||
souvenir. He sent the document to a friend who published it in an
|
||
electronic magazine called Phrack (an electronic hacker magazine
|
||
available on the internet). The phone company was furious that their
|
||
supposedly secure system had been breached and that proprietary
|
||
information was being spread throughout the hacker community. Not
|
||
only was this stolen/private property, the document contained
|
||
information pertaining to the 911 emergency services. Although the
|
||
document had been edited so that no harmful information was published,
|
||
the phone company was furious. Once a hacker has gained root or
|
||
super-user privileges at a phone company switching station there is
|
||
always the potential threat that they could do some very real damage
|
||
(intentionally or unintentionally). If a hacker could re-route a
|
||
judge's phone calls or have an enemies phone disconnected or make free
|
||
calls globally, what is to stop them from cutting off the 911
|
||
emergency systems??? This is why the U.S. Secret U.S. Service (the
|
||
branch of the government that is responsible for the prosecution of
|
||
most electronic crime) went so far as to break down doors of 15 year
|
||
olds with guns and haul them and all of their equipment away. One
|
||
hacker was reportedly banned from even going within 100 yards of a
|
||
computer terminal.
|
||
|
||
Our documentary will also explore the ramifications of the hacker's
|
||
actions. Many hackers have been arrested, imprisoned, had their
|
||
computers as well as their software confiscated. Are these arrests
|
||
always justified? Many innocent people have been questioned by the
|
||
Secret Service and FBI purely through suspicion in connection with
|
||
computer related crime. In fact, is was because of the FBI's
|
||
investigation of the alleged "theft" of Apple proprietary source code
|
||
and it's curious questioning of Mitch Kapor, founder of Lotus 1-2-3,
|
||
and John Perry Barlow, former Grateful Dead lyricist, that led to the
|
||
forming of the Electronic Frontier Foundation (EFF) (Sterling, 1992:
|
||
232-238). Phil Zimmerman, the creator of an electronic privacy
|
||
encryption program called PGP has been subpoenaed by the U.S.
|
||
government for creating a program that ensured legitimate privacy.
|
||
Many people have had their equipment confiscated without ever being
|
||
charged of a crime. Are fundamental human rights being broken because
|
||
of the fear of the unknown?
|
||
|
||
Is this fear really justified? If hackers can take control of local
|
||
switching stations (and they can) why don't they wreak havoc. If
|
||
there is such a threat to the general public then why don't hackers
|
||
cause more serious damage?
|
||
|
||
"Bellcore clearly believes that hackers are nothing short of
|
||
terrorists. A security alert from November 1990 warns that "the
|
||
potential for security incidents this holiday weekend is significantly
|
||
higher than normal because of the recent sentencing of the three
|
||
former Legion of Doom members. These incidents may include Social
|
||
Engineering (gaining information by posing as a bellcore employee over
|
||
the telephone), computer intrusion, as well as possible physical
|
||
intrusion."'*
|
||
|
||
But how do the hackers see themselves?? How do they justify breaking
|
||
into Bellcore electronically or physically. If hackers are such a
|
||
major threat then why do so many corporations using computers hooked
|
||
up to outside connections leave their electronic doors wide open?
|
||
|
||
As computers become more available and widespread throughout the
|
||
community, so does hacking. This documentary hopes to address the real
|
||
threats, as well as the hype. Is hacking "intellectual joyriding"?
|
||
Or serious criminal behavior.
|
||
|
||
By humanizing the hacker scene this documentary hopes to demystify the
|
||
sinister mythos surrounding what has been deemed by the media as 'the
|
||
outlaw hacker'. It is not the documentar's objective to make
|
||
judgements, only to try to understand.
|
||
|
||
The documentary will run approximately 30 minutes. Our objective will
|
||
be to film at various hacker conventions and meeting places in the
|
||
United States and Europe. We will be shooting on broadcast quality
|
||
video. The documentary crew will be leaving Los Angeles at the
|
||
beginning of December and going to wherever there are people who want
|
||
to get involved in the project. Ultimately, we hope to show the film
|
||
at conferences, festivals and perhaps on high quality t.v. (such as
|
||
Channel 4 in England or PBS in the U.S.). It will also be suitable for
|
||
classroom viewing and related educational purposes.
|
||
|
||
This documentary is about the hacker community itself. We are looking
|
||
for monetary donations from the underground or from people sympathetic
|
||
to the underground. In this way, we will be able to make the
|
||
documentary without corporate or film company control. Our group is
|
||
comprised of film makers who are involved in the scene itself. We are
|
||
looking also for any donation of services, i.e. Beta transfer time, an
|
||
off-on line editing suite, sound equipment, videotape, etc...
|
||
|
||
If anyone would like to get involved in the project in any capacity,
|
||
whether it be to go in front of the camera, or relate a story or a
|
||
hack anonymously to my e-mail address, or donate funds, or equipment
|
||
or editing time, please get in touch.
|
||
|
||
This documentary hopes to be an open forum for hackers to relate their
|
||
stories and ideas about the past/present/future. We hope to be able
|
||
to disseminate the hype from other sensationalized media who are only
|
||
looking for a good story and don't really care about the ramifications
|
||
of their actions.
|
||
|
||
Anyone who is interested in any aspect of this project, please contact
|
||
me Annaliza at annaliza@netcom.com
|
||
|
||
* Taken from 2600 Magazine - The Hacker Quarterly - Volume Nine,
|
||
Number Four - Winter 1992-93.
|
||
|
||
BIBLIOGRAPHY
|
||
|
||
Hafner, Katie, and John Markoff. 1991. _Cyberpunk: Outlaws and
|
||
Hackers on the Computer Frontier._ New York: Simon and Schuster.
|
||
|
||
Littman, Jonathan. 1993. "The Last Hacker." _The Los Angeles Times
|
||
Sunday Magazine_. September 12: 18 ff.
|
||
|
||
Sterling, Bruce. 1992. _The Hacker Crackdown_. New York: Bantam.
|
||
|
||
Stoll, Cliff. 1989. _The Cuckoo's Egg. New York: Doubleday.
|
||
|
||
------------------------------
|
||
|
||
End of Computer Underground Digest #5.82
|
||
************************************
|
||
|
||
|
||
|