19.12.2012 Views

Inventions and Inventors Volume 1 - Online Public Access Catalog

Inventions and Inventors Volume 1 - Online Public Access Catalog

Inventions and Inventors Volume 1 - Online Public Access Catalog

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Inventions</strong><br />

<strong>and</strong><br />

<strong>Inventors</strong>


This Page Intentionally Left Blank


MAGILL’S CHOICE<br />

<strong>Inventions</strong><br />

<strong>and</strong><br />

<strong>Inventors</strong><br />

<strong>Volume</strong> 1<br />

Abortion pill — Laminated glass<br />

1 – 458<br />

edited by<br />

Roger Smith<br />

Salem Press, Inc.<br />

Pasadena, California Hackensack, New Jersey


Copyright © 2002, by Salem Press, Inc.<br />

All rights in this book are reserved. No part of this work may be<br />

used or reproduced in any manner whatsoever or transmitted<br />

in any form or by any means, electronic or mechanical, including<br />

photocopy, recording, or any information storage <strong>and</strong> retrieval<br />

system, without written permission from the copyright<br />

owner except in the case of brief quotations embodied in critical<br />

articles <strong>and</strong> reviews. For information address the publisher, Salem<br />

Press, Inc., P.O. Box 50062, Pasadena, California 91115.<br />

Essays originally appeared in Twentieth Century: Great Events<br />

(1992, 1996), Twentieth Century: Great Scientific Achievements (1994),<br />

<strong>and</strong> Great Events from History II: Business <strong>and</strong> Commerce Series<br />

(1994). New material has been added.<br />

∞ The paper used in these volumes conforms to the American<br />

National St<strong>and</strong>ard for Permanence of Paper for Printed Library<br />

Materials, Z39.48-1992 (R1997).<br />

Library of Congress <strong>Catalog</strong>ing-in-<strong>Public</strong>ation Data<br />

<strong>Inventions</strong> <strong>and</strong> inventors / edited by Roger Smith<br />

p.cm. — (Magill’s choice)<br />

Includes bibliographical reference <strong>and</strong> index<br />

ISBN 1-58765-016-9 (set : alk. paper) — ISBN 1-58765-017-7<br />

(vol 1 : alk. paper) — ISBN 1-58765-018-5 (vol 2. : alk. paper)<br />

1. <strong>Inventions</strong>—History—20th century—Encyclopedias. 2. <strong>Inventors</strong>—Biography—Encyclopedias.<br />

I. Smith, Roger, 1953- .<br />

II. Series.<br />

T20 .I59 2001<br />

609—dc21 2001049412<br />

printed in the united states of america


Table of Contents<br />

Table of Contents<br />

Publisher’s Note . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix<br />

Editor’s Foreword . . . . . . . . . . . . . . . . . . . . . . . . . . . xi<br />

Abortion pill . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1<br />

Airplane . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6<br />

Alkaline storage battery . . . . . . . . . . . . . . . . . . . . . . . 11<br />

Ammonia . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16<br />

Amniocentesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20<br />

Antibacterial drugs . . . . . . . . . . . . . . . . . . . . . . . . . . 24<br />

Apple II computer. . . . . . . . . . . . . . . . . . . . . . . . . . . 28<br />

Aqualung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33<br />

Artificial blood . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38<br />

Artificial chromosome . . . . . . . . . . . . . . . . . . . . . . . . 41<br />

Artificial heart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45<br />

Artificial hormone. . . . . . . . . . . . . . . . . . . . . . . . . . . 50<br />

Artificial insemination . . . . . . . . . . . . . . . . . . . . . . . . 54<br />

Artificial kidney . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58<br />

Artificial satellite . . . . . . . . . . . . . . . . . . . . . . . . . . . 63<br />

Aspartame . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67<br />

Assembly line . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71<br />

Atomic bomb . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76<br />

Atomic clock . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80<br />

Atomic-powered ship. . . . . . . . . . . . . . . . . . . . . . . . . 84<br />

Autochrome plate . . . . . . . . . . . . . . . . . . . . . . . . . . . 88<br />

BASIC programming language . . . . . . . . . . . . . . . . . . . 92<br />

Bathyscaphe . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95<br />

Bathysphere . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100<br />

BINAC computer . . . . . . . . . . . . . . . . . . . . . . . . . . 104<br />

Birth control pill . . . . . . . . . . . . . . . . . . . . . . . . . . . 108<br />

Blood transfusion . . . . . . . . . . . . . . . . . . . . . . . . . . 113<br />

Breeder reactor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118<br />

Broadcaster guitar . . . . . . . . . . . . . . . . . . . . . . . . . . 122<br />

Brownie camera . . . . . . . . . . . . . . . . . . . . . . . . . . . 130<br />

Bubble memory . . . . . . . . . . . . . . . . . . . . . . . . . . . 138<br />

v


Table of Contents<br />

Bullet train . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142<br />

Buna rubber . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146<br />

CAD/CAM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151<br />

Carbon dating . . . . . . . . . . . . . . . . . . . . . . . . . . . . 158<br />

Cassette recording . . . . . . . . . . . . . . . . . . . . . . . . . . 163<br />

CAT scanner . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167<br />

Cell phone . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172<br />

Cloning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177<br />

Cloud seeding . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183<br />

COBOL computer language . . . . . . . . . . . . . . . . . . . . 187<br />

Color film . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192<br />

Color television . . . . . . . . . . . . . . . . . . . . . . . . . . . 196<br />

Colossus computer . . . . . . . . . . . . . . . . . . . . . . . . . 200<br />

Communications satellite . . . . . . . . . . . . . . . . . . . . . . 204<br />

Community antenna television . . . . . . . . . . . . . . . . . . 208<br />

Compact disc . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 217<br />

Compressed-air-accumulating power plant . . . . . . . . . . . 225<br />

Computer chips . . . . . . . . . . . . . . . . . . . . . . . . . . . 229<br />

Contact lenses . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235<br />

Coronary artery bypass surgery . . . . . . . . . . . . . . . . . . 240<br />

Cruise missile . . . . . . . . . . . . . . . . . . . . . . . . . . . . 244<br />

Cyclamate. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248<br />

Cyclotron . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 252<br />

Diesel locomotive . . . . . . . . . . . . . . . . . . . . . . . . . . 257<br />

Differential analyzer. . . . . . . . . . . . . . . . . . . . . . . . . 262<br />

Dirigible. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 267<br />

Disposable razor . . . . . . . . . . . . . . . . . . . . . . . . . . . 272<br />

Dolby noise reduction . . . . . . . . . . . . . . . . . . . . . . . . 279<br />

Electric clock . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 284<br />

Electric refrigerator . . . . . . . . . . . . . . . . . . . . . . . . . 289<br />

Electrocardiogram . . . . . . . . . . . . . . . . . . . . . . . . . . 293<br />

Electroencephalogram. . . . . . . . . . . . . . . . . . . . . . . . 298<br />

Electron microscope . . . . . . . . . . . . . . . . . . . . . . . . . 302<br />

Electronic synthesizer . . . . . . . . . . . . . . . . . . . . . . . . 307<br />

ENIAC computer . . . . . . . . . . . . . . . . . . . . . . . . . . 312<br />

vi


Table of Contents<br />

Fax machine . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 316<br />

Fiber-optics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 320<br />

Field ion microscope. . . . . . . . . . . . . . . . . . . . . . . . . 325<br />

Floppy disk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 330<br />

Fluorescent lighting . . . . . . . . . . . . . . . . . . . . . . . . . 335<br />

FM radio . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339<br />

Food freezing. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 343<br />

FORTRAN programming language . . . . . . . . . . . . . . . . 347<br />

Freeze-drying. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351<br />

Fuel cell . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355<br />

Gas-electric car . . . . . . . . . . . . . . . . . . . . . . . . . . . . 360<br />

Geiger counter . . . . . . . . . . . . . . . . . . . . . . . . . . . . 365<br />

Genetic “fingerprinting” . . . . . . . . . . . . . . . . . . . . . . 370<br />

Genetically engineered insulin . . . . . . . . . . . . . . . . . . . 374<br />

Geothermal power. . . . . . . . . . . . . . . . . . . . . . . . . . 378<br />

Gyrocompass . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 382<br />

Hard disk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 386<br />

Hearing aid . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 390<br />

Heart-lung machine . . . . . . . . . . . . . . . . . . . . . . . . . 394<br />

Heat pump . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 398<br />

Holography. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 402<br />

Hovercraft . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 407<br />

Hydrogen bomb . . . . . . . . . . . . . . . . . . . . . . . . . . . 412<br />

IBM Model 1401 computer . . . . . . . . . . . . . . . . . . . . . 417<br />

In vitro plant culture. . . . . . . . . . . . . . . . . . . . . . . . . 421<br />

Infrared photography . . . . . . . . . . . . . . . . . . . . . . . . 425<br />

Instant photography. . . . . . . . . . . . . . . . . . . . . . . . . 430<br />

Interchangeable parts . . . . . . . . . . . . . . . . . . . . . . . . 434<br />

Internal combustion engine. . . . . . . . . . . . . . . . . . . . . 442<br />

The Internet. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 446<br />

Iron lung . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 451<br />

Laminated glass . . . . . . . . . . . . . . . . . . . . . . . . . . . 454<br />

vii


This Page Intentionally Left Blank


Publisher’s Note<br />

Publisher’s Note<br />

To many people, the word “invention” brings to mind cleverly contrived<br />

gadgets <strong>and</strong> devices, such as safety pins, zippers, typewriters,<br />

<strong>and</strong> telephones—all of which have fascinating stories of invention<br />

behind them. However, the word actually has a much broader meaning,<br />

one that goes back to the Latin word invenire, for “to come upon.”<br />

In its broad sense, an invention can be any tangible device or contrivance,<br />

or even a process, that is brought into being by human imagination.<br />

It is in this broad sense that the term is used in <strong>Inventions</strong> <strong>and</strong> <strong>Inventors</strong>,<br />

the latest contribution to the Magill’s Choice reference books.<br />

This two-volume set contains articles on 195 twentieth century<br />

inventions, which span the full range of human imagination—from<br />

simple gadgets, such as disposable razors, to unimaginably complex<br />

medical breakthroughs, such as genetically engineered insulin.<br />

This set is not an encyclopedic catalog of the past century’s greatest<br />

inventions but rather a selective survey of noteworthy breakthroughs<br />

in the widest possible variety of fields.<br />

A combination of several features sets <strong>Inventions</strong> <strong>and</strong> <strong>Inventors</strong><br />

apart from other reference works on this subject: the diversity of its<br />

subject matter, the depth of its individual articles, <strong>and</strong> its emphasis<br />

on the people behind the inventions. The range of subjects covered<br />

here is unusually wide. In addition to articles on what might be considered<br />

“classic” inventions—such as airplanes, television, <strong>and</strong> satellites—the<br />

set has articles on inventions in fields as diverse as agriculture,<br />

biology, chemistry, computer science, consumer products,<br />

drugs <strong>and</strong> vaccines, energy, engineering, food science, genetic engineering,<br />

medical procedures, music, photography, physics, synthetics,<br />

transportation, <strong>and</strong> weapons technology.<br />

Most of this set’s essays appeared earlier in Twentieth Century:<br />

Great Events (1992, 1996) <strong>and</strong> Twentieth Century: Great Scientific<br />

Achievements (1994). Its longest essays are taken from Great Events<br />

from History II: Business <strong>and</strong> Commerce Series (1994). Information in<br />

the articles has been updated, <strong>and</strong> completely new bibliographical<br />

notes have been added to all of them. Half the essays also have original<br />

sidebars on people behind the inventions.<br />

ix


Publisher’s Note<br />

At least one thous<strong>and</strong> words in length, each essay opens with a<br />

brief summary of the invention <strong>and</strong> its significance, followed by an<br />

annotated list of important personages behind it—including scientists,<br />

engineers, technicians, <strong>and</strong> entrepreneurs. The essay then examines<br />

the background to the invention, its process of discovery<br />

<strong>and</strong> innovation, <strong>and</strong> its impact on the world. Half the articles have<br />

entirely new sidebars on individuals who played important roles in<br />

the inventions’ development <strong>and</strong> promotion.<br />

Users can find topics by using any of several different methods.<br />

Articles are alphabetically arranged under their titles, which use the<br />

names of the inventions themselves, such as “Abortion pill,” “Airplane,”<br />

“Alkaline storage battery,” “Ammonia,” <strong>and</strong> “Amniocentesis.”<br />

Many inventions are known by more than one name, however,<br />

<strong>and</strong> users may find what they are looking for in the general index,<br />

which lists topics under multiple terms.<br />

Several systems of cross-referencing direct users to articles of interest.<br />

Appended to every essay is a list of articles on related or similar<br />

inventions. Further help in can be found in appendices at the end<br />

of volume two. The first, a Time Line, lists essay topics chronologically,<br />

by the years in which the inventions were first made. The second,<br />

Topics by Category list, organizes essay topics under broader<br />

headings, with most topics appearing under at least two category<br />

headings. Allowing for the many topics counted more than once,<br />

these categories include Consumer products (36 essays), Electronics<br />

(28), Communications (27), Medicine (25), Measurement <strong>and</strong> detection<br />

(24), Computer science (23), Home products (20), Materials<br />

(18), Medical procedures (17), Synthetics (17), Photography (16), Energy<br />

(16), Engineering (16), Physics (13), Food science (13), Drugs<br />

<strong>and</strong> vaccines (13), Transportation (11), Weapons technology (11),<br />

Genetic engineering (11), Aviation <strong>and</strong> space (10), Biology (9),<br />

Chemistry (9), Exploration (8), Music (7), Earth science (6), Manufacturing<br />

(6), <strong>and</strong> Agriculture (5).<br />

More than one hundred scholars wrote the original articles used<br />

in these volumes. Because their names did not appear with their articles<br />

in the Twentieth Century sets, we cannot, unfortunately, list<br />

them here. However, we extend our thanks for their contributions.<br />

We also are indebted to Roger Smith for his help in assembling the<br />

topic list <strong>and</strong> in writing all the biographical sidebars.<br />

x


Editor’s Foreword<br />

The articles in <strong>Inventions</strong> <strong>and</strong> <strong>Inventors</strong> recount the birth <strong>and</strong> growth<br />

of important components in the technology of the twentieth centuries.<br />

They concern inventions ranging from processes, methods,<br />

sensors, <strong>and</strong> tests to appliances, tools, machinery, vehicles, electronics,<br />

<strong>and</strong> materials. To explain these various inventions, the essays<br />

deal with principles of physics, chemistry, engineering, biology, <strong>and</strong><br />

computers—all intended for general readers. From complex devices,<br />

such as electron microscopes, <strong>and</strong> phenomena difficult to define,<br />

such as the Internet, to things so familiar that they are seldom<br />

thought of as having individual histories at all, such as Pyrex glass<br />

<strong>and</strong> Velcro, all the inventions described here increased the richness<br />

of technological life. Some of these inventions, such as the rotarydial<br />

telephone, have passed out of common use, at least in the<br />

United States <strong>and</strong> Europe, while others, such as the computer, are<br />

now so heavily relied upon that mass technological culture could<br />

scarcely exist without them. Each article, then, is at the same time a<br />

historical sketch <strong>and</strong> technical explanation of an invention, written<br />

to inform <strong>and</strong>, I hope, intrigue.<br />

Brief biographical sidebars accompany half the articles. The<br />

sidebars outline the lives of people who are in some way responsible<br />

for the inventions discussed: the original inventor, a person who<br />

makes important refinements, an entrepreneur, or even a social crusader<br />

who fostered acceptance for a controversial invention, as Margaret<br />

Sanger did for the birth control pill. These little biographies,<br />

although offering only basic information, call forth the personal<br />

struggles behind inventions. And that is a facet to inventions that<br />

needs emphasizing, because it shows that technology, which can<br />

seem bewilderingly impersonal <strong>and</strong> complex, is always rooted in<br />

human need <strong>and</strong> desire.<br />

Roger Smith<br />

Portl<strong>and</strong>, Oregon<br />

xi


This Page Intentionally Left Blank


<strong>Inventions</strong><br />

<strong>and</strong><br />

<strong>Inventors</strong>


This Page Intentionally Left Blank


<strong>Inventions</strong><br />

<strong>and</strong><br />

<strong>Inventors</strong>


This Page Intentionally Left Blank


MAGILL’S CHOICE<br />

<strong>Inventions</strong><br />

<strong>and</strong><br />

<strong>Inventors</strong><br />

<strong>Volume</strong> 2<br />

Laser — Yellow fever vaccine<br />

Index<br />

459 – 936<br />

edited by<br />

Roger Smith<br />

Salem Press, Inc.<br />

Pasadena, California Hackensack, New Jersey


Copyright © 2002, by Salem Press, Inc.<br />

All rights in this book are reserved. No part of this work may be<br />

used or reproduced in any manner whatsoever or transmitted<br />

in any form or by any means, electronic or mechanical, including<br />

photocopy, recording, or any information storage <strong>and</strong> retrieval<br />

system, without written permission from the copyright<br />

owner except in the case of brief quotations embodied in critical<br />

articles <strong>and</strong> reviews. For information address the publisher, Salem<br />

Press, Inc., P.O. Box 50062, Pasadena, California 91115.<br />

Essays originally appeared in Twentieth Century: Great Events<br />

(1992, 1996), Twentieth Century: Great Scientific Achievements (1994),<br />

<strong>and</strong> Great Events from History II: Business <strong>and</strong> Commerce Series<br />

(1994). New material has been added.<br />

∞ The paper used in these volumes conforms to the American<br />

National St<strong>and</strong>ard for Permanence of Paper for Printed Library<br />

Materials, Z39.48-1992 (R1997).<br />

Library of Congress <strong>Catalog</strong>ing-in-<strong>Public</strong>ation Data<br />

<strong>Inventions</strong> <strong>and</strong> inventors / edited by Roger Smith<br />

p.cm. — (Magill’s choice)<br />

Includes bibliographical reference <strong>and</strong> index<br />

ISBN 1-58765-016-9 (set : alk. paper) — ISBN 1-58765-017-7<br />

(vol 1 : alk. paper) — ISBN 1-58765-018-5 (vol 2. : alk. paper)<br />

1. <strong>Inventions</strong>—History—20th century—Encyclopedias. 2. <strong>Inventors</strong>—Biography—Encyclopedias.<br />

I. Smith, Roger, 1953- .<br />

II. Series.<br />

T20 .I59 2001<br />

609—dc21 2001049412<br />

printed in the united states of america


Table of Contents<br />

Table of Contents<br />

Laser. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 459<br />

Laser-diode recording process . . . . . . . . . . . . . . . . . . . 464<br />

Laser eye surgery . . . . . . . . . . . . . . . . . . . . . . . . . . 468<br />

Laser vaporization . . . . . . . . . . . . . . . . . . . . . . . . . . 472<br />

Long-distance radiotelephony . . . . . . . . . . . . . . . . . . . 477<br />

Long-distance telephone . . . . . . . . . . . . . . . . . . . . . . 482<br />

Mammography. . . . . . . . . . . . . . . . . . . . . . . . . . . . 486<br />

Mark I calculator . . . . . . . . . . . . . . . . . . . . . . . . . . . 490<br />

Mass spectrograph. . . . . . . . . . . . . . . . . . . . . . . . . . 494<br />

Memory metal . . . . . . . . . . . . . . . . . . . . . . . . . . . . 498<br />

Microwave cooking . . . . . . . . . . . . . . . . . . . . . . . . . 502<br />

Neoprene . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 507<br />

Neutrino detector . . . . . . . . . . . . . . . . . . . . . . . . . . 511<br />

Nuclear magnetic resonance . . . . . . . . . . . . . . . . . . . . 516<br />

Nuclear power plant. . . . . . . . . . . . . . . . . . . . . . . . . 520<br />

Nuclear reactor. . . . . . . . . . . . . . . . . . . . . . . . . . . . 525<br />

Nylon . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 529<br />

Oil-well drill bit . . . . . . . . . . . . . . . . . . . . . . . . . . . 533<br />

Optical disk. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 537<br />

Orlon . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 541<br />

Pacemaker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 545<br />

Pap test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 549<br />

Penicillin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 553<br />

Personal computer. . . . . . . . . . . . . . . . . . . . . . . . . . 558<br />

Photoelectric cell . . . . . . . . . . . . . . . . . . . . . . . . . . . 562<br />

Photovoltaic cell . . . . . . . . . . . . . . . . . . . . . . . . . . . 567<br />

Plastic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 571<br />

Pocket calculator . . . . . . . . . . . . . . . . . . . . . . . . . . . 576<br />

Polio vaccine (Sabin). . . . . . . . . . . . . . . . . . . . . . . . . 581<br />

Polio vaccine (Salk) . . . . . . . . . . . . . . . . . . . . . . . . . 585<br />

Polyester . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 589<br />

xix


Table of Contents<br />

Polyethylene . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 593<br />

Polystyrene . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 597<br />

Propeller-coordinated machine gun . . . . . . . . . . . . . . . . 601<br />

Pyrex glass . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 606<br />

Radar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 611<br />

Radio . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 616<br />

Radio crystal sets . . . . . . . . . . . . . . . . . . . . . . . . . . 621<br />

Radio interferometer . . . . . . . . . . . . . . . . . . . . . . . . 625<br />

Refrigerant gas . . . . . . . . . . . . . . . . . . . . . . . . . . . . 630<br />

Reserpine . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 634<br />

Rice <strong>and</strong> wheat strains . . . . . . . . . . . . . . . . . . . . . . . 638<br />

Richter scale . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 645<br />

Robot (household) . . . . . . . . . . . . . . . . . . . . . . . . . . 650<br />

Robot (industrial) . . . . . . . . . . . . . . . . . . . . . . . . . . 654<br />

Rocket . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 658<br />

Rotary dial telephone . . . . . . . . . . . . . . . . . . . . . . . . 663<br />

SAINT. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 668<br />

Salvarsan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 673<br />

Scanning tunneling microscope . . . . . . . . . . . . . . . . . . 678<br />

Silicones. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 683<br />

Solar thermal engine. . . . . . . . . . . . . . . . . . . . . . . . . 687<br />

Sonar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 692<br />

Stealth aircraft . . . . . . . . . . . . . . . . . . . . . . . . . . . . 697<br />

Steelmaking process . . . . . . . . . . . . . . . . . . . . . . . . . 701<br />

Supercomputer. . . . . . . . . . . . . . . . . . . . . . . . . . . . 709<br />

Supersonic passenger plane . . . . . . . . . . . . . . . . . . . . 714<br />

Synchrocyclotron . . . . . . . . . . . . . . . . . . . . . . . . . . 720<br />

Synthetic amino acid . . . . . . . . . . . . . . . . . . . . . . . . 724<br />

Synthetic DNA . . . . . . . . . . . . . . . . . . . . . . . . . . . . 729<br />

Synthetic RNA . . . . . . . . . . . . . . . . . . . . . . . . . . . . 733<br />

Syphilis test. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 737<br />

Talking motion pictures . . . . . . . . . . . . . . . . . . . . . . . 741<br />

Teflon . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 746<br />

Telephone switching. . . . . . . . . . . . . . . . . . . . . . . . . 751<br />

Television . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 756<br />

Tevatron accelerator . . . . . . . . . . . . . . . . . . . . . . . . . 761<br />

xx


Table of Contents<br />

Thermal cracking process . . . . . . . . . . . . . . . . . . . . . . 765<br />

Tidal power plant . . . . . . . . . . . . . . . . . . . . . . . . . . 770<br />

Touch-tone telephone . . . . . . . . . . . . . . . . . . . . . . . . 774<br />

Transistor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 778<br />

Transistor radio . . . . . . . . . . . . . . . . . . . . . . . . . . . 786<br />

Tuberculosis vaccine. . . . . . . . . . . . . . . . . . . . . . . . . 791<br />

Tungsten filament . . . . . . . . . . . . . . . . . . . . . . . . . . 795<br />

Tupperware. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 799<br />

Turbojet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 807<br />

Typhus vaccine . . . . . . . . . . . . . . . . . . . . . . . . . . . . 811<br />

Ultracentrifuge . . . . . . . . . . . . . . . . . . . . . . . . . . . . 815<br />

Ultramicroscope . . . . . . . . . . . . . . . . . . . . . . . . . . . 819<br />

Ultrasound . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 823<br />

UNIVAC computer . . . . . . . . . . . . . . . . . . . . . . . . . 828<br />

Vacuum cleaner . . . . . . . . . . . . . . . . . . . . . . . . . . . 832<br />

Vacuum tube . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 837<br />

Vat dye . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 842<br />

Velcro . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 846<br />

Vending machine slug rejector . . . . . . . . . . . . . . . . . . . 850<br />

Videocassette recorder . . . . . . . . . . . . . . . . . . . . . . . 857<br />

Virtual machine . . . . . . . . . . . . . . . . . . . . . . . . . . . 861<br />

Virtual reality. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 866<br />

V-2 rocket . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 871<br />

Walkman cassette player . . . . . . . . . . . . . . . . . . . . . . 875<br />

Washing machine . . . . . . . . . . . . . . . . . . . . . . . . . . 883<br />

Weather satellite . . . . . . . . . . . . . . . . . . . . . . . . . . . 887<br />

Xerography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 891<br />

X-ray crystallography . . . . . . . . . . . . . . . . . . . . . . . . 896<br />

X-ray image intensifier . . . . . . . . . . . . . . . . . . . . . . . 901<br />

Yellow fever vaccine . . . . . . . . . . . . . . . . . . . . . . . . . 905<br />

Time Line . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 909<br />

Topics by Category . . . . . . . . . . . . . . . . . . . . . . . . . 915<br />

Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 923<br />

xxi


This Page Intentionally Left Blank


<strong>Inventions</strong><br />

<strong>and</strong><br />

<strong>Inventors</strong>


This Page Intentionally Left Blank


Abortion pill<br />

Abortion pill<br />

The invention: RU-486 was the first commercially available drug<br />

that prevented fertilized eggs from implanting themselves in the<br />

walls of women’s uteruses.<br />

The people behind the invention:<br />

Étienne-Émile Baulieu (1926- ), a French biochemist <strong>and</strong><br />

endocrinologist<br />

Georges Teutsch, a French chemist<br />

Alain Bélanger, a French chemist<br />

Daniel Philibert, a French physicist <strong>and</strong> pharmacologist<br />

Developing <strong>and</strong> Testing<br />

In 1980, Alain Bélanger, a research chemist, was working with<br />

Georges Teutsch at Roussel Uclaf, a French pharmaceutical company.<br />

Teutsch <strong>and</strong> Bélanger were interested in underst<strong>and</strong>ing how<br />

changes in steroids affect the chemicals’ ability to bind to their steroid<br />

receptors. (Receptors are molecules on cells that can bind with<br />

certain chemical substances such as hormones. Receptors therefore<br />

act as connecting links to promote or prevent specific bodily activities<br />

or processes.) Bélanger synthesized several steroids that bonded<br />

to steroid receptors. Among these steroids was a compound that<br />

came to be called “RU-486.”<br />

Another member of the research project, Daniel Philibert, found<br />

that RU-486 blocked the activities of progesterone by binding tightly<br />

to the progesterone receptor. Progesterone is a naturally occurring<br />

steroid hormone that prepares the wall of the uterus to accept a fertilized<br />

egg. Once this is done, the egg can become implanted <strong>and</strong><br />

can begin to develop. The hormone also prevents the muscles of the<br />

uterus from contracting, which might cause the uterus to reject the<br />

egg. Therefore RU-486, by acting as a kind of shield between hormone<br />

<strong>and</strong> receptor, essentially stopped the progesterone from doing<br />

its job.<br />

At the time, Teutsch’s group did not consider that RU-486 might<br />

be useful for deliberately interrupting human pregnancy. It was<br />

1


2 / Abortion pill<br />

Étienne-Émile Baulieu, a biochemist <strong>and</strong> endocrinologist <strong>and</strong> a consultant<br />

for Roussel Uclaf, who made this connection. He persuaded<br />

the company to test RU-486 for its effects on fertility control.<br />

Many tests were performed on rabbits, rats, <strong>and</strong> monkeys; they<br />

showed that, even in the presence of progesterone, RU-486 could<br />

prevent secretory tissue from forming in the uterus, could change<br />

the timing of the menstrual cycle, <strong>and</strong> could terminate a pregnancy—that<br />

is, cause an abortion. The compound also seemed to be<br />

nontoxic, even in high doses.<br />

In October of 1981, Baulieu began testing the drug with human<br />

volunteers. By 1985, major tests of RU-486 were being done in<br />

Étienne-Emile Baulieu<br />

Étienne-Émile Baulieu was born in Strasbourg, France, in<br />

1926. He moved to Paris for his advanced studies at the Faculty<br />

of Medicine <strong>and</strong> Faculty of Science of Pasteur College. He was<br />

an Intern of Paris from 1951 until he received a medical degree<br />

in 1955. He passed examinations qualifying him to become a<br />

teacher at state schools in 1958 <strong>and</strong> during the 1961-1962 academic<br />

year was a visiting scientist in Columbia University’s<br />

Department of Biochemistry.<br />

In 1963 Baulieu was made a Doctor of Science <strong>and</strong> appointed<br />

director of a research unit at France’s National Institute of<br />

Health <strong>and</strong> Medical Science, a position he held until he retired<br />

in 1997. He also served as Head of Service of Hormonal Biochemistry<br />

of the Hospital of Bicêtre (1970-1997), professor of<br />

biochemistry at University of Paris-South (1970-1993), <strong>and</strong> consultant<br />

for Roussel Uclaf (1963-1997).<br />

Among his many honors are the Gregory Pincus Memorial<br />

Award (1978), awards from the National Academy of Medicine,<br />

the Christopher Columbus Discovery Award in Biomedical Research<br />

(1992), the Joseph Bolivar DeLee Humanitarian Award<br />

(1994), <strong>and</strong> Comm<strong>and</strong>er of the Legion of Honor (1990). Although<br />

busy with research <strong>and</strong> teaching duties, Baulieu was on<br />

the editorial board of several French <strong>and</strong> international newspapers,<br />

a member of scientific councils, <strong>and</strong> a participant in the<br />

Special Program in Human Reproduction of the World Health<br />

Organization.


France, Great Britain, The Netherl<strong>and</strong>s, Sweden, <strong>and</strong> China. When a<br />

relatively low dose of RU-486 was given orally, there was an 85 percent<br />

success rate in ending pregnancy; the woman’s body expelled<br />

the embryo <strong>and</strong> all the endometrial surface. Researchers found that<br />

if a low dose of a prostagl<strong>and</strong>in (a hormonelike substance that<br />

causes the smooth muscles of the uterus to contract, thereby expelling<br />

the embryo) was given two days later, the success rate rose to 96<br />

percent. There were few side effects, <strong>and</strong> the low doses of RU-486<br />

did not interfere with the actions of other steroid hormones that are<br />

necessary to keep the body working.<br />

In the March, 1990, issue of The New Engl<strong>and</strong> Journal of Medicine,<br />

Baulieu <strong>and</strong> his coworkers reported that with one dose of RU-486,<br />

followed in thirty-six to forty-eight hours with a low dose of prostagl<strong>and</strong>in,<br />

96 percent of the 2,040 women they studied had a complete<br />

abortion with few side effects. The women were monitored after receiving<br />

the prostagl<strong>and</strong>in to watch for side effects, which included<br />

nausea, vomiting, abdominal pain, <strong>and</strong> diarrhea. When they returned<br />

for a later checkup, fewer than 2 percent of the women complained<br />

of side effects. The researchers used two different prostagl<strong>and</strong>ins;<br />

they found that one caused a quicker abortion but also<br />

brought about more pain <strong>and</strong> a longer period of bleeding.<br />

Using the Drug<br />

Abortion pill / 3<br />

In September, 1988, the French government approved the distribution<br />

of RU-486 for use in government-controlled clinics. The next<br />

month, however, Roussel Uclaf stopped selling the drug because<br />

people opposed to abortion did not want RU-486 to be available <strong>and</strong><br />

were threatening to boycott the company.<br />

Then, however, there were threats <strong>and</strong> pressure from the other<br />

side. For example, members of the World Congress of Obstetrics<br />

<strong>and</strong> Gynecology announced that they might boycott Roussel Uclaf<br />

if it did not make RU-486 available. The French government, which<br />

controlled a 36 percent interest in Roussel Uclaf, ordered the company<br />

to start distributing the drug once more.<br />

By the fall of 1989, more than one-fourth of all early abortions in<br />

France were being done with RU-486 <strong>and</strong> a prostagl<strong>and</strong>in. The French<br />

government began helping to pay the cost of using RU-486 in 1990.


4 / Abortion pill<br />

Testing for approval of RU-486 was completed in Great Britain<br />

<strong>and</strong> The Netherl<strong>and</strong>s, but Roussel Uclaf’s parent company, Hoechst<br />

AG, did not try to market the drug there or in any other country outside<br />

France. (In the United States, government regulations did not<br />

allow RU-486 to be tested using government funds.)<br />

Medical researchers believe that RU-486 may be useful not only<br />

for abortions but also in other ways. For example, it may help in<br />

treating certain breast cancers <strong>and</strong> other tumors. RU-486 is also being<br />

investigated as a possible treatment for glaucoma—to lower<br />

pressure in the eye that may be caused by a high level of steroid hormone.<br />

It may be useful in promoting the healing of skin wounds<br />

<strong>and</strong> softening the cervix at birth, easing delivery. Researchers hope<br />

as well that some form of RU-486 may prove useful as a contraceptive—that<br />

is, not to prevent a fertilized egg from implanting itself in<br />

the mother’s uterus but to prevent ovulation in the first place.<br />

Impact<br />

Groups opposed to abortion rights have spoken out against RU-<br />

486, while those who favor the right to abortion have urged its acceptance.<br />

The drug has been approved for use in China as well as in<br />

France. In the United States, however, the government has avoided<br />

giving its approval to the drug. Officials of the World Health Organization<br />

(WHO) have argued that RU-486 could prevent the deaths<br />

of women who undergo botched abortions. Under international<br />

law, WHO has the right to take control of the drug <strong>and</strong> make it available<br />

in poor countries at low cost. Because of the controversy surrounding<br />

the drug, however, WHO called for more testing to ensure<br />

that RU-486 is quite safe for women.<br />

See also Amniocentesis; Antibacterial drugs; Artificial hormone;<br />

Birth control pill; Salvarsan.<br />

Further Reading<br />

Baulieu, Etienne-Emile, <strong>and</strong> Mort Rosenblum. The “Abortion Pill”:<br />

RU-486, a Woman’s Choice. New York: Simon & Schuster, 1991.<br />

Butler, John Douglas, <strong>and</strong> David F. Walbert. Abortion, Medicine, <strong>and</strong><br />

the Law. 4th ed. New York: Facts on File, 1992.


Abortion pill / 5<br />

Lyall, Sarah. “Britain Allows Over-the-Counter Sales of Morning-<br />

After Pill.” New York Times (January 15, 2001).<br />

McCuen, Gary E. RU 486: The Abortion Pill Controversy. Hudson,<br />

Wis.: GEM <strong>Public</strong>ations, 1992.<br />

Nemecek, Sasha. “The Second Abortion Pill.” Scientific American<br />

283, no. 6 (December, 2000).<br />

Zimmerman, Rachel. “Ads for Controversial Abortion Pill Set to<br />

Appear in National Magazines.” Wall Street Journal (May 23,<br />

2001).


6<br />

Airplane<br />

Airplane<br />

The invention: The first heavier-than-air craft to fly, the airplane<br />

revolutionized transportation <strong>and</strong> symbolized the technological<br />

advances of the twentieth century.<br />

The people behind the invention:<br />

Wilbur Wright (1867-1912), an American inventor<br />

Orville Wright (1871-1948), an American inventor<br />

Octave Chanute (1832-1910), a French-born American civil<br />

engineer<br />

A Careful Search<br />

Although people have dreamed about flying since the time of the<br />

ancient Greeks, it was not until the late eighteenth century that hotair<br />

balloons <strong>and</strong> gliders made human flight possible. It was not until<br />

the late nineteenth century that enough experiments had been done<br />

with kites <strong>and</strong> gliders that people could begin to think seriously<br />

about powered, heavier-than-air flight. Two of these people were<br />

Wilbur <strong>and</strong> Orville Wright.<br />

The Wright brothers making their first successful powered flight, at Kitty Hawk, North<br />

Carolina. (Library of Congress)


The Wright brothers were more than just tinkerers who accidentally<br />

found out how to build a flying machine. In 1899, Wilbur wrote<br />

the Smithsonian Institution for a list of books to help them learn<br />

about flying. They used the research of people such as George<br />

Cayley, Octave Chanute, Samuel Langley, <strong>and</strong> Otto Lilienthal to<br />

help them plan their own experiments with birds, kites, <strong>and</strong> gliders.<br />

They even built their own wind tunnel. They never fully trusted the<br />

results of other people’s research, so they repeated the experiments<br />

of others <strong>and</strong> drew their own conclusions. They shared these results<br />

with Octave Chanute, who was able to offer them lots of good advice.<br />

They were continuing a tradition of excellence in engineering<br />

that began with careful research <strong>and</strong> avoided dangerous trial <strong>and</strong><br />

error.<br />

Slow Success<br />

Airplane / 7<br />

Before the brothers had set their minds to flying, they had built<br />

<strong>and</strong> repaired bicycles. This was a great help to them when they put<br />

their research into practice <strong>and</strong> actually built an airplane. From<br />

building bicycles, they knew how to work with wood <strong>and</strong> metal to<br />

make a lightweight but sturdy machine. Just as important, from riding<br />

bicycles, they got ideas about how an airplane needed to work.<br />

They could see that both bicycles <strong>and</strong> airplanes needed to be fast<br />

<strong>and</strong> light. They could also see that airplanes, like bicycles, needed to<br />

be kept under constant control to stay balanced, <strong>and</strong> that this control<br />

would probably take practice. This was a unique idea. Instead<br />

of building something solid that was controlled by levers <strong>and</strong> wheels<br />

like a car, the Wright brothers built a flexible airplane that was controlled<br />

partly by the movement of the pilot, like a bicycle.<br />

The result was the 1903 Wright Flyer. The Flyer had two sets of<br />

wings, one above the other, which were about 12 meters from tip to<br />

tip. They made their own 12-horsepower engine, as well as the two<br />

propellers the engine spun. The craft had skids instead of wheels.<br />

On December 14, 1903, the Wright brothers took the Wright Flyer to<br />

the shores of Kitty Hawk, North Carolina, where Wilbur Wright<br />

made the first attempt to fly the airplane.<br />

The first thing Wilbur found was that flying an airplane was not<br />

as easy as riding a bicycle. One wrong move sent him tumbling into


8 / Airplane<br />

The Wright Brothers<br />

Orville <strong>and</strong> his older brother Wilbur first got interested in<br />

aircraft when their father gave them a toy helicopter in 1878.<br />

Theirs was a large, supportive family. Their father, a minister,<br />

<strong>and</strong> their mother, a college graduate <strong>and</strong> inventor of household<br />

gadgets, encouraged all five of the children to be creative. Although<br />

Wilbur, born in 1867, was four years older than Orville,<br />

they were close as children. While in high school, they put out a<br />

weekly newspaper together, West Side News, <strong>and</strong> they opened<br />

their bicycle shop in 1892. Orville was the mechanically adept<br />

member of the team, the tinkerer; Wilbur was the deliberative<br />

one, the planner <strong>and</strong> designer.<br />

Since the bicycle business was seasonal, they had time to<br />

pursue their interest in aircraft, puzzling out the technical problems<br />

<strong>and</strong> studying the successes <strong>and</strong> failures of others. They<br />

started with gliders, flying their first, which had a five-foot<br />

wing span, in 1899. They developed their own technique to control<br />

the gliders, the “wing-warping technique,” after watching<br />

how birds fly. They attached wires to the trailing edges of the<br />

wings <strong>and</strong> pulled the wires to deform the wings’ shape. They<br />

built a sixteen-foot glider in 1900 <strong>and</strong> spent a vacation in North<br />

Carolina gaining flying experience. Further designs <strong>and</strong> many<br />

more tests followed, including more than two hundred shapes<br />

of wing studied in their home-built wind tunnel, before their<br />

first successful engine-powered flight in 1903.<br />

Neither man ever married. After Wilbur died of typhoid in<br />

1912, Orville was stricken by the loss of his brother but continued<br />

to run their business until 1915. He last piloted an airplane<br />

himself in 1918 <strong>and</strong> died thirty years later.<br />

Their first powered airplane, the Wright Flyer, lives on at the<br />

National Air <strong>and</strong> Space Museum in Washington, D.C. Small<br />

parts from the aircraft were taken to the Moon by Neil Armstrong<br />

<strong>and</strong> Edwin Aldrin when they made the first l<strong>and</strong>ing<br />

there in 1969.<br />

the s<strong>and</strong> only moments after takeoff. Wilbur was not seriously hurt,<br />

but a few more days were needed to repair the Wright Flyer.<br />

On December 17, 1903, at 10:35 a.m., after eight years of research<br />

<strong>and</strong> planning, Orville Wright took to the air for a historic twelve sec-


onds. He covered 37 meters of ground <strong>and</strong> 152 meters of air space.<br />

Both brothers took two flights that morning. On the fourth flight,<br />

Wilbur flew for fifty-nine seconds over 260 meters of ground <strong>and</strong><br />

through more than 800 meters of air space. After he had l<strong>and</strong>ed, a<br />

sudden gust of wind struck the plane, damaging it beyond repair.<br />

Yet no one was able to beat their record for three years.<br />

Impact<br />

Airplane / 9<br />

Those first flights in 1903 got little publicity. Only a few people,<br />

such as Octave Chanute, understood the significance of the Wright<br />

brothers’ achievement. For the next two years, they continued to<br />

work on their design, <strong>and</strong> by 1905 they had built the Wright Flyer III.<br />

Although Chanute tried to get them to enter flying contests, the<br />

brothers decided to be cautious <strong>and</strong> try to get their machine patented<br />

first, so that no one would be able to steal their ideas.<br />

News of their success spread slowly through the United States<br />

<strong>and</strong> Europe, giving hope to others who were working on airplanes<br />

of their own. When the Wright brothers finally went public with the<br />

Wright Flyer III, they inspired many new advances. By 1910, when<br />

the brothers started flying in air shows <strong>and</strong> contests, their feats were<br />

matched by another American, Glen Hammond Curtiss. The age of<br />

the airplane had arrived.<br />

Later in the decade, the Wright brothers began to think of military<br />

uses for their airplanes. They signed a contract with the U.S.<br />

Army Signal Corps <strong>and</strong> agreed to train military pilots.<br />

Aside from these achievements, the brothers from Dayton, Ohio,<br />

set the st<strong>and</strong>ard for careful research <strong>and</strong> practical experimentation.<br />

They taught the world not only how to fly but also how to design<br />

airplanes. Indeed, their methods of purposeful, meaningful, <strong>and</strong><br />

highly organized research had an impact not only on airplane design<br />

but also on the field of aviation science in general.<br />

See also Bullet train; Cruise missile; Dirigible; Gas-electric car;<br />

Propeller-coordinated machine gun; Rocket; Stealth aircraft; Supersonic<br />

passenger plane; Turbojet; V-2 rocket.


10 / Airplane<br />

Further Reading<br />

Brady, Tim. The American Aviation Experience: A History. Carbondale:<br />

Southern Illinois University Press, 2000.<br />

Chanute, Octave, Marvin Wilks, Orville Wright, <strong>and</strong> Wilbur Wright.<br />

The Papers of Wilbur <strong>and</strong> Orville Wright: Including the Chanute-<br />

Wright Letters <strong>and</strong> Other Papers of Octave Chanute. New York:<br />

McGraw-Hill, 2000.<br />

Culik, Fred, <strong>and</strong> Spencer Dunmore. On Great White Wings: The<br />

Wright Brothers <strong>and</strong> the Race for Flight. Toronto: McArthur, 2001.<br />

Howard, Fred. Wilbur <strong>and</strong> Orville: A Biography of the Wright Brothers.<br />

Mineola, N.Y.: Dover <strong>Public</strong>ations, 1998.


Alkaline storage battery<br />

Alkaline storage battery<br />

The invention: The nickel-iron alkaline battery was a lightweight,<br />

inexpensive portable power source for vehicles with electric motors.<br />

The people behind the invention:<br />

Thomas Alva Edison (1847-1931), American chemist, inventor,<br />

<strong>and</strong> industrialist<br />

Henry Ford (1863-1947), American inventor <strong>and</strong> industrialist<br />

Charles F. Kettering (1876-1958), American engineer <strong>and</strong><br />

inventor<br />

A Three-Way Race<br />

The earliest automobiles were little more than pairs of bicycles<br />

harnessed together within a rigid frame, <strong>and</strong> there was little agreement<br />

at first regarding the best power source for such contraptions.<br />

The steam engine, which was well established for railroad <strong>and</strong> ship<br />

transportation, required an external combustion area <strong>and</strong> a boiler.<br />

Internal combustion engines required h<strong>and</strong> cranking, which could<br />

cause injury if the motor backfired. Electric motors were attractive<br />

because they did not require the burning of fuel, but they required<br />

batteries that could store a considerable amount of energy <strong>and</strong><br />

could be repeatedly recharged. Ninety percent of the motorcabs in<br />

use in New York City in 1899 were electrically powered.<br />

The first practical storage battery, which was invented by the<br />

French physicist Gaston Planté in 1859, employed electrodes (conductors<br />

that bring electricity into <strong>and</strong> out of a conducting medium)<br />

of lead <strong>and</strong> lead oxide <strong>and</strong> a sulfuric acid electrolyte (a solution<br />

that conducts electricity). In somewhat improved form, this<br />

remained the only practical rechargeable battery at the beginning<br />

of the twentieth century. Edison considered the lead acid cell (battery)<br />

unsuitable as a power source for electric vehicles because using<br />

lead, one of the densest metals known, resulted in a heavy<br />

battery that added substantially to the power requirements of a<br />

motorcar. In addition, the use of an acid electrolyte required that<br />

11


12 / Alkaline storage battery<br />

the battery container be either nonmetallic or coated with a nonmetal<br />

<strong>and</strong> thus less dependable than a steel container.<br />

The Edison Battery<br />

In 1900, Edison began experiments aimed at developing a rechargeable<br />

battery with inexpensive <strong>and</strong> lightweight metal electrodes <strong>and</strong> an<br />

alkaline electrolyte so that a metal container could be used. He had already<br />

been involved in manufacturing the nonrechargeable battery<br />

known as the Lal<strong>and</strong>e cell, which had zinc <strong>and</strong> copper oxide electrodes<br />

<strong>and</strong> a highly alkaline sodium hydroxide electrolyte. Zinc electrodes<br />

could not be used in a rechargeable cell because the zinc would<br />

dissolve in the electrolyte. The copper electrode also turned out to be<br />

unsatisfactory. After much further experimentation, Edison settled<br />

on the nickel-iron system for his new storage battery. In this system,<br />

the power-producing reaction involved the conversion of nickel oxide<br />

to nickel hydroxide together with the oxidation of iron metal to<br />

iron oxide, with both materials in contact with a potassium hydroxide<br />

solution. When the battery was recharged, the nickel hydroxide<br />

was converted into oxide <strong>and</strong> the iron oxide was converted back to<br />

the pure metal.<br />

Although the basic ingredients<br />

of the Edison cell were<br />

inexpensive, they could not readily<br />

be obtained in adequate purity<br />

for battery use. Edison set<br />

up a new chemical works to<br />

prepare the needed materials.<br />

He purchased impure nickel alloy,<br />

which was then dissolved<br />

in acid, purified, <strong>and</strong> converted<br />

to the hydroxide. He prepared<br />

pure iron powder by using a<br />

multiple-step process. For use<br />

in the battery, the reactant powders<br />

had to be packed in pockets<br />

made of nickel-plated steel<br />

Thomas A. Edison. (Library of Congress)<br />

that had been perforated to al-


Alkaline storage battery / 13<br />

low the iron <strong>and</strong> nickel powders to come into contact with the electrolyte.<br />

Because the nickel compounds were poor electrical conductors,<br />

a flaky type of graphite was mixed with the nickel hydroxide at<br />

this stage.<br />

Sales of the new Edison storage battery began in 1904, but within<br />

six months it became apparent that the battery was subject to losses<br />

in power <strong>and</strong> a variety of other defects. Edison took the battery off<br />

Thomas Alva Edison<br />

Thomas Alva Edison (1847-1931) was America’s most famous<br />

<strong>and</strong> prolific inventor. His astonishing success story, rising<br />

from a home-schooled child who worked as a newsboy to<br />

a leader in American industry, was celebrated in children’s<br />

books, biographies, <strong>and</strong> movies. Corporations still bear his<br />

name, <strong>and</strong> his inventions <strong>and</strong> improvements of others’ inventions—such<br />

as the light bulb, phonograph, <strong>and</strong> motion picture—shaped<br />

the way Americans live, work, <strong>and</strong> entertain<br />

themselves. The U.S. Patent Office issued Edison 1,093 patents<br />

during his lifetime, the most granted to one person.<br />

Hailed as a genius, Edison himself emphasized the value of<br />

plain determination. Genius is one percent inspiration <strong>and</strong> 99<br />

percent perspiration, he insisted. He also understood the value<br />

of working with others. In fact, one of his greatest contributions<br />

to American technology involved organized research. At age<br />

twenty-three he sold the rights to his first major invention,<br />

an improved ticker-tape machine for Wall Street brokers, for<br />

$40,000. He invested the money in building an industrial research<br />

laboratory, the first ever. It led to his large facilities at<br />

Menlo Park, New Jersey, <strong>and</strong>, later, labs in other locations. At<br />

times as many as one hundred people worked for him, some of<br />

whom, such as Nikola Tesla <strong>and</strong> Reginald Fessenden, became<br />

celebrated inventors in their own right.<br />

At his labs Edison not only developed electrical items, such<br />

as the light bulb <strong>and</strong> storage battery; he also produced an efficient<br />

mimeograph <strong>and</strong> worked on innovations in metallurgy,<br />

organic chemistry, photography <strong>and</strong> motion pictures, <strong>and</strong> phonography.<br />

The phonograph, he once said, was his favorite invention.<br />

Edison never stopped working. He was still receiving patents<br />

the year he died.


14 / Alkaline storage battery<br />

the market in 1905 <strong>and</strong> offered full-price refunds for the defective<br />

batteries. Not a man to ab<strong>and</strong>on an invention, however, he spent the<br />

next five years examining the failed batteries <strong>and</strong> refining his design.<br />

He discovered that the repeated charging <strong>and</strong> discharging of<br />

the battery caused a shift in the distribution of the graphite in the<br />

nickel hydroxide electrode. By using a different type of graphite, he<br />

was able to eliminate this problem <strong>and</strong> produce a very dependable<br />

power source.<br />

The Ford Motor Company, founded by Henry Ford, a former<br />

Edison employee, began the large-scale production of gasolinepowered<br />

automobiles in 1903 <strong>and</strong> introduced the inexpensive, easyto-drive<br />

Model T in 1908. The introduction of the improved Edison<br />

battery in 1910 gave a boost to electric car manufacturers, but their<br />

new position in the market would be short-lived. In 1911, Charles<br />

Kettering invented an electric starter for gasoline-powered vehicles<br />

that eliminated the need for troublesome <strong>and</strong> risky h<strong>and</strong> cranking.<br />

By 1915, this device was available on all gasoline-powered automobiles,<br />

<strong>and</strong> public interest in electrically powered cars rapidly diminished.<br />

Although the Kettering starter required a battery, it required<br />

much less capacity than an electric motor would have <strong>and</strong> was almost<br />

ideally suited to the six-volt lead-acid battery.<br />

Impact<br />

Edison lost the race to produce an electrical power source that<br />

would meet the needs of automotive transportation. Instead, the internal<br />

combustion engine developed by Henry Ford became the st<strong>and</strong>ard.<br />

Interest in electrically powered transportation diminished as<br />

immense reserves of crude oil, from which gasoline could be obtained,<br />

were discovered first in the southwestern United States <strong>and</strong><br />

then on the Arabian peninsula. Nevertheless, the Edison cell found<br />

a variety of uses <strong>and</strong> has been manufactured continuously throughout<br />

most of the twentieth century much as Edison designed it.<br />

Electrically powered trucks proved to be well suited for local deliveries,<br />

<strong>and</strong> some department stores maintained fleets of such<br />

trucks into the mid-1920’s. Electrical power is still preferable to internal<br />

combustion for indoor use, where exhaust fumes are a significant<br />

problem, so forklifts in factories <strong>and</strong> passenger transport vehi-


cles at airports still make use of the Edison-type power source. The<br />

Edison battery also continues to be used in mines, in railway signals,<br />

in some communications equipment, <strong>and</strong> as a highly reliable<br />

source of st<strong>and</strong>by emergency power.<br />

See also Compressed-air-accumulating power plant; Internal<br />

combustion engine; Photoelectric cell; Photovoltaic cell.<br />

Further Reading<br />

Alkaline storage battery / 15<br />

Baldwin, Neil. Edison: Inventing the Century. Chicago: University of<br />

Chicago Press, 2001.<br />

Boyd, Thomas Alvin. Professional Amateur: The Biography of Charles<br />

Franklin Kettering. New York: Arno Press, 1972.<br />

Bryan, Ford R. Beyond the Model T: The Other Ventures of Henry Ford.<br />

Rev. ed. Detroit: Wayne State University Press, 1997.<br />

Cramer, Carol. Thomas Edison. San Diego, Calif.: Greenhaven Press,<br />

2001.<br />

Israel, Paul. Edison: A Life of Invention. New York: Wiley, 2000.


16<br />

Ammonia<br />

Ammonia<br />

The invention: The first successful method for converting nitrogen<br />

from the atmosphere <strong>and</strong> combining it with hydrogen to synthesize<br />

ammonia, a valuable compound used as a fertilizer.<br />

The person behind the invention:<br />

Fritz Haber (1868-1934), a German chemist who won the 1918<br />

Nobel Prize in Chemistry<br />

The Need for Nitrogen<br />

The nitrogen content of the soil, essential to plant growth, is<br />

maintained normally by the deposition <strong>and</strong> decay of old vegetation<br />

<strong>and</strong> by nitrates in rainfall. If, however, the soil is used extensively<br />

for agricultural purposes, more intensive methods must be used to<br />

maintain soil nutrients such as nitrogen. One such method is crop<br />

rotation, in which successive divisions of a farm are planted in rotation<br />

with clover, corn, or wheat, for example, or allowed to lie fallow<br />

for a year or so. The clover is able to absorb nitrogen from the air <strong>and</strong><br />

deposit it in the soil through its roots. As population has increased,<br />

however, farming has become more intensive, <strong>and</strong> the use of artificial<br />

fertilizers—some containing nitrogen—has become almost universal.<br />

Nitrogen-bearing compounds, such as potassium nitrate <strong>and</strong><br />

ammonium chloride, have been used for many years as artificial fertilizers.<br />

Much of the nitrate used, mainly potassium nitrate, came<br />

from Chilean saltpeter, of which a yearly amount of half a million<br />

tons was imported at the beginning of the twentieth century into<br />

Europe <strong>and</strong> the United States for use in agriculture. Ammonia was<br />

produced by dry distillation of bituminous coal <strong>and</strong> other lowgrade<br />

fuel materials. Originally, coke ovens discharged this valuable<br />

material into the atmosphere, but more economical methods<br />

were found later to collect <strong>and</strong> condense these ammonia-bearing<br />

vapors.<br />

At the beginning of the twentieth century, Germany had practically<br />

no source of fertilizer-grade nitrogen; almost all of its supply


came from the deserts of northern Chile. As dem<strong>and</strong> for nitrates increased,<br />

it became apparent that the supply from these vast deposits<br />

would not be enough. Other sources needed to be found, <strong>and</strong> the almost<br />

unlimited supply of nitrogen in the atmosphere (80 percent nitrogen)<br />

was an obvious source.<br />

Temperature <strong>and</strong> Pressure<br />

Ammonia / 17<br />

When Fritz Haber <strong>and</strong> coworkers began his experiments on ammonia<br />

production in 1904, Haber decided to repeat the experiments<br />

of the British chemist Sir William Ramsay <strong>and</strong> Sydney Young, who<br />

in 1884 had studied the decomposition of ammonia at about 800 degrees<br />

Celsius. They had found that a certain amount of ammonia<br />

was always left undecomposed. In other words, the reaction between<br />

ammonia <strong>and</strong> its constituent elements—nitrogen <strong>and</strong> hydrogen—had<br />

reached a state of equilibrium.<br />

Haber decided to determine the point at which this equilibrium<br />

took place at temperatures near 1,000 degrees Celsius. He tried several<br />

approaches, reacting pure hydrogen with pure nitrogen, <strong>and</strong><br />

starting with pure ammonia gas <strong>and</strong> using iron filings as a catalyst.<br />

(Catalytic agents speed up a reaction without affecting it otherwise).<br />

Having determined the point of equilibrium, he next tried different<br />

catalysts <strong>and</strong> found nickel to be as effective as iron, <strong>and</strong> calcium<br />

<strong>and</strong> manganese even better. At 1,000 degrees Celsius, the rate of reaction<br />

was enough to produce practical amounts of ammonia continuously.<br />

Further work by Haber showed that increasing the pressure also<br />

increased the percentage of ammonia at equilibrium. For example,<br />

at 300 degrees Celsius, the percentage of ammonia at equilibrium at<br />

1 atmosphere of pressure was very small, but at 200 atmospheres,<br />

the percentage of ammonia at equilibrium was far greater. A pilot<br />

plant was constructed <strong>and</strong> was successful enough to impress a<br />

chemical company, Badische Anilin-und Soda-Fabrik (BASF). BASF<br />

agreed to study Haber’s process <strong>and</strong> to investigate different catalysts<br />

on a large scale. Soon thereafter, the process became a commercial<br />

success.


(Nobel Foundation)<br />

18 / Ammonia<br />

Impact<br />

Fritz Haber<br />

Fritz Haber’s career is a warning to inventors: Beware of<br />

what you create, even if your intentions are honorable.<br />

Considered a leading chemist of his age, Haber was born in<br />

Breslau (now Wroclaw, Pol<strong>and</strong>) in 1868. A brilliant student, he<br />

earned a doctorate quickly, specializing in organic chemistry,<br />

<strong>and</strong> briefly worked as an industrial chemist. Although he soon<br />

took an academic job, throughout his career Haber believed<br />

that science must benefit society—new theoretical discoveries<br />

must find practical applications.<br />

Beginning in 1904, he applied new chemical techniques<br />

to fix atmospheric nitrogen in the form of ammonia.<br />

Nitrogen in the form of nitrates was urgently<br />

sought because nitrates were necessary to fertilize<br />

crops <strong>and</strong> natural sources were becoming rare. Only<br />

artificial nitrates could sustain the amount of agriculture<br />

needed to feed exp<strong>and</strong>ing populations. In 1908<br />

Haber succeeded in finding an efficient, cheap process<br />

to make ammonia <strong>and</strong> convert it to nitrates, <strong>and</strong><br />

by 1910 German manufacturers had built large plants<br />

to exploit his techniques. He was lauded as a great benefactor to<br />

humanity.<br />

However, his efforts to help Germany during World War I,<br />

even though he hated war, turned his life into a nightmare. His<br />

wife committed suicide because of his chlorine gas research,<br />

which also poisoned his international reputation <strong>and</strong> tainted<br />

his 1918 Nobel Prize in Chemistry. After the war he redirected<br />

his energies to helping Germany rebuild its economy. Eight<br />

years of experiments in extracting gold from seawater ended in<br />

failure, but he did raise the Kaiser Wilhelm Institute for Physical<br />

Chemistry, which he directed, to international prominence.<br />

Nonetheless, Haber had to flee Adolf Hitler’s Nazi regime in<br />

1933 <strong>and</strong> died a year later, better known for his war research<br />

than for his fundamental service to agriculture <strong>and</strong> industry.<br />

With the beginning of World War I, nitrates were needed more<br />

urgently for use in explosives than in agriculture. After the fall of<br />

Antwerp, 50,000 tons of Chilean saltpeter were discovered in the


harbor <strong>and</strong> fell into German h<strong>and</strong>s. Because the ammonia from<br />

Haber’s process could be converted readily into nitrates, it became<br />

an important war resource. Haber’s other contribution to the German<br />

war effort was his development of poison gas, which was used<br />

for the chlorine gas attack on Allied troops at Ypres in 1915. He also<br />

directed research on gas masks <strong>and</strong> other protective devices.<br />

At the end of the war, the 1918 Nobel Prize in Chemistry was<br />

awarded to Haber for his development of the process for making<br />

synthetic ammonia. Because the war was still fresh in everyone’s<br />

memory, it became one of the most controversial Nobel awards ever<br />

made. A headline in The New York Times for January 26, 1920, stated:<br />

“French Attack Swedes for Nobel Prize Award: Chemistry Honor<br />

Given to Dr. Haber, Inventor of German Asphyxiating Gas.” In a letter<br />

to the Times on January 28, 1920, the Swedish legation in Washington,<br />

D.C., defended the award.<br />

Haber left Germany in 1933 under duress from the anti-Semitic<br />

policies of the Nazi authorities. He was invited to accept a position<br />

with the University of Cambridge, Engl<strong>and</strong>, <strong>and</strong> died on a trip to<br />

Basel, Switzerl<strong>and</strong>, a few months later, a great man whose spirit had<br />

been crushed by the actions of an evil regime.<br />

See also Fuel cell; Refrigerant gas; Silicones; Thermal cracking<br />

process.<br />

Further Reading<br />

Ammonia / 19<br />

Goran, Morris Herbert. The Story of Fritz Haber. Norman: University<br />

of Oklahoma Press, 1967.<br />

Jansen, Sarah. “Chemical-Warfare Techniques for Insect Control: Insect<br />

‘Pests’ in Germany Before <strong>and</strong> After World War I.” Endeavour<br />

24, no. 1 (March, 2000).<br />

Smil, Vaclav. Enriching the Earth: Fritz Haber, Carl Bosch, <strong>and</strong> the Transformation<br />

of World Food Production. Cambridge, Mass.: MIT Press,<br />

2001.


20<br />

Amniocentesis<br />

Amniocentesis<br />

The invention: A technique for removing amniotic fluid from<br />

pregnant women, amniocentesis became a life-saving tool for diagnosing<br />

fetal maturity, health, <strong>and</strong> genetic defects.<br />

The people behind the invention:<br />

Douglas Bevis, an English physician<br />

Aubrey Milunsky (1936- ), an American pediatrician<br />

How Babies Grow<br />

For thous<strong>and</strong>s of years, the inability to see or touch a fetus in the<br />

uterus was a staggering problem in obstetric care <strong>and</strong> in the diagnosis<br />

of the future mental <strong>and</strong> physical health of human offspring. A<br />

beginning to the solution of this problem occurred on February 23,<br />

1952, when The Lancet published a study called “The Antenatal Prediction<br />

of a Hemolytic Disease of the Newborn.” This study, carried<br />

out by physician Douglas Bevis, described the use of amniocentesis<br />

to assess the risk factors found in the fetuses of Rh-negative women<br />

impregnated by Rh-positive men. The article is viewed by many as a<br />

l<strong>and</strong>mark in medicine that led to the wide use of amniocentesis as a<br />

tool for diagnosing fetal maturity, fetal health, <strong>and</strong> fetal genetic<br />

deects.<br />

At the beginning of a human pregnancy (conception) an egg <strong>and</strong><br />

a sperm unite to produce the fertilized egg that will become a new<br />

human being. After conception, the fertilized egg passes from the<br />

oviduct into the uterus, while dividing <strong>and</strong> becoming an organized<br />

cluster of cells capable of carrying out different tasks in the ninemonth-long<br />

series of events leading up to birth.<br />

About a week after conception, the cluster of cells, now a “vesicle”<br />

(a fluid-filled sac containing the new human cells), attaches<br />

to the uterine lining, penetrates it, <strong>and</strong> becomes intimately intertwined<br />

with uterine tissues. In time, the merger between the vesicle<br />

<strong>and</strong> the uterus results in formation of a placenta that connects the<br />

mother <strong>and</strong> the embryo, <strong>and</strong> an amniotic sac filled with the amniotic<br />

fluid in which the embryo floats.


Amniotic Sac<br />

Uterus<br />

Amniotic Fluid<br />

Placenta<br />

Physicians extract amniotic fluid directly from the<br />

womb <strong>and</strong> examine it to determine the health of the<br />

fetus.<br />

Eight weeks after conception,<br />

the embryo (now a<br />

fetus) is about 2.5 centimeters<br />

long <strong>and</strong> possesses<br />

all the anatomic elements<br />

it will have when it is<br />

born. At this time, about<br />

two <strong>and</strong> one-half months<br />

after her last menstruation,<br />

the expectant mother typically<br />

visits a physician <strong>and</strong><br />

finds out she is pregnant.<br />

Also at this time, expecting<br />

mothers often begin to<br />

worry about possible birth<br />

defects in the babies they<br />

carry. Diabetic mothers <strong>and</strong><br />

mothers older than thirty-<br />

five years have higher than usual chances of delivering babies who<br />

have birth defects.<br />

Many other factors inferred from the medical history an expecting<br />

mother provides to her physician can indicate the possible appearance<br />

of birth defects. In some cases, knowledge of possible<br />

physical problems in a fetus may allow their treatment in the uterus<br />

<strong>and</strong> save the newborn from problems that could persist throughout<br />

life or lead to death in early childhood. Information is obtained<br />

through the examination of the amniotic fluid in which the fetus is<br />

suspended throughout pregnancy. The process of obtaining this<br />

fluid is called “amniocentesis.”<br />

Diagnosing Diseases Before Birth<br />

Amniocentesis / 21<br />

Amniocentesis is carried out in several steps. First, the placenta<br />

<strong>and</strong> the fetus are located by the use of ultrasound techniques. Next,<br />

the expecting mother may be given a local anesthetic; a long needle<br />

is then inserted carefully into the amniotic sac. As soon as amniotic<br />

fluid is seen, a small sample (about four teaspoons) is drawn into a<br />

hypodermic syringe <strong>and</strong> the syringe is removed. Amniocentesis is


22 / Amniocentesis<br />

nearly painless, <strong>and</strong> most patients feel only a little abdominal pressure<br />

during the procedure.<br />

The amniotic fluid of early pregnancy resembles blood serum.<br />

As pregnancy continues, its content of substances from fetal urine<br />

<strong>and</strong> other fetal secretions increases. The fluid also contains fetal cells<br />

from skin <strong>and</strong> from the gastrointestinal, reproductive, <strong>and</strong> respiratory<br />

tracts. Therefore, it is of great diagnostic use. Immediately after<br />

the fluid is removed from the fetus, the fetal cells are separated out.<br />

Then, the cells are used for genetic analysis <strong>and</strong> the amniotic fluid is<br />

examined by means of various biochemical techniques.<br />

One important use of the amniotic fluid from amniocentesis is<br />

the determination of its lecithin <strong>and</strong> sphingomyelin content. Lecithins<br />

<strong>and</strong> sphingomyelins are two types of body lipids (fatty molecules)<br />

that are useful diagnostic tools. Lecithins are important because<br />

they are essential components of the so-called pulmonary<br />

surfactant of mature lungs. The pulmonary surfactant acts at lung<br />

surfaces to prevent the collapse of the lung air sacs (alveoli) when a<br />

person exhales.<br />

Subnormal lecithin production in a fetus indicates that it most<br />

likely will exhibit respiratory distress syndrome or a disease called<br />

“hyaline membrane disease” after birth. Both diseases can be fatal,<br />

so it is valuable to determine whether fetal lecithin levels are adequate<br />

for appropriate lung function in the newborn baby. This is<br />

particularly important in fetuses being carried by diabetic mothers,<br />

who frequently produce newborns with such problems. Often, when<br />

the risk of respiratory distress syndrome is identified through amniocentesis,<br />

the fetus in question is injected with hormones that help it<br />

produce mature lungs. This effect is then confirmed by the repeated<br />

use of amniocentesis. Many other problems can also be identified by<br />

the use of amniocentesis <strong>and</strong> corrected before the baby is born.<br />

Consequences<br />

In the years that have followed Bevis’s original observation, many<br />

improvements in the methodology of amniocentesis <strong>and</strong> in the techniques<br />

used in gathering <strong>and</strong> analyzing the genetic <strong>and</strong> biochemical<br />

information obtained have led to good results. Hundreds of debilitating<br />

hereditary diseases can be diagnosed <strong>and</strong> some ameliorated—by


the examination of amniotic fluid <strong>and</strong> fetal cells isolated by amniocentesis.<br />

For many parents who have had a child afflicted by some hereditary<br />

disease, the use of the technique has become a major consideration<br />

in family planning. Furthermore, many physicians recommend strongly<br />

that all mothers over the age of thirty-four be tested by amniocentesis<br />

to assist in the diagnosis of Down syndrome, a congenital but nonhereditary<br />

form of mental deficiency.<br />

There remains the question of whether such solutions are morally<br />

appropriate, but parents—<strong>and</strong> society—now have a choice resulting<br />

from the techniques that have developed since Bevis’s 1952<br />

observation. It is also hoped that these techniques will lead to<br />

means for correcting <strong>and</strong> preventing diseases <strong>and</strong> preclude the need<br />

for considering the therapeutic termination of any pregnancy.<br />

See also Abortion pill; Birth control pill; CAT scanner; Electrocardiogram;<br />

Electroencephalogram; Mammography; Nuclear magnetic<br />

resonance; Pap test; Ultrasound; X-ray image intensifier.<br />

Further Reading<br />

Amniocentesis / 23<br />

Milunsky, Aubrey. Genetic Disorders <strong>and</strong> the Fetus: Diagnosis, Prevention,<br />

<strong>and</strong> Treatment. 3d ed. Baltimore: Johns Hopkins University<br />

Press, 1992.<br />

Rapp, Rayna. Testing Women, Testing the Fetus: The Social Impact of<br />

Amniocentesis in America. New York: Routledge, 1999.<br />

Rothenberg, Karen H., <strong>and</strong> Elizabeth Jean Thomson. Women <strong>and</strong> Prenatal<br />

Testing: Facing the Challenges of Genetic Technology. Columbus:<br />

Ohio State University Press, 1994.<br />

Rothman, Barbara Katz. The Tentative Pregnancy: How Amniocentesis<br />

Changes the Experience of Motherhood. New York: Norton, 1993.


24<br />

Antibacterial drugs<br />

Antibacterial drugs<br />

The invention: Sulfonamides <strong>and</strong> other drugs that have proved effective<br />

in combating many previously untreatable bacterial diseases.<br />

The people behind the invention:<br />

Gerhard Domagk (1895-1964), a German physician who was<br />

awarded the 1939 Nobel Prize in Physiology or Medicine<br />

Paul Ehrlich (1854-1915), a German chemist <strong>and</strong> bacteriologist<br />

who was the cowinner of the 1908 Nobel Prize in Physiology<br />

or Medicine<br />

The Search for Magic Bullets<br />

Although quinine had been used to treat malaria long before the<br />

twentieth century, Paul Ehrlich, who discovered a large number of<br />

useful drugs, is usually considered the father of modern chemotherapy.<br />

Ehrlich was familiar with the technique of using dyes to stain<br />

microorganisms in order to make them visible under a microscope,<br />

<strong>and</strong> he suspected that some of these dyes might be used to poison<br />

the microorganisms responsible for certain diseases without hurting<br />

the patient. Ehrlich thus began to search for dyes that could act<br />

as “magic bullets” that would destroy microorganisms <strong>and</strong> cure<br />

diseases. From 1906 to 1910, Ehrlich tested numerous compounds<br />

that had been developed by the German dye industry. He eventually<br />

found that a number of complex trypan dyes would inhibit the<br />

protozoans that caused African sleeping sickness.<br />

Ehrlich <strong>and</strong> his coworkers also synthesized hundreds of organic<br />

compounds that contained arsenic. In 1910, he found that one of<br />

these compounds, salvarsan, was useful in curing syphilis, a sexually<br />

transmitted disease caused by the bacterium Treponema. This<br />

was an important discovery, because syphilis killed thous<strong>and</strong>s of<br />

people each year. Salvarsan, however, was often toxic to patients,<br />

because it had to be taken in large doses for as long as two years to<br />

effect a cure. Ehrlich thus searched for <strong>and</strong> found a less toxic arsenic<br />

compound, neosalvarsan, which replaced salvarsan in 1912.


In 1915, tartar emetic (a compound containing the metal antimony)<br />

was found to be useful in treating kala-azar, which was<br />

caused by a protozoan. Kala-azar affected millions of people in Africa,<br />

India, <strong>and</strong> Asia, causing much suffering <strong>and</strong> many deaths each<br />

year. Two years later, it was discovered that injection of tartar emetic<br />

into the blood of persons suffering from bilharziasis killed the<br />

flatworms infecting the bladder, liver, <strong>and</strong> spleen. In 1920, suramin,<br />

a colorless compound developed from trypan red, was introduced<br />

to treat African sleeping sickness. It was much less toxic to the patient<br />

than any of the drugs Ehrlich had developed, <strong>and</strong> a single dose<br />

would give protection for more than a month. From the dye methylene<br />

blue, chemists made mepacrine, a drug that was effective<br />

against the protozoans that cause malaria. This chemical was introduced<br />

in 1933 <strong>and</strong> used during World War II; its principal drawback<br />

was that it could cause a patient’s skin to become yellow.<br />

Well Worth the Effort<br />

Antibacterial drugs / 25<br />

Gerhard Domagk had been trained in medicine, but he turned to<br />

research in an attempt to discover chemicals that would inhibit or<br />

kill microorganisms. In 1927, he became director of experimental<br />

pathology <strong>and</strong> bacteriology at the Elberfeld laboratories of the German<br />

chemical firm I. G. Farbenindustrie. Ehrlich’s discovery that<br />

trypan dyes selectively poisoned microorganisms suggested to Domagk<br />

that he look for antimicrobials in a new group of chemicals<br />

known as azo dyes. A number of these dyes were synthesized<br />

from sulfonamides <strong>and</strong> purified by Fritz Mietzsch <strong>and</strong> Josef Klarer.<br />

Domagk found that many of these dyes protected mice infected<br />

with the bacteria Streptococcus pyogenes. In 1932, he discovered that<br />

one of these dyes was much more effective than any tested previously.<br />

This red azo dye containing a sulfonamide was named prontosil<br />

rubrum.<br />

From 1932 to 1935, Domagk began a rigorous testing program to<br />

determine the effectiveness <strong>and</strong> dangers of prontosil use at different<br />

doses in animals. Since all chemicals injected into animals or humans<br />

are potentially dangerous, Domagk determined the doses that<br />

harmed or killed. In addition, he worked out the lowest doses that<br />

would eliminate the pathogen. The firm supplied samples of the


26 / Antibacterial drugs<br />

drug to physicians to carry out clinical trials on humans. (Animal<br />

experimentation can give only an indication of which chemicals<br />

might be useful in humans <strong>and</strong> which doses are required.)<br />

Domagk thus learned which doses were effective <strong>and</strong> safe. This<br />

knowledge saved his daughter’s life. One day while knitting, Domagk’s<br />

daughter punctured her finger with a needle <strong>and</strong> was infected<br />

with a virulent bacteria, which quickly multiplied <strong>and</strong> spread<br />

from the wound into neighboring tissues. In an attempt to alleviate<br />

the swelling, the infected area was lanced <strong>and</strong> allowed to drain, but<br />

this did not stop the infection from spreading. The child became<br />

critically ill with developing septicemia, or blood poisoning.<br />

In those days, more than 75 percent of those who acquired blood<br />

infections died. Domagk realized that the chances for his daughter’s<br />

survival were poor. In desperation, he obtained some of the powdered<br />

prontosil that had worked so well on infected animals. He extrapolated<br />

from his animal experiments how much to give his<br />

daughter so that the bacteria would be killed but his daughter<br />

would not be poisoned. Within hours of the first treatment, her fever<br />

dropped, <strong>and</strong> she recovered completely after repeated doses of<br />

prontosil.<br />

Impact<br />

Directly <strong>and</strong> indirectly, Ehrlich’s <strong>and</strong> Domagk’s work served to<br />

usher in a new medical age. Prior to the discovery that prontosil<br />

could be use to treat bacterial infection <strong>and</strong> the subsequent development<br />

of a series of sulfonamides, or “sulfa drugs,” there was no<br />

chemical defense against this type of disease; as a result, illnesses<br />

such as streptococcal infection, gonorrhea, <strong>and</strong> pneumonia held terrors<br />

of which they have largely been shorn. Asmall injury could easily<br />

lead to death.<br />

By following the clues presented by the synthetic sulfa drugs <strong>and</strong><br />

how they worked to destroy bacteria, other scientists were able to<br />

develop an even more powerful type of drug, the antibiotic. When<br />

the American bacteriologist Rene Dubos discovered that natural organisms<br />

could also be used to fight bacteria, interest was renewed in<br />

an earlier discovery by the Scottish bacteriologist Sir Alex<strong>and</strong>er: the<br />

development of penicillin.


Antibiotics such as penicillin <strong>and</strong> streptomycin have become<br />

some of the most important tools in fighting disease. Antibiotics<br />

have replaced sulfa drugs for most uses, in part because they cause<br />

fewer side effects, but sulfa drugs are still used for a h<strong>and</strong>ful of purposes.<br />

Together, sulfonamides <strong>and</strong> antibiotics have offered the possibility<br />

of a cure to millions of people who previously would have<br />

had little chance of survival.<br />

See also Penicillin; Polio vaccine (Sabin); Polio vaccine (Salk);<br />

Salvarsan; Tuberculosis vaccine; Typhus vaccine; Yellow fever vaccine.<br />

Further Reading<br />

Antibacterial drugs / 27<br />

Alstaedter, Rosemarie. From Germanin to Acylureidopenicillin: Research<br />

That Made History: Documentation of a Scientific Revolution:<br />

Dedicated to Gerhardt Domagk on the Eighty-fifth Anniversary of His<br />

Birth. Leverkausen, West Germany: Bayer AG, 1980.<br />

Baumler, Ernst. Paul Ehrlich: Scientist for Life. New York: Holmes <strong>and</strong><br />

Meier, 1984.<br />

Galdston, Iago. Behind the Sulfa Drugs, a Short History of Chemotherapy.<br />

New York: D. Appleton-Century, 1943.<br />

Physiology or Medicine, 1922-1941. River Edge, N.J.: World Scientific,<br />

1999.


28<br />

Apple II computer<br />

Apple II computer<br />

The invention: The first commercially available, preassembled<br />

personal computer, the Apple II helped move computers out of<br />

the workplace <strong>and</strong> into the home.<br />

The people behind the invention:<br />

Stephen Wozniak (1950- ), cofounder of Apple <strong>and</strong> designer<br />

of the Apple II computer<br />

Steven Jobs (1955- ), cofounder of Apple<br />

Regis McKenna (1939- ), owner of the Silicon Valley public<br />

relations <strong>and</strong> advertising company that h<strong>and</strong>led the Apple<br />

account<br />

Chris Espinosa (1961- ), the high school student who wrote<br />

the BASIC program shipped with the Apple II<br />

R<strong>and</strong>y Wigginton (1960- ), a high school student <strong>and</strong> Apple<br />

software programmer<br />

Inventing the Apple<br />

As late as the 1960’s, not many people in the computer industry<br />

believed that a small computer could be useful to the average person.<br />

It was through the effort of two friends from the Silicon Valley—the<br />

high-technology area between San Francisco <strong>and</strong> San Jose—<br />

that the personal computer revolution was started.<br />

Both Steven Jobs <strong>and</strong> Stephen Wozniak had attended Homestead<br />

High School in Los Altos, California, <strong>and</strong> both developed early interests<br />

in technology, especially computers. In 1971, Wozniak built<br />

his first computer from spare parts. Shortly after this, he was introduced<br />

to Jobs. Jobs had already developed an interest in electronics<br />

(he once telephoned William Hewlett, cofounder of Hewlett-<br />

Packard, to ask for parts), <strong>and</strong> he <strong>and</strong> Wozniak became friends.<br />

Their first business together was the construction <strong>and</strong> sale of “blue<br />

boxes,” illegal devices that allowed the user to make long-distance<br />

telephone calls for free.<br />

After attending college, the two took jobs within the electronics<br />

industry. Wozniak began working at Hewlett-Packard, where he


studied calculator design, <strong>and</strong> Jobs took a job at Atari, the video<br />

company. The friendship paid off again when Wozniak, at Jobs’s request,<br />

designed the game “Breakout” for Atari, <strong>and</strong> the pair was<br />

paid seven hundred dollars.<br />

In 1975, the Altair computer, a personal computer in kit form,<br />

was introduced by Micro Instrumentation <strong>and</strong> Telemetry Systems<br />

(MITS). Shortly thereafter, the first personal computer club, the<br />

Homebrew Computer Club, began meeting in Menlo Park, near<br />

Stanford University. Wozniak <strong>and</strong> Jobs began attending the meeting<br />

regularly. Wozniak eagerly examined the Altairs that others<br />

brought. He thought that the design could be improved. In only a<br />

few more weeks, he produced a circuit board <strong>and</strong> interfaces that<br />

connected it to a keyboard <strong>and</strong> a video monitor. He showed the machine<br />

at a Homebrew meeting <strong>and</strong> distributed photocopies of the<br />

design.<br />

In this new machine, which he named an “Apple,” Jobs saw a big<br />

opportunity. He talked Wozniak into forming a partnership to develop<br />

personal computers. Jobs sold his car, <strong>and</strong> Wozniak sold his<br />

two Hewlett-Packard calculators; with the money, they ordered<br />

printed circuit boards made. Their break came when Paul Terrell, a<br />

retailer, was so impressed that he ordered fifty fully assembled Apples.<br />

Within thirty days, the computers were completed, <strong>and</strong> they<br />

sold for a fairly high price: $666.66.<br />

During the summer of 1976, Wozniak kept improving the Apple.<br />

The new computer would come with a keyboard, an internal power<br />

supply, a built-in computer language called the Beginner’s All-<br />

Purpose Symbolic Instruction Code” (BASIC), hookups for adding<br />

printers <strong>and</strong> other devices, <strong>and</strong> color graphics, all enclosed in a plastic<br />

case. The output would be seen on a television screen. The machine<br />

would sell for twelve hundred dollars.<br />

Selling the Apple<br />

Apple II computer / 29<br />

Regis McKenna was the head of the Regis McKenna <strong>Public</strong> Relations<br />

agency, the best of the public relations firms that served the<br />

high-technology industries of the valley, which Jobs wanted to h<strong>and</strong>le<br />

the Apple account. At first, McKenna rejected the offer, but<br />

Jobs’s constant pleading finally convinced him. The agency’s first


30 / Apple II computer<br />

Steven Jobs<br />

While IBM <strong>and</strong> other corporations were devoting massive<br />

resources <strong>and</strong> talent to designing a small computer in 1975,<br />

Steven Paul Jobs <strong>and</strong> Stephen Wozniak, members of the tiny<br />

Homebrew Computer Club, put together the first truly userfriendly<br />

personal computer in Wozniak’s home. Jobs admitted<br />

later that “Woz” was the engineering brains. Jobs himself was<br />

the brains of design <strong>and</strong> marketing. Both had to scrape together<br />

money for the project from their small salaries as low-level electronics<br />

workers. Within eight years, Jobs headed the most progressive<br />

company in the new personal computer industry <strong>and</strong><br />

was worth an estimated $210 million.<br />

Little in his background foretold such fast, large material<br />

success. Jobs was born in 1955 <strong>and</strong> became an orphan. Adopted<br />

by Paul <strong>and</strong> Clara Jobs, he grew up in California towns near the<br />

area that became known as Silicon Valley. He did not like school<br />

much <strong>and</strong> was considered a loner, albeit one who always had a<br />

distinctive way of thinking about things. Still in high school, he<br />

impressed William Hewlett, founder of Hewlett-Packard in<br />

Palo Alto, <strong>and</strong> won a summer job at the company, as well as<br />

some free equipment for one of his school projects.<br />

However, he dropped out of Reed College after one semester<br />

<strong>and</strong> became a hippie. He studied philosophy <strong>and</strong> Chinese<br />

<strong>and</strong> Indian mysticism. He became a vegetarian <strong>and</strong> practiced<br />

meditation. He even shaved his head <strong>and</strong> traveled to India on a<br />

spiritual pilgrimage. When he returned to America, however,<br />

he also returned to his interest in electronics <strong>and</strong> computers.<br />

Through various jobs at his original company, Apple, <strong>and</strong> elsewhere,<br />

he stayed there.<br />

contributions to Apple were the colorful striped Apple logo <strong>and</strong> a<br />

color ad in Playboy magazine.<br />

In February, 1977, the first Apple Computer office was opened in<br />

Cupertino, California. By this time, two of Wozniak’s friends from<br />

Homebrew, R<strong>and</strong>y Wigginton <strong>and</strong> Chris Espinosa—both high school<br />

students—had joined the company. Their specialty was writing software.<br />

Espinosa worked through his Christmas vacation so that BA-<br />

SIC (the built-in computer language) could ship with the computer.


The team pushed ahead to complete the new Apple in time to<br />

display it at the First West Coast Computer Faire in April, 1977. At<br />

this time, the name “Apple II” was chosen for the new model. The<br />

Apple II computer debuted at the convention <strong>and</strong> included many<br />

innovations. The “motherboard” was far simpler <strong>and</strong> more elegantly<br />

designed than that of any previous computer, <strong>and</strong> the ease of<br />

connecting the Apple II to a television screen made it that much<br />

more attractive to consumers.<br />

Consequences<br />

Apple II computer / 31<br />

The introduction of the Apple II computer launched what was to<br />

be a wave of new computers aimed at the home <strong>and</strong> small-business<br />

markets. Within a few months of the Apple II’s introduction, Commodore<br />

introduced its PET computer <strong>and</strong> T<strong>and</strong>y Corporation/Radio<br />

Shack brought out its TRS-80. Apple continued to increase the<br />

types of things that its computers could do <strong>and</strong> worked out a distribution<br />

deal with the new ComputerL<strong>and</strong> chain of stores.<br />

In December, 1977, Wozniak began work on creating a floppy<br />

disk system for the Apple II. (Afloppy disk is a small, flexible plastic<br />

disk coated with magnetic material. The magnetized surface enables<br />

computer data to be stored on the disk.) The cassette tape storage<br />

on which all personal computers then depended was slow <strong>and</strong><br />

unreliable. Floppy disks, which had been introduced for larger computers<br />

by the International Business Machines (IBM) Corporation in<br />

1970, were fast <strong>and</strong> reliable. As he did with everything that interested<br />

him, Wozniak spent almost all of his time learning about <strong>and</strong><br />

designing a floppy disk drive. When the final drive shipped in June,<br />

1978, it made possible development of more powerful software for<br />

the computer.<br />

By 1980, Apple had sold 130,000 Apple II’s. That year, the company<br />

went public, <strong>and</strong> Jobs <strong>and</strong> Wozniak, among others, became<br />

wealthy. Three years later, Apple became the youngest company to<br />

make the Fortune 500 list of the largest industrial companies. By<br />

then, IBM had entered the personal computer field <strong>and</strong> had begun<br />

to dominate it, but the Apple II’s earlier success ensured that personal<br />

computers would not be a market fad. By the end of the<br />

1980’s, 35 million personal computers would be in use.


32 / Apple II computer<br />

See also BINAC computer; Colossus computer; ENIAC computer;<br />

Floppy disk; Hard disk; IBM Model 1401 computer; Personal<br />

computer; UNIVAC computer.<br />

Further Reading<br />

Carlton, Jim. Apple: The Inside Story of Intrigue, Egomania, <strong>and</strong> Business<br />

Blunders. Rev. ed. London: R<strong>and</strong>om House, 1999.<br />

Gold, Rebecca. Steve Wozniak: A Wizard Called Woz. Minneapolis:<br />

Lerner, 1994.<br />

Linzmayer, Owen W. Apple Confidential: The Real Story of Apple Computer,<br />

Inc. San Francisco: No Starch Press, 1999.<br />

Moritz, Michael. The Little Kingdom: The Private Story of Apple Computer.<br />

New York: Morrow, 1984.<br />

Rose, Frank. West of Eden: The End of Innocence at Apple Computer.<br />

New York: Viking, 1989.


Aqualung<br />

Aqualung<br />

The invention: A device that allows divers to descend hundreds of<br />

meters below the surface of the ocean by enabling them to carry<br />

the oxygen they breathe with them.<br />

The people behind the invention:<br />

Jacques-Yves Cousteau (1910-1997), a French navy officer,<br />

undersea explorer, inventor, <strong>and</strong> author<br />

Émile Gagnan, a French engineer who invented an automatic<br />

air-regulating device<br />

The Limitations of Early Diving<br />

Undersea dives have been made since ancient times for the purposes<br />

of spying, recovering lost treasures from wrecks, <strong>and</strong> obtaining<br />

natural treasures (such as pearls). Many attempts have been made<br />

since then to prolong the amount of time divers could remain underwater.<br />

The first device, described by the Greek philosopher Aristotle<br />

in 335 b.c.e., was probably the ancestor of the modern snorkel. It was<br />

a bent reed placed in the mouth, with one end above the water.<br />

In addition to depth limitations set by the length of the reed,<br />

pressure considerations also presented a problem. The pressure on<br />

a diver’s body increases by about one-half pound per square centimeter<br />

for every meter ventured below the surface. After descending<br />

about 0.9 meter, inhaling surface air through a snorkel becomes difficult<br />

because the human chest muscles are no longer strong enough<br />

to inflate the chest. In order to breathe at or below this depth, a diver<br />

must breathe air that has been pressurized; moreover, that pressure<br />

must be able to vary as the diver descends or ascends.<br />

Few changes were possible in the technology of diving until air<br />

compressors were invented during the early nineteenth century.<br />

Fresh, pressurized air could then be supplied to divers. At first, the<br />

divers who used this method had to wear diving suits, complete<br />

with fishbowl-like helmets. This “tethered” diving made divers relatively<br />

immobile but allowed them to search for sunken treasure or<br />

do other complex jobs at great depths.<br />

33


34 / Aqualung<br />

The Development of Scuba Diving<br />

The invention of scuba gear gave divers more freedom to<br />

move about <strong>and</strong> made them less dependent on heavy equipment.<br />

(“Scuba” st<strong>and</strong>s for self-contained underwater breathing apparatus.)<br />

Its development occurred in several stages. In 1880, Henry<br />

Fleuss of Engl<strong>and</strong> developed an outfit that used a belt containing<br />

pure oxygen. Belt <strong>and</strong> diver were connected, <strong>and</strong> the diver breathed<br />

the oxygen over <strong>and</strong> over. A version of this system was used by the<br />

U.S. Navy in World War II spying efforts. Nevertheless, it had serious<br />

drawbacks: Pure oxygen was toxic to divers at depths greater<br />

than 9 meters, <strong>and</strong> divers could carry only enough oxygen for relatively<br />

short dives. It did have an advantage for spies, namely, that<br />

the oxygen—breathed over <strong>and</strong> over in a closed system—did not<br />

reach the surface in the form of telltale bubbles.<br />

The next stage of scuba development occurred with the design<br />

of metal tanks that were able to hold highly compressed air.<br />

This enabled divers to use air rather than the potentially toxic<br />

pure oxygen. More important, being hooked up to a greater supply<br />

of air meant that divers could stay under water longer. Initially,<br />

the main problem with the system was that the air flowed continuously<br />

through a mask that covered the diver’s entire face. This process<br />

wasted air, <strong>and</strong> the scuba divers expelled a continual stream<br />

of air bubbles that made spying difficult. The solution, according to<br />

Axel Madsen’s Cousteau (1986), was “a valve that would allow inhaling<br />

<strong>and</strong> exhaling through the same mouthpiece.”<br />

Jacques-Yves Cousteau’s father was an executive for Air Liquide—<br />

France’s main producer of industrial gases. He was able to direct<br />

Cousteau to Émile Gagnan, an engineer at thecompany’s Paris laboratory<br />

who had been developing an automatic gas shutoff valve for Air<br />

Liquide. This valve became the Cousteau-Gagnan regulator, a breathing<br />

device that fed air to the diver at just the right pressure whenever<br />

he or she inhaled.<br />

With this valve—<strong>and</strong> funding from Air Liquide—Cousteau <strong>and</strong><br />

Gagnan set out to design what would become the Aqualung. The<br />

first Aqualungs could be used at depths of up to 68.5 meters. During<br />

testing, however, the dangers of Aqualung diving became apparent.<br />

For example, unless divers ascended <strong>and</strong> descended in slow stages,


Jacques-Yves Cousteau<br />

Aqualung / 35<br />

The son of a businessman who liked to travel, Jacques-Yves<br />

Cousteau acquired the same w<strong>and</strong>erlust. Born in 1910 in Saint-<br />

André-de-Cubzac, France, he was a sickly child, but he learned<br />

to love swimming <strong>and</strong> the ocean. He also took an interest in<br />

movies, producing his first film when he was thirteen.<br />

Cousteau graduated from France’s naval academy, but his<br />

career as an officer ended with a nearly fatal car accident in<br />

1936. He went to Toulon, where he returned to his interests in<br />

the sea <strong>and</strong> photography, a period that culminated in his invention<br />

of the aqualung with Émile Gagnan in 1944. During<br />

World War II he also won a Légion d’honneur for<br />

his photographic espionage. The French Navy established<br />

the Underwater Research Group for Cousteau<br />

in 1944, <strong>and</strong> after the war the venture evolved into the<br />

freewheeling, worldwide voyages that Cousteau became<br />

famous for. Aboard the Calypso, a converted<br />

U.S. minesweeper, he <strong>and</strong> his crew conducted research<br />

<strong>and</strong> pioneered underwater photography. His<br />

1957 documentary The Silent World (based on a 1953<br />

book) won an Oscar <strong>and</strong> the Palm d’Or of the Cannes film<br />

festival. Subsequent movies <strong>and</strong> The Undersea World of Jacques<br />

Cousteau, a television series, established Cousteau as a leading<br />

environmentalist <strong>and</strong> science educator. His Cousteau Society,<br />

dedicated to exploring <strong>and</strong> protecting the oceans, attracted<br />

millions of members worldwide. Through it he launched another<br />

innovative technology, “Turbosails,” towering non-rotating<br />

cylinders that act as sails to reduce ships’ dependency on oilfueled<br />

engines. A new ship propelled by them, the Alcyone, eventually<br />

replaced the Calypso.<br />

Cousteau inspired legions of oceanographers <strong>and</strong> environmentalists<br />

while calling attention to pressing problems in the<br />

world’s oceans. Although his later years where marked by family<br />

tragedies <strong>and</strong> controversy, he was revered throughout the<br />

world <strong>and</strong> had received many honors when he died in 1997.<br />

it was likely that they would get “the bends” (decompression sickness),<br />

the feared disease of earlier, tethered deep-sea divers. Another<br />

problem was that, below 42.6 meters, divers encountered nitrogen<br />

narcosis. (This can lead to impaired judgment that may cause<br />

(Library of Congress)


36 / Aqualung<br />

fatal actions, including removing a mouthpiece or developing an<br />

overpowering desire to continue diving downward, to dangerous<br />

depths.)<br />

Cousteau believed that the Aqualung had tremendous military<br />

potential. During World War II, he traveled to London soon after the<br />

Norm<strong>and</strong>y invasion, hoping to persuade the Allied Powers of its<br />

usefulness. He was not successful. So Cousteau returned to Paris<br />

<strong>and</strong> convinced France’s new government to use Aqualungs to locate<br />

<strong>and</strong> neutralize underwater mines laid along the French coast by<br />

the German navy. Cousteau was commissioned to combine minesweeping<br />

with the study of the physiology of scuba diving. Further<br />

research revealed that the use of helium-oxygen mixtures increased<br />

to 76 meters the depth to which a scuba diver could go without suffering<br />

nitrogen narcosis.<br />

Impact<br />

One way to describe the effects of the development of the Aqualung<br />

is to summarize Cousteau’s continued efforts to the present. In<br />

1946, he <strong>and</strong> Philippe Tailliez established the Undersea Research<br />

Group of Toulon to study diving techniques <strong>and</strong> various aspects of<br />

life in the oceans. They studied marine life in the Red Sea from 1951<br />

to 1952. From 1952 to 1956, they engaged in an expedition supported<br />

by the National Geographic Society. By that time, the Research<br />

Group had developed many techniques that enabled them to<br />

identify life-forms <strong>and</strong> conditions at great depths.<br />

Throughout their undersea studies, Cousteau <strong>and</strong> his coworkers<br />

continued to develop better techniques for scuba diving, for recording<br />

observations by means of still <strong>and</strong> television photography, <strong>and</strong><br />

for collecting plant <strong>and</strong> animal specimens. In addition, Cousteau<br />

participated (with Swiss physicist Auguste Piccard) in the construction<br />

of the deep-submergence research vehicle, or bathyscaphe. In<br />

the 1960’s, he directed a program called Conshelf, which tested a<br />

human’s ability to live in a specially built underwater habitat. He<br />

also wrote <strong>and</strong> produced films on underwater exploration that attracted,<br />

entertained, <strong>and</strong> educated millions of people.<br />

Cousteau has won numerous medals <strong>and</strong> scientific distinctions.<br />

These include the Gold Medal of the National Geographic Society


(1963), the United Nations International Environment Prize (1977),<br />

membership in the American <strong>and</strong> Indian academies of science (1968<br />

<strong>and</strong> 1978, respectively), <strong>and</strong> honorary doctor of science degrees<br />

from the University of California, Berkeley (1970), Harvard University<br />

(1979), <strong>and</strong> Rensselaer Polytechnical Institute (1979).<br />

See also Bathyscaphe; Bathysphere.<br />

Further Reading<br />

Aqualung / 37<br />

Cousteau, Jacques Yves. The Silent World. New York: Harper &<br />

Brothers, 1952.<br />

_____. “Lord of The Depths. Time 153, no. 12 (March 29, 1999).<br />

_____, <strong>and</strong> James Dugan. The Living Sea. London: Elm Tree, 1988.<br />

Madsen, Axel. Cousteau: An Unauthorized Biography. New York:<br />

Beaufort Books, 1986.<br />

Munson, Richard. Cousteau: The Captain <strong>and</strong> His World. New York:<br />

Paragon House, 1991.<br />

Zanelli, Leo, <strong>and</strong> George T. Skuse. Sub-Aqua Illustrated Dictionary.<br />

New York: Oxford University Press, 1976.


38<br />

Artificial blood<br />

Artificial blood<br />

The invention: A perfluorocarbon emulsion that serves as a blood<br />

plasma substitute in the treatment of human patients.<br />

The person behind the invention:<br />

Ryoichi Naito (1906-1982), a Japanese physician<br />

Blood Substitutes<br />

The use of blood <strong>and</strong> blood products in humans is a very complicated<br />

issue. Substances present in blood serve no specific purpose<br />

<strong>and</strong> can be dangerous or deadly, especially when blood or blood<br />

products are taken from one person <strong>and</strong> given to another. This fact,<br />

combined with the necessity for long-term blood storage, a shortage<br />

of donors, <strong>and</strong> some patients’ refusal to use blood for religious reasons,<br />

brought about an intense search for a universal bloodlike substance.<br />

The life-sustaining properties of blood (for example, oxygen transport)<br />

can be entirely replaced by a synthetic mixture of known chemicals.<br />

Fluorocarbons are compounds that consist of molecules containing<br />

only fluorine <strong>and</strong> carbon atoms. These compounds are interesting<br />

to physiologists because they are chemically <strong>and</strong> pharmacologically<br />

inert <strong>and</strong> because they dissolve oxygen <strong>and</strong> other gases.<br />

Studies of fluorocarbons as blood substitutes began in 1966,<br />

when it was shown that a mouse breathing a fluorocarbon liquid<br />

treated with oxygen could survive. Subsequent research involved<br />

the use of fluorocarbons to play the role of red blood cells in transporting<br />

oxygen. Encouraging results led to the total replacement of<br />

blood in a rat, <strong>and</strong> the success of this experiment led in turn to trials<br />

in other mammals, culminating in 1979 with the use of fluorocarbons<br />

in humans.<br />

Clinical Studies<br />

The chemical selected for the clinical studies was Fluosol-DA,<br />

produced by the Japanese Green Cross Corporation. Fluosol-DA


consists of a 20 percent emulsion of two perfluorocarbons (perfluorodecalin<br />

<strong>and</strong> perfluorotripopylamine), emulsifiers, <strong>and</strong> salts<br />

that are included to give the chemical some of the properties of<br />

blood plasma. Fluosol-DA had been tested in monkeys, <strong>and</strong> it had<br />

shown a rapid reversible uptake <strong>and</strong> release of oxygen, a reasonably<br />

rapid excretion, no carcinogenicity or irreversible changes in the animals’<br />

systems, <strong>and</strong> the recovery of blood components to normal<br />

ranges within three weeks of administration.<br />

The clinical studies were divided into three phases. The first<br />

phase consisted of the administration of Fluosol-DA to normal human<br />

volunteers. Twelve healthy volunteers were administered the<br />

chemical, <strong>and</strong> the emulsion’s effects on blood pressure <strong>and</strong> composition<br />

<strong>and</strong> on heart, liver, <strong>and</strong> kidney functions were monitored. No<br />

adverse effects were found in any case. The first phase ended in<br />

March, 1979, <strong>and</strong> based on its positive results, the second <strong>and</strong> third<br />

phases were begun in April, 1979.<br />

Twenty-four Japanese medical institutions were involved in the<br />

next two phases. The reasons for the use of Fluosol-DA instead of<br />

blood in the patients involved were various, <strong>and</strong> they included refusal<br />

of transfusion for religious reasons, lack of compatible blood,<br />

“bloodless” surgery for protection from risk of hepatitis, <strong>and</strong> treatment<br />

of carbon monoxide intoxication.<br />

Among the effects noticed by the patients were the following: a<br />

small increase in blood pressure, with no corresponding effects on<br />

respiration <strong>and</strong> body temperature; an increase in blood oxygen content;<br />

bodily elimination of half the chemical within six to nineteen<br />

hours, depending on the initial dose administered; no change in<br />

red-cell count or hemoglobin content of blood; no change in wholeblood<br />

coagulation time; <strong>and</strong> no significant blood-chemistry changes.<br />

These results made the clinical trials a success <strong>and</strong> opened the door<br />

for other, more extensive ones.<br />

Impact<br />

Artificial blood / 39<br />

Perfluorocarbon emulsions were initially proposed as oxygencarrying<br />

resuscitation fluids, or blood substitutes, <strong>and</strong> the results of<br />

the pioneering studies show their success as such. Their success in<br />

this area, however, led to advanced studies <strong>and</strong> exp<strong>and</strong>ed use of


40 / Artificial blood<br />

these compounds in many areas of clinical medicine <strong>and</strong> biomedical<br />

research.<br />

Perfluorocarbon emulsions are useful in cancer therapy, because<br />

they increase the oxygenation of tumor cells <strong>and</strong> therefore sensitize<br />

them to the effects of radiation or chemotherapy. Perfluorocarbons<br />

can also be used as “contrasting agents” to facilitate magnetic resonance<br />

imaging studies of various tissues; for example, the uptake of<br />

particles of the emulsion by the cells of malignant tissues makes it<br />

possible to locate tumors. Perfluorocarbons also have a high nitrogen<br />

solubility <strong>and</strong> therefore can be used to alleviate the potentially<br />

fatal effects of decompression sickness by “mopping up” nitrogen<br />

gas bubbles from the circulation system. They can also be used to<br />

preserve isolated organs <strong>and</strong> amputated extremities until they can<br />

be reimplanted or reattached. In addition, the emulsions are used in<br />

cell cultures to regulate gas supply <strong>and</strong> to improve cell growth <strong>and</strong><br />

productivity.<br />

The biomedical applications of perfluorocarbon emulsions are<br />

multidisciplinary, involving areas as diverse as tissue imaging, organ<br />

preservation, cancer therapy, <strong>and</strong> cell culture. The successful<br />

clinical trials opened the door for new applications of these<br />

compounds, which rank among the most versatile compounds exploited<br />

by humankind.<br />

See also Artificial heart; Artificial hormone; Artificial kidney;<br />

Blood transfusion; Coronary artery bypass surgery; Electrocardiogram;<br />

Heart-lung machine.<br />

Further Reading<br />

“Artificial Blood Product May Debut in Two Years.” Health Care<br />

Strategic Management 18, no. 8 (August, 2000).<br />

“The Business of Blood: Ryoichi Naito <strong>and</strong> Fluosol-DA Artificial<br />

Blood.” Forbes 131 (January 17, 1983).<br />

Glanz, James. “Pulse Quickens in Search for Blood Substitute.” Research<br />

& Development 34, no. 10 (September, 1992).<br />

Tsuchida, E. Artificial Red Cells: Materials, Performances, <strong>and</strong> Clinical<br />

Study as Blood Substitutes. New York: Wiley, 1997.


Artificial chromosome<br />

Artificial chromosome<br />

The invention: Originally developed for use in the study of natural<br />

chromosome behavior, the artificial chromosome proved to be a<br />

valuable tool for recombinant DNA technology.<br />

The people behind the invention:<br />

Jack W. Szostak (1952- ), a British-born Canadian professor<br />

at Harvard Medical School<br />

Andrew W. Murray (1956- ), a graduate student<br />

The Value of Artificial Chromosomes<br />

The artificial chromosome gives biologists insight into the fundamental<br />

mechanisms by which cells replicate <strong>and</strong> plays an important<br />

role as a tool in genetic engineering technology. Soon after its invention<br />

in 1983 by Andrew W. Murray <strong>and</strong> Jack W. Szostak, the artificial<br />

chromosome was judged by scientists to be important <strong>and</strong> its value<br />

in the field of medicine was exploited.<br />

Chromosomes are essentially carriers of genetic information;<br />

that is, they possess the genetic code that is the blueprint for life. In<br />

higher organisms, the number <strong>and</strong> type of chromosomes that a cell<br />

contains in its nucleus are characteristic of the species. For example,<br />

each human cell has forty-six chromosomes, while the garden pea<br />

has fourteen <strong>and</strong> the guinea pig has sixty-four. The chromosome’s<br />

job in a dividing cell is to replicate <strong>and</strong> then distribute one copy of itself<br />

into each new “daughter” cell. This process, which is referred to<br />

as “mitosis” or “meiosis,” depending upon the actual mechanism<br />

by which the process occurs, is of supreme importance to the continuation<br />

of life.<br />

In 1953, when biophysicists James D. Watson <strong>and</strong> Francis Crick<br />

discovered the structure of deoxyribonucleic acid (DNA), an achievement<br />

for which they won the 1962 Nobel Prize in Physiology or<br />

Medicine, it was immediately apparent to them how the doublehelical<br />

form of DNA (which looks something like a twisted ladder)<br />

might explain the mechanism behind cell division. During DNA<br />

replication, the chromosome unwinds to expose the thin threads of<br />

41


42 / Artificial chromosome<br />

DNA. The two str<strong>and</strong>s of the double helix separate, <strong>and</strong> each acts as<br />

a template for the formation of a new complementary str<strong>and</strong>, thus<br />

forming two complete <strong>and</strong> identical chromosomes that can be distributed<br />

to each new cell. This distribution process, which is referred<br />

to as “segregation,” relies on the chromosomes being pulled along a<br />

microtubule framework in the cell called the “mitotic spindle.”<br />

Creating Artificial Chromosomes<br />

An artificial chromosome is a laboratory-designed chromosome<br />

that possesses only those functional elements its creators choose. In<br />

order to be a true working chromosome, however, it must, at minimum,<br />

maintain the machinery necessary for replication <strong>and</strong> segregation.<br />

By the early 1980’s, Murray <strong>and</strong> Szostak had recognized the possible<br />

advantages of using a simple, controlled model to study chromosome<br />

behavior, since there are several difficulties associated<br />

with studying chromosomes in their natural state. Since natural<br />

chromosomes are large <strong>and</strong> have poorly defined structures, it is almost<br />

impossible to sift out for study those elements that are essential<br />

for replication <strong>and</strong> segregation. Previous methods of altering a<br />

natural chromosome <strong>and</strong> observing the effects were difficult to use<br />

because the cells containing that altered chromosome usually died.<br />

Furthermore, even if the cell survived, analysis was complicated by<br />

the extensive amount of genetic information carried by the chromosome.<br />

Artificial chromosomes are simple <strong>and</strong> have known components,<br />

although the functions of those components may be poorly<br />

understood. In addition, since artificial chromosomes are extra chromosomes<br />

that are carried by the cell, their alteration does not kill the<br />

cell.<br />

Prior to the synthesis of the first artificial chromosome, the essential<br />

functional chromosomal elements of replication <strong>and</strong> segregation<br />

had to be identified <strong>and</strong> harvested. One of the three chromosomal<br />

elements thought to be required is the origin of replication,<br />

the site at which the synthesis of new DNA begins. The relatively<br />

weak interaction between DNA str<strong>and</strong>s at this site facilitates their<br />

separation, making possible—with the help of appropriate enzymes—<br />

the subsequent replication of the str<strong>and</strong>s into “sister chromatids.”


The second essential element is the “centromere,” a thinner segment<br />

of the chromosome that serves as the attachment site for the mitotic<br />

spindle. Sister chromatids are pulled into diametric ends of the dividing<br />

cell by the spindle apparatus, thus forming two identical<br />

daughter cells. The final functional elements are repetitive sequences<br />

of DNA called “telomeres,” which are located at both ends of the<br />

chromosome. The telomeres are needed to protect the terminal<br />

genes from degradation.<br />

With all the functional elements at their disposal, Murray <strong>and</strong><br />

Szostak proceeded to construct their first artificial chromosome.<br />

Once made, this chromosome would be inserted into yeast cells to<br />

replicate, since yeast cells are relatively simple <strong>and</strong> well characterized<br />

but otherwise resemble cells of higher organisms. Construction<br />

begins with a commonly used “bacterial plasmid,” a small, circular,<br />

autonomously replicating section of DNA. Enzymes are then called<br />

upon to create a gap in this “cloning vector” into which the three<br />

chromosomal elements are spliced. In addition, genes that confer<br />

some distinct trait, such as color, to yeast cells are also inserted, thus<br />

making it possible to determine which cells have actually taken up<br />

the new chromosome. Although their first attempt resulted in a<br />

chromosome that failed to segregate properly, by September, 1983,<br />

Murray <strong>and</strong> Szostak had announced in the prestigious British journal<br />

Nature their success in creating the first artificial chromosome.<br />

Consequences<br />

Artificial chromosome / 43<br />

One of the most exciting aspects of the artificial chromosome is<br />

its application to recombinant DNA technology, which involves creating<br />

novel genetic materials by combining segments of DNA from<br />

various sources. For example, the artificial yeast chromosome can<br />

be used as a cloning vector. In this process, a segment of DNA containing<br />

some desired gene is inserted into an artificial chromosome<br />

<strong>and</strong> is then allowed to replicate in yeast until large amounts of the<br />

gene are produced. David T. Burke, Georges F. Carle, <strong>and</strong> Maynard<br />

Victor Olson at Washington University in St. Louis have pioneered<br />

the technique of combining human genes with artificial yeast chromosomes<br />

<strong>and</strong> have succeeded in cloning large segments of human<br />

DNA.


44 / Artificial chromosome<br />

Although amplifying DNA in this manner has been done before,<br />

using bacterial plasmids as cloning vectors, the artificial yeast chromosome<br />

has the advantage of being able to hold much larger segments<br />

of DNA, thus allowing scientists to clone very large genes.<br />

This is of great importance, since the genes that cause diseases such<br />

as hemophilia <strong>and</strong> Duchenne’s muscular dystrophy are enormous.<br />

The most ambitious project for which the artificial yeast chromosome<br />

is being used is the national project whose intent is to clone the<br />

entire human genome.<br />

See also Artificial blood; Artificial hormone; Genetic “fingerprinting”;<br />

Genetically engineered insulin; In vitro plant culture;<br />

Synthetic DNA; Synthetic RNA.<br />

Further Reading<br />

“Evolving RNA with Enzyme-Like Action.” Science News 144 (August<br />

14, 1993).<br />

Freedman, David H. “Playing God: The H<strong>and</strong>made Cell.” Discover<br />

13, no. 8 (August, 1992).<br />

Varshavsky, Alex<strong>and</strong>er. “The 2000 Genetics Society of America<br />

Medal: Jack W. Szostak.” Genetics 157, no. 2 (February, 2001).


Artificial heart<br />

Artificial heart<br />

The invention: The first successful artificial heart, the Jarvik-7, has<br />

helped to keep patients suffering from otherwise terminal heart<br />

disease alive while they await human heart transplants.<br />

The people behind the invention:<br />

Robert Jarvik (1946- ), the main inventor of the Jarvik-7<br />

William Castle DeVries (1943- ), a surgeon at the University<br />

of Utah in Salt Lake City<br />

Barney Clark (1921-1983), a Seattle dentist, the first recipient of<br />

the Jarvik-7<br />

Early Success<br />

The Jarvik-7 artificial heart was designed <strong>and</strong> produced by researchers<br />

at the University of Utah in Salt Lake City; it is named for<br />

the leader of the research team, Robert Jarvik. An air-driven pump<br />

made of plastic <strong>and</strong> titanium, it is the size of a human heart. It is made<br />

up of two hollow chambers of polyurethane <strong>and</strong> aluminum, each<br />

containing a flexible plastic membrane. The heart is implanted in a<br />

human being but must remain connected to an external air pump by<br />

means of two plastic hoses. The hoses carry compressed air to the<br />

heart, which then pumps the oxygenated blood through the pulmonary<br />

artery to the lungs <strong>and</strong> through the aorta to the rest of the body.<br />

The device is expensive, <strong>and</strong> initially the large, clumsy air compressor<br />

had to be wheeled from room to room along with the patient.<br />

The device was new in 1982, <strong>and</strong> that same year Barney Clark, a<br />

dentist from Seattle, was diagnosed as having only hours to live.<br />

His doctor, cardiac specialist William Castle DeVries, proposed surgically<br />

implanting the Jarvik-7 heart, <strong>and</strong> Clark <strong>and</strong> his wife agreed.<br />

The Food <strong>and</strong> Drug Administration (FDA), which regulates the use<br />

of medical devices, had already given DeVries <strong>and</strong> his coworkers<br />

permission to implant up to seven Jarvik-7 hearts for permanent use.<br />

The operation was performed on Clark, <strong>and</strong> at first it seemed quite<br />

successful. Newspapers, radio, <strong>and</strong> television reported this medical<br />

breakthrough: the first time a severely damaged heart had been re-<br />

45


46 / Artificial heart<br />

William C. DeVries<br />

William Castle DeVries did not invent the artificial heart<br />

himself; however, he did develop the procedure to implant it.<br />

The first attempt took him seven <strong>and</strong> a half hours, <strong>and</strong> he<br />

needed fourteen assistants. A success, the surgery made DeVries<br />

one of the most talked-about doctors in the world.<br />

DeVries was born in Brooklyn, New York, in 1943. His father,<br />

a Navy physician, was killed in action a few months later, <strong>and</strong><br />

his mother, a nurse, moved with her son to Utah. As a child<br />

DeVries showed both considerable mechanical aptitude <strong>and</strong><br />

athletic prowess. He won an athletic scholarship to the University<br />

of Utah, graduating with honors in 1966. He entered the<br />

state medical school <strong>and</strong> there met Willem Kolff, a pioneer in<br />

designing <strong>and</strong> testing artificial organs. Under Kolff’s guidance,<br />

DeVries began performing experimental surgeries on animals<br />

to test prototype mechanical hearts. He finished medical school<br />

in 1970 <strong>and</strong> from 1971 until 1979 was an intern <strong>and</strong> then a resident<br />

in surgery at the Duke University Medical Center in North<br />

Carolina.<br />

DeVries returned to the University of Utah as an assistant<br />

professor of cardiovascular <strong>and</strong> thoracic surgery. In the meantime,<br />

Robert K. Jarvik had devised the Jarvik-7 artificial heart.<br />

DeVries experimented, implanting it in animals <strong>and</strong> cadavers<br />

until, following approval from the Federal Drug Administration,<br />

Barney Clark agreed to be the first test patient. He died 115<br />

days after the surgery, having never left the hospital. Although<br />

controversy arose over the ethics <strong>and</strong> cost of the procedure,<br />

more artificial heart implantations followed, many by DeVries.<br />

Long administrative delays getting patients approved for<br />

surgery at Utah frustrated DeVries, so he moved to Humana<br />

Hospital-Audubon in Louisville, Kentucky, in 1984 <strong>and</strong> then<br />

took a professorship at the University of Louisville. In 1988 he<br />

left experimentation for a traditional clinical practice. The FDA<br />

withdrew its approval for the Jarvik-7 in 1990.<br />

In 1999 DeVries retired from practice, but not from medicine.<br />

The next year he joined the Army Reserve <strong>and</strong> began teaching<br />

surgery at the Walter Reed Army Medical Center.<br />

placed by a totally artificial heart. It seemed DeVries had proved that<br />

an artificial heart could be almost as good as a human heart.<br />

Soon after Clark’s surgery, DeVries went on to implant the device


in several other patients with serious heart disease. For a time, all of<br />

them survived the surgery. As a result, DeVries was offered a position<br />

at Humana Hospital in Louisville, Kentucky. Humana offered<br />

to pay for the first one hundred implant operations.<br />

The Controversy Begins<br />

Artificial heart / 47<br />

In the three years after DeVries’s operation on Barney Clark,<br />

however, doubts <strong>and</strong> criticism arose. Of the people who by then had<br />

received the plastic <strong>and</strong> metal device as a permanent replacement<br />

for their own diseased hearts, three had died (including Clark) <strong>and</strong><br />

four had suffered serious strokes. The FDA asked Humana Hospital<br />

<strong>and</strong> Symbion (the company that manufactured the Jarvik-7) for<br />

complete, detailed histories of the artificial-heart recipients.<br />

It was determined that each of the patients who had died or been<br />

disabled had suffered from infection. Life-threatening infection, or<br />

“foreign-body response,” is a danger with the use of any artificial<br />

organ. The Jarvik-7, with its metal valves, plastic body, <strong>and</strong> Velcro<br />

attachments, seemed to draw bacteria like a magnet—<strong>and</strong> these<br />

bacteria proved resistant to even the most powerful antibiotics.<br />

By 1988, researchers had come to realize that severe infection was<br />

almost inevitable if a patient used the Jarvik-7 for a long period of<br />

time. As a result, experts recommended that the device be used for<br />

no longer than thirty days.<br />

Questions of values <strong>and</strong> morality also became part of the controversy<br />

surrounding the artificial heart. Some people thought that it<br />

was wrong to offer patients a device that would extend their lives<br />

but leave them burdened with hardship <strong>and</strong> pain. At times DeVries<br />

claimed that it was worth the price for patients to be able live another<br />

year; at other times, he admitted that if he thought a patient<br />

would have to spend the rest of his or her life in a hospital, he would<br />

think twice before performing the implant.<br />

There were also questions about “informed consent”—the patient’s<br />

underst<strong>and</strong>ing that a medical procedure has a high risk of<br />

failure <strong>and</strong> may leave the patient in misery even if it succeeds.<br />

Getting truly informed consent from a dying patient is tricky, because,<br />

underst<strong>and</strong>ably, the patient is probably willing to try anything.<br />

The Jarvik-7 raised several questions in this regard: Was the


48 / Artificial heart<br />

ordeal worth the risk? Was the patient’s suffering justifiable? Who<br />

should make the decision for or against the surgery: the patient, the<br />

researchers, or a government agency?<br />

Also there was the issue of cost. Should money be poured into expensive,<br />

high-technology devices such as the Jarvik heart, or should<br />

it be reserved for programs to help prevent heart disease in the first<br />

place? Expenses for each of DeVries’s patients had amounted to<br />

about one million dollars.<br />

Humana’s <strong>and</strong> DeVries’s earnings were criticized in particular.<br />

Once the first one hundred free Jarvik-7 implantations had been<br />

performed, Humana Hospital could expect to make large amounts<br />

of money on the surgery. By that time, Humana would have so<br />

much expertise in the field that, though the surgical techniques<br />

could not be patented, it was expected to have a practical monopoly.<br />

DeVries himself owned thous<strong>and</strong>s of shares of stock in Symbion.<br />

Many people wondered whether this was ethical.<br />

Consequences<br />

Given all the controversies, in December of 1985 a panel of experts<br />

recommended that the FDA allow the experiment to continue,<br />

but only with careful monitoring. Meanwhile, cardiac transplantation<br />

was becoming easier <strong>and</strong> more common. By the end of 1985, almost<br />

twenty-six hundred patients in various countries had received<br />

human heart transplants, <strong>and</strong> 76 percent of these patients had survived<br />

for at least four years. When the dem<strong>and</strong> for donor hearts exceeded<br />

the supply, physicians turned to the Jarvik device <strong>and</strong> other<br />

artificial hearts to help see patients through the waiting period.<br />

Experience with the Jarvik-7 made the world keenly aware of<br />

how far medical science still is from making the implantable permanent<br />

mechanical heart a reality. Nevertheless, the device was a<br />

breakthrough in the relatively new field of artificial organs. Since<br />

then, other artificial body parts have included heart valves, blood<br />

vessels, <strong>and</strong> inner ears that help restore hearing to the deaf.<br />

See also Artificial blood; Artificial kidney; Blood transfusion;<br />

Coronary artery bypass surgery; Electrocardiogram; Heart-lung<br />

machine; Pacemaker; Velcro.


Further Reading<br />

Artificial heart / 49<br />

Fox, Renee C., <strong>and</strong> Judith P. Swazy. Spare Parts: Organ Replacement in<br />

American Society. New York: Oxford University Press, 1992.<br />

Kunin, Calvin M., Joanne J. Debbins, <strong>and</strong> Julio C. Melo. “Infectious<br />

Complications in Four Long-Term Recipients of the Jarvik-7 Artificial<br />

Heart.” JAMA 259 (February 12, 1988).<br />

Kunzig, Robert. “The Beat Goes On.” Discover 21, no. 1 (January,<br />

2000).<br />

Lawrie, Gerald M. “Permanent Implantation of the Jarvik-7 Total<br />

Artificial Heart: A Clinical Perspective.” JAMA 259 (February 12,<br />

1988).


50<br />

Artificial hormone<br />

Artificial hormone<br />

The invention: Synthesized oxytocin, a small polypeptide hormone<br />

from the pituitary gl<strong>and</strong> that has shown how complex polypeptides<br />

<strong>and</strong> proteins may be synthesized <strong>and</strong> used in medicine.<br />

The people behind the invention:<br />

Vincent du Vigneaud (1901-1978), an American biochemist <strong>and</strong><br />

winner of the 1955 Nobel Prize in Chemistry<br />

Oliver Kamm (1888-1965), an American biochemist<br />

Sir Edward Albert Sharpey-Schafer (1850-1935), an English<br />

physiologist<br />

Sir Henry Hallett Dale (1875-1968), an English physiologist <strong>and</strong><br />

winner of the 1936 Nobel Prize in Physiology or Medicine<br />

John Jacob Abel (1857-1938), an American pharmacologist <strong>and</strong><br />

biochemist<br />

Body-Function Special Effects<br />

In Engl<strong>and</strong> in 1895, physician George Oliver <strong>and</strong> physiologist<br />

Edward Albert Sharpey-Schafer reported that a hormonal extract<br />

from the pituitary gl<strong>and</strong> of a cow produced a rise in blood pressure<br />

(a pressor effect) when it was injected into animals. In 1901, Rudolph<br />

Magnus <strong>and</strong> Sharpey-Schafer discovered that extracts from<br />

the pituitary also could restrict the flow of urine (an antidiuretic effect).<br />

This observation was related to the fact that when a certain<br />

section of the pituitary was removed surgically from an animal, the<br />

animal excreted an abnormally large amount of urine.<br />

In addition to the pressor <strong>and</strong> antidiuretic activities in the pituitary,<br />

two other effects were found in 1909. Sir Henry Hallett Dale,<br />

an English physiologist, was able to show that the extracts could<br />

cause the uterine muscle to contract (an oxytocic effect), <strong>and</strong> Isaac<br />

Ott <strong>and</strong> John C. Scott found that when lactating (milk-producing)<br />

animals were injected with the extracts, milk was released from the<br />

mammary gl<strong>and</strong>.<br />

Following the discovery of these various effects, attempts were<br />

made to concentrate <strong>and</strong> isolate the substance or substances that


were responsible. John Jacob Abel was able to concentrate the pressor<br />

activity at The Johns Hopkins University using heavy metal salts<br />

<strong>and</strong> extraction with organic solvents. The results of the early work,<br />

however, were varied. Some investigators came to the conclusion<br />

that only one substance was responsible for all the activities, while<br />

others concluded that two or more substances were likely to be involved.<br />

In 1928, Oliver Kamm <strong>and</strong> his coworkers at the drug firm of<br />

Parke, Davis <strong>and</strong> Company in Detroit reported a method for the<br />

separation of the four activities into two chemical fractions with<br />

high potency. One portion contained most of the pressor <strong>and</strong> antidiuretic<br />

activities, while the other contained the uterine-contracting<br />

<strong>and</strong> milk-releasing activities. Over the years, several names have<br />

been used for the two substances responsible for the effects. The generic<br />

name “vasopressin” generally has become the accepted term<br />

for the substance causing the pressor <strong>and</strong> antidiuretic effects, while<br />

the name “oxytocin” has been used for the other two effects. The<br />

two fractions that Kamm <strong>and</strong> his group had prepared were pure<br />

enough for the pharmaceutical firm to make them available for<br />

medical research related to obstetrics, surgical shock, <strong>and</strong> diabetes<br />

insipidus.<br />

A Complicated Synthesis<br />

Artificial hormone / 51<br />

The problem of these hormones <strong>and</strong> their nature interested Vincent<br />

du Vigneaud at the George Washington University School of<br />

Medicine. Working with Kamm, he was able to show that the sulfur<br />

content of both the oxytocin <strong>and</strong> the vasopressin fractions was a result<br />

of the amino acid cystine. This helped to strengthen the concept<br />

that these hormones were polypeptide, or proteinlike, substances.<br />

Du Vigneaud <strong>and</strong> his coworkers next tried to find a way of purifying<br />

oxytocin <strong>and</strong> vasopressin. This required not only the separation<br />

of the hormones themselves but also the separation from other impurities<br />

present in the preparations.<br />

During World War II (1939-1945) <strong>and</strong> shortly thereafter, other<br />

techniques were developed that would give du Vigneaud the tools<br />

he needed to complete the job of purifying <strong>and</strong> characterizing<br />

the two hormonal factors. One of the most important was the


52 / Artificial hormone<br />

countercurrent distribution method of chemist Lyman C. Craig at<br />

the Rockefeller Institute. Craig had developed an apparatus that<br />

could do multiple extractions, making possible separations of substances<br />

with similar properties. Du Vigneaud had used this technique<br />

in purifying his synthetic penicillin, <strong>and</strong> when he returned to<br />

the study of oxytocin <strong>and</strong> vasopressin in 1946, he used it on his purest<br />

preparations. The procedure worked well, <strong>and</strong> milligram quantities<br />

of pure oxytocin were available in 1949 for chemical characterization.<br />

Using the available techniques, Vigneaud <strong>and</strong> his coworkers<br />

were able to determine the structure of oxytocin. It was du Vigneaud’s<br />

goal to make synthetic oxytocin by duplicating the structure<br />

his group had worked out. Eventually, du Vigneaud’s synthetic<br />

oxytocin was obtained <strong>and</strong> the method published in the Journal of<br />

the American Chemical Society in 1953.<br />

Du Vigneaud’s oxytocin was next tested against naturally occurring<br />

oxytocin, <strong>and</strong> the two forms were found to act identically in every<br />

respect. In the final test, the synthetic form was found to induce<br />

labor when given intravenously to women about to give birth. Also,<br />

when microgram quantities of oxytocin were given intravenously<br />

to women who had recently given birth, milk was released from the<br />

mammary gl<strong>and</strong> in less than a minute.<br />

Consequences<br />

The work of du Vigneaud <strong>and</strong> his associates demonstrated for<br />

the first time that it was possible to synthesize peptides that have<br />

properties identical to the natural ones <strong>and</strong> that these can be useful<br />

in certain medical conditions. Oxytocin has been used in the last<br />

stages of labor during childbirth. Vasopressin has been used in the<br />

treatment of diabetes insipidus, when an individual has an insufficiency<br />

in the natural hormone, much as insulin is used by persons<br />

having diabetes mellitus.<br />

After receiving the Nobel Prize in Chemistry in 1955, du Vigneaud<br />

continued his work on synthesizing chemical variations of the two<br />

hormones. By making peptides that differed from oxytocin <strong>and</strong><br />

vasopressin by one or more amino acids, it was possible to study how<br />

the structure of the peptide was related to its physiological activity.


After the structure of insulin <strong>and</strong> some of the smaller proteins<br />

were determined, they, too, were synthesized, although with greater<br />

difficulty. Other methods of carrying out the synthesis of peptides<br />

<strong>and</strong> proteins have been developed <strong>and</strong> are used today. The production<br />

of biologically active proteins, such as insulin <strong>and</strong> growth hormone,<br />

has been made possible by efficient methods of biotechnology.<br />

The genes for these proteins can be put inside microorganisms,<br />

which then make them in addition to their own proteins. The microorganisms<br />

are then harvested <strong>and</strong> the useful protein hormones isolated<br />

<strong>and</strong> purified.<br />

See also Abortion pill; Artificial blood; Birth control pill; Genetically<br />

engineered insulin; Pap test.<br />

Further Reading<br />

Artificial hormone / 53<br />

Basa, Channa, <strong>and</strong> G. M. Anantharamaiah. Peptides: Design, Synthesis,<br />

<strong>and</strong> Biological Activity. Boston: Birkauser, 1994.<br />

Bodanszky, Miklos. “Vincent du Vigneaud, 1901-1978.” Nature 279,<br />

no. 5710 (1979).<br />

Vigneud, Vincent du. “A Trail of Sulfur Research from Insulin to<br />

Oxytocin” [Nobel lecture]. In Chemistry, 1942-1962. River Edge,<br />

N.J.: World Scientific, 1999.


54<br />

Artificial insemination<br />

Artificial insemination<br />

The invention: Practical techniques for the artificial insemination<br />

of farm animals that have revolutionized livestock breeding practices<br />

throughout the world.<br />

The people behind the invention:<br />

Lazzaro Spallanzani (1729-1799), an Italian physiologist<br />

Ilya Ivanovich Ivanov (1870-1932), a Soviet biologist<br />

R. W. Kunitsky, a Soviet veterinarian<br />

Reproduction Without Sex<br />

The tale is told of a fourteenth-century Arabian chieftain who<br />

sought to improve his mediocre breed of horses. Sneaking into the<br />

territory of a neighboring hostile tribe, he stimulated a prize stallion<br />

to ejaculate into a piece of cotton. Quickly returning home, he<br />

inserted this cotton into the vagina of his own mare, who subsequently<br />

gave birth to a high-quality horse. This may have been the<br />

first case of “artificial insemination,” the technique by which semen<br />

is introduced into the female reproductive tract without sexual<br />

contact.<br />

The first scientific record of artificial insemination comes from Italy<br />

in the 1770’s. Lazzaro Spallanzani was one of the foremost physiologists<br />

of his time, well known for having disproved the theory of<br />

spontaneous generation, which states that living organisms can<br />

spring “spontaneously” from lifeless matter. There was some disagreement<br />

at that time about the basic requirements for reproduction<br />

in animals. It was unclear if the sex act was necessary for an embryo<br />

to develop, or if it was sufficient that the sperm <strong>and</strong> eggs come<br />

into contact. Spallanzani began by studying animals in which union<br />

of the sperm <strong>and</strong> egg normally takes place outside the body of the<br />

female. He stimulated males <strong>and</strong> females to release their sperm <strong>and</strong><br />

eggs, then mixed these sex cells in a glass dish. In this way, he produced<br />

young frogs, toads, salam<strong>and</strong>ers, <strong>and</strong> silkworms.<br />

Next, Spallanzani asked whether the sex act was also unnecessary<br />

for reproduction in those species in which fertilization nor-


mally takes place inside the body of the female. He collected semen<br />

that had been ejaculated by a male spaniel <strong>and</strong>, using a syringe, injected<br />

the semen into the vagina of a female spaniel in heat. Two<br />

months later, she delivered a litter of three pups, which bore some<br />

resemblance to both the mother <strong>and</strong> the male that had provided the<br />

sperm.<br />

It was in animal breeding that Spallanzani’s techniques were to<br />

have their most dramatic application. In the 1880’s, an English dog<br />

breeder, Sir Everett Millais, conducted several experiments on artificial<br />

insemination. He was interested mainly in obtaining offspring<br />

from dogs that would not normally mate with one another because<br />

of difference in size. He followed Spallanzani’s methods to produce<br />

a cross between a short, low, basset hound <strong>and</strong> the much larger<br />

bloodhound.<br />

Long-Distance Reproduction<br />

Artificial insemination / 55<br />

Ilya Ivanovich Ivanov was a Soviet biologist who was commissioned<br />

by his government to investigate the use of artificial insemination<br />

on horses. Unlike previous workers who had used artificial<br />

insemination to get around certain anatomical barriers to fertilization,<br />

Ivanov began the use of artificial insemination to reproduce<br />

thoroughbred horses more effectively. His assistant in this work<br />

was the veterinarian R. W. Kunitsky.<br />

In 1901, Ivanov founded the Experimental Station for the Artificial<br />

Insemination of Horses. As its director, he embarked on a series<br />

of experiments to devise the most efficient techniques for breeding<br />

these animals. Not content with the demonstration that the technique<br />

was scientifically feasible, he wished to ensure further that it<br />

could be practiced by Soviet farmers.<br />

If sperm from a male were to be used to impregnate females in<br />

another location, potency would have to be maintained for a long<br />

time. Ivanov first showed that the secretions from the sex gl<strong>and</strong>s<br />

were not required for successful insemination; only the sperm itself<br />

was necessary. He demonstrated further that if a testicle were removed<br />

from a bull <strong>and</strong> kept cold, the sperm would remain alive.<br />

More useful than preservation of testicles would be preservation<br />

of the ejaculated sperm. By adding certain salts to the sperm-


56 / Artificial insemination<br />

containing fluids, <strong>and</strong> by keeping these at cold temperatures, Ivanov<br />

was able to preserve sperm for long periods.<br />

Ivanov also developed instruments to inject the sperm, to hold<br />

the vagina open during insemination, <strong>and</strong> to hold the horse in place<br />

during the procedure. In 1910, Ivanov wrote a practical textbook<br />

with technical instructions for the artificial insemination of horses.<br />

He also trained some three hundred veterinary technicians in the<br />

use of artificial insemination, <strong>and</strong> the knowledge he developed<br />

quickly spread throughout the Soviet Union. Artificial insemination<br />

became the major means of breeding horses.<br />

Until his death in 1932, Ivanov was active in researching many<br />

aspects of the reproductive biology of animals. He developed methods<br />

to treat reproductive diseases of farm animals <strong>and</strong> refined<br />

methods of obtaining, evaluating, diluting, preserving, <strong>and</strong> disinfecting<br />

sperm. He also began to produce hybrids between wild <strong>and</strong><br />

domestic animals in the hope of producing new breeds that would<br />

be able to withst<strong>and</strong> extreme weather conditions better <strong>and</strong> that<br />

would be more resistant to disease. His crosses included hybrids of<br />

ordinary cows with aurochs, bison, <strong>and</strong> yaks, as well as some more<br />

exotic crosses of zebras with horses.<br />

Ivanov also hoped to use artificial insemination to help preserve<br />

species that were in danger of becoming extinct. In 1926, he led an<br />

expedition to West Africa to experiment with the hybridization of<br />

different species of anthropoid apes.<br />

Impact<br />

The greatest beneficiaries of artificial insemination have been<br />

dairy farmers. Some bulls are able to sire genetically superior cows<br />

that produce exceptionally large volumes of milk. Under natural<br />

conditions, such a bull could father at most a few hundred offspring<br />

in its lifetime. Using artificial insemination, a prize bull can inseminate<br />

ten to fifteen thous<strong>and</strong> cows each year. Since frozen sperm may<br />

be purchased through the mail, this also means that dairy farmers<br />

no longer need to keep dangerous bulls on the farm. Artificial insemination<br />

has become the main method of reproduction of dairy<br />

cows, with about 150 million cows (as of 1992) produced this way<br />

throughout the world.


In the 1980’s, artificial insemination gained added importance as<br />

a method of breeding rare animals. Animals kept in zoo cages, animals<br />

that are unable to take part in normal mating, may still produce<br />

sperm that can be used to inseminate a female artificially.<br />

Some species require specific conditions of housing or diet for normal<br />

breeding to occur, conditions not available in all zoos. Such animals<br />

can still reproduce using artificial insemination.<br />

See also Abortion pill; Amniocentesis; Artificial chromosome;<br />

Birth control pill; Cloning; Genetic “fingerprinting”; Genetically engineered<br />

insulin; In vitro plant culture; Rice <strong>and</strong> wheat strains; Synthetic<br />

DNA.<br />

Further Reading<br />

Artificial insemination / 57<br />

Bearden, Henry Joe, <strong>and</strong> John W. Fuquay. Applied Animal Reproduction.<br />

5th ed. Upper Saddle River, N.J.: Prentice Hall, 2000.<br />

Foote, Robert H. Artificial Insemination to Cloning: Tracing Fifty Years<br />

of Research. Ithaca, N.Y.: Cornell University Press, 1998.<br />

Hafez, Elsayed Saad Eldin. Reproduction in Farm Animals. 6th ed.<br />

Philadelphia: Lea <strong>and</strong> Febiger, 1993.<br />

Herman, Harry August. Improving Cattle by the Millions: NAAB <strong>and</strong><br />

the Development <strong>and</strong> Worldwide Application of Artificial Insemination.<br />

Columbia: University of Missouri Press, 1981.


58<br />

Artificial kidney<br />

Artificial kidney<br />

The invention: A machine that removes waste end-products <strong>and</strong><br />

poisons out of the blood when human kidneys are not working<br />

properly.<br />

The people behind the invention:<br />

John Jacob Abel (1857-1938), a pharmacologist <strong>and</strong> biochemist<br />

known as the “father of American pharmacology”<br />

Willem Johan Kolff (1911- ), a Dutch American clinician who<br />

pioneered the artificial kidney <strong>and</strong> the artificial heart<br />

Cleansing the Blood<br />

In the human body, the kidneys are the dual organs that remove<br />

waste matter from the bloodstream <strong>and</strong> send it out of the system as<br />

urine. If the kidneys fail to work properly, this cleansing process<br />

must be done artifically—such as by a machine.<br />

John Jacob Abel was the first professor of pharmacology at Johns<br />

Hopkins University School of Medicine. Around 1912, he began to<br />

study the by-products of metabolism that are carried in the blood.<br />

This work was difficult, he realized, because it was nearly impossible<br />

to detect even the tiny amounts of the many substances in blood.<br />

Moreover, no one had yet developed a method or machine for taking<br />

these substances out of the blood.<br />

In devising a blood filtering system, Abel understood that he<br />

needed a saline solution <strong>and</strong> a membrane that would let some substances<br />

pass through but not others. Working with Leonard Rowntree<br />

<strong>and</strong> Benjamin B. Turner, he spent nearly two years figuring out<br />

how to build a machine that would perform dialysis—that is, remove<br />

metabolic by-products from blood. Finally their efforts succeeded.<br />

The first experiments were performed on rabbits <strong>and</strong> dogs. In<br />

operating the machine, the blood leaving the patient was sent flowing<br />

through a celloidin tube that had been wound loosely around a<br />

drum. An anticlotting substance (hirudin, taken out of leeches) was<br />

added to blood as the blood flowed through the tube. The drum,<br />

which was immersed in a saline <strong>and</strong> dextrose solution, rotated


slowly. As blood flowed through the immersed tubing, the pressure<br />

of osmosis removed urea <strong>and</strong> other substances, but not the plasma<br />

or cells, from the blood. The celloidin membranes allowed oxygen<br />

to pass from the saline <strong>and</strong> dextrose solution into the blood, so that<br />

purified, oxygenated blood then flowed back into the arteries.<br />

Abel studied the substances that his machine had removed from<br />

the blood, <strong>and</strong> he found that they included not only urea but also<br />

free amino acids. He quickly realized that his machine could be useful<br />

for taking care of people whose kidneys were not working properly.<br />

Reporting on his research, he wrote, “In the hope of providing a<br />

substitute in such emergencies, which might tide over a dangerous<br />

crisis...amethod has been devised by which the blood of a living<br />

animal may be submitted to dialysis outside the body, <strong>and</strong> again returned<br />

to the natural circulation.” Abel’s machine removed large<br />

quantities of urea <strong>and</strong> other poisonous substances fairly quickly, so<br />

that the process, which he called “vividiffusion,” could serve as an<br />

artificial kidney during cases of kidney failure.<br />

For his physiological research, Abel found it necessary to remove,<br />

study, <strong>and</strong> then replace large amounts of blood from living<br />

animals, all without dissolving the red blood cells, which carry oxygen<br />

to the body’s various parts. He realized that this process, which<br />

he called “plasmaphaeresis,” would make possible blood banks,<br />

where blood could be stored for emergency use.<br />

In 1914, Abel published these two discoveries in a series of three<br />

articles in the Journal of Pharmacology <strong>and</strong> Applied Therapeutics, <strong>and</strong><br />

he demonstrated his techniques in London, Engl<strong>and</strong>, <strong>and</strong> Groningen,<br />

The Netherl<strong>and</strong>s. Though he had suggested that his techniques<br />

could be used for medical purposes, he himself was interested<br />

mostly in continuing his biochemical research. So he turned to<br />

other projects in pharmacology, such as the crystallization of insulin,<br />

<strong>and</strong> never returned to studying vividiffusion.<br />

Refining the Technique<br />

Artificial kidney / 59<br />

Georg Haas, a German biochemist working in Giessen, West Germany,<br />

was also interested in dialysis; in 1915, he began to experiment<br />

with “blood washing.” After reading Abel’s 1914 writings,<br />

Haas tried substituting collodium for the celloidin that Abel had used


60 / Artificial kidney<br />

as a filtering membrane <strong>and</strong> using commercially prepared heparin<br />

instead of the homemade hirudin Abel had used to prevent blood<br />

clotting. He then used this machine on a patient <strong>and</strong> found that it<br />

showed promise, but he knew that many technical problems had to<br />

be worked out before the procedure could be used on many patients.<br />

In 1937, Willem Johan Kolff was a young physician at Groningen.<br />

He felt sad to see patients die from kidney failure, <strong>and</strong> he wanted to<br />

find a way to cure others. Having heard his colleagues talk about<br />

the possibility of using dialysis on human patients, he decided to<br />

build a dialysis machine.<br />

Kolff knew that cellophane was an excellent membrane for dialyzing,<br />

<strong>and</strong> that heparin was a good anticoagulant, but he also realized<br />

that his machine would need to be able to treat larger volumes<br />

of blood than Abel’s <strong>and</strong> Haas’s had. During World War II (1939-<br />

John Jacob Abel<br />

Born in 1857, John Jacob Abel grew up in Clevel<strong>and</strong>, Ohio,<br />

<strong>and</strong> then attended the University of Michigan. He graduated in<br />

1883 <strong>and</strong> studied for six years in Germany, which boasted the<br />

finest medical researchers of the times. He received a medical<br />

degree in 1888 in Strasbourg, transferred to Vienna, Austria, for<br />

more clinical experience, <strong>and</strong> then returned to the United States<br />

in 1891 to teach pharmacology at the University of Michigan.<br />

He had to organize his own laboratory, journal, <strong>and</strong> course of<br />

instruction. His efforts attracted the notice of Johns Hopkins<br />

University, which then had the nation’s most progressive medical<br />

school. In 1893 Abel moved there <strong>and</strong> became the first<br />

American to hold the title of professor of pharmacology. He remained<br />

at Johns Hopkins until his retirement in 1932.<br />

His biochemical research illuminated the complex interaction<br />

in the endocrine system. He isolated epinephrine (adrenaline),<br />

used his artificial kidney apparatus to demonstrate the<br />

presence of amino acids in the blood, <strong>and</strong> investigated pituitary<br />

gl<strong>and</strong> hormones <strong>and</strong> insulin.<br />

Abel died in 1938, but his influence did not. His many students<br />

took Abel’s interest in the biochemical basis of pharmacology<br />

to other universities <strong>and</strong> commercial laboratories, modernizing<br />

American drug research.


1945), with the help of the director of a nearby enamel factory, Kolff<br />

built an artificial kidney that was first tried on a patient on March<br />

17, 1943. Between March, 1943, <strong>and</strong> July 21, 1944, Kolff used his secretly<br />

constructed dialysis machines on fifteen patients, of whom<br />

only one survived. He published the results of his research in Acta<br />

Medica Sc<strong>and</strong>inavica. Even though most of his patients had not survived,<br />

he had collected information <strong>and</strong> developed the technique<br />

until he was sure dialysis would eventually work.<br />

Kolff brought machines to Amsterdam <strong>and</strong> The Hague <strong>and</strong> encouraged<br />

other physicians to try them; meanwhile, he continued to<br />

study blood dialysis <strong>and</strong> to improve his machines. In 1947, he<br />

brought improved machines to London <strong>and</strong> the United States. By<br />

the time he reached Boston, however, he had given away all of<br />

his machines. He did, however, explain the technique to John P.<br />

Merrill, a physician at the Harvard Medical School, who soon became<br />

the leading American developer of kidney dialysis <strong>and</strong> kidney-transplant<br />

surgery.<br />

Kolff himself moved to the United States, where he became an<br />

expert not only in artificial kidneys but also in artificial hearts. He<br />

helped develop the Jarvik-7 artificial heart (named for its chief inventor,<br />

Robert Jarvik), which was implanted in a patient in 1982.<br />

Impact<br />

Artificial kidney / 61<br />

Abel’s work showed that the blood carried some substances that<br />

had not been previously known <strong>and</strong> led to the development of the<br />

first dialysis machine for humans. It also encouraged interest in the<br />

possibility of organ transplants.<br />

After World War II, surgeons had tried to transplant kidneys from<br />

one animal to another, but after a few days the recipient began to reject<br />

the kidney <strong>and</strong> die. In spite of these failures, researchers in Europe <strong>and</strong><br />

America transplanted kidneys in several patients, <strong>and</strong> they used artificial<br />

kidneys to take care of the patients who were waiting for transplants.<br />

In 1954, Merrill—to whom Kolff had demonstrated an artificial<br />

kidney—successfully transplanted kidneys in identical twins.<br />

After immunosuppressant drugs (used to prevent the body<br />

from rejecting newly transplanted tissue) were discovered in 1962,<br />

transplantation surgery became much more practical. After kid-


62 / Artificial kidney<br />

ney transplants became common, the artificial kidney became simply<br />

a way of keeping a person alive until a kidney donor could be<br />

found.<br />

See also Artificial blood; Artificial heart; Blood transfusion; Genetically<br />

engineered insulin; Reserpine.<br />

Further Reading<br />

Cogan, Martin G., Patricia Schoenfeld, <strong>and</strong> Frank A. Gotch. Introduction<br />

to Dialysis. 2d ed. New York: Churchill Livingstone, 1991.<br />

DeJauregui, Ruth. One Hundred Medical Milestones That Shaped World<br />

History. San Mateo, Calif.: Bluewood Books, 1998.<br />

Noordwijk, Jacob van. Dialysing for Life: The Development of the Artificial<br />

Kidney. Boston: Kluwer Academic Publishers, 2001.


Artificial satellite<br />

Artificial satellite<br />

The invention: Sputnik I, the first object put into orbit around the<br />

earth, which began the exploration of space.<br />

The people behind the invention:<br />

Sergei P. Korolev (1907-1966), a Soviet rocket scientist<br />

Konstantin Tsiolkovsky (1857-1935), a Soviet schoolteacher <strong>and</strong><br />

the founder of rocketry in the Soviet Union<br />

Robert H. Goddard (1882-1945), an American scientist <strong>and</strong> the<br />

founder of rocketry in the United States<br />

Wernher von Braun (1912-1977), a German who worked on<br />

rocket projects<br />

Arthur C. Clarke (1917- ), the author of more than fifty<br />

books <strong>and</strong> the visionary behind telecommunications<br />

satellites<br />

A Shocking Launch<br />

In Russian, sputnik means “satellite” or “fellow traveler.” On October<br />

4, 1957, Sputnik 1, the first artificial satellite to orbit Earth, was<br />

placed into successful orbit by the Soviet Union. The launch of this<br />

small aluminum sphere, 0.58 meter in diameter <strong>and</strong> weighing 83.6<br />

kilograms, opened the doors to the frontiers of space.<br />

Orbiting Earth every 96 minutes, at 28,962 kilometers per hour,<br />

Sputnik 1 came within 215 kilometers of Earth at its closest point <strong>and</strong><br />

939 kilometers away at its farthest point. It carried equipment to<br />

measure the atmosphere <strong>and</strong> to experiment with the transmission<br />

of electromagnetic waves from space. Equipped with two radio<br />

transmitters (at different frequencies) that broadcast for twenty-one<br />

days, Sputnik 1 was in orbit for ninety-two days, until January 4,<br />

1958, when it disintegrated in the atmosphere.<br />

Sputnik 1 was launched using a Soviet intercontinental ballistic<br />

missile (ICBM) modified by Soviet rocket expert Sergei P. Korolev.<br />

After the launch of Sputnik 2, less than a month later, Chester<br />

Bowles, a former United States ambassador to India <strong>and</strong> Nepal,<br />

wrote: “Armed with a nuclear warhead, the rocket which launched<br />

63


64 / Artificial satellite<br />

Sergei P. Korolev<br />

Sergei P. Korolev’s rocket launched the Space Age: Sputnik I<br />

climbed into outer space aboard one of his R-7 missiles. Widely<br />

considered the Soviet Union’s premiere rocket scientist, he almost<br />

died in Joseph Stalin’s infamous Siberian prison camps<br />

before he could build the launchers that made his country a military<br />

superpower <strong>and</strong> pioneer of space exploration.<br />

Born in 1907, Korolev studied aeronautical engineering at<br />

the Kiev Polytechnic Institute. Upon graduation he helped found<br />

the Group for Investigation of Reactive Motion, which in the<br />

early 1930’s tested liquid-fuel rockets. His success attracted the<br />

military’s attention. It created the Reaction Propulsion Scientific<br />

Research Institute for him, <strong>and</strong> he was on the verge of testing<br />

a rocket-propelled airplane when he was arrested during a<br />

political purge in 1937 <strong>and</strong> sent as a prison laborer to the<br />

Kolyma gold mines. After Germany attacked Russia in World<br />

War II, Korolev was transferred to a prison research institute to<br />

help develop advanced aircraft.<br />

After World War II, rehabilitated in the eyes of the Soviet authorities,<br />

Korolev was placed in charge of long-range ballistic<br />

missile research. In 1953 he began to build the R-7 intercontinental<br />

ballistic missile (ICBM). While other design bureaus concentrated<br />

on developing the ICBM into a Cold War weapon,<br />

Korolev built rockets that explored the Moon with probes. His<br />

goal was to send cosmonauts there too. With his designs <strong>and</strong><br />

guidance, the Soviet space program proved that human space<br />

flight was possible in 1961, <strong>and</strong> so in 1962 he began development<br />

of the N-1, a booster that like the American Saturn V<br />

was powerful enough to send a crewed vehicle to the Moon.<br />

Tragically, Korolev died following minor surgery in 1966. The<br />

N-1 project was cancelled in 1971, along with Russian dreams of<br />

settling its citizens on the Moon.<br />

Sputnik 1 could destroy New York, Chicago, or Detroit 18 minutes<br />

after the button was pushed in Moscow.”<br />

Although the launch of Sputnik 1 came as a shock to the general<br />

public, it came as no surprise to those who followed rocketry. In<br />

June, 1957, the United States Air Force had issued a nonclassified<br />

memo stating that there was “every reason to believe that the Rus-


sian satellite shot would be made on the hundredth anniversary” of<br />

Konstantin Tsiolkovsky’s birth.<br />

Thous<strong>and</strong>s of Launches<br />

Artificial satellite / 65<br />

Rockets have been used since at least the twelfth century, when<br />

Europeans <strong>and</strong> the Chinese were using black powder devices. In<br />

1659, the Polish engineer Kazimir Semenovich published his Roketten<br />

für Luft und Wasser (rockets for air <strong>and</strong> water), which had a drawing<br />

of a three-stage rocket. Rockets were used <strong>and</strong> perfected for warfare<br />

during the nineteenth <strong>and</strong> twentieth centuries. Nazi Germany’s V-2<br />

rocket (thous<strong>and</strong>s of which were launched by Germany against Engl<strong>and</strong><br />

during the closing years of World War II) was the model for<br />

American <strong>and</strong> Soviet rocket designers between 1945 <strong>and</strong> 1957. In<br />

the Soviet Union, Tsiolkovsky had been thinking about <strong>and</strong> writing<br />

about space flight since the last decade of the nineteenth century,<br />

<strong>and</strong> in the United States, Robert H. Goddard had been thinking<br />

about <strong>and</strong> experimenting with rockets since the first decade of the<br />

twentieth century.<br />

Wernher von Braun had worked on rocket projects for Nazi Germany<br />

during World War II, <strong>and</strong>, as the war was ending in May, 1945,<br />

von Braun <strong>and</strong> several hundred other people involved in German<br />

rocket projects surrendered to American troops in Europe. Hundreds<br />

of other German rocket experts ended up in the Soviet Union<br />

to continue with their research. Tom Bower pointed out in his book<br />

The Paperclip Conspiracy: The Hunt for the Nazi Scientists (1987)—so<br />

named because American “recruiting officers had identified [Nazi]<br />

scientists to be offered contracts by slipping an ordinary paperclip<br />

onto their files”—that American rocketry research was helped<br />

tremendously by Nazi scientists who switched sides after World<br />

War II.<br />

The successful launch of Sputnik 1 convinced people that space<br />

travel was no longer simply science fiction. The successful launch of<br />

Sputnik 2 on November 3, 1957, carrying the first space traveler, a<br />

dog named Laika (who was euthanized in orbit because there were<br />

no plans to retrieve her), showed that the launch of Sputnik 1 was<br />

only the beginning of greater things to come.


66 / Artificial satellite<br />

Consequences<br />

After October 4, 1957, the Soviet Union <strong>and</strong> other nations launched<br />

more experimental satellites. On January 31, 1958, the United<br />

States sent up Explorer 1, after failing to launch a Vanguard satellite<br />

on December 6, 1957.<br />

Arthur C. Clarke, most famous for his many books of science fiction,<br />

published a technical paper in 1945 entitled “Extra-Terrestrial<br />

Relays: Can Rocket Stations Give World-Wide Radio Coverage?” In<br />

that paper, he pointed out that a satellite placed in orbit at the correct<br />

height <strong>and</strong> speed above the equator would be able to hover over<br />

the same spot on Earth. The placement of three such “geostationary”<br />

satellites would allow radio signals to be transmitted around<br />

the world. By the 1990’s, communications satellites were numerous.<br />

In the first twenty-five years after Sputnik 1 was launched, from<br />

1957 to 1982, more than two thous<strong>and</strong> objects were placed into various<br />

Earth orbits by more than twenty-four nations. On the average,<br />

something was launched into space every 3.82 days for this twentyfive-year<br />

period, all beginning with Sputnik 1.<br />

See also Communications satellite; Cruise missile; Rocket; V-2<br />

rocket; Weather satellite.<br />

Further Reading<br />

Dickson, Paul. Sputnik: The Shock of the Century. New York: Walker,<br />

2001.<br />

Heppenheimer, T. A. Countdown: A History of Space Flight. New York:<br />

John Wiley & Sons, 1997.<br />

Logsdon, John M., Roger D. Launius, <strong>and</strong> Robert W. Smith. Reconsidering<br />

Sputnik: Forty Years Since the Soviet Satellite. Australia:<br />

Harwood Academic, 2000.


Aspartame<br />

Aspartame<br />

The invention: An artificial sweetener with a comparatively natural<br />

taste widely used in carbonated beverages.<br />

The people behind the invention:<br />

Arthur H. Hayes, Jr. (1933- ), a physician <strong>and</strong> commissioner<br />

of the U.S. Food <strong>and</strong> Drug Administration (FDA)<br />

James M. Schlatter (1942- ), an American chemist<br />

Michael Sveda (1912- ), an American chemist <strong>and</strong> inventor<br />

Ludwig Frederick Audrieth (1901- ), an American chemist<br />

<strong>and</strong> educator<br />

Ira Remsen (1846-1927), an American chemist <strong>and</strong> educator<br />

Constantin Fahlberg (1850-1910), a German chemist<br />

Sweetness Without Calories<br />

People have sweetened food <strong>and</strong> beverages since before recorded<br />

history. The most widely used sweetener is sugar, or sucrose. The<br />

only real drawback to the use of sucrose is that it is a nutritive sweetener:<br />

In addition to adding a sweet taste, it adds calories. Because<br />

sucrose is readily absorbed by the body, an excessive amount can be<br />

life-threatening to diabetics. This fact alone would make the development<br />

of nonsucrose sweeteners attractive.<br />

There are three common nonsucrose sweeteners in use around<br />

the world: saccharin, cyclamates, <strong>and</strong> aspartame. Saccharin was the<br />

first of this group to be discovered, in 1879. Constantin Fahlberg<br />

synthesized saccharin based on the previous experimental work of<br />

Ira Remsen using toluene (derived from petroleum). This product<br />

was found to be three hundred to five hundred times as sweet as<br />

sugar, although some people could detect a bitter aftertaste.<br />

In 1944, the chemical family of cyclamates was discovered by<br />

Ludwig Frederick Audrieth <strong>and</strong> Michael Sveda. Although these<br />

compounds are only thirty to eighty times as sweet as sugar, there<br />

was no detectable aftertaste. By the mid-1960’s, cyclamates had<br />

resplaced saccharin as the leading nonnutritive sweetener in the<br />

United States. Although cyclamates are still in use throughout the<br />

67


68 / Aspartame<br />

world, in October, 1969, FDA removed them from the list of approved<br />

food additives because of tests that indicated possible health<br />

hazards.<br />

A Political Additive<br />

Aspartame is the latest in artificial sweeteners that are derived<br />

from natural ingredients—in this case, two amino acids, one from<br />

milk <strong>and</strong> one from bananas. Discovered by accident in 1965 by<br />

American chemist James M. Schlatter when he licked his fingers<br />

during an experiment, aspartame is 180 times as sweet as sugar. In<br />

1974, the FDA approved its use in dry foods such as gum <strong>and</strong> cereal<br />

<strong>and</strong> as a sugar replacement.<br />

Shortly after its approval for this limited application, the FDA<br />

held public hearings on the safety concerns raised by John W. Olney,<br />

a professor of neuropathology at Washington University in St. Louis.<br />

There was some indication that aspartame, when combined with<br />

the common food additive monosodium glutamate, caused brain<br />

damage in children. These fears were confirmed, but the risk of<br />

brain damage was limited to a small percentage of individuals with<br />

a rare genetic disorder. At this point, the public debate took a political<br />

turn: Senator William Proxmire charged FDA Commissioner Alex<strong>and</strong>er<br />

M. Schmidt with public misconduct. This controversy resulted<br />

in aspartame being taken off the market in 1975.<br />

In 1981, the new FDA commissioner, Arthur H. Hayes, Jr., resapproved<br />

aspartame for use in the same applications: as a tabletop<br />

sweetener, as a cold-cereal additive, in chewing gum, <strong>and</strong> for other<br />

miscellaneous uses. In 1983, the FDA approved aspartame for use in<br />

carbonated beverages, its largest application to date. Later safety<br />

studies revealed that children with a rare metabolic disease, phenylketonuria,<br />

could not ingest this sweetener without severe health<br />

risks because of the presence of phenylalanine in aspartame. This<br />

condition results in a rapid buildup in phenylalanine in the blood.<br />

Laboratories simulated this condition in rats <strong>and</strong> found that high<br />

doses of aspartame inhibited the synthesis of dopamine, a neurotransmitter.<br />

Once this happens, an increase in the frequency of seizures<br />

can occur. There was no direct evidence, however, that aspartame<br />

actually caused seizures in these experiments.


Many other compounds are being tested for use as sugar replacements,<br />

the sweetest being a relative of aspartame. This compound is<br />

seventeen thous<strong>and</strong> to fifty-two thous<strong>and</strong> times sweeter than sugar.<br />

Impact<br />

The business fallout from the approval of a new low-calorie<br />

sweetener occurred over a short span of time. In 1981, sales of this<br />

artificial sweetener by G. D. Searle <strong>and</strong> Company were $74 million.<br />

In 1983, sales rose to $336 million <strong>and</strong> exceeded half a billion dollars<br />

the following year. These figures represent sales of more than 2,500<br />

tons of this product. In 1985, 3,500 tons of aspartame were consumed.<br />

Clearly, this product’s introduction was a commercial success<br />

for Searle. During this same period, the percentage of reducedcalorie<br />

carbonated beverages containing saccharin declined from<br />

100 percent to 20 percent in an industry that had $4 billion in sales.<br />

Universally, consumers preferred products containing aspartame;<br />

the bitter aftertaste of saccharin was rejected in favor of the new, less<br />

powerful sweetener.<br />

There is a trade-off in using these products. The FDA found evidence<br />

linking both saccharin <strong>and</strong> cyclamates to an elevated incidence<br />

of cancer. Cyclamates were banned in the United States for<br />

this reason. <strong>Public</strong> resistance to this measure caused the agency to<br />

back away from its position. The rationale was that, compared to<br />

other health risks associated with the consumption of sugar (especially<br />

for diabetics <strong>and</strong> overweight persons), the chance of getting<br />

cancer was slight <strong>and</strong> therefore a risk that many people would<br />

choose to ignore. The total domination of aspartame in the sweetener<br />

market seems to support this assumption.<br />

See also Cyclamate; Genetically engineered insulin.<br />

Further Reading<br />

Aspartame / 69<br />

Blaylock, Russell L. Excitotoxins: The Taste That Kills. Santa Fe,<br />

N.Mex.: Health Press, 1998.<br />

Hull, Janet Starr. Sweet Poison: How the World’s Most Popular Artificial<br />

Sweetener Is Killing Us—My Story. Far Hills, N.J.: New<br />

Horizon Press, 1999.


70 / Aspartame<br />

Roberts, Hyman Jacob. Aspartame (NutraSweet®): Is It Safe? Philadelphia:<br />

Charles Press, 1990.<br />

Stegink, Lewis D., <strong>and</strong> Lloyd J. Filer, Aspartame: Physiology <strong>and</strong> Biochemistry.<br />

New York: M. Dekker, 1984.<br />

Stoddard, Mary Nash. Deadly Deception: Story of Aspartame, Shocking<br />

Expose of the World’s Most Controversial Sweetener. Dallas: Odenwald<br />

Press, 1998.


Assembly line<br />

Assembly line<br />

The invention: A manufacturing technique pioneered in the automobile<br />

industry by Henry Ford that lowered production costs<br />

<strong>and</strong> helped bring automobile ownership within the reach of millions<br />

of Americans in the early twentieth century.<br />

The people behind the invention:<br />

Henry Ford (1863-1947), an American carmaker<br />

Eli Whitney (1765-1825), an American inventor<br />

Elisha King Root (1808-1865), the developer of division of labor<br />

Oliver Evans (1755-1819), the inventor of power conveyors<br />

Frederick Winslow Taylor (1856-1915), an efficiency engineer<br />

A Practical Man<br />

Henry Ford built his first “horseless carriage” by h<strong>and</strong> in his<br />

home workshop in 1896. In 1903, the Ford Motor Company was<br />

born. Ford’s first product, the Model A, sold for less than one thous<strong>and</strong><br />

dollars, while other cars at that time were priced at five to ten<br />

thous<strong>and</strong> dollars each. When Ford <strong>and</strong> his partners tried, in 1905, to<br />

sell a more expensive car, sales dropped. Then, in 1907, Ford decided<br />

that the Ford Motor Company would build “a motor car for<br />

the great multitude.” It would be called the Model T.<br />

The Model T came out in 1908 <strong>and</strong> was everything that Henry Ford<br />

said it would be. Ford’s Model T was a low-priced (about $850), practical<br />

car that came in one color only: black. In the twenty years during<br />

which the Model T was built, the basic design never changed. Yet the<br />

price of the Model T, or “Tin Lizzie,” as it was affectionately called,<br />

dropped over the years to less than half that of the original Model T. As<br />

the price dropped, sales increased, <strong>and</strong> the Ford Motor Company<br />

quickly became the world’s largest automobile manufacturer.<br />

The last of more than 15 million Model T’s was made in 1927. Although<br />

it looked <strong>and</strong> drove almost exactly like the first Model T,<br />

these two automobiles were built in an entirely different way. The<br />

first was custom-built, while the last came off an assembly line.<br />

At first, Ford had built his cars in the same way everyone else<br />

71


72 / Assembly line<br />

did: one at a time. Skilled mechanics would work on a car from start<br />

to finish, while helpers <strong>and</strong> runners brought parts to these highly<br />

paid craftsmen as they were needed. After finishing one car, the mechanics<br />

<strong>and</strong> their helpers would begin the next.<br />

The Quest for Efficiency<br />

Custom-built products are good when there is little dem<strong>and</strong> <strong>and</strong><br />

buyers are willing to pay the high labor costs. This was not the case<br />

with the automobile. Ford realized that in order to make a large<br />

number of quality cars at a low price, he had to find a more efficient<br />

way to build cars. To do this, he looked to the past <strong>and</strong> the work of<br />

others. He found four ideas: interchangeable parts, continuous flow,<br />

division of labor, <strong>and</strong> elimination of wasted motion.<br />

Eli Whitney, the inventor of the cotton gin, was the first person to<br />

use interchangeable parts successfully in mass production. In 1798, the<br />

United States government asked Whitney to make several thous<strong>and</strong><br />

muskets in two years. Instead of finding <strong>and</strong> hiring gunsmiths to make<br />

the muskets by h<strong>and</strong>, Whitney used most of his time <strong>and</strong> money to design<br />

<strong>and</strong> build special machines that could make large numbers of<br />

Model-T assembly line in the Ford Motor Company’s Highl<strong>and</strong> Park Factory. (Library<br />

of Congress)


Assembly line / 73<br />

identical parts—one machine for each part that was needed to build a<br />

musket. These tools, <strong>and</strong> others Whitney made for holding, measuring,<br />

<strong>and</strong> positioning the parts, made it easy for semiskilled, <strong>and</strong> even<br />

unskilled, workers to build a large number of muskets.<br />

Production can be made more efficient by carefully arranging the<br />

different stages of production to create a “continuous flow.” Ford<br />

borrowed this idea from at least two places: the meat-packing<br />

houses of Chicago <strong>and</strong> an automatic grain mill run by Oliver Evans.<br />

Ford’s idea for a moving assembly line came from Chicago’s<br />

great meat-packing houses in the late 1860’s. Here, the bodies of animals<br />

were moved along an overhead rail past a number of workers,<br />

each of whom made a certain cut, or h<strong>and</strong>led one part of the packing<br />

job. This meant that many animals could be butchered <strong>and</strong> packaged<br />

in a single day.<br />

Ford looked to Oliver Evans for an automatic conveyor system.<br />

In 1783, Evans had designed <strong>and</strong> operated an automatic grain mill<br />

that could be run by only two workers. As one worker poured grain<br />

into a funnel-shaped container, called a “hopper,” at one end of the<br />

mill, a second worker filled sacks with flour at the other end. Everything<br />

in between was done automatically, as Evans’s conveyors<br />

passed the grain through the different steps of the milling process<br />

without any help.<br />

The idea of “division of labor” is simple: When one complicated<br />

job is divided into several easier jobs, some things can be made<br />

faster, with fewer mistakes, by workers who need fewer skills than<br />

ever before. Elisha King Root had used this principle to make the famous<br />

Colt “Six-Shooter.” In 1849, Root went to work for Samuel<br />

Colt at his Connecticut factory <strong>and</strong> proved to be a manufacturing<br />

genius. By dividing the work into very simple steps, with each step<br />

performed by one worker, Root was able to make many more guns<br />

in much less time.<br />

Before Ford applied Root’s idea to the making of engines, it took<br />

one worker one day to make one engine. By breaking down the<br />

complicated job of making an automobile engine into eighty-four<br />

simpler jobs, Ford was able to make the process much more efficient.<br />

By assigning one person to each job, Ford’s company was able<br />

to make 352 engines per day—an increase of more than 400 percent.<br />

Frederick Winslow Taylor has been called the “original efficiency


74 / Assembly line<br />

Henry Ford<br />

Henry Ford (1863-1947) was more of a synthesizer <strong>and</strong> innovator<br />

than an inventor. Others invented the gasoline-powered<br />

automobile <strong>and</strong> the techniques of mass production, but it was<br />

Ford who brought the two together. The result was the assembly<br />

line-produced Model T that the Ford Motor Company<br />

turned out in the millions from 1908 until 1927. And it changed<br />

America profoundly.<br />

Ford’s idea was to lower production costs enough so that<br />

practically everyone could afford a car, not just the wealthy. He<br />

succeeded brilliantly. The first Model T’s cost $850, rock bottom<br />

for the industry, <strong>and</strong> by 1927 the price was down to $290. Americans<br />

bought them up like no other technological marvel in the<br />

nation’s history. For years, out of every one hundred cars on the<br />

road almost forty of them were Model T’s. The basic version<br />

came with nothing on the dash board but an ignition switch,<br />

<strong>and</strong> the cars were quirky—so much so that an entire industry<br />

grew up to outfit them for the road <strong>and</strong> make sure they stayed<br />

running. Even then, they could only go up steep slopes backwards,<br />

<strong>and</strong> starting them was something of an art.<br />

Americans took the Model T to heart, affectionately nicknaming<br />

it the flivver <strong>and</strong> Tin Lizzie. This “democratization of<br />

the automobile,” as Ford called it, not only gave common people<br />

modern transportation <strong>and</strong> made them more mobile than<br />

every before; it started the American love affair with the car.<br />

Even after production stopped in 1927, the Model T Ford remained<br />

the archetype of American automobiles. As the great<br />

essayist E. B. White wrote in “Farewell My Lovely” (1936), his<br />

eulogy for the Model T, “…to a few million people who grew up<br />

with it, the old Ford practically was the American scene.”<br />

expert.” His idea was that inefficiency was caused by wasted time<br />

<strong>and</strong> wasted motion. So Taylor studied ways to eliminate wasted<br />

motion. He proved that, in the long run, doing a job too quickly was<br />

as bad as doing it too slowly. “Correct speed is the speed at which<br />

men can work hour after hour, day after day, year in <strong>and</strong> year out,<br />

<strong>and</strong> remain continuously in good health,” he said. Taylor also studied<br />

ways to streamline workers’ movements. In this way, he was<br />

able to keep wasted motion to a minimum.


Impact<br />

The changeover from custom production to mass production<br />

was an evolution rather than a revolution. Henry Ford applied the<br />

four basic ideas of mass production slowly <strong>and</strong> with care, testing<br />

each new idea before it was used. In 1913, the first moving assembly<br />

line for automobiles was being used to make Model T’s. Ford was<br />

able to make his Tin Lizzies faster than ever, <strong>and</strong> his competitors<br />

soon followed his lead. He had succeeded in making it possible for<br />

millions of people to buy automobiles.<br />

Ford’s work gave a new push to the Industrial Revolution. It<br />

showed Americans that mass production could be used to improve<br />

quality, cut the cost of making an automobile, <strong>and</strong> improve profits.<br />

In fact, the Model T was so profitable that in 1914 Ford was able to<br />

double the minimum daily wage of his workers, so that they too<br />

could afford to buy Tin Lizzies.<br />

Although Americans account for only about 6 percent of the<br />

world’s population, they now own about 50 percent of its wealth.<br />

There are more than twice as many radios in the United States as<br />

there are people. The roads are crowded with more than 180 million<br />

automobiles. Homes are filled with the sounds <strong>and</strong> sights emitting<br />

from more than 150 million television sets. Never have the people of<br />

one nation owned so much. Where did all the products—radios,<br />

cars, television sets—come from? The answer is industry, which still<br />

depends on the methods developed by Henry Ford.<br />

See also CAD/CAM; Color television; Interchangeable parts;<br />

Steelmaking process.<br />

Further Reading<br />

Assembly line / 75<br />

Abernathy, William, Kim Clark, <strong>and</strong> Alan Kantrow. Industrial Renaissance.<br />

New York: Basic Books, 1983.<br />

Bruchey, Stuart. Enterprise: The Dynamic Economy of a Free People.<br />

Cambridge, Mass.: Harvard University Press, 1990.<br />

Flink, James. The Car Culture. Cambridge, Mass.: MIT Press, 1975.<br />

Hayes, Robert. Restoring Our Competitive Edge. New York: Wiley, 1984.<br />

Olson, Sidney. Young Henry Ford: A Picture History of the First Forty<br />

Years. Detroit: Wayne State University Press, 1997.Wiley, 1984.


76<br />

Atomic bomb<br />

Atomic bomb<br />

The invention: A weapon of mass destruction created during<br />

World War II that utilized nuclear fission to create explosions<br />

equivalent to thous<strong>and</strong>s of tons of trinitrotoluene (TNT),<br />

The people behind the invention:<br />

J. Robert Oppenheimer (1904-1967), an American physicist<br />

Leslie Richard Groves (1896-1970), an American engineer <strong>and</strong><br />

Army general<br />

Enrico Fermi (1900-1954), an Italian American nuclear physicist<br />

Niels Bohr (1885-1962), a Danish physicist<br />

Energy on a Large Scale<br />

The first evidence of uranium fission (the splitting of uranium<br />

atoms) was observed by German chemists Otto Hahn <strong>and</strong> Fritz<br />

Strassmann in Berlin at the end of 1938. When these scientists discovered<br />

radioactive barium impurities in neutron-irradiated uranium,<br />

they wrote to their colleague Lise Meitner in Sweden. She <strong>and</strong><br />

her nephew, physicist Otto Robert Frisch, calculated the large release<br />

of energy that would be generated during the nuclear fission<br />

of certain elements. This result was reported to Niels Bohr in Copenhagen.<br />

Meanwhile, similar fission energies were measured by Frédéric<br />

Joliot <strong>and</strong> his associates in Paris, who demonstrated the release of<br />

up to three additional neutrons during nuclear fission. It was recognized<br />

immediately that if neutron-induced fission released enough<br />

additional neutrons to cause at least one more such fission, a selfsustaining<br />

chain reaction would result, yielding energy on a large<br />

scale.<br />

While visiting the United States from January to May of 1939,<br />

Bohr derived a theory of fission with John Wheeler of Princeton<br />

University. This theory led Bohr to predict that the common isotope<br />

uranium 238 (which constitutes 99.3 percent of naturally occurring<br />

uranium) would require fast neutrons for fission, but that the rarer<br />

uranium 235 would fission with neutrons of any energy. This meant


that uranium 235 would be far more suitable for use in any sort of<br />

bomb. Uranium bombardment in a cyclotron led to the discovery of<br />

plutonium in 1940 <strong>and</strong> the discovery that plutonium 239 was fissionable—<strong>and</strong><br />

thus potentially good bomb material. Uranium 238<br />

was then used to “breed” (create) plutonium 239, which was then<br />

separated from the uranium by chemical methods.<br />

During 1942, the Manhattan District of the Army Corps of Engineers<br />

was formed under General Leslie Richard Groves, an engineer<br />

<strong>and</strong> Army general who contracted with E. I. Du Pont de<br />

Nemours <strong>and</strong> Company to construct three secret atomic cities at a<br />

total cost of $2 billion. At Oak Ridge, Tennessee, twenty-five thous<strong>and</strong><br />

workers built a 1,000-kilowatt reactor as a pilot plant. Asecond<br />

city of sixty thous<strong>and</strong> inhabitants was built at Hanford, Washington,<br />

where three huge reactors <strong>and</strong> remotely controlled plutoniumextraction<br />

plants were completed in early 1945.<br />

A Sustained <strong>and</strong> Awesome Roar<br />

Atomic bomb / 77<br />

Studies of fast-neutron reactions for an atomic bomb were brought<br />

together in Chicago in June of 1942 under the leadership of J. Robert<br />

Oppenheimer. He soon became a personal adviser to Groves, who<br />

built for Oppenheimer a laboratory for the design <strong>and</strong> construction<br />

of the bomb at Los Alamos, New Mexico. In 1943, Oppenheimer<br />

gathered two hundred of the best scientists in what was by now being<br />

called the Manhattan Project to live <strong>and</strong> work in this third secret<br />

city.<br />

Two bomb designs were developed. A gun-type bomb called<br />

“Little Boy” used 15 kilograms of uranium 235 in a 4,500-kilogram<br />

cylinder about 2 meters long <strong>and</strong> 0.5 meter in diameter, in which a<br />

uranium bullet could be fired into three uranium target rings to<br />

form a critical mass. An implosion-type bomb called “Fat Man” had<br />

a 5-kilogram spherical core of plutonium about the size of an orange,<br />

which could be squeezed inside a 2,300-kilogram sphere<br />

about 1.5 meters in diameter by properly shaped explosives to make<br />

the mass critical in the shorter time required for the faster plutonium<br />

fission process.<br />

A flat scrub region 200 kilometers southeast of Alamogordo,<br />

called Trinity, was chosen for the test site, <strong>and</strong> observer bunkers


78 / Atomic bomb<br />

were built about 10 kilometers from a 30-meter steel tower. On July<br />

13, 1945, one of the plutonium bombs was assembled at the site; the<br />

next morning, it was raised to the top of the tower. Two days later,<br />

on July 16, after a short thunderstorm delay, the bomb was detonated<br />

at 5:30 a.m. The resulting implosion initiated a chain reaction<br />

of nearly 60 fission generations in about a microsecond. It produced<br />

an intense flash of light <strong>and</strong> a fireball that exp<strong>and</strong>ed to a diameter of<br />

about 600 meters in two seconds, rose to a height of more than 12 kilometers,<br />

<strong>and</strong> formed an ominous mushroom shape. Forty seconds<br />

later, an air blast hit the observer bunkers, followed by a sustained<br />

<strong>and</strong> awesome roar. Measurements confirmed that the explosion had<br />

the power of 18.6 kilotons of trinitrotoluene (TNT), nearly four<br />

times the predicted value.<br />

Impact<br />

On March 9, 1945, 325 American B-29 bombers dropped 2,000<br />

tons of incendiary bombs on Tokyo, resulting in 100,000 deaths from<br />

the fire storms that swept the city. Nevertheless, the Japanese military<br />

refused to surrender, <strong>and</strong> American military plans called for an<br />

invasion of Japan, with estimates of up to a half million American<br />

casualties, plus as many as 2 million Japanese casualties. On August<br />

6, 1945, after authorization by President Harry S. Truman, the<br />

B-29 Enola Gay dropped the uranium Little Boy bomb on Hiroshima<br />

at 8:15 a.m. On August 9, the remaining plutonium Fat Man bomb<br />

was dropped on Nagasaki. Approximately 100,000 people died at<br />

Hiroshima (out of a population of 400,000), <strong>and</strong> about 50,000 more<br />

died at Nagasaki. Japan offered to surrender on August 10, <strong>and</strong> after<br />

a brief attempt by some army officers to rebel, an official announcement<br />

by Emperor Hirohito was broadcast on August 15.<br />

The development of the thermonuclear fusion bomb, in which<br />

hydrogen isotopes could be fused together by the force of a fission<br />

explosion to produce helium nuclei <strong>and</strong> almost unlimited energy,<br />

had been proposed early in the Manhattan Project by physicist Edward<br />

Teller. Little effort was invested in the hydrogen bomb until<br />

after the surprise explosion of a Soviet atomic bomb in September,<br />

1949, which had been built with information stolen from the Manhattan<br />

Project. After three years of development under Teller’s


guidance, the first successful H-bomb was exploded on November<br />

1, 1952, obliterating the Elugelab atoll in the Marshall Isl<strong>and</strong>s of<br />

the South Pacific. The arms race then accelerated until each side had<br />

stockpiles of thous<strong>and</strong>s of H-bombs.<br />

The Manhattan Project opened a P<strong>and</strong>ora’s box of nuclear weapons<br />

that would plague succeeding generations, but it contributed<br />

more than merely weapons. About 19 percent of the electrical energy<br />

in the United States is generated by about 110 nuclear reactors<br />

producing more than 100,000 megawatts of power. More than 400<br />

reactors in thirty countries provide 300,000 megawatts of the world’s<br />

power. Reactors have made possible the widespread use of radioisotopes<br />

in medical diagnosis <strong>and</strong> therapy. Many of the techniques<br />

for producing <strong>and</strong> using these isotopes were developed by the hundreds<br />

of nuclear physicists who switched to the field of radiation<br />

biophysics after the war, ensuring that the benefits of their wartime<br />

efforts would reach the public.<br />

See also Airplane; Breeder reactor; Cruise missile; Hydrogen<br />

bomb; Rocket; Stealth aircraft; V-2 rocket.<br />

Further Reading<br />

Atomic bomb / 79<br />

Goudsmit, Samuel Abraham, <strong>and</strong> Albert E. Moyer. The History of<br />

Modern Physics, 1800-1950. Los Angeles: Tomash Publishers, 1983.<br />

Henshall, Phillip. The Nuclear Axis: Germany, Japan, <strong>and</strong> the Atom<br />

Bomb Race, 1939-1945. Stoud: Sutton, 2000.<br />

Krieger, David. Splitting the Atom: A Chronology of the Nuclear Age.<br />

Santa Barbara, Calif.: Nuclear Age Peace Foundation, 1998.<br />

Smith, June. How the Atom Bombs Began, 1939-1946. London: Brockwell,<br />

1988.


80<br />

Atomic clock<br />

Atomic clock<br />

The invention: A clock using the ammonia molecule as its oscillator<br />

that surpasses mechanical clocks in long-term stability, precision,<br />

<strong>and</strong> accuracy.<br />

The person behind the invention:<br />

Harold Lyons (1913-1984), an American physicist<br />

Time Measurement<br />

The accurate measurement of basic quantities, such as length,<br />

electrical charge, <strong>and</strong> temperature, is the foundation of science. The<br />

results of such measurements dictate whether a scientific theory is<br />

valid or must be modified or even rejected. Many experimental<br />

quantities change over time, but time cannot be measured directly.<br />

It must be measured by the occurrence of an oscillation or rotation,<br />

such as the twenty-four-hour rotation of the earth. For centuries, the<br />

rising of the Sun was sufficient as a timekeeper, but the need for<br />

more precision <strong>and</strong> accuracy increased as human knowledge grew.<br />

Progress in science can be measured by how accurately time has<br />

been measured at any given point. In 1713, the British government,<br />

after the disastrous sinking of a British fleet in 1707 because of a miscalculation<br />

of longitude, offered a reward of 20,000 pounds for the<br />

invention of a ship’s chronometer (a very accurate clock). Latitude<br />

is determined by the altitude of the Sun above the southern horizon<br />

at noon local time, but the determination of longitude requires an<br />

accurate clock set at Greenwich, Engl<strong>and</strong>, time. The difference between<br />

the ship’s clock <strong>and</strong> the local sun time gives the ship’s longitude.<br />

This permits the accurate charting of new l<strong>and</strong>s, such as those<br />

that were being explored in the eighteenth century. John Harrison,<br />

an English instrument maker, eventually built a chronometer that<br />

was accurate within one minute after five months at sea. He received<br />

his reward from Parliament in 1765.<br />

Atomic Clocks Provide Greater Stability<br />

A clock contains four parts: energy to keep the clock operating,<br />

an oscillator, an oscillation counter, <strong>and</strong> a display. A gr<strong>and</strong>father


Atomic clock / 81<br />

clock has weights that fall slowly, providing energy that powers the<br />

clock’s gears. The pendulum, a weight on the end of a rod, swings<br />

back <strong>and</strong> forth (oscillates) with a regular beat. The length of the rod<br />

determines the pendulum’s period of oscillation. The pendulum is<br />

attached to gears that count the oscillations <strong>and</strong> drive the display<br />

h<strong>and</strong>s.<br />

There are limits to a mechanical clock’s accuracy <strong>and</strong> stability.<br />

The length of the rod changes as the temperature changes, so the<br />

period of oscillation changes. Friction in the gears changes as they<br />

wear out. Making the clock smaller increases its accuracy, precision,<br />

<strong>and</strong> stability. Accuracy is how close the clock is to telling the actual<br />

time. Stability indicates how the accuracy changes over time, while<br />

precision is the number of accurate decimal places in the display. A<br />

gr<strong>and</strong>father clock, for example, might be accurate to ten seconds per<br />

day <strong>and</strong> precise to a second, while having a stability of minutes per<br />

week.<br />

Applying an electrical signal to a quartz crystal will make the<br />

crystal oscillate at its natural vibration frequency, which depends on<br />

its size, its shape, <strong>and</strong> the way in which it was cut from the larger<br />

crystal. Since the faster a clock’s oscillator vibrates, the more precise<br />

the clock, a crystal-based clock is more precise than a large pendulum<br />

clock. By keeping the crystal under constant temperature, the<br />

clock is kept accurate, but it eventually loses its stability <strong>and</strong> slowly<br />

wears out.<br />

In 1948, Harold Lyons <strong>and</strong> his colleagues at the National Bureau<br />

of St<strong>and</strong>ards (NBS) constructed the first atomic clock, which used<br />

the ammonia molecule as its oscillator. Such a clock is called an<br />

atomic clock because, when it operates, a nitrogen atom vibrates.<br />

The pyramid-shaped ammonia molecule is composed of a triangular<br />

base; there is a hydrogen atom at each corner <strong>and</strong> a nitrogen<br />

atom at the top of the pyramid. The nitrogen atom does not remain<br />

at the top; if it absorbs radio waves of the right energy <strong>and</strong> frequency,<br />

it passes through the base to produce an upside-down pyramid<br />

<strong>and</strong> then moves back to the top. This oscillation frequency occurs<br />

at 23,870 megacycles (1 megacycle equals 1 million cycles) per<br />

second.<br />

Lyons’s clock was actually a quartz-ammonia clock, since the signal<br />

from a quartz crystal produced radio waves of the crystal’s fre-


82 / Atomic clock<br />

quency that were fed into an ammonia-filled tube. If the radio<br />

waves were at 23,870 megacycles, the ammonia molecules absorbed<br />

the waves; a detector sensed this, <strong>and</strong> it sent no correction signal to<br />

the crystal. If radio waves deviated from 23,870 megacycles, the ammonia<br />

did not absorb them, the detector sensed the unabsorbed radio<br />

waves, <strong>and</strong> a correction signal was sent to the crystal. The<br />

atomic clock’s accuracy <strong>and</strong> precision were comparable to those of a<br />

quartz-based clock—one part in a hundred million—but the atomic<br />

clock was more stable because molecules do not wear out.<br />

The atomic clock’s accuracy was improved by using cesium<br />

133 atoms as the source of oscillation. These atoms oscillate at<br />

9,192,631,770 plus or minus 20 cycles per second. They are accurate<br />

to a billionth of a second per day <strong>and</strong> precise to nine decimal places.<br />

A cesium clock is stable for years. Future developments in atomic<br />

clocks may see accuracies of one part in a million billions.<br />

Impact<br />

The development of stable, very accurate atomic clocks has farreaching<br />

implications for many areas of science. Global positioning<br />

satellites send signals to receivers on ships <strong>and</strong> airplanes. By timing<br />

the signals, the receiver’s position is calculated to within several<br />

meters of its true location.<br />

Chemists are interested in finding the speed of chemical reactions,<br />

<strong>and</strong> atomic clocks are used for this purpose. The atomic clock<br />

led to the development of the maser (an acronym for microwave amplification<br />

by stimulated emission of radiation), which is used to<br />

amplify weak radio signals, <strong>and</strong> the maser led to the development<br />

of the laser, a light-frequency maser that has more uses than can be<br />

listed here.<br />

Atomic clocks have been used to test Einstein’s theories of relativity<br />

that state that time on a moving clock, as observed by a stationary<br />

observer, slows down, <strong>and</strong> that a clock slows down near a<br />

large mass (because of the effects of gravity). Under normal conditions<br />

of low velocities <strong>and</strong> low mass, the changes in time are very<br />

small, but atomic clocks are accurate <strong>and</strong> stable enough to detect<br />

even these small changes. In such experiments, three sets of clocks<br />

were used—one group remained on Earth, one was flown west


around the earth on a jet, <strong>and</strong> the last set was flown east. By comparing<br />

the times of the in-flight sets with the stationary set, the<br />

predicted slowdowns of time were observed <strong>and</strong> the theories were<br />

verified.<br />

See also Carbon dating; Cyclotron; Electric clock; Laser; Synchrocyclotron;<br />

Tevatron accelerator.<br />

Further Reading<br />

Atomic clock / 83<br />

Audoin, Claude, <strong>and</strong> Bernard Guinot. The Measurement of Time:<br />

Time, Frequency, <strong>and</strong> the Atomic Clock. New York: Cambridge University<br />

Press, 2001.<br />

Barnett, Jo Ellen. Time’s Pendulum: The Quest to Capture Time—From<br />

Sundials to Atomic Clocks. New York: Plenum Trade, 1998.<br />

Bendick, Jeanne. The First Book of Time. New York: F. Watts, 1970.<br />

“Ultra-Accurate Atomic Clock Unveiled at NIST Laboratory.” Research<br />

<strong>and</strong> Development 42, no. 2 (February, 2000).


84<br />

Atomic-powered ship<br />

Atomic-powered ship<br />

The invention: The world’s first atomic-powered merchant ship<br />

demonstrated a peaceful use of atomic power.<br />

The people behind the invention:<br />

Otto Hahn (1879-1968), a German chemist<br />

Enrico Fermi (1901-1954), an Italian American physicist<br />

Dwight D. Eisenhower (1890-1969), president of the United<br />

States, 1953-1961<br />

Splitting the Atom<br />

In 1938, Otto Hahn, working at the Kaiser Wilhelm Institute for<br />

Chemistry, discovered that bombarding uranium atoms with neutrons<br />

causes them to split into two smaller, lighter atoms. A large<br />

amount of energy is released during this process, which is called<br />

“fission.” When one kilogram of uranium is fissioned, it releases the<br />

same amount of energy as does the burning of 3,000 metric tons of<br />

coal. The fission process also releases new neutrons.<br />

Enrico Fermi suggested that these new neutrons could be used to<br />

split more uranium atoms <strong>and</strong> produce a chain reaction. Fermi <strong>and</strong><br />

his assistants produced the first human-made chain reaction at the<br />

University of Chicago on December 2, 1942. Although the first use<br />

of this new energy source was the atomic bombs that were used to<br />

defeat Japan in World War II, it was later realized that a carefully<br />

controlled chain reaction could produce useful energy. The submarine<br />

Nautilus, launched in 1954, used the energy released from fission<br />

to make steam to drive its turbines.<br />

U.S. President Dwight David Eisenhower proposed his “Atoms<br />

for Peace” program in December, 1953. On April 25, 1955, President<br />

Eisenhower announced that the “Atoms for Peace” program would<br />

be exp<strong>and</strong>ed to include the design <strong>and</strong> construction of an atomicpowered<br />

merchant ship, <strong>and</strong> he signed the legislation authorizing<br />

the construction of the ship in 1956.


Savannah’s Design <strong>and</strong> Construction<br />

Atomic-powered ship / 85<br />

A contract to design an atomic-powered merchant ship was<br />

awarded to George G. Sharp, Inc., on April 4, 1957. The ship was to<br />

carry approximately one hundred passengers (later reduced to sixty<br />

to reduce the ship’s cost) <strong>and</strong> 10,886 metric tons of cargo while making<br />

a speed of 21 knots, about 39 kilometers per hour. The ship was<br />

to be 181 meters long <strong>and</strong> 23.7 meters wide. The reactor was to provide<br />

steam for a 20,000-horsepower turbine that would drive the<br />

ship’s propeller. Most of the ship’s machinery was similar to that of<br />

existing ships; the major difference was that steam came from a reactor<br />

instead of a coal- or oil-burning boiler.<br />

New York Shipbuilding Corporation of Camden, New Jersey,<br />

won the contract to build the ship on November 16, 1957. States Marine<br />

Lines was selected in July, 1958, to operate the ship. It was christened<br />

Savannah <strong>and</strong> launched on July 21, 1959. The name Savannah<br />

was chosen to honor the first ship to use steam power while crossing<br />

an ocean. This earlier Savannah was launched in New York City<br />

in 1818.<br />

Ships are normally launched long before their construction is<br />

complete, <strong>and</strong> the new Savannah was no exception. It was finally<br />

turned over to States Marine Lines on May 1, 1962. After extensive<br />

testing by its operators <strong>and</strong> delays caused by labor union disputes,<br />

it began its maiden voyage from Yorktown, Virginia, to Savannah,<br />

Georgia, on August 20, 1962. The original budget for design <strong>and</strong><br />

construction was $35 million, but by this time, the actual cost was<br />

about $80 million.<br />

Savannah‘s nuclear reactor was fueled with about 7,000 kilograms<br />

(15,400 pounds) of uranium. Uranium consists of two forms,<br />

or “isotopes.” These are uranium 235, which can fission, <strong>and</strong> uranium<br />

238, which cannot. Naturally occurring uranium is less than 1<br />

percent uranium 235, but the uranium in Savannah‘s reactor had<br />

been enriched to contain nearly 5 percent of this isotope. Thus, there<br />

was less than 362 kilograms of usable uranium in the reactor. The<br />

ship was able to travel about 800,000 kilometers on this initial fuel<br />

load. Three <strong>and</strong> a half million kilograms of water per hour flowed<br />

through the reactor under a pressure of 5,413 kilograms per square<br />

centimeter. It entered the reactor at 298.8 degrees Celsius <strong>and</strong> left at


86 / Atomic-powered ship<br />

317.7 degrees Celsius. Water leaving the reactor passed through a<br />

heat exchanger called a “steam generator.” In the steam generator,<br />

reactor water flowed through many small tubes. Heat passed through<br />

the walls of these tubes <strong>and</strong> boiled water outside them. About<br />

113,000 kilograms of steam per hour were produced in this way at a<br />

pressure of 1,434 kilograms per square centimeter <strong>and</strong> a temperature<br />

of 240.5 degrees Celsius.<br />

Labor union disputes dogged Savannah‘s early operations, <strong>and</strong> it<br />

did not start its first trans-Atlantic crossing until June 8, 1964. Savannah<br />

was never a money maker. Even in the 1960’s, the trend was toward<br />

much bigger ships. It was announced that the ship would be<br />

retired in August, 1967, but that did not happen. It was finally put<br />

out of service in 1971. Later, Savannah was placed on permanent display<br />

at Charleston, South Carolina.<br />

Consequences<br />

Following the United States’ lead, Germany <strong>and</strong> Japan built<br />

atomic-powered merchant ships. The Soviet Union is believed to<br />

have built several atomic-powered icebreakers. Germany’s Otto<br />

Hahn, named for the scientist who first split the atom, began service<br />

in 1968, <strong>and</strong> Japan’s Mutsuai was under construction as Savannah retired.<br />

Numerous studies conducted in the early 1970’s claimed to prove<br />

that large atomic-powered merchant ships were more profitable<br />

than oil-fired ships of the same size. Several conferences devoted to<br />

this subject were held, but no new ships were built.<br />

Although the U.S. Navy has continued to use reactors to power<br />

submarines, aircraft carriers, <strong>and</strong> cruisers, atomic power has not<br />

been widely used for merchant-ship propulsion. Labor union problems<br />

such as those that haunted Savannah, high insurance costs, <strong>and</strong><br />

high construction costs are probably the reasons. <strong>Public</strong> opinion, after<br />

the reactor accidents at Three Mile Isl<strong>and</strong> (in 1979) <strong>and</strong> Chernobyl<br />

(in 1986) is also a factor.<br />

See also Gyrocompass; Hovercraft; Nuclear reactor; Supersonic<br />

passenger plane.


Further Reading<br />

Atomic-powered ship / 87<br />

Epstein, Sam Epstein, Beryl (William) Epstein, <strong>and</strong> Raymond<br />

Burns. Enrico Fermi, Father of Atomic Power. Champaign, Ill. Garrard<br />

Publishing, 1970.<br />

Hahn, Otto, <strong>and</strong> Wily Ley. Otto Hahn: A Scientific Autobiography.<br />

New York: C. Scribner’s Sons, 1966.<br />

Hoffman, Klaus. Otto Hahn: Achievement <strong>and</strong> Responsibility. New<br />

York: Springer, 2001.<br />

“The Race to Power Bigger, Faster Ships.” Business Week 2305 (November<br />

10, 1973).<br />

“Underway on Nuclear Power.” All H<strong>and</strong>s 979 (November, 1998).


88<br />

Autochrome plate<br />

Autochrome plate<br />

The invention: The first commercially successful process in which<br />

a single exposure in a regular camera produced a color image.<br />

The people behind the invention:<br />

Louis Lumière (1864-1948), a French inventor <strong>and</strong> scientist<br />

Auguste Lumière (1862-1954), an inventor, physician, physicist,<br />

chemist, <strong>and</strong> botanist<br />

Alphonse Seyewetz, a skilled scientist <strong>and</strong> assistant of the<br />

Lumière brothers<br />

Adding Color<br />

In 1882, Antoine Lumière, painter, pioneer photographer, <strong>and</strong> father<br />

of Auguste <strong>and</strong> Louis, founded a factory to manufacture photographic<br />

gelatin dry-plates. After the Lumière brothers took over the<br />

factory’s management, they exp<strong>and</strong>ed production to include roll<br />

film <strong>and</strong> printing papers in 1887 <strong>and</strong> also carried out joint research<br />

that led to fundamental discoveries <strong>and</strong> improvements in photographic<br />

development <strong>and</strong> other aspects of photographic chemistry.<br />

While recording <strong>and</strong> reproducing the actual colors of a subject<br />

was not possible at the time of photography’s inception (about<br />

1822), the first practical photographic process, the daguerreotype,<br />

was able to render both striking detail <strong>and</strong> good tonal quality. Thus,<br />

the desire to produce full-color images, or some approximation to<br />

realistic color, occupied the minds of many photographers <strong>and</strong> inventors,<br />

including Louis <strong>and</strong> Auguste Lumière, throughout the<br />

nineteenth century.<br />

As researchers set out to reproduce the colors of nature, the first<br />

process that met with any practical success was based on the additive<br />

color theory expounded by the Scottish physicist James Clerk<br />

Maxwell in 1861. He believed that any color can be created by<br />

adding together red, green, <strong>and</strong> blue light in certain proportions.<br />

Maxwell, in his experiments, had taken three negatives through<br />

screens or filters of these additive primary colors. He then took<br />

slides made from these negatives <strong>and</strong> projected the slides through


Antoine Lumière <strong>and</strong> Sons<br />

Autochrome plate / 89<br />

Antoine Lumière was explosive in temperament, loved a<br />

good fight, <strong>and</strong> despised Americans. With these qualities—<strong>and</strong><br />

his sons to take care of the practicalities—he turned France into<br />

a leader of the early photography <strong>and</strong> film industries.<br />

Lumière was born into a family of wine growers in 1840 <strong>and</strong><br />

trained to be a sign painter. Bored with his job, he learned the<br />

new art of photography, set up a studio in Lyon, <strong>and</strong> began to<br />

experiment with ways to make his own photographic plates.<br />

Failures led to frustration, <strong>and</strong> frustration ignited his temper,<br />

which often ended in his smashing the furniture <strong>and</strong> glassware<br />

nearby. His sons, Auguste, born 1862, <strong>and</strong> Louis, born 1864,<br />

came to the rescue. Louis, a science whiz as a teenager, succeeded<br />

where his father had failed. The dry plate he invented,<br />

Blue Label, was the most sensitive yet. The Lumières set up a<br />

factory to manufacture the plates <strong>and</strong> quickly found themselves<br />

wealthy, but the old man’s love of extravagant spending<br />

<strong>and</strong> parties led them to the door of bankruptcy in 1882. His sons<br />

had to take control to save the family finances.<br />

The father, an ardent French patriot, soon threw himself into<br />

a new crusade. American tariffs made it impossible for the<br />

Lumières to make a profit selling their photographic plates in<br />

the United States, which so angered the old man that he<br />

looked for revenge. He found it in the form of Thomas Edison’s<br />

Kinetoscope in 1894. He got hold of samples, <strong>and</strong> soon<br />

the family factory was making motion picture film of its own<br />

<strong>and</strong> could undersell Edison in France. Louis also invented a<br />

projector, adapted from a sewing machine, that made it possible<br />

for movies to be shown to audiences.<br />

Before Antoine Lumière died in Paris in 1911, he had the satisfaction<br />

of seeing his beloved France producing better, cheaper<br />

photographic products than those available from America, as<br />

well as becoming a pioneer in film making.<br />

the same filters onto a screen so that their images were superimposed.<br />

As a result, he found that it was possible to reproduce the exact<br />

colors as well as the form of an object.<br />

Unfortunately, since colors could not be printed in their tonal<br />

relationships on paper before the end of the nineteenth century,


90 / Autochrome plate<br />

Maxwell’s experiment was unsuccessful. Although Frederick E.<br />

Ives of Philadelphia, in 1892, optically united three transparencies<br />

so that they could be viewed in proper alignment by looking through<br />

a peephole, viewing the transparencies was still not as simple as<br />

looking at a black-<strong>and</strong>-white photograph.<br />

The Autochrome Plate<br />

The first practical method of making a single photograph that<br />

could be viewed without any apparatus was devised by John Joly of<br />

Dublin in 1893. Instead of taking three separate pictures through<br />

three colored filters, he took one negative through one filter minutely<br />

checkered with microscopic areas colored red, green, <strong>and</strong><br />

blue. The filter <strong>and</strong> the plate were exactly the same size <strong>and</strong> were<br />

placed in contact with each other in the camera. After the plate was<br />

developed, a transparency was made, <strong>and</strong> the filter was permanently<br />

attached to it. The black-<strong>and</strong>-white areas of the picture allowed<br />

more or less light to shine through the filters; if viewed from a<br />

proper distance, the colored lights blended to form the various colors<br />

of nature.<br />

In sum, the potential principles of additive color <strong>and</strong> other methods<br />

<strong>and</strong> their potential applications in photography had been discovered<br />

<strong>and</strong> even experimentally demonstrated by 1880. Yet a practical<br />

process of color photography utilizing these principles could<br />

not be produced until a truly panchromatic emulsion was available,<br />

since making a color print required being able to record the primary<br />

colors of the light cast by the subject.<br />

Louis <strong>and</strong> Auguste Lumière, along with their research associate<br />

Alphonse Seyewetz, succeeded in creating a single-plate process<br />

based on this method in 1903. It was introduced commercially as the<br />

autochrome plate in 1907 <strong>and</strong> was soon in use throughout the<br />

world. This process is one of many that take advantage of the limited<br />

resolving power of the eye. Grains or dots too small to be recognized<br />

as separate units are accepted in their entirety <strong>and</strong>, to the<br />

sense of vision, appear as tones <strong>and</strong> continuous color.


Impact<br />

While the autochrome plate remained one of the most popular<br />

color processes until the 1930’s, soon this process was superseded by<br />

subtractive color processes. Leopold Mannes <strong>and</strong> Leopold Godowsky,<br />

both musicians <strong>and</strong> amateur photographic researchers who eventually<br />

joined forces with Eastman Kodak research scientists, did the<br />

most to perfect the Lumière brothers’ advances in making color<br />

photography practical. Their collaboration led to the introduction in<br />

1935 of Kodachrome, a subtractive process in which a single sheet of<br />

film is coated with three layers of emulsion, each sensitive to one<br />

primary color. A single exposure produces a color image.<br />

Color photography is now commonplace. The amateur market is<br />

enormous, <strong>and</strong> the snapshot is almost always taken in color. Commercial<br />

<strong>and</strong> publishing markets use color extensively. Even photography<br />

as an art form, which was done in black <strong>and</strong> white through<br />

most of its history, has turned increasingly to color.<br />

See also Color film; Instant photography; Xerography.<br />

Further Reading<br />

Autochrome plate / 91<br />

Collins, Douglas. The Story of Kodak. New York: Harry N. Abrams,<br />

1990.<br />

Glendinning, Peter. Color Photography: History, Theory, <strong>and</strong> Darkroom<br />

Technique. Englewood Cliffs, N.J.: Prentice-Hall, 1985.<br />

Lartigue, Jacques-Henri, <strong>and</strong> Georges Herscher. The Autochromes of<br />

J. H. Lartigue, 1912-1927. New York: Viking Press, 1981.<br />

Tolstoy, Ivan. James Clerk Maxwell: A Biography. Chicago: University<br />

of Chicago Press, 1982.<br />

Wood, John. The Art of the Autochrome: The Birth of Color Photography.<br />

Iowa City: University of Iowa Press, 1993.


92<br />

BASIC programming language<br />

BASIC programming language<br />

The invention: An interactive computer system <strong>and</strong> simple programming<br />

language that made it easier for nontechnical people<br />

to use computers.<br />

The people behind the invention:<br />

John G. Kemeny (1926-1992), the chairman of Dartmouth’s<br />

mathematics department<br />

Thomas E. Kurtz (1928- ), the director of the Kiewit<br />

Computation Center at Dartmouth<br />

Bill Gates (1955- ), a cofounder <strong>and</strong> later chairman of the<br />

board <strong>and</strong> chief operating officer of the Microsoft<br />

Corporation<br />

The Evolution of Programming<br />

The first digital computers were developed during World War II<br />

(1939-1945) to speed the complex calculations required for ballistics,<br />

cryptography, <strong>and</strong> other military applications. Computer technology<br />

developed rapidly, <strong>and</strong> the 1950’s <strong>and</strong> 1960’s saw computer systems<br />

installed throughout the world. These systems were very large<br />

<strong>and</strong> expensive, requiring many highly trained people for their operation.<br />

The calculations performed by the first computers were determined<br />

solely by their electrical circuits. In the 1940’s, The American<br />

mathematician John von Neumann <strong>and</strong> others pioneered the idea of<br />

computers storing their instructions in a program, so that changes<br />

in calculations could be made without rewiring their circuits. The<br />

programs were written in machine language, long lists of zeros <strong>and</strong><br />

ones corresponding to on <strong>and</strong> off conditions of circuits. During the<br />

1950’s, “assemblers” were introduced that used short names for<br />

common sequences of instructions <strong>and</strong> were, in turn, transformed<br />

into the zeros <strong>and</strong> ones intelligible to the computer. The late 1950’s<br />

saw the introduction of high-level languages, notably Formula Translation<br />

(FORTRAN), Common Business Oriented Language (COBOL),<br />

<strong>and</strong> Algorithmic Language (ALGOL), which used English words to


communicate instructions to the computer. Unfortunately, these<br />

high-level languages were complicated; they required some knowledge<br />

of the computer equipment <strong>and</strong> were designed to be used by<br />

scientists, engineers, <strong>and</strong> other technical experts.<br />

Developing BASIC<br />

BASIC programming language / 93<br />

John G. Kemeny was chairman of the department of mathematics<br />

at Dartmouth College in Hanover, New Hampshire. In 1962,<br />

Thomas E. Kurtz, Dartmouth’s computing director, approached<br />

Kemeny with the idea of implementing a computer system at Dartmouth<br />

College. Both men were dedicated to the idea that liberal arts<br />

students should be able to make use of computers. Although the English<br />

comm<strong>and</strong>s of FORTRAN <strong>and</strong> ALGOL were a tremendous improvement<br />

over the cryptic instructions of assembly language, they<br />

were both too complicated for beginners. Kemeny convinced Kurtz<br />

that they needed a completely new language, simple enough for beginners<br />

to learn quickly, yet flexible enough for many different<br />

kinds of applications.<br />

The language they developed was known as the “Beginner’s Allpurpose<br />

Symbolic Instruction Code,” or BASIC. The original language<br />

consisted of fourteen different statements. Each line of a<br />

BASIC program was preceded by a number. Line numbers were referenced<br />

by control flow statements, such as, “IFX=9THEN GOTO<br />

200.” Line numbers were also used as an editing reference. If line 30<br />

of a program contained an error, the programmer could make the<br />

necessary correction merely by retyping line 30.<br />

Programming in BASIC was first taught at Dartmouth in the fall<br />

of 1964. Students were ready to begin writing programs after two<br />

hours of classroom lectures. By June of 1968, more than 80 percent of<br />

the undergraduates at Dartmouth could write a BASIC program.<br />

Most of them were not science majors <strong>and</strong> used their programs in<br />

conjunction with other nontechnical courses.<br />

Kemeny <strong>and</strong> Kurtz, <strong>and</strong> later others under their supervision,<br />

wrote more powerful versions of BASIC that included support for<br />

graphics on video terminals <strong>and</strong> structured programming. The creators<br />

of BASIC, however, always tried to maintain their original design<br />

goal of keeping BASIC simple enough for beginners.


94 / BASIC programming language<br />

Consequences<br />

Kemeny <strong>and</strong> Kurtz encouraged the widespread adoption of BA-<br />

SIC by allowing other institutions to use their computer system <strong>and</strong><br />

by placing BASIC in the public domain. Over time, they shaped BA-<br />

SIC into a powerful language with numerous features added in response<br />

to the needs of its users. What Kemeny <strong>and</strong> Kurtz had not<br />

foreseen was the advent of the microprocessor chip in the early<br />

1970’s, which revolutionized computer technology. By 1975, microcomputer<br />

kits were being sold to hobbyists for well under a thous<strong>and</strong><br />

dollars. The earliest of these was the Altair.<br />

That same year, prelaw student William H. Gates (1955- ) was<br />

persuaded by a friend, Paul Allen, to drop out of Harvard University<br />

<strong>and</strong> help create a version of BASIC that would run on the Altair.<br />

Gates <strong>and</strong> Allen formed a company, Microsoft Corporation, to sell<br />

their BASIC interpreter, which was designed to fit into the tiny<br />

memory of the Altair. It was about as simple as the original Dartmouth<br />

BASIC but had to depend heavily on the computer hardware.<br />

Most computers purchased for home use still include a version<br />

of Microsoft Corporation’s BASIC.<br />

See also BINAC computer; COBOL computer language; FOR-<br />

TRAN programming language; SAINT; Supercomputer.<br />

Further Reading<br />

Kemeney, John G., <strong>and</strong> Thomas E. Kurtz. True BASIC: The Structured<br />

Language System for the Future. Reference Manual. West Lebanon,<br />

N.H.: True BASIC, 1988.<br />

Kurtz, Thomas E., <strong>and</strong> John G. Kemeney. BASIC. 5th ed. Hanover,<br />

N.H., 1970.<br />

Spencer, Donald D. Great Men <strong>and</strong> Women of Computing. 2d ed. Ormond<br />

Beach, Fla.: Camelot Publishing, 1999.


Bathyscaphe<br />

Bathyscaphe<br />

The invention: A submersible vessel capable of exploring the<br />

deepest trenches of the world’s oceans.<br />

The people behind the invention:<br />

William Beebe (1877-1962), an American biologist <strong>and</strong> explorer<br />

Auguste Piccard (1884-1962), a Swiss-born Belgian physicist<br />

Jacques Piccard (1922- ), a Swiss ocean engineer<br />

Early Exploration of the Deep Sea<br />

The first human penetration of the deep ocean was made by William<br />

Beebe in 1934, when he descended 923 meters into the Atlantic<br />

Ocean near Bermuda. His diving chamber was a 1.5-meter steel ball<br />

that he named Bathysphere, from the Greek word bathys (deep) <strong>and</strong><br />

the word sphere, for its shape. He found that a sphere resists pressure<br />

in all directions equally <strong>and</strong> is not easily crushed if it is constructed<br />

of thick steel. The bathysphere weighed 2.5 metric tons. It<br />

had no buoyancy <strong>and</strong> was lowered from a surface ship on a single<br />

2.2-centimeter cable; a broken cable would have meant certain<br />

death for the bathysphere’s passengers.<br />

Numerous deep dives by Beebe <strong>and</strong> his engineer colleague, Otis<br />

Barton, were the first uses of submersibles for science. Through two<br />

small viewing ports, they were able to observe <strong>and</strong> photograph<br />

many deep-sea creatures in their natural habitats for the first time.<br />

They also made valuable observations on the behavior of light as<br />

the submersible descended, noting that the green surface water became<br />

pale blue at 100 meters, dark blue at 200 meters, <strong>and</strong> nearly<br />

black at 300 meters. A technique called “contour diving” was particularly<br />

dangerous. In this practice, the bathysphere was slowly<br />

towed close to the seafloor. On one such dive, the bathysphere narrowly<br />

missed crashing into a coral crag, but the explorers learned a<br />

great deal about the submarine geology of Bermuda <strong>and</strong> the biology<br />

of a coral-reef community. Beebe wrote several popular <strong>and</strong> scientific<br />

books about his adventures that did much to arouse interest in<br />

the ocean.<br />

95


96 / Bathyscaphe<br />

Testing the Bathyscaphe<br />

The next important phase in the exploration of the deep ocean<br />

was led by the Swiss physicist Auguste Piccard. In 1948, he launched<br />

a new type of deep-sea research craft that did not require a cable <strong>and</strong><br />

that could return to the surface by means of its own buoyancy. He<br />

called the craft a bathyscaphe, which is Greek for “deep boat.”<br />

Piccard began work on the bathyscaphe in 1937, supported by a<br />

grant from the Belgian National Scientific Research Fund. The German<br />

occupation of Belgium early in World War II cut the project<br />

short, but Piccard continued his work after the war. The finished<br />

bathyscaphe was named FNRS 2, for the initials of the Belgian fund<br />

that had sponsored the project. The vessel was ready for testing in<br />

the fall of 1948.<br />

The first bathyscaphe, as well as later versions, consisted of<br />

two basic components: first, a heavy steel cabin to accommodate<br />

observers, which looked somewhat like an enlarged version of<br />

Beebe’s bathysphere; <strong>and</strong> second, a light container called a float,<br />

filled with gasoline, that provided lifting power because it was<br />

lighter than water. Enough iron shot was stored in silos to cause<br />

the vessel to descend. When this ballast was released, the gasoline<br />

in the float gave the bathyscaphe sufficient buoyancy to return to<br />

the surface.<br />

Piccard’s bathyscaphe had a number of ingenious devices. Jacques-<br />

Yves Cousteau, inventor of the Aqualung six years earlier, contributed<br />

a mechanical claw that was used to take samples of rocks, sediment,<br />

<strong>and</strong> bottom creatures. A seven-barreled harpoon gun, operated<br />

by water pressure, was attached to the sphere to capture<br />

specimens of giant squids or other large marine animals for study.<br />

The harpoons had electrical-shock heads to stun the “sea monsters,”<br />

<strong>and</strong> if that did not work, the harpoon could give a lethal injection of<br />

strychnine poison. Inside the sphere were various instruments for<br />

measuring the deep-sea environment, including a Geiger counter<br />

for monitoring cosmic rays. The air-purification system could support<br />

two people for up to twenty-four hours. The bathyscaphe had a<br />

radar mast to broadcast its location as soon as it surfaced. This was<br />

essential because there was no way for the crew to open the sphere<br />

from the inside.


Auguste Piccard<br />

Bathyscaphe / 97<br />

Auguste Piccard used balloons to set records in altitude both<br />

above sea level <strong>and</strong> below sea level. However, setting records<br />

was not his purpose: He went where no one had gone before for<br />

the sake of science.<br />

Born in Basel, Switzerl<strong>and</strong>, in 1884, Auguste <strong>and</strong><br />

his twin brother, Jean-Félix Piccard, studied in Zurich.<br />

After university in 1913, Auguste, a physicist,<br />

<strong>and</strong> Jean-Félix, a chemist, took up hot-air ballooning,<br />

<strong>and</strong> they joined the balloon section of the Swiss Army<br />

in 1915.<br />

Auguste moved to Brussels, Belgium, in 1922 to<br />

take a professorship of applied physics, <strong>and</strong> there he<br />

continued his ballooning. His subject of interest was<br />

cosmic rays, <strong>and</strong> in order to study them he had to get above the<br />

thick lower layer of atmosphere. Accordingly, he designed hydrogen-filled<br />

balloons that could reach high altitude. A ballshaped,<br />

pressurized gondola carried him, his instruments, <strong>and</strong><br />

one colleague to 51,775 feet altitude in 1931 <strong>and</strong> to 53,152 feet in<br />

1932. Both were records.<br />

Auguste, working with his son Jacques, then turned his attention<br />

to the sea. In order to explore the largely unknown world<br />

underwater, he built the bathyscaphe. It was really just another<br />

type of balloon, one which was made of steel <strong>and</strong> carried him<br />

inside. His dives with his son in various models of bathyscaphe<br />

set record after record. Their 1953 dive down 10,300 feet into the<br />

Mediterranean Sea was the deepest until Jacques, accompanied<br />

by a U.S. Navy officer, descended to the deepest spot on Earth<br />

seven years later.<br />

The FNRS 2 was first tested off the Cape Verde Isl<strong>and</strong>s with the<br />

assistance of the French navy. Although Piccard descended to only<br />

25 meters, the dive demonstrated the potential of the bathyscaphe.<br />

On the second dive, the vessel was severely damaged by waves, <strong>and</strong><br />

further tests were suspended. A redesigned <strong>and</strong> rebuilt bathyscaphe,<br />

renamed FNRS 3 <strong>and</strong> operated by the French navy, descended to a<br />

depth of 4,049 meters off Dakar, Senegal, on the west coast of Africa<br />

in early 1954.<br />

In August, 1953, Auguste Piccard, with his son Jacques, launched a<br />

(Library of Congress)


98 / Bathyscaphe<br />

greatly improved bathyscaphe, the Trieste, which they named for the<br />

Italian city in which it was built. In September of the same year, the<br />

Trieste successfully dived to 3,150 meters in the Mediterranean Sea. The<br />

Piccards glimpsed, for the first time, animals living on the seafloor at<br />

that depth. In 1958, the U.S. Navy purchased the Trieste <strong>and</strong> transported<br />

it to California, where it was equipped with a new cabin designed<br />

to enable the vessel to reach the seabed of the great oceanic<br />

trenches. Several successful descents were made in the Pacific by<br />

Jacques Piccard, <strong>and</strong> on January 23, 1960, Piccard, accompanied by<br />

Lieutenant Donald Walsh of the U.S. Navy, dived a record 10,916 meters<br />

to the bottom of the Mariana Trench near the isl<strong>and</strong> of Guam.<br />

Impact<br />

The oceans have always raised formidable barriers to humanity’s<br />

curiosity <strong>and</strong> underst<strong>and</strong>ing. In 1960, two events demonstrated the<br />

ability of humans to travel underwater for prolonged periods <strong>and</strong> to<br />

observe the extreme depths of the ocean. The nuclear submarine<br />

Triton circumnavigated the world while submerged, <strong>and</strong> Jacques<br />

Piccard <strong>and</strong> Lieutenant Donald Walsh descended nearly 11 kilometers<br />

to the bottom of the ocean’s greatest depression aboard the<br />

Trieste. After sinking for four hours <strong>and</strong> forty-eight minutes, the<br />

Trieste l<strong>and</strong>ed in the Challenger Deep of the Mariana Trench, the<br />

deepest known spot on the ocean floor. The explorers remained on<br />

the bottom for only twenty minutes, but they answered one of the<br />

biggest questions about the sea: Can animals live in the immense<br />

cold <strong>and</strong> pressure of the deep trenches? Observations of red shrimp<br />

<strong>and</strong> flatfishes proved that the answer was yes.<br />

The Trieste played another important role in undersea exploration<br />

when, in 1963, it located <strong>and</strong> photographed the wreckage of the<br />

nuclear submarine Thresher. The Thresher had mysteriously disappeared<br />

on a test dive off the New Engl<strong>and</strong> coast, <strong>and</strong> the Navy had<br />

been unable to find a trace of the lost submarine using surface vessels<br />

equipped with sonar <strong>and</strong> remote-control cameras on cables.<br />

Only the Trieste could actually search the bottom. On its third dive,<br />

the bathyscaphe found a piece of the wreckage, <strong>and</strong> it eventually<br />

photographed a 3,000-meter trail of debris that led to Thresher‘s hull,<br />

at a depth of 2.5 kilometers.


These exploits showed clearly that scientific submersibles could<br />

be used anywhere in the ocean. Piccard’s work thus opened the last<br />

geographic frontier on Earth.<br />

See also Aqualung; Bathysphere; Sonar; Ultrasound.<br />

Further Reading<br />

Bathyscaphe / 99<br />

Ballard, Robert D., <strong>and</strong> Will Hively. The Eternal Darkness: A Personal<br />

History of Deep-Sea Exploration. Princeton, N.J.: Princeton University<br />

Press, 2000.<br />

Piccard, Jacques, <strong>and</strong> Robert S. Dietz. Seven Miles Down: The Story of<br />

the Bathyscaphe Trieste. New York: Longmans, 1962.<br />

Welker, Robert Henry. Natural Man: The Life of William Beebe. Bloomington:<br />

Indiana University Press, 1975.


100<br />

Bathysphere<br />

Bathysphere<br />

The invention: The first successful chamber for manned deep-sea<br />

diving missions.<br />

The people behind the invention:<br />

William Beebe (1877-1962), an American naturalist <strong>and</strong> curator<br />

of ornithology<br />

Otis Barton (1899- ), an American engineer<br />

John Tee-Van (1897-1967), an American general associate with<br />

the New York Zoological Society<br />

Gloria Hollister Anable (1903?-1988), an American research<br />

associate with the New York Zoological Society<br />

Inner Space<br />

Until the 1930’s, the vast depths of the oceans had remained<br />

largely unexplored, although people did know something of the<br />

ocean’s depths. Soundings <strong>and</strong> nettings of the ocean bottom had<br />

been made many times by a number of expeditions since the 1870’s.<br />

Diving helmets had allowed humans to descend more than 91 meters<br />

below the surface, <strong>and</strong> the submarine allowed them to reach a<br />

depth of nearly 120 meters. There was no firsth<strong>and</strong> knowledge,<br />

however, of what it was like in the deepest reaches of the ocean: inner<br />

space.<br />

The person who gave the world the first account of life at great<br />

depths was William Beebe. When he announced in 1926 that he was<br />

attempting to build a craft to explore the ocean, he was already a<br />

well-known naturalist. Although his only degrees had been honorary<br />

doctorates, he was graduated as a special student in the Department<br />

of Zoology of Columbia University in 1898. He began his lifelong<br />

association with the New York Zoological Society in 1899.<br />

It was during a trip to the Galápagos Isl<strong>and</strong>s off the west coast of<br />

South America that Beebe turned his attention to oceanography. He<br />

became the first scientist to use a diving helmet in fieldwork, swimming<br />

in the shallow waters. He continued this shallow-water work<br />

at the new station he established in 1928, with the permission of En-


glish authorities, on the tiny isl<strong>and</strong> of Nonesuch in the Bermudas.<br />

Beebe realized, however, that he had reached the limits of the current<br />

technology <strong>and</strong> that to study the animal life of the ocean depths<br />

would require a new approach.<br />

A New Approach<br />

Bathysphere / 101<br />

While he was considering various cylindrical designs for a new<br />

deep-sea exploratory craft, Beebe was introduced to Otis Barton.<br />

Barton, a young New Engl<strong>and</strong>er who had been trained as an engineer<br />

at Harvard University, had turned to the problems of ocean<br />

diving while doing postgraduate work at Columbia University. In<br />

December, 1928, Barton brought his blueprints to Beebe. Beebe immediately<br />

saw that Barton’s design was what he was looking for,<br />

<strong>and</strong> the two went ahead with the construction of Barton’s craft.<br />

The “bathysphere,” as Beebe named the device, weighed 2,268<br />

kilograms <strong>and</strong> had a diameter of 1.45 meters <strong>and</strong> steel walls 3.8 centimeters<br />

thick. The door, weighing 180 kilograms, would be fastened<br />

over a manhole with ten bolts. Four windows, made of fused<br />

quartz, were ordered from the General Electric Company at a cost of<br />

$500 each. A 250-watt water spotlight lent by the Westinghouse<br />

Company provided the exterior illumination, <strong>and</strong> a telephone lent<br />

by the Bell Telephone Laboratory provided a means of communicating<br />

with the surface. The breathing apparatus consisted of two oxygen<br />

tanks that allowed 2 liters of oxygen per minute to escape into<br />

the sphere. During the dive, the carbon dioxide <strong>and</strong> moisture were<br />

removed, respectively, by trays containing soda lime <strong>and</strong> calcium<br />

chloride. A winch would lower the bathysphere on a steel cable.<br />

In early July, 1930, after several test dives, the first manned dive<br />

commenced. Beebe <strong>and</strong> Barton descended to a depth of 244 meters.<br />

A short circuit in one of the switches showered them with sparks<br />

momentarily, but the descent was largely a success. Beebe <strong>and</strong><br />

Barton had descended farther than any human.<br />

Two more days of diving yielded a final dive record of 435 meters<br />

below sea level. Beebe <strong>and</strong> the other members of his staff (ichthyologist<br />

John Tee-Van <strong>and</strong> zoologist Gloria Hollister Anable) saw many<br />

species of fish <strong>and</strong> other marine life that previously had been seen<br />

only after being caught in nets. These first dives proved that an un-


102 / Bathysphere<br />

dersea exploratory craft had potential value, at least for deep water.<br />

After 1932, the bathysphere went on display at the Century of Progress<br />

Exhibition in Chicago.<br />

In late 1933, the National Geographic Society offered to sponsor<br />

another series of dives. Although a new record was not a stipulation,<br />

Beebe was determined to supply one. The bathysphere was<br />

completely refitted before the new dives.<br />

An unmanned test dive to 920 meters was made on August 7,<br />

1934, once again off Nonesuch Isl<strong>and</strong>. Minor adjustments were<br />

made, <strong>and</strong> on the morning of August 11, the first dive commenced,<br />

attaining a depth of 765 meters <strong>and</strong> recording a number of new scientific<br />

observations. Several days later, on August 15, the weather<br />

was again right for the dive.<br />

This dive also paid rich dividends in the number of species of<br />

deep-sea life observed. Finally, with only a few turns of cable left on<br />

the winch spool, the bathysphere reached a record depth of 923 meters—almost<br />

a kilometer below the ocean’s surface.<br />

Impact<br />

Barton continued to work on the bathysphere design for some<br />

years. It was not until 1948, however, that his new design, the<br />

benthoscope, was finally constructed. It was similar in basic design<br />

to the bathysphere, though the walls were increased to withst<strong>and</strong><br />

greater pressures. Other improvements were made, but the essential<br />

strengths <strong>and</strong> weaknesses remained. On August 16, 1949, Barton,<br />

diving alone, broke the record he <strong>and</strong> Beebe had set earlier,<br />

reaching a depth of 1,372 meters off the coast of Southern California.<br />

The bathysphere effectively marked the end of the tethered exploration<br />

of the deep, but it pointed the way to other possibilities.<br />

The first advance in this area came in 1943, when undersea explorer<br />

Jacques-Yves Cousteau <strong>and</strong> engineer Émile Gagnan developed the<br />

Aqualung underwater breathing apparatus, which made possible<br />

unfettered <strong>and</strong> largely unencumbered exploration down to about<br />

60 meters. This was by no means deep diving, but it was clearly a<br />

step along the lines that Beebe had envisioned for underwater research.<br />

A further step came in the development of the bathyscaphe by


Auguste Piccard, the renowned Swiss physicist, who, in the 1930’s,<br />

had conquered the stratosphere in high-altitude balloons. The bathyscaphe<br />

was a balloon that operated in reverse. A spherical steel passenger<br />

cabin was attached beneath a large float filled with gasoline<br />

for buoyancy. Several tons of iron pellets held by electromagnets<br />

acted as ballast. The bathyscaphe would sink slowly to the bottom<br />

of the ocean, <strong>and</strong> when its passengers wished to return, the ballast<br />

would be dumped. The craft would then slowly rise to the surface.<br />

On September 30, 1953, Piccard touched bottom off the coast of Italy,<br />

some 3,000 meters below sea level.<br />

See also Aqualung; Bathyscaphe; Sonar; Ultrasound.<br />

Further Reading<br />

Bathysphere / 103<br />

Ballard, Robert D., <strong>and</strong> Will Hively. The Eternal Darkness: A Personal<br />

History of Deep-Sea Exploration. Princeton, N.J.: Princeton University<br />

Press, 2000.<br />

Forman, Will. The History of American Deep Submersible Operations,<br />

1775-1995. Flagstaff, Ariz.: Best, 1999.<br />

Welker, Robert Henry. Natural Man: The Life of William Beebe. Bloomington:<br />

Indiana University Press, 1975.


104<br />

BINAC computer<br />

BINAC computer<br />

The invention: The world’s first electronic general-purpose digital<br />

computer.<br />

The people behind the invention:<br />

John Presper Eckert (1919-1995), an American electrical engineer<br />

John W. Mauchly (1907-1980), an American physicist<br />

John von Neumann (1903-1957), a Hungarian American<br />

mathematician<br />

Alan Mathison Turing (1912-1954), an English mathematician<br />

Computer Evolution<br />

In the 1820’s, there was a need for error-free mathematical <strong>and</strong><br />

astronomical tables for use in navigation, unreliable versions of<br />

which were being produced by human “computers.” The problem<br />

moved English mathematician <strong>and</strong> inventor Charles Babbage to design<br />

<strong>and</strong> partially construct some of the earliest prototypes of modern<br />

computers, with substantial but inadequate funding from the<br />

British government. In the 1880’s, the search by the U.S. Bureau of<br />

the Census for a more efficient method of compiling the 1890 census<br />

led American inventor Herman Hollerith to devise a punched-card<br />

calculator, a machine that reduced by several years the time required<br />

to process the data.<br />

The emergence of modern electronic computers began during<br />

World War II (1939-1945), when there was an urgent need in the<br />

American military for reliable <strong>and</strong> quickly produced mathematical<br />

tables that could be used to aim various types of artillery. The calculation<br />

of very complex tables had progressed somewhat since<br />

Babbage’s day, <strong>and</strong> the human computers were being assisted by<br />

mechanical calculators. Still, the growing dem<strong>and</strong> for increased accuracy<br />

<strong>and</strong> efficiency was pushing the limits of these machines.<br />

Finally, in 1946, following three years of intense work at the University<br />

of Pennsylvania’s Moore School of Engineering, John Presper<br />

Eckert <strong>and</strong> John W. Mauchly presented their solution to the problems<br />

in the form of the Electronic Numerical Integrator <strong>and</strong> Calcula-


tor (ENIAC) the world’s first electronic general-purpose digital<br />

computer.<br />

The ENIAC, built under a contract with the Army’s Ballistic Research<br />

Laboratory, became a great success for Eckert <strong>and</strong> Mauchly,<br />

but even before it was completed, they were setting their sights on<br />

loftier targets. The primary drawback of the ENIAC was the great<br />

difficulty involved in programming it. Whenever the operators<br />

needed to instruct the machine to shift from one type of calculation<br />

to another, they had to reset a vast array of dials <strong>and</strong> switches, unplug<br />

<strong>and</strong> replug numerous cables, <strong>and</strong> make various other adjustments<br />

to the multiple pieces of hardware involved. Such a mode of<br />

operation was deemed acceptable for the ENIAC because, in computing<br />

firing tables, it would need reprogramming only occasionally.<br />

Yet if instructions could be stored in a machine’s memory, along<br />

with the data, such a machine would be able to h<strong>and</strong>le a wide range<br />

of calculations with ease <strong>and</strong> efficiency.<br />

The Turing Concept<br />

BINAC computer / 105<br />

The idea of a stored-program computer had first appeared in a<br />

paper published by English mathematician Alan Mathison Turing<br />

in 1937. In this paper, Turing described a hypothetical machine of<br />

quite simple design that could be used to solve a wide range of logical<br />

<strong>and</strong> mathematical problems. One significant aspect of this imaginary<br />

Turing machine was that the tape that would run through it<br />

would contain both information to be processed <strong>and</strong> instructions on<br />

how to process it. The tape would thus be a type of memory device,<br />

storing both the data <strong>and</strong> the program as sets of symbols that the<br />

machine could “read” <strong>and</strong> underst<strong>and</strong>. Turing never attempted to<br />

construct this machine, <strong>and</strong> it was not until 1946 that he developed a<br />

design for an electronic stored-program computer, a prototype of<br />

which was built in 1950.<br />

In the meantime, John von Neumann, a Hungarian American<br />

mathematician acquainted with Turing’s ideas, joined Eckert <strong>and</strong><br />

Mauchly in 1944 <strong>and</strong> contributed to the design of ENIAC’s successor,<br />

the Electronic Discrete Variable Automatic Computer (EDVAC), another<br />

project financed by the Army. The EDVAC was the first computer<br />

designed to incorporate the concept of the stored program.


106 / BINAC computer<br />

In March of 1946, Eckert <strong>and</strong> Mauchly, frustrated by a controversy<br />

over patent rights for the ENIAC, resigned from the<br />

Moore School. Several months later, they formed the Philadelphiabased<br />

Electronic Control Company on the strength of a contract<br />

from the National Bureau of St<strong>and</strong>ards <strong>and</strong> the Census Bureau to<br />

build a much gr<strong>and</strong>er computer, the Universal Automatic Computer<br />

(UNIVAC). They thus ab<strong>and</strong>oned the EDVAC project, which<br />

was finally completed by the Moore School in 1952, but they incorporated<br />

the main features of the EDVAC into the design of the<br />

UNIVAC.<br />

Building the UNIVAC, however, proved to be much more involved<br />

<strong>and</strong> expensive than anticipated, <strong>and</strong> the funds provided by<br />

the original contract were inadequate. Eckert <strong>and</strong> Mauchly, therefore,<br />

took on several other smaller projects in an effort to raise<br />

funds. On October 9, 1947, they signed a contract with the Northrop<br />

Corporation of Hawthorne, California, to produce a relatively small<br />

computer to be used in the guidance system of a top-secret missile<br />

called the Snark, which Northrop was building for the Air Force.<br />

This computer, the Binary Automatic Computer (BINAC), turned<br />

out to be Eckert <strong>and</strong> Mauchly’s first commercial sale <strong>and</strong> the first<br />

stored-program computer completed in the United States.<br />

The BINAC was designed to be at least a preliminary version of a<br />

compact, airborne computer. It had two main processing units.<br />

These contained a total of fourteen hundred vacuum tubes, a drastic<br />

reduction from the eighteen thous<strong>and</strong> used in the ENIAC. There<br />

were also two memory units, as well as two power supplies, an input<br />

converter unit, <strong>and</strong> an input console, which used either a typewriter<br />

keyboard or an encoded magnetic tape (the first time such<br />

tape was used for computer input). Because of its dual processing,<br />

memory, <strong>and</strong> power units, the BINAC was actually two computers,<br />

each of which would continually check its results against those of<br />

the other in an effort to identify errors.<br />

The BINAC became operational in August, 1949. <strong>Public</strong> demonstrations<br />

of the computer were held in Philadelphia from August 18<br />

through August 20.


Impact<br />

The design embodied in the BINAC is the real source of its significance.<br />

It demonstrated successfully the benefits of the dual processor<br />

design for minimizing errors, a feature adopted in many subsequent<br />

computers. It showed the suitability of magnetic tape as an<br />

input-output medium. Its most important new feature was its ability<br />

to store programs in its relatively spacious memory, the principle<br />

that Eckert, Mauchly, <strong>and</strong> von Neumann had originally designed<br />

into the EDVAC. In this respect, the BINAC was a direct descendant<br />

of the EDVAC.<br />

In addition, the stored-program principle gave electronic computers<br />

new powers, quickness, <strong>and</strong> automatic control that, as they<br />

have continued to grow, have contributed immensely to the aura of<br />

intelligence often associated with their operation.<br />

The BINAC successfully demonstrated some of these impressive<br />

new powers in August of 1949 to eager observers from a number of<br />

major American corporations. It helped to convince many influential<br />

leaders of the commercial segment of society of the promise of<br />

electronic computers. In doing so, the BINAC helped to ensure the<br />

further evolution of computers.<br />

See also Apple II computer; BINAC computer; Colossus computer;<br />

ENIAC computer; IBM Model 1401 computer; Personal computer;<br />

Supercomputer; UNIVAC computer.<br />

Further Reading<br />

BINAC computer / 107<br />

Macrae, Norman. John von Neumann: The Scientific Genius Who Pioneered<br />

the Modern Computer, Game Theory, Nuclear Deterrence, <strong>and</strong><br />

Much More. New York: Pantheon Books, 1992.<br />

Spencer, Donald D. Great Men <strong>and</strong> Women of Computing. 2d ed. Ormond<br />

Beach, Fla.: Camelot Publishing, 1999.<br />

Zientara, Marguerite. The History of Computing: A Biographical Portrait<br />

of the Visionaries Who Shaped the Destiny of the Computer Industry.<br />

Framingham, Mass.: CW Communications, 1981.


108<br />

Birth control pill<br />

Birth control pill<br />

The invention: An orally administered drug that inhibits ovulation<br />

in women, thereby greatly reducing the chance of pregnancy.<br />

The people behind the invention:<br />

Gregory Pincus (1903-1967), an American biologist<br />

Min-Chueh Chang (1908-1991), a Chinese-born reproductive<br />

biologist<br />

John Rock (1890-1984), an American gynecologist<br />

Celso-Ramon Garcia (1921- ), a physician<br />

Edris Rice-Wray (1904- ), a physician<br />

Katherine Dexter McCormick (1875-1967), an American<br />

millionaire<br />

Margaret Sanger (1879-1966), an American activist<br />

An Ardent Crusader<br />

Margaret Sanger was an ardent crusader for birth control <strong>and</strong><br />

family planning. Having decided that a foolproof contraceptive was<br />

necessary, Sanger met with her friend, the wealthy socialite Katherine<br />

Dexter McCormick. A 1904 graduate in biology from the Massachusetts<br />

Institute of Technology, McCormick had the knowledge<br />

<strong>and</strong> the vision to invest in biological research. Sanger arranged a<br />

meeting between McCormick <strong>and</strong> Gregory Pincus, head of the<br />

Worcester Institutes of Experimental Biology. After listening to Sanger’s<br />

pleas for an effective contraceptive <strong>and</strong> McCormick’s offer of financial<br />

backing, Pincus agreed to focus his energies on finding a pill<br />

that would prevent pregnancy.<br />

Pincus organized a team to conduct research on both laboratory<br />

animals <strong>and</strong> humans. The laboratory studies were conducted under<br />

the direction of Min-Chueh Chang, a Chinese-born scientist who<br />

had been studying sperm biology, artificial insemination, <strong>and</strong> in vitro<br />

fertilization. The goal of his research was to see whether pregnancy<br />

might be prevented by manipulation of the hormones usually<br />

found in a woman.


It was already known that there was one time when a woman<br />

could not become pregnant—when she was already pregnant. In<br />

1921, Ludwig Haberl<strong>and</strong>t, an Austrian physiologist, had transplanted<br />

the ovaries from a pregnant rabbit into a nonpregnant one.<br />

The latter failed to produce ripe eggs, showing that some substance<br />

from the ovaries of a pregnant female prevents ovulation. This substance<br />

was later identified as the hormone progesterone by George<br />

W. Corner, Jr., <strong>and</strong> Willard M. Allen in 1928.<br />

If progesterone could inhibit ovulation during pregnancy, maybe<br />

progesterone treatment could prevent ovulation in nonpregnant females<br />

as well. In 1937, this was shown to be the case by scientists<br />

from the University of Pennsylvania, who prevented ovulation in<br />

rabbits with injections of progesterone. It was not until 1951, however,<br />

when Carl Djerassi <strong>and</strong> other chemists devised inexpensive<br />

ways of producing progesterone in the laboratory, that serious consideration<br />

was given to the medical use of progesterone. The synthetic<br />

version of progesterone was called “progestin.”<br />

Testing the Pill<br />

Birth control pill / 109<br />

In the laboratory, Chang tried more than two hundred different<br />

progesterone <strong>and</strong> progestin compounds, searching for one that<br />

would inhibit ovulation in rabbits <strong>and</strong> rats. Finally, two compounds<br />

were chosen: progestins derived from the root of a wild Mexican<br />

yam. Pincus arranged for clinical tests to be carried out by Celso-<br />

Ramon Garcia, a physician, <strong>and</strong> John Rock, a gynecologist.<br />

Rock had already been conducting experiments with progesterone<br />

as a treatment for infertility. The treatment was effective in some<br />

women but required that large doses of expensive progesterone be<br />

injected daily. Rock was hopeful that the synthetic progestin that<br />

Chang had found effective in animals would be helpful in infertile<br />

women as well. With Garcia <strong>and</strong> Pincus, Rock treated another<br />

group of fifty infertile women with the synthetic progestin. After<br />

treatment ended, seven of these previously infertile women became<br />

pregnant within half a year. Garcia, Pincus, <strong>and</strong> Rock also took several<br />

physiological measurements of the women while they were<br />

taking the progestin <strong>and</strong> were able to conclude that ovulation did<br />

not occur while the women were taking the progestin pill.


(Library of Congess)<br />

110 / Birth control pill<br />

Margaret Sanger<br />

Margaret Louise Higgins saw her mother die at the age of<br />

only fifty. The cause was tuberculosis, but Margaret, the sixth of<br />

eleven children, was convinced her mother’s string of pregnancies<br />

was what killed her. Her crusade to liberate women from<br />

the burden of unwanted, dangerous pregnancies lasted the rest<br />

of her life.<br />

Born in Corning, New York, in 1879, she went to Claverack<br />

College <strong>and</strong> Hudson River Institute <strong>and</strong> joined a nursing program<br />

at White Plains Hospital, graduating in 1900.<br />

Two years later she married William Sanger, an architect<br />

<strong>and</strong> painter. They moved into New York City in<br />

1910 <strong>and</strong> became part of Greenwich Village’s community<br />

of left-wing intellectuals, artists, <strong>and</strong> activists,<br />

such as John Reed, Upton Sinclair, <strong>and</strong> Emma<br />

Goldman. She used her free time to support liberal<br />

reform causes, participating in labor actions of the Industrial<br />

Workers of the World. Working as a visiting<br />

nurse, she witnessed the health problems among poor<br />

women caused by poor hygiene <strong>and</strong> frequent pregnancies.<br />

In 1912 she test this began a newspaper column, “What<br />

Every Girl Should Know,” about reproductive health <strong>and</strong> education.<br />

The authorities tried to suppress some of the columns as<br />

obscene—for instance, one explaining venereal disease—but<br />

Sanger was undaunted. In 1914, she launched The Woman Rebel,<br />

a magazine promoting women’s liberation <strong>and</strong> birth control.<br />

From then on, although threatened with legal action <strong>and</strong> jail,<br />

she vigorously fought the political battles for birth control She<br />

published books, lectured, took part in demonstrations, opened<br />

a birth control clinic in Brooklyn (the nation’s first), started the<br />

Birth Control Federation of American (later renamed Planned<br />

Parenthood Federation of America), <strong>and</strong> traveled overseas to<br />

promote birth control in order to improve the st<strong>and</strong>ard of living<br />

in Third World countries <strong>and</strong> to curb population growth.<br />

Sanger was not an inventor, but she contributed ideas to the<br />

invention of various birth control devices <strong>and</strong> in the 1950’s<br />

found the money needed for the research <strong>and</strong> development of<br />

oral contraceptives at the Worcester Foundation for Experimental<br />

Biology, which produced the first birth control pill. She died<br />

in Tucson, Arizona, in 1966.


Having shown that the hormone could effectively prevent ovulation<br />

in both animals <strong>and</strong> humans, the investigators turned their attention<br />

back to birth control. They were faced with several problems:<br />

whether side effects might occur in women using progestins for a<br />

long time, <strong>and</strong> whether women would remember to take the pill day<br />

after day, for months or even years. To solve these problems, the birth<br />

control pill was tested on a large scale. Because of legal problems in<br />

the United States, Pincus decided to conduct the test in Puerto Rico.<br />

The test started in April of 1956. Edris Rice-Wray, a physician,<br />

was responsible for the day-to-day management of the project. As<br />

director of the Puerto Rico Family Planning Association, she had<br />

seen firsth<strong>and</strong> the need for a cheap, reliable contraceptive. The<br />

women she recruited for the study were married women from a<br />

low-income population living in a housing development in Río<br />

Piedras, a suburb of San Juan. Word spread quickly, <strong>and</strong> soon<br />

women were volunteering to take the pill that would prevent pregnancy.<br />

In the first study, 221 women took a pill containing 10 milligrams<br />

of progestin <strong>and</strong> 0.15 milligrams of estrogen. (The estrogen<br />

was added to help control breakthrough bleeding.)<br />

Results of the test were reported in 1957. Overall, the pill proved<br />

highly effective in preventing conception. None of the women<br />

who took the pill according to directions became pregnant, <strong>and</strong><br />

most women who wanted to get pregnant after stopping the pill<br />

had no difficulty. Nevertheless, 17 percent of the women had some<br />

unpleasant reactions, such as nausea or dizziness. The scientists<br />

believed that these mild side effects, as well as one death from congestive<br />

heart failure, were unrelated to the use of the pill.<br />

Even before the final results were announced, additional field<br />

tests were begun. In 1960, the U.S. Food <strong>and</strong> Drug Administration<br />

(FDA) approved the use of the pill developed by Pincus <strong>and</strong> his collaborators<br />

as an oral contraceptive.<br />

Consequences<br />

Birth control pill / 111<br />

Within two years of approval by the FDA, more than a million<br />

women in the United States were using the birth control pill. New<br />

contraceptives were developed in the 1960’s <strong>and</strong> 1970’s, but the<br />

birth control pill remains the most widely used method of prevent-


112 / Birth control pill<br />

ing pregnancy. More than 60<br />

million women use the pill<br />

worldwide.<br />

The greatest impact of the<br />

pill has been in the social <strong>and</strong><br />

political world. Before Sanger<br />

began the push for the pill,<br />

birth control was regarded often<br />

as socially immoral <strong>and</strong><br />

often illegal as well. Women<br />

in those post-World War II<br />

years were expected to have<br />

a lifelong career as a mother<br />

to their many children.<br />

With the advent of the pill,<br />

a radical change occurred<br />

in society’s attitude toward<br />

women’s work. Women had in-<br />

creased freedom to work <strong>and</strong> enter careers previously closed to them<br />

because of fears that they might get pregnant. Women could control<br />

more precisely when they would get pregnant <strong>and</strong> how many children<br />

they would have. The women’s movement of the 1960’s—with its<br />

change to more liberal social <strong>and</strong> sexual values—gained much of its<br />

strength from the success of the birth control pill.<br />

See also Abortion pill; Amniocentesis; Artificial hormone; Genetically<br />

engineered insulin; Mammography; Syphilis test; Ultrasound.<br />

Further Reading<br />

Dispensers designed to help users keep track of<br />

the days on which they take their pills. (Image<br />

Club Graphics)<br />

DeJauregui, Ruth. One Hundred Medical Milestones That Shaped World<br />

History. San Mateo, Calif.: Bluewood Books, 1998.<br />

Tone, Andrea. Devices <strong>and</strong> Desires: A History of Contraceptives in America.<br />

New York: Hill <strong>and</strong> Wang, 2001.<br />

Watkins, Elizabeth Siegel. On the Pill: A Social History of Oral Contraceptives,<br />

1950-1970. Baltimore: Johns Hopkins University Press, 1998.


Blood transfusion<br />

Blood transfusion<br />

The invention: A technique that greatly enhanced surgery patients’<br />

chances of survival by replenishing the blood they lose in<br />

surgery with a fresh supply.<br />

The people behind the invention:<br />

Charles Drew (1904-1950), American pioneer in blood<br />

transfusion techniques<br />

George Washington Crile (1864-1943), an American surgeon,<br />

author, <strong>and</strong> brigadier general in the U.S. Army Medical<br />

Officers’ Reserve Corps<br />

Alexis Carrel (1873-1944), a French surgeon<br />

Samuel Jason Mixter (1855-1923), an American surgeon<br />

Nourishing Blood Transfusions<br />

113<br />

It is impossible to say when <strong>and</strong> where the idea of blood transfusion<br />

first originated, although descriptions of this procedure are<br />

found in ancient Egyptian <strong>and</strong> Greek writings. The earliest documented<br />

case of a blood transfusion is that of Pope Innocent VII. In<br />

April, 1492, the pope, who was gravely ill, was transfused with the<br />

blood of three young boys. As a result, all three boys died without<br />

bringing any relief to the pope.<br />

In the centuries that followed, there were occasional descriptions<br />

of blood transfusions, but it was not until the middle of the seventeenth<br />

century that the technique gained popularity following the<br />

English physician <strong>and</strong> anatomist William Harvey’s discovery of the<br />

circulation of the blood in 1628. In the medical thought of those<br />

times, blood transfusion was considered to have a nourishing effect<br />

on the recipient. In many of those experiments, the human recipient<br />

received animal blood, usually from a lamb or a calf. Blood transfusion<br />

was tried as a cure for many different diseases, mainly those<br />

that caused hemorrhages, as well as for other medical problems <strong>and</strong><br />

even for marital problems.<br />

Blood transfusions were a dangerous procedure, causing many<br />

deaths of both donor <strong>and</strong> recipient as a result of excessive blood


114 / Blood transfusion<br />

loss, infection, passage of blood clots into the circulatory systems of<br />

the recipients, passage of air into the blood vessels (air embolism),<br />

<strong>and</strong> transfusion reaction as a result of incompatible blood types. In<br />

the mid-nineteenth century, blood transfusions from animals to humans<br />

stopped after it was discovered that the serum of one species<br />

agglutinates <strong>and</strong> dissolves the blood cells of other species. A sharp<br />

drop in the use of blood transfusion came with the introduction of<br />

physiologic salt solution in 1875. Infusion of salt solution was simple<br />

<strong>and</strong> was safer than blood transfusion.<br />

Direct-Connection Blood Transfusions<br />

In 1898, when George Washington Crile began his work on blood<br />

transfusions, the major obstacle he faced was solving the problem of<br />

blood clotting during transfusions. He realized that salt solutions<br />

were not helpful in severe cases of blood loss, when there is a need to<br />

restore the patient to consciousness, steady the heart action, <strong>and</strong> raise<br />

the blood pressure. At that time, he was experimenting with indirect<br />

blood transfusions by drawing the blood of the donor into a vessel,<br />

then transferring it into the recipient’s vein by tube, funnel, <strong>and</strong> cannula,<br />

the same technique used in the infusion of saline solution.<br />

The solution to the problem of blood clotting came in 1902 when<br />

Alexis Carrel developed the technique of surgically joining blood<br />

vessels without exposing the blood to air or germs, either of which<br />

can lead to clotting. Crile learned this technique from Carrel <strong>and</strong><br />

used it to join the peripheral artery in the donor to a peripheral vein<br />

of the recipient. Since the transfused blood remained sealed in the<br />

inner lining of the vessels, blood clotting did not occur.<br />

The first human blood transfusion of this type was performed by<br />

Crile in December, 1905. The patient, a thirty-five-year-old woman,<br />

was transfused by her husb<strong>and</strong> but died a few hours after the procedure.<br />

The second, but first successful, transfusion was performed on<br />

August 8, 1906. The patient, a twenty-three-year-old male, suffered<br />

from severe hemorrhaging following surgery to remove kidney<br />

stones. After all attempts to stop the bleeding were exhausted with<br />

no results, <strong>and</strong> the patient was dangerously weak, transfusion was<br />

considered as a last resort. One of the patient’s brothers was the do-


Charles Drew<br />

Blood transfusion / 115<br />

While he was still in medical school, Charles Richard Drew<br />

saw a man’s life saved with a blood transfusion. He also saw patients<br />

die because suitable donors could not be found. Impressed<br />

by both the life-saving power of transfusions <strong>and</strong> the<br />

dire need for more of them, Drew devoted his career to improving<br />

the nation’s blood supply. His inventions saved untold<br />

thous<strong>and</strong>s of lives, especially during World War II, before artificial<br />

blood was developed.<br />

Born in 1904 in Washington, D.C., Drew was a star athlete in<br />

high school, in Amherst College—from which he graduated in<br />

1926—<strong>and</strong> even in medical school at McGill University<br />

in Montreal from 1928 to 1933. He returned to the<br />

U.S. capital to become a resident in Freedmen’s Hospital<br />

of Howard University. While there he invented a<br />

method for separating plasma from whole blood <strong>and</strong><br />

discovered that it was not necessary to recombine the<br />

plasma <strong>and</strong> red blood cells for transfusion. Plasma<br />

alone was sufficient, <strong>and</strong> by drying or <strong>and</strong> freezing it,<br />

the plasma remained fresh enough over long periods<br />

to act as an emergency reserve. In 1938 Drew took a<br />

fellowship in blood research at Columbia Presbyterian Hospital<br />

in New York City. Employing his plasma preservation methods,<br />

he opened the first blood bank <strong>and</strong> wrote a dissertation on his<br />

techniques. He became the first African American to earn a<br />

Doctor of Science degree from Columbia University in 1940.<br />

He organized another blood bank, this one in Great Britain,<br />

<strong>and</strong> in 1941 was appointed director of the American Red Cross<br />

blood donor project. However, Drew learned to his disgust that<br />

the Red Cross <strong>and</strong> U.S. government would not allow blood<br />

from African Americans <strong>and</strong> Caucasians to be mixed in the<br />

blood bank. There was no scientific reason for such segregation.<br />

Bias prevailed. Drew angrily denounced the policy at a press<br />

conference <strong>and</strong> resigned from the Red Cross.<br />

He went back to Howard University as head of surgery <strong>and</strong>,<br />

later, director of Freedmen’s Hospital. Drew died in 1950 following<br />

an automobile accident.<br />

nor. Following the transfusion, the patient showed remarkable recovery<br />

<strong>and</strong> was strong enough to withst<strong>and</strong> surgery to remove the<br />

kidney <strong>and</strong> stop the bleeding. When his condition deteriorated a<br />

(Associated Publishers)


116 / Blood transfusion<br />

few days later, another transfusion was done. This time, too, he<br />

showed remarkable improvement, which continued until his complete<br />

recovery.<br />

For his first transfusions, Crile used the Carrel suture method,<br />

which required using very fine needles <strong>and</strong> thread. It was a very<br />

delicate <strong>and</strong> time-consuming procedure. At the suggestion of Samuel<br />

Jason Mixter, Crile developed a new method using a short tubal<br />

device with an attached h<strong>and</strong>le to connect the blood vessels. By this<br />

method, 3 or 4 centimeters of the vessels to be connected were surgically<br />

exposed, clamped, <strong>and</strong> cut, just as under the previous method.<br />

Yet, instead of suturing of the blood vessels, the recipient’s vein was<br />

passed through the tube <strong>and</strong> then cuffed back over the tube <strong>and</strong> tied<br />

to it. Then the donor’s artery was slipped over the cuff. The clamps<br />

were opened, <strong>and</strong> blood was allowed to flow from the donor to the<br />

recipient. In order to accommodate different-sized blood vessels,<br />

tubes of four different sizes were made, ranging in diameter from<br />

1.5 to 3 millimeters.<br />

Impact<br />

Crile’s method was the preferred method of blood transfusion<br />

for a number of years. Following the publication of his book on<br />

transfusion, a number of modifications to the original method were<br />

published in medical journals. In 1913, Edward Lindeman developed<br />

a method of transfusing blood simply by inserting a needle<br />

through the patient’s skin <strong>and</strong> into a surface vein, making it for the<br />

first time a nonsurgical method. This method allowed one to measure<br />

the exact quantity of blood transfused. It also allowed the donor<br />

to serve in multiple transfusions. This development opened the<br />

field of transfusions to all physicians. Lindeman’s needle <strong>and</strong> syringe<br />

method also eliminated another major drawback of direct<br />

blood transfusion: the need to have both donor <strong>and</strong> recipient right<br />

next to each other.<br />

See also Coronary artery bypass surgery; Electrocardiogram;<br />

Electroencephalogram; Heart-lung machine.


Further Reading<br />

Blood transfusion / 117<br />

English, Peter C. Shock, Physiological Surgery, <strong>and</strong> George Washington<br />

Crile: Medical Innovation in the Progressive Era. Westport, Conn.:<br />

Greenwood Press, 1980.<br />

Le Vay, David, <strong>and</strong> Roy Porter. Alexis Carrel: The Perfectibility of Man.<br />

Rockville, Md.: Kabel Publishers, 1996.<br />

Malinin, Theodore I. Surgery <strong>and</strong> Life: The Extraordinary Career of<br />

Alexis Carrel. New York: Harcourt Brace Jovanovich, 1979.<br />

May, Angelo M., <strong>and</strong> Alice G. May. The Two Lions of Lyons: The Tale of<br />

Two Surgeons, Alexis Carrel <strong>and</strong> René Leriche. Rockville, Md.: Kabel<br />

Publishers, 1992.


118<br />

Breeder reactor<br />

Breeder reactor<br />

The invention: A plant that generates electricity from nuclear fission<br />

while creating new fuel.<br />

The person behind the invention:<br />

Walter Henry Zinn (1906-2000), the first director of the Argonne<br />

National Laboratory<br />

Producing Electricity with More Fuel<br />

The discovery of nuclear fission involved both the discovery that<br />

the nucleus of a uranium atom would split into two lighter elements<br />

when struck by a neutron <strong>and</strong> the observation that additional neutrons,<br />

along with a significant amount of energy, were released at<br />

the same time. These neutrons might strike other atoms <strong>and</strong> cause<br />

them to fission (split) also. That, in turn, would release more energy<br />

<strong>and</strong> more neutrons, triggering a chain reaction as the process continued<br />

to repeat itself, yielding a continuing supply of heat.<br />

Besides the possibility that an explosive weapon could be constructed,<br />

early speculation about nuclear fission included its use in<br />

the generation of electricity. The occurrence of World War II (1939-<br />

1945) meant that the explosive weapon would be developed first.<br />

Both the weapons technology <strong>and</strong> the basic physics for the electrical<br />

reactor had their beginnings in Chicago with the world’s first nuclear<br />

chain reaction. The first self-sustaining nuclear chain reaction occurred<br />

in a laboratory at the University of Chicago on December 2, 1942.<br />

It also became apparent at that time that there was more than one<br />

way to build a bomb. At this point, two paths were taken: One was<br />

to build an atomic bomb with enough fissionable uranium in it to<br />

explode when detonated, <strong>and</strong> another was to generate fissionable<br />

plutonium <strong>and</strong> build a bomb. Energy was released in both methods,<br />

but the second method also produced another fissionable substance.<br />

The observation that plutonium <strong>and</strong> energy could be produced together<br />

meant that it would be possible to design electric power systems<br />

that would produce fissionable plutonium in quantities as large<br />

as, or larger than, the amount of fissionable material consumed. This


is the breeder concept, the idea that while using up fissionable uranium<br />

235, another fissionable element can be made. The full development<br />

of this concept for electric power was delayed until the end of<br />

World War II.<br />

Electricity from Atomic Energy<br />

Breeder reactor / 119<br />

On August 1, 1946, the Atomic Energy Commission (AEC) was<br />

established to control the development <strong>and</strong> explore the peaceful<br />

uses of nuclear energy. The Argonne National Laboratory was assigned<br />

the major responsibilities for pioneering breeder reactor<br />

technologies. Walter Henry Zinn was the laboratory’s first director.<br />

He led a team that planned a modest facility (Experimental Breeder<br />

Reactor I, or EBR-I) for testing the validity of the breeding principle.<br />

Planning for this had begun in late 1944 <strong>and</strong> grew as a natural extension<br />

of the physics that developed the plutonium atomic bomb.<br />

The conceptual design details for a breeder-electric reactor were<br />

reasonably complete by late 1945. On March 1, 1949, the AEC announced<br />

the selection of a site in Idaho for the National Reactor Station<br />

(later to be named the Idaho National Engineering Laboratory,<br />

or INEL). Construction at the INEL site in Arco, Idaho, began in October,<br />

1949. Critical mass was reached in August, 1951. (“Critical<br />

mass” is the amount <strong>and</strong> concentration of fissionable material required<br />

to produce a self-sustaining chain reaction.)<br />

The system was brought to full operating power, 1.1 megawatts<br />

of thermal power, on December 19, 1951. The next day, December<br />

20, at 11:00 a.m., steam was directed to a turbine generator. At 1:23<br />

p.m., the generator was connected to the electrical grid at the site,<br />

<strong>and</strong> “electricity flowed from atomic energy,” in the words of Zinn’s<br />

console log of that day. Approximately 200 kilowatts of electric<br />

power were generated most of the time that the reactor was run.<br />

This was enough to satisfy the needs of the EBR-I facilities. The reactor<br />

was shut down in 1964 after five years of use primarily as a test<br />

facility. It had also produced the first pure plutonium.<br />

With the first fuel loading, a conversion ratio of 1.01 was achieved,<br />

meaning that more new fuel was generated than was consumed by<br />

about 1 percent. When later fuel loadings were made with plutonium,<br />

the conversion ratios were more favorable, reaching as high


120 / Breeder reactor<br />

as 1.27. EBR-I was the first reactor to generate its own fuel <strong>and</strong> the<br />

first power reactor to use plutonium for fuel.<br />

The use of EBR-I also included pioneering work on fuel recovery<br />

<strong>and</strong> reprocessing. During its five-year lifetime, EBR-I operated with<br />

four different fuel loadings, each designed to establish specific<br />

benchmarks of breeder technology. This reactor was seen as the first<br />

in a series of increasingly large reactors in a program designed to<br />

develop breeder technology. The reactor was replaced by EBR-II,<br />

which had been proposed in 1953 <strong>and</strong> was constructed from 1955 to<br />

1964. EBR-II was capable of producing 20 megawatts of electrical<br />

power. It was approximately fifty times more powerful than EBR-I<br />

but still small compared to light-water commercial reactors of 600 to<br />

1,100 megawatts in use toward the end of the twentieth century.<br />

Consequences<br />

The potential for peaceful uses of nuclear fission were dramatized<br />

with the start-up of EBR-I in 1951: It was the first in the world<br />

to produce electricity, while also being the pioneer in a breeder reactor<br />

program. The breeder program was not the only reactor program<br />

being developed, however, <strong>and</strong> it eventually gave way to the<br />

light-water reactor design for use in the United States. Still, if energy<br />

resources fall into short supply, it is likely that the technologies first<br />

developed with EBR-I will find new importance. In France <strong>and</strong> Japan,<br />

commercial reactors make use of breeder reactor technology;<br />

these reactors require extensive fuel reprocessing.<br />

Following the completion of tests with plutonium loading in 1964,<br />

EBR-I was shut down <strong>and</strong> placed in st<strong>and</strong>by status. In 1966, it was declared<br />

a national historical l<strong>and</strong>mark under the stewardship of the<br />

U.S. Department of the Interior. The facility was opened to the public<br />

in June, 1975.<br />

See also Atomic bomb; Geothermal power; Nuclear power<br />

plant; Nuclear reactor; Solar thermal engine; Tidal power plant.


Further Reading<br />

Breeder reactor / 121<br />

“Breeder Trouble.” Technology Review 91, no. 5 (July, 1988).<br />

Hippel, Frank von, <strong>and</strong> Suzanne Jones. “Birth of the Breeder.” Bulletin<br />

of the Atomic Scientists 53, no. 5 (September/October, 1997).<br />

Krieger, David. Splitting the Atom: A Chronology of the Nuclear Age.<br />

Santa Barbara, Calif.: Nuclear Age Peace foundation, 1998.


122<br />

Broadcaster guitar<br />

Broadcaster guitar<br />

The invention: The first commercially manufactured solid-body<br />

electric guitar, the Broadcaster revolutionized the guitar industry<br />

<strong>and</strong> changed the face of popular music<br />

The people behind the invention:<br />

Leo Fender (1909-1991), designer of affordable <strong>and</strong> easily massproduced<br />

solid-body electric guitars<br />

Les Paul (Lester William Polfuss, 1915- ), a legendary<br />

guitarist <strong>and</strong> designer of solid-body electric guitars<br />

Charlie Christian (1919-1942), an influential electric jazz<br />

guitarist of the 1930’s<br />

Early Electric Guitars<br />

It has been estimated that between 1931 <strong>and</strong> 1937, approximately<br />

twenty-seven hundred electric guitars <strong>and</strong> amplifiers were sold in<br />

the United States. The Electro String Instrument Company, run<br />

by Adolph Rickenbacker <strong>and</strong> his designer partners, George Beauchamp<br />

<strong>and</strong> Paul Barth, produced two of the first commercially manufactured<br />

electric guitars—the Rickenbacker A-22 <strong>and</strong> A-25—in<br />

1931. The Rickenbacker models were what are known as “lap steel”<br />

or Hawaiian guitars. A Hawaiian guitar is played with the instrument<br />

lying flat across a guitarist’s knees. By the mid-1930’s, the Gibson<br />

company had introduced an electric Spanish guitar, the ES-150.<br />

Legendary jazz guitarist Charlie Christian made this model famous<br />

while playing for Benny Goodman’s orchestra. Christian was the<br />

first electric guitarist to be heard by a large American audience.<br />

He became an inspiration for future electric guitarists, because he<br />

proved that the electric guitar could have its own unique solo<br />

sound. Along with Christian, the other electric guitar figures who<br />

put the instrument on the musical map were blues guitarist T-Bone<br />

Walker, guitarist <strong>and</strong> inventor Les Paul, <strong>and</strong> engineer <strong>and</strong> inventor<br />

Leo Fender.<br />

Early electric guitars were really no more than acoustic guitars,<br />

with the addition of one or more pickups, which convert string vi-


ations to electrical signals that can be played through a speaker.<br />

Amplification of a guitar made it a more assertive musical instrument.<br />

The electrification of the guitar ultimately would make it<br />

more flexible, giving it a more prominent role in popular music. Les<br />

Paul, always a compulsive inventor, began experimenting with<br />

ways of producing an electric solid-body guitar in the late 1930’s. In<br />

1929, at the age of thirteen, he had amplified his first acoustic guitar.<br />

Another influential inventor of the 1940’s was Paul Bigsby. He built<br />

a prototype solid-body guitar for country music star Merle Travis in<br />

1947. It was Leo Fender who revolutionized the electric guitar industry<br />

by producing the first commercially viable solid-body electric<br />

guitar, the Broadcaster, in 1948.<br />

Leo Fender<br />

Broadcaster guitar / 123<br />

Leo Fender was born in the Anaheim, California, area in 1909. As<br />

a teenager, he began to build <strong>and</strong> repair guitars. By the 1930’s,<br />

Fender was building <strong>and</strong> renting out public address systems for<br />

group gatherings. In 1937, after short tenures of employment with<br />

the Division of Highways <strong>and</strong> the U.S. Tire Company, he opened a<br />

radio repair company in Fullerton, California. Always looking to<br />

exp<strong>and</strong> <strong>and</strong> invent new <strong>and</strong> exciting electrical gadgets, Fender <strong>and</strong><br />

Clayton Orr “Doc” Kauffman started the K&FCompany in 1944.<br />

Kauffman was a musician <strong>and</strong> a former employee of the Electro<br />

String Instrument Company. TheK&FCompany lasted until 1946<br />

<strong>and</strong> produced steel guitars <strong>and</strong> amplifiers. After that partnership<br />

ended, Fender founded the Fender Electric Instruments Company.<br />

With the help of George Fullerton, who joined the company in<br />

1948, Fender developed the Fender Broadcaster. The body of the<br />

Broadcaster was made of a solid plank of ash wood. The corners of<br />

the ash body were rounded. There was a cutaway located under the<br />

joint with the solid maple neck, making it easier for the guitarist to<br />

access the higher frets. The maple neck was bolted to the body of the<br />

guitar, which was unusual, since most guitar necks prior to the<br />

Broadcaster had been glued to the body. Frets were positioned directly<br />

into designed cuts made in the maple of the neck. The guitar<br />

had two pickups.<br />

The Fender Electric Instruments Company made fewer than one


124 / Broadcaster guitar<br />

thous<strong>and</strong> Broadcasters. In 1950, the name of the guitar was changed<br />

from the Broadcaster to the Telecaster, as the Gretsch company had<br />

already registered the name Broadcaster for some of its drums <strong>and</strong><br />

banjos. Fender decided not to fight in court over use of the name.<br />

Leo Fender has been called the Henry Ford of the solid-body<br />

electric guitar, <strong>and</strong> the Telecaster became known as the Model T of<br />

the industry. The early Telecasters sold for $189.50. Besides being inexpensive,<br />

the Telecaster was a very durable instrument. Basically,<br />

the Telecaster was a continuation of the Broadcaster. Fender did not<br />

file for a patent on its unique bridge pickup until January 13, 1950,<br />

<strong>and</strong> he did not file for a patent on the Telecaster’s unique body<br />

shape until April 3, 1951.<br />

In the music industry during the late 1940’s, it was important for<br />

a company to unveil new instruments at trade shows. At this time,<br />

there was only one important trade show, sponsored by the National<br />

Association of Music Merchants. The Broadcaster was first<br />

sprung on the industry at the 1948 trade show in Chicago. The industry<br />

had seen nothing like this guitar ever before. This new guitar<br />

existed only to be amplified; it was not merely an acoustic guitar<br />

that had been converted.<br />

Impact<br />

The Telecaster, as it would be called after 1950, remained in continuous<br />

production for more years than any other guitar of its type<br />

<strong>and</strong> was one of the industry’s best sellers. From the beginning, it<br />

looked <strong>and</strong> sounded unique. The electrified acoustic guitars had a<br />

mellow woody tone, whereas the Telecaster had a clean twangy<br />

tone. This tone made it popular with country <strong>and</strong> blues guitarists.<br />

The Telecaster could also be played at higher volume than previous<br />

electric guitars.<br />

Because Leo Fender attempted something revolutionary by introducing<br />

an electric solid-body guitar, there was no guarantee that<br />

his business venture would succeed. Fender Electric Instruments<br />

Company had fifteen employees in 1947. At times, during the early<br />

years of the company, it looked as though Fender’s dreams would<br />

not come to fruition, but the company persevered <strong>and</strong> grew. Between<br />

1948 <strong>and</strong> 1955 with an increase of employees, the company


was able to produce ten thous<strong>and</strong> Broadcaster/Telecaster guitars.<br />

Fender had taken a big risk, but it paid off enormously. Between<br />

1958 <strong>and</strong> the mid-1970’s, Fender produced more than 250,000 Telecasters.<br />

Other guitar manufacturers were placed in a position of<br />

having to catch up. Fender had succeeded in developing a process<br />

by which electric solid-body guitars could be manufactured profitably<br />

on a large scale.<br />

Early Guitar Pickups<br />

Broadcaster guitar / 125<br />

The first pickups used on a guitar can be traced back to the 1920’s<br />

<strong>and</strong> the efforts of Lloyd Loar, but there was not strong interest on the<br />

part of the American public for the guitar to be amplified. The public<br />

did not become intrigued until the 1930’s. Charlie Christian’s<br />

electric guitar performances with Benny Goodman woke up the<br />

public to the potential of this new <strong>and</strong> exciting sound. It was not until<br />

the 1950’s, though, that the electric guitar became firmly established.<br />

Leo Fender was the right man in the right place. He could not<br />

have known that his Fender guitars would help to usher in a whole<br />

new musical l<strong>and</strong>scape. Since the electric guitar was the newest<br />

member of the family of guitars, it took some time for musical audiences<br />

to fully appreciate what it could do. The electric solid-body<br />

guitar has been called a dangerous, uncivilized instrument. The<br />

youth culture of the 1950’s found in this new guitar a voice for their<br />

rebellion. Fender unleashed a revolution not only in the construction<br />

of a guitar but also in the way popular music would be approached<br />

henceforth.<br />

Because of the ever-increasing dem<strong>and</strong> for the Fender product,<br />

Fender Sales was established as a separate distribution company in<br />

1953 by Don R<strong>and</strong>all. Fender Electric Instruments Company had fifteen<br />

employees in 1947, but by 1955, the company employed fifty<br />

people. By 1960, the number of employees had risen to more than<br />

one hundred. Before Leo Fender sold the company to CBS on January<br />

4, 1965, for $13 million, the company occupied twenty-seven<br />

buildings <strong>and</strong> employed more than five hundred workers.<br />

Always interested in finding new ways of designing a more nearly<br />

perfect guitar, Leo Fender again came up with a remarkable guitar in<br />

1954, with the Stratocaster. There was talk in the guitar industry that


126 / Broadcaster guitar<br />

Charlie Christian<br />

Charlie Christian (1919-1942) did not invent the electric guitar,<br />

but he did pioneer its use. He was born to music, <strong>and</strong> for<br />

jazz aficionados he quickly developed into a legend, not only<br />

establishing a new solo instrument but also helping to invent a<br />

whole new type of jazz.<br />

Christian grew up in Texas, surrounded by a family of professional<br />

musicians. His parents <strong>and</strong> two brothers played trumpet,<br />

guitar, <strong>and</strong> piano, <strong>and</strong> sang, <strong>and</strong> Charlie was quick to imitate<br />

them. As a boy he made his own guitars out of cigar boxes<br />

<strong>and</strong>, according to a childhood friend, novelist Ralph Ellison,<br />

wowed his friends at school with his riffs. When he first heard<br />

an electric guitar in the mid-1930’s, he made that his own, too.<br />

The acoustic guitar had been only a backup instrument in<br />

jazz because it was too quiet to soar in solos. In 1935, Eddie Durham<br />

found that electric guitars could swing side by side with<br />

louder instruments. Charlie, already an experienced performer<br />

with acoustic guitar <strong>and</strong> bass, immediately recognized the power<br />

<strong>and</strong> range of subtle expression possible with the electrified instrument.<br />

He bought a Gibson ES-150 <strong>and</strong> began to make musical<br />

history with his improvisations.<br />

He impressed producer John Hammond, who introduced<br />

him to big-b<strong>and</strong> leader Benny Goodman in 1939. Notoriously<br />

hard to please, Goodman rejected Christian after an audition.<br />

However, Hammond later sneaked him on stage while the<br />

Goodman b<strong>and</strong> was performing. Outraged, Goodman segued<br />

into a tune he was sure Christian did not know, “Rose Room.”<br />

Christian was undaunted. He delivered an astonishingly inventive<br />

solo, <strong>and</strong> Goodman was won over despite himself. Christian’s<br />

ensuing tenure with Goodman’s b<strong>and</strong> brought electric<br />

guitar solos into the limelight.<br />

However, it was during after-hours jam sessions at the Hotel<br />

Cecil in New York that Christian left his stylistic imprint on<br />

jazz. Including such jazz greats as Joe Guy, Thelonious Monk,<br />

<strong>and</strong> Kenny Clarke, the groups played around with new sounds.<br />

Out of these sessions bebop was born, <strong>and</strong> Christian was a central<br />

figure. Sick with tuberculosis, he had to quit playing in 1941<br />

<strong>and</strong> died the following spring, only twenty-five years old.


Fender had gone too far with the introduction of the Stratocaster, but<br />

it became a huge success because of its versatility. It was the first commercial<br />

solid-body electric guitar to have three pickups <strong>and</strong> a vibrato<br />

bar. It was also easier to play than the Telecaster because of its double<br />

cutaway, contoured body, <strong>and</strong> scooped back. The Stratocaster sold<br />

for $249.50. Since its introduction, the Stratocaster has undergone<br />

some minor changes, but Fender <strong>and</strong> his staff basically got it right the<br />

first time.<br />

The Gibson company entered the solid-body market in 1952 with<br />

the unveiling of the “Les Paul” model. After the Telecaster, the Les<br />

Paul guitar was the next significant solid-body to be introduced. Les<br />

Paul was a legendary guitarist who also had been experimenting<br />

with electric guitar designs for many years. The Gibson designers<br />

came up with a striking model that produced a thick rounded tone.<br />

Over the years, the Les Paul model has won a loyal following.<br />

The Precision Bass<br />

Broadcaster guitar / 127<br />

In 1951, Leo Fender introduced another revolutionary guitar, the<br />

Precision bass. At a cost of $195.50, the first electric bass would go on<br />

to dominate the market. The Fender company has manufactured numerous<br />

guitar models over the years, but the three that st<strong>and</strong> above<br />

all others in the field are the Telecaster, the Precision bass, <strong>and</strong> the<br />

Stratocaster. The Telecaster is considered to be more of a workhorse,<br />

whereas the Stratocaster is thought of as the thoroughbred of electric<br />

guitars. The Precision bass was in its own right a revolutionary guitar.<br />

With a styling that had been copied from the Telecaster, the Precision<br />

freed musicians from bulky oversized acoustic basses, which<br />

were prone to feedback. The name Precision had meaning. Fender’s<br />

electric bass made it possible, with its frets, for the precise playing of<br />

notes; many acoustic basses were fretless. The original Precision bass<br />

model was manufactured from 1951 to 1954. The next version lasted<br />

from 1954 until June of 1957. The Precision bass that went into production<br />

in June, 1957, with its split humbucking pickup, continued to<br />

be the st<strong>and</strong>ard electric bass on the market into the 1990’s.<br />

By 1964, the Fender Electric Instruments Company had grown<br />

enormously. In addition to Leo Fender, a number of crucial people<br />

worked for the organization, including George Fullerton <strong>and</strong> Don


128 / Broadcaster guitar<br />

R<strong>and</strong>all. Fred Tavares joined the company’s research <strong>and</strong> development<br />

team in 1953. In May, 1954, Forrest White became Fender’s<br />

plant manager. All these individuals played vital roles in the success<br />

of Fender, but the driving force behind the scene was always<br />

Leo Fender. As Fender’s health deteriorated, R<strong>and</strong>all commenced<br />

negotiations with CBS to sell the Fender company. In January, 1965,<br />

CBS bought Fender for $13 million. Eventually, Leo Fender regained<br />

his health, <strong>and</strong> he was hired as a technical adviser by CBS/Fender.<br />

He continued in this capacity until 1970. He remained determined<br />

to create more guitar designs of note. Although he never again produced<br />

anything that could equal his previous success, he never<br />

stopped trying to attain a new perfection of guitar design.<br />

Fender died on March 21, 1991, in Fullerton, California. He had<br />

suffered for years from Parkinson’s disease, <strong>and</strong> he died of complications<br />

from the disease. He is remembered for his Broadcaster/<br />

Telecaster, Precision bass, <strong>and</strong> Stratocaster, which revolutionized<br />

popular music. Because the Fender company was able to mass produce<br />

these <strong>and</strong> other solid-body electric guitars, new styles of music<br />

that relied on the sound made by an electric guitar exploded onto<br />

the scene. The electric guitar manufacturing business grew rapidly<br />

after Fender introduced mass production. Besides American companies,<br />

there are guitar companies that have flourished in Europe<br />

<strong>and</strong> Japan.<br />

The marriage between rock music <strong>and</strong> solid-body electric guitars<br />

was initiated by the Fender guitars. The Telecaster, Precision bass,<br />

<strong>and</strong> Stratocaster become synonymous with the explosive character<br />

of rock <strong>and</strong> roll music. The multi-billion-dollar music business can<br />

point to Fender as the pragmatic visionary who put the solid-body<br />

electric guitar into the forefront of the musical scene. His innovative<br />

guitars have been used by some of the most important guitarists of<br />

the rock era, including Jimi Hendrix, Eric Clapton, <strong>and</strong> Jeff Beck.<br />

More important, Fender guitars have remained bestsellers with<br />

the public worldwide. Amateur musicians purchased them by the<br />

thous<strong>and</strong>s for their own entertainment. Owning <strong>and</strong> playing a<br />

Fender guitar, or one of the other electric guitars that followed, allowed<br />

these amateurs to feel closer to their musician idols. A large<br />

market for sheet music from popular artists also developed.<br />

In 1992, Fender was inducted into the Rock <strong>and</strong> Roll Hall of


Fame. He is one of the few non-musicians ever to be inducted. The<br />

sound of an electric guitar is the sound of exuberance, <strong>and</strong> since the<br />

Broadcaster was first unveiled in 1948, that sound has grown to be<br />

pervasive <strong>and</strong> enormously profitable.<br />

See also Cassette recording; Dolby noise reduction; Electronic<br />

synthesizer.<br />

Further Reading<br />

Broadcaster guitar / 129<br />

Bacon, Tony, <strong>and</strong> Paul Day. The Fender Book. San Francisco: GPI<br />

Books, 1992.<br />

Brosnac, Donald, ed. Guitars Made by the Fender Company. Westport,<br />

Conn.: Bold Strummer, 1986.<br />

Freeth, Nick. The Electric Guitar. Philadelphia: Courage Books, 1999.<br />

Trynka, Paul. The Electric Guitar: An Illustrated History. San Francisco:<br />

Chronicle Books, 1995.<br />

Wheeler, Tom. American Guitars: An Illustrated History. New York:<br />

Harper & Row, 1982.<br />

_____. “Electric Guitars.” In The Guitar Book: A H<strong>and</strong>book for Electric<br />

<strong>and</strong> Acoustic Guitarists. New York: Harper & Row, 1974.


130<br />

Brownie camera<br />

Brownie camera<br />

The invention: The first inexpensive <strong>and</strong> easy-to-use camera available<br />

to the general public, the Brownie revolutionized photography<br />

by making it possible for every person to become a photographer.<br />

The people behind the invention:<br />

George Eastman (1854-1932), founder of the Eastman Kodak<br />

Company<br />

Frank A. Brownell, a camera maker for the Kodak Company<br />

who designed the Brownie<br />

Henry M. Reichenbach, a chemist who worked with Eastman to<br />

develop flexible film<br />

William H. Walker, a Rochester camera manufacturer who<br />

collaborated with Eastman<br />

A New Way to Take Pictures<br />

In early February of 1900, the first shipments of a new small box<br />

camera called the Brownie reached Kodak dealers in the United<br />

States <strong>and</strong> Engl<strong>and</strong>. George Eastman, eager to put photography<br />

within the reach of everyone, had directed Frank Brownell to design<br />

a small camera that could be manufactured inexpensively but that<br />

would still take good photographs.<br />

Advertisements for the Brownie proclaimed that everyone—<br />

even children—could take good pictures with the camera. The<br />

Brownie was aimed directly at the children’s market, a fact indicated<br />

by its box, which was decorated with drawings of imaginary<br />

elves called “Brownies” created by the Canadian illustrator Palmer<br />

Cox. Moreover, the camera cost only one dollar.<br />

The Brownie was made of jute board <strong>and</strong> wood, with a hinged<br />

back fastened by a sliding catch. It had an inexpensive two-piece<br />

glass lens <strong>and</strong> a simple rotary shutter that allowed both timed <strong>and</strong><br />

instantaneous exposures to be made. With a lens aperture of approximately<br />

f14 <strong>and</strong> a shutter speed of approximately 1/50 of a second,<br />

the Brownie was certainly capable of taking acceptable snap-


Brownie camera / 131<br />

shots. It had no viewfinder; however, an optional clip-on reflecting<br />

viewfinder was available. The camera came loaded with a six-exposure<br />

roll of Kodak film that produced square negatives 2.5 inches on<br />

a side. This film could be developed, printed, <strong>and</strong> mounted for forty<br />

cents, <strong>and</strong> a new roll could be purchased for fifteen cents.<br />

George Eastman’s first career choice had been banking, but when<br />

he failed to receive a promotion he thought he deserved, he decided<br />

to devote himself to his hobby, photography. Having worked with a<br />

rigorous wet-plate process, he knew why there were few amateur<br />

photographers at the time—the whole process, from plate preparation<br />

to printing, was too expensive <strong>and</strong> too much trouble. Even so,<br />

he had already begun to think about the commercial possibilities of<br />

photography; after reading of British experiments with dry-plate<br />

technology, he set up a small chemical laboratory <strong>and</strong> came up with<br />

a process of his own. The Eastman Dry Plate Company became one<br />

of the most successful producers of gelatin dry plates.<br />

Dry-plate photography had attracted more amateurs, but it was<br />

still a complicated <strong>and</strong> expensive hobby. Eastman realized that the<br />

number of photographers would have to increase considerably if<br />

the market for cameras <strong>and</strong> supplies were to have any potential. In<br />

the early 1880’s, Eastman first formulated the policies that would<br />

make the Eastman Kodak Company so successful in years to come:<br />

mass production, low prices, foreign <strong>and</strong> domestic distribution, <strong>and</strong><br />

selling through extensive advertising <strong>and</strong> by demonstration.<br />

In his efforts to exp<strong>and</strong> the amateur market, Eastman first tackled<br />

the problem of the glass-plate negative, which was heavy, fragile,<br />

<strong>and</strong> expensive to make. By 1884, his experiments with paper<br />

negatives had been successful enough that he changed the name of<br />

his company to The Eastman Dry Plate <strong>and</strong> Film Company. Since<br />

flexible roll film needed some sort of device to hold it steady in the<br />

camera’s focal plane, Eastman collaborated with William Walker<br />

to develop the Eastman-Walker roll-holder. Eastman’s pioneering<br />

manufacture <strong>and</strong> use of roll films led to the appearance on the market<br />

in the 1880’s of a wide array of h<strong>and</strong> cameras from a number of<br />

different companies. Such cameras were called “detective cameras”<br />

because they were small <strong>and</strong> could be used surreptitiously. The<br />

most famous of these, introduced by Eastman in 1888, was named<br />

the “Kodak”—a word he coined to be terse, distinctive, <strong>and</strong> easily


132 / Brownie camera<br />

pronounced in any language. This camera’s simplicity of operation<br />

was appealing to the general public <strong>and</strong> stimulated the growth of<br />

amateur photography.<br />

The Camera<br />

The Kodak was a box about seven inches long <strong>and</strong> four inches<br />

wide, with a one-speed shutter <strong>and</strong> a fixed-focus lens that produced<br />

reasonably sharp pictures. It came loaded with enough roll film to<br />

make one hundred exposures. The camera’s initial price of twentyfive<br />

dollars included the cost of processing the first roll of film; the<br />

camera also came with a leather case <strong>and</strong> strap. After the film was<br />

exposed, the camera was mailed, unopened, to the company’s plant<br />

in Rochester, New York, where the developing <strong>and</strong> printing were<br />

done. For an additional ten dollars, the camera was reloaded <strong>and</strong><br />

sent back to the customer.<br />

The Kodak was advertised in mass-market publications, rather<br />

than in specialized photographic journals, with the slogan: “You<br />

press the button, we do the rest.” With his introduction of a camera<br />

that was easy to use <strong>and</strong> a service that eliminated the need to know<br />

anything about processing negatives, Eastman revolutionized the<br />

photographic market. Thous<strong>and</strong>s of people no longer depended<br />

upon professional photographers for their portraits but instead<br />

learned to make their own. In 1892, the Eastman Dry Plate <strong>and</strong> Film<br />

Company became the Eastman Kodak Company, <strong>and</strong> by the mid-<br />

1890’s, one hundred thous<strong>and</strong> Kodak cameras had been manufactured<br />

<strong>and</strong> sold, half of them in Europe by Kodak Limited.<br />

Having popularized photography with the first Kodak, in 1900<br />

Eastman turned his attention to the children’s market with the introduction<br />

of the Brownie. The first five thous<strong>and</strong> cameras sent to<br />

dealers were sold immediately; by the end of the following year, almost<br />

a quarter of a million had been sold. The Kodak Company organized<br />

Brownie camera clubs <strong>and</strong> held competitions specifically<br />

for young photographers. The Brownie came with an instruction<br />

booklet that gave children simple directions for taking successful<br />

pictures, <strong>and</strong> “The Brownie Boy,” an appealing youngster who<br />

loved photography, became a st<strong>and</strong>ard feature of Kodak’s advertisements.


Impact<br />

Brownie camera / 133<br />

Eastman followed the success of the first Brownie by introducing<br />

several additional models between 1901 <strong>and</strong> 1917. Each was a more<br />

elaborate version of the original. These Brownie box cameras were<br />

on the market until the early 1930’s, <strong>and</strong> their success inspired other<br />

companies to manufacture box cameras of their own. In 1906, the<br />

Ansco company produced the Buster Brown camera in three sizes<br />

that corresponded to Kodak’s Brownie camera range; in 1910 <strong>and</strong><br />

1914, Ansco made three more versions. The Seneca company’s<br />

Scout box camera, in three sizes, appeared in 1913, <strong>and</strong> Sears Roebuck’s<br />

Kewpie cameras, in five sizes, were sold beginning in 1916.<br />

In Engl<strong>and</strong>, the Houghtons company introduced its first Scout camera<br />

in 1901, followed by another series of four box cameras in 1910<br />

sold under the Ensign trademark. Other English manufacturers of<br />

box cameras included the James Sinclair company, with its Traveller<br />

Una of 1909, <strong>and</strong> the Thornton-Pickard company, with a Filma camera<br />

marketed in four sizes in 1912.<br />

After World War I ended, several series of box cameras were<br />

manufactured in Germany by companies that had formerly concentrated<br />

on more advanced <strong>and</strong> expensive cameras. The success of<br />

box cameras in other countries, led by Kodak’s Brownie, undoubtedly<br />

prompted this trend in the German photographic industry. The<br />

Ernemann Film K series of cameras in three sizes, introduced in<br />

1919, <strong>and</strong> the all-metal Trapp Little Wonder of 1922 are examples of<br />

popular German box cameras.<br />

In the early 1920’s, camera manufacturers began making boxcamera<br />

bodies from metal rather than from wood <strong>and</strong> cardboard.<br />

Machine-formed metal was less expensive than the traditional h<strong>and</strong>worked<br />

materials. In 1924, Kodak’s two most popular Brownie sizes<br />

appeared with aluminum bodies.<br />

In 1928, Kodak Limited of Engl<strong>and</strong> added two important new<br />

features to the Brownie—a built-in portrait lens, which could be<br />

brought in front of the taking lens by pressing a lever, <strong>and</strong> camera<br />

bodies in a range of seven different fashion colors. The Beau<br />

Brownie cameras, made in 1930, were the most popular of all the<br />

colored box cameras. The work of Walter Dorwin Teague, a leading<br />

American designer, these cameras had an Art Deco geometric pat-


134 / Brownie camera<br />

tern on the front panel, which was enameled in a color matching the<br />

leatherette covering of the camera body. Several other companies,<br />

including Ansco, again followed Kodak’s lead <strong>and</strong> introduced their<br />

own lines of colored cameras.<br />

In the 1930’s, several new box cameras with interesting features appeared,<br />

many manufactured by leading film companies. In France, the<br />

Lumiere Company advertised a series of box cameras—the Luxbox,<br />

Scoutbox, <strong>and</strong> Lumibox—that ranged from a basic camera to one with<br />

an adjustable lens <strong>and</strong> shutter. In 1933, the German Agfa company restyled<br />

its entire range of box cameras, <strong>and</strong> in 1939, the Italian Ferrania<br />

company entered the market with box cameras in two sizes. In 1932,<br />

Kodak redesigned its Brownie series to take the new 620 roll film,<br />

which it had just introduced. This film <strong>and</strong> the new Six-20 Brownies inspired<br />

other companies to experiment with variations of their own;<br />

some box cameras, such as the Certo Double-box, the Coronet Every<br />

Distance, <strong>and</strong> the Ensign E-20 cameras, offered a choice of two picture<br />

formats.<br />

Another new trend was a move toward smaller-format cameras<br />

using st<strong>and</strong>ard 127 roll film. In 1934, Kodak marketed the small<br />

Baby Brownie. Designed by Teague <strong>and</strong> made from molded black<br />

plastic, this little camera with a folding viewfinder sold for only one<br />

dollar—the price of the original Brownie in 1900.<br />

The Baby Brownie, the first Kodak camera made of molded plastic,<br />

heralded the move to the use of plastic in camera manufacture.<br />

Soon many others, such as the Altissa series of box cameras <strong>and</strong> the<br />

Voigtl<strong>and</strong>er Brilliant V/6 camera, were being made from this new<br />

material.<br />

Later Trends<br />

By the late 1930’s, flashbulbs had replaced flash powder for taking<br />

pictures in low light; again, the Eastman Kodak Company led<br />

the way in introducing this new technology as a feature on the inexpensive<br />

box camera. The Falcon Press-Flash, marketed in 1939, was<br />

the first mass-produced camera to have flash synchronization <strong>and</strong><br />

was followed the next year by the Six-20 Flash Brownie, which had a<br />

detachable flash gun. In the early 1940’s, other companies, such as<br />

Agfa-Ansco, introduced this feature on their own box cameras.


George Eastman<br />

Brownie camera / 135<br />

Frugal, bold, practical, generous to those who were loyal,<br />

impatient with dissent, <strong>and</strong> possessing a steely determination,<br />

George Eastman (1854-1932) rose to become one of the richest<br />

people of his generation. He abhorred poverty <strong>and</strong> did his best<br />

to raise others from it as well.<br />

At age fourteen, when his father died, Eastman<br />

had to drop out of school to support his mother <strong>and</strong><br />

sister. The missed opportunity for an education <strong>and</strong><br />

the struggle to earn a living shaped his outlook. He<br />

worked at an insurance agency <strong>and</strong> then at a bank,<br />

keeping careful record of the money he earned. By<br />

the time he was twenty-five he had saved three thous<strong>and</strong><br />

dollars <strong>and</strong> found his job as a banker to be unrewarding.<br />

As a teenager, he had taught himself photography.<br />

However, that was only a start. He taught himself the physics<br />

<strong>and</strong> chemistry of photography too—<strong>and</strong> enough French <strong>and</strong><br />

German to read the latest foreign scientific journals. His purpose<br />

was practical, to make cameras cheap <strong>and</strong> easy to use so<br />

that average people could own them. This launched him on the<br />

career of invention <strong>and</strong> business that took him away from banking<br />

<strong>and</strong> made his fortune. At the same time he remembered his<br />

origins <strong>and</strong> family. Out of his first earnings, he bought photographs<br />

for his mother <strong>and</strong> a favorite teacher. He never stopped<br />

giving. At the company he founded, he gave substantial raises<br />

to employees, reduced their hours, <strong>and</strong> installed safety equipment,<br />

a medical department, <strong>and</strong> a lunch room. He gave millions<br />

to the Hampton Institute, Tuskegee Institute, Massachusetts<br />

Institute of Technology, <strong>and</strong> University of Rochester, while<br />

also establishing dental clinics for the poor.<br />

In his old age he found he could no longer keep up with his<br />

younger scientific <strong>and</strong> business colleagues. In 1932, leaving behind<br />

a note that asked, simply, “My work is done, why wait?”<br />

he committed suicide. Even then he continued to give. His will<br />

left most of his vast fortune to charities.<br />

In the years after World War II, the box camera evolved into an<br />

eye-level camera, making it more convenient to carry <strong>and</strong> use.<br />

Many amateur photographers, however, still had trouble h<strong>and</strong>ling<br />

(Smithsonian Institution)


136 / Brownie camera<br />

paper-backed roll film <strong>and</strong> were taking their cameras back to dealers<br />

to be unloaded <strong>and</strong> reloaded. Kodak therefore developed a new<br />

system of film loading, using the Kodapak cartridge, which could<br />

be mass-produced with a high degree of accuracy by precision plastic-molding<br />

techniques. To load the camera, the user simply opened<br />

the camera back <strong>and</strong> inserted the cartridge. This new film was introduced<br />

in 1963, along with a series of Instamatic cameras designed<br />

for its use. Both were immediately successful.<br />

The popularity of the film cartridge ended the long history of the<br />

simple <strong>and</strong> inexpensive roll film camera. The last English Brownie<br />

was made in 1967, <strong>and</strong> the series of Brownies made in the United<br />

States was discontinued in 1970. Eastman’s original marketing strategy<br />

of simplifying photography in order to increase the dem<strong>and</strong> for<br />

cameras <strong>and</strong> film continued, however, with the public’s acceptance<br />

of cartridge-loading cameras such as the Instamatic.<br />

From the beginning, Eastman had recognized that there were<br />

two kinds of photographers other than professionals. The first, he<br />

declared, were the true amateurs who devoted time enough to acquire<br />

skill in the complex processing procedures of the day. The second<br />

were those who merely wanted personal pictures or memorabilia<br />

of their everyday lives, families, <strong>and</strong> travels. The second class,<br />

he observed, outnumbered the first by almost ten to one. Thus, it<br />

was to this second kind of amateur photographer that Eastman had<br />

appealed, both with his first cameras <strong>and</strong> with his advertising slogan,<br />

“You press the button, we do the rest.” Eastman had done<br />

much more than simply invent cameras <strong>and</strong> films; he had invented<br />

a system <strong>and</strong> then developed the means for supporting that system.<br />

This is essentially what the Eastman Kodak Company continued to<br />

accomplish with the series of Instamatics <strong>and</strong> other descendants of<br />

the original Brownie. In the decade between 1963 <strong>and</strong> 1973, for example,<br />

approximately sixty million Instamatics were sold throughout<br />

the world.<br />

The research, manufacturing, <strong>and</strong> marketing activities of the<br />

Eastman Kodak Company have been so complex <strong>and</strong> varied that no<br />

one would suggest that the company’s prosperity rests solely on the<br />

success of its line of inexpensive cameras <strong>and</strong> cartridge films, although<br />

these have continued to be important to the company. Like<br />

Kodak, however, most large companies in the photographic indus-


try have exp<strong>and</strong>ed their research to satisfy the ever-growing dem<strong>and</strong><br />

from amateurs. The amateurism that George Eastman recognized<br />

<strong>and</strong> encouraged at the beginning of the twentieth century<br />

thus still flourished at its end.<br />

See also Autochrome plate; Color film; Instant photography.<br />

Further Reading<br />

Brownie camera / 137<br />

Brooke-Ball, Peter. George Eastman <strong>and</strong> Kodak. Watford: Exley, 1994.<br />

Collins, Douglas. The Story of Kodak. New York: Harry N. Abrams,<br />

1990.<br />

Freund, Gisele. Photography <strong>and</strong> Society. Boston: David R. Godine,<br />

1980.<br />

Wade, John. A Short History of the Camera. Watford, Engl<strong>and</strong>: Fountain<br />

Press, 1979.<br />

West, Nancy Martha. Kodak <strong>and</strong> the Lens of Nostalgia. Charlottesville:<br />

University Press of Virginia, 2000.


138<br />

Bubble memory<br />

Bubble memory<br />

The invention: An early nonvolatile medium for storing information<br />

on computers.<br />

The person behind the invention:<br />

Andrew H. Bobeck (1926- ), a Bell Telephone Laboratories<br />

scientist<br />

Magnetic Technology<br />

The fanfare over the commercial prospects of magnetic bubbles<br />

was begun on August 8, 1969, by a report appearing in both The New<br />

York Times <strong>and</strong> The Wall Street Journal. The early 1970’s would see the<br />

anticipation mount (at least in the computer world) with each prediction<br />

of the benefits of this revolution in information storage technology.<br />

Although it was not disclosed to the public until August of 1969,<br />

magnetic bubble technology had held the interest of a small group<br />

of researchers around the world for many years. The organization<br />

that probably can claim the greatest research advances with respect<br />

to computer applications of magnetic bubbles is Bell Telephone<br />

Laboratories (later part of American Telephone <strong>and</strong> Telegraph). Basic<br />

research into the properties of certain ferrimagnetic materials<br />

started at Bell Laboratories shortly after the end of World War II<br />

(1939-1945).<br />

Ferrimagnetic substances are typically magnetic iron oxides. Research<br />

into the properties of these <strong>and</strong> related compounds accelerated<br />

after the discovery of ferrimagnetic garnets in 1956 (these are a<br />

class of ferrimagnetic oxide materials that have the crystal structure<br />

of garnet). Ferrimagnetism is similar to ferromagnetism, the phenomenon<br />

that accounts for the strong attraction of one magnetized<br />

body for another. The ferromagnetic materials most suited for bubble<br />

memories contain, in addition to iron, the element yttrium or a<br />

metal from the rare earth series.<br />

It was a fruitful collaboration between scientist <strong>and</strong> engineer,<br />

between pure <strong>and</strong> applied science, that produced this promising


eakthrough in data storage technology. In 1966, Bell Laboratories<br />

scientist Andrew H. Bobeck <strong>and</strong> his coworkers were the first to realize<br />

the data storage potential offered by the strange behavior of thin<br />

slices of magnetic iron oxides under an applied magnetic field. The<br />

first U.S. patent for a memory device using magnetic bubbles was<br />

filed by Bobeck in the fall of 1966 <strong>and</strong> issued on August 5, 1969.<br />

Bubbles Full of Memories<br />

Bubble memory / 139<br />

The three basic functional elements of a computer are the central<br />

processing unit, the input/output unit, <strong>and</strong> memory. Most implementations<br />

of semiconductor memory require a constant power<br />

source to retain the stored data. If the power is turned off, all stored<br />

data are lost. Memory with this characteristic is called “volatile.”<br />

Disks <strong>and</strong> tapes, which are typically used for secondary memory,<br />

are “nonvolatile.” Nonvolatile memory relies on the orientation of<br />

magnetic domains, rather than on electrical currents, to sustain its<br />

existence.<br />

One can visualize by analogy how this will work by taking a<br />

group of permanent bar magnets that are labeled with N for north at<br />

one end <strong>and</strong> S for south at the other. If an arrow is painted starting<br />

from the north end with the tip at the south end on each magnet, an<br />

orientation can then be assigned to a magnetic domain (here one<br />

whole bar magnet). Data are “stored” with these bar magnets by arranging<br />

them in rows, some pointing up, some pointing down. Different<br />

arrangements translate to different data. In the binary world<br />

of the computer, all information is represented by two states. A<br />

stored data item (known as a “bit,” or binary digit) is either on or off,<br />

up or down, true or false, depending on the physical representation.<br />

The “on” state is commonly labeled with the number 1 <strong>and</strong> the “off”<br />

state with the number 0. This is the principle behind magnetic disk<br />

<strong>and</strong> tape data storage.<br />

Now imagine a thin slice of a certain type of magnetic material in<br />

the shape of a 3-by-5-inch index card. Under a microscope, using a<br />

special source of light, one can see through this thin slice in many regions<br />

of the surface. Darker, snakelike regions can also be seen, representing<br />

domains of an opposite orientation (polarity) to the transparent<br />

regions. If a weak external magnetic field is then applied by


140 / Bubble memory<br />

placing a permanent magnet of the same shape as the card on the<br />

underside of the slice, a strange thing happens to the dark serpentine<br />

pattern—the long domains shrink <strong>and</strong> eventually contract into<br />

“bubbles,” tiny magnetized spots. Viewed from the side of the slice,<br />

the bubbles are cylindrically shaped domains having a polarity opposite<br />

to that of the material on which they rest. The presence or absence<br />

of a bubble indicates eithera0ora1bit. Data bits are stored by<br />

moving the bubbles in the thin film. As long as the field is applied<br />

by the permanent magnet substrate, the data will be retained. The<br />

bubble is thus a nonvolatile medium for data storage.<br />

Consequences<br />

Magnetic bubble memory created quite a stir in 1969 with its<br />

splashy public introduction. Most of the manufacturers of computer<br />

chips immediately instituted bubble memory development projects.<br />

Texas Instruments, Philips, Hitachi, Motorola, Fujitsu, <strong>and</strong> International<br />

Business Machines (IBM) joined the race with Bell Laboratories<br />

to mass-produce bubble memory chips. Texas Instruments<br />

became the first major chip manufacturer to mass-produce bubble<br />

memories in the mid-to-late 1970’s. By 1990, however, almost all the<br />

research into magnetic bubble technology had shifted to Japan.<br />

Hitachi <strong>and</strong> Fujitsu began to invest heavily in this area.<br />

Mass production proved to be the most difficult task. Although<br />

the materials it uses are different, the process of producing magnetic<br />

bubble memory chips is similar to the process applied in producing<br />

semiconductor-based chips such as those used for r<strong>and</strong>om access<br />

memory (RAM). It is for this reason that major semiconductor manufacturers<br />

<strong>and</strong> computer companies initially invested in this technology.<br />

Lower fabrication yields <strong>and</strong> reliability issues plagued<br />

early production runs, however, <strong>and</strong>, although these problems<br />

have mostly been solved, gains in the performance characteristics of<br />

competing conventional memories have limited the impact that<br />

magnetic bubble technology has had on the marketplace. The materials<br />

used for magnetic bubble memories are costlier <strong>and</strong> possess<br />

more complicated structures than those used for semiconductor or<br />

disk memory.<br />

Speed <strong>and</strong> cost of materials are not the only bases for compari-


son. It is possible to perform some elementary logic with magnetic<br />

bubbles. Conventional semiconductor-based memory offers storage<br />

only. The capability of performing logic with magnetic bubbles<br />

puts bubble technology far ahead of other magnetic technologies<br />

with respect to functional versatility.<br />

A small niche market for bubble memory developed in the 1980’s.<br />

Magnetic bubble memory can be found in intelligent terminals, desktop<br />

computers, embedded systems, test equipment, <strong>and</strong> similar microcomputer-based<br />

systems.<br />

See also Computer chips; Floppy disk; Hard disk; Optical disk;<br />

Personal computer.<br />

Further Reading<br />

Bubble memory / 141<br />

“Bubble Memory’s Ruggedness Revives Interest for Military Use.”<br />

Aviation Week <strong>and</strong> Space Technology 130, no. 3 (January 16, 1989).<br />

Graff, Gordon. “Better Bubbles.” Popular Science 232, no. 2 (February,<br />

1988).<br />

McLeod, Jonah. “Will Bubble Memories Make a Comeback?” Electronics<br />

61, no. 14 (August, 1988).<br />

Nields, Megan. “Bubble Memory Bursts into Niche Markets.” Mini-<br />

Micro Systems 20, no. 5 (May, 1987).


142<br />

Bullet train<br />

Bullet train<br />

The invention: An ultrafast passenger railroad system capable of<br />

moving passengers at speeds double or triple those of ordinary<br />

trains.<br />

The people behind the invention:<br />

Ikeda Hayato (1899-1965), Japanese prime minister from 1960 to<br />

1964, who pushed for the expansion of public expenditures<br />

Shinji Sogo (1901-1971), the president of the Japanese National<br />

Railways, the “father of the bullet train”<br />

Building a Faster Train<br />

By 1900, Japan had a world-class railway system, a logical result<br />

of the country’s dense population <strong>and</strong> the needs of its modernizing<br />

economy. After 1907, the government controlled the system<br />

through the Japanese National Railways (JNR). In 1938, JNR engineers<br />

first suggested the idea of a train that would travel 125 miles<br />

per hour from Tokyo to the southern city of Shimonoseki. Construction<br />

of a rapid train began in 1940 but was soon stopped because of<br />

World War II.<br />

The 311-mile railway between Tokyo <strong>and</strong> Osaka, the Tokaido<br />

Line, has always been the major line in Japan. By 1957, a business express<br />

along the line operated at an average speed of 57 miles per<br />

hour, but the double-track line was rapidly reaching its transport capacity.<br />

The JNR established two investigative committees to explore<br />

alternative solutions. In 1958, the second committee recommended<br />

the construction of a high-speed railroad on a separate double track,<br />

to be completed in time for the Tokyo Olympics of 1964. The Railway<br />

Technical Institute of the JNR concluded that it was feasible to<br />

design a line that would operate at an average speed of about 130<br />

miles per hour, cutting time for travel between Tokyo <strong>and</strong> Osaka<br />

from six hours to three hours.<br />

By 1962, about 17 miles of the proposed line were completed for<br />

test purposes. During the next two years, prototype trains were<br />

tested to correct flaws <strong>and</strong> make improvements in the design. The en-


tire project was completed on schedule in July, 1964, with total construction<br />

costs of more than $1 billion, double the original estimates.<br />

The Speeding Bullet<br />

Bullet train / 143<br />

Service on the Shinkansen, or New Trunk Line, began on October<br />

1, 1964, ten days before the opening of the Olympic Games.<br />

Commonly called the “bullet train” because of its shape <strong>and</strong> speed,<br />

the Shinkansen was an instant success with the public, both in Japan<br />

<strong>and</strong> abroad. As promised, the time required to travel between Tokyo<br />

<strong>and</strong> Osaka was cut in half. Initially, the system provided daily<br />

services of sixty trains consisting of twelve cars each, but the number<br />

of scheduled trains was almost doubled by the end of the year.<br />

The Shinkansen was able to operate at its unprecedented speed<br />

because it was designed <strong>and</strong> operated as an integrated system,<br />

making use of countless technological <strong>and</strong> scientific developments.<br />

Tracks followed the st<strong>and</strong>ard gauge of 56.5 inches, rather than the<br />

more narrow gauge common in Japan. For extra strength, heavy<br />

Japanese bullet trains. (PhotoDisc)


144 / Bullet train<br />

welded rails were attached directly onto reinforced concrete slabs.<br />

The minimum radius of a curve was 8,200 feet, except where sharper<br />

curves were m<strong>and</strong>ated by topography. In many ways similar to<br />

modern airplanes, the railway cars were made airtight in order to<br />

prevent ear discomfort caused by changes in pressure when trains<br />

enter tunnels.<br />

The Shinkansen trains were powered by electric traction motors,<br />

with four 185-kilowatt motors on each car—one motor attached to<br />

each axle. This design had several advantages: It provided an even<br />

distribution of axle load for reducing strain on the tracks; it allowed<br />

the application of dynamic brakes (where the motor was used for<br />

braking) on all axles; <strong>and</strong> it prevented the failure of one or two units<br />

from interrupting operation of the entire train. The 25,000-volt electrical<br />

current was carried by trolley wire to the cars, where it was<br />

rectified into a pulsating current to drive the motors.<br />

The Shinkansen system established a casualty-free record because<br />

of its maintenance policies combined with its computerized<br />

Centralized Traffic Control system. The control room at Tokyo Station<br />

was designed to maintain timely information about the location<br />

of all trains <strong>and</strong> the condition of all routes. Although train operators<br />

had some discretion in determining speed, automatic brakes<br />

also operated to ensure a safe distance between trains. At least once<br />

each month, cars were thoroughly inspected; every ten days, an inspection<br />

train examined the conditions of tracks, communication<br />

equipment, <strong>and</strong> electrical systems.<br />

Impact<br />

<strong>Public</strong> usage of the Tokyo-Osaka bullet train increased steadily<br />

because of the system’s high speed, comfort, punctuality, <strong>and</strong> superb<br />

safety record. Businesspeople were especially happy that the<br />

rapid service allowed them to make the round-trip without the necessity<br />

of an overnight stay, <strong>and</strong> continuing modernization soon allowed<br />

nonstop trains to make a one-way trip in two <strong>and</strong> one-half<br />

hours, requiring speeds of 160 miles per hour in some stretches. By<br />

the early 1970’s, the line was transporting a daily average of 339,000<br />

passengers in 240 trains, meaning that a train departed from Tokyo<br />

about every ten minutes.


The popularity of the Shinkansen system quickly resulted in dem<strong>and</strong>s<br />

for its extension into other densely populated regions. In<br />

1972, a 100-mile stretch between Osaka <strong>and</strong> Okayama was opened<br />

for service. By 1975, the line was further extended to Hakata on the<br />

isl<strong>and</strong> of Kyushu, passing through the Kammon undersea tunnel.<br />

The cost of this 244-mile stretch was almost $2.5 billion. In 1982,<br />

lines were completed from Tokyo to Niigata <strong>and</strong> from Tokyo to<br />

Morioka. By 1993, the system had grown to 1,134 miles of track.<br />

Since high usage made the system extremely profitable, the sale of<br />

the JNR to private companies in 1987 did not appear to produce adverse<br />

consequences.<br />

The economic success of the Shinkansen had a revolutionary effect<br />

on thinking about the possibilities of modern rail transportation,<br />

leading one authority to conclude that the line acted as “a<br />

savior of the declining railroad industry.” Several other industrial<br />

countries were stimulated to undertake large-scale railway projects;<br />

France, especially, followed Japan’s example by constructing highspeed<br />

electric railroads from Paris to Nice <strong>and</strong> to Lyon. By the mid-<br />

1980’s, there were experiments with high-speed trains based on<br />

magnetic levitation <strong>and</strong> other radical innovations, but it was not<br />

clear whether such designs would be able to compete with the<br />

Shinkansen model.<br />

See also Airplane; Atomic-powered ship; Diesel locomotive; Supersonic<br />

passenger plane.<br />

Further Reading<br />

Bullet train / 145<br />

French, Howard W. “Japan’s New Bullet Train Draws Fire.” New<br />

York Times (September 24, 2000).<br />

Frew, Tim. Locomotives: From the Steam Locomotive to the Bullet Train.<br />

New York: Mallard Press, 1990.<br />

Holley, David. “Faster Than a Speeding Bullet: High-Speed Trains<br />

Are Japan’s Pride, Subject of Debate.” Los Angeles Times (April 10,<br />

1994).<br />

O’Neill, Bill. “Beating the Bullet Train.” New Scientist 140, no. 1893<br />

(October 2, 1993).<br />

Raoul, Jean-Claude. “How High-Speed Trains Make Tracks.” Scientific<br />

American 277 (October, 1997).


146<br />

Buna rubber<br />

Buna rubber<br />

The invention: The first practical synthetic rubber product developed,<br />

Buna inspired the creation of other other synthetic substances<br />

that eventually replaced natural rubber in industrial applications.<br />

The people behind the invention:<br />

Charles de la Condamine (1701-1774), a French naturalist<br />

Charles Goodyear (1800-1860), an American inventor<br />

Joseph Priestley (1733-1804), an English chemist<br />

Charles Greville Williams (1829-1910), an English chemist<br />

A New Synthetic Rubber<br />

The discovery of natural rubber is often credited to the French<br />

scientist Charles de la Condamine, who, in 1736, sent the French<br />

Academy of Science samples of an elastic material used by Peruvian<br />

Indians to make balls that bounced. The material was primarily a<br />

curiosity until 1770, when Joseph Priestley, an English chemist, discovered<br />

that it rubbed out pencil marks, after which he called it<br />

“rubber.” Natural rubber, made from the sap of the rubber tree<br />

(Hevea brasiliensis), became important after Charles Goodyear discovered<br />

in 1830 that heating rubber with sulfur (a process called<br />

“vulcanization”) made it more elastic <strong>and</strong> easier to use. Vulcanized<br />

natural rubber came to be used to make raincoats, rubber b<strong>and</strong>s,<br />

<strong>and</strong> motor vehicle tires.<br />

Natural rubber is difficult to obtain (making one tire requires<br />

the amount of rubber produced by one tree in two years), <strong>and</strong> wars<br />

have often cut off supplies of this material to various countries.<br />

Therefore, efforts to manufacture synthetic rubber began in the<br />

late eighteenth century. Those efforts followed the discovery by<br />

English chemist Charles Greville Williams <strong>and</strong> others in the 1860’s<br />

that natural rubber was composed of thous<strong>and</strong>s of molecules of a<br />

chemical called isoprene that had been joined to form giant, necklace-like<br />

molecules. The first successful synthetic rubber, Buna,<br />

was patented by Germany’s I. G. Farben Industrie in 1926. The suc-


Buna rubber / 147<br />

cess of this rubber led to the development of many other synthetic<br />

rubbers, which are now used in place of natural rubber in many<br />

applications.<br />

Charles Goodyear<br />

It was an accident that finally showed Charles Goodyear<br />

(1800-1860) how to make rubber into a durable, practical material.<br />

For years he had been experimenting at home looking for<br />

ways to improve natural rubber—<strong>and</strong> producing stenches that<br />

drove his family <strong>and</strong> neighbors to distraction—when in 1839 he<br />

dropped a piece of rubber mixed with sulfur onto a hot stove.<br />

When he examined the charred specimen, he discovered it was<br />

not sticky, as hot natural rubber always is, <strong>and</strong> when he took it<br />

outside into the cold, it did not become brittle.<br />

The son of an inventor, Goodyear invented much<br />

more than his vulcanizing process for rubber. He also<br />

patented a spring-lever faucet, pontoon boat, hay fork,<br />

<strong>and</strong> air pump, but he was never successful in making<br />

money from his inventions. Owner of a hardware<br />

store, he went broke during a financial panic in 1830<br />

<strong>and</strong> had to spend time in debtor’s prison. He was<br />

never financially stable afterwards, often having to<br />

borrow money <strong>and</strong> sell his family’s belongings to<br />

support his experiments. And he had a large family—twelve<br />

children, of whom only half lived beyond childhood.<br />

Even vulcanized rubber did not make Goodyear’s fortune.<br />

He delayed patenting it until Thomas Hancock, an Englishman,<br />

replicated Goodyear’s method of vulcanizing <strong>and</strong> began producing<br />

rubber in Engl<strong>and</strong>. Goodyear sued <strong>and</strong> lost. Others stole<br />

his method, <strong>and</strong> although he won one large case, legal expenses<br />

took away most of the settlement. He borrowed more <strong>and</strong> more<br />

money to advertise his product, with some success. For example,<br />

Emperor Napoleon III awarded Goodyear the Cross of the<br />

Legion of Honor for his display at the 1851 Crystal Palace Exhibition<br />

in London. Nevertheless, Goodyear died deeply in debt.<br />

Despite all the imitators, vulcanized rubber remained associated<br />

with Goodyear. Thirty-eight years after he died, the<br />

world’s larger rubber manufacturer took his name for the company’s<br />

title.<br />

(Smithsonian Institution)


148 / Buna rubber<br />

From Erasers to Gas Pumps<br />

Natural rubber belongs to the group of chemicals called “polymers.”<br />

A polymer is a giant molecule that is made up of many simpler<br />

chemical units (“monomers”) that are attached chemically to<br />

form long strings. In natural rubber, the monomer is isoprene<br />

(dimethylbutadiene). The first efforts to make a synthetic rubber<br />

used the discovery that isoprene could be made <strong>and</strong> converted<br />

into an elastic polymer. The synthetic rubber that was created from<br />

isoprene was, however, inferior to natural rubber. The first Buna<br />

rubber, which was patented by I. G. Farben in 1926, was better, but it<br />

was still less than ideal. Buna rubber was made by polymerizing the<br />

monomer butadiene in the presence of sodium. The name Buna<br />

comes from the first two letters of the words “butadiene” <strong>and</strong> “natrium”<br />

(German for sodium). Natural <strong>and</strong> Buna rubbers are called<br />

homopolymers because they contain only one kind of monomer.<br />

The ability of chemists to make Buna rubber, along with its successful<br />

use, led to experimentation with the addition of other monomers<br />

to isoprene-like chemicals used to make synthetic rubber.<br />

Among the first great successes were materials that contained two<br />

alternating monomers; such materials are called “copolymers.” If<br />

the two monomers are designated A <strong>and</strong> B, part of a polymer molecule<br />

can be represented as (ABABABABABABABABAB). Numerous<br />

synthetic copolymers, which are often called “elastomers,” now<br />

replace natural rubber in applications where they have superior<br />

properties. All elastomers are rubbers, since objects made from<br />

them both stretch greatly when pulled <strong>and</strong> return quickly to their<br />

original shape when the tension is released.<br />

Two other well-known rubbers developed by I. G. Farben are the<br />

copolymers called Buna-N <strong>and</strong> Buna-S. These materials combine butadiene<br />

<strong>and</strong> the monomers acrylonitrile <strong>and</strong> styrene, respectively.<br />

Many modern motor vehicle tires are made of synthetic rubber that<br />

differs little from Buna-S rubber. This rubber was developed after<br />

the United States was cut off in the 1940’s, during World War II,<br />

from its Asian source of natural rubber. The solution to this problem<br />

was the development of a synthetic rubber industry based on GR-S<br />

rubber (government rubber plus styrene), which was essentially<br />

Buna-S rubber. This rubber is still widely used.


Buna-S rubber is often made by mixing butadiene <strong>and</strong> styrene in<br />

huge tanks of soapy water, stirring vigorously, <strong>and</strong> heating the mixture.<br />

The polymer contains equal amounts of butadiene <strong>and</strong> styrene<br />

(BSBSBSBSBSBSBSBS). When the molecules of the Buna-S polymer<br />

reach the desired size, the polymerization is stopped <strong>and</strong> the rubber<br />

is coagulated (solidified) chemically. Then, water <strong>and</strong> all the unused<br />

starting materials are removed, after which the rubber is dried <strong>and</strong><br />

shipped to various plants for use in tires <strong>and</strong> other products. The<br />

major difference between Buna-S <strong>and</strong> GR-S rubber is that the method<br />

of making GR-S rubber involves the use of low temperatures.<br />

Buna-N rubber is made in a fashion similar to that used for Buna-<br />

S, using butadiene <strong>and</strong> acrylonitrile. Both Buna-N <strong>and</strong> the related<br />

neoprene rubber, invented by Du Pont, are very resistant to gasoline<br />

<strong>and</strong> other liquid vehicle fuels. For this reason, they can be used in<br />

gas-pump hoses. All synthetic rubbers are vulcanized before they<br />

are used in industry.<br />

Impact<br />

Buna rubber / 149<br />

Buna rubber became the basis for the development of the other<br />

modern synthetic rubbers. These rubbers have special properties<br />

that make them suitable for specific applications. One developmental<br />

approach involved the use of chemically modified butadiene in<br />

homopolymers such as neoprene. Made of chloroprene (chlorobutadiene),<br />

neoprene is extremely resistant to sun, air, <strong>and</strong> chemicals.<br />

It is so widely used in machine parts, shoe soles, <strong>and</strong> hoses that<br />

more than 400 million pounds are produced annually.<br />

Another developmental approach involved copolymers that alternated<br />

butadiene with other monomers. For example, the successful<br />

Buna-N rubber (butadiene <strong>and</strong> acrylonitrile) has properties<br />

similar to those of neoprene. It differs sufficiently from neoprene,<br />

however, to be used to make items such as printing press rollers.<br />

About 200 million pounds of Buna-N are produced annually. Some<br />

4 billion pounds of the even more widely used polymer Buna-S/<br />

GR-S are produced annually, most of which is used to make tires.<br />

Several other synthetic rubbers have significant industrial applications,<br />

<strong>and</strong> efforts to make copolymers for still other purposes continue.


150 / Buna rubber<br />

See also Neoprene; Nylon; Orlon; Plastic; Polyester; Polyethylene;<br />

Polystyrene; Silicones; Teflon; Velcro.<br />

Further Reading<br />

Herbert, Vernon. Synthetic Rubber: A Project That Had to Succeed.<br />

Westport, Conn.: Greenwood Press, 1985.<br />

Mossman, S. T. I., <strong>and</strong> Peter John Turnbull Morris. The Development of<br />

Plastics. Cambridge: Royal Society of Chemistry, 1994.<br />

Von Hagen, Victor Wolfgang. South America Called Them: Explorations<br />

of the Great Naturalists, La Condamine, Humboldt, Darwin,<br />

Spruce. New York: A. A. Knopf, 1945.


CAD/CAM<br />

CAD/CAM<br />

The invention: Computer-Aided Design (CAD) <strong>and</strong> Computer-<br />

Aided Manufacturing (CAM) enhanced flexibility in engineering<br />

design, leading to higher quality <strong>and</strong> reduced time for manufacturing<br />

The people behind the invention:<br />

Patrick Hanratty, a General Motors Research Laboratory<br />

worker who developed graphics programs<br />

Jack St. Clair Kilby (1923- ), a Texas Instruments employee<br />

who first conceived of the idea of the integrated circuit<br />

Robert Noyce (1927-1990), an Intel Corporation employee who<br />

developed an improved process of manufacturing<br />

integrated circuits on microchips<br />

Don Halliday, an early user of CAD/CAM who created the<br />

Made-in-America car in only four months by using CAD<br />

<strong>and</strong> project management software<br />

Fred Borsini, an early user of CAD/CAM who demonstrated<br />

its power<br />

Summary of Event<br />

151<br />

Computer-Aided Design (CAD) is a technique whereby geometrical<br />

descriptions of two-dimensional (2-D) or three-dimensional (3-<br />

D) objects can be created <strong>and</strong> stored, in the form of mathematical<br />

models, in a computer system. Points, lines, <strong>and</strong> curves are represented<br />

as graphical coordinates. When a drawing is requested from<br />

the computer, transformations are performed on the stored data,<br />

<strong>and</strong> the geometry of a part or a full view from either a two- or a<br />

three-dimensional perspective is shown. CAD systems replace the<br />

tedious process of manual drafting, <strong>and</strong> computer-aided drawing<br />

<strong>and</strong> redrawing that can be retrieved when needed has improved<br />

drafting efficiency. A CAD system is a combination of computer<br />

hardware <strong>and</strong> software that facilitates the construction of geometric<br />

models <strong>and</strong>, in many cases, their analysis. It allows a wide variety of<br />

visual representations of those models to be displayed.


152 / CAD/CAM<br />

Computer-Aided Manufacturing (CAM) refers to the use of computers<br />

to control, wholly or partly, manufacturing processes. In<br />

practice, the term is most often applied to computer-based developments<br />

of numerical control technology; robots <strong>and</strong> flexible manufacturing<br />

systems (FMS) are included in the broader use of CAM<br />

systems. A CAD/CAM interface is envisioned as a computerized<br />

database that can be accessed <strong>and</strong> enriched by either design or manufacturing<br />

professionals during various stages of the product development<br />

<strong>and</strong> production cycle.<br />

In CAD systems of the early 1990’s, the ability to model solid objects<br />

became widely available. The use of graphic elements such as<br />

lines <strong>and</strong> arcs <strong>and</strong> the ability to create a model by adding <strong>and</strong> subtracting<br />

solids such as cubes <strong>and</strong> cylinders are the basic principles of<br />

CAD <strong>and</strong> of simulating objects within a computer. CAD systems enable<br />

computers to simulate both taking things apart (sectioning)<br />

<strong>and</strong> putting things together for assembly. In addition to being able<br />

to construct prototypes <strong>and</strong> store images of different models, CAD<br />

systems can be used for simulating the behavior of machines, parts,<br />

<strong>and</strong> components. These abilities enable CAD to construct models<br />

that can be subjected to nondestructive testing; that is, even before<br />

engineers build a physical prototype, the CAD model can be subjected<br />

to testing <strong>and</strong> the results can be analyzed. As another example,<br />

designers of printed circuit boards have the ability to test their<br />

circuits on a CAD system by simulating the electrical properties of<br />

components.<br />

During the 1950’s, the U.S. Air Force recognized the need for reducing<br />

the development time for special aircraft equipment. As a<br />

result, the Air Force commissioned the Massachusetts Institute of<br />

Technology to develop numerically controlled (NC) machines that<br />

were programmable. A workable demonstration of NC machines<br />

was made in 1952; this began a new era for manufacturing. As the<br />

speed of an aircraft increased, the cost of manufacturing also increased<br />

because of stricter technical requirements. This higher cost<br />

provided a stimulus for the further development of NC technology,<br />

which promised to reduce errors in design before the prototype<br />

stage.<br />

The early 1960’s saw the development of mainframe computers.<br />

Many industries valued computing technology for its speed <strong>and</strong> for


CAD/CAM / 153<br />

its accuracy in lengthy <strong>and</strong> tedious numerical operations in design,<br />

manufacturing, <strong>and</strong> other business functional areas. Patrick<br />

Hanratty, working for General Motors Research Laboratory, saw<br />

other potential applications <strong>and</strong> developed graphics programs for<br />

use on mainframe computers. The use of graphics in software aided<br />

the development of CAD/CAM, allowing visual representations of<br />

models to be presented on computer screens <strong>and</strong> printers.<br />

The 1970’s saw an important development in computer hardware,<br />

namely the development <strong>and</strong> growth of personal computers<br />

(PCs). Personal computers became smaller as a result of the development<br />

of integrated circuits. Jack St. Clair Kilby, working for Texas<br />

Instruments, first conceived of the integrated circuit; later, Robert<br />

Noyce, working for Intel Corporation, developed an improved process<br />

of manufacturing integrated circuits on microchips. Personal<br />

computers using these microchips offered both speed <strong>and</strong> accuracy<br />

at costs much lower than those of mainframe computers.<br />

Five companies offered integrated commercial computer-aided<br />

design <strong>and</strong> computer-aided manufacturing systems by the first half<br />

of 1973. Integration meant that both design <strong>and</strong> manufacturing<br />

were contained in one system. Of these five companies—Applicon,<br />

Computervision, Gerber Scientific, Manufacturing <strong>and</strong> Consulting<br />

Services (MCS), <strong>and</strong> United Computing—four offered turnkey systems<br />

exclusively. Turnkey systems provide design, development,<br />

training, <strong>and</strong> implementation for each customer (company) based<br />

on the contractual agreement; they are meant to be used as delivered,<br />

with no need for the purchaser to make significant adjustments<br />

or perform programming.<br />

The 1980’s saw a proliferation of mini- <strong>and</strong> microcomputers with<br />

a variety of platforms (processors) with increased speed <strong>and</strong> better<br />

graphical resolution. This made the widespread development of<br />

computer-aided design <strong>and</strong> computer-aided manufacturing possible<br />

<strong>and</strong> practical. Major corporations spent large research <strong>and</strong> development<br />

budgets developing CAD/CAM systems that would<br />

automate manual drafting <strong>and</strong> machine tool movements. Don Halliday,<br />

working for Truesports Inc., provided an early example of the<br />

benefits of CAD/CAM. He created the Made-in-America car in only<br />

four months by using CAD <strong>and</strong> project management software. In<br />

the late 1980’s, Fred Borsini, the president of Leap Technologies in


154 / CAD/CAM<br />

Michigan, brought various products to market in record time through<br />

the use of CAD/CAM.<br />

In the early 1980’s, much of the CAD/CAM industry consisted of<br />

software companies. The cost for a relatively slow interactive system<br />

in 1980 was close to $100,000. The late 1980’s saw the demise of<br />

minicomputer-based systems in favor of Unix work stations <strong>and</strong><br />

PCs based on 386 <strong>and</strong> 486 microchips produced by Intel. By the time<br />

of the International Manufacturing Technology show in September,<br />

1992, the industry could show numerous CAD/CAM innovations<br />

including tools, CAD/CAM models to evaluate manufacturability<br />

in early design phases, <strong>and</strong> systems that allowed use of the same<br />

data for a full range of manufacturing functions.<br />

Impact<br />

In 1990, CAD/CAM hardware sales by U.S. vendors reached<br />

$2.68 billion. In software alone, $1.42 billion worth of CAD/CAM<br />

products <strong>and</strong> systems were sold worldwide by U.S. vendors, according<br />

to International Data Corporation figures for 1990. CAD/<br />

CAM systems were in widespread use throughout the industrial<br />

world. Development lagged in advanced software applications,<br />

particularly in image processing, <strong>and</strong> in the communications software<br />

<strong>and</strong> hardware that ties processes together.<br />

A reevaluation of CAD/CAM systems was being driven by the<br />

industry trend toward increased functionality of computer-driven<br />

numerically controlled machines. Numerical control (NC) software<br />

enables users to graphically define the geometry of the parts in a<br />

product, develop paths that machine tools will follow, <strong>and</strong> exchange<br />

data among machines on the shop floor. In 1991, NC configuration<br />

software represented 86 percent of total CAM sales. In 1992,<br />

the market shares of the five largest companies in the CAD/CAM<br />

market were 29 percent for International Business Machines, 17 percent<br />

for Intergraph, 11 percent for Computervision, 9 percent for<br />

Hewlett-Packard, <strong>and</strong> 6 percent for Mentor Graphics.<br />

General Motors formed a joint venture with Ford <strong>and</strong> Chrysler to<br />

develop a common computer language in order to make the next<br />

generation of CAD/CAM systems easier to use. The venture was<br />

aimed particularly at problems that posed barriers to speeding up


CAD/CAM / 155<br />

the design of new automobiles. The three car companies all had sophisticated<br />

computer systems that allowed engineers to design<br />

parts on computers <strong>and</strong> then electronically transmit specifications<br />

to tools that make parts or dies.<br />

CAD/CAM technology was expected to advance on many fronts.<br />

As of the early 1990’s, different CAD/CAM vendors had developed<br />

systems that were often incompatible with one another, making it<br />

difficult to transfer data from one system to another. Large corporations,<br />

such as the major automakers, developed their own interfaces<br />

<strong>and</strong> network capabilities to allow different systems to communicate.<br />

Major users of CAD/CAM saw consolidation in the industry<br />

through the establishment of st<strong>and</strong>ards as being in their interests.<br />

Resellers of CAD/CAM products also attempted to redefine<br />

their markets. These vendors provide technical support <strong>and</strong> service<br />

to users. The sale of CAD/CAM products <strong>and</strong> systems offered substantial<br />

opportunities, since dem<strong>and</strong> remained strong. Resellers<br />

worked most effectively with small <strong>and</strong> medium-sized companies,<br />

which often were neglected by the primary sellers of CAD/CAM<br />

equipment because they did not generate a large volume of business.<br />

Some projections held that by 1995 half of all CAD/CAM systems<br />

would be sold through resellers, at a cost of $10,000 or less for<br />

each system. The CAD/CAM market thus was in the process of dividing<br />

into two markets: large customers (such as aerospace firms<br />

<strong>and</strong> automobile manufacturers) that would be served by primary<br />

vendors, <strong>and</strong> small <strong>and</strong> medium-sized customers that would be serviced<br />

by resellers.<br />

CAD will find future applications in marketing, the construction<br />

industry, production planning, <strong>and</strong> large-scale projects such as shipbuilding<br />

<strong>and</strong> aerospace. Other likely CAD markets include hospitals,<br />

the apparel industry, colleges <strong>and</strong> universities, food product<br />

manufacturers, <strong>and</strong> equipment manufacturers. As the linkage between<br />

CAD <strong>and</strong> CAM is enhanced, systems will become more productive.<br />

The geometrical data from CAD will be put to greater use<br />

by CAM systems.<br />

CAD/CAM already had proved that it could make a big difference<br />

in productivity <strong>and</strong> quality. Customer orders could be changed<br />

much faster <strong>and</strong> more accurately than in the past, when a change<br />

could require a manual redrafting of a design. Computers could do


156 / CAD/CAM<br />

automatically in minutes what once took hours manually. CAD/<br />

CAM saved time by reducing, <strong>and</strong> in some cases eliminating, human<br />

error. Many flexible manufacturing systems (FMS) had machining<br />

centers equipped with sensing probes to check the accuracy<br />

of the machining process. These self-checks can be made part of numerical<br />

control (NC) programs. With the technology of the early<br />

1990’s, some experts estimated that CAD/CAM systems were in<br />

many cases twice as productive as the systems they replaced; in the<br />

long run, productivity is likely to improve even more, perhaps up to<br />

three times that of older systems or even higher. As costs for CAD/<br />

CAM systems concurrently fall, the investment in a system will be<br />

recovered more quickly. Some analysts estimated that by the mid-<br />

1990’s, the recovery time for an average system would be about<br />

three years.<br />

Another frontier in the development of CAD/CAM systems is<br />

expert (or knowledge-based) systems, which combine data with a<br />

human expert’s knowledge, expressed in the form of rules that the<br />

computer follows. Such a system will analyze data in a manner<br />

mimicking intelligence. For example, a 3-D model might be created<br />

from st<strong>and</strong>ard 2-D drawings. Expert systems will likely play a<br />

pivotal role in CAM applications. For example, an expert system<br />

could determine the best sequence of machining operations to produce<br />

a component.<br />

Continuing improvements in hardware, especially increased<br />

speed, will benefit CAD/CAM systems. Software developments,<br />

however, may produce greater benefits. Wider use of CAD/CAM<br />

systems will depend on the cost savings from improvements in<br />

hardware <strong>and</strong> software as well as on the productivity of the systems<br />

<strong>and</strong> the quality of their product. The construction, apparel,<br />

automobile, <strong>and</strong> aerospace industries have already experienced<br />

increases in productivity, quality, <strong>and</strong> profitability through the use<br />

of CAD/CAM. A case in point is Boeing, which used CAD from<br />

start to finish in the design of the 757.<br />

See also Differential analyzer; Mark I calculator; Personal computer;<br />

SAINT; Virtual machine; Virtual reality.


Further Reading<br />

CAD/CAM / 157<br />

Groover, Mikell P., <strong>and</strong> Emory W. Zimmers, Jr. CAD/CAM: Computer-Aided<br />

Design <strong>and</strong> Manufacturing. Englewood Cliffs, N.J.:<br />

Prentice-Hall, 1984.<br />

Jurgen, Ronald K. Computers <strong>and</strong> Manufacturing Productivity. New<br />

York: Institute of Electrical <strong>and</strong> Electronics Engineers, 1987.<br />

McMahon, Chris, <strong>and</strong> Jimmie Browne. CAD/CAM: From Principles to<br />

Practice. Reading, Mass.: Addison-Wesley, 1993.<br />

_____. CAD/CAM: Principles, Practice, <strong>and</strong> Manufacturing Management.<br />

2d ed. Harlow, Engl<strong>and</strong>: Addison-Wesley, 1998.<br />

Medl<strong>and</strong>, A. J., <strong>and</strong> Piers Burnett. CAD/CAM in Practice. New York:<br />

John Wiley & Sons, 1986.


158<br />

Carbon dating<br />

Carbon dating<br />

The invention: A technique that measures the radioactive decay of<br />

carbon 14 in organic substances to determine the ages of artifacts<br />

as old as ten thous<strong>and</strong> years.<br />

The people behind the invention:<br />

Willard Frank Libby (1908-1980), an American chemist who won<br />

the 1960 Nobel Prize in Chemistry<br />

Charles Wesley Ferguson (1922-1986), a scientist who<br />

demonstrated that carbon 14 dates before 1500 b.c. needed to<br />

be corrected<br />

One in a Trillion<br />

Carbon dioxide in the earth’s atmosphere contains a mixture of<br />

three carbon isotopes (isotopes are atoms of the same element that<br />

contain different numbers of neutrons), which occur in the following<br />

percentages: about 99 percent carbon 12, about 1 percent carbon<br />

13, <strong>and</strong> approximately one atom in a trillion of radioactive carbon<br />

14. Plants absorb carbon dioxide from the atmosphere during photosynthesis,<br />

<strong>and</strong> then animals eat the plants, so all living plants <strong>and</strong><br />

animals contain a small amount of radioactive carbon.<br />

When a plant or animal dies, its radioactivity slowly decreases as<br />

the radioactive carbon 14 decays. The time it takes for half of any radioactive<br />

substance to decay is known as its “half-life.” The half-life<br />

for carbon 14 is known to be about fifty-seven hundred years. The<br />

carbon 14 activity will drop to one-half after one half-life, onefourth<br />

after two half-lives, one-eighth after three half-lives, <strong>and</strong> so<br />

forth. After ten or twenty half-lives, the activity becomes too low to<br />

be measurable. Coal <strong>and</strong> oil, which were formed from organic matter<br />

millions of years ago, have long since lost any carbon 14 activity.<br />

Wood samples from an Egyptian tomb or charcoal from a prehistoric<br />

fireplace a few thous<strong>and</strong> years ago, however, can be dated with<br />

good reliability from the leftover radioactivity.<br />

In the 1940’s, the properties of radioactive elements were still<br />

being discovered <strong>and</strong> were just beginning to be used to solve problems.<br />

Scientists still did not know the half-life of carbon 14, <strong>and</strong> ar-


chaeologists still depended mainly on historical evidence to determine<br />

the ages of ancient objects.<br />

In early 1947, Willard Frank Libby started a crucial experiment in<br />

testing for radioactive carbon. He decided to test samples of methane<br />

gas from two different sources. One group of samples came<br />

from the sewage disposal plant at Baltimore, Maryl<strong>and</strong>, which was<br />

rich in fresh organic matter. The other sample of methane came from<br />

an oil refinery, which should have contained only ancient carbon<br />

from fossils whose radioactivity should have completely decayed.<br />

The experimental results confirmed Libby’s suspicions: The methane<br />

from fresh sewage was radioactive, but the methane from oil<br />

was not. Evidently, radioactive carbon was present in fresh organic<br />

material, but it decays away eventually.<br />

Tree-Ring Dating<br />

Carbon dating / 159<br />

In order to establish the validity of radiocarbon dating, Libby analyzed<br />

known samples of varying ages. These included tree-ring<br />

samples from the years 575 <strong>and</strong> 1075 <strong>and</strong> one redwood from 979<br />

b.c.e., as well as artifacts from Egyptian tombs going back to about<br />

3000 b.c.e. In 1949, he published an article in the journal Science that<br />

contained a graph comparing the historical ages <strong>and</strong> the measured<br />

radiocarbon ages of eleven objects. The results were accurate within<br />

10 percent, which meant that the general method was sound.<br />

The first archaeological object analyzed by carbon dating, obtained<br />

from the Metropolitan Museum of Art in New York, was a<br />

piece of cypress wood from the tomb of King Djoser of Egypt. Based<br />

on historical evidence, the age of this piece of wood was about fortysix<br />

hundred years. A small sample of carbon obtained from this<br />

wood was deposited on the inside of Libby’s radiation counter, giving<br />

a count rate that was about 40 percent lower than that of modern<br />

organic carbon. The resulting age of the wood calculated from its residual<br />

radioactivity was about thirty-eight hundred years, a difference<br />

of eight hundred years. Considering that this was the first object<br />

to be analyzed, even such a rough agreement with the historic<br />

age was considered to be encouraging.<br />

The validity of radiocarbon dating depends on an important assumption—namely,<br />

that the abundance of carbon 14 in nature has


160 / Carbon dating<br />

Willard Frank Libby<br />

Born in 1908, Willard Frank Libby came from a family of<br />

farmers in Gr<strong>and</strong> View, Colorado. They moved to Sebastopol,<br />

California, where Libby went through public school. He entered<br />

the University of California, Berkeley, in 1927, earning a<br />

bachelor of science degree in 1931 <strong>and</strong> a doctorate in 1933. He<br />

stayed on at Berkeley as an instructor of chemistry until he won<br />

the first of his three Guggenheim Fellowships in 1941. He<br />

moved to Princeton University to study, but World War II cut<br />

short his fellowship. Instead, he joined the Manhattan Project,<br />

helping design the atomic bomb at Columbia University’s Division<br />

of War Research.<br />

After the war Libby became a professor of chemistry at the<br />

University of Chicago, where he conducted his research on carbon-14<br />

dating. A leading expert in radiochemistry, he also investigated<br />

isotope tracers <strong>and</strong> the effects of fallout. However,<br />

his career saw as much public service as research. In 1954 President<br />

Dwight Eisenhower appointed him to the Atomic Energy<br />

Commission as its first chemist, <strong>and</strong> Libby directed the administration’s<br />

international Atoms for Peace program. He resigned<br />

in 1959 to take an appointment at the University of California,<br />

Los Angeles, as professor of chemistry <strong>and</strong> then in 1962 as director<br />

of the Institute of Geophysics <strong>and</strong> Planetary Physics, a<br />

position he held until he died in 1980.<br />

Libby received the Nobel Prize in Chemistry in 1960 for developing<br />

carbon-14 dating. Among his many other honors were<br />

the American Chemical Society’s Willard Gibbs Award in 1958,<br />

the Albert Einstein Medal in 1959, <strong>and</strong> the Day Medal of the<br />

Geological Society of America in 1961. He was a member of the<br />

Advisory Board of the Guggenheim Memorial Foundation, the<br />

Office of Civil <strong>and</strong> Defense Mobilization, the National Science<br />

Foundation’s General Commission on Science, <strong>and</strong> the<br />

Academic Institution <strong>and</strong> also a director of Douglas Aircraft<br />

Company.<br />

been constant for many thous<strong>and</strong>s of years. If carbon 14 was less<br />

abundant at some point in history, organic samples from that era<br />

would have started with less radioactivity. When analyzed today,<br />

their reduced activity would make them appear to be older than<br />

they really are.


Charles Wesley Ferguson from the Tree-Ring Research Laboratory<br />

at the University of Arizona tackled this problem. He measured<br />

the age of bristlecone pine trees both by counting the rings <strong>and</strong> by<br />

using carbon 14 methods. He found that carbon 14 dates before<br />

1500 b.c.e. needed to be corrected. The results show that radiocarbon<br />

dates are older than tree-ring counting dates by as much as several<br />

hundred years for the oldest samples. He knew that the number<br />

of tree rings had given him the correct age of the pines, because trees<br />

accumulate one ring of growth for every year of life. Apparently, the<br />

carbon 14 content in the atmosphere has not been constant. Fortunately,<br />

tree-ring counting gives reliable dates that can be used to<br />

correct radiocarbon measurements back to about 6000 b.c.e.<br />

Impact<br />

Carbon dating / 161<br />

Some interesting samples were dated by Libby’s group. The<br />

Dead Sea Scrolls had been found in a cave by an Arab shepherd in<br />

1947, but some Bible scholars at first questioned whether they were<br />

genuine. The linen wrapping from the Book of Isaiah was tested for<br />

carbon 14, giving a date of 100 b.c.e., which helped to establish its<br />

authenticity. Human hair from an Egyptian tomb was determined<br />

to be nearly five thous<strong>and</strong> years old. Well-preserved s<strong>and</strong>als from a<br />

cave in eastern Oregon were determined to be ninety-three hundred<br />

years old. A charcoal sample from a prehistoric site in western<br />

South Dakota was found to be about seven thous<strong>and</strong> years old.<br />

The Shroud of Turin, located in Turin, Italy, has been a controversial<br />

object for many years. It is a linen cloth, more than four meters<br />

long, which shows the image of a man’s body, both front <strong>and</strong> back.<br />

Some people think it may have been the burial shroud of Jesus<br />

Christ after his crucifixion. A team of scientists in 1978 was permitted<br />

to study the shroud, using infrared photography, analysis of<br />

possible blood stains, microscopic examination of the linen fibers,<br />

<strong>and</strong> other methods. The results were ambiguous. A carbon 14 test<br />

was not permitted because it would have required cutting a piece<br />

about the size of a h<strong>and</strong>kerchief from the shroud.<br />

A new method of measuring carbon 14 was developed in the late<br />

1980’s. It is called “accelerator mass spectrometry,” or AMS. Unlike<br />

Libby’s method, it does not count the radioactivity of carbon. In-


162 / Carbon dating<br />

stead, a mass spectrometer directly measures the ratio of carbon 14<br />

to ordinary carbon. The main advantage of this method is that the<br />

sample size needed for analysis is about a thous<strong>and</strong> times smaller<br />

than before. The archbishop of Turin permitted three laboratories<br />

with the appropriate AMS apparatus to test the shroud material.<br />

The results agreed that the material was from the fourteenth century,<br />

not from the time of Christ. The figure on the shroud may be a<br />

watercolor painting on linen.<br />

Since Libby’s pioneering experiments in the late 1940’s, carbon<br />

14 dating has established itself as a reliable dating technique for archaeologists<br />

<strong>and</strong> cultural historians. Further improvements are expected<br />

to increase precision, to make it possible to use smaller samples,<br />

<strong>and</strong> to extend the effective time range of the method back to<br />

fifty thous<strong>and</strong> years or earlier.<br />

See also Atomic clock; Geiger counter; Richter scale.<br />

Further Reading<br />

Goldberg, Paul, Vance T. Holliday, <strong>and</strong> C. Reid Ferring. Earth Sciences<br />

<strong>and</strong> Archaeology. New York: Kluwer Academic Plenum, 2001.<br />

Libby, Willard Frank. “Radiocarbon Dating” [Nobel lecture]. In<br />

Chemistry, 1942-1962. River Edge, N.J.: World Scientific, 1999.<br />

Lowe, John J. Radiocarbon Dating: Recent Applications <strong>and</strong> Future Potential.<br />

New York: John Wiley <strong>and</strong> Sons, 1996.


Cassette recording<br />

Cassette recording<br />

The invention: Self-contained system making it possible to record<br />

<strong>and</strong> repeatedly play back sound without having to thread tape<br />

through a machine.<br />

The person behind the invention:<br />

Fritz Pfleumer, a German engineer whose work on audiotapes<br />

paved the way for audiocassette production<br />

Smaller Is Better<br />

163<br />

The introduction of magnetic audio recording tape in 1929 was<br />

met with great enthusiasm, particularly in the entertainment industry,<br />

<strong>and</strong> specifically among radio broadcasters. Although somewhat<br />

practical methods for recording <strong>and</strong> storing sound for later playback<br />

had been around for some time, audiotape was much easier to<br />

use, store, <strong>and</strong> edit, <strong>and</strong> much less expensive to produce.<br />

It was Fritz Pfleumer, a German engineer, who in 1929 filed the<br />

first audiotape patent. His detailed specifications indicated that<br />

tape could be made by bonding a thin coating of oxide to strips of either<br />

paper or film. Pfleumer also suggested that audiotape could be<br />

attached to filmstrips to provide higher-quality sound than was<br />

available with the film sound technologies in use at that time. In<br />

1935, the German electronics firm AEG produced a reliable prototype<br />

of a record-playback machine based on Pfleumer’s idea. By<br />

1947, the American company 3M had refined the concept to the<br />

point where it was able to produce a high-quality tape using a plastic-based<br />

backing <strong>and</strong> red oxide. The tape recorded <strong>and</strong> reproduced<br />

sound with a high degree of clarity <strong>and</strong> dynamic range <strong>and</strong> would<br />

soon become the st<strong>and</strong>ard in the industry.<br />

Still, the tape was sold <strong>and</strong> used in a somewhat inconvenient<br />

open-reel format. The user had to thread it through a machine <strong>and</strong><br />

onto a take-up reel. This process was somewhat cumbersome <strong>and</strong><br />

complicated for the layperson. For many years, sound-recording<br />

technology remained a tool mostly for professionals.<br />

In 1963, the first audiocassette was introduced by the Nether-


164 / Cassette recording<br />

l<strong>and</strong>s-based Philips NV company. This device could be inserted into<br />

a machine without threading. Rewind <strong>and</strong> fast-forward were faster,<br />

<strong>and</strong> it made no difference where the tape was stopped prior to the<br />

ejection of the cassette. By contrast, open-reel audiotape required<br />

that the tape be wound fully onto one or the other of the two reels<br />

before it could be taken off the machine.<br />

Technical advances allowed the cassette tape to be much narrower<br />

than the tape used in open reels <strong>and</strong> also allowed the tape<br />

speed to be reduced without sacrificing sound quality. Thus, the<br />

cassette was easier to carry around, <strong>and</strong> more sound could be recorded<br />

on a cassette tape. In addition, the enclosed cassette decreased<br />

wear <strong>and</strong> tear on the tape <strong>and</strong> protected it from contamination.<br />

Creating a Market<br />

One of the most popular uses for audiocassettes was to record<br />

music from radios <strong>and</strong> other audio sources for later playback. During<br />

the 1970’s, many radio stations developed “all music” formats<br />

in which entire albums were often played without interruption.<br />

That gave listeners an opportunity to record the music for later<br />

playback. At first, the music recording industry complained about<br />

this practice, charging that unauthorized recording of music from<br />

the radio was a violation of copyright laws. Eventually, the issue<br />

died down as the same companies began to recognize this new, untapped<br />

market for recorded music on cassette.<br />

Audiocassettes, all based on the original Philips design, were being<br />

manufactured by more than sixty companies within only a few<br />

years of their introduction. In addition, spin-offs of that design were<br />

being used in many specialized applications, including dictation,<br />

storage of computer information, <strong>and</strong> surveillance. The emergence<br />

of videotape resulted in a number of formats for recording <strong>and</strong><br />

playing back video based on the same principle. Although each is<br />

characterized by different widths of tape, each uses the same technique<br />

for tape storage <strong>and</strong> transport.<br />

The cassette has remained a popular means of storing <strong>and</strong> retrieving<br />

information on magnetic tape for more than a quarter of a<br />

century. During the early 1990’s, digital technologies such as audio<br />

CDs (compact discs) <strong>and</strong> the more advanced CD-ROM (compact


discs that reproduce sound, text, <strong>and</strong> images via computer) were beginning<br />

to store information in revolutionary new ways. With the<br />

development of this increasingly sophisticated technology, need for<br />

the audiocassette, once the most versatile, reliable, portable, <strong>and</strong><br />

economical means of recording, storing, <strong>and</strong> playing-back sound,<br />

became more limited.<br />

Consequences<br />

Cassette recording / 165<br />

The cassette represented a new level of convenience for the audiophile,<br />

resulting in a significant increase in the use of recording<br />

technology in all walks of life. Even small children could operate<br />

cassette recorders <strong>and</strong> players, which led to their use in schools for a<br />

variety of instructional tasks <strong>and</strong> in the home for entertainment. The<br />

recording industry realized that audiotape cassettes would allow<br />

consumers to listen to recorded music in places where record players<br />

were impractical: in automobiles, at the beach, even while camping.<br />

The industry also saw the need for widespread availability of<br />

music <strong>and</strong> information on cassette tape. It soon began distributing<br />

albums on audiocassette in addition to the long-play vinyl discs,<br />

<strong>and</strong> recording sales increased substantially. This new technology<br />

put recorded music into automobiles for the first time, again resulting<br />

in a surge in sales for recorded music. Eventually, information,<br />

including language instruction <strong>and</strong> books-on-tape, became popular<br />

commuter fare.<br />

With the invention of the microchip, audiotape players became<br />

available in smaller <strong>and</strong> smaller sizes, making them truly portable.<br />

Audiocassettes underwent another explosion in popularity during<br />

the early 1980’s, when the Sony Corporation introduced the<br />

Walkman, an extremely compact, almost weightless cassette player<br />

that could be attached to clothing <strong>and</strong> used with lightweight earphones<br />

virtually anywhere. At the same time, cassettes were suddenly<br />

being used with microcomputers for backing up magnetic<br />

data files.<br />

Home video soon exploded onto the scene, bringing with it new<br />

applications for cassettes. As had happened with audiotape, video<br />

camera-recorder units, called “camcorders,” were miniaturized to<br />

the point where 8-millimeter videocassettes capable of recording up


166 / Cassette recording<br />

to 90 minutes of live action <strong>and</strong> sound were widely available. These<br />

cassettes closely resembled the audiocassette first introduced in<br />

1963.<br />

See also Compact disc; Dolby noise reduction; Electronic synthesizer;<br />

FM radio; Transistor radio; Walkman cassette player.<br />

Further Reading<br />

Miller, Christopher. “The One Hundred Greatest <strong>Inventions</strong>: Audio<br />

<strong>and</strong> Video.” Popular Science 254, no. 4 (April, 1999).<br />

Praag, Phil van. Evolution of the Audio Recorder. Waukesha, Wis.: EC<br />

Designs, 1997.<br />

Stark, Craig. “Thirty Five Years of Tape Recording.” Stereo Review 58<br />

(September, 1993).


CAT scanner<br />

CAT scanner<br />

The invention: A technique that collects X-ray data from solid,<br />

opaque masses such as human bodies <strong>and</strong> uses a computer to<br />

construct a three-dimensional image.<br />

The people behind the invention:<br />

Godfrey Newbold Hounsfield (1919- ), an English<br />

electronics engineer who shared the 1979 Nobel Prize in<br />

Physiology or Medicine<br />

Allan M. Cormack (1924-1998), a South African-born American<br />

physicist who shared the 1979 Nobel Prize in Physiology or<br />

Medicine<br />

James Ambrose, an English radiologist<br />

A Significant Merger<br />

167<br />

Computerized axial tomography (CAT) is a technique that collects<br />

X-ray data from an opaque, solid mass such as a human body<br />

<strong>and</strong> uses a sophisticated computer to assemble those data into a<br />

three-dimensional image. This sophisticated merger of separate<br />

technologies led to another name for CAT, computer-assisted tomography<br />

(it came to be called computed tomography, or CT). CAT<br />

is a technique of medical radiology, an area of medicine that began<br />

after the German physicist Wilhelm Conrad Röntgen’s 1895 discovery<br />

of the high-energy electromagnetic radiations he named “X<br />

rays.” Röntgen <strong>and</strong> others soon produced X-ray images of parts of<br />

the human body, <strong>and</strong> physicians were quick to learn that these images<br />

were valuable diagnostic aids.<br />

In the late 1950’s <strong>and</strong> early 1960’s, Allan M. Cormack, a physicist<br />

at Tufts University in Massachusetts, pioneered a mathematical<br />

method for obtaining detailed X-ray absorption patterns in opaque<br />

samples meant to model biological samples. His studies used narrow<br />

X-ray beams that were passed through samples at many different angles.<br />

Because the technique probed test samples from many different<br />

points of reference, it became possible—by using the proper mathematics—to<br />

reconstruct the interior structure of a thin slice of the object<br />

being studied.


168 / CAT scanner<br />

Cormack published his data but received almost no recognition<br />

because computers that could analyze the data in an effective fashion<br />

had not yet been developed. Nevertheless, X-ray tomography—<br />

the process of using X-rays to produce detailed images of thin<br />

sections of solid objects—had been born. It remained for Godfrey<br />

Newbold Hounsfield of Engl<strong>and</strong>’s Electrical <strong>and</strong> Musical Instruments<br />

(EMI) Limited (independently, <strong>and</strong> reportedly with no<br />

knowledge of Cormack’s work) to design the first practical CAT<br />

scanner.<br />

A Series of Thin Slices<br />

Hounsfield, like Cormack, realized that X-ray tomography was<br />

the most practical approach to developing a medical body imager. It<br />

could be used to divide any three-dimensional object into a series of<br />

thin slices that could be reconstructed into images by using appropriate<br />

computers. Hounsfield developed another mathematical approach<br />

to the method. He estimated that the technique would make<br />

possible the very accurate reconstruction of images of thin body sections<br />

with a sensitivity well above that of the X-ray methodology<br />

then in use. Moreover, he proposed that his method would enable<br />

Medical technicians studying CAT scan results. (PhotoDisc)


Godfrey Newbold Hounsfield<br />

CAT scanner / 169<br />

On his family farm outside Newark, Nottinghamshire, Engl<strong>and</strong>,<br />

Godfrey Newbold Hounsfield (born 1919), the youngest<br />

of five children, was usually left to his own devices. The farm,<br />

he later wrote, offered an infinite variety of diversions, <strong>and</strong> his<br />

favorites were the many mechanical <strong>and</strong> electrical gadgets. By<br />

his teen years, he was making his own gadgets, such as an electrical<br />

recording machine, <strong>and</strong> experimenting with homemade<br />

gliders <strong>and</strong> water-propelled rockets. All these childhood projects<br />

taught him the fundamentals of practical reasoning.<br />

During World War II he joined the Royal Air Force, where<br />

his talent with gadgets got him a position as an instructor at the<br />

school for radio mechanics. There, on his own, he built his an oscilloscope<br />

<strong>and</strong> demonstration equipment. This initiative caught<br />

the eye of a high-ranking officer, who after the war arranged a<br />

scholarship so that Hounsfield could attend the Faraday Electrical<br />

Engineering College in London. Upon graduating in 1951,<br />

he took a research position with Electrical <strong>and</strong> Musical Instruments,<br />

Limited (EMI). His first assignments involved radar <strong>and</strong><br />

guided weapons, but he also developed an interest in computers<br />

<strong>and</strong> in 1958 led the design team that put together Engl<strong>and</strong>’s<br />

first all-transistor computer, the EMIDEC 1100. This experience,<br />

in turn, prepared him to follow through on his idea for computed<br />

tomography, which came to him in 1967.<br />

EMI released its first CT scanner in 1971, <strong>and</strong> it so impressed<br />

the medical world that in 1979 Hounsfield <strong>and</strong> Allan M. Cormack<br />

shared the Nobel Prize in Physiology or Medicine for the<br />

invention. Hounsfield, who continued to work on improved<br />

computed tomography <strong>and</strong> other diagnostic imagining techniques,<br />

was knighted in 1981.<br />

researchers <strong>and</strong> physicians to distinguish between normal <strong>and</strong> diseased<br />

tissue. Hounsfield was correct about that.<br />

The prototype instrument that Hounsfield developed was quite<br />

slow, requiring nine days to scan an object. Soon, he modified the<br />

scanner so that its use took only nine hours, <strong>and</strong> he obtained successful<br />

tomograms of preserved human brains <strong>and</strong> the fresh brains<br />

of cattle. The further development of the CAT scanner then pro-


170 / CAT scanner<br />

ceeded quickly, yielding an instrument that required four <strong>and</strong> onehalf<br />

minutes to gather tomographic data <strong>and</strong> twenty minutes to<br />

produce the tomographic image.<br />

In late 1971, the first clinical CAT scanner was installed at Atkinson<br />

Morley’s Hospital in Wimbledon, Engl<strong>and</strong>. By early 1972,<br />

the first patient, a woman with a suspected brain tumor, had been<br />

examined, <strong>and</strong> the resultant tomogram identified a dark, circular<br />

cyst in her brain. Additional data collection from other patients<br />

soon validated the technique. Hounsfield <strong>and</strong> EMI patented the<br />

CAT scanner in 1972, <strong>and</strong> the findings were reported at that year’s<br />

annual meeting of the British Institute of Radiology.<br />

Hounsfield published a detailed description of the instrument in<br />

1973. Hounsfield’s clinical collaborator, James Ambrose, published<br />

on the clinical aspects of the technique. Neurologists all around the<br />

world were ecstatic about the new tool that allowed them to locate<br />

tissue abnormalities with great precision.<br />

The CAT scanner consisted of an X-ray generator, a scanner unit<br />

composed of an X-ray tube <strong>and</strong> a detector in a circular chamber<br />

about which they could be rotated, a computer that could process<br />

all the data obtained, <strong>and</strong> a cathode-ray tube on which tomograms<br />

were viewed. To produce tomograms, the patient was placed on a<br />

couch, head inside the scanner chamber, <strong>and</strong> the emitter-detector<br />

was rotated 1 degree at a time. At each position, 160 readings were<br />

taken, converted to electrical signals, <strong>and</strong> fed into the computer. In<br />

the 180 degrees traversed, 28,800 readings were taken <strong>and</strong> processed.<br />

The computer then converted the data into a tomogram (a<br />

cross-sectional representation of the brain that shows the differences<br />

in tissue density). A Polaroid picture of the tomogram was<br />

then taken <strong>and</strong> interpreted by the physician in charge.<br />

Consequences<br />

Many neurologists agree that CAT is the most important method<br />

developed in the twentieth century to facilitate diagnosis of disorders<br />

of the brain. Even the first scanners could distinguish between<br />

brain tumors <strong>and</strong> blood clots <strong>and</strong> help physicians to diagnose a variety<br />

of brain-related birth defects. In addition, the scanners are believed<br />

to have saved many lives by allowing physicians to avoid


the dangerous exploratory brain surgery once required in many<br />

cases <strong>and</strong> by replacing more dangerous techniques, such as pneumoencephalography,<br />

which required a physician to puncture the<br />

head for diagnostic purposes.<br />

By 1975, improvements, including quicker reaction time <strong>and</strong><br />

more complex emitter-detector systems, made it possible for EMI to<br />

introduce full-body CAT scanners to the world market. Then it became<br />

possible to examine other parts of the body—including the<br />

lungs, the heart, <strong>and</strong> the abdominal organs—for cardiovascular<br />

problems, tumors, <strong>and</strong> other structural health disorders. The technique<br />

became so ubiquitous that many departments of radiology<br />

changed their names to departments of medical imaging.<br />

The use of CAT scanners has not been problem-free. Part of<br />

the reason for this is the high cost of the devices—ranging from<br />

about $300,000 for early models to $1 million for modern instruments—<strong>and</strong><br />

resultant claims by consumer advocacy groups that<br />

the scanners are unnecessarily expensive toys for physicians.<br />

Still, CAT scanners have become important everyday diagnostic<br />

tools in many areas of medicine. Furthermore, continuation of the<br />

efforts of Hounsfield <strong>and</strong> others has led to more improvements of<br />

CAT scanners <strong>and</strong> to the use of nonradiologic nuclear magnetic resonance<br />

imaging in such diagnoses.<br />

See also Amniocentesis; Electrocardiogram; Electroencephalogram;<br />

Mammography; Nuclear magnetic resonance; Pap test; Ultrasound;<br />

X-ray image intensifier.<br />

Further Reading<br />

CAT scanner / 171<br />

Gambarelli, J. Computerized Axial Tomography: An Anatomic Atlas of<br />

Serial Sections of the Human Body: Anatomy—Radiology—Scanner.<br />

New York: Springer Verlag, 1977.<br />

Raju, Tones N. K. “The Nobel Chronicles.” Lancet 354, no. 9190 (November<br />

6, 1999).<br />

Thomas, Robert McG., Jr. “Allan Cormack, Seventy Four, Nobelist<br />

Who Helped Invent CAT Scan.” New York Times (May 9, 1998).


172<br />

Cell phone<br />

Cell phone<br />

The invention: Mobile telephone system controlled by computers<br />

to use a region’s radio frequencies, or channels, repeatedly,<br />

thereby accommodating large numbers of users.<br />

The people behind the invention:<br />

William Oliver Baker (1915- ), the president of Bell<br />

Laboratories<br />

Richard H. Fefrenkiel, the head of the mobile systems<br />

engineering department at Bell<br />

The First Radio Telephones<br />

The first recorded attempt to use radio technology to provide direct<br />

access to a telephone system took place in 1920. It was not until<br />

1946, however, that Bell Telephone established the first such commercial<br />

system in St. Louis. The system had a number of disadvantages;<br />

users had to contact an operator who did the dialing <strong>and</strong> the<br />

connecting, <strong>and</strong> the use of a single radio frequency prevented simultaneous<br />

talking <strong>and</strong> listening. In 1949, a system was developed<br />

that used two radio frequencies (a “duplex pair”), permitting both<br />

the mobile unit <strong>and</strong> the base station to transmit <strong>and</strong> receive simultaneously<br />

<strong>and</strong> making a more normal sort of telephone conversation<br />

possible. This type of service, known as Mobile Telephone Service<br />

(MTS), was the norm in the field for many years.<br />

The history of MTS is one of continuously increasing business usage.<br />

The development of the transistor made possible the design <strong>and</strong><br />

manufacture of reasonably light, compact, <strong>and</strong> reliable equipment,<br />

but the expansion of MTS was slowed by the limited number of radio<br />

frequencies; there is nowhere near enough space on the radio spectrum<br />

for each user to have a separate frequency. In New York City, for<br />

example, New York Telephone Company was limited to just twelve<br />

channels for its more than seven hundred mobile subscribers, meaning<br />

that only twelve conversations could be carried on at once. In addition,<br />

because of possible interference, none of those channels could<br />

be reused in nearby cities; only fifty-four channels were available na-


A dominant trend in cell phone design is smaller<br />

<strong>and</strong> lighter units. (PhotoDisc)<br />

tionwide. By the late 1970’s,<br />

most of the systems in major<br />

cities were considered full, <strong>and</strong><br />

new subscribers were placed<br />

on a waiting list; some people<br />

had been waiting for as long<br />

as ten years to become subscribers.<br />

Mobile phone users<br />

commonly experienced long<br />

delays in getting poor-quality<br />

channels.<br />

The Cellular<br />

Breakthrough<br />

Cell phone / 173<br />

In 1968, the Federal Communications<br />

Commission (FCC)<br />

requested proposals for the<br />

creation of high-capacity, spectrum-efficient<br />

mobile systems.<br />

Bell Telephone had already<br />

been lobbying for the creation<br />

of such a system for some years. In the early 1970’s, both Motorola <strong>and</strong><br />

Bell Telephone proposed the use of cellular technology to solve the<br />

problems posed by mobile telephone service. Cellular systems involve<br />

the use of a computer to make it possible to use an area’s frequencies,<br />

or channels, repeatedly, allowing such systems to accommodate many<br />

more users.<br />

A two-thous<strong>and</strong>-customer, 2100-square-mile cellular telephone<br />

system called the Advanced Mobile Phone Service, built by the<br />

AMPS Corporation, an AT&T subsidiary, became operational in<br />

Chicago in 1978. The Illinois Bell Telephone Company was allowed<br />

to make a limited commercial offering <strong>and</strong> obtained about fourteen<br />

hundred subscribers. American Radio Telephone Service was allowed<br />

to conduct a similar test in the Baltimore/Washington area.<br />

These first systems showed the technological feasibility <strong>and</strong> affordability<br />

of cellular service.<br />

In 1979, Bell Labs of Murray Hill, New Jersey, received a patent


174 / Cell phone<br />

William Oliver Baker<br />

For great discoveries <strong>and</strong> inventions to be possible in the<br />

world of high technology, inventors need great facilities—laboratories<br />

<strong>and</strong> workshops—with brilliant colleagues. These must<br />

be managed by imaginative administrators.<br />

One of the best was William Oliver Baker (b. 1915), who rose<br />

to become president of the legendary Bell Labs. Baker started out<br />

as one of the most promising scientists of his generation. After<br />

earning a Ph.D. in chemistry at Princeton University, he joined<br />

the research section at Bell Telephone Laboratories in 1939. He<br />

studied the physics <strong>and</strong> chemistry of polymers, especially for use<br />

in electronics <strong>and</strong> telecommunications. During his research career<br />

he helped develop synthetic rubber <strong>and</strong> radar, found uses<br />

for polymers in communications <strong>and</strong> power cables, <strong>and</strong> participated<br />

in the discovery of microgels. In 1954 he ranked among the<br />

top-ten scientists in American industry <strong>and</strong> asked to chair a National<br />

Research Council committee studying heat shields for<br />

missiles <strong>and</strong> satellites.<br />

Administration suited him. The following year he took over<br />

as leader of research at Bell Labs <strong>and</strong> served as president from<br />

1973 until 1979. Under his direction, basic discoveries <strong>and</strong> inventions<br />

poured out of the lab that later transformed the way<br />

people live <strong>and</strong> work: satellite communications, principles for<br />

programming high-speed computers, the technology for modern<br />

electronic communications, the superconducting solenoid,<br />

the maser, <strong>and</strong> the laser. His scientists won Nobel Prizes <strong>and</strong> legions<br />

of other honors, as did Baker himself, who received dozens<br />

of medals, awards, <strong>and</strong> honorary degrees. Moreover, he<br />

was an original member of the President’s Science Advisory<br />

Board, became the first chair of the National Science Information<br />

Council, <strong>and</strong> served on the National Science Board. His<br />

influence on American science <strong>and</strong> technology was deep <strong>and</strong><br />

lasting.<br />

for such a system. The inventor was Richard H. Fefrenkiel, head of<br />

the mobile systems engineering department under the leadership<br />

of Labs president William Baker. The patented method divides a<br />

city into small coverage areas called “cells,” each served by lowpower<br />

transmitter-receivers. When a vehicle leaves the coverage


of one cell, calls are switched to the antenna <strong>and</strong> channels of an adjacent<br />

cell; a conversation underway is automatically transferred<br />

<strong>and</strong> continues without interruption. A channel used in one cell can<br />

be reused a few cells away for a different conversation. In this way,<br />

a few hundred channels can serve hundreds of thous<strong>and</strong>s of users.<br />

Computers control the call-transfer process, effectively reducing<br />

the amount of radio spectrum required. Cellular systems thus actually<br />

use radio frequencies to transmit conversations, but because<br />

the equipment is so telephone-like, “cellular telephone” (or “cell<br />

phone”) became the accepted term for the new technology.<br />

Each AMPS cell station is connected by wire to a central switching<br />

office, which determines when a mobile phone should be transferred<br />

to another cell as the transmitter moves out of range during a<br />

conversation. It does this by monitoring the strength of signals received<br />

from the mobile unit by adjacent cells, “h<strong>and</strong>ing off” the call<br />

when a new cell receives a stronger signal; this change is imperceptible<br />

to the user.<br />

Impact<br />

Cell phone / 175<br />

In 1982, the FCC began accepting applications for cellular system<br />

licenses in the thirty largest U.S. cities. By the end of 1984, there<br />

were about forty thous<strong>and</strong> cellular customers in nearly two dozen<br />

cities. Cellular telephone ownership boomed to 9 million by 1992.<br />

As cellular telephones became more common, they also became<br />

cheaper <strong>and</strong> more convenient to buy <strong>and</strong> to use. New systems<br />

developed in the 1990’s continued to make smaller, lighter, <strong>and</strong><br />

cheaper cellular phones even more accessible. Since the cellular telephone<br />

was made possible by the marriage of communications <strong>and</strong><br />

computers, advances in both these fields have continued to change<br />

the industry at a rapid rate.<br />

Cellular phones have proven ideal for many people who need or<br />

want to keep in touch with others at all times. They also provide<br />

convenient emergency communication devices for travelers <strong>and</strong><br />

field-workers. On the other h<strong>and</strong>, ownership of a cellular phone can<br />

also have its drawbacks; many users have found that they can never<br />

be out of touch—even when they would rather be.


176 / Cell phone<br />

See also Internet; Long-distance telephone; Rotary dial telephone;<br />

Telephone switching; Touch-tone telephone.<br />

Further Reading<br />

Carlo, George Louis, <strong>and</strong> Martin Schram. Cell Phones: Invisible Hazards<br />

in the Wireless Age. New York: Carroll <strong>and</strong> Graf, 2001.<br />

“The Cellular Phone.” Newsweek 130, 24A (Winter 1997/1998).<br />

Oliphant, Malcolm W. “How Mobile Telephony Got Going.” IEEE<br />

Spectrum 36, no. 8 (August, 1999).<br />

Young, Peter. Person to Person: The International Impact of the Telephone.<br />

Cambridge: Granta Editions, 1991.


Cloning<br />

Cloning<br />

The invention: Experimental technique for creating exact duplicates<br />

of living organisms by recreating their DNA.<br />

The people behind the invention:<br />

Ian Wilmut, an embryologist with the Roslin Institute<br />

Keith H. S. Campbell, an experiment supervisor with the Roslin<br />

Institute<br />

J. McWhir, a researcher with the Roslin Institute<br />

W. A. Ritchie, a researcher with the Roslin Institute<br />

Making Copies<br />

177<br />

On February 22, 1997, officials of the Roslin Institute, a biological<br />

research institution near Edinburgh, Scotl<strong>and</strong>, held a press conference<br />

to announce startling news: They had succeeded in creating<br />

a clone—a biologically identical copy—from cells taken from<br />

an adult sheep. Although cloning had been performed previously<br />

with simpler organisms, the Roslin Institute experiment marked<br />

the first time that a large, complex mammal had been successfully<br />

cloned.<br />

Cloning, or the production of genetically identical individuals,<br />

has long been a staple of science fiction <strong>and</strong> other popular literature.<br />

Clones do exist naturally, as in the example of identical twins. Scientists<br />

have long understood the process by which identical twins<br />

are created, <strong>and</strong> agricultural researchers have often dreamed of a<br />

method by which cheap identical copies of superior livestock could<br />

be created.<br />

The discovery of the double helix structure of deoxyribonucleic<br />

acid (DNA), or the genetic code, by James Watson <strong>and</strong> Francis Crick<br />

in the 1950’s led to extensive research into cloning <strong>and</strong> genetic engineering.<br />

Using the discoveries of Watson <strong>and</strong> Crick, scientists were<br />

soon able to develop techniques to clone laboratory mice; however,<br />

the cloning of complex, valuable animals such as livestock proved<br />

to be hard going.<br />

Early versions of livestock cloning were technical attempts at dupli-


178 / Cloning<br />

Ian Wilmut<br />

Ian Wilmut was born in Hampton Lucey, not far from Warwick<br />

in central Engl<strong>and</strong>, in 1944. He found his life’s calling in embryology—<strong>and</strong><br />

especially animal genetic engineering— while he<br />

was studying at the University of Nottingham, where his mentor<br />

was G. Eric Lamming, a leading expert on reproduction. After<br />

receiving his undergraduate degree, he attended Darwin<br />

College, Cambridge University. He completed his doctorate in<br />

1973 upon submitting a thesis about freezing boar sperm. This<br />

came after he produced a viable calf, named Frosty, from the<br />

frozen semen, the first time anyone had done so.<br />

Soon afterward he joined the Animal Breeding Research Station,<br />

which later became the Roslin Institute in Roslin, Scotl<strong>and</strong>.<br />

He immersed himself in research, seldom working fewer than<br />

nine hours a day. During the 1980’s he experimented with the<br />

insertion of genes into sheep embryos but concluded that cloning<br />

would be less time-consuming <strong>and</strong> less prone to failure.<br />

Joined by Keith Campbell in 1990, he cloned two Welsh mountain<br />

sheep from differentiated embryo cells, a feat similar to<br />

those of other reproductive experimenters. However, Dolly,<br />

who was cloned from adult cells, shook the world when her<br />

birth was announced in 1997. That same year Wilmut <strong>and</strong><br />

Campbell produced another cloned sheep, Polly. Cloned from<br />

fetal skin cells, she was genetically altered to carry a human<br />

gene.<br />

Wilmut’s technique for cloning from adult cells, which the<br />

laboratory patented, was a fundamentally new method of reproduction,<br />

but he had a loftier purpose in mind than simply<br />

establishing a first. He believed that animals genetically engineered<br />

to include human genes can produce proteins needed by<br />

people who because of genetic diseases cannot make the proteins<br />

themselves. The production of new treatments for old diseases,<br />

he told an astonished public after the revelation of Dolly,<br />

was his goal.<br />

cating the natural process of fertilized egg splitting that leads to the<br />

birth of identical twins. Artificially inseminated eggs were removed,<br />

split, <strong>and</strong> then reinserted into surrogate mothers. This method proved<br />

to be overly costly for commercial purposes, a situation aggravated by<br />

a low success rate.


Nuclear Transfer<br />

Model of a double helix. (PhotoDisc)<br />

Cloning / 179<br />

Researchers at the Roslin Institute found these earlier attempts to<br />

be fundamentally flawed. Even if the success rate could be improved,<br />

the number of clones created (of sheep, in this case) would<br />

still be limited. The Scots, led by embryologist Ian Wilmut <strong>and</strong> experiment<br />

supervisor Keith Campbell, decided to take an entirely<br />

different approach. The result was the first live birth of a mammal<br />

produced through a process known as “nuclear transfer.”<br />

Nuclear transfer involves the replacement of the nucleus of an<br />

immature egg with a nucleus taken from another cell. Previous attempts<br />

at nuclear transfer had cells from a single embryo divided<br />

up <strong>and</strong> implanted into an egg. Because a sheep embryo has only<br />

about forty usable cells, this method also proved limiting.<br />

The Roslin team therefore decided to grow their own cells in a<br />

laboratory culture. They took more mature embryonic cells than<br />

those previously used, <strong>and</strong> they experimented with the use of a nutrient<br />

mixture. One of their breakthroughs occurred when they discovered<br />

that these “cell lines” grew much more quickly when certain<br />

nutrients were absent.


180 / Cloning<br />

Using this technique, the Scots were able to produce a theoretically<br />

unlimited number of genetically identical cell lines. The next<br />

step was to transfer the cell lines of the sheep into the nucleus of unfertilized<br />

sheep eggs.<br />

First, 277 nuclei with a full set of chromosomes were transferred<br />

to the unfertilized eggs. An electric shock was then used to cause the<br />

eggs to begin development, the shock performing the duty of fertilization.<br />

Of these eggs, twenty-nine developed enough to be inserted<br />

into surrogate mothers.<br />

All the embryos died before birth except one: a ewe the scientists<br />

named “Dolly.” Her birth on July 5, 1996, was witnessed by only a<br />

veterinarian <strong>and</strong> a few researchers. Not until the clone had survived<br />

the critical earliest stages of life was the success of the experiment<br />

disclosed; Dolly was more than seven months old by the time her<br />

birth was announced to a startled world.<br />

Impact<br />

The news that the cloning of sophisticated organisms had left the<br />

realm of science fiction <strong>and</strong> become a matter of accomplished scientific<br />

fact set off an immediate uproar. Ethicists <strong>and</strong> media commentators<br />

quickly began to debate the moral consequences of the use—<br />

<strong>and</strong> potential misuse—of the technology. Politicians in numerous<br />

countries responded to the news by calling for legal restrictions on<br />

cloning research. Scientists, meanwhile, speculated about the possible<br />

benefits <strong>and</strong> practical limitations of the process.<br />

The issue that stirred the imagination of the broader public <strong>and</strong><br />

sparked the most spirited debate was the possibility that similar experiments<br />

might soon be performed using human embryos. Although<br />

most commentators seemed to agree that such efforts would<br />

be profoundly immoral, many experts observed that they would be<br />

virtually impossible to prevent. “Could someone do this tomorrow<br />

morning on a human embryo?” Arthur L. Caplan, the director of the<br />

University of Pennsylvania’s bioethics center, asked reporters. “Yes.<br />

It would not even take too much science. The embryos are out<br />

there.”<br />

Such observations conjured visions of a future that seemed marvelous<br />

to some, nightmarish to others. Optimists suggested that the


est <strong>and</strong> brightest of humanity could be forever perpetuated, creating<br />

an endless supply of Albert Einsteins <strong>and</strong> Wolfgang Amadeus<br />

Mozarts. Pessimists warned of a world overrun by clones of selfserving<br />

narcissists <strong>and</strong> petty despots, or of the creation of a secondary<br />

class of humans to serve as organ donors for their progenitors.<br />

The Roslin Institute’s researchers steadfastly proclaimed their<br />

own opposition to human experimentation. Moreover, most scientists<br />

were quick to point out that such scenarios were far from realization,<br />

noting the extremely high failure rate involved in the creation<br />

of even a single sheep. In addition, most experts emphasized<br />

more practical possible uses of the technology: improving agricultural<br />

stock by cloning productive <strong>and</strong> disease-resistant animals, for<br />

example, or regenerating endangered or even extinct species. Even<br />

such apparently benign schemes had their detractors, however, as<br />

other observers remarked on the potential dangers of thus narrowing<br />

a species’ genetic pool.<br />

Even prior to the Roslin Institute’s announcement, most European<br />

nations had adopted a bioethics code that flatly prohibited genetic<br />

experiments on human subjects. Ten days after the announcement,<br />

U.S. president Bill Clinton issued an executive order that<br />

banned the use of federal money for human cloning research, <strong>and</strong><br />

he called on researchers in the private sector to refrain from such experiments<br />

voluntarily. Nevertheless, few observers doubted that<br />

Dolly’s birth marked only the beginning of an intriguing—<strong>and</strong> possibly<br />

frightening—new chapter in the history of science.<br />

See also Amniocentesis; Artificial chromosome; Artificial insemination;<br />

Genetic “fingerprinting”; In vitro plant culture; Rice <strong>and</strong><br />

wheat strains.<br />

Further Reading<br />

Cloning / 181<br />

Facklam, Margery, Howard Facklam, <strong>and</strong> Paul Facklam. From Cell to<br />

Clone: The Story of Genetic Engineering. New York: Harcourt Brace<br />

Jovanovich, 1979.<br />

Gillis, Justin. “Cloned Cows Are Fetching Big Bucks: Dozens of Genetic<br />

Duplicates Ready to Take Up Residence on U.S. Farms.”<br />

Washington Post (March 25, 2001).


182 / Cloning<br />

Kolata, Gina Bari. Clone: The Road to Dolly, <strong>and</strong> the Path Ahead. New<br />

York: William Morrow, 1998.<br />

Regalado, Antonio. “Clues Are Sought for Cloning’s Fail Rate: Researchers<br />

Want to Know Exactly How an Egg Reprograms Adult<br />

DNA.” Wall Street Journal (November 24, 2000).<br />

Winslow, Ron. “Scientists Clone Pigs, Lifting Prospects of Replacement<br />

Organs for Humans.” Wall Street Journal (August 17, 2000).


Cloud seeding<br />

Cloud seeding<br />

The invention: Technique for inducing rainfall by distributing dry<br />

ice or silver nitrate into reluctant rainclouds.<br />

The people behind the invention:<br />

Vincent Joseph Schaefer (1906-1993), an American chemist <strong>and</strong><br />

meteorologist<br />

Irving Langmuir (1881-1957), an American physicist <strong>and</strong><br />

chemist who won the 1932 Nobel Prize in Chemistry<br />

Bernard Vonnegut (1914-1997), an American physical chemist<br />

<strong>and</strong> meteorologist<br />

Praying for Rain<br />

183<br />

Beginning in 1943, an intense interest in the study of clouds developed<br />

into the practice of weather “modification.” Working for<br />

the General Electric Research Laboratory, Nobel laureate Irving<br />

Langmuir <strong>and</strong> his assistant researcher <strong>and</strong> technician, Vincent Joseph<br />

Schaefer, began an intensive study of precipitation <strong>and</strong> its<br />

causes.<br />

Past research <strong>and</strong> study had indicated two possible ways that<br />

clouds produce rain. The first possibility is called “coalescing,” a<br />

process by which tiny droplets of water vapor in a cloud merge after<br />

bumping into one another <strong>and</strong> become heavier <strong>and</strong> fatter until they<br />

drop to earth. The second possibility is the “Bergeron process” of<br />

droplet growth, named after the Swedish meteorologist Tor Bergeron.<br />

Bergeron’s process relates to supercooled clouds, or clouds<br />

that are at or below freezing temperatures <strong>and</strong> yet still contain both<br />

ice crystals <strong>and</strong> liquid water droplets. The size of the water droplets<br />

allows the droplets to remain liquid despite freezing temperatures;<br />

while small droplets can remain liquid only down to 4 degrees Celsius,<br />

larger droplets may not freeze until reaching −15 degrees<br />

Celsius. Precipitation occurs when the ice crystals become heavy<br />

enough to fall. If the temperature at some point below the cloud is<br />

warm enough, it will melt the ice crystals before they reach the<br />

earth, producing rain. If the temperature remains at the freezing


184 / Cloud seeding<br />

point, the ice crystals retain their form <strong>and</strong> fall as snow.<br />

Schaefer used a deep-freezing unit in order to observe water<br />

droplets in pure cloud form. In order to observe the droplets better,<br />

Schaefer lined the chest with black velvet <strong>and</strong> concentrated a beam<br />

of light inside. The first agent he introduced inside the supercooled<br />

freezer was his own breath. When that failed to form the desired ice<br />

crystals, he proceeded to try other agents. His hope was to form ice<br />

crystals that would then cause the moisture in the surrounding air<br />

to condense into more ice crystals, which would produce a miniature<br />

snowfall.<br />

He eventually achieved success when he tossed a h<strong>and</strong>ful of dry<br />

ice inside <strong>and</strong> was rewarded with the long-awaited snow. The<br />

freezer was set at the freezing point of water, 0 degrees Celsius, but<br />

not all the particles were ice crystals, so when the dry ice was introduced<br />

all the stray water droplets froze instantly, producing ice<br />

crystals, or snowflakes.<br />

Planting the First Seeds<br />

On November 13, 1946, Schaefer took to the air over Mount<br />

Greylock with several pounds of dry ice in order to repeat the experiment<br />

in nature. After he had finished sprinkling, or seeding, a<br />

supercooled cloud, he instructed the pilot to fly underneath the<br />

cloud he had just seeded. Schaefer was greeted by the sight of snow.<br />

By the time it reached the ground, it had melted into the first-ever<br />

human-made rainfall.<br />

Independently of Schaefer <strong>and</strong> Langmuir, another General Electric<br />

scientist, Bernard Vonnegut, was also seeking a way to cause<br />

rain. He found that silver iodide crystals, which have the same size<br />

<strong>and</strong> shape as ice crystals, could “fool” water droplets into condensing<br />

on them. When a certain chemical mixture containing silver iodide<br />

is heated on a special burner called a “generator,” silver iodide<br />

crystals appear in the smoke of the mixture. Vonnegut’s discovery<br />

allowed seeding to occur in a way very different from seeding with<br />

dry ice, but with the same result. Using Vonnegut’s process, the<br />

seeding is done from the ground. The generators are placed outside<br />

<strong>and</strong> the chemicals are mixed. As the smoke wafts upward, it carries<br />

the newly formed silver iodide crystals with it into the clouds.


The results of the scientific experiments by Langmuir, Vonnegut,<br />

<strong>and</strong> Schaefer were alternately hailed <strong>and</strong> rejected as legitimate.<br />

Critics argue that the process of seeding is too complex <strong>and</strong><br />

would have to require more than just the addition of dry ice or silver<br />

nitrate in order to produce rain. One of the major problems surrounding<br />

the question of weather modification by cloud seeding is<br />

the scarcity of knowledge about the earth’s atmosphere. A journey<br />

begun about fifty years ago is still a long way from being completed.<br />

Impact<br />

Although the actual statistical <strong>and</strong> other proofs needed to support<br />

cloud seeding are lacking, the discovery in 1946 by the General<br />

Electric employees set off a wave of interest <strong>and</strong> dem<strong>and</strong> for information<br />

that far surpassed the interest generated by the discovery of<br />

nuclear fission shortly before. The possibility of ending drought<br />

<strong>and</strong>, in the process, hunger excited many people. The discovery also<br />

prompted both legitimate <strong>and</strong> false “rainmakers” who used the information<br />

gathered by Schaefer, Langmuir, <strong>and</strong> Vonnegut to set up<br />

cloud-seeding businesses. Weather modification, in its current stage<br />

of development, cannot be used to end worldwide drought. It does,<br />

however, have beneficial results in some cases on the crops of<br />

smaller farms that have been affected by drought.<br />

In order to underst<strong>and</strong> the advances made in weather modification,<br />

new instruments are needed to record accurately the results of<br />

further experimentation. The storm of interest—both favorable <strong>and</strong><br />

nonfavorable—generated by the discoveries of Schaefer, Langmuir,<br />

<strong>and</strong> Vonnegut has had <strong>and</strong> will continue to have far-reaching effects<br />

on many aspects of society.<br />

See also Airplane; Artificial insemination; In vitro plant culture;<br />

Weather satellite.<br />

Further Reading<br />

Cloud seeding / 185<br />

Cole, Stephen. “Mexico Results Spur New Looking at Rainmaking.”<br />

Washington Post (January 22, 2001).


186 / Cloud seeding<br />

Havens, Barrington S., James E. Jiusto, <strong>and</strong> Bernard Vonnegut. Early<br />

History of Cloud Seeding. Socorro, N.Mex.: Langmuir Laboratory,<br />

New Mexico Institute of Mining <strong>and</strong> Technology, 1978.<br />

“Science <strong>and</strong> Technology: Cloudbusting.” The Economist (August 21,<br />

1999).<br />

Villiers, Marq de. Water: The Fate of Our Most Precious Resource. Boston:<br />

Houghton Mifflin, 2000.


COBOL computer language<br />

COBOL computer language<br />

The invention: The first user-friendly computer programming language,<br />

COBOL was originally designed to solve ballistics problems.<br />

The people behind the invention:<br />

Grace Murray Hopper (1906-1992), an American<br />

mathematician<br />

Howard Hathaway Aiken (1900-1973), an American<br />

mathematician<br />

Plain Speaking<br />

187<br />

Grace Murray Hopper, a mathematician, was a faculty member<br />

at Vassar College when World War II (1939-1945) began. She enlisted<br />

in the Navy <strong>and</strong> in 1943 was assigned to the Bureau of Ordnance<br />

Computation Project, where she worked on ballistics problems.<br />

In 1944, the Navy began using one of the first electronic<br />

computers, the Automatic Sequence Controlled Calculator (ASCC),<br />

designed by an International Business Machines (IBM) Corporation<br />

team of engineers headed by Howard Hathaway Aiken, to solve<br />

ballistics problems. Hopper became the third programmer of the<br />

ASCC.<br />

Hopper’s interest in computer programming continued after<br />

the war ended. By the early 1950’s, Hopper’s work with programming<br />

languages had led to her development of FLOW-MATIC, the<br />

first English-language data processing compiler. Hopper’s work<br />

on FLOW-MATIC paved the way for her later work with COBOL<br />

(Common Business Oriented Language).<br />

Until Hopper developed FLOW-MATIC, digital computer programming<br />

was all machine-specific <strong>and</strong> was written in machine<br />

code. A program designed for one computer could not be used on<br />

another. Every program was both machine-specific <strong>and</strong> problemspecific<br />

in that the programmer would be told what problem the<br />

machine was going to be asked <strong>and</strong> then would write a completely<br />

new program for that specific problem in the machine code.


188 / COBOL computer language<br />

Grace Murray Hopper<br />

Grace Brewster Murray was born in New York City in 1906.<br />

As a child she revered her great-gr<strong>and</strong>father, a U.S. Navy admiral,<br />

<strong>and</strong> her gr<strong>and</strong>father, an engineer. Her career melded their<br />

professions.<br />

She studied mathematics <strong>and</strong> physics at Vassar College,<br />

earning a bachelor’s degree in 1928 <strong>and</strong> a master’s degree in<br />

1930, when she married Vincent Foster Hopper. She accepted a<br />

teaching post at Vassar but continued her studies, completing a<br />

doctorate at Yale University in 1934. In 1943 she left academia<br />

for the Navy <strong>and</strong> was assigned to the Bureau of Ordnance Computation<br />

Project at Harvard University. She worked on the nation’s<br />

first modern computer, the Mark I, <strong>and</strong> contributed to the<br />

development of major new models afterward, including Sperry<br />

Corporation’s ENIAC <strong>and</strong> UNIVAC. While still with the Navy<br />

project at Harvard, Hopper participated in a minor incident<br />

that forever marked computer slang. One day a moth became<br />

caught in a switch, causing the computer to malfunction. She<br />

<strong>and</strong> other technicians found it <strong>and</strong> ever after referred to correcting<br />

mechanical glitches as “debugging.”<br />

Hopper joined Sperry Corporation after the war <strong>and</strong> carried<br />

out her seminal work with the FLOW-MATIC <strong>and</strong> COBOL<br />

computer languages. Meanwhile, she retained her commission<br />

in the Naval Reserves, helping the service incorporate computers<br />

<strong>and</strong> COBOL into its armaments <strong>and</strong> administration systems.<br />

She retired from the Navy in 1966 <strong>and</strong> from Sperry in<br />

1971, but the Navy soon had her out of retirement on temporary<br />

active duty to help with its computer systems. After her second<br />

retirement, the Navy, grateful for her tireless service, promoted<br />

her to rear admiral in 1985, the nation’s first woman admiral.<br />

She was also awarded the Distinguished Service Cross by the<br />

Department of Defense, the National Medal of Technology, <strong>and</strong><br />

the Legion of Merit. She became an inductee into the Engineering<br />

<strong>and</strong> Science Hall of Fame in 1991. Hopper, nicknamed<br />

Amazing Grace, died a year later.<br />

Machine code was based on the programmer’s knowledge of the<br />

physical characteristics of the computer as well as the requirements of<br />

the problem to be solved; that is, the programmer had to know what<br />

was happening within the machine as it worked through a series of


calculations, which relays tripped when <strong>and</strong> in what order, <strong>and</strong> what<br />

mathematical operations were necessary to solve the problem. Programming<br />

was therefore a highly specialized skill requiring a unique<br />

combination of linguistic, reasoning, engineering, <strong>and</strong> mathematical<br />

abilities that not even all the mathematicians <strong>and</strong> electrical engineers<br />

who designed <strong>and</strong> built the early computers possessed.<br />

While every computer still operates in response to the programming,<br />

or instructions, built into it, which are formatted in machine<br />

code, modern computers can accept programs written in nonmachine<br />

code—that is, in various automatic programming languages. They<br />

are able to accept nonmachine code programs because specialized<br />

programs now exist to translate those programs into the appropriate<br />

machine code. These translating programs are known as “compilers,”<br />

or “assemblers,” <strong>and</strong> FLOW-MATIC was the first such program.<br />

Hopper developed FLOW-MATIC after realizing that it would<br />

be necessary to eliminate unnecessary steps in programming to<br />

make computers more efficient. FLOW-MATIC was based, in part,<br />

on Hopper’s recognition that certain elements, or comm<strong>and</strong>s, were<br />

common to many different programming applications. Hopper theorized<br />

that it would not be necessary to write a lengthy series of instructions<br />

in machine code to instruct a computer to begin a series of<br />

operations; instead, she believed that it would be possible to develop<br />

comm<strong>and</strong>s in an assembly language in such a way that a programmer<br />

could write one comm<strong>and</strong>, such as the word add, that<br />

would translate into a sequence of several comm<strong>and</strong>s in machine<br />

code. Hopper’s successful development of a compiler to translate<br />

programming languages into machine code thus meant that programming<br />

became faster <strong>and</strong> easier. From assembly languages such<br />

as FLOW-MATIC, it was a logical progression to the development of<br />

high-level computer languages, such as FORTRAN (Formula Translation)<br />

<strong>and</strong> COBOL.<br />

The Language of Business<br />

COBOL computer language / 189<br />

Between 1955 (when FLOW-MATIC was introduced) <strong>and</strong> 1959, a<br />

number of attempts at developing a specific business-oriented language<br />

were made. IBM <strong>and</strong> Remington R<strong>and</strong> believed that the only<br />

way to market computers to the business community was through


190 / COBOL computer language<br />

the development of a language that business people would be<br />

comfortable using. Remington R<strong>and</strong> officials were especially committed<br />

to providing a language that resembled English. None of<br />

the attempts to develop a business-oriented language succeeded,<br />

however, <strong>and</strong> by 1959 Hopper <strong>and</strong> other members of the U.S. Department<br />

of Defense had persuaded representatives of various companies<br />

of the need to cooperate.<br />

On May 28 <strong>and</strong> 29, 1959, a conference sponsored by the Department<br />

of Defense was held at the Pentagon to discuss the problem of<br />

establishing a common language for the adaptation of electronic<br />

computers for data processing. As a result, the first distribution of<br />

COBOL was accomplished on December 17, 1959. Although many<br />

people were involved in the development of COBOL, Hopper played<br />

a particularly important role. She not only found solutions to technical<br />

problems but also succeeded in selling the concept of a common<br />

language from an administrative <strong>and</strong> managerial point of view. Hopper<br />

recognized that while the companies involved in the commercial<br />

development of computers were in competition with one another, the<br />

use of a common, business-oriented language would contribute to<br />

the growth of the computer industry as a whole, as well as simplify<br />

the training of computer programmers <strong>and</strong> operators.<br />

Consequences<br />

COBOL was the first compiler developed for business data processing<br />

operations. Its development simplified the training required<br />

for computer users in business applications <strong>and</strong> demonstrated that<br />

computers could be practical tools in government <strong>and</strong> industry as<br />

well as in science. Prior to the development of COBOL, electronic<br />

computers had been characterized as expensive, oversized adding<br />

machines that were adequate for performing time-consuming mathematics<br />

but lacked the flexibility that business people required.<br />

In addition, the development of COBOL freed programmers not<br />

only from the need to know machine code but also from the need to<br />

underst<strong>and</strong> the physical functioning of the computers they were using.<br />

Programming languages could be written that were both machine-independent<br />

<strong>and</strong> almost universally convertible from one<br />

computer to another.


Finally, because Hopper <strong>and</strong> the other committee members worked<br />

under the auspices of the Department of Defense, the software<br />

was not copyrighted, <strong>and</strong> in a short period of time COBOL became<br />

widely available to anyone who wanted to use it. It diffused rapidly<br />

throughout the industry <strong>and</strong> contributed to the widespread adaptation<br />

of computers for use in countless settings.<br />

See also BASIC programming language; Colossus computer;<br />

ENIAC computer; FORTRAN programming language; SAINT.<br />

Further Reading<br />

COBOL computer language / 191<br />

Cohen, Bernard I., Gregory W. Welch, <strong>and</strong> Robert V. D. Campbell.<br />

Makin’ Numbers: Howard Aiken: <strong>and</strong> the Computer. Cambridge,<br />

Mass.: MIT Press, 1999.<br />

Cohen, Bernard I. Howard Aiken: Portrait of a Computer Pioneer. Cambridge,<br />

Mass.: MIT Press, 1999.<br />

Ferguson, David E. “The Roots of COBOL.” Systems 3X World <strong>and</strong> As<br />

World 17, no. 7 (July, 1989).<br />

Yount, Lisa. A to Z of Women in Science <strong>and</strong> Math. New York: Facts on<br />

File, 1999.


192<br />

Color film<br />

Color film<br />

The invention: Aphotographic medium used to take full-color pictures.<br />

The people behind the invention:<br />

Rudolf Fischer (1881-1957), a German chemist<br />

H. Siegrist (1885-1959), a German chemist <strong>and</strong> Fischer’s<br />

collaborator<br />

Benno Homolka (1877-1949), a German chemist<br />

The Process Begins<br />

Around the turn of the twentieth century, Arthur-Louis Ducos du<br />

Hauron, a French chemist <strong>and</strong> physicist, proposed a tripack (threelayer)<br />

process of film development in which three color negatives<br />

would be taken by means of superimposed films. This was a subtractive<br />

process. (In the “additive method” of making color pictures,<br />

the three colors are added in projection—that is, the colors are formed<br />

by the mixture of colored light of the three primary hues. In the<br />

“subtractive method,” the colors are produced by the superposition<br />

of prints.) In Ducos du Hauron’s process, the blue-light negative<br />

would be taken on the top film of the pack; a yellow filter below it<br />

would transmit the yellow light, which would reach a green-sensitive<br />

film <strong>and</strong> then fall upon the bottom of the pack, which would be sensitive<br />

to red light. Tripacks of this type were unsatisfactory, however,<br />

because the light became diffused in passing through the emulsion<br />

layers, so the green <strong>and</strong> red negatives were not sharp.<br />

To obtain the real advantage of a tripack, the three layers must<br />

be coated one over the other so that the distance between the bluesensitive<br />

<strong>and</strong> red-sensitive layers is a small fraction of a thous<strong>and</strong>th<br />

of an inch. Tripacks of this type were suggested by the early pioneers<br />

of color photography, who had the idea that the packs would<br />

be separated into three layers for development <strong>and</strong> printing. The<br />

manipulation of such systems proved to be very difficult in practice.<br />

It was also suggested, however, that it might be possible to develop<br />

such tripacks as a unit <strong>and</strong> then, by chemical treatment, convert the<br />

silver images into dye images.


Fischer’s Theory<br />

One of the earliest subtractive tripack methods that seemed to<br />

hold great promise was that suggested by Rudolf Fischer in 1912. He<br />

proposed a tripack that would be made by coating three emulsions<br />

on top of one another; the lowest one would be red-sensitive, the<br />

middle one would be green-sensitive, <strong>and</strong> the top one would be bluesensitive.<br />

Chemical substances called “couplers,” which would produce<br />

dyes in the development process, would be incorporated into<br />

the layers. In this method, the molecules of the developing agent, after<br />

becoming oxidized by developing the silver image, would react<br />

with the unoxidized form (the coupler) to produce the dye image.<br />

The two types of developing agents described by Fischer are<br />

paraminophenol <strong>and</strong> paraphenylenediamine (or their derivatives).<br />

The five types of dye that Fischer discovered are formed when silver<br />

images are developed by these two developing agents in the presence<br />

of suitable couplers. The five classes of dye he used (indophenols,<br />

indoanilines, indamines, indothiophenols, <strong>and</strong> azomethines)<br />

were already known when Fischer did his work, but it was he who<br />

discovered that the photographic latent image could be used to promote<br />

their formulation from “coupler” <strong>and</strong> “developing agent.”<br />

The indoaniline <strong>and</strong> azomethine types have been found to possess<br />

the necessary properties, but the other three suffer from serious defects.<br />

Because only p-phenylenediamine <strong>and</strong> its derivatives can be<br />

used to form the indoaniline <strong>and</strong> azomethine dyes, it has become<br />

the most widely used color developing agent.<br />

Impact<br />

Color film / 193<br />

In the early 1920’s, Leopold Mannes <strong>and</strong> Leopold Godowsky<br />

made a great advance beyond the Fischer process. Working on a<br />

new process of color photography, they adopted coupler development,<br />

but instead of putting couplers into the emulsion as Fischer<br />

had, they introduced them during processing. Finally, in 1935, the<br />

film was placed on the market under the name “Kodachrome,” a<br />

name that had been used for an early two-color process.<br />

The first use of the new Kodachrome process in 1935 was for 16millimeter<br />

film. Color motion pictures could be made by the Koda-


194 / Color film<br />

chrome process as easily as black-<strong>and</strong>-white pictures, because the<br />

complex work involved (the color development of the film) was<br />

done under precise technical control. The definition (quality of the<br />

image) given by the process was soon sufficient to make it practical<br />

for 8-millimeter pictures, <strong>and</strong> in 1936, Kodachrome film was introduced<br />

in a 35-millimeter size for use in popular miniature cameras.<br />

Soon thereafter, color processes were developed on a larger scale<br />

<strong>and</strong> new color materials were rapidly introduced. In 1940, the Kodak<br />

Research Laboratories worked out a modification of the Fischer<br />

process in which the couplers were put into the emulsion layers.<br />

These couplers are not dissolved in the gelatin layer itself, as the<br />

Fischer couplers are, but are carried in small particles of an oily material<br />

that dissolves the couplers, protects them from the gelatin,<br />

<strong>and</strong> protects the silver bromide from any interaction with the couplers.<br />

When development takes place, the oxidation product of the<br />

developing agent penetrates into the organic particles <strong>and</strong> reacts<br />

with the couplers so that the dyes are formed in small particles that<br />

are dispersed throughout the layers. In one form of this material,<br />

Ektachrome (originally intended for use in aerial photography), the<br />

film is reversed to produce a color positive. It is first developed with<br />

a black-<strong>and</strong>-white developer, then reexposed <strong>and</strong> developed with a<br />

color developer that recombines with the couplers in each layer to<br />

produce the appropriate dyes, all three of which are produced simultaneously<br />

in one development.<br />

In summary, although Fischer did not succeed in putting his theory<br />

into practice, his work still forms the basis of most modern color<br />

photographic systems. Not only did he demonstrate the general<br />

principle of dye-coupling development, but the art is still mainly<br />

confined to one of the two types of developing agent, <strong>and</strong> two of the<br />

five types of dye, described by him.<br />

See also Autochrome plate; Brownie camera; Infrared photography;<br />

Instant photography.<br />

Further Reading<br />

Collins, Douglas. The Story of Kodak. New York: Harry N. Abrams,<br />

1990.


Color film / 195<br />

Glendinning, Peter. Color Photography: History, Theory, <strong>and</strong> Darkroom<br />

Technique. Englewood Cliffs, N.J.: Prentice-Hall, 1985.<br />

Wood, John. The Art of the Autochrome: The Birth of Color Photography.<br />

Iowa City: University of Iowa Press, 1993.


196<br />

Color television<br />

Color television<br />

The invention: System for broadcasting full-color images over the<br />

airwaves.<br />

The people behind the invention:<br />

Peter Carl Goldmark (1906-1977), the head of the CBS research<br />

<strong>and</strong> development laboratory<br />

William S. Paley (1901-1990), the businessman who took over<br />

CBS<br />

David Sarnoff (1891-1971), the founder of RCA<br />

The Race for St<strong>and</strong>ardization<br />

Although by 1928 color television had already been demonstrated<br />

in Scotl<strong>and</strong>, two events in 1940 mark that year as the beginning<br />

of color television. First, on February 12, 1940, the Radio Corporation<br />

of America (RCA) demonstrated its color television system<br />

privately to a group that included members of the Federal Communications<br />

Commission (FCC), an administrative body that had the<br />

authority to set st<strong>and</strong>ards for an electronic color system. The demonstration<br />

did not go well; indeed, David Sarnoff, the head of RCA,<br />

canceled a planned public demonstration <strong>and</strong> returned his engineers<br />

to the Princeton, New Jersey, headquarters of RCA’s laboratories.<br />

Next, on September 1, 1940, the Columbia Broadcasting System<br />

(CBS) took the first step to develop a color system that would become<br />

the st<strong>and</strong>ard for the United States. On that day, CBS demonstrated<br />

color television to the public, based on the research of an engineer,<br />

Peter Carl Goldmark. Goldmark placed a set of spinning<br />

filters in front of the black-<strong>and</strong>-white television images, breaking<br />

them down into three primary colors <strong>and</strong> producing color television.<br />

The audience saw what was called “additive color.”<br />

Although Goldmark had been a researcher at CBS since January,<br />

1936, he did not attempt to develop a color television system until<br />

March, 1940, after watching the Technicolor motion picture Gone<br />

with the Wind (1939). Inspired, Goldmark began to tinker in his tiny


CBS laboratory in the headquarters building in New York City.<br />

If a decision had been made in 1940, the CBS color st<strong>and</strong>ard<br />

would have been accepted as the national st<strong>and</strong>ard. The FCC was,<br />

at that time, more concerned with trying to establish a black-<strong>and</strong>white<br />

st<strong>and</strong>ard for television. Color television seemed decades away.<br />

In 1941, the FCC decided to adopt st<strong>and</strong>ards for black-<strong>and</strong>-white<br />

television only, leaving the issue of color unresolved—<strong>and</strong> the<br />

doors to the future of color broadcasting wide open. Control of a potentially<br />

lucrative market as well as personal rivalry threw William<br />

S. Paley, the head of CBS, <strong>and</strong> Sarnoff into a race for the control of<br />

color television. Both companies would pay dearly in terms of<br />

money <strong>and</strong> time, but it would take until the 1960’s before the United<br />

States would become a nation of color television watchers.<br />

RCA was at the time the acknowledged leader in the development<br />

of black-<strong>and</strong>-white television. CBS engineers soon discovered,<br />

however, that their company’s color system would not work when<br />

combined with RCA black-<strong>and</strong>-white televisions. In other words,<br />

customers would need one set for black-<strong>and</strong>-white <strong>and</strong> one for<br />

color. Moreover, since the color system of CBS needed more broadcast<br />

frequency space than the black-<strong>and</strong>-white system in use, CBS<br />

was forced to ask the FCC to allocate new channel space in the<br />

ultrahigh frequency (UHF) b<strong>and</strong>, which was then not being used. In<br />

contrast, RCA scientists labored to make a compatible color system<br />

that required no additional frequency space.<br />

No Time to Wait<br />

Color television / 197<br />

Following the end of World War II, in 1945, the suburbanites who<br />

populated new communities in America’s cities wanted television sets<br />

right away; they did not want to wait for the government to decide on<br />

a color st<strong>and</strong>ard <strong>and</strong> then wait again while manufacturers redesigned<br />

assembly lines to make color sets. Rich with savings accumulated during<br />

the prosperity of the war years, Americans wanted to spend their<br />

money. After the war, the FCC saw no reason to open up proceedings<br />

regarding color systems. Black-<strong>and</strong>-white was operational; customers<br />

were waiting in line for the new electronic marvel. To give its engineers<br />

time to create a compatible color system, RCA skillfully lobbied the<br />

members of the FCC to take no action.


198 / Color television<br />

There were other problems with the CBS mechanical color television.<br />

It was noisy <strong>and</strong> large, <strong>and</strong> its color balance was hard to maintain.<br />

CBS claimed that through further engineering work, it would<br />

improve the actual sets. Yet RCA was able to convince other manufacturers<br />

to support it in preference to CBS principally because of its<br />

proven manufacturing track record.<br />

In 1946, RCA demonstrated a new electronic color receiver with<br />

three picture tubes, one for each of the primary colors. Color reproduction<br />

was fairly true; although any movement on the screen<br />

caused color blurring, there was little flicker. It worked, however,<br />

<strong>and</strong> thus ended the invention phase of color television begun in<br />

1940. The race for st<strong>and</strong>ardization would require seven more years<br />

of corporate struggle before the RCA system would finally win<br />

adoption as the national st<strong>and</strong>ard in 1953.<br />

Impact<br />

Through the 1950’s, black-<strong>and</strong>-white television remained the order<br />

of the day. Through the later years of the decade, only the National<br />

Broadcasting Company (NBC) television network was regularly<br />

airing programs in color. Full production <strong>and</strong> presentation of<br />

shows in color during prime time did not come until the mid-1960’s;<br />

most industry observers date 1972 as the true arrival of color television.<br />

By 1972, color sets were found in more than half the homes in the<br />

United States. At that point, since color was so widespread, TV<br />

Guide stopped tagging color program listings with a special symbol<br />

<strong>and</strong> instead tagged only black-<strong>and</strong>-white shows, as it does to this<br />

day. Gradually, only cheap, portable sets were made for black-<strong>and</strong>white<br />

viewing, while color sets came in all varieties from tiny h<strong>and</strong>held<br />

pocket televisions to mammoth projection televisions.<br />

See also Autochrome plate; Community antenna television;<br />

Communications satellite; Fiber-optics; FM radio; Radio; Television;<br />

Transistor; Videocassette recorder.


Further Reading<br />

Color television / 199<br />

Burns, R. W. Television: An International History of the Formative Years.<br />

London: Institution of Electrical Engineers in association with<br />

the Science Museum, 1998.<br />

Fisher, David E., <strong>and</strong> Marshall Fisher. Tube: The Invention of Television.<br />

Washington, D.C.: Counterpoint, 1996.<br />

Lewis, Tom. Empire of the Air: The Men Who Made Radio. New York:<br />

HarperPerennial, 1993.<br />

Lyons, Eugene. David Sarnoff: A Biography. New York: Harper <strong>and</strong><br />

Row, 1967.


200<br />

Colossus computer<br />

Colossus computer<br />

The invention: The first all-electronic calculating device, the Colossus<br />

computer was built to decipher German military codes<br />

during World War II.<br />

The people behind the invention:<br />

Thomas H. Flowers, an electronics expert<br />

Max H. A. Newman (1897-1984), a mathematician<br />

Alan Mathison Turing (1912-1954), a mathematician<br />

C. E. Wynn-Williams, a member of the Telecommunications<br />

Research Establishment<br />

An Undercover Operation<br />

In 1939, during World War II (1939-1945), a team of scientists,<br />

mathematicians, <strong>and</strong> engineers met at Bletchley Park, outside London,<br />

to discuss the development of machines that would break the<br />

secret code used in Nazi military communications. The Germans<br />

were using a machine called “Enigma” to communicate in code between<br />

headquarters <strong>and</strong> field units. Polish scientists, however, had<br />

been able to examine a German Enigma <strong>and</strong> between 1928 <strong>and</strong> 1938<br />

were able to break the codes by using electromechanical codebreaking<br />

machines called “bombas.” In 1938, the Germans made the<br />

Enigma more complicated, <strong>and</strong> the Polish were no longer able to<br />

break the codes. In 1939, the Polish machines <strong>and</strong> codebreaking<br />

knowledge passed to the British.<br />

Alan Mathison Turing was one of the mathematicians gathered<br />

at Bletchley Park to work on codebreaking machines. Turing was<br />

one of the first people to conceive of the universality of digital computers.<br />

He first mentioned the “Turing machine” in 1936 in an article<br />

published in the Proceedings of the London Mathematical Society.<br />

The Turing machine, a hypothetical device that can solve any<br />

problem that involves mathematical computation, is not restricted<br />

to only one task—hence the universality feature.<br />

Turing suggested an improvement to the Bletchley codebreaking<br />

machine, the “Bombe,” which had been modeled on the Polish


omba. This improvement increased the computing power of the<br />

machine. The new codebreaking machine replaced the tedious<br />

method of decoding by h<strong>and</strong>, which in addition to being slow,<br />

was ineffective in dealing with complicated encryptions that were<br />

changed daily.<br />

Building a Better Mousetrap<br />

Colossus computer / 201<br />

The Bombe was very useful. In 1942, when the Germans started<br />

using a more sophisticated cipher machine known as the “Fish,”<br />

Max H. A. Newman, who was in charge of one subunit at Bletchley<br />

Park, believed that an automated device could be designed to break<br />

the codes produced by the Fish. Thomas H. Flowers, who was in<br />

charge of a switching group at the Post Office Research Station at<br />

Dollis Hill, had been approached to build a special-purpose electromechanical<br />

device for Bletchley Park in 1941. The device was not<br />

useful, <strong>and</strong> Flowers was assigned to other problems.<br />

Flowers began to work closely with Turing, Newman, <strong>and</strong> C. E.<br />

Wynn-Williams of the Telecommunications Research Establishment<br />

(TRE) to develop a machine that could break the Fish codes. The<br />

Dollis Hill team worked on the tape driving <strong>and</strong> reading problems,<br />

<strong>and</strong> Wynn-Williams’s team at TRE worked on electronic counters<br />

<strong>and</strong> the necessary circuitry. Their efforts produced the “Heath Robinson,”<br />

which could read two thous<strong>and</strong> characters per second. The<br />

Heath Robinson used vacuum tubes, an uncommon component in<br />

the early 1940’s. The vacuum tubes performed more reliably <strong>and</strong><br />

rapidly than the relays that had been used for counters. Heath Robinson<br />

<strong>and</strong> the companion machines proved that high-speed electronic<br />

devices could successfully do cryptoanalytic work (solve decoding<br />

problems).<br />

Entirely automatic in operation once started, the Heath Robinson<br />

was put together at Bletchley Park in the spring of 1943. The Heath<br />

Robinson became obsolete for codebreaking shortly after it was put<br />

into use, so work began on a bigger, faster, <strong>and</strong> more powerful machine:<br />

the Colossus.<br />

Flowers led the team that designed <strong>and</strong> built the Colossus in<br />

eleven months at Dollis Hill. The first Colossus (Mark I) was a bigger,<br />

faster version of the Heath Robinson <strong>and</strong> read about five thou-


202 / Colossus computer<br />

s<strong>and</strong> characters per second. Colossus had approximately fifteen<br />

hundred vacuum tubes, which was the largest number that had<br />

ever been used at that time. Although Turing <strong>and</strong> Wynn-Williams<br />

were not directly involved with the design of the Colossus, their<br />

previous work on the Heath Robinson was crucial to the project,<br />

since the first Colossus was based on the Heath Robinson.<br />

Colossus became operational at Bletchley Park in December,<br />

1943, <strong>and</strong> Flowers made arrangements for the manufacture of its<br />

components in case other machines were required. The request for<br />

additional machines came in March, 1944. The second Colossus, the<br />

Mark II, was extensively redesigned <strong>and</strong> was able to read twentyfive<br />

thous<strong>and</strong> characters per second because it was capable of performing<br />

parallel operations (carrying out several different operations<br />

at once, instead of one at a time); it also had a short-term<br />

memory. The Mark II went into operation on June 1, 1944. More<br />

machines were made, each with further modifications, until there<br />

were ten. The Colossus machines were special-purpose, programcontrolled<br />

electronic digital computers, the only known electronic<br />

programmable computers in existence in 1944. The use of electronics<br />

allowed for a tremendous increase in the internal speed of the<br />

machine.<br />

Impact<br />

The Colossus machines gave Britain the best codebreaking machines<br />

of World War II <strong>and</strong> provided information that was crucial<br />

for the Allied victory. The information decoded by Colossus, the actual<br />

messages, <strong>and</strong> their influence on military decisions would remain<br />

classified for decades after the war.<br />

The later work of several of the people involved with the Bletchley<br />

Park projects was important in British computer development<br />

after the war. Newman’s <strong>and</strong> Turing’s postwar careers were closely<br />

tied to emerging computer advances. Newman, who was interested<br />

in the impact of computers on mathematics, received a grant from<br />

the Royal Society in 1946 to establish a calculating machine laboratory<br />

at Manchester University. He was also involved with postwar<br />

computer growth in Britain.<br />

Several other members of the Bletchley Park team, including Tu-


ing, joined Newman at Manchester in 1948. Before going to Manchester<br />

University, however, Turing joined Britain’s National Physical<br />

Laboratory (NPL). At NPL, Turing worked on an advanced<br />

computer known as the Pilot Automatic Computing Engine (Pilot<br />

ACE). While at NPL, Turing proposed the concept of a stored program,<br />

which was a controversial but extremely important idea in<br />

computing. A “stored” program is one that remains in residence inside<br />

the computer, making it possible for a particular program <strong>and</strong><br />

data to be fed through an input device simultaneously. (The Heath<br />

Robinson <strong>and</strong> Colossus machines were limited by utilizing separate<br />

input tapes, one for the program <strong>and</strong> one for the data to be analyzed.)<br />

Turing was among the first to explain the stored-program<br />

concept in print. He was also among the first to imagine how subroutines<br />

could be included in a program. (A subroutine allows separate<br />

tasks within a large program to be done in distinct modules; in<br />

effect, it is a detour within a program. After the completion of the<br />

subroutine, the main program takes control again.)<br />

See also Apple II computer; Differential analyzer; ENIAC computer;<br />

IBM Model 1401 computer; Personal computer; Supercomputer;<br />

UNIVAC computer.<br />

Further Reading<br />

Colossus computer / 203<br />

Carter, Frank. Codebreaking with the Colossus Computer: Finding the K-<br />

Wheel Patterns—An Account of Some of the Techniques Used. Milton<br />

Keynes, Engl<strong>and</strong>: Bletchley Park Trust, 1997.<br />

Gray, Paul. “Computer Scientist: Alan Turing.” Time 153, no. 12<br />

(March 29, 1999).<br />

Hodges, Andrew. Alan Turing: The Enigma. New York: Walker, 2000.<br />

Sale, Tony. The Colossus Computer, 1943-1996: And How It Helped to<br />

Break the German Lorenz Cipher in World War II. Cleobury Mortimer:<br />

M&M Baldwin, 1998.


204<br />

Communications satellite<br />

Communications satellite<br />

The invention: Telstar I, the world’s first commercial communications<br />

satellite, opened the age of live, worldwide television by<br />

connecting the United States <strong>and</strong> Europe.<br />

The people behind the invention:<br />

Arthur C. Clarke (1917- ), a British science-fiction writer<br />

who in 1945 first proposed the idea of using satellites as<br />

communications relays<br />

John R. Pierce (1910- ), an American engineer who worked<br />

on the Echo <strong>and</strong> Telstar satellite communications projects<br />

Science Fiction?<br />

In 1945, Arthur C. Clarke suggested that a satellite orbiting high<br />

above the earth could relay television signals between different stations<br />

on the ground, making for a much wider range of transmission<br />

than that of the usual ground-based systems. Writing in the<br />

February, 1945, issue of Wireless World, Clarke said that satellites<br />

“could give television <strong>and</strong> microwave coverage to the entire<br />

planet.”<br />

In 1956, John R. Pierce at the Bell Telephone Laboratories of the<br />

American Telephone & Telegraph Company (AT&T) began to urge<br />

the development of communications satellites. He saw these satellites<br />

as a replacement for the ocean-bottom cables then being used to<br />

carry transatlantic telephone calls. In 1950, about one-<strong>and</strong>-a-half<br />

million transatlantic calls were made, <strong>and</strong> that number was expected<br />

to grow to three million by 1960, straining the capacity of the<br />

existing cables; in 1970, twenty-one million calls were made.<br />

Communications satellites offered a good, cost-effective alternative<br />

to building more transatlantic telephone cables. On January 19,<br />

1961, the Federal Communications Commission (FCC) gave permission<br />

for AT&T to begin Project Telstar, the first commercial communications<br />

satellite bridging the Atlantic Ocean. AT&T reached an<br />

agreement with the National Aeronautics <strong>and</strong> Space Administration<br />

(NASA) in July, 1961, in which AT&T would pay $3 million for


each Telstar launch. The Telstar project involved about four hundred<br />

scientists, engineers, <strong>and</strong> technicians at the Bell Telephone<br />

Laboratories, twenty more technical personnel at AT&T headquarters,<br />

<strong>and</strong> the efforts of more than eight hundred other companies<br />

that provided equipment or services.<br />

Telstar 1 was shaped like a faceted sphere, was 88 centimeters in<br />

diameter, <strong>and</strong> weighed 80 kilograms. Most of its exterior surface<br />

(sixty of the seventy-four facets) was covered by 3,600 solar cells to<br />

convert sunlight into 15 watts of electricity to power the satellite.<br />

Each solar cell was covered with artificial sapphire to reduce the<br />

damage caused by radiation. The main instrument was a two-way<br />

radio able to h<strong>and</strong>le six hundred telephone calls at a time or one<br />

television channel.<br />

The signal that the radio would send back to Earth was very<br />

weak—less than one-thirtieth the energy used by a household light<br />

bulb. Large ground antennas were needed to receive Telstar’s faint<br />

signal. The main ground station was built by AT&T in Andover,<br />

Maine, on a hilltop informally called “Space Hill.” A horn-shaped<br />

antenna, weighing 380 tons, with a length of 54 meters <strong>and</strong> an open<br />

end with an area of 1,097 square meters, was mounted so that it<br />

could rotate to track Telstar across the sky. To protect it from wind<br />

<strong>and</strong> weather, the antenna was built inside an inflated dome, 64 meters<br />

in diameter <strong>and</strong> 49 meters tall. It was, at the time, the largest inflatable<br />

structure ever built. A second, smaller horn antenna in<br />

Holmdel, New Jersey, was also used.<br />

International Cooperation<br />

Communications satellite / 205<br />

In February, 1961, the governments of the United States <strong>and</strong> Engl<strong>and</strong><br />

agreed to let the British Post Office <strong>and</strong> NASA work together<br />

to test experimental communications satellites. The British Post Office<br />

built a 26-meter-diameter steerable dish antenna of its own design<br />

at Goonhilly Downs, near Cornwall, Engl<strong>and</strong>. Under a similar<br />

agreement, the French National Center for Telecommunications<br />

Studies constructed a ground station, almost identical to the Andover<br />

station, at Pleumeur-Bodou, Brittany, France.<br />

After testing, Telstar 1 was moved to Cape Canaveral, Florida,<br />

<strong>and</strong> attached to the Thor-Delta launch vehicle built by the Douglas


206 / Communications satellite<br />

Aircraft Company. The Thor-Delta was launched at 3:35 a.m. eastern<br />

st<strong>and</strong>ard time (EST) on July 10, 1962. Once in orbit, Telstar 1 took<br />

157.8 minutes to circle the globe. The satellite came within range of<br />

the Andover station on its sixth orbit, <strong>and</strong> a television test pattern<br />

was transmitted to the satellite at 6:26 p.m. EST. At 6:30 p.m. EST, a<br />

tape-recorded black-<strong>and</strong>-white image of the American flag with the<br />

Andover station in the background, transmitted from Andover to<br />

Holmdel, opened the first television show ever broadcast by satellite.<br />

Live pictures of U.S. vice president Lyndon B. Johnson <strong>and</strong><br />

other officials gathered at Carnegie Institution in Washington, D.C.,<br />

followed on the AT&T program carried live on all three American<br />

networks.<br />

Up to the moment of launch, it was uncertain if the French station<br />

would be completed in time to participate in the initial test. At 6:47<br />

p.m. EST, however, Telstar’s signal was picked up by the station in<br />

Pleumeur-Bodou, <strong>and</strong> Johnson’s image became the first television<br />

transmission to cross the Atlantic. Pictures received at the French<br />

station were reported to be so clear that they looked like they had<br />

been sent from only forty kilometers away. Because of technical difficulties,<br />

the English station was unable to receive a clear signal.<br />

The first formal exchange of programming between the United<br />

States <strong>and</strong> Europe occurred on July 23, 1962. This special eighteenminute<br />

program, produced by the European Broadcasting Union,<br />

consisted of live scenes from major cities throughout Europe <strong>and</strong><br />

was transmitted from Goonhilly Downs, where the technical difficulties<br />

had been corrected, to Andover via Telstar.<br />

On the previous orbit, a program entitled “America, July 23,<br />

1962,” showing scenes from fifty television cameras around the<br />

United States, was beamed from Andover to Pleumeur-Bodou <strong>and</strong><br />

seen by an estimated one hundred million viewers throughout Europe.<br />

Consequences<br />

Telstar 1 <strong>and</strong> the communications satellites that followed it revolutionized<br />

the television news <strong>and</strong> sports industries. Before, television<br />

networks had to ship film across the oceans, meaning delays of<br />

hours or days between the time an event occurred <strong>and</strong> the broadcast


of pictures of that event on television on another continent. Now,<br />

news of major significance, as well as sporting events, can be viewed<br />

live around the world. The impact on international relations also<br />

was significant, with world opinion becoming able to influence the<br />

actions of governments <strong>and</strong> individuals, since those actions could<br />

be seen around the world as the events were still in progress.<br />

More powerful launch vehicles allowed new satellites to be placed<br />

in geosynchronous orbits, circling the earth at a speed the same as<br />

the earth’s rotation rate. When viewed from the ground, these satellites<br />

appeared to remain stationary in the sky. This allowed continuous<br />

communications <strong>and</strong> greatly simplified the ground antenna<br />

system. By the late 1970’s, private individuals had built small antennas<br />

in their backyards to receive television signals directly from the<br />

satellites.<br />

See also Artificial satellite; Cruise missile; Rocket; Weather satellite.<br />

Further Reading<br />

Communications satellite / 207<br />

McAleer, Neil. Odyssey: The Authorised Biography of Arthur C. Clarke.<br />

London: Victor Gollancz, 1992.<br />

Pierce, John Robinson. The Beginnings of Satellite Communications.<br />

San Francisco: San Francisco Press, 1968.<br />

_____. Science, Art, <strong>and</strong> Communication. New York: C. N. Potter, 1968.


208<br />

Community antenna television<br />

Community antenna television<br />

The invention: A system for connecting households in isolated areas<br />

to common antennas to improve television reception, community<br />

antenna television was a forerunner of modern cabletelevision<br />

systems.<br />

The people behind the invention:<br />

Robert J. Tarlton, the founder of CATV in eastern Pennsylvania<br />

Ed Parsons, the founder of CATV in Oregon<br />

Ted Turner (1938- ), founder of the first cable superstation,<br />

WTBS<br />

Growing Dem<strong>and</strong> for Television<br />

Television broadcasting in the United States began in the late<br />

1930’s. After delays resulting from World War II, it exploded into<br />

the American public’s consciousness. The new medium relied primarily<br />

on existing broadcasting stations that quickly converted<br />

from radio to television formats. Consequently, the reception of television<br />

signals was centralized in large cities. The dem<strong>and</strong> for television<br />

quickly swept across the country. Ownership of television receivers<br />

increased dramatically, <strong>and</strong> those who could not afford their<br />

own flocked to businesses, usually taverns, or to the homes of<br />

friends with sets. People in urban areas had more opportunities to<br />

view the new medium <strong>and</strong> had the advantage of more broadcasts<br />

within the range of reception. Those in outlying regions were not so<br />

fortunate, as they struggled to see fuzzy pictures <strong>and</strong> were, in some<br />

cases, unable to receive a signal at all.<br />

The situation for outlying areas worsened in 1948, when the Federal<br />

Communications Commission (FCC) implemented a ban on all<br />

new television stations while it considered how to exp<strong>and</strong> the television<br />

market <strong>and</strong> how to deal with a controversy over color reception.<br />

This left areas without nearby stations in limbo, while people<br />

in areas with established stations reaped the benefits of new programming.<br />

The ban would remain in effect until 1952, when new<br />

stations came under construction across the country.


Poor reception in some areas <strong>and</strong> the FCC ban on new station<br />

construction together set the stage for the development of Community<br />

Antenna Television (CATV). CATV did not have a glamorous<br />

beginning. Late in 1949, two different men, frustrated by the slow<br />

movement of television to outlying areas, set up what would become<br />

the foundation of the multimillion-dollar cable industry.<br />

Robert J. Tarlton was a radio salesman in Lansford, Pennsylvania,<br />

about sixty-five miles from Philadelphia. He wanted to move<br />

into television sales but lived in an area with poor reception. Together<br />

with friends, he founded Panther Valley Television <strong>and</strong> set<br />

up a master antenna in a mountain range that blocked the reception<br />

of Philadelphia-based broadcasting. For an installation fee of $125<br />

<strong>and</strong> a fee of $3 per month, Panther Valley Television offered residents<br />

clear reception of the three Philadelphia stations via a coaxial<br />

cable wired to their homes. At the same time, Ed Parsons, of KAST<br />

radio in Astoria, Oregon, linked homes via coaxial cables to a master<br />

antenna set up to receive remote broadcasts. Both systems offered<br />

three channels, the major network affiliates, to subscribers. By<br />

1952, when the FCC ban was lifted, some seventy CATV systems<br />

provided small <strong>and</strong> rural communities with the wonders of television.<br />

That same year, the National Cable Television Association was<br />

formed to represent the interests of the young industry.<br />

Early systems could carry only one to three channels. In 1953,<br />

CATV began to use microwave relays, which could import distant<br />

signals to add more variety <strong>and</strong> pushed system capability to twelve<br />

channels. A system of towers began sprouting up across the country.<br />

These towers could relay a television signal from a powerful<br />

originating station to each cable system’s main antenna. This further<br />

opened the reception available to subscribers.<br />

Pay Television<br />

Community antenna television / 209<br />

The notion of pay television also began at this time. In 1951, the<br />

FCC authorized a test of Zenith Radio Corporation’s Phonevision in<br />

Chicago. Scrambled images could be sent as electronic impulses<br />

over telephone lines, then unscrambled by devices placed in subscribers’<br />

homes. Subscribers could order a film over the telephone<br />

for a minimal cost, usually $1. Advertisers for the system promoted


210 / Community antenna television<br />

the idea of films for the “sick, aged, <strong>and</strong> sitterless.” This early test<br />

was a forerunner of the premium, or pay, channels of later decades.<br />

Network opposition to CATV came in the late 1950’s. RCA chairman<br />

David Sarnoff warned against a pay television system that<br />

could soon fall under government regulation, as in the case of utilities.<br />

In April, 1959, the FCC found no basis for asserting jurisdiction<br />

or authority over CATV. This left the industry open to tremendous<br />

growth.<br />

By 1960, the industry included 640 systems with 700,000 subscribers.<br />

Ten years later, 2,490 systems were in operation, serving<br />

more than 4.5 million households. This accelerated growth came at<br />

a price. In April, 1965, the FCC reversed itself <strong>and</strong> asserted authority<br />

over microwave-fed CATV. A year later, the entire cable system<br />

came under FCC control. The FCC quickly restricted the use of distant<br />

signals in the largest hundred markets.<br />

The FCC movement to control cable systems stemmed from the<br />

agency’s desire to balance the television market. From the onset of<br />

television broadcasting, the FCC strived to maintain a balanced programming<br />

schedule. The goal was to create local markets in which<br />

local affiliate stations prospered from advertising <strong>and</strong> other community<br />

support <strong>and</strong> would not be unduly harmed by competition<br />

from larger metropolitan stations. In addition, growth of the industry<br />

ideally was to be uniform, with large <strong>and</strong> small cities receiving<br />

equal consideration. Cable systems, particularly those that could receive<br />

distant signals via microwave relay, upset the balance. For example,<br />

a small Ohio town could receive New York channels as well<br />

as Chicago channels via cable, as opposed to receiving only the<br />

channels from one city.<br />

The balance was further upset with the creation of a new communications<br />

satellite, COMSAT, in 1963. This technology allowed a<br />

signal to be sent to the satellite, retransmitted back to Earth, <strong>and</strong><br />

then picked up by a receiving station. This further increased the<br />

range of cable offerings <strong>and</strong> moved the transmission of television<br />

signals to a national scale, as microwave-relayed transmissions<br />

worked best in a regional scope. These two factors led the FCC to<br />

freeze the cable industry from new development <strong>and</strong> construction<br />

in December, 1968. After 1972, when the cable freeze was lifted, the<br />

greatest impact of CATV would be felt.


Ted Turner<br />

Community antenna television / 211<br />

“The whole idea of gr<strong>and</strong> things always turned me on,” Ted<br />

Turner said in a 1978 Playboy magazine interview. Irrepressible,<br />

tenacious, <strong>and</strong> flamboyant, Turner was groomed from childhood<br />

for gr<strong>and</strong>ness.<br />

Born Robert Edward Turner III in 1938 in Cincinnati, Ohio,<br />

he was raised by a harsh, dem<strong>and</strong>ing father who sent him to<br />

military preparatory schools <strong>and</strong> insisted he study business at<br />

Brown University instead of attending the U.S. Naval Academy,<br />

as the son wanted. Known as “Terrible Ted” in<br />

school for his high-energy, maverick ways, he became<br />

an champion debater, expert sailor, <strong>and</strong> natural<br />

leader. When the Turner Advertising Company failed<br />

in 1960, <strong>and</strong> his father committed suicide, young<br />

Turner took it over <strong>and</strong> parlayed it into an empire, acquiring<br />

or creating television stations <strong>and</strong> revolutionizing<br />

how they were broadcast to Americans.<br />

From then on he acquired, innovated, <strong>and</strong>, often,<br />

shocked. He bought the Atlanta Braves baseball team<br />

<strong>and</strong> Hawks basketball team, often angering sports<br />

executives with his recruiting methods, earning the<br />

nicknames “Mouth of the South” <strong>and</strong> “Captain Outrageous”<br />

for his assertiveness. He won the prestigious America’s Cup in<br />

1977 at the helm of the yacht Courageous. He bought Metro-<br />

Golden-Mayer/United Artists <strong>and</strong> incensed movie purists by<br />

having black-<strong>and</strong>-white classics “colorized.” In 1995 he concluded<br />

a $7.5 billion merger of Turner Broadcasting <strong>and</strong> Time<br />

Warner <strong>and</strong> set about an insult-slinging business war with another<br />

media tycoon, Rupert Murdoch. Meanwhile, he went<br />

through three marriages, the last to movie star Jane Fonda, <strong>and</strong><br />

became the largest private l<strong>and</strong>holder in the nation, with luxury<br />

homes in six states.<br />

However, Turner’s life was not all acquisition. He started a<br />

charitable foundation <strong>and</strong> sponsored the Olympics-like Goodwill<br />

Games between the United States <strong>and</strong> the Soviet Union to<br />

improve relations, for which Time magazine named him its man<br />

of the year in 1991. However, Turner’s gr<strong>and</strong>est shocker came<br />

in 1997 when he promised to donate $1 billion—$100 million<br />

each year for a decade—to the United Nations to help in feeding<br />

the poor, resettling refugees, <strong>and</strong> eradicating l<strong>and</strong> mines.<br />

And he publicly challenged other super-rich people to use their<br />

vast wealth similarly.<br />

(George Bennett)


212 / Community antenna television<br />

Impact<br />

The founding of cable television had a two-tier effect on the<br />

American public. The immediate impact of CATV was the opening<br />

of television to areas cut off from network broadcasting as a result of<br />

distance or topographical obstructions. Cable brought television to<br />

those who would have otherwise missed the early years of the medium.<br />

As technology furthered the capabilities of the industry, a second<br />

impact emerged. Along with the 1972 lifting of the ban on cable expansion,<br />

the FCC established strict guidelines for the advancement<br />

of the industry. Issuing a 500-page blueprint for the expansion of cable,<br />

the FCC included limits on the use of imported distant signals,<br />

required the blacking out of some specific programs (films <strong>and</strong> serials,<br />

for example), <strong>and</strong> limited pay cable to films that were more than<br />

two years old <strong>and</strong> to sports.<br />

Another component of the guidelines required all systems that<br />

went into operation after March, 1972 (<strong>and</strong> all systems by March,<br />

1977), to provide public access channels for education <strong>and</strong> local<br />

government. In addition, channels were to be made available for<br />

lease. These access channels opened information to subscribers that<br />

would not normally be available. Local governments <strong>and</strong> school<br />

boards began to broadcast meetings, <strong>and</strong> even high school athletics<br />

soon appeared via public access channels. These channels also<br />

provided space to local educational institutions for home-based<br />

courses in a variety of disciplines.<br />

Cable Communications Policy Act<br />

Further FCC involvement came in the 1984 Cable Communications<br />

Policy Act, which deregulated the industry <strong>and</strong> opened the<br />

door for more expansion. This act removed local control over cable<br />

service rates <strong>and</strong> virtually made monopolies out of local providers<br />

by limiting competition. The late 1980’s brought a new technology,<br />

fiber optics, which promised to further advance the industry by increasing<br />

the quality of cable services <strong>and</strong> channel availability.<br />

One area of the cable industry, pay television, took off in the<br />

1970’s <strong>and</strong> early 1980’s. The first major pay channel was developed


Community antenna television / 213<br />

by the media giant Time-Life. It inaugurated Home Box Office<br />

(HBO) in 1975 as the first national satellite interconnected network.<br />

Early HBO programming primarily featured films but included no<br />

films less than two years old (meeting the 1972 FCC guidelines), no<br />

serials, <strong>and</strong> no advertisements. Other premium movie channels followed,<br />

including Showtime, Cinemax, <strong>and</strong> The Movie Channel. By<br />

the late 1970’s, cable systems offered multiple premium channels to<br />

their subscribers.<br />

Superstations were another component of the cable industry that<br />

boomed in the 1970’s <strong>and</strong> 1980’s. The first, WTBS, was owned <strong>and</strong><br />

operated by Ted Turner <strong>and</strong> broadcast from Atlanta, Georgia. It emphasized<br />

films <strong>and</strong> reruns of old television series. Cable systems<br />

that broadcast WTBS were asked to allocate the signal to channel 17,<br />

thus creating uniformity across the country for the superstation.<br />

Chicago’s WGN <strong>and</strong> New York City’s WOR soon followed, gaining<br />

access to homes across the nation via cable. Both these superstations<br />

emphasized sporting events in the early years <strong>and</strong> exp<strong>and</strong>ed to include<br />

films <strong>and</strong> other entertainment in the 1980’s.<br />

Both pay channels <strong>and</strong> superstations transmitted via satellites<br />

(WTBS leased space from RCA, for example) <strong>and</strong> were picked up by<br />

cable systems across the country. Other stations with broadcasts intended<br />

solely for the cable industry opened in the 1980’s. Ted Turner<br />

started the Cable News Network in 1980 <strong>and</strong> followed with the allnews<br />

network Headline News. He added another channel with the<br />

Turner Network Television (TNT) in 1988. Other 1980’s additions<br />

included The Disney Channel, ESPN, The Entertainment Channel,<br />

The Discovery Channel, <strong>and</strong> Lifetime. The Cable-Satellite <strong>Public</strong> Affairs<br />

Network (C-SPAN) enhanced the cable industry’s presence in<br />

Washington, D.C., by broadcasting sessions of the House of Representatives.<br />

Specialized networks for particular audiences also developed.<br />

Music Television (MTV), featuring songs played along with video<br />

sequences, premiered in 1981. Nickelodeon, a children’s channel,<br />

<strong>and</strong> VH-1, a music channel aimed at baby boomers rather than<br />

MTV’s teenage audience, reflected the movement toward specialization.<br />

Other specialized channels, such as the Sci-Fi Channel <strong>and</strong> the<br />

Comedy Channel, went even further in targeting specific audiences.


214 / Community antenna television<br />

Cable <strong>and</strong> the <strong>Public</strong><br />

The impact on the American public was tremendous. Information<br />

<strong>and</strong> entertainment became available around the clock. Cable<br />

provided a new level of service, information, <strong>and</strong> entertainment unavailable<br />

to nonsubscribers. One phenomenon that exploded in the<br />

late 1980’s was home shopping. Via The Home Shopping Club <strong>and</strong><br />

QVC, two shopping channels offered through cable television, the<br />

American public could order a full range of products. Everything<br />

from jewelry to tools <strong>and</strong> home cleaning supplies to clothing <strong>and</strong><br />

electronics was available to anyone with a credit card. Americans<br />

could now go shopping from home.<br />

The cable industry was not without its competitors <strong>and</strong> critics. In<br />

the 1980’s, the videocassette recorder (VCR) opened the viewing<br />

market. Prerecorded cassettes of recent film releases as well as classics<br />

were made available for purchase or for a small rental fee. National<br />

chains of video rental outlets, such as Blockbuster Video <strong>and</strong><br />

Video Towne, offered thous<strong>and</strong>s of titles for rent. Libraries also began<br />

to stock films. This created competition for the cable industry, in<br />

particular the premium movie channels. To combat this competition,<br />

channels began to offer original productions unavailable on<br />

videocassette. The combined effect of the cable industry <strong>and</strong> the<br />

videocassette market was devastating to the motion picture industry.<br />

The wide variety of programming available at home encouraged<br />

the American public, especially baby boomers with children,<br />

to stay home <strong>and</strong> watch cable or rented films instead of going to theaters.<br />

Critics of the cable industry seized on the violence, sexual content,<br />

<strong>and</strong> graphic language found in some of cable’s offerings. One<br />

parent responded by developing a lockout device that could make<br />

certain channels unavailable to children. Some premium channels<br />

developed an after-hours programming schedule that aired adulttheme<br />

programming only late at night. Another criticism stemmed<br />

from the repetition common on pay channels. As a result of the limited<br />

supply of <strong>and</strong> large dem<strong>and</strong> for films, pay channels were forced<br />

to repeat programs several times within a month <strong>and</strong> to rebroadcast<br />

films that were several years old. This led consumers to question the<br />

value of the additional monthly fee paid for such channels. To com-


at the problem, premium channels increased efforts aimed at original<br />

production <strong>and</strong> added more films that had not been box-office hits.<br />

By the early 1990’s, as some eleven thous<strong>and</strong> cable systems were<br />

serving 56.2 million subscribers, a new cry for regulation began. Debates<br />

over services <strong>and</strong> increasingly high rates led the FCC <strong>and</strong><br />

Congress to investigate the industry, opening the door for new<br />

guidelines on the cable industry. The non-cable networks—American<br />

Broadcasting Company (ABC), Columbia Broadcasting System<br />

(CBS), National Broadcasting Company (NBC), <strong>and</strong> newcomer<br />

Fox—stressed their concerns about the cable industry. These networks<br />

provided free programming, <strong>and</strong> cable systems profited<br />

from inclusion of network programming. Television industry representatives<br />

expressed the opinion that cable providers should pay<br />

for the privilege of retransmitting network broadcasts.<br />

The impact on cable’s subscribers, especially concerning monthly<br />

cable rates, came under heavy debate in public <strong>and</strong> government forums.<br />

The administration in Washington, D.C., expressed concern<br />

that cable rates had risen too quickly <strong>and</strong> for no obvious reason other<br />

than profit-seeking by what were essentially monopolistic local cable<br />

systems. What was clear was that the cable industry had transformed<br />

the television experience <strong>and</strong> was going to remain a powerful force<br />

within the medium. Regulators <strong>and</strong> television industry leaders were<br />

left to determine how to maintain an equitable coexistence within the<br />

medium.<br />

See also Color television; Communications satellite; Fiber-optics;<br />

Telephone switching; Television.<br />

Further Reading<br />

Community antenna television / 215<br />

Baldwin, Thomas F., <strong>and</strong> D. Stevens McVoy. Cable Communication.<br />

Englewood Cliffs, N.J.: Prentice-Hall, 1983.<br />

Brenner, Daniel L., <strong>and</strong> Monroe E. Price. Cable Television <strong>and</strong> Other<br />

Nonbroadcast Video: Law <strong>and</strong> Policy. New York: Clark Boardman,<br />

1986.<br />

Burns, R. W. Television: An International History of the Formative Years.<br />

London: Institution of Electrical Engineers in Association with<br />

the Science Museum, 1998.


216 / Community antenna television<br />

Coleman, Wim. The Age of Broadcasting: Television. Carlisle, Mass.:<br />

Discovery Enterprises, 1997.<br />

Negrine, Ralph M., ed. Cable Television <strong>and</strong> the Future of Broadcasting.<br />

New York: St. Martin’s Press, 1985.<br />

Sconce, Jeffrey. Haunted Media: Electronic Presence from Telegraphy to<br />

Television. Durham, N.C.: Duke University Press, 2000.<br />

Whittemore, Hank. CNN: The Inside Story. Boston: Little, Brown,<br />

1990.


Compact disc<br />

Compact disc<br />

The invention: A plastic disk on which digitized music or computer<br />

data is stored.<br />

The people behind the invention:<br />

Akio Morita (1921- ), a Japanese physicist <strong>and</strong> engineer<br />

who was a cofounder of Sony<br />

Wisse Dekker (1924- ), a Dutch businessman who led the<br />

Philips company<br />

W. R. Bennett (1904-1983), an American engineer who was a<br />

pioneer in digital communications <strong>and</strong> who played an<br />

important part in the Bell Laboratories research program<br />

Digital Recording<br />

217<br />

The digital system of sound recording, like the analog methods<br />

that preceded it, was developed by the telephone companies to improve<br />

the quality <strong>and</strong> speed of telephone transmissions. The system<br />

of electrical recording introduced by Bell Laboratories in the 1920s<br />

was part of this effort. Even Edison’s famous invention of the phonograph<br />

in 1877 was originally conceived as an accompaniment to<br />

the telephone. Although developed within the framework of telephone<br />

communications, these innovations found wide applications<br />

in the entertainment industry.<br />

The basis of the digital recording system was a technique of sampling<br />

the electrical waveforms of sound called PCM, or pulse code<br />

modulation. PCM measures the characteristics of these waves <strong>and</strong><br />

converts them into numbers. This technique was developed at Bell<br />

Laboratories in the 1930’s to transmit speech. At the end of World<br />

War II, engineers of the Bell System began to adapt PCM technology<br />

for ordinary telephone communications.<br />

The problem of turning sound waves into numbers was that of<br />

finding a method that could quickly <strong>and</strong> reliably manipulate millions<br />

of them. The answer to this problem was found in electronic computers,<br />

which used binary code to h<strong>and</strong>le millions of computations in a<br />

few seconds. The rapid advance of computer technology <strong>and</strong> the


218 / Compact disc<br />

semiconductor circuits that gave computers the power to h<strong>and</strong>le<br />

complex calculations provided the means to bring digital sound technology<br />

into commercial use. In the 1960’s, digital transmission <strong>and</strong><br />

switching systems were introduced to the telephone network.<br />

Pulse coded modulation of audio signals into digital code achieved<br />

st<strong>and</strong>ards of reproduction that exceeded even the best analog system,<br />

creating an enormous dynamic range of sounds with no distortion<br />

or background noise. The importance of digital recording went<br />

beyond the transmission of sound because it could be applied to all<br />

types of magnetic recording in which the source signal is transformed<br />

into an electric current. There were numerous commercial<br />

applications for such a system, <strong>and</strong> several companies began to explore<br />

the possibilities of digital recording in the 1970’s.<br />

Researchers at the Sony, Matsushita, <strong>and</strong> Mitsubishi electronics<br />

companies in Japan produced experimental digital recording systems.<br />

Each developed its own PCM processor, an integrated circuit<br />

that changes audio signals into digital code. It does not continuously<br />

transform sound but instead samples it by analyzing thous<strong>and</strong>s<br />

of minute slices of it per second. Sony’s PCM-F1 was the first<br />

analog-to-digital conversion chip to be produced. This gave Sony a<br />

lead in the research into <strong>and</strong> development of digital recording.<br />

All three companies had strong interests in both audio <strong>and</strong> video<br />

electronics equipment <strong>and</strong> saw digital recording as a key technology<br />

because it could deal with both types of information simultaneously.<br />

They devised recorders for use in their manufacturing operations.<br />

After using PCM techniques to turn sound into digital code, they recorded<br />

this information onto tape, using not magnetic audio tape but<br />

the more advanced video tape, which could h<strong>and</strong>le much more information.<br />

The experiments with digital recording occurred simultaneously<br />

with the accelerated development of video recording technology<br />

<strong>and</strong> owed much to the enhanced capabilities of video recorders.<br />

At this time, videocassette recorders were being developed in<br />

several corporate laboratories in Japan <strong>and</strong> Europe. The Sony Corporation<br />

was one of the companies developing video recorders at this<br />

time. Its U-matic machines were successfully used to record digitally.<br />

In 1972, the Nippon Columbia Company began to make its master recordings<br />

digitally on an Ampex video recording machine.


Links Among New Technologies<br />

Compact disc / 219<br />

There were powerful links between the new sound recording<br />

systems <strong>and</strong> the emerging technologies of storing <strong>and</strong> retrieving<br />

video images. The television had proved to be the most widely used<br />

<strong>and</strong> profitable electronic product of the 1950’s, but with the market<br />

for color television saturated by the end of the 1960’s, manufacturers<br />

had to look for a replacement product. A machine to save <strong>and</strong> replay<br />

television images was seen as the ideal companion to the family<br />

TV set. The great consumer electronics companies—General<br />

Electric <strong>and</strong> RCA in the United States, Philips <strong>and</strong> Telefunken in Europe,<br />

<strong>and</strong> Sony <strong>and</strong> Matsushita in Japan—began experimental programs<br />

to find a way to save video images.<br />

RCA’s experimental teams took the lead in developing an optical<br />

videodisc system, called Selectavision, that used an electronic stylus<br />

to read changes in capacitance on the disc. The greatest challenge to<br />

them came from the Philips company of Holl<strong>and</strong>. Its optical videodisc<br />

used a laser beam to read information on a revolving disc, in<br />

which a layer of plastic contained coded information. With the aid<br />

of the engineering department of the Deutsche Grammophon record<br />

company, Philips had an experimental laser disc in h<strong>and</strong> by<br />

1964.<br />

The Philips Laservision videodisc was not a commercial success,<br />

but it carried forward an important idea. The research <strong>and</strong> engineering<br />

work carried out in the laboratories at Eindhoven in Holl<strong>and</strong><br />

proved that the laser reader could do the job. More important,<br />

Philips engineers had found that this fragile device could be mass<br />

produced as a cheap <strong>and</strong> reliable component of a commercial product.<br />

The laser optical decoder was applied to reading the binary<br />

codes of digital sound. By the end of the 1970’s, Philips engineers<br />

had produced a working system.<br />

Ten years of experimental work on the Laservision system proved<br />

to be a valuable investment for the Philips corporation. Around<br />

1979, it started to work on a digital audio disc (DAD) playback system.<br />

This involved more than the basic idea of converting the output<br />

of the PCM conversion chip onto a disc. The lines of pits on the<br />

compact disc carry a great amount of information: the left- <strong>and</strong><br />

right-h<strong>and</strong> tracks of the stereo system are identified, <strong>and</strong> a sequence


220 / Compact disc<br />

of pits also controls the motor speed <strong>and</strong> corrects any error in the laser<br />

reading of the binary codes.<br />

This research was carried out jointly with the Sony Corporation<br />

of Japan, which had produced a superior method of encoding digital<br />

sound with its PCM chips. The binary codes that carried the information<br />

were manipulated by Sony’s sixteen-bit microprocessor.<br />

Its PCM chip for analog-to-digital conversion was also employed.<br />

Together, Philips <strong>and</strong> Sony produced a commercial digital playback<br />

record that they named the compact disc. The name is significant, as<br />

it does more than indicate the size of the disc—it indicates family<br />

ties with the highly successful compact cassette. Philips <strong>and</strong> Sony<br />

had already worked to establish this st<strong>and</strong>ard in the magnetic tape<br />

format <strong>and</strong> aimed to make their compact disc the st<strong>and</strong>ard for digital<br />

sound reproduction.<br />

Philips <strong>and</strong> Sony began to demonstrate their compact digital disc<br />

(CD) system to representatives of the audio industry in 1981. They<br />

were not alone in digital recording. The Japanese Victor Company, a<br />

subsidiary of Matsushita, had developed a version of digital recording<br />

from its VHD video disc design. It was called audio high density<br />

disc (AHD). Instead of<br />

the small CD disc, the AHD<br />

system used a ten-inch vinyl<br />

disc. Each digital recording<br />

system used a different<br />

PCM chip with a<br />

different rate of sampling<br />

the audio signal.<br />

The recording <strong>and</strong> electronics<br />

industries’ decision<br />

to st<strong>and</strong>ardize on the Philips/Sony<br />

CD system was<br />

therefore a major victory for<br />

these companies <strong>and</strong> an important<br />

event in the digital<br />

era of sound recording.<br />

Sony had found out the<br />

hard way that the technical<br />

performance of an innova-<br />

Although not much larger than a 3.25-inch floppy<br />

disk, a compact disk can store more than five hundred<br />

times as much data. (PhotoDisc)


tion is irrelevant when compared with the politics of turning it into<br />

an industrywide st<strong>and</strong>ard. Although the pioneer in videocassette<br />

recorders, Sony had been beaten by its rival, Matsushita, in establishing<br />

the video recording st<strong>and</strong>ard. This mistake was not repeated<br />

in the digital st<strong>and</strong>ards negotiations, <strong>and</strong> many companies were<br />

persuaded to license the new technology. In 1982, the technology<br />

was announced to the public. The following year, the compact disc<br />

was on the market.<br />

The Apex of Sound Technology<br />

Compact disc / 221<br />

The compact disc represented the apex of recorded sound technology.<br />

Simply put, here at last was a system of recording in which<br />

there was no extraneous noise—no surface noise of scratches <strong>and</strong><br />

pops, no tape hiss, no background hum—<strong>and</strong> no damage was done<br />

to the recording as it was played. In principle, a digital recording<br />

will last forever, <strong>and</strong> each play will sound as pure as the first. The<br />

compact disc could also play much longer than the vinyl record or<br />

long-playing cassette tape.<br />

Despite these obvious technical advantages, the commercial success<br />

of digital recording was not ensured. There had been several<br />

other advanced systems that had not fared well in the marketplace,<br />

<strong>and</strong> the conspicuous failure of quadrophonic sound in the 1970’s<br />

had not been forgotten within the industry of recorded sound. Historically,<br />

there were two key factors in the rapid acceptance of a new<br />

system of sound recording <strong>and</strong> reproduction: a library of prerecorded<br />

music to tempt the listener into adopting the system <strong>and</strong> a<br />

continual decrease in the price of the playing units to bring them<br />

within the budgets of more buyers.<br />

By 1984, there were about a thous<strong>and</strong> titles available on compact<br />

disc in the United States; that number had doubled by 1985. Although<br />

many of these selections were classical music—it was naturally<br />

assumed that audiophiles would be the first to buy digital<br />

equipment—popular music was well represented. The first CD available<br />

for purchase was an album by popular entertainer Billy Joel.<br />

The first CD-playing units cost more than $1,000, but Akio Morita<br />

of Sony was determined that the company should reduce the<br />

price of players even if it meant selling them below cost. Sony’s


222 / Compact disc<br />

Akio Morita<br />

Akio Morita was born in Nagoya, Japan, in 1921 into a family<br />

owning one of the country’s oldest, most prosperous sake<br />

breweries. As the eldest son, Morita was expected to take over<br />

its management from his father. However, business did not interest<br />

him as a child. Electronics did, especially radios. He made<br />

his own radio <strong>and</strong> phonograph <strong>and</strong> resolved to be a scientist.<br />

He succeeded, but in an ironic twist, he also became one of the<br />

twentieth century’s most successful businessmen.<br />

After taking a degree in physics from Osaka Imperial University<br />

in 1944, he worked at the Naval Research Center. There<br />

he met Masaru Ibuka. Although Ibuka was twelve years older<br />

<strong>and</strong> much more reserved in temperament, the two became fast<br />

friends. After World War II, they borrowed the equivalent of<br />

about $500 from Morita’s father <strong>and</strong> opened the Tokyo Telecommunications<br />

Company, making voltmeters <strong>and</strong>, later, tape<br />

recorders.<br />

To help along sluggish sales, Morita visited local schools to<br />

demonstrate the tape recorder’s usefulness in teaching. He was<br />

so successful that a third of Japan’s elementary schools bought<br />

them. From then on, Morita, as vice president of the company,<br />

was the lead man in marketing <strong>and</strong> sales strategy. He bought<br />

rights from West Electric Company to manufacture transistors<br />

in 1954, <strong>and</strong> soon the company was turning out transistor radios.<br />

Sales soared. They changed the name to Sony (based on<br />

the Latin word for sound, sonus) because it was more memorable.<br />

Despite an American bias against Japanese products—<br />

which many Americans regarded as shoddy imitations—Morita<br />

launched Sony America in 1960. In 1963 Sony became the first<br />

Japanese company to sell its stock in America <strong>and</strong> in 1970 the<br />

first to be listed on the New York Stock Exchange, opening an<br />

American factory two years later. Morita became president of<br />

Sony Corporation in 1971 <strong>and</strong> board chairman in 1976.<br />

In 1984 Sony earnings exceeded $5 billion, a ten-million percent<br />

increase in worth in less than forty years. As important for<br />

Japanese industry <strong>and</strong> national honor, Morita <strong>and</strong> Sony moved<br />

Japanese electronics into leading edge of technical sophistication<br />

<strong>and</strong> craftsmanship.


audio engineering department improved the performance of the<br />

players while reducing size <strong>and</strong> cost. By 1984, Sony had a small CD<br />

unit on the market for $300. Several of Sony’s competitors, including<br />

Matsushita, had followed its lead into digital reproduction.<br />

There were several compact disc players available in 1985 that cost<br />

less than $500. Sony quickly applied digital technology to the popular<br />

personal stereo <strong>and</strong> to automobile sound systems. Sales of CD<br />

units increased roughly tenfold from 1983 to 1985.<br />

Impact on Vinyl Recording<br />

Compact disc / 223<br />

When the compact disc was announced in 1982, the vinyl record<br />

was the leading form of recorded sound, with 273 million units sold<br />

annually compared to 125 million prerecorded cassette tapes. The<br />

compact disc sold slowly, beginning with 800,000 units shipped in<br />

1983 <strong>and</strong> rising to 53 million in 1986. By that time, the cassette tape<br />

had taken the lead, with slightly fewer than 350 million units. The<br />

vinyl record was in decline, with only about 110 million units<br />

shipped. Compact discs first outsold vinyl records in 1988. In the ten<br />

years from 1979 to 1988, the sales of vinyl records dropped nearly 80<br />

percent. In 1989, CDs accounted for more than 286 million sales, but<br />

cassettes still led the field with total sales of 446 million. The compact<br />

disc finally passed the cassette in total sales in 1992, when more<br />

than 300 million CDs were shipped, an increase of 22 percent over<br />

the figure for 1991.<br />

The introduction of digital recording had an invigorating effect<br />

on the industry of recorded sound, which had been unable to fully<br />

recover from the slump of the late 1970’s. Sales of recorded music<br />

had stagnated in the early 1980’s, <strong>and</strong> an industry accustomed to<br />

steady increases in output became eager to find a new product or<br />

style of music to boost its sales. The compact disc was the product to<br />

revitalize the market for both recordings <strong>and</strong> players. During the<br />

1980’s, worldwide sales of recorded music jumped from $12 billion<br />

to $22 billion, with about half of the sales volume accounted for by<br />

digital recordings by the end of the decade.<br />

The success of digital recording served in the long run to undermine<br />

the commercial viability of the compact disc. This was a playonly<br />

technology, like the vinyl record before it. Once users had be-


224 / Compact disc<br />

come accustomed to the pristine digital sound, they clamored for<br />

digital recording capability. The alliance of Sony <strong>and</strong> Philips broke<br />

down in the search for a digital tape technology for home use. Sony<br />

produced a digital tape system called DAT, while Philips responded<br />

with a digital version of its compact audio tape called DCC. Sony<br />

answered the challenge of DCC with its Mini Disc (MD) product,<br />

which can record <strong>and</strong> replay digitally.<br />

The versatility of digital recording has opened up a wide range of<br />

consumer products. Compact disc technology has been incorporated<br />

into the computer, in which CD-ROM readers convert the digital<br />

code of the disc into sound <strong>and</strong> images. Many home computers have<br />

the capability to record <strong>and</strong> replay sound digitally. Digital recording<br />

is the basis for interactive audio/video computer programs in which<br />

the user can interface with recorded sound <strong>and</strong> images. Philips has<br />

established a strong foothold in interactive digital technology with its<br />

CD-I (compact disc interactive) system, which was introduced in<br />

1990. This acts as a multimedia entertainer, providing sound, moving<br />

images, games, <strong>and</strong> interactive sound <strong>and</strong> image publications such as<br />

encyclopedias. The future of digital recording will be broad-based<br />

systems that can record <strong>and</strong> replay a wide variety of sounds <strong>and</strong> images<br />

<strong>and</strong> that can be manipulated by users of home computers.<br />

See also Cassette recording; Dolby noise reduction; Electronic<br />

synthesizer; FM radio; Laser-diode recording process; Optical disk;<br />

Transistor; Videocassette recorder; Walkman cassette player.<br />

Further Reading<br />

Copel<strong>and</strong>, Peter. Sound Recordings. London: British Library, 1991.<br />

Heerding, A. A Company of Many Parts. Cambridge: Cambridge University<br />

Press, 1998.<br />

Marshall, David V. Akio Morita <strong>and</strong> Sony. Watford: Exley, 1995.<br />

Morita, Akio, with Edwin M. Reingold, <strong>and</strong> Mitsuko Shimomura.<br />

Made in Japan: Akio Morita <strong>and</strong> Sony. London: HarperCollins, 1994.<br />

Nathan, John. Sony: The Private Life. Boston, Mass.: Houghton Mifflin,<br />

1999.<br />

Schlender, Brenton R. “How Sony Keeps the Magic Going.” Fortune<br />

125 (February 24, 1992).


Compressed-air-accumulating<br />

power plant<br />

Compressed-air-accumulating power plant<br />

The invention: Plants that can be used to store energy in the form<br />

of compressed air when electric power dem<strong>and</strong> is low <strong>and</strong> use it<br />

to produce energy when power dem<strong>and</strong> is high.<br />

The organization behind the invention:<br />

Nordwestdeutsche Kraftwerke, a Germany company<br />

Power, Energy Storage, <strong>and</strong> Compressed Air<br />

225<br />

Energy, which can be defined as the capacity to do work, is essential<br />

to all aspects of modern life. One familiar kind of energy, which<br />

is produced in huge amounts by power companies, is electrical energy,<br />

or electricity. Most electricity is produced in a process that consists<br />

of two steps. First, a fossil fuel such as coal is burned <strong>and</strong> the resulting<br />

heat is used to make steam. Then, the steam is used to<br />

operate a turbine system that produces electricity. Electricity has<br />

myriad applications, including the operation of heaters, home appliances<br />

of many kinds, industrial machinery, computers, <strong>and</strong> artificial<br />

illumination systems.<br />

An essential feature of electricity manufacture is the production<br />

of the particular amount of electricity that is needed at a given time.<br />

If moment-to-moment energy requirements are not met, the city or<br />

locality involved will experience a “blackout,” the most obvious<br />

feature of which is the loss of electrical lighting. To prevent blackouts,<br />

it is essential to store extra electricity at times when power production<br />

exceeds power dem<strong>and</strong>s. Then, when power dem<strong>and</strong>s exceed<br />

the capacity to make energy by normal means, stored energy<br />

can be used to make up the difference.<br />

One successful modern procedure for such storage is the compressed-air-accumulation<br />

process, pioneered by the Nordwestdeutsche<br />

Kraftwerke company’s compressed-air-accumulating power<br />

plant, which opened in December, 1978. The plant, which is<br />

located in Huntorf, Germany (at the time, West Germany), makes<br />

compressed air during periods of low electricity dem<strong>and</strong>, stores the


226 / Compressed-air-accumulating power plant<br />

air in an underground cavern, <strong>and</strong> uses it to produce extra electricity<br />

during periods of high dem<strong>and</strong>.<br />

Plant Operation <strong>and</strong> Components<br />

The German 300-megawatt compressed-air-accumulating power<br />

plant in Huntorf produces extra electricity from stored compressed<br />

air that will provide up to four hours per day of local peak electricity<br />

needs. The energy-storage process, which is vital to meeting very<br />

high peak electric power dem<strong>and</strong>s, is viable for electric power<br />

plants whose total usual electric outputs range from 25 megawatts<br />

to the 300 megawatts produced at Huntorf. It has been suggested,<br />

however, that the process is most suitable for 25- to 50-megawatt<br />

plants.<br />

The energy-storage procedure used at Huntorf is quite simple.<br />

All the surplus electricity that is made in nonpeak-dem<strong>and</strong> periods<br />

is utilized to drive an air compressor. The compressor pumps air<br />

from the surrounding atmosphere into an airtight underground<br />

storage cavern. When extra electricity is required, the stored compressed<br />

air is released <strong>and</strong> passed through a heating unit to be<br />

warmed, after which it is used to run gas-turbine systems that produce<br />

electricity. This sequence of events is the same as that used in<br />

any gas-turbine generating system; the only difference is that the<br />

compressed air can be stored for any desired period of time rather<br />

than having to be used immediately.<br />

One requirement of any compressed-air-accumulating power<br />

plant is an underground storage chamber. The Huntorf plant utilizes<br />

a cavern that was hollowed out some 450 meters below the surface<br />

of the earth. The cavern was created by drilling a hole into an<br />

underground salt deposit <strong>and</strong> pumping in water. The water dissolved<br />

the salt, <strong>and</strong> the resultant saltwater solution (brine) was<br />

pumped out of the deposit. The process of pumping in water <strong>and</strong> removing<br />

brine was continued until the cavern reached the desired<br />

size. This type of storage cavern is virtually leak-free. The preparation<br />

of such underwater salt-dome caverns has been performed<br />

roughly since the middle of the twentieth century. Until the Huntorf<br />

endeavor, such caves were used to stockpile petroleum <strong>and</strong> natural<br />

gas for later use. It is also possible to use mined, hard-rock caverns


for compressed-air accumulation when it is necessary to compress<br />

air to pressures higher than those that can be maintained effectively<br />

in a salt-dome cavern.<br />

The essential machinery that must be added to conventional<br />

power plants to turn them into compressed-air-accumulating power<br />

plants are motor-driven air compressors <strong>and</strong> gas turbine generating<br />

systems. This equipment must be connected appropriately so that<br />

in the storage mode, the overall system will compress air for storage<br />

in the underground cavern, <strong>and</strong> in the power-production mode, the<br />

system will produce electricity from the stored compressed air.<br />

Large compressed-air-accumulating power plants require specially<br />

constructed machinery. For example, the compressors that<br />

are used at Huntorf were developed specifically for that plant by<br />

Sulzer, a Swiss company. When the capacity of such plants is no<br />

higher than 50 megawatts, however, st<strong>and</strong>ard, readily available<br />

components can be used. This means that relatively small compressed-air-accumulating<br />

power plants can be constructed for a reasonable<br />

cost.<br />

Consequences<br />

Compressor<br />

Compressed-air-accumulating power plant / 227<br />

Air<br />

Electricity<br />

Out<br />

Electricity<br />

In<br />

Exhaust Stack<br />

Recuperator<br />

Clutch Motor Clutch<br />

Generator<br />

Turbine<br />

Valve Valve<br />

Combuster<br />

Burning Fuel<br />

Schematic of a compressed-air-accumulating power plant.<br />

The development of compressed-air-accumulating power plants<br />

has had a significant impact on the electric power industry, adding to<br />

its capacity to store energy. The main storage methods available prior<br />

to the development of compressed-air-accumulation methodology<br />

were batteries <strong>and</strong> water that was pumped uphill (hydro-storage). Battery<br />

technology is expensive, <strong>and</strong> its capacity is insufficient for major,<br />

long-term power storage. Hydro-storage is a more viable technology.


228 / Compressed-air-accumulating power plant<br />

Compressed-air energy-storage systems have several advantages<br />

over hydro-storage. First, they can be used in areas where flat terrain<br />

makes it impossible to use hydro-storage. Second, compressedair<br />

storage is more efficient than hydro-storage. Finally, the fact that<br />

st<strong>and</strong>ard plant components can be used, along with several other<br />

factors, means that 25- to 50-megawatt compressed-air storage plants<br />

can be constructed much more quickly <strong>and</strong> cheaply than comparable<br />

hydro-storage plants.<br />

The attractiveness of compressed-air-accumulating power plants<br />

has motivated efforts to develop hard-rock cavern construction<br />

techniques that cut costs <strong>and</strong> make it possible to use high-pressure<br />

air storage. In addition, aquifers (underground strata of porous rock<br />

that normally hold groundwater) have been used successfully for<br />

compressed-air storage. It is expected that compressed-air-accumulating<br />

power plants will be widely used in the future, which will<br />

help to decrease pollution <strong>and</strong> cut the use of fossil fuels.<br />

See also Alkaline storage battery; Breeder reactor; Fuel cell; Geothermal<br />

power; Heat pump; Nuclear power plant; Tidal power<br />

plant.<br />

Further Reading<br />

“Compressed Air Stores Electricity.” Popular Science 242, no. 5 (May,<br />

1993).<br />

Lee, Daehee. “Power to Spare: Compressed Air Energy Storage.”<br />

Mechanical Engineering 113, no. 7 (July, 1991).<br />

Shepard, Sam, <strong>and</strong> Septimus van der Linden. “Compressed Air Energy<br />

Storage Adapts Proven Technology to Address Market Opportunities.”<br />

Power Engineering 105, no. 4 (April, 2001).<br />

Zink, John C. “Who Says You Can’t Store Electricity?” Power Engineering<br />

101, no. 3 (March, 1997).


Computer chips<br />

Computer chips<br />

The invention: Also known as a microprocessor, a computer chip<br />

combines the basic logic circuits of a computer on a single silicon<br />

chip.<br />

The people behind the invention:<br />

Robert Norton Noyce (1927-1990), an American physicist<br />

William Shockley (1910-1989), an American coinventor of the<br />

transistor who was a cowinner of the 1956 Nobel Prize in<br />

Physics<br />

Marcian Edward Hoff, Jr. (1937- ), an American engineer<br />

Jack St. Clair Kilby (1923- ), an American researcher <strong>and</strong><br />

assistant vice president of Texas Instruments<br />

The Shockley Eight<br />

229<br />

The microelectronics industry began shortly after World War II<br />

with the invention of the transistor. While radar was being developed<br />

during the war, it was discovered that certain crystalline substances,<br />

such as germanium <strong>and</strong> silicon, possess unique electrical<br />

properties that make them excellent signal detectors. This class of<br />

materials became known as “semiconductors,” because they are<br />

neither conductors nor insulators of electricity.<br />

Immediately after the war, scientists at Bell Telephone Laboratories<br />

began to conduct research on semiconductors in the hope that<br />

they might yield some benefits for communications. The Bell physicists<br />

learned to control the electrical properties of semiconductor<br />

crystals by “doping” (treating) them with minute impurities. When<br />

two thin wires for current were attached to this material, a crude device<br />

was obtained that could amplify the voice. The transistor, as<br />

this device was called, was developed late in 1947. The transistor<br />

duplicated many functions of vacuum tubes; it was also smaller, required<br />

less power, <strong>and</strong> generated less heat. The three Bell Laboratories<br />

scientists who guided its development—William Shockley,<br />

Walter H. Brattain, <strong>and</strong> John Bardeen—won the 1956 Nobel Prize in<br />

Physics for their work.


230 / Computer chips<br />

Shockley left Bell Laboratories <strong>and</strong> went to Palo Alto, California,<br />

where he formed his own company, Shockley Semiconductor Laboratories,<br />

which was a subsidiary of Beckman Instruments. Palo Alto<br />

is the home of Stanford University, which, in 1954, set aside 655<br />

acres of l<strong>and</strong> for a high-technology industrial area known as Stanford<br />

Research Park. One of the first small companies to lease a site<br />

there was Hewlett-Packard. Many others followed, <strong>and</strong> the surrounding<br />

area of Santa Clara County gave rise in the 1960’s <strong>and</strong><br />

1970’s to a booming community of electronics firms that became<br />

known as “Silicon Valley.” On the strength of his prestige, Shockley<br />

recruited eight young scientists from the eastern United States to<br />

work for him. One was Robert Norton Noyce, an Iowa-bred physicist<br />

with a doctorate from the Massachusetts Institute of Technology.<br />

Noyce came to Shockley’s company in 1956.<br />

The “Shockley Eight,” as they became known in the industry,<br />

soon found themselves at odds with their boss over issues of research<br />

<strong>and</strong> development. Seven of the dissenting scientists negotiated<br />

with industrialist Sherman Fairchild, <strong>and</strong> they convinced the<br />

remaining holdout, Noyce, to join them as their leader. The Shock-<br />

Despite their tiny size, individual computer chips contain the basic logic circuits of entire<br />

computers. (PhotoDisc)


Jack St. Clair Kilby<br />

Computer chips / 231<br />

Maybe the original, deepest inspiration for the integrated<br />

circuit chip was topographical: As a boy Jack Kilby (b.1923) often<br />

accompanied his father, an electrical engineer, on trips over<br />

the circuit of roads through his flat home state, Kansas.<br />

In any case, he learned to love things electrical, <strong>and</strong> radios<br />

especially, from his father. Young Kilby had just started studying<br />

at the University of Illinois on his way to a degree in electrical<br />

engineering, when World War II started. He joined the<br />

Office of Strategic Services (OSS), which sent him into Japaneseoccupied<br />

territory to train local freedom fighters. He found the<br />

radios given to him to be heavy <strong>and</strong> unreliable, so he got hold of<br />

components on his own <strong>and</strong> built better, smaller radios.<br />

The “better, smaller” theme stayed with him. His first job<br />

out of college was with Centralab in Milwaukee, Wisconsin,<br />

where he designed ever smaller circuits. However, the bulky,<br />

hot vacuum tubes then in use limited miniaturization. In 1952,<br />

Centralab <strong>and</strong> Kilby eagerly incorporated the newly invented<br />

transistors into their designs. Kilby found, however, that all the<br />

electrical connections needed to hook up transistors <strong>and</strong> wires<br />

in a complex circuit also limited miniaturization.<br />

He moved to Texas Instruments in 1958. The company was<br />

working on a modular approach to miniaturization with snaptogether<br />

st<strong>and</strong>ardized parts. Kilby had a better idea: place everything<br />

for a specific circuit on a chip of silicon. Along with<br />

many other inventors, Kilby was soon looking for ways to put<br />

this new integrated circuit to work. He experimented with their<br />

use in computers <strong>and</strong> in generating solar power. He helped to<br />

develop the first h<strong>and</strong>-held calculator. Soon integrated circuits<br />

were in practically every electronic gadget, so that by the year<br />

2000 his invention supported an electronic equipment industry<br />

that earned more than a trillion dollars a year.<br />

Among his many awards, Kilby shared the 2000 Nobel Prize<br />

in Physics with Zhores I. Alferov <strong>and</strong> Herbert Kroemer, both of<br />

whom also miniaturized electronics.<br />

ley Eight defected in 1957 to form a new company, Fairchild Semiconductor,<br />

in nearby Mountain View, California. Shockley’s company,<br />

which never recovered from the loss of these scientists, soon<br />

went out of business.


232 / Computer chips<br />

Integrating Circuits<br />

Research efforts at Fairchild Semiconductor <strong>and</strong> Texas Instruments,<br />

in Dallas, Texas, focused on putting several transistors on<br />

one piece, or “chip,” of silicon. The first step involved making miniaturized<br />

electrical circuits. Jack St. Clair Kilby, a researcher at Texas<br />

Instruments, succeeded in making a circuit on a chip that consisted<br />

of tiny resistors, transistors, <strong>and</strong> capacitors, all of which were connected<br />

with gold wires. He <strong>and</strong> his company filed for a patent on<br />

this “integrated circuit” in February, 1959. Noyce <strong>and</strong> his associates<br />

at Fairchild Semiconductor followed in July of that year with an integrated<br />

circuit manufactured by means of a “planar process,”<br />

which involved laying down several layers of semiconductor that<br />

were isolated by layers of insulating material. Although Kilby <strong>and</strong><br />

Noyce are generally recognized as coinventors of the integrated circuit,<br />

Kilby alone received a membership in the National <strong>Inventors</strong><br />

Hall of Fame for his efforts.<br />

Consequences<br />

By 1968, Fairchild Semiconductor had grown to a point at which<br />

many of its key Silicon Valley managers had major philosophical<br />

differences with the East Coast management of their parent company.<br />

This led to a major exodus of top-level management <strong>and</strong> engineers.<br />

Many started their own companies. Noyce, Gordon E. Moore,<br />

<strong>and</strong> Andrew Grove left Fairchild to form a new company in Santa<br />

Clara called Intel with $2 million that had been provided by venture<br />

capitalist Arthur Rock. Intel’s main business was the manufacture<br />

of computer memory integrated circuit chips. By 1970, Intel was<br />

able to develop <strong>and</strong> bring to market a r<strong>and</strong>om-access memory<br />

(RAM) chip that was subsequently purchased in large quantities by<br />

several major computer manufacturers, providing large profits for<br />

Intel.<br />

In 1969, Marcian Edward Hoff, Jr., an Intel research <strong>and</strong> development<br />

engineer, met with engineers from Busicom, a Japanese firm.<br />

These engineers wanted Intel to design a set of integrated circuits for<br />

Busicom’s desktop calculators, but Hoff told them their specifications<br />

were too complex. Nevertheless, Hoff began to think about the possi-


Circuitry of a typical computer chip. (PhotoDisc)<br />

Computer chips / 233<br />

bility of incorporating all the logic circuits of a computer central processing<br />

unit (CPU) into one chip. He began to design a chip called a<br />

“microprocessor,” which, when combined with a chip that would<br />

hold a program <strong>and</strong> one that would hold data, would become a small,<br />

general-purpose computer. Noyce encouraged Hoff <strong>and</strong> his associates<br />

to continue his work on the microprocessor, <strong>and</strong> Busicom contracted<br />

with Intel to produce the chip. Frederico Faggin, who was hired from<br />

Fairchild, did the chip layout <strong>and</strong> circuit drawings.<br />

In January, 1971, the Intel team finished its first working microprocessor,<br />

the 4004. The following year, Intel made a higher-capacity<br />

microprocessor, the 8008, for Computer Terminals Corporation.<br />

That company contracted with Texas Instruments to produce a chip<br />

with the same specifications as the 8008, which was produced in<br />

June, 1972. Other manufacturers soon produced their own microprocessors.<br />

The Intel microprocessor became the most widely used computer<br />

chip in the budding personal computer industry <strong>and</strong> may<br />

take significant credit for the PC “revolution” that soon followed.<br />

Microprocessors have become so common that people use them every<br />

day without realizing it. In addition to being used in computers,


234 / Computer chips<br />

the microprocessor has found its way into automobiles, microwave<br />

ovens, wristwatches, telephones, <strong>and</strong> many other ordinary items.<br />

See also Bubble memory; Floppy disk; Hard disk; Optical disk;<br />

Personal computer; Virtual machine.<br />

Further Reading<br />

Ceruzzi, Paul E. A History of Modern Computing. Cambridge, Mass.:<br />

MIT Press, 2000.<br />

Reid, T. R. The Chip: How Two Americans Invented the Microchip <strong>and</strong><br />

Launched a Revolution. New York: R<strong>and</strong>om House, 2001.<br />

Slater, Robert. Portraits in Silicon. Cambridge, Mass.: MIT Press,<br />

1987.


Contact lenses<br />

Contact lenses<br />

The invention: Small plastic devices that fit under the eyelids, contact<br />

lenses, or “contacts,” frequently replace the more familiar<br />

eyeglasses that many people wear to correct vision problems.<br />

The people behind the invention:<br />

Leonardo da Vinci (1452-1519), an Italian artist <strong>and</strong> scientist<br />

Adolf Eugen Fick (1829-1901), a German glassblower<br />

Kevin Tuohy, an American optician<br />

Otto Wichterle (1913- ), a Czech chemist<br />

William Feinbloom (1904-1985), an American optometrist<br />

An Old Idea<br />

235<br />

There are two main types of contact lenses: hard <strong>and</strong> soft. Both<br />

types are made of synthetic polymers (plastics). The basic concept of<br />

the contact lens was conceived by Leonardo da Vinci in 1508. He<br />

proposed that vision could be improved if small glass ampules<br />

filled with water were placed in front of each eye. Nothing came of<br />

the idea until glass scleral lenses were invented by the German<br />

glassblower Adolf Fick. Fick’s large, heavy lenses covered the pupil<br />

of the eye, its colored iris, <strong>and</strong> part of the sclera (the white of the<br />

eye). Fick’s lenses were not useful, since they were painful to wear.<br />

In the mid-1930’s, however, plastic scleral lenses were developed<br />

by various organizations <strong>and</strong> people, including the German company<br />

I. G. Farben <strong>and</strong> the American optometrist William Feinbloom.<br />

These lenses were light <strong>and</strong> relatively comfortable; they<br />

could be worn for several hours at a time.<br />

In 1945, the American optician Kevin Tuohy developed corneal<br />

lenses, which covered only the cornea of the eye. Reportedly,<br />

Tuohy’s invention was inspired by the fact that his nearsighted wife<br />

could not bear scleral lenses but hated to wear eyeglasses. Tuohy’s<br />

lenses were hard contact lenses made of rigid plastic, but they were<br />

much more comfortable than scleral lenses <strong>and</strong> could be worn for<br />

longer periods of time. Soon after, other people developed soft contact<br />

lenses, which cover both the cornea <strong>and</strong> the iris. At present,


236 / Contact lenses<br />

many kinds of contact lenses are available. Both hard <strong>and</strong> soft contact<br />

lenses have advantages for particular uses.<br />

Eyes, Tears, <strong>and</strong> Contact Lenses<br />

The camera-like human eye automatically focuses itself <strong>and</strong> adjusts<br />

to the prevailing light intensity. In addition, it never runs out of<br />

“film” <strong>and</strong> makes a continuous series of visual images. In the process<br />

of seeing, light enters the eye <strong>and</strong> passes through the clear,<br />

dome-shaped cornea, through the hole (the pupil) in the colored<br />

iris, <strong>and</strong> through the clear eye lens, which can change shape by<br />

means of muscle contraction. The lens focuses the light, which next<br />

passes across the jellylike “vitreous humor” <strong>and</strong> hits the retina.<br />

There, light-sensitive retinal cells send visual images to the optic<br />

nerve, which transmits them to the brain for interpretation.<br />

Many people have 20/20 (normal) vision, which means that they<br />

can clearly see letters on a designated line of a st<strong>and</strong>ard eye chart<br />

placed 20 feet away. Nearsighted (myopic) people have vision of<br />

20/40 or worse. This means that, 20 feet from the eye chart, they see<br />

clearly what people with 20/20 vision can see clearly at a greater<br />

distance.<br />

Myopia (nearsightedness) is one of the four most common visual<br />

defects. The others are hyperopia, astigmatism, <strong>and</strong> presbyopia. All<br />

are called “refractive errors” <strong>and</strong> are corrected with appropriate<br />

eyeglasses or contact lenses. Myopia, which occurs in 30 percent of<br />

humans, occurs when the eyeball is too long for the lens’s focusing<br />

ability <strong>and</strong> images of distant objects focus before they reach the retina,<br />

causing blurry vision. Hyperopia, or farsightedness, occurs<br />

when the eyeballs are too short. In hyperopia, the eye’s lenses cannot<br />

focus images of nearby objects by the time those images reach<br />

the retina, resulting in blurry vision. A more common condition is<br />

astigmatism, in which incorrectly shaped corneas make all objects<br />

appear blurred. Finally, presbyopia, part of the aging process,<br />

causes the lens of the eye to lose its elasticity. It causes progressive<br />

difficulty in seeing nearby objects. In myopic, hyperopic, or astigmatic<br />

people, bifocal (two-lens) systems are used to correct presbyopia,<br />

whereas monofocal systems are used to correct presbyopia in<br />

people whose vision is otherwise normal.


William Feinbloom<br />

Contact lenses / 237<br />

William Feinbloom started his career in eye care when he<br />

was only three, helping his father, an optometrist, in his practice.<br />

Born in Brooklyn, New York, in 1904, Feinbloom studied at<br />

the Columbia School of Optometry <strong>and</strong> graduated at nineteen.<br />

He later earned degrees in physics, mathematics, biophysics,<br />

<strong>and</strong> psychology, all of it to help him treat people who suffered<br />

visual impairments. His many achievements on the behalf of<br />

the partially sighted won him professional accolades as the “father<br />

of low vision.”<br />

In 1932, while working in a clinic, Feinbloom produced the<br />

first of his special vision-enhancing inventions. He ground<br />

three-power lenses, imitating the primary lens of a refracting<br />

telescope, <strong>and</strong> fit them in a frame for an elderly patient whose<br />

vision could not be treated. The patient was again able to see,<br />

<strong>and</strong> when news of this miracle later reached Pope Pius XI, he<br />

sent a special blessing to Feinbloom. He soon opened his own<br />

practice <strong>and</strong> during the next fifty years invented a series of new<br />

lenses for people with macular degeneration <strong>and</strong> other vision<br />

diseases, as well as making the first set of contact lenses in<br />

America.<br />

In 1978 Feinbloom bequeathed his practice to the Pennsylvania<br />

College of Optometry, which named it the William Feinbloom<br />

Vision Rehabilitation Center. Every year the William<br />

Feinbloom Award honors a vision-care specialist who has improved<br />

the delivery <strong>and</strong> quality of optometric service. Feinbloom<br />

died in 1985.<br />

Modern contact lenses, which many people prefer to eyeglasses,<br />

are used to correct all common eye defects as well as many others<br />

not mentioned here. The lenses float on the layer of tears that is<br />

made continuously to nourish the eye <strong>and</strong> keep it moist. They fit under<br />

the eyelids <strong>and</strong> either over the cornea or over both the cornea<br />

<strong>and</strong> the iris, <strong>and</strong> they correct visual errors by altering the eye’s focal<br />

length enough to produce 20/20 vision. In addition to being more attractive<br />

than eyeglasses, contact lenses correct visual defects more effectively<br />

than eyeglasses can. Some soft contact lenses (all are made<br />

of flexible plastics) can be worn almost continuously. Hard lenses are


238 / Contact lenses<br />

made of more rigid plastic <strong>and</strong> last longer, though they can usually be<br />

worn only for six to nine hours at a time. The choice of hard or soft<br />

lenses must be made on an individual basis.<br />

The disadvantages of contact lenses include the fact that they must<br />

be cleaned frequently to prevent eye irritation. Furthermore, people<br />

who do not produce adequate amounts of tears (a condition called<br />

“dry eyes”) cannot wear them. Also, arthritis, many allergies, <strong>and</strong><br />

poor manual dexterity caused by old age or physical problems make<br />

many people poor c<strong>and</strong>idates for contact lenses.<br />

Impact<br />

The invention of Plexiglas hard scleral contact lenses set the stage<br />

for the development of the widely used corneal hard lenses by Tuohy.<br />

The development of soft contact lenses available to the general public<br />

began in Czechoslovakia in the 1960’s. It led to the sale, starting in the<br />

1970’s, of the popular, soft<br />

contact lenses pioneered by<br />

Otto Wichterle. The Wichterle<br />

lenses, which cover<br />

both the cornea <strong>and</strong> the iris,<br />

are made of a plastic called<br />

HEMA (short for hydroxyethylmethylmethacrylate).<br />

These very thin lenses<br />

have disadvantages that include<br />

the requirement of<br />

disinfection between uses,<br />

incomplete astigmatism correction,<br />

low durability, <strong>and</strong><br />

the possibility of chemical<br />

combination with some<br />

medications, which can<br />

damage the eyes. Therefore,<br />

much research is being<br />

carried out to improve<br />

Contact lenses are placed directly on the surface of<br />

the eye. (Digital Stock)<br />

them. For this reason, <strong>and</strong><br />

because of the continued


popularity of hard lenses, new kinds of soft <strong>and</strong> hard lenses are continually<br />

coming on the market.<br />

See also Artificial heart; Disposable razor; Hearing aid; Laser eye<br />

surgery; Pacemaker.<br />

Further Reading<br />

Contact lenses / 239<br />

“The Contact Lens.” Newsweek 130 (Winter, 1997/1998).<br />

Hemphill, Clara. “A Quest for Better Vision: Spectacles over the<br />

Centuries.” New York Times (August 8, 2000).<br />

Koetting, Robert A. History of the Contact Lens. Irvine, Calif.:<br />

Allergan, 1978.<br />

Lubick, Naomi. “The Hard <strong>and</strong> the Soft.” Scientific American 283, no.<br />

4 (October, 2000).


240<br />

Coronary artery bypass surgery<br />

Coronary artery bypass surgery<br />

The invention: The most widely used procedure of its type, coronary<br />

bypass surgery uses veins from legs to improve circulation<br />

to the heart.<br />

The people behind the invention:<br />

Rene Favaloro (1923-2000), a heart surgeon<br />

Donald B. Effler (1915- ), a member of the surgical team<br />

that performed the first coronary artery bypass operation<br />

F. Mason Sones (1918- ), a physician who developed an<br />

improved technique of X-raying the heart’s arteries<br />

Fighting Heart Disease<br />

In the mid-1960’s, the leading cause of death in the United States<br />

was coronary artery disease, claiming nearly 250 deaths per 100,000<br />

people. Because this number was so alarming, much research was<br />

being conducted on the heart. Most of the public’s attention was focused<br />

on heart transplants performed separately by the famous surgeons<br />

Christiaan Barnard <strong>and</strong> Michael DeBakey. Yet other, less dramatic<br />

procedures were being developed <strong>and</strong> studied.<br />

A major problem with coronary artery disease, besides the threat<br />

of death, is chest pain, or angina. Individuals whose arteries are<br />

clogged with fat <strong>and</strong> cholesterol are frequently unable to deliver<br />

enough oxygen to their heart muscles. This may result in angina,<br />

which causes enough pain to limit their physical activities. Some of<br />

the heart research in the mid-1960’s was an attempt to find a surgical<br />

procedure that would eliminate angina in heart patients. The<br />

various surgical procedures had varying success rates.<br />

In the late 1950’s <strong>and</strong> early 1960’s, a team of physicians in Clevel<strong>and</strong><br />

was studying surgical procedures that would eliminate angina.<br />

The team was composed of Rene Favaloro, Donald B. Effler, F.<br />

Mason Sones, <strong>and</strong> Laurence Groves. They were working on the concept,<br />

proposed by Dr. Arthur M. Vineberg from McGill University<br />

in Montreal, of implanting a healthy artery from the chest into the<br />

heart. This bypass procedure would provide the heart with another


Bypass<br />

Graft<br />

Blockage<br />

Before bypass surgery (left) the blockage in the artery<br />

threatens to cut off bloodflow; after surgery to<br />

graft a piece of vein (right), the blood can flow<br />

around the blockage.<br />

Coronary artery bypass surgery / 241<br />

source of blood, resulting<br />

in enough oxygen to overcome<br />

the angina. Yet Vineberg’s<br />

surgery was often<br />

ineffective because it was<br />

hard to determine exactly<br />

where to implant the new<br />

artery.<br />

New Techniques<br />

In order to make Vineberg’s<br />

proposed operation<br />

successful, better diagnostic<br />

tools were needed. This was<br />

accomplished by the work<br />

of Sones. He developed a diagnostic procedure, called “arteriography,”<br />

whereby a catheter was inserted into an artery in the arm,<br />

which he ran all the way into the heart. He then injected a dye into the<br />

coronary arteries <strong>and</strong> photographed them with a high-speed motionpicture<br />

camera. This provided an image of the heart, which made it<br />

easy to determine where the blockages were in the coronary arteries.<br />

Using this tool, the team tried several new techniques. First, the<br />

surgeons tried to ream out the deposits found in the narrow portion<br />

of the artery. They found, however, that this actually reduced<br />

blood flow. Second, they tried slitting the length of the blocked<br />

area of the artery <strong>and</strong> suturing a strip of tissue that would increase<br />

the diameter of the opening. This was also ineffective because it often<br />

resulted in turbulent blood flow. Finally, the team attempted to<br />

reroute the flow of blood around the blockage by suturing in other<br />

tissue, such as a portion of a vein from the upper leg. This bypass<br />

procedure removed that part of the artery that was clogged <strong>and</strong> replaced<br />

it with a clear vessel, thereby restoring blood flow through<br />

the artery. This new method was introduced by Favaloro in 1967.<br />

In order for Favaloro <strong>and</strong> other heart surgeons to perform coronary<br />

artery surgery successfully, several other medical techniques<br />

had to be developed. These included extracorporeal circulation <strong>and</strong><br />

microsurgical techniques.


242 / Coronary artery bypass surgery<br />

Extracorporeal circulation is the process of diverting the patient’s<br />

blood flow from the heart <strong>and</strong> into a heart-lung machine.<br />

This procedure was developed in 1953 by U.S. surgeon John H.<br />

Gibbon, Jr. Since the blood does not flow through the heart, the<br />

heart can be temporarily stopped so that the surgeons can isolate<br />

the artery <strong>and</strong> perform the surgery on motionless tissue.<br />

Microsurgery is necessary because some of the coronary arteries<br />

are less than 1.5 millimeters in diameter. Since these arteries<br />

had to be sutured, optical magnification <strong>and</strong> very delicate <strong>and</strong> sophisticated<br />

surgical tools were required. After performing this surgery<br />

on numerous patients, follow-up studies were able to determine<br />

the surgery’s effectiveness. Only then was the value of coronary artery<br />

bypass surgery recognized as an effective procedure for reducing angina<br />

in heart patients.<br />

Consequences<br />

According to the American Heart Association, approximately<br />

332,000 bypass surgeries were performed in the United States in<br />

1987, an increase of 48,000 from 1986. These figures show that the<br />

work by Favaloro <strong>and</strong> others has had a major impact on the<br />

health of United States citizens. The future outlook is also positive.<br />

It has been estimated that five million people had coronary<br />

artery disease in 1987. Of this group, an estimated 1.5 million had<br />

heart attacks <strong>and</strong> 500,000 died. Of those living, many experienced<br />

angina. Research has developed new surgical procedures <strong>and</strong><br />

new drugs to help fight coronary artery disease. Yet coronary artery<br />

bypass surgery is still a major form of treatment.<br />

See also Artificial blood; Artificial heart; Blood transfusion;<br />

Electrocardiogram; Heart-lung machine; Pacemaker.<br />

Further Reading<br />

Bing, Richard J. Cardiology: The Evolution of the Science <strong>and</strong> the<br />

Art. 2d ed. New Brunswick, N.J.: Rutgers University Press,<br />

1999.


Coronary artery bypass surgery / 243<br />

Faiola, Anthony. “Doctor’s Suicide Strikes at Heart of Argentina’s<br />

Health Care Crisis: Famed Cardiac Surgeon Championed<br />

the Poor.” Washington Post (August 25, 2000).<br />

Favaloro, René G. The Challenging Dream of Heart Surgery: From the<br />

Pampas to Clevel<strong>and</strong>. Boston: Little, Brown, 1994.


244<br />

Cruise missile<br />

Cruise missile<br />

The invention: Aircraft weapons system that makes it possible to<br />

attack both l<strong>and</strong> <strong>and</strong> sea targets with extreme accuracy without<br />

endangering the lives of the pilots.<br />

The person behind the invention:<br />

Rear Admiral Walter M. Locke (1930- ), U.S. Navy project<br />

manager<br />

From the Buzz Bombs of World War II<br />

During World War II, Germany developed <strong>and</strong> used two different<br />

types of missiles: ballistic missiles <strong>and</strong> cruise missiles. A ballistic<br />

missile is one that does not use aerodynamic lift in order to fly. It is<br />

fired into the air by powerful jet engines <strong>and</strong> reaches a high altitude;<br />

when its engines are out of fuel, it descends on its flight path toward<br />

its target. The German V-2 was the first ballistic missile. The United<br />

States <strong>and</strong> other countries subsequently developed a variety of<br />

highly sophisticated <strong>and</strong> accurate ballistic missiles.<br />

The other missile used by Germany was a cruise missile called<br />

the V-1, which was also called the flying bomb or the buzz bomb.<br />

The V-1 used aerodynamic lift in order to fly, just as airplanes do. It<br />

flew relatively low <strong>and</strong> was slow; by the end of the war, the British,<br />

against whom it was used, had developed techniques for countering<br />

it, primarily by shooting it down.<br />

After World War II, both the United States <strong>and</strong> the Soviet Union<br />

carried on the Germans’ development of both ballistic <strong>and</strong> cruise<br />

missiles. The United States discontinued serious work on cruise<br />

missile technology during the 1950’s: The development of ballistic<br />

missiles of great destructive capability had been very successful.<br />

Ballistic missiles armed with nuclear warheads had become the basis<br />

for the U.S. strategy of attempting to deter enemy attacks with<br />

the threat of a massive missile counterattack. In addition, aircraft<br />

carriers provided an air-attack capability similar to that of cruise<br />

missiles. Finally, cruise missiles were believed to be too vulnerable<br />

to being shot down by enemy aircraft or surface-to-air missiles.


While ballistic missiles are excellent for attacking large, fixed targets,<br />

they are not suitable for attacking moving targets. They can be<br />

very accurately aimed, but since they are not very maneuverable<br />

during their final descent, they are limited in their ability to change<br />

course to hit a moving target, such as a ship.<br />

During the 1967 war, the Egyptians used a Soviet-built cruise<br />

missile to sink the Israeli ship Elath. The U.S. military, primarily the<br />

Navy <strong>and</strong> the Air Force, took note of the Egyptian success <strong>and</strong><br />

within a few years initiated cruise missile development programs.<br />

The Development of Cruise Missiles<br />

Cruise missile / 245<br />

The United States probably could have developed cruise missiles<br />

similar to 1990’s models as early as the 1960’s, but it would have required<br />

a huge effort. The goal was to develop missiles that could be<br />

launched from ships <strong>and</strong> planes using existing launching equipment,<br />

could fly long distances at low altitudes at fairly high speeds,<br />

<strong>and</strong> could reach their targets with a very high degree of accuracy. If<br />

the missiles flew too slowly, they would be fairly easy to shoot<br />

down, like the German V-1’s. If they flew at too high an altitude,<br />

they would be vulnerable to the same type of surface-based missiles<br />

that shot down Gary Powers, the pilot of the U.S. U2 spyplane, in<br />

1960. If they were inaccurate, they would be of little use.<br />

The early Soviet cruise missiles were designed to meet their performance<br />

goals without too much concern about how they would<br />

be launched. They were fairly large, <strong>and</strong> the ships that launched<br />

them required major modifications. The U.S. goal of being able to<br />

launch using existing equipment, without making major modifications<br />

to the ships <strong>and</strong> planes that would launch them, played a major<br />

part in their torpedo-like shape: Sea-Launched Cruise Missiles<br />

(SLCMs) had to fit in the submarine’s torpedo tubes, <strong>and</strong> Air-<br />

Launched Cruise Missiles (ALCMs) were constrained to fit in rotary<br />

launchers. The size limitation also meant that small, efficient jet engines<br />

would be required that could fly the long distances required<br />

without needing too great a fuel load. Small, smart computers were<br />

needed to provide the required accuracy. The engine <strong>and</strong> computer<br />

technologies began to be available in the 1970’s, <strong>and</strong> they blossomed<br />

in the 1980’s.


246 / Cruise missile<br />

The U.S. Navy initiated cruise missile development efforts in<br />

1972; the Air Force followed in 1973. In 1977, the Joint Cruise Missile<br />

Project was established, with the Navy taking the lead. Rear<br />

Admiral Walter M. Locke was named project manager. The goal<br />

was to develop air-, sea-, <strong>and</strong> ground-launched cruise missiles.<br />

By coordinating activities, encouraging competition, <strong>and</strong><br />

requiring the use of common components wherever possible, the<br />

cruise missile development program became a model for future<br />

weapon-system development efforts. The primary contractors<br />

included Boeing Aerospace Company, General Dynamics, <strong>and</strong><br />

McDonnell Douglas.<br />

In 1978, SLCMs were first launched from submarines. Over the<br />

next few years, increasingly dem<strong>and</strong>ing tests were passed by several<br />

versions of cruise missiles. By the mid-1980’s, both antiship <strong>and</strong><br />

antil<strong>and</strong> missiles were available. An antil<strong>and</strong> version could be guided<br />

to its target with extreme accuracy by comparing a map programmed<br />

into its computer to the picture taken by an on-board video camera.<br />

The typical cruise missile is between 18 <strong>and</strong> 21 feet long, about 21<br />

inches in diameter, <strong>and</strong> has a wingspan of between 8 <strong>and</strong> 12 feet.<br />

Cruise missiles travel slightly below the speed of sound <strong>and</strong> have a<br />

range of around 1,350 miles (antil<strong>and</strong>) or 250 miles (antiship). Both<br />

conventionally armed <strong>and</strong> nuclear versions have been fielded.<br />

Consequences<br />

Cruise missiles have become an important part of the U.S. arsenal.<br />

They provide a means of attacking targets on l<strong>and</strong> <strong>and</strong> water<br />

without having to put an aircraft pilot’s life in danger. Their value<br />

was demonstrated in 1991 during the Persian Gulf War. One of their<br />

uses was to “soften up” defenses prior to sending in aircraft, thus reducing<br />

the risk to pilots. Overall estimates are that about 85 percent<br />

of cruise missiles used in the Persian Gulf War arrived on target,<br />

which is an outst<strong>and</strong>ing record. It is believed that their extreme accuracy<br />

also helped to minimize noncombatant casualties.<br />

See also Airplane; Atomic bomb; Hydrogen bomb; Rocket;<br />

Stealth aircraft; V-2 rocket.


Further Reading<br />

Cruise missile / 247<br />

Collyer, David G. Buzz Bomb. Deal, Kent, Engl<strong>and</strong>: Kent Aviation<br />

Historical Research Society, 1994.<br />

McDaid, Hugh, <strong>and</strong> David Oliver. Robot Warriors: The Top Secret History<br />

of the Pilotless Plane. London: Orion Media, 1997.<br />

Macknight, Nigel. Tomahawk Cruise Missile. Osceola, Wis.: Motorbooks<br />

International, 1995.<br />

Werrell, Kenneth P. The Evolution of the Cruise Missile. Maxwell Air<br />

Force Base, Ala.: Air University Press, 1997.


248<br />

Cyclamate<br />

Cyclamate<br />

The invention: An artificial sweetener introduced to the American<br />

market in 1950 under the tradename Sucaryl.<br />

The person behind the invention:<br />

Michael Sveda (1912-1999), an American chemist<br />

A Foolhardy Experiment<br />

The first synthetic sugar substitute, saccharin, was developed in<br />

1879. It became commercially available in 1907 but was banned for<br />

safety reasons in 1912. Sugar shortages during World War I (1914-<br />

1918) resulted in its reintroduction. Two other artificial sweeteners,<br />

Dulcin <strong>and</strong> P-4000, were introduced later but were banned in 1950<br />

for causing cancer in laboratory animals.<br />

In 1937, Michael Sveda was a young chemist working on his<br />

Ph.D. at the University of Illinois. A flood in the Ohio valley had ruined<br />

the local pipe-tobacco crop, <strong>and</strong> Sveda, a smoker, had been<br />

forced to purchase cigarettes. One day while in the laboratory,<br />

Sveda happened to brush some loose tobacco from his lips <strong>and</strong> noticed<br />

that his fingers tasted sweet. Having a curious, if rather foolhardy,<br />

nature, Sveda tasted the chemicals on his bench to find which<br />

one was responsible for the taste. The culprit was the forerunner of<br />

cyclohexylsulfamate, the material that came to be known as “cyclamate.”<br />

Later, on reviewing his career, Sveda explained the serendipitous<br />

discovery with the comment: “God looks after ...fools, children,<br />

<strong>and</strong> chemists.”<br />

Sveda joined E. I. Du Pont de Nemours <strong>and</strong> Company in 1939<br />

<strong>and</strong> assigned the patent for cyclamate to his employer. In June of<br />

1950, after a decade of testing on animals <strong>and</strong> humans, Abbott Laboratories<br />

announced that it was launching Sveda’s artificial sweetener<br />

under the trade name Sucaryl. Du Pont followed with its<br />

sweetener product, Cyclan. A Time magazine article in 1950 announced<br />

the new product <strong>and</strong> noted that Abbott had warned that<br />

because the product was a sodium salt, individuals with kidney<br />

problems should consult their doctors before adding it to their food.


Cyclamate had no calories, but it was thirty to forty times sweeter<br />

than sugar. Unlike saccharin, cyclamate left no unpleasant aftertaste.<br />

The additive was also found to improve the flavor of some<br />

foods, such as meat, <strong>and</strong> was used extensively to preserve various<br />

foods. By 1969, about 250 food products contained cyclamates, including<br />

cakes, puddings, canned fruit, ice cream, salad dressings,<br />

<strong>and</strong> its most important use, carbonated beverages.<br />

It was originally thought that cyclamates were harmless to the<br />

human body. In 1959, the chemical was added to the GRAS (generally<br />

recognized as safe) list. Materials on this list, such as sugar, salt,<br />

pepper, <strong>and</strong> vinegar, did not have to be rigorously tested before being<br />

added to food. In 1964, however, a report cited evidence that cyclamates<br />

<strong>and</strong> saccharin, taken together, were a health hazard. Its<br />

publication alarmed the scientific community. Numerous investigations<br />

followed.<br />

Shooting Themselves in the Foot<br />

Cyclamate / 249<br />

Initially, the claims against cyclamate had been that it caused diarrhea<br />

or prevented drugs from doing their work in the body.<br />

By 1969, these claims had begun to include the threat of cancer.<br />

Ironically, the evidence that sealed the fate of the artificial sweetener<br />

was provided by Abbott itself.<br />

A private Long Isl<strong>and</strong> company had been hired by Abbott to conduct<br />

an extensive toxicity study to determine the effects of longterm<br />

exposure to the cyclamate-saccharin mixtures often found in<br />

commercial products. The team of scientists fed rats daily doses of<br />

the mixture to study the effect on reproduction, unborn fetuses, <strong>and</strong><br />

fertility. In each case, the rats were declared to be normal. When the<br />

rats were killed at the end of the study, however, those that had been<br />

exposed to the higher doses showed evidence of bladder tumors.<br />

Abbott shared the report with investigators from the National Cancer<br />

Institute <strong>and</strong> then with the U.S. Food <strong>and</strong> Drug Administration<br />

(FDA).<br />

The doses required to produce the tumors were equivalent to an<br />

individual drinking 350 bottles of diet cola a day. That was more<br />

than one hundred times greater than that consumed even by those<br />

people who consumed a high amount of cyclamate. A six-person


250 / Cyclamate<br />

panel of scientists met to review the data <strong>and</strong> urged the ban of all cyclamates<br />

from foodstuffs. In October, 1969, amid enormous media<br />

coverage, the federal government announced that cyclamates were<br />

to be withdrawn from the market by the beginning of 1970.<br />

In the years following the ban, the controversy continued. Doubt<br />

was cast on the results of the independent study linking sweetener<br />

use to tumors in rats, because the study was designed not to evaluate<br />

cancer risks but to explain the effects of cyclamate use over<br />

many years. Bladder parasites, known as “nematodes,” found in the<br />

rats may have affected the outcome of the tests. In addition, an impurity<br />

found in some of the saccharin used in the study may have<br />

led to the problems observed. Extensive investigations such as the<br />

three-year project conducted at the National Cancer Research Center<br />

in Heidelberg, Germany, found no basis for the widespread ban.<br />

In 1972, however, rats fed high doses of saccharin alone were<br />

found to have developed bladder tumors. At that time, the sweetener<br />

was removed from the GRAS list. An outright ban was averted<br />

by the m<strong>and</strong>atory use of labels alerting consumers that certain<br />

products contained saccharin.<br />

Impact<br />

The introduction of cyclamate heralded the start of a new industry.<br />

For individuals who had to restrict their sugar intake for health<br />

reasons, or for those who wished to lose weight, there was now an<br />

alternative to giving up sweet food.<br />

The Pepsi-Cola company put a new diet drink formulation on<br />

the market almost as soon as the ban was instituted. In fact, it ran<br />

advertisements the day after the ban was announced showing the<br />

Diet Pepsi product boldly proclaiming “Sugar added—No Cyclamates.”<br />

Sveda, the discoverer of cyclamates, was not impressed with the<br />

FDA’s decision on the sweetener <strong>and</strong> its h<strong>and</strong>ling of subsequent investigations.<br />

He accused the FDA of “a massive cover-up of elemental<br />

blunders” <strong>and</strong> claimed that the original ban was based on sugar<br />

politics <strong>and</strong> bad science.<br />

For the manufacturers of cyclamate, meanwhile, the problem lay<br />

with the wording of the Delaney amendment, the legislation that


egulates new food additives. The amendment states that the manufacturer<br />

must prove that its product is safe, rather than the FDAhaving<br />

to prove that it is unsafe. The onus was on Abbott Laboratories<br />

to deflect concerns about the safety of the product, <strong>and</strong> it remained<br />

unable to do so.<br />

See also Aspartame; Genetically engineered insulin.<br />

Further Reading<br />

Cyclamate / 251<br />

Kaufman, Leslie. “Michael Sveda, the Inventor of Cyclamates, Dies<br />

at Eighty Seven.” New York Times (August 21, 1999).<br />

Lawler, Philip F. Sweet Talk: Media Coverage of Artificial Sweeteners.<br />

Washington, D.C.: Media Institute, 1986.<br />

Remington, Dennis W. The Bitter Truth About Artificial Sweeteners.<br />

Provo, Utah: Vitality House, 1987.<br />

Whelan, Elizabeth M. “The Bitter Truth About a Sweetener Scare.”<br />

Wall Street Journal (August 26, 1999).


252<br />

Cyclotron<br />

Cyclotron<br />

The invention: The first successful magnetic resonance accelerator<br />

for protons, the cyclotron gave rise to the modern era of particle<br />

accelerators, which are used by physicists to study the structure<br />

of atoms.<br />

The people behind the invention:<br />

Ernest Orl<strong>and</strong>o Lawrence (1901-1958), an American nuclear<br />

physicist who was awarded the 1939 Nobel Prize in Physics<br />

M. Stanley Livingston (1905-1986), an American nuclear<br />

physicist<br />

Niels Edlefsen (1893-1971), an American physicist<br />

David Sloan (1905- ), an American physicist <strong>and</strong> electrical<br />

engineer<br />

The Beginning of an Era<br />

The invention of the cyclotron by Ernest Orl<strong>and</strong>o Lawrence<br />

marks the beginning of the modern era of high-energy physics. Although<br />

the energies of newer accelerators have increased steadily,<br />

the principles incorporated in the cyclotron have been fundamental<br />

to succeeding generations of accelerators, many of which were also<br />

developed in Lawrence’s laboratory. The care <strong>and</strong> support for such<br />

machines have also given rise to “big science”: the massing of scientists,<br />

money, <strong>and</strong> machines in support of experiments to discover<br />

the nature of the atom <strong>and</strong> its constituents.<br />

At the University of California, Lawrence took an interest in the<br />

new physics of the atomic nucleus, which had been developed by<br />

the British physicist Ernest Rutherford <strong>and</strong> his followers in Engl<strong>and</strong>,<br />

<strong>and</strong> which was attracting more attention as the development<br />

of quantum mechanics seemed to offer solutions to problems that<br />

had long preoccupied physicists. In order to explore the nucleus of<br />

the atom, however, suitable probes were required. An artificial<br />

means of accelerating ions to high energies was also needed.<br />

During the late 1920’s, various means of accelerating alpha particles,<br />

protons (hydrogen ions), <strong>and</strong> electrons had been tried, but


none had been successful in causing a nuclear transformation when<br />

Lawrence entered the field. The high voltages required exceeded<br />

the resources available to physicists. It was believed that more than<br />

a million volts would be required to accelerate an ion to sufficient<br />

energies to penetrate even the lightest atomic nuclei. At such voltages,<br />

insulators broke down, releasing sparks across great distances.<br />

European researchers even attempted to harness lightning to accomplish<br />

the task, with fatal results.<br />

Early in April, 1929, Lawrence discovered an article by a German<br />

electrical engineer that described a linear accelerator of ions that<br />

worked by passing an ion through two sets of electrodes, each of<br />

which carried the same voltage <strong>and</strong> increased the energy of the ions<br />

correspondingly. By spacing the electrodes appropriately <strong>and</strong> using<br />

an alternating electrical field, this “resonance acceleration” of ions<br />

could speed subatomic particles to many times the energy applied<br />

in each step, overcoming the problems presented when one tried to<br />

apply a single charge to an ion all at once. Unfortunately, the spacing<br />

of the electrodes would have to be increased as the ions were accelerated,<br />

since they would travel farther between each alternation<br />

of the phase of the accelerating charge, making an accelerator impractically<br />

long in those days of small-scale physics.<br />

Fast-Moving Streams of Ions<br />

Cyclotron / 253<br />

Lawrence knew that a magnetic field would cause the ions to be<br />

deflected <strong>and</strong> form a curved path. If the electrodes were placed<br />

across the diameter of the circle formed by the ions’ path, they<br />

should spiral out as they were accelerated, staying in phase with the<br />

accelerating charge until they reached the periphery of the magnetic<br />

field. This, it seemed to him, afforded a means of producing indefinitely<br />

high voltages without using high voltages by recycling the accelerated<br />

ions through the same electrodes. Many scientists doubted<br />

that such a method would be effective. No mechanism was known<br />

that would keep the circulating ions in sufficiently tight orbits to<br />

avoid collisions with the walls of the accelerating chamber. Others<br />

tried unsuccessfully to use resonance acceleration.<br />

A graduate student, M. Stanley Livingston, continued Lawrence’s<br />

work. For his dissertation project, he used a brass cylinder 10 centi-


254 / Cyclotron<br />

Ernest Orl<strong>and</strong>o Lawrence<br />

A man of great energy <strong>and</strong> gusty temper, Ernest Orl<strong>and</strong>o<br />

Lawrence danced for joy when one of his cyclotrons accelerated<br />

a particle to more than the one million electron volts. That<br />

amount of power was important, according to contemporary<br />

theorists, because it was enough to penetrate the nucleus of a<br />

target atom. For giving physicists a tool with which to examine<br />

the subatomic realm, Lawrence received the 1939 Nobel Prize in<br />

Physics, among many other honors.<br />

The gr<strong>and</strong>son of Norwegian immigrants, Lawrence was<br />

born in Canton, South Dakota, in 1901. After high school, he<br />

went to St. Olaf’s College, the University of South Dakota, the<br />

University of Minnesota, <strong>and</strong> Yale University, where he completed<br />

a doctorate in physics in 1925. After post-graduate fellowships<br />

at Yale, he became a professor at the University of<br />

California, Berkeley, the youngest on campus. In 1936 the university<br />

made him director of its radiation laboratory. Now<br />

named the Lawrence-Livermore National Laboratory, it stayed<br />

at the forefront of physics <strong>and</strong> high-technology research ever<br />

since.<br />

Before World War II Lawrence <strong>and</strong> his brother, Dr. John<br />

Lawrence, also at the university, worked together to find practical<br />

biological <strong>and</strong> medical applications for the radioisotopes<br />

made in Lawrence’s particle accelerators. During the war Lawrence<br />

participated in the Manhattan Project, which made the<br />

atomic bomb. He was a passionate anticommunist <strong>and</strong> after the<br />

war argued before Congress for funds to develop death rays<br />

<strong>and</strong> radiation bombs from research with his cyclotrons; however,<br />

he was also an American delegate to the Geneva Conference<br />

in 1958, which sought a ban on atomic bomb tests.<br />

Lawrence helped solve the mystery of cosmic particles, invented<br />

a method for measuring ultra-small time intervals, <strong>and</strong><br />

calculated with high precision the ratio of the charge of an electron<br />

to its mass, a fundamental constant of nature. Lawrence<br />

died in 1958 in Palo Alto, California.<br />

meters in diameter sealed with wax to hold a vacuum, a half-pillbox<br />

of copper mounted on an insulated stem to serve as the electrode,<br />

<strong>and</strong> a Hartley radio frequency oscillator producing 10 watts. The<br />

hydrogen molecular ions were produced by a thermionic cathode


(mounted near the center of the apparatus) from hydrogen gas admitted<br />

through an aperture in the side of the cylinder after a vacuum<br />

had been produced by a pump. Once formed, the oscillating<br />

electrical field drew out the ions <strong>and</strong> accelerated them as they<br />

passed through the cylinder. The accelerated ions spiraled out in a<br />

magnetic field produced by a 10-centimeter electromagnet to a collector.<br />

By November, 1930, Livingston had observed peaks in the<br />

collector current as he tuned the magnetic field through the value<br />

calculated to produce acceleration.<br />

Borrowing a stronger magnet <strong>and</strong> tuning his radio frequency oscillator<br />

appropriately, Livingston produced 80,000-electronvolt ions<br />

at his collector on January 2, 1931, thus demonstrating the principle<br />

of magnetic resonance acceleration.<br />

Impact<br />

Cyclotron / 255<br />

Demonstration of the principle led to the construction of a succession<br />

of large cyclotrons, beginning with a 25-centimeter cyclotron<br />

developed in the spring <strong>and</strong> summer of 1931 that produced<br />

one-million-electronvolt protons. With the support of the Research<br />

Corporation, Lawrence secured a large electromagnet that had been<br />

developed for radio transmission <strong>and</strong> an unused laboratory to<br />

house it: the Radiation Laboratory.<br />

The 69-centimeter cyclotron built with the magnet was used to<br />

explore nuclear physics. It accelerated deuterons, ions of heavy<br />

water or deuterium that contain, in addition to the proton, the neutron,<br />

which was discovered by Sir James Chadwick in 1932. The accelerated<br />

deuteron, which injected neutrons into target atoms, was<br />

used to produce a wide variety of artificial radioisotopes. Many of<br />

these, such as technetium <strong>and</strong> carbon 14, were discovered with the<br />

cyclotron <strong>and</strong> were later used in medicine.<br />

By 1939, Lawrence had built a 152-centimeter cyclotron for medical<br />

applications, including therapy with neutron beams. In that<br />

year, he won the Nobel Prize in Physics for the invention of the cyclotron<br />

<strong>and</strong> the production of radioisotopes. During World War II,<br />

Lawrence <strong>and</strong> the members of his Radiation Laboratory developed<br />

electromagnetic separation of uranium ions to produce the uranium<br />

235 required for the atomic bomb. After the war, the 467-centimeter


256 / Cyclotron<br />

cyclotron was completed as a synchrocyclotron, which modulated<br />

the frequency of the accelerating fields to compensate for the increasing<br />

mass of ions as they approached the speed of light. The<br />

principle of synchronous acceleration, invented by Lawrence’s associate,<br />

the American physicist Edwin Mattison McMillan, became<br />

fundamental to proton <strong>and</strong> electron synchrotrons.<br />

The cyclotron <strong>and</strong> the Radiation Laboratory were the center of<br />

accelerator physics throughout the 1930’s <strong>and</strong> well into the postwar<br />

era. The invention of the cyclotron not only provided a new tool for<br />

probing the nucleus but also gave rise to new forms of organizing<br />

scientific work <strong>and</strong> to applications in nuclear medicine <strong>and</strong> nuclear<br />

chemistry. Cyclotrons were built in many laboratories in the United<br />

States, Europe, <strong>and</strong> Japan, <strong>and</strong> they became a st<strong>and</strong>ard tool of nuclear<br />

physics.<br />

See also Atomic bomb; Electron microscope; Field ion microscope;<br />

Geiger counter; Hydrogen bomb; Mass spectrograph; Neutrino<br />

detector; Scanning tunneling microscope; Synchrocyclotron;<br />

Tevatron accelerator.<br />

Further Reading<br />

Childs, Herbert. An American Genius: The Life of Ernest Orl<strong>and</strong>o Lawrence.<br />

New York: Dutton, 1968.<br />

Close, F. E., Michael Marten, <strong>and</strong> Christine Sutton. The Particle Explosion.<br />

New York: Oxford University Press, 1994.<br />

Pais, Abraham. Inward Bound: Of Matter <strong>and</strong> Forces in the Physical<br />

World. New York: Clarendon Press, 1988.<br />

Wilson, Elizabeth K. “Fifty Years of Heavy Chemistry.” Chemical <strong>and</strong><br />

Engineering News 78, no. 13 (March 27, 2000).


Diesel locomotive<br />

Diesel locomotive<br />

The invention: An internal combustion engine in which ignition is<br />

achieved by the use of high-temperature compressed air, rather<br />

than a spark plug.<br />

The people behind the invention:<br />

Rudolf Diesel (1858-1913), a German engineer <strong>and</strong> inventor<br />

Sir Dugold Clark (1854-1932), a British engineer<br />

Gottlieb Daimler (1834-1900), a German engineer<br />

Henry Ford (1863-1947), an American automobile magnate<br />

Nikolaus Otto (1832-1891), a German engineer <strong>and</strong> Daimler’s<br />

teacher<br />

A Beginning in Winterthur<br />

257<br />

By the beginning of the twentieth century, new means of providing<br />

society with power were needed. The steam engines that were<br />

used to run factories <strong>and</strong> railways were no longer sufficient, since<br />

they were too heavy <strong>and</strong> inefficient. At that time, Rudolf Diesel, a<br />

German mechanical engineer, invented a new engine. His diesel engine<br />

was much more efficient than previous power sources. It also<br />

appeared that it would be able to run on a wide variety of fuels,<br />

ranging from oil to coal dust. Diesel first showed that his engine was<br />

practical by building a diesel-driven locomotive that was tested in<br />

1912.<br />

In the 1912 test runs, the first diesel-powered locomotive was operated<br />

on the track of the Winterthur-Romanston rail line in Switzerl<strong>and</strong>.<br />

The locomotive was built by a German company, Gesellschaft<br />

für Thermo-Lokomotiven, which was owned by Diesel <strong>and</strong><br />

his colleagues. Immediately after the test runs at Winterthur proved<br />

its efficiency, the locomotive—which had been designed to pull express<br />

trains on Germany’s Berlin-Magdeburg rail line—was moved<br />

to Berlin <strong>and</strong> put into service. It worked so well that many additional<br />

diesel locomotives were built. In time, diesel engines were<br />

also widely used to power many other machines, including those<br />

that ran factories, motor vehicles, <strong>and</strong> ships.


258 / Diesel locomotive<br />

Rudolf Diesel<br />

Unbending, suspicious of others, but also exceptionally intelligent,<br />

Rudolf Christian Karl Diesel led a troubled life <strong>and</strong><br />

came to a mysterious end. His parents, expatriate Germans,<br />

lived in Paris when he was born, 1858, <strong>and</strong> he spent his early<br />

childhood there. In 1870, just as he was starting his formal education,<br />

his family fled to Engl<strong>and</strong> on the outbreak of the Franco-<br />

Prussian War, which turned the French against Germans. In Engl<strong>and</strong>,<br />

Diesel spent much of his spare time in museums, educating<br />

himself. His father, a leather craftsman, was unable to support<br />

his family, so as a teenager Diesel was packed off to<br />

Augsburg, Germany, where he was largely on his own. Although<br />

these experiences made him fluent in English, French,<br />

<strong>and</strong> German, his was not a stable or happy childhood.<br />

He threw himself into his studies, finishing his high school<br />

education three years ahead of schedule, <strong>and</strong> entered the Technical<br />

College of Munich, where he was the star student. Once,<br />

during his school years, he saw a demonstration of a Chinese<br />

firestick. The firestick was a tube with a plunger. When a small<br />

piece of flammable material was put in one end <strong>and</strong> the plunger<br />

pushed down rapidly toward it, the heat of the compressed air<br />

in the tube ignited the material. The demonstration later inspired<br />

Diesel to adapt the principle to an engine.<br />

His was the first engine to run successfully with compressed<br />

air fuel ignition, but it was not the first design. So although he<br />

received the patent for the diesel engine, he had to fight challenges<br />

in court from other inventors over licensing rights. He always<br />

won, but the strain of litigation worsened his tendency to<br />

stubborn self-reliance, <strong>and</strong> this led him into difficulties. The<br />

first compression engines were unreliable <strong>and</strong> unwieldy, but<br />

Diesel rebuffed all suggestions for modifications, requiring that<br />

builders follow his original design. His attitude led to delays in<br />

development of the engine <strong>and</strong> lost him financial support.<br />

In 1913, while crossing the English Channel aboard a ship,<br />

Diesel disappeared. His body was never found, <strong>and</strong> although<br />

the authorities concluded that Diesel committed suicide, no one<br />

knows what happened.


Diesels, Diesels Everywhere<br />

Diesel locomotive / 259<br />

In the 1890’s, the best engines available were steam engines that<br />

were able to convert only 5 to 10 percent of input heat energy to useful<br />

work. The burgeoning industrial society <strong>and</strong> a widespread network<br />

of railroads needed better, more efficient engines to help businesses<br />

make profits <strong>and</strong> to speed up the rate of transportation<br />

available for moving both goods <strong>and</strong> people, since the maximum<br />

speed was only about 48 kilometers per hour. In 1894, Rudolf Diesel,<br />

then thirty-five years old, appeared in Augsburg, Germany, with a<br />

new engine that he believed would demonstrate great efficiency.<br />

The diesel engine demonstrated at Augsburg ran for only a<br />

short time. It was, however, more efficient than other existing engines.<br />

In addition, Diesel predicted that his engines would move<br />

trains faster than could be done by existing engines <strong>and</strong> that they<br />

would run on a wide variety of fuels. Experimentation proved the<br />

truth of his claims; even the first working motive diesel engine (the<br />

one used in the Winterthur test) was capable of pulling heavy<br />

freight <strong>and</strong> passenger trains at maximum speeds of up to 160 kilometers<br />

per hour.<br />

By 1912, Diesel, a millionaire, saw the wide use of diesel locomotives<br />

in Europe <strong>and</strong> the United States <strong>and</strong> the conversion of hundreds<br />

of ships to diesel power. Rudolf Diesel’s role in the story ends<br />

here, a result of his mysterious death in 1913—believed to be a suicide<br />

by the authorities—while crossing the English Channel on the<br />

steamer Dresden. Others involved in the continuing saga of diesel<br />

engines were the Britisher Sir Dugold Clerk, who improved diesel<br />

design, <strong>and</strong> the American Adolphus Busch (of beer-brewing fame),<br />

who bought the North American rights to the diesel engine.<br />

The diesel engine is related to automobile engines invented by<br />

Nikolaus Otto <strong>and</strong> Gottlieb Daimler. The st<strong>and</strong>ard Otto-Daimler (or<br />

Otto) engine was first widely commercialized by American auto<br />

magnate Henry Ford. The diesel <strong>and</strong> Otto engines are internalcombustion<br />

engines. This means that they do work when a fuel is<br />

burned <strong>and</strong> causes a piston to move in a tight-fitting cylinder. In diesel<br />

engines, unlike Otto engines, the fuel is not ignited by a spark<br />

from a spark plug. Instead, ignition is accomplished by the use of<br />

high-temperature compressed air.


260 / Diesel locomotive<br />

In common “two-stroke” diesel engines, pioneered by Sir Dugold<br />

Clerk, a starter causes the engine to make its first stroke. This<br />

draws in air <strong>and</strong> compresses the air sufficiently to raise its temperature<br />

to 900 to 1,000 degrees Fahrenheit. At this point, fuel (usually<br />

oil) is sprayed into the cylinder, ignites, <strong>and</strong> causes the piston to<br />

make its second, power-producing stroke. At the end of that stroke,<br />

more air enters as waste gases leave the cylinder; air compression<br />

occurs again; <strong>and</strong> the power-producing stroke repeats itself. This<br />

process then occurs continuously, without restarting.<br />

Impact<br />

Intake Compression Power Exhaust<br />

The four strokes of a diesel engine. (Robert Bosch Corporation)<br />

Proof of the functionality of the first diesel locomotive set the<br />

stage for the use of diesel engines to power many machines. Although<br />

Rudolf Diesel did not live to see it, diesel engines were<br />

widely used within fifteen years after his death. At first, their main<br />

applications were in locomotives <strong>and</strong> ships. Then, because diesel<br />

engines are more efficient <strong>and</strong> more powerful than Otto engines,<br />

they were modified for use in cars, trucks, <strong>and</strong> buses.<br />

At present, motor vehicle diesel engines are most often used in<br />

buses <strong>and</strong> long-haul trucks. In contrast, diesel engines are not as<br />

popular in automobiles as Otto engines, although European auto-


makers make much wider use of diesel engines than American<br />

automakers do. Many enthusiasts, however, view diesel automobiles<br />

as the wave of the future. This optimism is based on the durability<br />

of the engine, its great power, <strong>and</strong> the wide range <strong>and</strong> economical<br />

nature of the fuels that can be used to run it. The drawbacks<br />

of diesels include the unpleasant odor <strong>and</strong> high pollutant content of<br />

their emissions.<br />

Modern diesel engines are widely used in farm <strong>and</strong> earth-moving<br />

equipment, including balers, threshers, harvesters, bulldozers,rock<br />

crushers, <strong>and</strong> road graders. Construction of the Alaskan oil pipeline<br />

relied heavily on equipment driven by diesel engines. Diesel engines<br />

are also commonly used in sawmills, breweries, coal mines,<br />

<strong>and</strong> electric power plants.<br />

Diesel’s brainchild has become a widely used power source, just<br />

as he predicted. It is likely that the use of diesel engines will continue<br />

<strong>and</strong> will exp<strong>and</strong>, as the dem<strong>and</strong>s of energy conservation require<br />

more efficient engines <strong>and</strong> as moves toward fuel diversification<br />

require engines that can be used with various fuels.<br />

See also Bullet train; Gas-electric car; Internal combustion engine.<br />

Further Reading<br />

Diesel locomotive / 261<br />

Cummins, C. Lyle. Diesel’s Engine. Wilsonville, Oreg.: Carnot Press,<br />

1993.<br />

Diesel, Eugen. From Engines to Autos: Five Pioneers in Engine Development<br />

<strong>and</strong> Their Contributions to the Automotive Industry. Chicago:<br />

H. Regnery, 1960.<br />

Nitske, Robert W., <strong>and</strong> Charles Morrow Wilson. Rudolf Diesel: Pioneer<br />

of the Age of Power. Norman: University of Oklahoma Press,<br />

1965.


262<br />

Differential analyzer<br />

Differential analyzer<br />

The invention: An electromechanical device capable of solving differential<br />

equations.<br />

The people behind the invention:<br />

Vannevar Bush (1890-1974), an American electrical engineer<br />

Harold L. Hazen (1901-1980), an American electrical engineer<br />

Electrical Engineering Problems Become More Complex<br />

After World War I, electrical engineers encountered increasingly<br />

difficult differential equations as they worked on vacuum-tube circuitry,<br />

telephone lines, <strong>and</strong>, particularly, long-distance power transmission<br />

lines. These calculations were lengthy <strong>and</strong> tedious. Two of<br />

the many steps required to solve them were to draw a graph manually<br />

<strong>and</strong> then to determine the area under the curve (essentially, accomplishing<br />

the mathematical procedure called integration).<br />

In 1925, Vannevar Bush, a faculty member in the Electrical Engineering<br />

Department at the Massachusetts Institute of Technology<br />

(MIT), suggested that one of his graduate students devise a machine<br />

to determine the area under the curve. They first considered a mechanical<br />

device but later decided to seek an electrical solution. Realizing<br />

that a watt-hour meter such as that used to measure electricity<br />

in most homes was very similar to the device they needed, Bush <strong>and</strong><br />

his student refined the meter <strong>and</strong> linked it to a pen that automatically<br />

recorded the curve.<br />

They called this machine the Product Integraph, <strong>and</strong> MIT students<br />

began using it immediately. In 1927, Harold L. Hazen, another<br />

MIT faculty member, modified it in order to solve the more complex<br />

second-order differential equations (it originally solved only firstorder<br />

equations).<br />

The Differential Analyzer<br />

The original Product Integraph had solved problems electrically,<br />

<strong>and</strong> Hazen’s modification had added a mechanical integrator. Al-


Differential analyzer / 263<br />

though the revised Product Integraph was useful in solving the<br />

types of problems mentioned above, Bush thought the machine<br />

could be improved by making it an entirely mechanical integrator,<br />

rather than a hybrid electrical <strong>and</strong> mechanical device.<br />

In late 1928, Bush received funding from MIT to develop an entirely<br />

mechanical integrator, <strong>and</strong> he completed the resulting Differential<br />

Analyzer in 1930. This machine consisted of numerous interconnected<br />

shafts on a long, tablelike framework, with drawing<br />

boards flanking one side <strong>and</strong> six wheel-<strong>and</strong>-disk integrators on the<br />

other. Some of the drawing boards were configured to allow an operator<br />

to trace a curve with a pen that was linked to the Analyzer,<br />

thus providing input to the machine. The other drawing boards<br />

were configured to receive output from the Analyzer via a pen that<br />

drew a curve on paper fastened to the drawing board.<br />

The wheel-<strong>and</strong>-disk integrator, which Hazen had first used in<br />

the revised Product Integraph, was the key to the operation of the<br />

Differential Analyzer. The rotational speed of the horizontal disk<br />

was the input to the integrator, <strong>and</strong> it represented one of the variables<br />

in the equation. The smaller wheel rolled on the top surface of<br />

the disk, <strong>and</strong> its speed, which was different from that of the disk,<br />

represented the integrator’s output. The distance from the wheel to<br />

the center of the disk could be changed to accommodate the equation<br />

being solved, <strong>and</strong> the resulting geometry caused the two shafts<br />

to turn so that the output was the integral of the input. The integrators<br />

were linked mechanically to other devices that could add, subtract,<br />

multiply, <strong>and</strong> divide. Thus, the Differential Analyzer could<br />

solve complex equations involving many different mathematical<br />

operations. Because all the linkages <strong>and</strong> calculating devices were<br />

mechanical, the Differential Analyzer actually acted out each calculation.<br />

Computers of this type, which create an analogy to the physical<br />

world, are called analog computers.<br />

The Differential Analyzer fulfilled Bush’s expectations, <strong>and</strong> students<br />

<strong>and</strong> researchers found it very useful. Although each different<br />

problem required Bush’s team to set up a new series of mechanical<br />

linkages, the researchers using the calculations viewed this as a minor<br />

inconvenience. Students at MIT used the Differential Analyzer<br />

in research for doctoral dissertations, master’s theses, <strong>and</strong> bachelor’s<br />

theses. Other researchers worked on a wide range of problems


264 / Differential analyzer<br />

Vannevar Bush<br />

One of the most politically powerful scientists of the twentieth<br />

century, Vannevar Bush was born in 1890 in Everett, Massachusetts.<br />

He studied at Tufts College in Boston, not only earning<br />

two degrees in engineering but also registering his first<br />

patent while still an undergraduate. He worked for General<br />

Electric Company briefly after college <strong>and</strong> then conducted research<br />

on submarine detection for the U.S. Navy during World<br />

War I.<br />

After the war he became a professor of electrical power<br />

transmission (<strong>and</strong> later dean of the engineering school) at the<br />

Massachusetts Institute of Technology (MIT). He also acted as a<br />

consultant for industry <strong>and</strong> started companies of his own, including<br />

(with two others) Raytheon Corporation. While at MIT<br />

he developed the Product Integraph <strong>and</strong> Differential Analyzer<br />

to aid in solving problems related to electrical power transmission.<br />

Starting in 1939, Bush became a key science administrator.<br />

He was president of the Carnegie Foundation from 1939 until<br />

1955, chaired the National Advisory Committee for Aeronautics<br />

from 1939 until 1941, in 1940 was appointed chairman of the<br />

President’s National Defense Research Committee, <strong>and</strong> from<br />

1941 until 1946 was director of the Office of Scientific Research<br />

<strong>and</strong> Development. This meant he was President Franklin Roosevelt’s<br />

science adviser during World War II <strong>and</strong> oversaw wartime<br />

military research, including involvement in the Manhattan<br />

Project that build the first atomic bombs. After the war he<br />

worked for peaceful application of atomic power <strong>and</strong> was instrumental<br />

in inaugurating the National Science Foundation,<br />

which he directed, in 1950. Between 1957 <strong>and</strong> 1959 he served as<br />

chairman of MIT Corporation, retaining an honorary chairmanship<br />

thereafter.<br />

All these political <strong>and</strong> administrative roles meant he exercised<br />

enormous influence in deciding which scientific projects<br />

were supported financially. Having received many honorary<br />

degrees <strong>and</strong> awards, including the National Medal of Science<br />

(1964), Bush died in 1974.


with the Differential Analyzer, mostly in electrical engineering, but<br />

also in atomic physics, astrophysics, <strong>and</strong> seismology. An English researcher,<br />

Douglas Hartree, visited Bush’s laboratory in 1933 to learn<br />

about the Differential Analyzer <strong>and</strong> to use it in his own work on the<br />

atomic field of mercury. When he returned to Engl<strong>and</strong>, he built several<br />

analyzers based on his knowledge of MIT’s machine. The U.S.<br />

Army also built a copy in order to carry out the complex calculations<br />

required to create artillery firing tables (which specified the<br />

proper barrel angle to achieve the desired range). Other analyzers<br />

were built by industry <strong>and</strong> universities around the world.<br />

Impact<br />

Differential analyzer / 265<br />

As successful as the Differential Analyzer had been, Bush wanted<br />

to make another, better analyzer that would be more precise, more<br />

convenient to use, <strong>and</strong> more mathematically flexible. In 1932, Bush<br />

began seeking money for his new machine, but because of the Depression<br />

it was not until 1936 that he received adequate funding for<br />

the Rockefeller Analyzer, as it came to be known. Bush left MIT in<br />

1938, but work on the Rockefeller Analyzer continued. It was first<br />

demonstrated in 1941, <strong>and</strong> by 1942, it was being used in the war effort<br />

to calculate firing tables <strong>and</strong> design radar antenna profiles. At<br />

the end of the war, it was the most important computer in existence.<br />

All the analyzers, which were mechanical computers, faced serious<br />

limitations in speed because of the momentum of the machinery,<br />

<strong>and</strong> in precision because of slippage <strong>and</strong> wear. The digital computers<br />

that were being developed after World War II (even at MIT)<br />

were faster, more precise, <strong>and</strong> capable of executing more powerful<br />

operations because they were electrical computers. As a result, during<br />

the 1950’s, they eclipsed differential analyzers such as those<br />

built by Bush. Descendants of the Differential Analyzer remained in<br />

use as late as the 1990’s, but they played only a minor role.<br />

See also Colossus computer; ENIAC computer; Mark I calculator;<br />

Personal computer; SAINT; UNIVAC computer.


266 / Differential analyzer<br />

Further Reading<br />

Bush, Vannevar. Pieces of the Action. New York: Morrow, 1970.<br />

Marcus, Alan I., <strong>and</strong> Howard P. Segal. Technology in America. Fort<br />

Worth, Tex.: Harcourt Brace College, 1999.<br />

Spencer, Donald D. Great Men <strong>and</strong> Women of Computing. Ormond<br />

Beach, Fla.: Camelot Publishing, 1999.<br />

Zachary, G. Pascal. Endless Frontier: Vannevar Bush, Engineer of the<br />

American Century. Cambridge, Mass.: MIT Press, 1999.


Dirigible<br />

Dirigible<br />

The invention: A rigid lighter-than-air aircraft that played a major<br />

role in World War I <strong>and</strong> in international air traffic until a disastrous<br />

accident destroyed the industry.<br />

The people behind the invention:<br />

Ferdin<strong>and</strong> von Zeppelin (1838-1917), a retired German general<br />

Theodor Kober (1865-1930), Zeppelin’s private engineer<br />

Early Competition<br />

267<br />

When the Montgolfier brothers launched the first hot-air balloon<br />

in 1783, engineers—especially those in France—began working on<br />

ways to use machines to control the speed <strong>and</strong> direction of balloons.<br />

They thought of everything: rowing through the air with silk-covered<br />

oars; building movable wings; using a rotating fan, an airscrew, or a<br />

propeller powered by a steam engine (1852) or an electric motor<br />

(1882). At the end of the nineteenth century, the internal combustion<br />

engine was invented. It promised higher speeds <strong>and</strong> more power.<br />

Up to this point, however, the balloons were not rigid.<br />

A rigid airship could be much larger than a balloon <strong>and</strong> could fly<br />

farther. In 1890, a rigid airship designed by David Schwarz of<br />

Dalmatia was tested in St. Petersburg, Russia. The test failed because<br />

there were problems with inflating the dirigible. A second<br />

test, in Berlin in 1897, was only slightly more successful, since the<br />

hull leaked <strong>and</strong> the flight ended in a crash.<br />

Schwarz’s airship was made of an entirely rigid aluminum cylinder.<br />

Ferdin<strong>and</strong> von Zeppelin had a different idea: His design was<br />

based on a rigid frame. Zeppelin knew about balloons from having<br />

fought in two wars in which they were used: the American Civil<br />

War of 1861-1865 <strong>and</strong> the Franco-Prussian War of 1870-1871. He<br />

wrote down his first “thoughts about an airship” in his diary on<br />

March 25, 1874, inspired by an article about flying <strong>and</strong> international<br />

mail. Zeppelin soon lost interest in this idea of civilian uses for an<br />

airship <strong>and</strong> concentrated instead on the idea that dirigible balloons<br />

might become an important part of modern warfare. He asked the


268 / Dirigible<br />

German government to fund his research, pointing out that France<br />

had a better military air force than Germany did. Zeppelin’s patriotism<br />

was what kept him trying, in spite of money problems <strong>and</strong><br />

technical difficulties.<br />

In 1893, in order to get more money, Zeppelin tried to persuade<br />

the German military <strong>and</strong> engineering experts that his invention was<br />

practical. Even though a government committee decided that his<br />

work was worth a small amount of funding, the army was not sure<br />

that Zeppelin’s dirigible was worth the cost. Finally, the committee<br />

chose Schwarz’s design. In 1896, however, Zeppelin won the support<br />

of the powerful Union of German Engineers, which in May,<br />

1898, gave him 800,000 marks to form a stock company called the<br />

Association for the Promotion of Airship Flights. In 1899, Zeppelin<br />

began building his dirigible in Manzell at Lake Constance. In July,<br />

1900, the airship was finished <strong>and</strong> ready for its first test flight.<br />

Several Attempts<br />

Zeppelin, together with his engineer, Theodor Kober, had worked<br />

on the design since May, 1892, shortly after Zeppelin’s retirement<br />

from the army. They had finished the rough draft by 1894, <strong>and</strong><br />

though they made some changes later, this was the basic design of<br />

the Zeppelin. An improved version was patented in December,<br />

1897.<br />

In the final prototype, called the LZ 1, the engineers tried to make<br />

the airship as light as possible. They used a light internal combustion<br />

engine <strong>and</strong> designed a frame made of the light metal aluminum.<br />

The airship was 128 meters long <strong>and</strong> had a diameter of 11.7<br />

meters when inflated. Twenty-four zinc-aluminum girders ran the<br />

length of the ship, being drawn together at each end. Sixteen rings<br />

held the body together. The engineers stretched an envelope of<br />

smooth cotton over the framework to reduce wind resistance <strong>and</strong> to<br />

protect the gas bags from the sun’s rays. Seventeen gas bags made of<br />

rubberized cloth were placed inside the framework. Together they<br />

held more than 120,000 cubic meters of hydrogen gas, which would<br />

lift 11,090 kilograms. Two motor gondolas were attached to the<br />

sides, each with a 16-horsepower gasoline engine, spinning four<br />

propellers.


Count Ferdin<strong>and</strong> von Zeppelin<br />

Dirigible / 269<br />

The Zeppelin, the first lighter-than-air craft that was powered<br />

<strong>and</strong> steerable, began as a retirement project.<br />

Count Ferdin<strong>and</strong> von Zeppelin was born near Lake Constance<br />

in southern Germany in 1838 <strong>and</strong> grew up in a family<br />

long used to aristocratic privilege <strong>and</strong> government service. After<br />

studying engineering at the University of Tübingen, he was<br />

commissioned as a lieutenant of engineers. In 1863 he traveled<br />

to the United States <strong>and</strong>, armed with a letter of introduction<br />

from President Abraham Lincoln, toured the Union emplacements.<br />

The observation balloons then used to see behind enemy<br />

lines impressed him. He learned all he could about them <strong>and</strong><br />

even flew up in one to seven hundred feet.<br />

His enthusiasm for airships stayed with him throughout his<br />

career, but he was not really able to apply himself to the problem<br />

until he retired (as a brigadier general) in 1890. Then he<br />

concentrated on the struggle to line up financing <strong>and</strong> attract talented<br />

help. He found investors for 90 percent of the money he<br />

needed <strong>and</strong> got the rest from his wife’s inheritance. The first<br />

LZ’s (Luftschiff Zeppelin) had troubles, but setbacks did not stop<br />

him. He was a stubborn, determined man. By the time he died<br />

in 1917 near Berlin he had seen ninety-two airships built. And<br />

because his design was so thoroughly associated with lighterthan-air<br />

vessels in the mind of the German public, they have<br />

ever after been known as zeppelins. However, he had already<br />

recognized their vulnerability as military aircraft, his main interest,<br />

<strong>and</strong> so he had turned his attention to designs for large<br />

airplanes as bombers.<br />

The test flight did not go well. The two main questions—whether<br />

the craft was strong enough <strong>and</strong> fast enough—could not be answered<br />

because little things kept going wrong; for example, a crankshaft<br />

broke <strong>and</strong> a rudder jammed. The first flight lasted no more<br />

than eighteen minutes, with a maximum speed of 13.7 kilometers<br />

per hour. During all three test flights, the airship was in the air for a<br />

total of only two hours, going no faster than 28.2 kilometers per<br />

hour.<br />

Zeppelin had to drop the project for some years because he ran<br />

out of money, <strong>and</strong> his company was dissolved. The LZ 1 was


270 / Dirigible<br />

wrecked in the spring of 1901. A second airship was tested in November,<br />

1905, <strong>and</strong> January, 1906. Both tests were unsuccessful, <strong>and</strong><br />

in the end the ship was destroyed during a storm.<br />

By 1906, however, the German government was convinced of the<br />

military usefulness of the airship, though it would not give money<br />

to Zeppelin unless he agreed to design one that could stay in the air<br />

for at least twenty-four hours. The third Zeppelin failed this test in<br />

the autumn of 1907. Finally, in the summer of 1908, the LZ 4 not only<br />

proved itself to the military but also attracted great publicity. It flew<br />

for more than twenty-four hours <strong>and</strong> reached a speed of more than<br />

60 kilometers per hour. Caught in a storm at the end of this flight,<br />

the airship was forced to l<strong>and</strong> <strong>and</strong> exploded, but money came from<br />

all over Germany to build another.<br />

Impact<br />

Most rigid airships were designed <strong>and</strong> flown in Germany. Of the<br />

161 that were built between 1900 <strong>and</strong> 1938, 139 were made in Germany,<br />

<strong>and</strong> 119 were based on the Zeppelin design.<br />

More than 80 percent of the airships were built for the military.<br />

The Germans used more than one hundred for gathering information<br />

<strong>and</strong> for bombing during World War I (1914-1918). Starting in<br />

May, 1915, airships bombed Warsaw, Pol<strong>and</strong>; Bucharest, Romania;<br />

Salonika, Greece; <strong>and</strong> London, Engl<strong>and</strong>. This was mostly a fear tactic,<br />

since the attacks did not cause great damage, <strong>and</strong> the English antiaircraft<br />

defense improved quickly. By 1916, the German army had<br />

lost so many airships that it stopped using them, though the navy<br />

continued.<br />

Airships were first used for passenger flights in 1910. By 1914,<br />

the Delag (German Aeronautic Stock Company) used seven passenger<br />

airships for sightseeing trips around German cities. There were<br />

still problems with engine power <strong>and</strong> weather forecasting, <strong>and</strong> it<br />

was difficult to move the airships on the ground. After World War I,<br />

the Zeppelins that were left were given to the Allies as payment,<br />

<strong>and</strong> the Germans were not allowed to build airships for their own<br />

use until 1925.<br />

In the 1920’s <strong>and</strong> 1930’s, it became cheaper to use airplanes for


short flights, so airships were useful mostly for long-distance flight.<br />

A British airship made the first transatlantic flight in 1919. The British<br />

hoped to connect their empire by means of airships starting in<br />

1924, but the 1930 crash of the R-101, in which most of the leading<br />

English aeronauts were killed, brought that hope to an end.<br />

The United States Navy built the Akron (1931) <strong>and</strong> the Macon<br />

(1933) for long-range naval reconnaissance, but both airships crashed.<br />

Only the Germans continued to use airships on a regular basis. In<br />

1929, the world tour of the Graf Zeppelin was a success. Regular<br />

flights between Germany <strong>and</strong> South America started in 1932, <strong>and</strong> in<br />

1936, German airships bearing Nazi swastikas flew to Lakehurst,<br />

New Jersey. The tragic explosion of the hydrogen-filled Hindenburg<br />

in 1937, however, brought the era of the rigid airship to a close. The<br />

U.S. secretary of the interior vetoed the sale of nonflammable helium,<br />

fearing that the Nazis would use it for military purposes, <strong>and</strong><br />

the German government had to stop transatlantic flights for safety<br />

reasons. In 1940, the last two remaining rigid airships were destroyed.<br />

See also Airplane; Gyrocompass; Stealth aircraft; Supersonic<br />

passenger plane; Turbojet.<br />

Further Reading<br />

Dirigible / 271<br />

Brooks, Peter. Zeppelin: Rigid Airships, 1893-1940. London: Putman,<br />

1992.<br />

Chant, Christopher. The Zeppelin: The History of German Airships from<br />

1900-1937. New York: Barnes <strong>and</strong> Noble Books, 2000.<br />

Griehl, Manfred, <strong>and</strong> Joachim Dressel. Zeppelin! The German Airship<br />

Story. New York: Sterling Publishing, 1990.<br />

Syon, Guillaume de. Zeppelin!: Germany <strong>and</strong> the Airship, 1900-1939.<br />

Baltimore: John Hopkins University Press, 2001.


272<br />

Disposable razor<br />

Disposable razor<br />

The invention: An inexpensive shaving blade that replaced the traditional<br />

straight-edged razor <strong>and</strong> transformed shaving razors<br />

into a frequent household purchase item.<br />

The people behind the invention:<br />

King Camp Gillette (1855-1932), inventor of the disposable razor<br />

Steven Porter, the machinist who created the first three<br />

disposable razors for King Camp Gillette<br />

William Emery Nickerson (1853-1930), an expert machine<br />

inventor who created the machines necessary for mass<br />

production<br />

Jacob Heilborn, an industrial promoter who helped Gillette start<br />

his company <strong>and</strong> became a partner<br />

Edward J. Stewart, a friend <strong>and</strong> financial backer of Gillette<br />

Henry Sachs, an investor in the Gillette Safety Razor Company<br />

John Joyce, an investor in the Gillette Safety Razor Company<br />

William Painter (1838-1906), an inventor who inspired Gillette<br />

George Gillette, an inventor, King Camp Gillette’s father<br />

A Neater Way to Shave<br />

In 1895, King Camp Gillette thought of the idea of a disposable razor<br />

blade. Gillette spent years drawing different models, <strong>and</strong> finally<br />

Steven Porter, a machinist <strong>and</strong> Gillette’s associate, created from those<br />

drawings the first three disposable razors that worked. Gillette soon<br />

founded the Gillette Safety Razor Company, which became the leading<br />

seller of disposable razor blades in the United States.<br />

George Gillette, King Camp Gillette’s father, had been a newspaper<br />

editor, a patent agent, <strong>and</strong> an inventor. He never invented a very<br />

successful product, but he loved to experiment. He encouraged all<br />

of his sons to figure out how things work <strong>and</strong> how to improve on<br />

them. King was always inventing something new <strong>and</strong> had many<br />

patents, but he was unsuccessful in turning them into profitable<br />

businesses.<br />

Gillette worked as a traveling salesperson for Crown Cork <strong>and</strong>


Seal Company. William Painter, one of Gillette’s friends <strong>and</strong> the inventor<br />

of the crown cork, presented Gillette with a formula for making<br />

a fortune: Invent something that would constantly need to be replaced.<br />

Painter’s crown cork was used to cap beer <strong>and</strong> soda bottles.<br />

It was a tin cap covered with cork, used to form a tight seal over a<br />

bottle. Soda <strong>and</strong> beer companies could use a crown cork only once<br />

<strong>and</strong> needed a steady supply.<br />

King took Painter’s advice <strong>and</strong> began thinking of everyday items<br />

that needed to be replaced often. After owning a Star safety razor<br />

for some time, King realized that the razor blade had not been improved<br />

for a long time. He studied all the razors on the market <strong>and</strong><br />

found that both the common straight razor <strong>and</strong> the safety razor featured<br />

a heavy V-shaped piece of steel, sharpened on one side. King<br />

reasoned that a thin piece of steel sharpened on both sides would<br />

create a better shave <strong>and</strong> could be thrown away once it became dull.<br />

The idea of the disposable razor had been born.<br />

Gillette made several drawings of disposable razors. He then<br />

made a wooden model of the razor to better explain his idea.<br />

Gillette’s first attempt to construct a working model was unsuccessful,<br />

as the steel was too flimsy. Steven Porter, a Boston machinist, decided<br />

to try to make Gillette’s razor from his drawings. He produced<br />

three razors, <strong>and</strong> in the summer of 1899 King was the first<br />

man to shave with a disposable razor.<br />

Changing Consumer Opinion<br />

Disposable razor / 273<br />

In the early 1900’s, most people considered a razor to be a oncein-a-lifetime<br />

purchase. Many fathers h<strong>and</strong>ed down their razors to<br />

their sons. Straight razors needed constant <strong>and</strong> careful attention to<br />

keep them sharp. The thought of throwing a razor in the garbage after<br />

several uses was contrary to the general public’s idea of a razor.<br />

If Gillette’s razor had not provided a much less painful <strong>and</strong> faster<br />

shave, it is unlikely that the disposable would have been a success.<br />

Even with its advantages, public opinion against the product was<br />

still difficult to overcome.<br />

Financing a company to produce the razor proved to be a major<br />

obstacle. King did not have the money himself, <strong>and</strong> potential investors<br />

were skeptical. Skepticism arose both because of public percep-


274 / Disposable razor<br />

tions of the product <strong>and</strong> because of its manufacturing process. Mass<br />

production appeared to be impossible, but the disposable razor<br />

would never be profitable if produced using the methods used to<br />

manufacture its predecessor.<br />

William Emery Nickerson, an expert machine inventor, had looked<br />

at Gillette’s razor <strong>and</strong> said it was impossible to create a machine to<br />

produce it. He was convinced to reexamine the idea <strong>and</strong> finally created<br />

a machine that would create a workable blade. In the process,<br />

Nickerson changed Gillette’s original model. He improved the h<strong>and</strong>le<br />

<strong>and</strong> frame so that it would better support the thin steel blade.<br />

In the meantime, Gillette was busy getting his patent assigned to<br />

the newly formed American Safety Razor Company, owned by<br />

Gillette, Jacob Heilborn, Edward J. Stewart, <strong>and</strong> Nickerson. Gillette<br />

owned considerably more shares than anyone else. Henry Sachs<br />

provided additional capital, buying shares from Gillette.<br />

The stockholders decided to rename the company the Gillette<br />

Safety Razor Company. It soon spent most of its money on machinery<br />

<strong>and</strong> lacked the capital it needed to produce <strong>and</strong> advertise its<br />

product. The only offer the company had received was from a group<br />

of New York investors who were willing to give $125,000 in exchange<br />

for 51 percent of the company. None of the directors wanted<br />

to lose control of the company, so they rejected the offer.<br />

John Joyce, a friend of Gillette, rescued the financially insecure<br />

new company. He agreed to buy $100,000 worth of bonds from the<br />

company for sixty cents on the dollar, purchasing the bonds gradually<br />

as the company needed money. He also received an equivalent<br />

amount of company stock. After an investment of $30,000, Joyce<br />

had the option of backing out. This deal enabled the company to<br />

start manufacturing <strong>and</strong> advertising.<br />

Impact<br />

The company used $18,000 to perfect the machinery to produce<br />

the disposable razor blades <strong>and</strong> razors. Originally the directors<br />

wanted to sell each razor with twenty blades for three dollars. Joyce<br />

insisted on a price of five dollars. In 1903, five dollars was about<br />

one-third of the average American’s weekly salary, <strong>and</strong> a highquality<br />

straight razor could be purchased for about half that price.


Disposable razor / 275<br />

The other directors were skeptical, but Joyce threatened to buy up<br />

all the razors for three dollars <strong>and</strong> sell them himself for five dollars.<br />

Joyce had the financial backing to make this promise good, so the directors<br />

agreed to the higher price.<br />

The Gillette Safety Razor Company contracted with Townsend &<br />

Hunt for exclusive sales. The contract stated that Townsend & Hunt<br />

would buy 50,000 razors with twenty blades each during a period of<br />

slightly more than a year <strong>and</strong> would purchase 100,000 sets per year<br />

for the following four years. The first advertisement for the product<br />

appeared in System Magazine in early fall of 1903, offering the razors<br />

by mail order. By the end of 1903, only fifty-one razors had been<br />

sold.<br />

Since Gillette <strong>and</strong> most of the directors of the company were not<br />

salaried, Gillette had needed to keep his job as salesman with<br />

Crown Cork <strong>and</strong> Seal. At the end of 1903, he received a promotion<br />

that meant relocation from Boston to London. Gillette did not want<br />

to go <strong>and</strong> pleaded with the other directors, but they insisted that the<br />

company could not afford to put him on salary. The company decided<br />

to reduce the number of blades in a set from twenty to twelve<br />

in an effort to increase profits without noticeably raising the cost of a<br />

set. Gillette resigned the title of company president <strong>and</strong> left for Engl<strong>and</strong>.<br />

Shortly thereafter, Townsend & Hunt changed its name to the<br />

Gillette Sales Company, <strong>and</strong> three years later the sales company<br />

sold out to the parent company for $300,000. Sales of the new type<br />

of razor were increasing rapidly in the United States, <strong>and</strong> Joyce<br />

wanted to sell patent rights to European companies for a small percentage<br />

of sales. Gillette thought that that would be a horrible mistake<br />

<strong>and</strong> quickly traveled back to Boston. He had two goals: to stop<br />

the sale of patent rights, based on his conviction that the foreign<br />

market would eventually be very lucrative, <strong>and</strong> to become salaried<br />

by the company. Gillette accomplished both these goals <strong>and</strong> soon<br />

moved back to Boston.<br />

Despite the fact that Joyce <strong>and</strong> Gillette had been good friends for<br />

a long time, their business views often differed. Gillette set up a<br />

holding company in an effort to gain back controlling interest in the<br />

Gillette Safety Razor Company. He borrowed money <strong>and</strong> convinced<br />

his allies in the company to invest in the holding company, eventu-


276 / Disposable razor<br />

ally regaining control. He was reinstated as president of the company.<br />

One clear disagreement was that Gillette wanted to relocate the company<br />

to Newark, New Jersey, <strong>and</strong> Joyce thought that that would be a<br />

waste of money. Gillette authorized company funds to be invested in<br />

a Newark site. The idea was later dropped, costing the company a<br />

large amount of capital. Gillette was not a very wise businessman<br />

King Camp Gillette<br />

At age sixteen, King Camp Gillette (1855-1932) saw all of his<br />

family’s belongings consumed in the Great Chicago Fire. He<br />

had to drop out of school because of it <strong>and</strong> earn his own living.<br />

The catastrophe <strong>and</strong> the sudden loss of security that followed<br />

shaped his ambitions. He was not about to risk destitution ever<br />

again.<br />

He made himself a successful traveling salesman but still<br />

felt he was earning too little. So he turned his mind to inventions,<br />

hoping to get rich quick. The disposable razor was his<br />

only venture, but it was enough. After its long preparation for<br />

marketing Gillette’s invention <strong>and</strong> some subsequent turmoil<br />

among its board of directors, the Gillette Safety Razor Company<br />

was a phenomenal success <strong>and</strong> a bonanza for Gillette. He<br />

became wealthy. He retired in 1913, just ten years after the company<br />

opened, his security assured.<br />

His mother had written cookbooks, one of which was a bestseller.<br />

As an adult, Gillette got the writing bug himself <strong>and</strong><br />

wrote four books, but his theme was far loftier than cooking—<br />

social theory <strong>and</strong> security for the masses. Like Karl Marx he argued<br />

that economic competition squ<strong>and</strong>ers human resources<br />

<strong>and</strong> leads to deprivation, which in turn leads to crime. So, he<br />

reasoned, getting rid of economic competition will end misery<br />

<strong>and</strong> crime. He recommended that a centralized agency plan<br />

production <strong>and</strong> oversee distribution, a recommendation that<br />

America resoundingly ignored. However, other ideas of his<br />

eventually found acceptance, such as air conditioning for workers<br />

<strong>and</strong> government assistance for the unemployed.<br />

In 1922 Gillette moved to Los Angeles, California, <strong>and</strong> devoted<br />

himself to raising oranges <strong>and</strong> collecting his share of the<br />

company profits. However, he seldom felt free enough with his<br />

money to donate it to charity or finance social reform.


Disposable razor / 277<br />

<strong>and</strong> made many costly mistakes. Joyce even accused him of deliberately<br />

trying to keep the stock price low so that Gillette could purchase<br />

more stock. Joyce eventually bought out Gillette, who retained<br />

his title as president but had little say about company<br />

business.<br />

With Gillette out of a management position, the company became<br />

more stable <strong>and</strong> more profitable. The biggest problem the<br />

company faced was that it would soon lose its patent rights. After<br />

the patent expired, the company would have competition. The company<br />

decided that it could either cut prices (<strong>and</strong> therefore profits) to<br />

compete with the lower-priced disposables that would inevitably<br />

enter the market, or it could create a new line of even better razors.<br />

The company opted for the latter strategy. Weeks before the patent<br />

expired, the Gillette Safety Razor Company introduced a new line<br />

of razors.<br />

Both World War I <strong>and</strong> World War II were big boosts to the company,<br />

which contracted with the government to supply razors to almost<br />

all the troops. This transaction created a huge increase in sales<br />

<strong>and</strong> introduced thous<strong>and</strong>s of young men to the Gillette razor. Many<br />

of them continued to use Gillettes after returning from the war.<br />

Aside from the shaky start of the company, its worst financial difficulties<br />

were during the Great Depression. Most Americans simply<br />

could not afford Gillette blades, <strong>and</strong> many used a blade for an extended<br />

time <strong>and</strong> then resharpened it rather than throwing it away. If<br />

it had not been for the company’s foreign markets, the company<br />

would not have shown a profit during the Great Depression.<br />

Gillette’s obstinancy about not selling patent rights to foreign investors<br />

proved to be an excellent decision.<br />

The company advertised through sponsoring sporting events,<br />

including the World Series. Gillette had many celebrity endorsements<br />

from well-known baseball players. Before it became too expensive<br />

for one company to sponsor an entire event, Gillette had<br />

exclusive advertising during the World Series, various boxing<br />

matches, the Kentucky Derby, <strong>and</strong> football bowl games. Sponsoring<br />

these events was costly, but sports spectators were the typical<br />

Gillette customers.<br />

The Gillette Company created many products that complemented<br />

razors <strong>and</strong> blades, including shaving cream, women’s ra-


278 / Disposable razor<br />

zors, <strong>and</strong> electric razors. The company exp<strong>and</strong>ed into new products<br />

including women’s cosmetics, writing utensils, deodorant, <strong>and</strong><br />

wigs. One of the main reasons for obtaining a more diverse product<br />

line was that a one-product company is less stable, especially in a<br />

volatile market. The Gillette Company had learned that lesson in<br />

the Great Depression. Gillette continued to thrive by following the<br />

principles the company had used from the start. The majority of<br />

Gillette’s profits came from foreign markets, <strong>and</strong> its employees<br />

looked to improve products <strong>and</strong> find opportunities in other departments<br />

as well as their own.<br />

See also Contact lenses; Memory metal; Steelmaking process.<br />

Further Reading<br />

Adams, Russell B., Jr. King C. Gillette: The Man <strong>and</strong> His Wonderful<br />

Shaving Device. Boston: Little, Brown, 1978.<br />

Dowling, Tim. Inventor of the Disposable Culture: King Camp Gillette,<br />

1855-1932. London: Short, 2001.<br />

“Gillette: Blade-runner.” The Economist 327 (April 10, 1993).<br />

Killgren, Lucy. “Nicking Gillette.” Marketing Week 22 (June 17, 1999).<br />

McKibben, Gordon. Cutting Edge: Gillette’s Journey to Global Leadership.<br />

Boston, Mass.: Harvard Business School Press, 1998.<br />

Thomas, Robert J. New Product Success Stories: Lessons from Leading<br />

Innovators. New York: John Wiley, 1995.<br />

Zeien, Alfred M. The Gillette Company. New York: Newcomen Society<br />

of the United States, 1999.


Dolby noise reduction<br />

Dolby noise reduction<br />

The invention: Electronic device that reduces the signal-to-noise<br />

ratio of sound recordings <strong>and</strong> greatly improves the sound quality<br />

of recorded music.<br />

The people behind the invention:<br />

Emil Berliner (1851-1929), a German inventor<br />

Ray Milton Dolby (1933- ), an American inventor<br />

Thomas Alva Edison (1847-1931), an American inventor<br />

Phonographs, Tapes, <strong>and</strong> Noise Reduction<br />

279<br />

The main use of record, tape, <strong>and</strong> compact disc players is to listen<br />

to music, although they are also used to listen to recorded speeches,<br />

messages, <strong>and</strong> various forms of instruction. Thomas Alva Edison<br />

invented the first sound-reproducing machine, which he called the<br />

“phonograph,” <strong>and</strong> patented it in 1877. Ten years later, a practical<br />

phonograph (the “gramophone”) was marketed by a German, Emil<br />

Berliner. Phonographs recorded sound by using diaphragms that<br />

vibrated in response to sound waves <strong>and</strong> controlled needles that cut<br />

grooves representing those vibrations into the first phonograph records,<br />

which in Edison’s machine were metal cylinders <strong>and</strong> in Berliner’s<br />

were flat discs. The recordings were then played by reversing<br />

the recording process: Placing a needle in the groove in the recorded<br />

cylinder or disk caused the diaphragm to vibrate, re-creating the<br />

original sound that had been recorded.<br />

In the 1920’s, electrical recording methods developed that produced<br />

higher-quality recordings, <strong>and</strong> then, in the 1930’s, stereophonic<br />

recording was developed by various companies, including<br />

the British company Electrical <strong>and</strong> Musical Industries (EMI). Almost<br />

simultaneously, the technology of tape recording was developed.<br />

By the 1940’s, long-playing stereo records <strong>and</strong> tapes were<br />

widely available. As recording techniques improved further, tapes<br />

became very popular, <strong>and</strong> by the 1960’s, they had evolved into both<br />

studio master recording tapes <strong>and</strong> the audio cassettes used by consumers.


280 / Dolby noise reduction<br />

Hisses <strong>and</strong> other noises associated with sound recording <strong>and</strong> its<br />

environment greatly diminished the quality of recorded music. In<br />

1967, Ray Dolby invented a noise reducer, later named “Dolby A,”<br />

that could be used by recording studios to reduce tape signal-tonoise<br />

ratios. Several years later, his “Dolby B” system, designed<br />

for home use, became st<strong>and</strong>ard equipment in all types of playback<br />

machines. Later, Dolby <strong>and</strong> others designed improved noisesuppression<br />

systems.<br />

Recording <strong>and</strong> Tape Noise<br />

Sound is made up of vibrations of varying frequencies—sound<br />

waves—that sound recorders can convert into grooves on plastic records,<br />

varying magnetic arrangements on plastic tapes covered<br />

with iron particles, or tiny pits on compact discs. The following discussion<br />

will focus on tape recordings, for which the original Dolby<br />

noise reducers were designed.<br />

Tape recordings are made by a process that converts sound<br />

waves into electrical impulses that cause the iron particles in a tape<br />

to reorganize themselves into particular magnetic arrangements.<br />

The process is reversed when the tape is played back. In this process,<br />

the particle arrangements are translated first into electrical impulses<br />

<strong>and</strong> then into sound that is produced by loudspeakers.<br />

Erasing a tape causes the iron particles to move back into their original<br />

spatial arrangement.<br />

Whenever a recording is made, undesired sounds such as hisses,<br />

hums, pops, <strong>and</strong> clicks can mask the nuances of recorded sound, annoying<br />

<strong>and</strong> fatiguing listeners. The first attempts to do away with<br />

undesired sounds (noise) involved making tapes, recording devices,<br />

<strong>and</strong> recording studios quieter. Such efforts did not, however,<br />

remove all undesired sounds.<br />

Furthermore, advances in recording technology increased the<br />

problem of noise by producing better instruments that “heard” <strong>and</strong><br />

transmitted to recordings increased levels of noise. Such noise is often<br />

caused by the components of the recording system; tape hiss is<br />

an example of such noise. This type of noise is most discernible in<br />

quiet passages of recordings, because loud recorded sounds often<br />

mask it.


Ray Dolby<br />

Dolby noise reduction / 281<br />

Ray Dolby, born in Portl<strong>and</strong>, Oregon, in 1933, became an<br />

electronics engineer while still in high school in 1952. That is<br />

when he began working part time for Ampex Corporation,<br />

helping develop the first videotape recorder. He was responsible<br />

for the electronics in the Ampex VTR, which was marketed<br />

in 1956. The next year he finished a bachelor of science degree at<br />

Stanford University, won a Marshall Scholarship <strong>and</strong> National<br />

Science Foundation grant, <strong>and</strong> went to Cambridge University<br />

in Engl<strong>and</strong> for graduate studies. He received a Ph.D. in 1961<br />

<strong>and</strong> a fellowship to Pembroke College, during which he also<br />

consulted for the United Kingdom Atomic Energy Authority.<br />

After two years in India as a United Nations adviser, he set<br />

up Dolby Laboratories in London. It was there that he produced<br />

the sound suppression equipment that made him famous to audiophiles<br />

<strong>and</strong> movie goers, particularly in the 1970’s for the<br />

Dolby stereo (“surround sound”) that enlivened such blockbusters<br />

as Star Wars. In 1976 he moved to San Francisco <strong>and</strong><br />

opened new offices for his company. The holder of more than<br />

fifty patents, Dolby published monographs on videotape recording,<br />

long wavelength X-ray analysis, <strong>and</strong> noise reduction.<br />

He is among the most honored scientists in the recording industry.<br />

Among many other awards, he received an Oscar, Emmy,<br />

Samuel L. Warner Memorial Award, gold <strong>and</strong> silver medals<br />

from the Audio Engineering Society, <strong>and</strong> the National Medal of<br />

Technology. Engl<strong>and</strong> made him an honorary Officer of the Most<br />

Excellent Order of the British Empire, <strong>and</strong> Cambridge University<br />

<strong>and</strong> York University awarded him honorary doctorates.<br />

Because of the problem of noise in quiet passages of recorded<br />

sound, one early attempt at noise suppression involved the reduction<br />

of noise levels by using “dynaural” noise suppressors. These<br />

devices did not alter the loud portions of a recording; instead, they<br />

reduced the very high <strong>and</strong> very low frequencies in the quiet passages<br />

in which noise became most audible. The problem with such<br />

devices was, however, that removing the high <strong>and</strong> low frequencies<br />

could also affect the desirable portions of the recorded sound.<br />

These suppressors could not distinguish desirable from undesirable<br />

sounds. As recording techniques improved, dynaural noise sup-


282 / Dolby noise reduction<br />

pressors caused more <strong>and</strong> more problems, <strong>and</strong> their use was finally<br />

discontinued.<br />

Another approach to noise suppression is sound compression<br />

during the recording process. This compression is based on the fact<br />

that most noise remains at a constant level throughout a recording,<br />

regardless of the sound level of a desired signal (such as music). To<br />

carry out sound compression, the lowest-level signals in a recording<br />

are electronically elevated above the sound level of all noise. Musical<br />

nuances can be lost when the process is carried too far, because<br />

the maximum sound level is not increased by devices that use<br />

sound compression. To return the music or other recorded sound to<br />

its normal sound range for listening, devices that “exp<strong>and</strong>” the recorded<br />

music on playback are used. Two potential problems associated<br />

with the use of sound compression <strong>and</strong> expansion are the difficulty<br />

of matching the two processes <strong>and</strong> the introduction into the<br />

recording of noise created by the compression devices themselves.<br />

In 1967, Ray Dolby developed Dolby A to solve these problems as<br />

they related to tape noise (but not to microphone signals) in the recording<br />

<strong>and</strong> playing back of studio master tapes. The system operated<br />

by carrying out ten-decibel compression during recording <strong>and</strong><br />

then restoring (noiselessly) the range of the music on playback. This<br />

was accomplished by exp<strong>and</strong>ing the sound exactly to its original<br />

range. Dolby A was very expensive <strong>and</strong> was thus limited to use in recording<br />

studios. In the early 1970’s, however, Dolby invented the less<br />

expensive Dolby B system, which was intended for consumers.<br />

Consequences<br />

The development of Dolby A <strong>and</strong> Dolby B noise-reduction systems<br />

is one of the most important contributions to the high-quality<br />

recording <strong>and</strong> reproduction of sound. For this reason, Dolby A<br />

quickly became st<strong>and</strong>ard in the recording industry. In similar fashion,<br />

Dolby B was soon incorporated into virtually every highfidelity<br />

stereo cassette deck to be manufactured.<br />

Dolby’s discoveries spurred advances in the field of noise reduction.<br />

For example, the German company Telefunken <strong>and</strong> the Japanese<br />

companies Sanyo <strong>and</strong> Toshiba, among others, developed their<br />

own noise-reduction systems. Dolby Laboratories countered by


producing an improved system: Dolby C. The competition in the<br />

area of noise reduction continues, <strong>and</strong> it will continue as long as<br />

changes in recording technology produce new, more sensitive recording<br />

equipment.<br />

See also Cassette recording; Compact disc; Electronic synthesizer;<br />

FM radio; Radio; Transistor; Transistor radio; Walkman cassette<br />

player.<br />

Further Reading<br />

Dolby noise reduction / 283<br />

Alkin, E. G. M. Sound Recording <strong>and</strong> Reproduction. 3d ed. Boston: Focal<br />

Press, 1996.<br />

Baldwin, Neil. Edison: Inventing the Century. Chicago: University of<br />

Chicago Press, 2001.<br />

Wile, Frederic William. Emile Berliner, Maker of the Microphone. New<br />

York: Arno Press, 1974.


284<br />

Electric clock<br />

Electric clock<br />

The invention: Electrically powered time-keeping device with a<br />

quartz resonator that has led to the development of extremely accurate,<br />

relatively inexpensive electric clocks that are used in computers<br />

<strong>and</strong> microprocessors.<br />

The person behind the invention:<br />

Warren Alvin Marrison (1896-1980), an American scientist<br />

From Complex Mechanisms to Quartz Crystals<br />

William Alvin Marrison’s fabrication of the electric clock began a<br />

new era in time-keeping. Electric clocks are more accurate <strong>and</strong> more<br />

reliable than mechanical clocks, since they have fewer moving parts<br />

<strong>and</strong> are less likely to malfunction.<br />

An electric clock is a device that generates a string of electric<br />

pulses. The most frequently used electric clocks are called “free running”<br />

<strong>and</strong> “periodic,” which means that they generate a continuous<br />

sequence of electric pulses that are equally spaced. There are various<br />

kinds of electronic “oscillators” (materials that vibrate) that can<br />

be used to manufacture electric clocks.<br />

The material most commonly used as an oscillator in electric<br />

clocks is crystalline quartz. Because quartz (silicon dioxide) is a<br />

completely oxidized compound (which means that it does not deteriorate<br />

readily) <strong>and</strong> is virtually insoluble in water, it is chemically<br />

stable <strong>and</strong> resists chemical processes that would break down other<br />

materials. Quartz is a “piezoelectric” material, which means that it<br />

is capable of generating electricity when it is subjected to pressure<br />

or stress of some kind. In addition, quartz has the advantage of generating<br />

electricity at a very stable frequency, with little variation. For<br />

these reasons, quartz is an ideal material to use as an oscillator.<br />

The Quartz Clock<br />

A quartz clock is an electric clock that makes use of the piezoelectric<br />

properties of a quartz crystal. When a quartz crystal vibrates, a


Early electric clock. (PhotoDisc)<br />

Electric clock / 285<br />

difference of electric potential is produced between two of its faces.<br />

The crystal has a natural frequency (rate) of vibration that is determined<br />

by its size <strong>and</strong> shape. If the crystal is placed in an oscillating<br />

electric circuit that has a frequency that is nearly the same as that of<br />

the crystal, it will vibrate at its natural frequency <strong>and</strong> will cause the<br />

frequency of the entire circuit to match its own frequency.<br />

Piezoelectricity is electricity, or “electric polarity,” that is caused<br />

by the application of mechanical pressure on a “dielectric” material<br />

(one that does not conduct electricity), such as a quartz crystal. The<br />

process also works in reverse; if an electric charge is applied to the<br />

dielectric material, the material will experience a mechanical distortion.<br />

This reciprocal relationship is called “the piezoelectric effect.”<br />

The phenomenon of electricity being generated by the application<br />

of mechanical pressure is called the direct piezoelectric effect, <strong>and</strong><br />

the phenomenon of mechanical stress being produced as a result of<br />

the application of electricity is called the converse piezoelectric<br />

effect.<br />

When a quartz crystal is used to create an oscillator, the natural<br />

frequency of the crystal can be used to produce other frequencies<br />

that can power clocks. The natural frequency of a quartz crystal is<br />

nearly constant if precautions are taken when it is cut <strong>and</strong> polished<br />

<strong>and</strong> if it is maintained at a nearly constant temperature <strong>and</strong> pressure.<br />

After a quartz crystal has been used for some time, its fre-


286 / Electric clock<br />

Warren Alvin Marrison<br />

Born in Invenary, Canada, in 1896, Warren Alvin Marrison<br />

completed high school at Kingston Collegiate Institute in Ontario<br />

<strong>and</strong> attended Queen’s University in Kingston, where he<br />

studied science. World War I interrupted his studies, <strong>and</strong> while<br />

serving in the Royal Flying Corps as an electronics researcher,<br />

he began his life-long interest in radio. He graduated from university<br />

with a degree in engineering physics in 1920, transferred<br />

to Harvard University in 1921, <strong>and</strong> earned a master’s degree.<br />

After his studies, he worked for the Western Electric Company<br />

in New York, helping to develop a method to record<br />

sound on film. He moved to the company’s Bell Laboratory in<br />

1925 <strong>and</strong> studied how to produce frequency st<strong>and</strong>ards for radio<br />

transmissions. This research led him to use quartz crystals as<br />

oscillators, <strong>and</strong> he was able to step down the frequency enough<br />

that it could power a motor. Because the motor revolved at the<br />

same rate as the crystal’s frequency, he could determine the<br />

number of vibrations per time unit of the crystal <strong>and</strong> set a frequency<br />

st<strong>and</strong>ard. However, because the vibrations were constant<br />

over time, the crystal also measured time, <strong>and</strong> a new type<br />

of clock was born.<br />

For his work, Marrison received the British Horological Institute’s<br />

Gold Medal in 1947 <strong>and</strong> the Clockmakers’ Company’s<br />

Tompion Medal in 1955. He died in California in 1980.<br />

quency usually varies slowly as a result of physical changes. If allowances<br />

are made for such changes, quartz-crystal clocks such as<br />

those used in laboratories can be manufactured that will accumulate<br />

errors of only a few thous<strong>and</strong>ths of a second per month. The<br />

quartz crystals that are typically used in watches, however, may accumulate<br />

errors of tens of seconds per year.<br />

There are other materials that can be used to manufacture accurate<br />

electric clocks. For example, clocks that use the element rubidium<br />

typically would accumulate errors no larger than a few tenthous<strong>and</strong>ths<br />

of a second per year, <strong>and</strong> those that use the element cesium<br />

would experience errors of only a few millionths of a second<br />

per year. Quartz is much less expensive than rarer materials such as


ubidium <strong>and</strong> cesium, <strong>and</strong> it is easy to use in such common applications<br />

as computers. Thus, despite their relative inaccuracy, electric<br />

quartz clocks are extremely useful <strong>and</strong> popular, particularly for applications<br />

that require accurate timekeeping over a relatively short<br />

period of time. In such applications, quartz clocks may be adjusted<br />

periodically to correct for accumulated errors.<br />

Impact<br />

The electric quartz clock has contributed significantly to the development<br />

of computers <strong>and</strong> microprocessors. The computer’s control<br />

unit controls <strong>and</strong> synchronizes all data transfers <strong>and</strong> transformations<br />

in the computer system <strong>and</strong> is the key subsystem in the<br />

computer itself. Every action that the computer performs is implemented<br />

by the control unit.<br />

The computer’s control unit uses inputs from a quartz clock to<br />

derive timing <strong>and</strong> control signals that regulate the actions in the system<br />

that are associated with each computer instruction. The control<br />

unit also accepts, as input, control signals generated by other devices<br />

in the computer system.<br />

The other primary impact of the quartz clock is in making the<br />

construction of multiphase clocks a simple task. A multiphase<br />

clock is a clock that has several outputs that oscillate at the same<br />

frequency. These outputs may generate electric waveforms of different<br />

shapes or of the same shape, which makes them useful for<br />

various applications. It is common for a computer to incorporate a<br />

single-phase quartz clock that is used to generate a two-phase<br />

clock.<br />

See also Atomic clock; Carbon dating; Electric refrigerator; Fluorescent<br />

lighting; Microwave cooking; Television; Vacuum cleaner;<br />

Washing machine.<br />

Further Reading<br />

Electric clock / 287<br />

Barnett, Jo Ellen. Time’s Pendulum: From Sundials to Atomic Clocks, the<br />

Fascinating History of Time Keeping <strong>and</strong> How Our Discoveries<br />

Changed the World. San Diego: Harcourt Brace, 1999.


288 / Electric clock<br />

Dennis, Maggie, <strong>and</strong> Carlene Stephens. “Engineering Time: Inventing<br />

the Electronic Wristwatch.” British Journal for the History<br />

of Science 33, no. 119 (December, 2000).<br />

Ganeri, Anita. From C<strong>and</strong>le to Quartz Clock: The Story of Time <strong>and</strong><br />

Timekeeping. London: Evna Brothers, 1996.<br />

Thurber, Karl. “All the Time in the World.” Popular Electronics 14, no.<br />

10 (October, 1997).


Electric refrigerator<br />

Electric refrigerator<br />

The invention: An electrically powered <strong>and</strong> hermetically sealed<br />

food-storage appliance that replaced iceboxes, improved production,<br />

<strong>and</strong> lowered food-storage costs.<br />

The people behind the invention:<br />

Marcel Audiffren, a French monk<br />

Christian Steenstrup (1873-1955), an American engineer<br />

Fred Wolf, an American engineer<br />

Ice Preserves America’s Food<br />

289<br />

Before the development of refrigeration in the United States, a<br />

relatively warm climate made it difficult to preserve food. Meat<br />

spoiled within a day <strong>and</strong> milk could spoil within an hour after milking.<br />

In early America, ice was stored below ground in icehouses that<br />

had roofs at ground level. George Washington had a large icehouse<br />

at his Mount Vernon estate. By 1876, America was consuming more<br />

than 2 million tons of ice each year, which required 4,000 horses <strong>and</strong><br />

10,000 men to deliver.<br />

Several related inventions were needed before mechanical refrigeration<br />

was developed. James Watt invented the condenser, an important<br />

refrigeration system component, in 1769. In 1805, Oliver Evans<br />

presented the idea of continuous circulation of a refrigerant in a<br />

closed cycle. In this closed cooling cycle, a liquid refrigerant evaporates<br />

to a gas at low temperature, absorbing heat from its environment<br />

<strong>and</strong> thereby producing “cold,” which is circulated around an<br />

enclosed cabinet. To maintain this cooling cycle, the refrigerant gas<br />

must be returned to liquid form through condensation by compression.<br />

The first closed-cycle vapor-compression refrigerator, which<br />

was patented by Jacob Perkins in 1834, used ether as a refrigerant.<br />

Iceboxes were used in homes before refrigerators were developed.<br />

Ice was cut from lakes <strong>and</strong> rivers in the northern United States<br />

or produced by ice machines in the southern United States. An ice<br />

machine using air was patented by John Gorrie at New Orleans in<br />

1851. Ferdin<strong>and</strong> Carre introduced the first successful commercial


290 / Electric refrigerator<br />

ice machine, which used ammonia as a refrigerant, in 1862, but it<br />

was too large for home use <strong>and</strong> produced only a pound of ice per<br />

hour. Ice machinery became very dependable after 1890 but was<br />

plagued by low efficiency. Very warm summers in 1890 <strong>and</strong> 1891 cut<br />

natural ice production dramatically <strong>and</strong> increased dem<strong>and</strong> for mechanical<br />

ice production. Ice consumption continued to increase after<br />

1890; by 1914, 21 million tons of ice were used annually. The high<br />

prices charged for ice <strong>and</strong> the extremely low efficiency of home iceboxes<br />

gradually led the public to dem<strong>and</strong> a substitute for ice refrigeration.<br />

Refrigeration for the Home<br />

Domestic refrigeration required a compact unit with a built-in<br />

electric motor that did not require supervision or maintenance.<br />

Marcel Audiffren, a French monk, conceived the idea of an electric<br />

refrigerator for home use around 1910. The first electric refrigerator,<br />

which was invented by Fred Wolf in 1913, was called the Domelre,<br />

which stood for domestic electric refrigerator. This machine used<br />

condensation equipment that was housed in the home’s basement.<br />

In 1915, Alfred Mellowes built the first refrigerator to contain all of<br />

its components; this machine was known as Guardian’s Frigerator.<br />

General Motors acquired Guardian in 1918 <strong>and</strong> began to mass produce<br />

refrigerators. Guardian was renamed Frigidaire in 1919. In<br />

1918, the Kelvinator Company, run by Edmund Copel<strong>and</strong>, built the<br />

first refrigerator with automatic controls, the most important of<br />

which was the thermostatic switch. Despite these advances, by 1920<br />

only a few thous<strong>and</strong> homes had refrigerators, which cost about<br />

$1,000 each.<br />

The General Electric Company (GE) purchased the rights to the<br />

General Motors refrigerator, which was based on an improved<br />

design submitted by one of its engineers, Christian Steenstrup.<br />

Steenstrup’s innovative design included a motor <strong>and</strong> reciprocating<br />

compressor that were hermetically sealed with the refrigerant.<br />

This unit, known as the GE Monitor Top, was first produced in<br />

1927. A patent on this machine was filed for in 1926 <strong>and</strong> granted to<br />

Steenstrup in 1930. Steenstrup became chief engineer of GE’s electric<br />

refrigeration department <strong>and</strong> accumulated thirty-nine addi-


Electric refrigerator / 291<br />

tional patents in refrigeration over the following years. By 1936, he<br />

had more than one hundred patents to his credit in refrigeration <strong>and</strong><br />

other areas.<br />

Further refinement of the refrigerator evolved with the development<br />

of Freon, a nonexplosive, nontoxic, <strong>and</strong> noncorrosive refrigerant<br />

discovered by Thomas Midgely, Jr., in 1928. Freon used lower<br />

pressures than ammonia did, which meant that lighter materials<br />

<strong>and</strong> lower temperatures could be used in refrigeration.<br />

During the years following the introduction of the Monitor Top,<br />

the cost of refrigerators dropped from $1,000 in 1918 to $400 in 1926,<br />

<strong>and</strong> then to $170 in 1935. Sales of units increased from 200,000 in<br />

1926 to 1.5 million in 1935.<br />

Initially, refrigerators were sold separately from their cabinets,<br />

which commonly were used wooden iceboxes. Frigidaire began<br />

making its own cabinets in 1923, <strong>and</strong> by 1930, refrigerators that<br />

combined machinery <strong>and</strong> cabinet were sold.<br />

Throughout the 1930’s, refrigerators were well-insulated, hermetically<br />

sealed steel units that used evaporator coils to cool the<br />

food compartment. The refrigeration system was transferred from<br />

on top of to below the food storage area, which made it possible to<br />

raise the food storage area to a more convenient level. Special light<br />

bulbs that produced radiation to kill taste- <strong>and</strong> odor-bearing bacteria<br />

were used in refrigerators. Other developments included sliding<br />

shelves, shelves in doors, rounded <strong>and</strong> styled cabinet corners, ice<br />

cube trays, <strong>and</strong> even a built-in radio.<br />

The freezing capacity of early refrigerators was inadequate. Only<br />

a package or two of food could be kept cool at a time, ice cubes<br />

melted, <strong>and</strong> only a minimal amount of food could be kept frozen.<br />

The two-temperature refrigerator consisting of one compartment<br />

providing normal cooling <strong>and</strong> a separate compartment for freezing<br />

was developed by GE in 1939. Evaporator coils for cooling were<br />

placed within the refrigerator walls, providing more cooling capacity<br />

<strong>and</strong> more space for food storage. Frigidaire introduced a Cold<br />

Wall compartment, while White-Westinghouse introduced a Colder<br />

Cold system. After World War II, GE introduced the refrigeratorfreezer<br />

combination.


292 / Electric refrigerator<br />

Impact<br />

Audiffren, Wolf, Steenstrup, <strong>and</strong> others combined the earlier inventions<br />

of Watt, Perkins, <strong>and</strong> Carre with the development of electric<br />

motors to produce the electric refrigerator. The development of<br />

domestic electric refrigeration had a tremendous effect on the quality<br />

of home life. Reliable, affordable refrigeration allowed consumers<br />

a wider selection of food <strong>and</strong> increased flexibility in their daily<br />

consumption. The domestic refrigerator with increased freezer capacity<br />

spawned the growth of the frozen food industry. Without the<br />

electric refrigerator, households would still depend on unreliable<br />

supplies of ice.<br />

See also Fluorescent lighting; Food freezing; Freeze-drying; Microwave<br />

cooking; Refrigerant gas; Robot (household); Tupperware;<br />

Vacuum cleaner; Washing machine.<br />

Further Reading<br />

Anderson, Oscar Edward. Refrigeration in America: A History of a New<br />

Technology <strong>and</strong> Its Impact. Princeton: Princeton University Press,<br />

1953.<br />

Donaldson, Barry, Bernard Nagengast, <strong>and</strong> Gershon Meckler. Heat<br />

<strong>and</strong> Cold: Mastering the Great Indoors: A Selective History of Heating,<br />

Ventilation, Air-Conditioning <strong>and</strong> Refrigeration from the Ancients to<br />

the 1930’s. Atlanta, Ga.: American Society of Heating, Refrigerating<br />

<strong>and</strong> Air-Conditioning Engineers, 1994.<br />

Woolrich, Willis Raymond. The Men Who Created Cold: A History of<br />

Refrigeration. New York: Exposition Press, 1967.


Electrocardiogram<br />

Electrocardiogram<br />

The invention: Device for analyzing the electrical currents of the<br />

human heart.<br />

The people behind the invention:<br />

Willem Einthoven (1860-1927), a Dutch physiologist <strong>and</strong><br />

winner of the 1924 Nobel Prize in Physiology or Medicine<br />

Augustus D. Waller (1856-1922), a German physician <strong>and</strong><br />

researcher<br />

Sir Thomas Lewis (1881-1945), an English physiologist<br />

Horse Vibrations<br />

293<br />

In the late 1800’s, there was substantial research interest in the<br />

electrical activity that took place in the human body. Researchers<br />

studied many organs <strong>and</strong> systems in the body, including the nerves,<br />

eyes, lungs, muscles, <strong>and</strong> heart. Because of a lack of available technology,<br />

this research was tedious <strong>and</strong> frequently inaccurate. Therefore,<br />

the development of the appropriate instrumentation was as<br />

important as the research itself.<br />

The initial work on the electrical activity of the heart (detected<br />

from the surface of the body) was conducted by Augustus D. Waller<br />

<strong>and</strong> published in 1887. Many credit him with the development of<br />

the first electrocardiogram. Waller used a Lippmann’s capillary<br />

electrometer (named for its inventor, the French physicist Gabriel-<br />

Jonas Lippmann) to determine the electrical charges in the heart <strong>and</strong><br />

called his recording a “cardiograph.” The recording was made by<br />

placing a series of small tubes on the surface of the body. The tubes<br />

contained mercury <strong>and</strong> sulfuric acid. As an electrical current passed<br />

through the tubes, the mercury would exp<strong>and</strong> <strong>and</strong> contract. The resulting<br />

images were projected onto photographic paper to produce<br />

the first cardiograph. Yet Waller had only limited sucess with the<br />

device <strong>and</strong> eventually ab<strong>and</strong>oned it.<br />

In the early 1890’s, Willem Einthoven, who became a good friend<br />

of Waller, began using the same type of capillary tube to study the<br />

electrical currents of the heart. Einthoven also had a difficult time


294 / Electrocardiogram<br />

working with the instrument. His laboratory was located in an old<br />

wooden building near a cobblestone street. Teams of horses pulling<br />

heavy wagons would pass by <strong>and</strong> cause his laboratory to vibrate.<br />

This vibration affected the capillary tube, causing the cardiograph<br />

to be unclear. In his frustration, Einthoven began to modify his laboratory.<br />

He removed the floorboards <strong>and</strong> dug a hole some ten to fifteen<br />

feet deep. He lined the walls with large rocks to stabilize his instrument.<br />

When this failed to solve the problem, Einthoven, too,<br />

ab<strong>and</strong>oned the Lippmann’s capillary tube. Yet Einthoven did not<br />

ab<strong>and</strong>on the idea, <strong>and</strong> he began to experiment with other instruments.<br />

Electrocardiographs over the Phone<br />

In order to continue his research on the electrical currents of the<br />

heart, Einthoven began to work with a new device, the d’Arsonval<br />

galvanometer (named for its inventor, the French biophysicist<br />

Arsène d’Arsonval). This instrument had a heavy coil of wire suspended<br />

between the poles of a horseshoe magnet. Changes in electrical<br />

activity would cause the coil to move; however, Einthoven<br />

found that the coil was too heavy to record the small electrical<br />

changes found in the heart. Therefore, he modified the instrument<br />

by replacing the coil with a silver-coated quartz thread (string).<br />

The movements could be recorded by transmitting the deflections<br />

through a microscope <strong>and</strong> projecting them on photographic film.<br />

Einthoven called the new instrument the “string galvanometer.”<br />

In developing his string galvanomter, Einthoven was influenced<br />

by the work of one of his teachers, Johannes Bosscha. In the 1850’s,<br />

Bosscha had published a study describing the technical complexities<br />

of measuring very small amounts of electricity. He proposed the<br />

idea that a galvanometer modified with a needle hanging from a<br />

silk thread would be more sensitive in measuring the tiny electric<br />

currents of the heart.<br />

By 1905, Einthoven had improved the string galvanometer to<br />

the point that he could begin using it for clinical studies. In 1906,<br />

he had his laboratory connected to the hospital in Leiden by a telephone<br />

wire. With this arrangement, Einthoven was able to study in<br />

his laboratory electrocardiograms derived from patients in the


Willem Einthoven<br />

Electrocardiogram / 295<br />

Willem Einthoven was born in 1860 on the Isl<strong>and</strong> of Java,<br />

now part of Indonesia. His father was a Dutch army medical officer,<br />

<strong>and</strong> his mother was the daughter of the Finance Director<br />

for the Dutch East Indies. When his father died in 1870, his<br />

mother moved with her six children to Utrecht, Holl<strong>and</strong>.<br />

Einthoven entered the University of Utrecht in 1878 intending<br />

to become a physician like his father, but physics <strong>and</strong> physiology<br />

attracted him more. During his education two research<br />

projects that he conducted brought him notoriety. The first involved<br />

the articulation of the elbow, which he undertook after a<br />

sports injury of his own elbow. (He remained an avid participant<br />

in sports his whole life.) The second, which earned him his<br />

doctorate in 1885, examined stereoscopy <strong>and</strong> color variation.<br />

Because of the keen investigative abilities these studies displayed,<br />

he was at once appointed professor of physiology at the<br />

University of Leiden. He took up the position the next year, after<br />

qualifying as a general practitioner.<br />

Einthoven conducted research into asthma <strong>and</strong> the optics<br />

<strong>and</strong> electrical activity of vision before turning his attention to<br />

the heart. He developed the electrocardiogram in order to measure<br />

the heart’s electrical activity accurately <strong>and</strong> tested its applications<br />

<strong>and</strong> capacities with many students <strong>and</strong> visiting scientists,<br />

helping thereby to widen interest in it as a diagnostic tool.<br />

For this work he received the 1924 Nobel Prize in Physiology or<br />

Medicine.<br />

In his later years, Einthoven studied problems in acoustics<br />

<strong>and</strong> the electrical activity of the sympathetic nervous system.<br />

He died in Leiden in 1927.<br />

hospital, which was located a mile away. With this source of subjects,<br />

Einthoven was able to use his galvanometer to study many<br />

heart problems. As a result of these studies, Einthoven identified<br />

the following heart problems: blocks in the electrical conduction<br />

system of the heart; premature beats of the heart, including two<br />

premature beats in a row; <strong>and</strong> enlargements of the various chambers<br />

of the heart. He was also able to study how the heart behaved<br />

during the administration of cardiac drugs.


296 / Electrocardiogram<br />

A major researcher who communicated with Einthoven about<br />

the electrocardiogram was Sir Thomas Lewis, who is credited with<br />

developing the electrocardiogram into a useful clinical tool. One of<br />

Lewis’s important accomplishments was his identification of atrial<br />

fibrillation, the overactive state of the upper chambers of the heart.<br />

During World War I, Lewis was involved with studying soldiers’<br />

hearts. He designed a series of graded exercises, which he used to<br />

test the soldiers’ ability to perform work. From this study, Lewis<br />

was able to use similar tests to diagnose heart disease <strong>and</strong> to screen<br />

recruits who had heart problems.<br />

Impact<br />

As Einthoven published additional studies on the string galvanometer<br />

in 1903, 1906, <strong>and</strong> 1908, greater interest in his instrument<br />

was generated around the world. In 1910, the instrument, now<br />

called the “electrocardiograph,” was installed in the United States.<br />

It was the foundation of a new laboratory for the study of heart disease<br />

at Johns Hopkins University.<br />

As time passed, the use of the electrocardiogram—or “EKG,” as<br />

it is familiarly known—increased substantially. The major advantage<br />

of the EKG is that it can be used to diagnose problems in the<br />

heart without incisions or the use of needles. It is relatively painless<br />

for the patient; in comparison with other diagnostic techniques,<br />

moreover, it is relatively inexpensive.<br />

Recent developments in the use of the EKG have been in the area<br />

of stress testing. Since many heart problems are more evident during<br />

exercise, when the heart is working harder, EKGs are often<br />

given to patients as they exercise, generally on a treadmill. The clinician<br />

gradually increases the intensity of work the patient is doing<br />

while monitoring the patient’s heart. The use of stress testing has<br />

helped to make the EKG an even more valuable diagnostic tool.<br />

See also Amniocentesis; Artificial heart; Blood transfusion; CAT<br />

scanner; Coronary artery bypass surgery; Electroencephalogram;<br />

Heart-lung machine; Mammography; Nuclear magnetic resonance;<br />

Pacemaker; Ultrasound; X-ray image intensifier.


Further Reading<br />

Electrocardiogram / 297<br />

Cline, Barbara Lovett. Men Who Made a New Physics: Physicists <strong>and</strong><br />

the Quantum Theory. Chicago: University of Chicago Press, 1987.<br />

Hollman, Arthur. Sir Thomas Lewis: Pioneer Cardiologist <strong>and</strong> Clinical<br />

Scientist. New York: Springer, 1997.<br />

Lewis, Thomas. Collected Works on Heart Disease. 1912. Reprint. New<br />

York: Classics of Cardiology Library, 1991.<br />

Snellen, H. A. Two Pioneers of Electrocardiography: The Correspondence<br />

Between Einthoven <strong>and</strong> Lewis from 1908-1926. Rotterdam: Donker<br />

Academic <strong>Public</strong>ations, 1983.<br />

_____. Willem Einthoven, 1860-1927, Father of Electrocardiography: Life<br />

<strong>and</strong> Work, Ancestors <strong>and</strong> Contemporaries. Boston: Kluwer Academic<br />

Publishers, 1995.


298<br />

Electroencephalogram<br />

Electroencephalogram<br />

The invention: A system of electrodes that measures brain wave<br />

patterns in humans, making possible a new era of neurophysiology.<br />

The people behind the invention:<br />

Hans Berger (1873-1941), a German psychiatrist <strong>and</strong> research<br />

scientist<br />

Richard Caton (1842-1926), an English physiologist <strong>and</strong> surgeon<br />

The Electrical Activity of the Brain<br />

Hans Berger’s search for the human electroencephalograph (English<br />

physiologist Richard Caton had described the electroencephalogram,<br />

or “brain wave,” in rabbits <strong>and</strong> monkeys in 1875) was motivated<br />

by his desire to find a physiological method that might be<br />

applied successfully to the study of the long-st<strong>and</strong>ing problem of<br />

the relationship between the mind <strong>and</strong> the brain. His scientific career,<br />

therefore, was directed toward revealing the psychophysical<br />

relationship in terms of principles that would be rooted firmly in the<br />

natural sciences <strong>and</strong> would not have to rely upon vague philosophical<br />

or mystical ideas.<br />

During his early career, Berger attempted to study psychophysical<br />

relationships by making plethysmographic measurements of<br />

changes in the brain circulation of patients with skull defects. In<br />

plethysmography, an instrument is used to indicate <strong>and</strong> record by<br />

tracings the variations in size of an organ or part of the body. Later,<br />

Berger investigated temperature changes occurring in the human<br />

brain during mental activity <strong>and</strong> the action of psychoactive drugs.<br />

He became disillusioned, however, by the lack of psychophysical<br />

underst<strong>and</strong>ing generated by these investigations.<br />

Next, Berger turned to the study of the electrical activity of the<br />

brain, <strong>and</strong> in the 1920’s he set out to search for the human electroencephalogram.<br />

He believed that the electroencephalogram would finally<br />

provide him with a physiological method capable of furnishing<br />

insight into mental functions <strong>and</strong> their disturbances.


Electroencephalogram / 299<br />

Berger made his first unsuccessful attempt at recording the electrical<br />

activity of the brain in 1920, using the scalp of a bald medical<br />

student. He then attempted to stimulate the cortex of patients with<br />

skull defects by using a set of electrodes to apply an electrical current<br />

to the skin covering the defect. The main purpose of these<br />

stimulation experiments was to elicit subjective sensations. Berger<br />

hoped that eliciting these sensations might give him some clue<br />

about the nature of the relationship between the physiochemical<br />

events produced by the electrical stimulus <strong>and</strong> the mental processes<br />

revealed by the patients’ subjective experience. The availability<br />

of many patients with skull defects—in whom the pulsating<br />

surface of the brain was separated from the stimulating electrodes<br />

by only a few millimeters of tissue—reactivated Berger’s interest<br />

in recording the brain’s electrical activity.<br />

Hans Berger<br />

Hans Berger, the father of electroencephalography, was born<br />

in Neuses bei Coburn, Germany, in 1873. He entered the University<br />

of Jena in 1892 as a medical student <strong>and</strong> became an assistant<br />

in the psychiatric clinic in 1897. In 1912 he was appointed<br />

the clinic’s chief doctor <strong>and</strong> then its director <strong>and</strong> a university<br />

professor of psychiatry. In 1919 he was chosen as rector of the<br />

university.<br />

Berger hoped to settle the long-st<strong>and</strong>ing philosophical question<br />

about the brain <strong>and</strong> the mind by finding observable physical<br />

processes that correlated with thought <strong>and</strong> feelings. He<br />

started off by studying the blood circulation in the head <strong>and</strong><br />

brain temperature. Even though this work founded psychophysiology,<br />

he failed to find objective evidence of subjective<br />

states until he started examining fluctuations in the electrical<br />

potential of the brain in 1924. His 1929 paper describing the<br />

electroencephalograph later provided medicine with a basic diagnostic<br />

tool, but the instrument proved to be a very confusing<br />

probe of the human psyche for him. His colleagues in psychiatry<br />

<strong>and</strong> medicine did not accept his relationships of physical<br />

phenomena <strong>and</strong> mental states.<br />

Berger retired as professor emeritus in 1938 <strong>and</strong> died three<br />

years later in Jena.


300 / Electroencephalogram<br />

Small, Tremulous Movements<br />

Berger used several different instruments in trying to detect<br />

brain waves, but all of them used a similar method of recording.<br />

Electrical oscillations deflected a mirror upon which a light beam<br />

was projected. The deflections of the light beam were proportional<br />

to the magnitude of the electrical signals. The movement of the spot<br />

of the light beam was recorded on photographic paper moving at a<br />

speed no greater than 3 centimeters per second.<br />

In July, 1924, Berger observed small, tremulous movements of<br />

the instrument while recording from the skin overlying a bone defect<br />

in a seventeen-year-old patient. In his first paper on the electroencephalogram,<br />

Berger described this case briefly as his first successful<br />

recording of an electroencephalogram. At the time of these<br />

early studies, Berger already had used the term “electroencephalogram”<br />

in his diary. Yet for several years he had doubts about the origin<br />

of the electrical signals he recorded. As late as 1928, he almost<br />

ab<strong>and</strong>oned his electrical recording studies.<br />

The publication of Berger’s first paper on the human encephalogram<br />

in 1929 had little impact on the scientific world. It was either<br />

ignored or regarded with open disbelief. At this time, even<br />

when Berger himself was not completely free of doubts about the<br />

validity of his findings, he managed to continue his work. He published<br />

additional contributions to the study of the electroencephalogram<br />

in a series of fourteen papers. As his research progressed,<br />

Berger became increasingly confident <strong>and</strong> convinced of the significance<br />

of his discovery.<br />

Impact<br />

The long-range impact of Berger’s work is incontestable. When<br />

Berger published his last paper on the human encephalogram in<br />

1938, the new approach to the study of brain function that he inaugurated<br />

in 1929 had gathered momentum in many centers, both in<br />

Europe <strong>and</strong> in the United States. As a result of his pioneering work,<br />

a new diagnostic method had been introduced into medicine. Physiology<br />

had acquired a new investigative tool. Clinical neurophysiology<br />

had been liberated from its dependence upon the functional


anatomical approach, <strong>and</strong> electrophysiological exploration of complex<br />

functions of the central nervous system had begun in earnest.<br />

Berger’s work had finally received its well-deserved recognition.<br />

Many of those who undertook the study of the electroencephalogram<br />

were able to bring a far greater technical knowledge of<br />

neurophysiology to bear upon the problems of the electrical activity<br />

of the brain. Yet the community of neurological scientists has not<br />

ceased to look with respect to the founder of electroencephalography,<br />

who, despite overwhelming odds <strong>and</strong> isolation, opened a new<br />

area of neurophysiology.<br />

See also Amniocentesis; CAT scanner; Electrocardiogram; Mammography;<br />

Nuclear magnetic resonance; Ultrasound; X-ray image<br />

intensifier.<br />

Further Reading<br />

Electroencephalogram / 301<br />

Barlow, John S. The Electroencephalogram: Its Patterns <strong>and</strong> Origins.<br />

Cambridge, Mass.: MIT Press, 1993.<br />

Berger, Hans. Hans Berger on the Electroencephalogram of Man. New<br />

York: Elsevier, 1969.


302<br />

Electron microscope<br />

Electron microscope<br />

The invention: A device for viewing extremely small objects that<br />

uses electron beams <strong>and</strong> “electron lenses” instead of the light<br />

rays <strong>and</strong> optical lenses used by ordinary microscopes.<br />

The people behind the invention:<br />

Ernst Ruska (1906-1988), a German engineer, researcher, <strong>and</strong><br />

inventor who shared the 1986 Nobel Prize in Physics<br />

Hans Busch (1884-1973), a German physicist<br />

Max Knoll (1897-1969), a German engineer <strong>and</strong> professor<br />

Louis de Broglie (1892-1987), a French physicist who won the<br />

1929 Nobel Prize in Physics<br />

Reaching the Limit<br />

The first electron microscope was constructed by Ernst Ruska<br />

<strong>and</strong> Max Knoll in 1931. Scientists who look into the microscopic<br />

world always dem<strong>and</strong> microscopes of higher <strong>and</strong> higher resolution<br />

(resolution is the ability of an optical instrument to distinguish<br />

closely spaced objects). As early as 1834, George Airy, the eminent<br />

British astronomer, theorized that there should be a natural limit to<br />

the resolution of optical microscopes. In 1873, two Germans, Ernst<br />

Abbe, cofounder of the Karl Zeiss Optical Works at Jena, <strong>and</strong> Hermann<br />

von Helmholtz, the famous physicist <strong>and</strong> philosopher, independently<br />

published papers on this issue. Both arrived at the same<br />

conclusion as Airy: Light is limited by the size of its wavelength.<br />

Specifically, light cannot resolve smaller than one-half the height of<br />

its wavelength.<br />

One solution to this limitation was to experiment with light, or<br />

electromagnetic radiation, or shorter <strong>and</strong> shorter wavelengths.<br />

At the beginning of the twentieth century, Joseph Edwin Barnard<br />

experimented on microscopes using ultraviolet light. Such instruments,<br />

however, only modestly improved the resolution. In<br />

1912, German physicist Max von Laue considered using X rays.<br />

At the time, however, it was hard to turn “X-ray microscopy” into<br />

a physical reality. The wavelengths of X rays are exceedingly


short, but for the most part they are used to penetrate matter, not<br />

to illuminate objects. It appeared that microscopes had reached<br />

their limit.<br />

Matter Waves<br />

Electron microscope / 303<br />

In a new microscopy, then, light—even electromagnetic radiation<br />

in general—as the medium that traditionally carried image information,<br />

had to be replaced by a new medium. In 1924, French<br />

theoretical physicist Louis de Broglie advanced a startling hypothesis:<br />

Matter on the scale of subatomic particles possesses wave<br />

characteristics. De Broglie also concluded that the speed of lowmass<br />

subatomic particles, such as electrons, is related to wavelength.<br />

Specifically, higher speeds correspond to shorter wavelengths.<br />

When Knoll <strong>and</strong> Ruska built the first electron microscope in 1931,<br />

they had never heard about de Broglie’s “matter wave.” Ruska recollected<br />

that when, in 1932, he <strong>and</strong> Knoll first learned about de<br />

Broglie’s idea, he realized that those matter waves would have to be<br />

many times shorter in wavelength than light waves.<br />

The core component of the new instrument was the electron<br />

beam, or “cathode ray,” as it was usually called then. The cathoderay<br />

tube was invented in 1857 <strong>and</strong> was the source of a number of<br />

discoveries, including X rays. In 1896, Olaf Kristian Birkel<strong>and</strong>, a<br />

Norwegian scientist, after experimenting with the effect of parallel<br />

magnetic fields on the electron beam of the cathode-ray tube, concluded<br />

that cathode rays that are concentrated on a focal point by<br />

means of a magnet are as effective as parallel light rays that are concentrated<br />

by means of a lens.<br />

From around 1910, German physicist Hans Busch was the leading<br />

researcher in the field. In 1926, he published his theory on the<br />

trajectories of electrons in magnetic fields. His conclusions confirmed<br />

<strong>and</strong> exp<strong>and</strong>ed upon those of Birkel<strong>and</strong>. As a result, Busch<br />

has been recognized as the founder of a new field later known<br />

as “electron optics.” His theoretical study showed, among other<br />

things, that the analogy between light <strong>and</strong> lenses on the one h<strong>and</strong>,<br />

<strong>and</strong> electron beams <strong>and</strong> electromagnetic lenses, on the other h<strong>and</strong>,<br />

was accurate.


304 / Electron microscope<br />

Ernst Ruska<br />

Ernst August Friedrich Ruska was born in 1906 in Heidelberg<br />

to Professor Julius Ruska <strong>and</strong> his wife, Elisabeth. In 1925<br />

he left home for the Technical College of Munich, moving two<br />

years later to the Technical College of Berlin <strong>and</strong> gaining practical<br />

training at nearby Siemens <strong>and</strong> Halsk Limited. During his<br />

university days he became interested in vacuum tube technology<br />

<strong>and</strong> worked at the Institute of High Voltage, participating<br />

in the development of a high performance cathode ray oscilloscope.<br />

His interests also lay with the theory <strong>and</strong> application of electron<br />

optics. In 1929, as part of his graduate work, Ruska published<br />

a proof of Hans Busch’s theory explaining possible lenslike<br />

effects of a magnetic field on an electron stream, which led<br />

to the invention of the polschuh lens. It formed the core of the<br />

electron microscope that Ruska built with his mentor, Max<br />

Kroll, in 1931.<br />

Ruska completed his doctoral studies in 1934, but he had already<br />

found work in industry, believing that further technical<br />

development of electron microscopes was beyond the means of<br />

university laboratories. He worked for Fernseh Limited from<br />

1933 to 1937 <strong>and</strong> for Siemens from 1937 to 1955. Following<br />

World War II he helped set up the Institute of Electron Optics<br />

<strong>and</strong> worked in the Faculty of Medicine <strong>and</strong> Biology of the German<br />

Academy of Sciences. He joined the Fritz Haber Institute<br />

of the Max Planck Society in Berlin in 1949 <strong>and</strong> took over as director<br />

of its Institute for Electron Microscopy in 1955, keeping<br />

the position until he retired in 1974.<br />

His life-long work with electron microscopy earned Ruska<br />

half of the 1986 Nobel Prize in Physics. He died two years later.<br />

To honor his memory, European manufacturers of electron microscopes<br />

instituted the Ernst Ruska Prizes, one for researchers<br />

of materials <strong>and</strong> optics <strong>and</strong> one for biomedical researchers.<br />

Beginning in 1928, Ruska, as a graduate student at the Berlin Institute<br />

of Technology, worked on refining Busch’s work. He found<br />

that the energy of the electrons in the beam was not uniform. This<br />

nonuniformity meant that the images of microscopic objects would<br />

ultimately be fuzzy. Knoll <strong>and</strong> Ruska were able to work from the


ecognition of this problem to the design <strong>and</strong> materialization of a<br />

concentrated electron “writing spot” <strong>and</strong> to the actual construction<br />

of the electron microscope. By April, 1931, they had established a<br />

technological l<strong>and</strong>mark with the “first constructional realization of<br />

an electron microscope.”<br />

Impact<br />

The world’s first electron microscope, which took its first photographic<br />

record on April 7, 1931, was rudimentary. Its two-stage total<br />

magnification was only sixteen times larger than the sample. Since<br />

Ruska <strong>and</strong> Knoll’s creation, however, progress in electron microscopy<br />

has been spectacular. Such an achievement is one of the prominent<br />

examples that illustrate the historically unprecedented pace of<br />

science <strong>and</strong> technology in the twentieth century.<br />

In 1935, for the first time, the electron microscope surpassed<br />

the optical microscope in resolution. The problem of damaging<br />

the specimen by the heating effects of the electron beam proved<br />

to be more difficult to resolve. In 1937, a team at the University of<br />

Toronto constructed the first generally usable electron microscope.<br />

In 1942, a group headed by James Hillier at the Radio Corporation<br />

of America produced commercial transmission electron<br />

microscopes. In 1939 <strong>and</strong> 1940, research papers on electron microscopes<br />

began to appear in Sweden, Canada, the United States,<br />

<strong>and</strong> Japan; from 1944 to 1947, papers appeared in Switzerl<strong>and</strong>,<br />

France, the Soviet Union, The Netherl<strong>and</strong>s, <strong>and</strong> Engl<strong>and</strong>. Following<br />

research work in laboratories, commercial transmission electron<br />

microscopes using magnetic lenses with short focal lengths<br />

also appeared in these countries.<br />

See also Cyclotron; Field ion microscope; Geiger counter; Mass<br />

spectrograph; Neutrino detector; Scanning tunneling microscope;<br />

Synchrocyclotron; Tevatron accelerator; Ultramicroscope.<br />

Further Reading<br />

Electron microscope / 305<br />

Cline, Barbara Lovett. Men Who Made a New Physics: Physicists <strong>and</strong><br />

the Quantum Theory. Chicago: University of Chicago Press, 1987.


306 / Electron microscope<br />

Hawkes, P. W. The Beginnings of Electron Microscopy. Orl<strong>and</strong>o: Academic<br />

Press, 1985.<br />

Marton, Ladislaus. Early History of the Electron Microscope. 2d ed. San<br />

Francisco: San Francisco Press, 1994.<br />

Rasmussen, Nicolas. Picture Control: The Electron Microscope <strong>and</strong> the<br />

Transformation of Biology in America, 1940-1960. Stanford, Calif.:<br />

Stanford University Press, 1997.


Electronic synthesizer<br />

Electronic synthesizer<br />

The invention: Portable electronic device that both simulates the<br />

sounds of acoustic instruments <strong>and</strong> creates entirely new sounds.<br />

The person behind the invention:<br />

Robert A. Moog (1934- ), an American physicist, engineer,<br />

<strong>and</strong> inventor<br />

From Harmonium to Synthesizer<br />

307<br />

The harmonium, or acoustic reed organ, is commonly viewed as<br />

having evolved into the modern electronic synthesizer that can be<br />

used to create many kinds of musical sounds, from the sounds of<br />

single or combined acoustic musical instruments to entirely original<br />

sounds. The first instrument to be called a synthesizer was patented<br />

by the Frenchman J. A. Dereux in 1949. Dereux’s synthesizer, which<br />

amplified the acoustic properties of harmoniums, led to the development<br />

of the recording organ.<br />

Next, several European <strong>and</strong> American inventors altered <strong>and</strong><br />

augmented the properties of such synthesizers. This stage of the<br />

process was followed by the invention of electronic synthesizers,<br />

which initially used electronically generated sounds to imitate<br />

acoustic instruments. It was not long, however, before such synthesizers<br />

were used to create sounds that could not be produced by any<br />

other instrument. Among the early electronic synthesizers were<br />

those made in Germany by Herbert Elmert <strong>and</strong> Robert Beyer in<br />

1953, <strong>and</strong> the American Olsen-Belar synthesizers, which were developed<br />

in 1954. Continual research produced better <strong>and</strong> better versions<br />

of these large, complex electronic devices.<br />

Portable synthesizers, which are often called “keyboards,” were<br />

then developed for concert <strong>and</strong> home use. These instruments became<br />

extremely popular, especially in rock music. In 1964, Robert A.<br />

Moog, an electronics professor, created what are thought by many<br />

to be the first portable synthesizers to be made available to the public.<br />

Several other well-known portable synthesizers, such as ARP<br />

<strong>and</strong> Buchla synthesizers, were also introduced at about the same


308 / Electronic synthesizer<br />

time. Currently, many companies manufacture studio-quality synthesizers<br />

of various types.<br />

Synthesizer Components <strong>and</strong> Operation<br />

Modern synthesizers make music electronically by building up<br />

musical phrases via numerous electronic circuits <strong>and</strong> combining<br />

those phrases to create musical compositions. In addition to duplicating<br />

the sounds of many instruments, such synthesizers also enable<br />

their users to create virtually any imaginable sound. Many<br />

sounds have been created on synthesizers that could not have been<br />

created in any other way.<br />

Synthesizers use sound-processing <strong>and</strong> sound-control equipment<br />

that controls “white noise” audio generators <strong>and</strong> oscillator circuits.<br />

This equipment can be manipulated to produce a huge variety of<br />

sound frequencies <strong>and</strong> frequency mixtures in the same way that a<br />

beam of white light can be manipulated to produce a particular<br />

color or mixture of colors.<br />

Once the desired products of a synthesizer’s noise generator <strong>and</strong><br />

oscillators are produced, percussive sounds that contain all or many<br />

audio frequencies are mixed with many chosen individual sounds<br />

<strong>and</strong> altered by using various electronic processing components. The<br />

better the quality of the synthesizer, the more processing components<br />

it will possess. Among these components are sound amplifiers,<br />

sound mixers, sound filters, reverberators, <strong>and</strong> sound combination<br />

devices.<br />

Sound amplifiers are voltage-controlled devices that change the<br />

dynamic characteristics of any given sound made by a synthesizer.<br />

Sound mixers make it possible to combine <strong>and</strong> blend two or more<br />

manufactured sounds while controlling their relative volumes.<br />

Sound filters affect the frequency content of sound mixtures by increasing<br />

or decreasing the amplitude of the sound frequencies<br />

within particular frequency ranges, which are called “b<strong>and</strong>s.”<br />

Sound filters can be either b<strong>and</strong>-pass filters or b<strong>and</strong>-reject filters.<br />

They operate by increasing or decreasing the amplitudes of sound<br />

frequencies within given ranges (such as treble or bass). Reverberators<br />

(or “reverb” units) produce artificial echoes that can have significant<br />

musical effects. There are also many other varieties of sound-


Robert Moog<br />

Electronic synthesizer / 309<br />

Robert Moog, born in 1934, grew up in the Queens borough<br />

of New York City, a tough area for a brainy kid. To avoid the<br />

bullies who picked on him because he was a nerd, Moog spent a<br />

lot of time helping his father with his hobby, electronics. At<br />

fourteen, he built his own theremin, an eerie-sounding forerunner<br />

of electric instruments.<br />

Moog’s mother, meanwhile, force-fed him piano lessons. He<br />

liked science better <strong>and</strong> majored in physics at Queens College<br />

<strong>and</strong> then Cornell University, but he did not forget the music.<br />

While in college, he designed a kit for making theremins <strong>and</strong><br />

advertised it, selling enough of them to run up a sizable bankroll.<br />

Also while in college, Moog, acting on a suggestion from a<br />

composer, put together the first easy-to-play electronic synthesizer.<br />

Other music synthesizers already existed, but they were<br />

large, complex, <strong>and</strong> expensive—suitable only for recording studios.<br />

When Moog unveiled his synthesizer in 1965, it was portable,<br />

sold for one-tenth the price, <strong>and</strong> gave musicians virtually<br />

an orchestra at their fingertips. It became a stage instrument.<br />

Walter Carlos used a Moog synthesizer in 1969 for his album<br />

Switched-on Bach, electronic renditions of Johann Sebastian Bach’s<br />

concertos. It was a hit <strong>and</strong> won a Grammy award. The album<br />

made Moog <strong>and</strong> his new instrument famous. Its reputation<br />

grew when the Beatles used it for “Because” on Abbey Road <strong>and</strong><br />

Carlos recorded the score for Stanley Kubrick’s classic movie A<br />

Clockwork Orange on a Moog. With the introduction of the even<br />

more portable Minimoog, the popularity of synthesizers soared,<br />

especially among rock musicians but also in jazz <strong>and</strong> other<br />

styles.<br />

Moog sold his company <strong>and</strong> moved to North Carolina in<br />

1978. There he started another company, Big Briar, devoted to<br />

designing special instruments, such as a keyboard that can be<br />

played with as much expressive subtlety as a violin <strong>and</strong> an interactive<br />

piano.<br />

processing elements, among them sound-envelope generators,<br />

spatial locators, <strong>and</strong> frequency shifters. Ultimately, the soundcombination<br />

devices put together the results of the various groups<br />

of audio generating <strong>and</strong> processing elements, shaping the sound<br />

that has been created into its final form.


310 / Electronic synthesizer<br />

A variety of control elements are used to integrate the operation<br />

of synthesizers. Most common is the keyboard, which provides the<br />

name most often used for portable electronic synthesizers. Portable<br />

synthesizer keyboards are most often pressure-sensitive devices<br />

(meaning that the harder one presses the key, the louder the resulting<br />

sound will be) that resemble the black-<strong>and</strong>-white keyboards of<br />

more conventional musical instruments such as the piano <strong>and</strong> the<br />

organ. These synthesizer keyboards produce two simultaneous outputs:<br />

control voltages that govern the pitches of oscillators, <strong>and</strong> timing<br />

pulses that sustain synthesizer responses for as long as a particular<br />

key is depressed.<br />

Unseen but present are the integrated voltage controls that control<br />

overall signal generation <strong>and</strong> processing. In addition to voltage<br />

controls <strong>and</strong> keyboards, synthesizers contain buttons <strong>and</strong> other<br />

switches that can transpose their sound ranges <strong>and</strong> other qualities.<br />

Using the appropriate buttons or switches makes it possible for a<br />

single synthesizer to imitate different instruments—or groups of instruments—at<br />

different times. Other synthesizer control elements<br />

include sample-<strong>and</strong>-hold devices <strong>and</strong> r<strong>and</strong>om voltage sources that<br />

make it possible to sustain particular musical effects <strong>and</strong> to add various<br />

effects to the music that is being played, respectively.<br />

Electronic synthesizers are complex <strong>and</strong> flexible instruments.<br />

The various types <strong>and</strong> models of synthesizers make it possible to<br />

produce many different kinds of music, <strong>and</strong> many musicians use a<br />

variety of keyboards to give them great flexibility in performing<br />

<strong>and</strong> recording.<br />

Impact<br />

The development <strong>and</strong> wide dissemination of studio <strong>and</strong> portable<br />

synthesizers has led to their frequent use to combine the sound<br />

properties of various musical instruments; a single musician can<br />

thus produce, inexpensively <strong>and</strong> with a single instrument, sound<br />

combinations that previously could have been produced only by a<br />

large number of musicians playing various instruments. (Underst<strong>and</strong>ably,<br />

many players of acoustic instruments have been upset by<br />

this development, since it means that they are hired to play less often<br />

than they were before synthesizers were developed.) Another


consequence of synthesizer use has been the development of entirely<br />

original varieties of sound, although this area has been less<br />

thoroughly explored, for commercial reasons. The development of<br />

synthesizers has also led to the design of other new electronic music-making<br />

techniques <strong>and</strong> to the development of new electronic<br />

musical instruments.<br />

Opinions about synthesizers vary from person to person—<strong>and</strong>,<br />

in the case of certain illustrious musicians, from time to time. One<br />

well-known musician initially proposed that electronic synthesizers<br />

would replace many or all conventional instruments, particularly<br />

pianos. Two decades later, though, this same musician noted<br />

that not even the best modern synthesizers could match the quality<br />

of sound produced by pianos made by manufacturers such as<br />

Steinway <strong>and</strong> Baldwin.<br />

See also Broadcaster guitar; Cassette recording; Compact disc;<br />

Dolby noise reduction; Transistor.<br />

Further Reading<br />

Electronic synthesizer / 311<br />

Hopkin, Bart. Gravikords, Whirlies <strong>and</strong> Pyrophones: Experimental Musical<br />

Instruments. Roslyn, N.Y.: Ellipsis Arts, 1996.<br />

Koener, Brendan I.. “Back to Music’s Future.” U.S. News & World Report<br />

122, no. 8 (March 3, 1997).<br />

Nunziata, Susan. “Moog Keyboard Offers Human Touch.” Billboard<br />

104, no. 7 (February 15, 1992).<br />

Shapiro, Peter. Modulations: A History of Electronic Music: Throbbing<br />

Words on Sound. New York: Caipirinha Productions, 2000.


312<br />

ENIAC computer<br />

ENIAC computer<br />

The invention: The first general-purpose electronic digital computer.<br />

The people behind the invention:<br />

John Presper Eckert (1919-1995), an electrical engineer<br />

John William Mauchly (1907-1980), a physicist, engineer, <strong>and</strong><br />

professor<br />

John von Neumann (1903-1957), a Hungarian American<br />

mathematician, physicist, <strong>and</strong> logician<br />

Herman Heine Goldstine (1913- ), an army mathematician<br />

Arthur Walter Burks (1915- ), a philosopher, engineer, <strong>and</strong><br />

professor<br />

John Vincent Atanasoff (1903-1995), a mathematician <strong>and</strong><br />

physicist<br />

A Technological Revolution<br />

The Electronic Numerical Integrator <strong>and</strong> Calculator (ENIAC) was<br />

the first general-purpose electronic digital computer. By demonstrating<br />

the feasibility <strong>and</strong> value of electronic digital computation, it initiated<br />

the computer revolution. The ENIAC was developed during<br />

World War II (1939-1945) at the Moore School of Electrical Engineering<br />

by a team headed by John William Mauchly <strong>and</strong> John Presper<br />

Eckert, who were working on behalf of the U.S. Ordnance Ballistic<br />

Research Laboratory (BRL) at the Aberdeen Proving Ground in<br />

Maryl<strong>and</strong>. Early in the war, the BRL’s need to generate ballistic firing<br />

tables already far outstripped the combined abilities of the available<br />

differential analyzers <strong>and</strong> teams of human computers.<br />

In 1941, Mauchly had seen the special-purpose electronic computer<br />

developed by John Vincent Atanasoff to solve sets of linear<br />

equations. Atanasoff’s computer was severely limited in scope <strong>and</strong><br />

was never fully completed. The functioning prototype, however,<br />

helped convince Mauchly of the feasibility of electronic digital computation<br />

<strong>and</strong> so led to Mauchly’s formal proposal in April, 1943, to<br />

develop the general-purpose ENIAC. The BRL, in desperate need of<br />

computational help, agreed to fund the project, with Lieutenant


Herman Heine Goldstine overseeing it for the U.S. Army.<br />

This first substantial electronic computer was designed, built,<br />

<strong>and</strong> debugged within two <strong>and</strong> one-half years. Even given the highly<br />

talented team, it could be done only by taking as few design risks as<br />

possible. The ENIAC ended up as an electronic version of prior<br />

computers: Its functional organization was similar to that of the<br />

differential analyzer, while it was programmed via a plugboard<br />

(which was something like a telephone switchboard), much like the<br />

earlier electromechanical calculators made by the International Business<br />

Machines (IBM) Corporation. Another consequence was that<br />

the internal representation of numbers was decimal rather than the<br />

now-st<strong>and</strong>ard binary, since the familiar electromechanical computers<br />

used decimal digits.<br />

Although the ENIAC was completed only after the end of the<br />

war, it was used primarily for military purposes. In fact, the first<br />

production run on the system was a two-month calculation needed<br />

for the design of the hydrogen bomb. John von Neumann, working<br />

as a consultant to both the Los Alamos Scientific Laboratory <strong>and</strong> the<br />

ENIAC project, arranged for the production run immediately prior<br />

to ENIAC’s formal dedication in 1946.<br />

A Very Fast Machine<br />

ENIAC computer / 313<br />

The ENIAC was an impressive machine: It contained 18,000 vacuum<br />

tubes, weighed 27 metric tons, <strong>and</strong> occupied a large room. The<br />

final cost to the U.S. Army was about $486,000. For this price, the<br />

army received a machine that computed up to a thous<strong>and</strong> times<br />

faster than its electromechanical precursors; for example, addition<br />

<strong>and</strong> subtraction required only 200 microseconds (200 millionths of a<br />

second). At its dedication ceremony, the ENIAC was fast enough to<br />

calculate a fired shell’s trajectory faster than the shell itself took to<br />

reach its target.<br />

The machine also was much more complex than any predecessor<br />

<strong>and</strong> employed a risky new technology in vacuum tubes; this caused<br />

much concern about its potential reliability. In response to this concern,<br />

Eckert, the lead engineer, imposed strict safety factors on all<br />

components, requiring the design to use components at a level well<br />

below the manufacturers’ specified limits. The result was a machine


314 / ENIAC computer<br />

that ran for as long as three days without a hardware malfunction.<br />

Programming the ENIAC was effected by setting switches <strong>and</strong><br />

physically connecting accumulators, function tables (a kind of manually<br />

set read-only memory), <strong>and</strong> control units. Connections were<br />

made via cables running between plugboards. This was a laborious<br />

<strong>and</strong> error-prone process, often requiring a one-day set time.<br />

The team recognized this problem, <strong>and</strong> in early 1945, Eckert,<br />

Mauchly, <strong>and</strong> Neumann worked on the design of a new machine.<br />

Their basic idea was to treat both program <strong>and</strong> data in the same way,<br />

<strong>and</strong> in particular to store them in the same high-speed memory; in<br />

other words, they planned to produce a stored-program computer.<br />

Neumann described <strong>and</strong> explained this design in his “First Draft of<br />

a Report on the EDVAC” (EDVAC is an acronym for Electronic Discrete<br />

Variable Automatic Computer). In his report, Neumann contributed<br />

new design techniques <strong>and</strong> provided the first general, comprehensive<br />

description of the stored-program architecture.<br />

After the delivery of the ENIAC, Neumann suggested that it<br />

could be wired up so that a set of instructions would be permanently<br />

available <strong>and</strong> could be selected by entries in the function tables.<br />

Engineers implemented the idea, providing sixty instructions<br />

that could be invoked from the programs stored into the function tables.<br />

Despite slowing down the computer’s calculations, this technique<br />

was so superior to plugboard programming that it was used<br />

exclusively thereafter. In this way, the ENIAC was converted into a<br />

kind of primitive stored-program computer.<br />

Impact<br />

The ENIAC’s electronic speed <strong>and</strong> the stored-program design of<br />

the EDVAC posed a serious engineering challenge: to produce a<br />

computer memory that would be large, inexpensive, <strong>and</strong> fast. Without<br />

such fast memories, the electronic control logic would spend<br />

most of its time idling. Vacuum tubes themselves (used in the control)<br />

were not an effective answer because of their large power requirements<br />

<strong>and</strong> heat generation.<br />

The EDVAC design draft proposed using mercury delay lines,<br />

which had been used earlier in radars. These delay lines converted<br />

an electronic signal into a slower acoustic signal in a mercury solu-


tion; for continuous storage, the signal picked up at the other end<br />

was regenerated <strong>and</strong> sent back into the mercury. Maurice Vincent<br />

Wilkes at the University of Cambridge was the first to complete<br />

such a system, in May, 1949. One month earlier, Frederick Call<strong>and</strong><br />

Williams <strong>and</strong> Tom Kilburn at Manchester University had brought<br />

their prototype computer into operation, which used cathode-ray<br />

tubes (CRTs) for its main storage. Thus, Engl<strong>and</strong> took an early lead<br />

in developing computing systems, largely because of a more immediate<br />

practical design approach.<br />

In the meantime, Eckert <strong>and</strong> Mauchly formed the Electronic Control<br />

Company (later the Eckert-Mauchly Computer Corporation).<br />

They produced the Binary Automatic Computer (BINAC) in 1949<br />

<strong>and</strong> the Universal Automatic Computer (UNIVAC) I in 1951; both<br />

machines used mercury storage.<br />

The memory problem that the ENIAC introduced was finally resolved<br />

with the invention of the magnetic core in the early 1950’s.<br />

Core memory was installed on the ENIAC <strong>and</strong> soon on all new machines.<br />

The ENIAC continued in operation until October, 1955,<br />

when parts of it were retired to the Smithsonian Institution. The<br />

ENIAC proved the viability of digital electronics <strong>and</strong> led directly to<br />

the development of stored-program computers. Its impact can be<br />

seen in every modern digital computer.<br />

See also Apple II computer; BINAC computer; Colossus computer;<br />

IBM Model 1401 computer; Personal computer; Supercomputer;<br />

UNIVAC computer.<br />

Further Reading<br />

ENIAC computer / 315<br />

Burks, Alice R., <strong>and</strong> Arthur W. Burks. The First Electronic Computer:<br />

The Atanasoff Story. Ann Arbor: University of Michigan Press,<br />

1990.<br />

McCarney, Scott. ENIAC: The Triumphs <strong>and</strong> Tragedies of the World’s<br />

First Computer. New York: Berkley Books, 2001.<br />

Slater, Robert. Portraits in Silicon. Cambridge, Mass.: MIT Press,<br />

1989.<br />

Stern, Nancy B. From ENIAC to UNIVAC: An Appraisal of the Eckert-<br />

Mauchly Computers. Bedford, Mass.: Digital Press, 1981.


316<br />

Fax machine<br />

Fax machine<br />

The invention: Originally known as the “facsimile machine,” a<br />

machine that converts written <strong>and</strong> printed images into electrical<br />

signals that can be sent via telephone, computer, or radio.<br />

The person behind the invention:<br />

Alex<strong>and</strong>er Bain (1818-1903), a Scottish inventor<br />

Sending Images<br />

The invention of the telegraph <strong>and</strong> telephone during the latter<br />

half of the nineteenth century gave people the ability to send information<br />

quickly over long distances. With the invention of radio <strong>and</strong><br />

television technologies, voices <strong>and</strong> moving pictures could be seen<br />

around the world as well. Oddly, however, the facsimile process—<br />

which involves the transmission of pictures, documents, or other<br />

physical data over distance—predates all these modern devices,<br />

since a simple facsimile apparatus (usually called a fax machine)<br />

was patented in 1843 by Alex<strong>and</strong>er Bain. This early device used a<br />

pendulum to synchronize the transmitting <strong>and</strong> receiving units; it<br />

did not convert the image into an electrical format, however, <strong>and</strong> it<br />

was quite crude <strong>and</strong> impractical. Nevertheless, it reflected the desire<br />

to send images over long distances, which remained a technological<br />

goal for more than a century.<br />

Facsimile machines developed in the period around 1930 enabled<br />

news services to provide newspapers around the world with<br />

pictures for publication. It was not until the 1970’s, however, that<br />

technological advances made small fax machines available for everyday<br />

office use.<br />

Scanning Images<br />

Both the fax machines of the 1930’s <strong>and</strong> those of today operate on<br />

the basis of the same principle: scanning. In early machines, an image<br />

(a document or a picture) was attached to a roller, placed in the<br />

fax machine, <strong>and</strong> rotated at a slow <strong>and</strong> fixed speed (which must be


Fax machine / 317<br />

the same at each end of the link) in a bright light. Light from the image<br />

was reflected from the document in varying degrees, since dark<br />

areas reflect less light than lighter areas do. A lens moved across the<br />

page one line at a time, concentrating <strong>and</strong> directing the reflected<br />

light to a photoelectric tube. This tube would respond to the change<br />

in light level by varying its electric output, thus converting the image<br />

into an output signal whose intensity varied with the changing<br />

light <strong>and</strong> dark spots of the image. Much like the signal from a microphone<br />

or television camera, this modulated (varying) wave could<br />

then be broadcast by radio or sent over telephone lines to a receiver<br />

that performed a reverse function. At the receiving end, a light bulb<br />

was made to vary its intensity to match the varying intensity of the<br />

incoming signal. The output of the light bulb was concentrated<br />

through a lens onto photographically sensitive paper, thus re-creating<br />

the original image as the paper was rotated.<br />

Early fax machines were bulky <strong>and</strong> often difficult to operate.<br />

Advances in semiconductor <strong>and</strong> computer technology in the 1970’s,<br />

however, made the goal of creating an easy-to-use <strong>and</strong> inexpensive<br />

fax machine realistic. Instead of a photoelectric tube that consumes<br />

a relatively large amount of electrical power, a row of small photodiode<br />

semiconductors is used to measure light intensity. Instead of a<br />

power-consuming light source, low-power light-emitting diodes<br />

(LEDs) are used. Some 1,728 light-sensitive diodes are placed in a<br />

row, <strong>and</strong> the image to be scanned is passed over them one line at a<br />

time. Each diode registers either a dark or a light portion of the image.<br />

As each diode is checked in sequence, it produces a signal for<br />

one picture element, also known as a “pixel” or “pel.” Because<br />

many diodes are used, there is no need for a focusing lens; the diode<br />

bar is as wide as the page being scanned, <strong>and</strong> each pixel represents a<br />

portion of a line on that page.<br />

Since most fax transmissions take place over public telephone<br />

system lines, the signal from the photodiodes is transmitted by<br />

means of a built-in computer modem in much the same format that<br />

computers use to transmit data over telephone lines. The receiving<br />

fax uses its modem to convert the audible signal into a sequence that<br />

varies in intensity in proportion to the original signal. This varying<br />

signal is then sent in proper sequence to a row of 1,728 small wires<br />

over which a chemically treated paper is passed. As each wire re-


318 / Fax machine<br />

ceives a signal that represents a black portion of the scanned image,<br />

the wire heats <strong>and</strong>, in contact with the paper, produces a black dot<br />

that corresponds to the transmitted pixel. As the page is passed over<br />

these wires one line at a time, the original image is re-created.<br />

Consequences<br />

The fax machine has long been in use in many commercial <strong>and</strong><br />

scientific fields. Weather data in the form of pictures are transmitted<br />

from orbiting satellites to ground stations; newspapers receive photographs<br />

from international news sources via fax; <strong>and</strong>, using a very<br />

expensive but very high-quality fax device, newspapers <strong>and</strong> magazines<br />

are able to transmit full-size proof copies of each edition to<br />

printers thous<strong>and</strong>s of miles away so that a publication edited in one<br />

country can reach newsst<strong>and</strong>s around the world quickly.<br />

With the technological advances that have been made in recent<br />

years, however, fax transmission has become a part of everyday life,<br />

particularly in business <strong>and</strong> research environments. The ability to<br />

send quickly a copy of a letter, document, or report over thous<strong>and</strong>s<br />

of miles means that information can be shared in a matter of minutes<br />

rather than in a matter of days. In fields such as advertising <strong>and</strong><br />

architecture, it is often necessary to send pictures or drawings to remote<br />

sites. Indeed, the fax machine has played an important role in<br />

providing information to distant observers of political unrest when<br />

other sources of information (such as radio, television, <strong>and</strong> newspapers)<br />

are shut down.<br />

In fact, there has been a natural coupling of computers, modems,<br />

<strong>and</strong> fax devices. Since modern faxes are sent as computer data over<br />

phone lines, specialized <strong>and</strong> inexpensive modems (which allow<br />

two computers to share data) have been developed that allow any<br />

computer user to send <strong>and</strong> receive faxes without bulky machines.<br />

For example, a document—including drawings, pictures, or graphics<br />

of some kind—is created in a computer <strong>and</strong> transmitted directly<br />

to another fax machine. That computer can also receive a fax transmission<br />

<strong>and</strong> either display it on the computer’s screen or print it on<br />

the local printer. Since fax technology is now within the reach of almost<br />

anyone who is interested in using it, there is little doubt that it<br />

will continue to grow in popularity.


See also Communications satellite; Instant photography; Internet;<br />

Personal computer; Xerography.<br />

Further Reading<br />

Fax machine / 319<br />

Bain, Alex<strong>and</strong>er, <strong>and</strong> Leslie William Davidson. Autobiography. New<br />

York: Longmans, Green, 1973.<br />

Cullen, Scott. “Telecommunications in the Office.” Office Systems 16,<br />

no. 12 (December, 1999).<br />

Holtzmann, Gerald J. “Just the Fax.” Inc. 20, no. 13 (September 15,<br />

1998).<br />

Hunkin, Tim. “Just Give Me the Fax.” New Scientist 137, no. 1860<br />

(February 13, 1993).


320<br />

Fiber-optics<br />

Fiber-optics<br />

The invention: The application of glass fibers to electronic communications<br />

<strong>and</strong> other fields to carry large volumes of information<br />

quickly, smoothly, <strong>and</strong> cheaply over great distances.<br />

The people behind the invention:<br />

Samuel F. B. Morse (1791-1872), the American artist <strong>and</strong><br />

inventor who developed the electromagnetic telegraph<br />

system<br />

Alex<strong>and</strong>er Graham Bell (1847-1922), the Scottish American<br />

inventor <strong>and</strong> educator who invented the telephone <strong>and</strong> the<br />

photophone<br />

Theodore H. Maiman (1927- ), the American physicist <strong>and</strong><br />

engineer who invented the solid-state laser<br />

Charles K. Kao (1933- ), a Chinese-born electrical engineer<br />

Zhores I. Alferov (1930- ), a Russian physicist <strong>and</strong><br />

mathematician<br />

The Singing Sun<br />

In 1844, Samuel F. B. Morse, inventor of the telegraph, sent his famous<br />

message, “What hath God wrought?” by electrical impulses<br />

traveling at the speed of light over a 66-kilometer telegraph wire<br />

strung between Washington, D.C., <strong>and</strong> Baltimore. Ever since that<br />

day, scientists have worked to find faster, less expensive, <strong>and</strong> more<br />

efficient ways to convey information over great distances.<br />

At first, the telegraph was used to report stock-market prices <strong>and</strong><br />

the results of political elections. The telegraph was quite important<br />

in the American Civil War (1861-1865). The first transcontinental<br />

telegraph message was sent by Stephen J. Field, chief justice of the<br />

California Supreme Court, to U.S. president Abraham Lincoln on<br />

October 24, 1861. The message declared that California would remain<br />

loyal to the Union. By 1866, telegraph lines had reached all<br />

across the North American continent <strong>and</strong> a telegraph cable had<br />

been laid beneath the Atlantic Ocean to link the Old World with the<br />

New World.


Zhores I. Alferov<br />

Fiber-optics / 321<br />

To create a telephone system that transmitted with light,<br />

perfecting fiber-optic cables was only half the solution. There<br />

also had to be a small, reliable, energy-efficient light source. In<br />

the 1960’s engineers realized that lasers were the best c<strong>and</strong>idate.<br />

However, early gas lasers were bulky, <strong>and</strong> semiconductor<br />

lasers, while small, were temperamental <strong>and</strong> had to be cooled<br />

in liquid nitrogen. Nevertheless, the race was on to devise a<br />

semiconductor laser that produced a continuous beam <strong>and</strong> did<br />

not need to be cooled. The race was between a Bell Labs team in<br />

the United States <strong>and</strong> a Russian team led by Zhores I. Alferov,<br />

neither of which knew much about the other.<br />

Alferov was born in 1930 in Vitebsk, Byelorussia, then part<br />

of the Soviet Union. He earned a degree in electronics from the<br />

V. I. Ulyanov (Lenin) Electrotechnical Institute in Leningrad<br />

(now St. Petersburg). As part of his graduate studies, he became<br />

a researcher at the A. F. Ioffe Physico-Technical Institute in the<br />

same city, receiving a doctorate in physics <strong>and</strong> mathematics in<br />

1970. By then he was one of the world’s leading experts in semiconductor<br />

lasers.<br />

Alferov found that he could improve the laser’s performance<br />

by s<strong>and</strong>wiching very thin layers of gallium arsenide <strong>and</strong><br />

metal, insulated in silicon, in such a way that electrons flowed<br />

only along a 0.03 millimeter strip, producing light in the process.<br />

This double heterojunction narrow-stripe laser was the answer,<br />

producing a steady beam at room temperature. Alferov<br />

published his results a month before the American team came<br />

up with almost precisely the same solution.<br />

The question of who was first was not settled until much<br />

later, during which time both Bell Labs <strong>and</strong> Alferov’s institute<br />

went on to further refinements of the technology. Alferov rose<br />

to become a dean at the St. Petersburg Technical University <strong>and</strong><br />

vice-president of the Russian Academy of Sciences. In 2000 he<br />

shared the Nobel Prize in Physics.<br />

Another American inventor made the leap from the telegraph to<br />

the telephone. Alex<strong>and</strong>er Graham Bell, a teacher of the deaf, was interested<br />

in the physical way speech works. In 1875, he started experimenting<br />

with ways to transmit sound vibrations electrically. He realized<br />

that an electrical current could be adjusted to resemble the


322 / Fiber-optics<br />

vibrations of speech. Bell patented his invention on March 7, 1876.<br />

On July 9, 1877, he founded the Bell Telephone Company.<br />

In 1880, Bell invented a device called the “photophone.” He used<br />

it to demonstrate that speech could be transmitted on a beam of<br />

light. Light is a form of electromagnetic energy. It travels in a vibrating<br />

wave. When the amplitude (height) of the wave is adjusted, a<br />

light beam can be made to carry messages. Bell’s invention included<br />

a thin mirrored disk that converted sound waves directly into a<br />

beam of light. At the receiving end, a selenium resistor connected to<br />

a headphone converted the light back into sound. “I have heard a<br />

ray of sun laugh <strong>and</strong> cough <strong>and</strong> sing,” Bell wrote of his invention.<br />

Although Bell proved that he could transmit speech over distances<br />

of several hundred meters with the photophone, the device<br />

was awkward <strong>and</strong> unreliable, <strong>and</strong> it never became popular as the<br />

telephone did. Not until one hundred years later did researchers find<br />

important practical uses for Bell’s idea of talking on a beam of light.<br />

Two other major discoveries needed to be made first: development<br />

of the laser <strong>and</strong> of high-purity glass. Theodore H. Maiman, an<br />

American physicist <strong>and</strong> electrical engineer at Hughes Research Laboratories<br />

in Malibu, California, built the first laser. The laser produces<br />

an intense, narrowly focused beam of light that can be adjusted to<br />

carry huge amounts of information. The word itself is an acronym for<br />

light amplification by the stimulated emission of radiation.<br />

It soon became clear, though, that even bright laser light can be<br />

broken up <strong>and</strong> absorbed by smog, fog, rain, <strong>and</strong> snow. So in 1966,<br />

Charles K. Kao, an electrical engineer at the St<strong>and</strong>ard Telecommunications<br />

Laboratories in Engl<strong>and</strong>, suggested that glass fibers could<br />

be used to transmit message-carrying beams of laser light without<br />

disruption from weather.<br />

Fiber Optics Are Tested<br />

Optical glass fiber is made from common materials, mostly silica,<br />

soda, <strong>and</strong> lime. The inside of a delicate silica glass tube is coated<br />

with a hundred or more layers of extremely thin glass. The tube is<br />

then heated to 2,000 degrees Celsius <strong>and</strong> collapsed into a thin glass<br />

rod, or preform. The preform is then pulled into thin str<strong>and</strong>s of fiber.<br />

The fibers are coated with plastic to protect them from being nicked<br />

or scratched, <strong>and</strong> then they are covered in flexible cable.


Fiber-optics / 323<br />

The earliest glass fibers<br />

contained many impurities<br />

<strong>and</strong> defects, so they did not<br />

carry light well. Signal repeaters<br />

were needed every<br />

few meters to energize<br />

(amplify) the fading pulses<br />

of light. In 1970, however,<br />

researchers at the Corning<br />

Glass Works in New York<br />

developed a fiber pure<br />

enough to carry light at<br />

Fiber optic str<strong>and</strong>s. (PhotoDisc)<br />

least one kilometer without<br />

amplification.<br />

The telephone industry<br />

quickly became involved in the new fiber-optics technology. Researchers<br />

believed that a bundle of optical fibers as thin as a pencil<br />

could carry several hundred telephone calls at the same time. Optical<br />

fibers were first tested by telephone companies in big cities,<br />

where the great volume of calls often overloaded st<strong>and</strong>ard underground<br />

phone lines.<br />

On May 11, 1977, American Telephone & Telegraph Company<br />

(AT&T), along with Illinois Bell Telephone, Western Electric, <strong>and</strong><br />

Bell Telephone Laboratories, began the first commercial test of fiberoptics<br />

telecommunications in downtown Chicago. The system consisted<br />

of a 2.4-kilometer cable laid beneath city streets. The cable,<br />

only 1.3 centimeters in diameter, linked an office building in the<br />

downtown business district with two telephone exchange centers.<br />

Voice <strong>and</strong> video signals were coded into pulses of laser light <strong>and</strong><br />

transmitted through the hair-thin glass fibers. The tests showed that<br />

a single pair of fibers could carry nearly six hundred telephone conversations<br />

at once very reliably <strong>and</strong> at a reasonable cost.<br />

Six years later, in October, 1983, Bell Laboratories succeeded in<br />

transmitting the equivalent of six thous<strong>and</strong> telephone signals through<br />

an optical fiber cable that was 161 kilometers long. Since that time,<br />

countries all over the world, from Engl<strong>and</strong> to Indonesia, have developed<br />

optical communications systems.


324 / Fiber-optics<br />

Consequences<br />

Fiber optics has had a great impact on telecommunications. A single<br />

fiber can now carry thous<strong>and</strong>s of conversations with no electrical<br />

interference. These fibers are less expensive, weigh less, <strong>and</strong> take up<br />

much less space than copper wire. As a result, people can carry on<br />

conversations over long distances without static <strong>and</strong> at a low cost.<br />

One of the first uses of fiber optics <strong>and</strong> perhaps its best-known<br />

application is the fiberscope, a medical instrument that permits internal<br />

examination of the human body without surgery or X-ray<br />

techniques. The fiberscope, or endoscope, consists of two fiber<br />

bundles. One of the fiber bundles transmits bright light into the patient,<br />

while the other conveys a color image back to the eye of the<br />

physician. The fiberscope has been used to look for ulcers, cancer,<br />

<strong>and</strong> polyps in the stomach, intestine, <strong>and</strong> esophagus of humans.<br />

Medical instruments, such as forceps, can be attached to the fiberscope,<br />

allowing the physician to perform a range of medical procedures,<br />

such as clearing a blocked windpipe or cutting precancerous<br />

polyps from the colon.<br />

See also Cell phone; Community antenna television; Communications<br />

satellite; FM radio; Laser; Long-distance radiotelephony;<br />

Long-distance telephone; Telephone switching.<br />

Further Reading<br />

Carey, John, <strong>and</strong> Neil Gross. “The Light Fantastic: Optoelectronics<br />

May Revolutionize Computers—<strong>and</strong> a Lot More.” Business Week<br />

(May 10, 1993).<br />

Free, John. “Fiber Optics Head for Home.” Popular Science 238<br />

(March, 1991).<br />

Hecht, Jeff. City of Light: The Story of Fiber Optics. Oxford: Oxford<br />

University Press, 1999.<br />

Paul, Noel C. “Laying Down the Line with Huge Projects to Circle<br />

the Globe in Fiber Optic Cable.” Christian Science Monitor (March<br />

29, 2001).<br />

Shinal, John G., with Timothy J. Mullaney. “At the Speed of Light.”<br />

Business Week (October 9, 2000).


Field ion microscope<br />

Field ion microscope<br />

The invention: A microscope that uses ions formed in high-voltage<br />

electric fields to view atoms on metal surfaces.<br />

The people behind the invention:<br />

Erwin Wilhelm Müller (1911-1977), a physicist, engineer, <strong>and</strong><br />

research professor<br />

J. Robert Oppenheimer (1904-1967), an American physicist<br />

To See Beneath the Surface<br />

325<br />

In the early twentieth century, developments in physics, especially<br />

quantum mechanics, paved the way for the application of<br />

new theoretical <strong>and</strong> experimental knowledge to the problem of<br />

viewing the atomic structure of metal surfaces. Of primary importance<br />

were American physicist George Gamow’s 1928 theoretical<br />

explanation of the field emission of electrons by quantum mechanical<br />

means <strong>and</strong> J. Robert Oppenheimer’s 1928 prediction of the<br />

quantum mechanical ionization of hydrogen in a strong electric<br />

field.<br />

In 1936, Erwin Wilhelm Müller developed his field emission microscope,<br />

the first in a series of instruments that would exploit<br />

these developments. It was to be the first instrument to view<br />

atomic structures—although not the individual atoms themselves—<br />

directly. Müller’s subsequent field ion microscope utilized the<br />

same basic concepts used in the field emission microscope yet<br />

proved to be a much more powerful <strong>and</strong> versatile instrument. By<br />

1956, Müller’s invention allowed him to view the crystal lattice<br />

structure of metals in atomic detail; it actually showed the constituent<br />

atoms.<br />

The field emission <strong>and</strong> field ion microscopes make it possible to<br />

view the atomic surface structures of metals on fluorescent screens.<br />

The field ion microscope is the direct descendant of the field emission<br />

microscope. In the case of the field emission microscope, the<br />

images are projected by electrons emitted directly from the tip of a<br />

metal needle, which constitutes the specimen under investigation.


326 / Field ion microscope<br />

These electrons produce an image of the atomic lattice structure of<br />

the needle’s surface. The needle serves as the electron-donating<br />

electrode in a vacuum tube, also known as the “cathode.” A fluorescent<br />

screen that serves as the electron-receiving electrode, or “anode,”<br />

is placed opposite the needle. When sufficient electrical voltage<br />

is applied across the cathode <strong>and</strong> anode, the needle tip emits<br />

electrons, which strike the screen. The image produced on the<br />

screen is a projection of the electron source—the needle surface’s<br />

atomic lattice structure.<br />

Müller studied the effect of needle shape on the performance of<br />

the microscope throughout much of 1937. When the needles had<br />

been properly shaped, Müller was able to realize magnifications of<br />

up to 1 million times. This magnification allowed Müller to view<br />

what he called “maps” of the atomic crystal structure of metals,<br />

since the needles were so small that they were often composed of<br />

only one simple crystal of the material. While the magnification<br />

may have been great, however, the resolution of the instrument was<br />

severely limited by the physics of emitted electrons, which caused<br />

the images Müller obtained to be blurred.<br />

Improving the View<br />

In 1943, while working in Berlin, Müller realized that the resolution<br />

of the field emission microscope was limited by two factors.<br />

The electron velocity, a particle property, was extremely high <strong>and</strong><br />

uncontrollably r<strong>and</strong>om, causing the micrographic images to be<br />

blurred. In addition, the electrons had an unsatisfactorily high wavelength.<br />

When Müller combined these two factors, he was able to determine<br />

that the field emission microscope could never depict single<br />

atoms; it was a physical impossibility for it to distinguish one<br />

atom from another.<br />

By 1951, this limitation led him to develop the technology behind<br />

the field ion microscope. In 1952, Müller moved to the United States<br />

<strong>and</strong> founded the Pennsylvania State University Field Emission Laboratory.<br />

He perfected the field ion microscope between 1952 <strong>and</strong><br />

1956.<br />

The field ion microscope utilized positive ions instead of electrons<br />

to create the atomic surface images on the fluorescent screen.


Erwin Müller<br />

Field ion microscope / 327<br />

Erwin Müller’s scientific goal was to see an individual atom,<br />

<strong>and</strong> to that purpose he invented ever more powerful microscopes.<br />

He was born in Berlin, Germany, in 1911 <strong>and</strong> attended<br />

the city’s Technische Hochschule, earning a diploma in engineering<br />

in 1935 <strong>and</strong> a doctorate in physics in 1936.<br />

Following his studies he worked as an industrial researcher.<br />

Still a neophyte scientist, he discovered the principle of the field<br />

emission microscope <strong>and</strong> was able to produce an image of a<br />

structure only two nanometers in diameter on the surface of a<br />

cathode. In 1941 Müller discovered field desorption by reversing<br />

the polarity of the electron emitter at very low temperatures<br />

so that surface atoms evaporated in the electric field. In 1947 he<br />

left industry <strong>and</strong> began an academic career, teaching physical<br />

chemistry at the Altenburg Engineering School. The following<br />

year he was appointed a department head at the Fritz Haber Institute.<br />

While there, he found that by having a cathode absorb<br />

gas ions <strong>and</strong> then re-emit them he could produce greater magnification.<br />

In 1952 Müller became a professor at Pennsylvania State<br />

University. Applying the new field-ion emission principle, he<br />

was able to achieve his goal, images of individual atoms, in<br />

1956. Almost immediately chemists <strong>and</strong> physicists adopted the<br />

field-ion microscope to conduct basic research concerning the<br />

underlying behavior of field ionization <strong>and</strong> interactions among<br />

absorbed atoms. He further aided such research by coupling a<br />

field-ion microscope <strong>and</strong> mass spectrometer, calling the combination<br />

an atom-probe field-ion microscope; it could both magnify<br />

<strong>and</strong> chemically analyze atoms.<br />

Müller died in 1977. He received the National Medal of Science<br />

posthumously, one of many honors for his contributions to<br />

microscopy.<br />

When an easily ionized gas—at first hydrogen, but usually helium,<br />

neon, or argon—was introduced into the evacuated tube, the emitted<br />

electrons ionized the gas atoms, creating a stream of positively<br />

charged particles, much as Oppenheimer had predicted in 1928.<br />

Müller’s use of positive ions circumvented one of the resolution<br />

problems inherent in the use of imaging electrons. Like the electrons,<br />

however, the positive ions traversed the tube with unpredict-


328 / Field ion microscope<br />

ably r<strong>and</strong>om velocities. Müller eliminated this problem by cryogenically<br />

cooling the needle tip with a supercooled liquefied gas such as<br />

nitrogen or hydrogen.<br />

By 1956, Müller had perfected the means of supplying imaging<br />

positive ions by filling the vacuum tube with an extremely small<br />

quantity of an inert gas such as helium, neon, or argon. By using<br />

such a gas, Müller was assured that no chemical reaction would occur<br />

between the needle tip <strong>and</strong> the gas; any such reaction would alter<br />

the surface atomic structure of the needle <strong>and</strong> thus alter the resulting<br />

microscopic image. The imaging ions allowed the field ion<br />

microscope to image the emitter surface to a resolution of between<br />

two <strong>and</strong> three angstroms, making it ten times more accurate than its<br />

close relative, the field emission microscope.<br />

Consequences<br />

The immediate impact of the field ion microscope was its influence<br />

on the study of metallic surfaces. It is a well-known fact of materials<br />

science that the physical properties of metals are influenced<br />

by the imperfections in their constituent lattice structures. It was not<br />

possible to view the atomic structure of the lattice, <strong>and</strong> thus the finest<br />

detail of any imperfection, until the field ion microscope was developed.<br />

The field ion microscope is the only instrument powerful<br />

enough to view the structural flaws of metal specimens in atomic<br />

detail.<br />

Although the instrument may be extremely powerful, the extremely<br />

large electrical fields required in the imaging process preclude<br />

the instrument’s application to all but the heartiest of metallic<br />

specimens. The field strength of 500 million volts per centimeter<br />

exerts an average stress on metal specimens in the range of almost<br />

1 ton per square millimeter. Metals such as iron <strong>and</strong> platinum can<br />

withst<strong>and</strong> this strain because of the shape of the needles into which<br />

they are formed. Yet this limitation of the instrument makes it extremely<br />

difficult to examine biological materials, which cannot withst<strong>and</strong><br />

the amount of stress that metals can. A practical by-product in<br />

the study of field ionization—field evaporation—eventually permitted<br />

scientists to view large biological molecules.<br />

Field evaporation also allowed surface scientists to view the


atomic structures of biological molecules. By embedding molecules<br />

such as phthalocyanine within the metal needle, scientists have<br />

been able to view the atomic structures of large biological molecules<br />

by field evaporating much of the surrounding metal until the biological<br />

material remains at the needle’s surface.<br />

See also Cyclotron; Electron microscope; Mass spectrograph;<br />

Neutrino detector; Scanning tunneling microscope; Sonar; Synchrocyclotron;<br />

Tevatron accelerator; Ultramicroscope.<br />

Further Reading<br />

Field ion microscope / 329<br />

Gibson, J. M. “Tools for Probing ‘Atomic’ Action.” IEEE Spectrum 22,<br />

no. 12 (December, 1985).<br />

Kunetka, James W. Oppenheimer: The Years of Risk. Englewood Cliffs,<br />

N.J.: Prentice-Hall, 1982.<br />

Schweber, Silvan S. In the Shadow of the Bomb: Bethe, Oppenheimer, <strong>and</strong><br />

the Moral Responsibility of the Scientist. Princeton, N.J.: Princeton<br />

University Press, 2000.<br />

Tsong, Tien Tzou. Atom-Probe Field Ion Microscopy: Field Ion Emission<br />

<strong>and</strong> Surfaces <strong>and</strong> Interfaces at Atomic Resolution. New York: Cambridge<br />

University Press, 1990.


330<br />

Floppy disk<br />

Floppy disk<br />

The invention: Inexpensive magnetic medium for storing <strong>and</strong><br />

moving computer data.<br />

The people behind the invention:<br />

Andrew D. Booth (1918- ), an English inventor who<br />

developed paper disks as a storage medium<br />

Reynold B. Johnson (1906-1998), a design engineer at IBM’s<br />

research facility who oversaw development of magnetic disk<br />

storage devices<br />

Alan Shugart (1930- ), an engineer at IBM’s research<br />

laboratory who first developed the floppy disk as a means of<br />

mass storage for mainframe computers<br />

First Tries<br />

When the International Business Machines (IBM) Corporation<br />

decided to concentrate on the development of computers for business<br />

use in the 1950’s, it faced a problem that had troubled the earliest<br />

computer designers: how to store data reliably <strong>and</strong> inexpensively.<br />

In the early days of computers (the early 1940’s), a number of<br />

ideas were tried. The English inventor Andrew D. Booth produced<br />

spinning paper disks on which he stored data by means of punched<br />

holes, only to ab<strong>and</strong>on the idea because of the insurmountable engineering<br />

problems he foresaw.<br />

The next step was “punched” cards, an idea first used when the<br />

French inventor Joseph-Marie Jacquard invented an automatic weaving<br />

loom for which patterns were stored in pasteboard cards. The<br />

idea was refined by the English mathematician <strong>and</strong> inventor Charles<br />

Babbage for use in his “analytical engine,” an attempt to build a kind<br />

of computing machine. Although it was simple <strong>and</strong> reliable, it was<br />

not fast enough, nor did it store enough data, to be truly practical.<br />

The Ampex Corporation demonstrated its first magnetic audiotape<br />

recorder after World War II (1939-1945). Shortly after that, the<br />

Binary Automatic Computer (BINAC) was introduced with a storage<br />

device that appeared to be a large tape recorder. A more ad-


vanced machine, the Universal Automatic Computer (UNIVAC),<br />

used metal tape instead of plastic (plastic was easily stretched or<br />

even broken). Unfortunately, metal tape was considerably heavier,<br />

<strong>and</strong> its edges were razor-sharp <strong>and</strong> thus dangerous. Improvements<br />

in plastic tape eventually produced sturdy media, <strong>and</strong> magnetic<br />

tape became (<strong>and</strong> remains) a practical medium for storage of computer<br />

data.<br />

Still later designs combined Booth’s spinning paper disks with<br />

magnetic technology to produce rapidly rotating “drums.” Whereas<br />

a tape might have to be fast-forwarded nearly to its end to locate a<br />

specific piece of data, a drum rotating at speeds up to 12,500 revolutions<br />

per minute (rpm) could retrieve data very quickly <strong>and</strong><br />

could store more than 1 million bits (or approximately 125 kilobytes)<br />

of data.<br />

In May, 1955, these drums evolved, under the direction of Reynold<br />

B. Johnson, into IBM’s hard disk unit. The hard disk unit consisted<br />

of fifty platters, each 2 feet in diameter, rotating at 1,200 rpm. Both<br />

sides of the disk could be used to store information. When the operator<br />

wished to access the disk, at his or her comm<strong>and</strong> a read/write<br />

head was moved to the right disk <strong>and</strong> to the side of the disk that<br />

held the desired data. The operator could then read data from or record<br />

data onto the disk. To speed things even more, the next version<br />

of the device, similar in design, employed one hundred read/write<br />

heads—one for each of its fifty double-sided disks. The only remaining<br />

disadvantage was its size, which earned IBM’s first commercial<br />

unit the nickname “jukebox.”<br />

The First Floppy<br />

Floppy disk / 331<br />

The floppy disk drive developed directly from hard disk technology.<br />

It did not take shape until the late 1960’s under the direction of<br />

Alan Shugart (it was announced by IBM as a ready product in 1970).<br />

First created to help restart the operating systems of mainframe<br />

computers that had gone dead, the floppy seemed in some ways to<br />

be a step back, for it operated more slowly than a hard disk drive<br />

<strong>and</strong> did not store as much data. Initially, it consisted of a single thin<br />

plastic disk eight inches in diameter <strong>and</strong> was developed without the<br />

protective envelope in which it is now universally encased. The ad-


332 / Floppy disk<br />

dition of that jacket gave the floppy its single greatest advantage<br />

over the hard disk: portability with reliability.<br />

Another advantage soon became apparent: The floppy is resilient<br />

to damage. In a hard disk drive, the read/write heads must<br />

hover thous<strong>and</strong>ths of a centimeter over the disk surface in order to<br />

attain maximum performance. Should even a small particle of dust<br />

get in the way, or should the drive unit be bumped too hard, the<br />

head may “crash” into the surface of the disk <strong>and</strong> ruin its magnetic<br />

coating; the result is a permanent loss of data. Because the floppy<br />

operates with the read-write head in contact with the flexible plastic<br />

disk surface, individual particles of dust or other contaminants are<br />

not nearly as likely to cause disaster.<br />

As a result of its advantages, the floppy disk was the logical<br />

choice for mass storage in personal computers (PCs), which were<br />

developed a few years after the floppy disk’s introduction. The<br />

floppy is still an important storage device even though hard disk<br />

drives for PCs have become less expensive. Moreover, manufacturers<br />

continually are developing new floppy formats <strong>and</strong> new floppy<br />

disks that can hold more data.<br />

Three-<strong>and</strong>-one-half-inch disks improved on the design of earlier floppies by protecting their<br />

magnetic media within hard plastic shells <strong>and</strong> using sliding metal flanges to protect the surfaces<br />

on which recording heads make contact. (PhotoDisc)


Consequences<br />

Floppy disk / 333<br />

Personal computing would have developed very differently were<br />

it not for the availability of inexpensive floppy disk drives. When<br />

IBM introduced its PC in 1981, the machine provided as st<strong>and</strong>ard<br />

equipment a connection for a cassette tape recorder as a storage device;<br />

a floppy disk was only an option (though an option few did not<br />

take). The awkwardness of tape drives—their slow speed <strong>and</strong> sequential<br />

nature of storing data—presented clear obstacles to the acceptance<br />

of the personal computer as a basic information tool. By<br />

contrast, the floppy drive gives computer users relatively fast storage<br />

at low cost.<br />

Floppy disks provided more than merely economical data storage.<br />

Since they are built to be removable (unlike hard drives), they<br />

represented a basic means of transferring data between machines.<br />

Indeed, prior to the popularization of local area networks (LANs),<br />

the floppy was known as a “sneaker” network: One merely carried<br />

the disk by foot to another computer.<br />

Floppy disks were long the primary means of distributing new<br />

software to users. Even the very flexible floppy showed itself to be<br />

quite resilient to the wear <strong>and</strong> tear of postal delivery. Later, the 3.5inch<br />

disk improved upon the design of the original 8-inch <strong>and</strong> 5.25inch<br />

floppies by protecting the disk medium within a hard plastic<br />

shell <strong>and</strong> by using a sliding metal door to protect the area where the<br />

read/write heads contact the disk.<br />

By the late 1990’s, floppy disks were giving way to new datastorage<br />

media, particularly CD-ROMs—durable laser-encoded disks<br />

that hold more than 700 megabytes of data. As the price of blank<br />

CDs dropped dramatically, floppy disks tended to be used mainly<br />

for short-term storage of small amounts of data. Floppy disks were<br />

also being used less <strong>and</strong> less for data distribution <strong>and</strong> transfer, as<br />

computer users turned increasingly to sending files via e-mail on<br />

the Internet, <strong>and</strong> software providers made their products available<br />

for downloading on Web sites.<br />

See also Bubble memory; Compact disc; Computer chips; Hard<br />

disk; Optical disk; Personal computer.


334 / Floppy disk<br />

Further Reading<br />

Br<strong>and</strong>el, Mary. “IBM Fashions the Floppy.” Computerworld 33, no. 23<br />

(June 7, 1999).<br />

Chposky, James, <strong>and</strong> Ted Leonsis. Blue Magic: The People, Power, <strong>and</strong><br />

Politics Behind the IBM Personal Computer. New York: Facts on File,<br />

1988.<br />

Freiberger, Paul, <strong>and</strong> Michael Swaine. Fire in the Valley: The Making of<br />

the Personal Computer. New York: McGraw-Hill, 2000.<br />

Grossman. Wendy. Remembering the Future: Interviews from Personal<br />

Computer World. New York: Springer, 1997.


Fluorescent lighting<br />

Fluorescent lighting<br />

The invention: A form of electrical lighting that uses a glass tube<br />

coated with phosphor that gives off a cool bluish light <strong>and</strong> emits<br />

ultraviolet radiation.<br />

The people behind the invention:<br />

Vincenzo Cascariolo (1571-1624), an Italian alchemist <strong>and</strong><br />

shoemaker<br />

Heinrich Geissler (1814-1879), a German glassblower<br />

Peter Cooper Hewitt (1861-1921), an American electrical<br />

engineer<br />

Celebrating the “Twelve Greatest <strong>Inventors</strong>”<br />

335<br />

On the night of November 23, 1936, more than one thous<strong>and</strong> industrialists,<br />

patent attorneys, <strong>and</strong> scientists assembled in the main<br />

ballroom of the Mayflower Hotel in Washington, D.C., to celebrate<br />

the one hundredth anniversary of the U.S. Patent Office. A transport<br />

liner over the city radioed the names chosen by the Patent Office as<br />

America’s “Twelve Greatest <strong>Inventors</strong>,” <strong>and</strong>, as the distinguished<br />

group strained to hear those names, “the room was flooded for a<br />

moment by the most brilliant light yet used to illuminate a space<br />

that size.”<br />

Thus did The New York Times summarize the commercial introduction<br />

of the fluorescent lamp. The twelve inventors present were<br />

Thomas Alva Edison, Robert Fulton, Charles Goodyear, Charles<br />

Hall, Elias Howe, Cyrus Hall McCormick, Ottmar Mergenthaler,<br />

Samuel F. B. Morse, George Westinghouse, Wilbur Wright, <strong>and</strong> Eli<br />

Whitney. There was, however, no name to bear the honor for inventing<br />

fluorescent lighting. That honor is shared by many who participated<br />

in a very long series of discoveries.<br />

The fluorescent lamp operates as a low-pressure, electric discharge<br />

inside a glass tube that contains a droplet of mercury <strong>and</strong> a<br />

gas, commonly argon. The inside of the glass tube is coated with<br />

fine particles of phosphor. When electricity is applied to the gas, the<br />

mercury gives off a bluish light <strong>and</strong> emits ultraviolet radiation.


336 / Fluorescent lighting<br />

When bathed in the strong ultraviolet radiation emitted by the mercury,<br />

the phosphor fluoresces (emits light).<br />

The setting for the introduction of the fluorescent lamp began at<br />

the beginning of the 1600’s, when Vincenzo Cascariolo, an Italian<br />

shoemaker <strong>and</strong> alchemist, discovered a substance that gave off a<br />

bluish glow in the dark after exposure to strong sunlight. The fluorescent<br />

substance was apparently barium sulfide <strong>and</strong> was so unusual<br />

for that time <strong>and</strong> so valuable that its formulation was kept secret<br />

for a long time. Gradually, however, scholars became aware of<br />

the preparation secrets of the substance <strong>and</strong> studied it <strong>and</strong> other luminescent<br />

materials.<br />

Further studies in fluorescent lighting were made by the German<br />

physicist Johann Wilhelm Ritter. He observed the luminescence of<br />

phosphors that were exposed to various “exciting” lights. In 1801,<br />

he noted that some phosphors shone brightly when illuminated by<br />

light that the eye could not see (ultraviolet light). Ritter thus discovered<br />

the ultraviolet region of the light spectrum. The use of phosphors<br />

to transform ultraviolet light into visible light was an important<br />

step in the continuing development of the fluorescent lamp.<br />

Further studies in fluorescent lighting were made by the German<br />

physicist Johann Wilhelm Ritter. He observed the luminescence of<br />

phosphors that were exposed to various “exciting” lights. In 1801,<br />

he noted that some phosphors shone brightly when illuminated by<br />

light that the eye could not see (ultraviolet light). Ritter thus discovered<br />

the ultraviolet region of the light spectrum. The use of phosphors<br />

to transform ultraviolet light into visible light was an important<br />

step in the continuing development of the fluorescent lamp.<br />

The British mathematician <strong>and</strong> physicist Sir George Gabriel Stokes<br />

studied the phenomenon as well. It was he who, in 1852, termed the<br />

afterglow “fluorescence.”<br />

Geissler Tubes<br />

While these advances were being made, other workers were trying<br />

to produce a practical form of electric light. In 1706, the English<br />

physicist Francis Hauksbee devised an electrostatic generator, which<br />

is used to accelerate charged particles to very high levels of electrical<br />

energy. He then connected the device to a glass “jar,” used a vac-


uum pump to evacuate the jar to a low pressure, <strong>and</strong> tested his<br />

generator. In so doing, Hauksbee obtained the first human-made<br />

electrical glow discharge by “capturing lightning” in a jar.<br />

In 1854, Heinrich Geissler, a glassblower <strong>and</strong> apparatus maker,<br />

opened his shop in Bonn, Germany, to make scientific instruments;<br />

in 1855, he produced a vacuum pump that used liquid mercury as<br />

an evacuation fluid. That same year, Geissler made the first gaseous<br />

conduction lamps while working in collaboration with the German<br />

scientist Julius Plücker. Plücker referred to these lamps as “Geissler<br />

tubes.” Geissler was able to create red light with neon gas filling a<br />

lamp <strong>and</strong> light of nearly all colors by using certain types of gas<br />

within each of the lamps. Thus, both the neon sign business <strong>and</strong> the<br />

science of spectroscopy were born.<br />

Geissler tubes were studied extensively by a variety of workers.<br />

At the beginning of the twentieth century, the practical American<br />

engineer Peter Cooper Hewitt put these studies to use by marketing<br />

the first low-pressure mercury vapor lamps. The lamps were quite<br />

successful, although they required high voltage for operation, emitted<br />

an eerie blue-green, <strong>and</strong> shone dimly by comparison with their<br />

eventual successor, the fluorescent lamp. At about the same time,<br />

systematic studies of phosphors had finally begun.<br />

By the 1920’s, a number of investigators had discovered that the<br />

low-pressure mercury vapor discharge marketed by Hewitt was an<br />

extremely efficient method for producing ultraviolet light, if the<br />

mercury <strong>and</strong> rare gas pressures were properly adjusted. With a<br />

phosphor to convert the ultraviolet light back to visible light, the<br />

Hewitt lamp made an excellent light source.<br />

Impact<br />

Fluorescent lighting / 337<br />

The introduction of fluorescent lighting in 1936 presented the<br />

public with a completely new form of lighting that had enormous<br />

advantages of high efficiency, long life, <strong>and</strong> relatively low cost.<br />

By 1938, production of fluorescent lamps was well under way. By<br />

April, 1938, four sizes of fluorescent lamps in various colors had<br />

been offered to the public <strong>and</strong> more than two hundred thous<strong>and</strong><br />

lamps had been sold.<br />

During 1939 <strong>and</strong> 1940, two great expositions—the New York


338 / Fluorescent lighting<br />

World’s Fair <strong>and</strong> the San Francisco International Exposition—<br />

helped popularize fluorescent lighting. Thous<strong>and</strong>s of tubular fluorescent<br />

lamps formed a great spiral in the “motor display salon,”<br />

the car showroom of the General Motors exhibit at the New York<br />

World’s Fair. Fluorescent lamps lit the Polish Restaurant <strong>and</strong> hung<br />

in vertical clusters on the flagpoles along the Avenue of the Flags at<br />

the fair, while two-meter-long, upright fluorescent tubes illuminated<br />

buildings at the San Francisco International Exposition.<br />

When the United States entered World War II (1939-1945), the<br />

dem<strong>and</strong> for efficient factory lighting soared. In 1941, more than<br />

twenty-one million fluorescent lamps were sold. Technical advances<br />

continued to improve the fluorescent lamp. By the 1990’s,<br />

this type of lamp supplied most of the world’s artificial lighting.<br />

See also Electric clock; Electric refrigerator; Microwave cooking;<br />

Television; Tungsten filament; Vacuum cleaner; Washing machine.<br />

Further Reading<br />

Bowers, B. “New Lamps for Old: The Story of Electric Lighting.” IEE<br />

Review 41, no. 6 (November 16, 1995).<br />

Dake, Henry Carl, <strong>and</strong> Jack De Ment. Fluorescent Light <strong>and</strong> Its Applications,<br />

Including Location <strong>and</strong> Properties of Fluorescent Materials.<br />

Brooklyn, N.Y.: Chemical Publishing, 1941.<br />

“EPA Sees the Light on Fluorescent Bulbs.” Environmental Health<br />

Perspectives 107, no. 12 (December, 1999).<br />

Harris, J. B. “Electric Lamps, Past <strong>and</strong> Present.” Engineering Science<br />

<strong>and</strong> Education Journal 2, no. 4 (August, 1993).<br />

“How Fluorescent Lighting Became Smaller.” Consulting-Specifying<br />

Engineer 23, no. 2 (February, 1998).


FM radio<br />

FM radio<br />

The invention: A method of broadcasting radio signals by modulating<br />

the frequency, rather than the amplitude, of radio waves,<br />

FM radio greatly improved the quality of sound transmission.<br />

The people behind the invention:<br />

Edwin H. Armstrong (1890-1954), the inventor of FM radio<br />

broadcasting<br />

David Sarnoff (1891-1971), the founder of RCA<br />

An Entirely New System<br />

339<br />

Because early radio broadcasts used amplitude modulation (AM)<br />

to transmit their sounds, they were subject to a sizable amount of interference<br />

<strong>and</strong> static. Since good AM reception relies on the amount<br />

of energy transmitted, energy sources in the atmosphere between<br />

the station <strong>and</strong> the receiver can distort or weaken the original signal.<br />

This is particularly irritating for the transmission of music.<br />

Edwin H. Armstrong provided a solution to this technological<br />

constraint. A graduate of Columbia University, Armstrong made a<br />

significant contribution to the development of radio with his basic<br />

inventions for circuits for AM receivers. (Indeed, the monies Armstrong<br />

received from his earlier inventions financed the development<br />

of the frequency modulation, or FM, system.) Armstrong was<br />

one among many contributors to AM radio. For FM broadcasting,<br />

however, Armstrong must be ranked as the most important inventor.<br />

During the 1920’s, Armstrong established his own research laboratory<br />

in Alpine, New Jersey, across the Hudson River from New<br />

York City. With a small staff of dedicated assistants, he carried out<br />

research on radio circuitry <strong>and</strong> systems for nearly three decades. At<br />

that time, Armstrong also began to teach electrical engineering at<br />

Columbia University.<br />

From 1928 to 1933, Armstrong worked diligently at his private<br />

laboratory at Columbia University to construct a working model of<br />

an FM radio broadcasting system. With the primitive limitations<br />

then imposed on the state of vacuum tube technology, a number of


340 / FM radio<br />

Armstrong’s experimental circuits required as many as one hundred<br />

tubes. Between July, 1930, <strong>and</strong> January, 1933, Armstrong filed<br />

four basic FM patent applications. All were granted simultaneously<br />

on December 26, 1933.<br />

Armstrong sought to perfect FM radio broadcasting, not to offer<br />

radio listeners better musical reception but to create an entirely<br />

new radio broadcasting system. On November 5, 1935, Armstrong<br />

made his first public demonstration of FM broadcasting in New<br />

York City to an audience of radio engineers. An amateur station<br />

based in suburban Yonkers, New York, transmitted these first signals.<br />

The scientific world began to consider the advantages <strong>and</strong><br />

disadvantages of Armstrong’s system; other laboratories began to<br />

craft their own FM systems.<br />

Corporate Conniving<br />

Because Armstrong had no desire to become a manufacturer or<br />

broadcaster, he approached David Sarnoff, head of the Radio Corporation<br />

of America (RCA). As the owner of the top manufacturer<br />

of radio sets <strong>and</strong> the top radio broadcasting network, Sarnoff was<br />

interested in all advances of radio technology. Armstrong first demonstrated<br />

FM radio broadcasting for Sarnoff in December, 1933.<br />

This was followed by visits from RCA engineers, who were sufficiently<br />

impressed to recommend to Sarnoff that the company conduct<br />

field tests of the Armstrong system.<br />

In 1934, Armstrong, with the cooperation of RCA, set up a test<br />

transmitter at the top of the Empire State Building, sharing facilities<br />

with the experimental RCA television transmitter. From 1934 through<br />

1935, tests were conducted using the Empire State facility, to mixed<br />

reactions of RCA’s best engineers. AM radio broadcasting already<br />

had a performance record of nearly two decades. The engineers<br />

wondered if this new technology could replace something that had<br />

worked so well.<br />

This less-than-enthusiastic evaluation fueled the skepticism of<br />

RCA lawyers <strong>and</strong> salespeople. RCA had too much invested in the<br />

AM system, both as a leading manufacturer <strong>and</strong> as the dominant<br />

owner of the major radio network of the time, the National Broadcasting<br />

Company (NBC). Sarnoff was in no rush to adopt FM. To


change systems would risk the millions of dollars RCA was making<br />

as America emerged from the Great Depression.<br />

In 1935, Sarnoff advised Armstrong that RCA would cease any<br />

further research <strong>and</strong> development activity in FM radio broadcasting.<br />

(Still, engineers at RCA laboratories continued to work on FM<br />

to protect the corporate patent position.) Sarnoff declared to the<br />

press that his company would push the frontiers of broadcasting by<br />

concentrating on research <strong>and</strong> development of radio with pictures,<br />

that is, television. As a tangible sign, Sarnoff ordered that Armstrong’s<br />

FM radio broadcasting tower be removed from the top of<br />

the Empire State Building.<br />

Armstrong was outraged. By the mid-1930’s, the development of<br />

FM radio broadcasting had become a mission for Armstrong. For<br />

the remainder of his life, Armstrong devoted his considerable talents<br />

to the promotion of FM radio broadcasting.<br />

Impact<br />

FM radio / 341<br />

After the break with Sarnoff, Armstrong proceeded with plans to<br />

develop his own FM operation. Allied with two of RCA’s biggest<br />

manufacturing competitors, Zenith <strong>and</strong> General Electric, Armstrong<br />

pressed ahead. In June of 1936, at a Federal Communications Commission<br />

(FCC) hearing, Armstrong proclaimed that FM broadcasting<br />

was the only static-free, noise-free, <strong>and</strong> uniform system—both<br />

day <strong>and</strong> night—available. He argued, correctly, that AM radio broadcasting<br />

had none of these qualities.<br />

During World War II (1939-1945), Armstrong gave the military<br />

permission to use FM with no compensation. That patriotic gesture<br />

cost Armstrong millions of dollars when the military soon became<br />

all FM. It did, however, exp<strong>and</strong> interest in FM radio broadcasting.<br />

World War II had provided a field test of equipment <strong>and</strong> use.<br />

By the 1970’s, FM radio broadcasting had grown tremendously.<br />

By 1972, one in three radio listeners tuned into an FM station some<br />

time during the day. Advertisers began to use FM radio stations to<br />

reach the young <strong>and</strong> affluent audiences that were turning to FM stations<br />

in greater numbers.<br />

By the late 1970’s, FM radio stations were outnumbering AM stations.<br />

By 1980, nearly half of radio listeners tuned into FM stations


342 / FM radio<br />

on a regular basis. A decade later, FM radio listening accounted for<br />

more than two-thirds of audience time. Armstrong’s predictions<br />

that listeners would prefer the clear, static-free sounds offered by<br />

FM radio broadcasting had come to pass by the mid-1980’s, nearly<br />

fifty years after Armstrong had commenced his struggle to make<br />

FM radio broadcasting a part of commercial radio.<br />

See also Community antenna television; Communications satellite;<br />

Dolby noise reduction; Fiber-optics; Radio; Radio crystal sets;<br />

Television; Transistor radio.<br />

Further Reading<br />

Lewis, Tom. Empire of the Air: The Men Who Made Radio. New York:<br />

HarperPerennial, 1993.<br />

Sobel, Robert. RCA. New York: Stein <strong>and</strong> Day, 1986.<br />

Streissguth, Thomas. Communications: Sending the Message. Minneapolis,<br />

Minn.: Oliver Press, 1997.


Food freezing<br />

Food freezing<br />

The invention: It was long known that low temperatures helped to<br />

protect food against spoiling; the invention that made frozen<br />

food practical was a method of freezing items quickly. Clarence<br />

Birdseye’s quick-freezing technique made possible a revolution<br />

in food preparation, storage, <strong>and</strong> distribution.<br />

The people behind the invention:<br />

Clarence Birdseye (1886-1956), a scientist <strong>and</strong> inventor<br />

Donald K. Tressler (1894-1981), a researcher at Cornell<br />

University<br />

Am<strong>and</strong>a Theodosia Jones (1835-1914), a food-preservation<br />

pioneer<br />

Feeding the Family<br />

343<br />

In 1917, Clarence Birdseye developed a means of quick-freezing<br />

meat, fish, vegetables, <strong>and</strong> fruit without substantially changing<br />

their original taste. His system of freezing was called by Fortune<br />

magazine “one of the most exciting <strong>and</strong> revolutionary ideas in the<br />

history of food.” Birdseye went on to refine <strong>and</strong> perfect his method<br />

<strong>and</strong> to promote the frozen foods industry until it became a commercial<br />

success nationwide.<br />

It was during a trip to Labrador, where he worked as a fur trader,<br />

that Birdseye was inspired by this idea. Birdseye’s new wife <strong>and</strong><br />

five-week-old baby had accompanied him there. In order to keep<br />

his family well fed, he placed barrels of fresh cabbages in salt water<br />

<strong>and</strong> then exposed the vegetables to freezing winds. Successful at<br />

preserving vegetables, he went on to freeze a winter’s supply of<br />

ducks, caribou, <strong>and</strong> rabbit meat.<br />

In the following years, Birdseye experimented with many freezing<br />

techniques. His equipment was crude: an electric fan, ice, <strong>and</strong> salt<br />

water. His earliest experiments were on fish <strong>and</strong> rabbits, which he<br />

froze <strong>and</strong> packed in old c<strong>and</strong>y boxes. By 1924, he had borrowed<br />

money against his life insurance <strong>and</strong> was lucky enough to find three<br />

partners willing to invest in his new General Seafoods Company


344 / Food freezing<br />

(later renamed General Foods), located in Gloucester, Massachusetts.<br />

Although it was Birdseye’s genius that put the principles of<br />

quick-freezing to work, he did not actually invent quick-freezing.<br />

The scientific principles involved had been known for some time.<br />

As early as 1842, a patent for freezing fish had been issued in Engl<strong>and</strong>.<br />

Nevertheless, the commercial exploitation of the freezing<br />

process could not have happened until the end of the 1800’s, when<br />

mechanical refrigeration was invented. Even then, Birdseye had to<br />

overcome major obstacles.<br />

Finding a Niche<br />

By the 1920’s, there still were few mechanical refrigerators in<br />

American homes. It would take years before adequate facilities for<br />

food freezing <strong>and</strong> retail distribution would be established across the<br />

United States. By the late 1930’s, frozen foods had, indeed, found its<br />

role in commerce but still could not compete with canned or fresh<br />

foods. Birdseye had to work tirelessly to promote the industry, writing<br />

<strong>and</strong> delivering numerous lectures <strong>and</strong> articles to advance its<br />

popularity. His efforts were helped by scientific research conducted<br />

at Cornell University by Donald K. Tressler <strong>and</strong> by C. R. Fellers of<br />

what was then Massachusetts State College. Also, during World<br />

War II (1939-1945), more Americans began to accept the idea: Rationing,<br />

combined with a shortage of canned foods, contributed to<br />

the dem<strong>and</strong> for frozen foods. The armed forces made large purchases<br />

of these items as well.<br />

General Foods was the first to use a system of extremely rapid<br />

freezing of perishable foods in packages. Under the Birdseye system,<br />

fresh foods, such as berries or lobster, were packaged snugly in convenient<br />

square containers. Then, the packages were pressed between<br />

refrigerated metal plates under pressure at 50 degrees below zero.<br />

Two types of freezing machines were used. The “double belt” freezer<br />

consisted of two metal belts that moved through a 15-meter freezing<br />

tunnel, while a special salt solution was sprayed on the surfaces of<br />

the belts. This double-belt freezer was used only in permanent installations<br />

<strong>and</strong> was soon replaced by the “multiplate” freezer, which was<br />

portable <strong>and</strong> required only 11.5 square meters of floor space compared<br />

to the double belt’s 152 square meters.


Am<strong>and</strong>a Theodosia Jones<br />

Food freezing / 345<br />

Am<strong>and</strong>a Theodosia Jones (1835-1914) was close to her brother.<br />

When he suddenly died while they were at school <strong>and</strong> she was<br />

left to contact relatives <strong>and</strong> make the necessary arrangements<br />

for his remains, she was devastated. She had a nervous breakdown<br />

at seventeen <strong>and</strong> could not believe he was entirely gone.<br />

She was sure that he remained an active presence in her life, <strong>and</strong><br />

she became a spiritualist <strong>and</strong> medium so that they could talk<br />

during séances.<br />

Jones always claimed she did not come up with the idea for<br />

the vacuum packing method for preserving food, an important<br />

technique before freezing foods became practicable. It was her<br />

brother who gave it to her. She did the actual experimental<br />

work herself, however, <strong>and</strong> with the aid of Leroy C. Cooley got<br />

the first of their seven patents for food processing. In 1873 she<br />

launched The Women’s Canning <strong>and</strong> Preserving Company, <strong>and</strong><br />

it was more than just a company. It was a mission. All the officers,<br />

stockholders, <strong>and</strong> employees were women. “This is a<br />

woman’s industry,” she insisted, <strong>and</strong> ran the company so that it<br />

was a training school for working women.<br />

In the 1880’s, the spirit of invention moved Jones again. Concerned<br />

about the high rate of accidents among oil drillers, she<br />

examined the problem. Simply add a safety valve to pipes to<br />

control the release of the crude oil, she told drillers in Pennsylvania.<br />

The idea had not occurred to them, but they tried it, <strong>and</strong><br />

it so improved safety that Jones won wide praise.<br />

The multiplate freezer also made it possible to apply the technique<br />

of quick-freezing to seasonal crops. People were able to transport<br />

these freezers easily from one harvesting field to another,<br />

where they were used to freeze crops such as peas fresh off the vine.<br />

The h<strong>and</strong>y multiplate freezer consisted of an insulated cabinet<br />

equipped with refrigerated metal plates. Stacked one above the<br />

other, these plates were capable of being opened <strong>and</strong> closed to receive<br />

food products <strong>and</strong> to compress them with evenly distributed<br />

pressure. Each aluminum plate had internal passages through which<br />

ammonia flowed <strong>and</strong> exp<strong>and</strong>ed at a temperature of −3.8 degrees<br />

Celsius, thus causing the foods to freeze.<br />

A major benefit of the new frozen foods was that their taste <strong>and</strong>


346 / Food freezing<br />

vitamin content were not lost. Ordinarily, when food is frozen<br />

slowly, ice crystals form, which slowly rupture food cells, thus altering<br />

the taste of the food. With quick-freezing, however, the food<br />

looks, tastes, <strong>and</strong> smells like fresh food. Quick-freezing also cuts<br />

down on bacteria.<br />

Impact<br />

During the months between one food harvest <strong>and</strong> the next, humankind<br />

requires trillions of pounds of food to survive. In many<br />

parts of the world, an adequate supply of food is available; elsewhere,<br />

much food goes to waste <strong>and</strong> many go hungry. Methods of<br />

food preservation such as those developed by Birdseye have done<br />

much to help those who cannot obtain proper fresh foods. Preserving<br />

perishable foods also means that they will be available in<br />

greater quantity <strong>and</strong> variety all year-round. In all parts of the world,<br />

both tropical <strong>and</strong> arctic delicacies can be eaten in any season of the<br />

year.<br />

With the rise in popularity of frozen “fast” foods, nutritionists<br />

began to study their effect on the human body. Research has shown<br />

that fresh is the most beneficial. In an industrial nation with many<br />

people, the distribution of fresh commodities is, however, difficult.<br />

It may be many decades before scientists know the long-term effects<br />

on generations raised primarily on frozen foods.<br />

See also Electric refrigerator; Freeze-drying; Microwave cooking;<br />

Polystyrene; Refrigerant gas; Tupperware.<br />

Further Reading<br />

Altman, Linda Jacobs. Women <strong>Inventors</strong>. New York: Facts on File,<br />

1997.<br />

Tressler, Donald K. The Memoirs of Donald K. Tressler. Westport,<br />

Conn.: Avi Publishing, 1976.<br />

_____, <strong>and</strong> Clifford F. Evers. The Freezing Preservation of Foods. New<br />

York: Avi Publishing, 1943.


FORTRAN programming<br />

language<br />

FORTRAN programming language<br />

The invention: The first major computer programming language,<br />

FORTRAN supported programming in a mathematical language<br />

that was natural to scientists <strong>and</strong> engineers <strong>and</strong> achieved unsurpassed<br />

success in scientific computation.<br />

The people behind the invention:<br />

John Backus (1924- ), an American software engineer <strong>and</strong><br />

manager<br />

John W. Mauchly (1907-1980), an American physicist <strong>and</strong><br />

engineer<br />

Herman Heine Goldstine (1913- ), a mathematician <strong>and</strong><br />

computer scientist<br />

John von Neumann (1903-1957), a Hungarian American<br />

mathematician <strong>and</strong> physicist<br />

Talking to Machines<br />

347<br />

Formula Translation, or FORTRAN—the first widely accepted<br />

high-level computer language—was completed by John Backus<br />

<strong>and</strong> his coworkers at the International Business Machines (IBM)<br />

Corporation in April, 1957. Designed to support programming<br />

in a mathematical language that was natural to scientists <strong>and</strong> engineers,<br />

FORTRAN achieved unsurpassed success in scientific<br />

computation.<br />

Computer languages are means of specifying the instructions<br />

that a computer should execute <strong>and</strong> the order of those instructions.<br />

Computer languages can be divided into categories of progressively<br />

higher degrees of abstraction. At the lowest level is binary<br />

code, or machine code: Binary digits, or “bits,” specify in<br />

complete detail every instruction that the machine will execute.<br />

This was the only language available in the early days of computers,<br />

when such machines as the ENIAC (Electronic Numerical Integrator<br />

<strong>and</strong> Calculator) required h<strong>and</strong>-operated switches <strong>and</strong><br />

plugboard connections. All higher levels of language are imple-


348 / FORTRAN programming language<br />

mented by having a program translate instructions written in the<br />

higher language into binary machine language (also called “object<br />

code”). High-level languages (also called “programming languages”)<br />

are largely or entirely independent of the underlying<br />

machine structure. FORTRAN was the first language of this type<br />

to win widespread acceptance.<br />

The emergence of machine-independent programming languages<br />

was a gradual process that spanned the first decade of electronic<br />

computation. One of the earliest developments was the invention of<br />

“flowcharts,” or “flow diagrams,” by Herman Heine Goldstine <strong>and</strong><br />

John von Neumann in 1947. Flowcharting became the most influential<br />

software methodology during the first twenty years of<br />

computing.<br />

Short Code was the first language to be implemented that contained<br />

some high-level features, such as the ability to use mathematical<br />

equations. The idea came from John W. Mauchly, <strong>and</strong> it was<br />

implemented on the BINAC (Binary Automatic Computer) in 1949<br />

with an “interpreter”; later, it was carried over to the UNIVAC (Universal<br />

Automatic Computer) I. Interpreters are programs that do<br />

not translate comm<strong>and</strong>s into a series of object-code instructions; instead,<br />

they directly execute (interpret) those comm<strong>and</strong>s. Every time<br />

the interpreter encounters a comm<strong>and</strong>, that comm<strong>and</strong> must be interpreted<br />

again. “Compilers,” however, convert the entire comm<strong>and</strong><br />

into object code before it is executed.<br />

Much early effort went into creating ways to h<strong>and</strong>le commonly<br />

encountered problems—particularly scientific mathematical<br />

calculations. A number of interpretive languages arose to<br />

support these features. As long as such complex operations had<br />

to be performed by software (computer programs), however, scientific<br />

computation would be relatively slow. Therefore, Backus<br />

lobbied successfully for a direct hardware implementation of these<br />

operations on IBM’s new scientific computer, the 704. Backus then<br />

started the Programming Research Group at IBM in order to develop<br />

a compiler that would allow programs to be written in a<br />

mathematically oriented language rather than a machine-oriented<br />

language. In November of 1954, the group defined an initial version<br />

of FORTRAN.


A More <strong>Access</strong>ible Language<br />

Before FORTRAN was developed, a computer had to perform a<br />

whole series of tasks to make certain types of mathematical calculations.<br />

FORTRAN made it possible for the same calculations to be<br />

performed much more easily. In general, FORTRAN supported constructs<br />

with which scientists were already acquainted, such as functions<br />

<strong>and</strong> multidimensional arrays. In defining a powerful notation<br />

that was accessible to scientists <strong>and</strong> engineers, FORTRAN opened<br />

up programming to a much wider community.<br />

Backus’s success in getting the IBM 704’s hardware to support<br />

scientific computation directly, however, posed a major challenge:<br />

Because such computation would be much faster, the object code<br />

produced by FORTRAN would also have to be much faster. The<br />

lower-level compilers preceding FORTRAN produced programs<br />

that were usually five to ten times slower than their h<strong>and</strong>-coded<br />

counterparts; therefore, efficiency became the primary design objective<br />

for Backus. The highly publicized claims for FORTRAN met<br />

with widespread skepticism among programmers. Much of the<br />

team’s efforts, therefore, went into discovering ways to produce the<br />

most efficient object code.<br />

The efficiency of the compiler produced by Backus, combined<br />

with its clarity <strong>and</strong> ease of use, guaranteed the system’s success. By<br />

1959, many IBM 704 users programmed exclusively in FORTRAN.<br />

By 1963, virtually every computer manufacturer either had delivered<br />

or had promised a version of FORTRAN.<br />

Incompatibilities among manufacturers were minimized by the<br />

popularity of IBM’s version of FORTRAN; every company wanted<br />

to be able to support IBM programs on its own equipment. Nevertheless,<br />

there was sufficient interest in obtaining a st<strong>and</strong>ard for<br />

FORTRAN that the American National St<strong>and</strong>ards Institute adopted<br />

a formal st<strong>and</strong>ard for it in 1966. A revised st<strong>and</strong>ard was adopted in<br />

1978, yielding FORTRAN 77.<br />

Consequences<br />

FORTRAN programming language / 349<br />

In demonstrating the feasibility of efficient high-level languages,<br />

FORTRAN inaugurated a period of great proliferation of program-


350 / FORTRAN programming language<br />

ming languages. Most of these languages attempted to provide similar<br />

or better high-level programming constructs oriented toward a<br />

different, nonscientific programming environment. COBOL, for example,<br />

st<strong>and</strong>s for “Common Business Oriented Language.”<br />

FORTRAN, while remaining the dominant language for scientific<br />

programming, has not found general acceptance among nonscientists.<br />

An IBM project established in 1963 to extend FORTRAN<br />

found the task too unwieldy <strong>and</strong> instead ended up producing an entirely<br />

different language, PL/I, which was delivered in 1966. In the<br />

beginning, Backus <strong>and</strong> his coworkers believed that their revolutionary<br />

language would virtually eliminate the burdens of coding <strong>and</strong><br />

debugging. Instead, FORTRAN launched software as a field of<br />

study <strong>and</strong> an industry in its own right.<br />

In addition to stimulating the introduction of new languages,<br />

FORTRAN encouraged the development of operating systems. Programming<br />

languages had already grown into simple operating systems<br />

called “monitors.” Operating systems since then have been<br />

greatly improved so that they support, for example, simultaneously<br />

active programs (multiprogramming) <strong>and</strong> the networking (combining)<br />

of multiple computers.<br />

See also BASIC programming language; COBOL computer language;<br />

SAINT.<br />

Further Reading<br />

Goff, Leslie. “Born of Frustration.” Computerworld 33, no. 6 (February<br />

8, 1999).<br />

Moreau, René. The Computer Comes of Age: The People, the Hardware,<br />

<strong>and</strong> the Software. Cambridge, Mass.: MIT Press, 1984.<br />

Slater, Robert. Portraits in Silicon. Cambridge, Mass.: MIT Press,<br />

1987.<br />

Stern, Nancy B. From ENIAC to UNIVAC: An Appraisal of the Eckert-<br />

Mauchly Computers. Bedford, Mass.: Digital Press., 1981.


Freeze-drying<br />

Freeze-drying<br />

The invention: Method for preserving foods <strong>and</strong> other organic<br />

matter by freezing them <strong>and</strong> using a vacuum to remove their<br />

water content without damaging their solid matter.<br />

The people behind the invention:<br />

Earl W. Flosdorf (1904- ), an American physician<br />

Ronald I. N. Greaves (1908- ), an English pathologist<br />

Jacques Arsène d’Arsonval (1851-1940), a French physicist<br />

Freeze-Drying for Preservation<br />

351<br />

Drying, or desiccation, is known to preserve biomaterials, including<br />

foods. In freeze-drying, water is evaporated in a frozen<br />

state in a vacuum, by means of sublimation (the process of changing<br />

a solid to a vapor without first changing it to a liquid).<br />

In 1811, John Leslie had first caused freezing by means of the<br />

evaporation <strong>and</strong> sublimation of ice. In 1813, William Wollaston<br />

demonstrated this process to the Royal Society of London. It does<br />

not seem to have occurred to either Leslie or Wollaston to use sublimation<br />

for drying. That distinction goes to Richard Altmann, a<br />

German histologist, who dried pieces of frozen tissue in 1890.<br />

Later, in 1903, Vansteenberghe freeze-dried the rabies virus. In<br />

1906, Jacques Arsène d’Arsonval removed water at a low temperature<br />

for distillation.<br />

Since water removal is the essence of drying, d’Arsonval is often<br />

credited with the discovery of freeze-drying, but the first clearly recorded<br />

use of sublimation for preservation was by Leon Shackell in<br />

1909. His work was widely recognized, <strong>and</strong> he freeze-dried a variety<br />

of biological materials. The first patent for freeze-drying was issued<br />

to Henri Tival, a French inventor, in 1927. In 1934, William<br />

Elser received patents for a modern freeze-drying apparatus that<br />

supplied heat for sublimation.<br />

In 1933, Earl W. Flosdorf had freeze-dried human blood serum<br />

<strong>and</strong> plasma for clinical use. The subsequent efforts of Flosdorf led to<br />

commercial freeze-drying applications in the United States.


352 / Freeze-drying<br />

Freeze-Drying of Foods<br />

With the freeze-drying technique fairly well established for biological<br />

products, it was a natural extension for Flosdorf to apply the<br />

technique to the drying of foods. As early as 1935, Flosdorf experimented<br />

with the freeze-drying of fruit juices <strong>and</strong> milk. An early British<br />

patent was issued to Franklin Kidd, a British inventor, in 1941 for<br />

the freeze-drying of foods. An experimental program on the freezedrying<br />

of food was also initiated at the Low Temperature Research<br />

Station at Cambridge University in Engl<strong>and</strong>, but until World War II,<br />

freeze-drying was only an occasionally used scientific tool.<br />

It was the desiccation of blood plasma from the frozen state, performed<br />

by the American Red Cross for the U.S. armed forces, that<br />

provided the first spectacular, extensive use of freeze-drying. This<br />

work demonstrated the vast potential of freeze-drying for commercial<br />

applications. In 1949, Flosdorf published the first book on<br />

freeze-drying, which laid the foundation for freeze-drying of foods<br />

<strong>and</strong> remains one of the most important contributions to large-scale<br />

operations in the field. In the book, Flosdorf described the freezedrying<br />

of fruit juices, milk, meats, oysters, clams, fish fillets, coffee<br />

<strong>and</strong> tea extracts, fruits, vegetables, <strong>and</strong> other products. Flosdorf also<br />

devoted an entire chapter to describing the equipment used for both<br />

batch <strong>and</strong> continuous processing, <strong>and</strong> he discussed cost analysis.<br />

The holder of more than fifteen patents covering various aspects of<br />

freeze-drying, Flosdorf dominated the move toward commercialization<br />

in the United States.<br />

Simultaneously, researchers in Engl<strong>and</strong> were developing freezedrying<br />

applications under the leadership of Ronald I. N. Greaves.<br />

The food crisis during World War II had led to the recognition that<br />

dried foods cut the costs of transporting, storing, <strong>and</strong> packaging<br />

foods in times of emergency. Thus, in 1951, the British Ministry of<br />

Food Research was established at Aberdeen, Scotl<strong>and</strong>. Scientists at<br />

Aberdeen developed a vacuum contact plate freeze-dryer that improved<br />

product quality <strong>and</strong> reduced the time required for rehydration<br />

(replacement of the water removed in the freeze-drying<br />

process so that the food can be used).<br />

In 1954, trials of initial freeze-drying, followed by the ordinary<br />

process of vacuum drying, were carried out. The abundance of


membranes within plant <strong>and</strong> animal tissues was a major obstacle to<br />

the movement of water vapor, thus limiting the drying rate. In 1956,<br />

two Canadian scientists developed a new method of improving the<br />

freeze-drying rate for steaks by impaling the steaks on spiked heater<br />

plates. This idea was adapted in 1957 by interposing sheets of exp<strong>and</strong>ed<br />

metal, instead of spikes, between the drying surfaces of the<br />

frozen food <strong>and</strong> the heating platens. Because of the substantially<br />

higher freeze-drying rates that it achieved, the process was called<br />

“accelerated freeze-drying.”<br />

In 1960, Greaves described an ingenious method of freeze-drying<br />

liquids. It involved continuously scraping the dry layer during its<br />

formation. This led to a continuous process for freeze-drying liquids.<br />

During the remainder of the 1960’s, freeze-drying applications<br />

proliferated with the advent of several techniques for controlling<br />

<strong>and</strong> improving the effectiveness of the freeze-drying process.<br />

Impact<br />

Freeze-drying / 353<br />

Flosdorf’s vision <strong>and</strong> ingenuity in applying freeze-drying to<br />

foods has revolutionized food preservation. He was also responsible<br />

for making a laboratory technique a tremendous commercial<br />

success.<br />

Freeze-drying is important because it stops the growth of microorganisms,<br />

inhibits deleterious chemical reactions, <strong>and</strong> facilitates<br />

distribution <strong>and</strong> storage. Freeze-dried foods are easily prepared for<br />

consumption by adding water (rehydration). When freeze-dried<br />

properly, most foods, either raw or cooked, can be rehydrated<br />

quickly to yield products that are equal in quality to their frozen<br />

counterparts. Freeze-dried products retain most of their nutritive<br />

qualities <strong>and</strong> have a long storage life, even at room temperature.<br />

Freeze-drying is not, however, without disadvantages. The major<br />

disadvantage is the high cost of processing. Thus, to this day, the<br />

great potential of freeze-drying has not been fully realized. The drying<br />

of cell-free materials, such as coffee <strong>and</strong> tea extracts, has been extremely<br />

successful, but the obstacles imposed by the cell membranes<br />

in foods such as fruits, vegetables, <strong>and</strong> meats have limited<br />

the application to expensive specialty items such as freeze-dried<br />

soups <strong>and</strong> to foods for armies, campers, <strong>and</strong> astronauts. Future eco-


354 / Freeze-drying<br />

nomic changes may create a situation in which the high cost of<br />

freeze-drying is more than offset by the cost of transportation <strong>and</strong><br />

storage.<br />

See also Electric refrigerator; Food freezing; Polystyrene; Tupperware.<br />

Further Reading<br />

Comello, Vic. “Improvements in Freeze Drying Exp<strong>and</strong> Application<br />

Base.” Research <strong>and</strong> Development 42, no. 5 (May, 2000).<br />

Flosdorf, Earl William. Freeze-Drying: Drying by Sublimation. New<br />

York: Reinhold, 1949.<br />

Noves, Robert. Freeze Drying of Foods <strong>and</strong> Biologicals, 1968. Park<br />

Ridge, N.J.: Noyes Development Corporation, 1968.


Fuel cell<br />

Fuel cell<br />

The invention: An electrochemical cell that directly converts energy<br />

from reactions between oxidants <strong>and</strong> fuels, such as liquid<br />

hydrogen, into electrical energy.<br />

The people behind the invention:<br />

Francis Thomas Bacon (1904-1992), an English engineer<br />

Sir William Robert Grove (1811-1896), an English inventor<br />

Georges Leclanché (1839-1882), a French engineer<br />

Aless<strong>and</strong>ro Volta (1745-1827), an Italian physicist<br />

The Earth’s Resources<br />

Because of the earth’s rapidly increasing population <strong>and</strong> the<br />

dwindling of fossil fuels (natural gas, coal, <strong>and</strong> petroleum), there is<br />

a need to design <strong>and</strong> develop new ways to obtain energy <strong>and</strong> to encourage<br />

its intelligent use. The burning of fossil fuels to create energy<br />

causes a slow buildup of carbon dioxide in the atmosphere,<br />

creating pollution that poses many problems for all forms of life on<br />

this planet. Chemical <strong>and</strong> electrical studies can be combined to create<br />

electrochemical processes that yield clean energy.<br />

Because of their very high rate of efficiency <strong>and</strong> their nonpolluting<br />

nature, fuel cells may provide the solution to the problem of<br />

finding sufficient energy sources for humans. The simple reaction of<br />

hydrogen <strong>and</strong> oxygen to form water in such a cell can provide an<br />

enormous amount of clean (nonpolluting) energy. Moreover, hydrogen<br />

<strong>and</strong> oxygen are readily available.<br />

Studies by Aless<strong>and</strong>ro Volta, Georges Leclanché, <strong>and</strong> William<br />

Grove preceded the work of Bacon in the development of the fuel<br />

cell. Bacon became interested in the idea of a hydrogen-oxygen fuel<br />

cell in about 1932. His original intent was to develop a fuel cell that<br />

could be used in commercial applications.<br />

The Fuel Cell Emerges<br />

355<br />

In 1800, the Italian physicist Aless<strong>and</strong>ro Volta experimented<br />

with solutions of chemicals <strong>and</strong> metals that were able to conduct


356 / Fuel cell<br />

electricity. He found that two pieces of metal <strong>and</strong> such a solution<br />

could be arranged in such a way as to produce an electric current.<br />

His creation was the first electrochemical battery, a device that produced<br />

energy from a chemical reaction. Studies in this area were<br />

continued by various people, <strong>and</strong> in the late nineteenth century,<br />

Georges Leclanché invented the dry cell battery, which is now commonly<br />

used.<br />

The work of William Grove followed that of Leclanché. His first<br />

significant contribution was the Grove cell, an improved form of the<br />

cells described above, which became very popular. Grove experimented<br />

with various forms of batteries <strong>and</strong> eventually invented the<br />

“gas battery,” which was actually the earliest fuel cell. It is worth<br />

noting that his design incorporated separate test tubes of hydrogen<br />

<strong>and</strong> oxygen, which he placed over strips of platinum.<br />

After studying the design of Grove’s fuel cell, Bacon decided<br />

that, for practical purposes, the use of platinum <strong>and</strong> other precious<br />

metals should be avoided. By 1939, he had constructed a cell in<br />

which nickel replaced the platinum used.<br />

The theory behind the fuel cell can be described in the following<br />

way. If a mixture of hydrogen <strong>and</strong> oxygen is ignited, energy is released<br />

in the form of a violent explosion. In a fuel cell, however, the<br />

reaction takes place in a controlled manner. Electrons lost by the hydrogen<br />

gas flow out of the fuel cell <strong>and</strong> return to be taken up by the<br />

oxygen in the cell. The electron flow provides electricity to any device<br />

that is connected to the fuel cell, <strong>and</strong> the water that the fuel cell<br />

produces can be purified <strong>and</strong> used for drinking.<br />

Bacon’s studies were interrupted by World War II. After the war<br />

was over, however, Bacon continued his work. Sir Eric Keightley<br />

Rideal of Cambridge University in Engl<strong>and</strong> supported Bacon’s<br />

studies; later, others followed suit. In January, 1954, Bacon wrote an<br />

article entitled “Research into the Properties of the Hydrogen/ Oxygen<br />

Fuel Cell” for a British journal. He was surprised at the speed<br />

with which news of the article spread throughout the scientific<br />

world, particularly in the United States.<br />

After a series of setbacks, Bacon demonstrated a forty-cell unit<br />

that had increased power. This advance showed that the fuel cell<br />

was not merely an interesting toy; it had the capacity to do useful<br />

work. At this point, the General Electric Company (GE), an Ameri-


can corporation, sent a representative to Engl<strong>and</strong> to offer employment<br />

in the United States to senior members of Bacon’s staff. Three scientists<br />

accepted the offer.<br />

A high point in Bacon’s career was the announcement that the<br />

American Pratt <strong>and</strong> Whitney Aircraft company had obtained an order<br />

to build fuel cells for the Apollo project, which ultimately put<br />

two men on the Moon in 1969. Toward the end of his career in 1978,<br />

Bacon hoped that commercial applications for his fuel cells would<br />

be found.<br />

Impact<br />

H 2<br />

- +<br />

Electrolyte<br />

Porous<br />

Electrodes<br />

Anode Cathode<br />

Parts of a basic fuel cell<br />

Fuel cell / 357<br />

Because they are lighter <strong>and</strong> more efficient than batteries, fuel<br />

cells have proved to be useful in the space program. Beginning with<br />

the Gemini 5 spacecraft, alkaline fuel cells (in which a water solution<br />

of potassium hydroxide, a basic, or alkaline, chemical, is placed)<br />

have been used for more than ten thous<strong>and</strong> hours in space. The fuel<br />

cells used aboard the space shuttle deliver the same amount of power<br />

as batteries weighing ten times as much. On a typical seven-day<br />

mission, the shuttle’s fuel cells consume 680 kilograms (1,500 pounds)<br />

of hydrogen <strong>and</strong> generate 719 liters (190 gallons) of water that can<br />

be used for drinking.<br />

Major technical <strong>and</strong> economic problems must be overcome in order<br />

to design fuel cells for practical applications, but some important<br />

advancements have been made. Afew test vehicles that use fuel<br />

O 2


358 / Fuel cell<br />

Francis Bacon<br />

Born in Billericay, Engl<strong>and</strong>, in 1904, Francis Thomas Bacon<br />

completed secondary school at the prestigious Eton College<br />

<strong>and</strong> then attended Trinity College, Cambridge University. In<br />

1932 he started his long search for a practical fuel cell based<br />

upon the oxygen-hydrogen (Hydrox) reaction with an alkaline<br />

electrolyte <strong>and</strong> inexpensive nickel electrodes. In 1940 the British<br />

Admiralty set him up in full-time experimental work at King’s<br />

College, London, <strong>and</strong> then moved him to the Anti-Submarine<br />

Experimental Establishment because the Royal Navy wanted<br />

fuel cells for their submarines.<br />

After World War II Cambridge University appointed him to<br />

the faculty at the Department of Chemical Engineering, <strong>and</strong> he<br />

worked intensively on his fuel cell research. In 1959 he proved<br />

the worth of his work by producing a fuel cell capable of powering<br />

a small truck. It was not until the 1990’s, however, that fuel<br />

cells were taken seriously as the main power source for automobiles.<br />

In 1998, for instance, Icel<strong>and</strong> enlisted the help of Daimler-<br />

Chrysler, Shell Oil, <strong>and</strong> Norsk Hydro to convert all its transportation<br />

vehicles, including its fishing boats, to fuel cell power,<br />

part of its long-range plans for a completely “hydrogen economy.”<br />

Meanwhile, Bacon had the satisfaction of seeing his invention<br />

become a power source for American space vehicles<br />

<strong>and</strong> stations. He died in 1992 in Cambridge.<br />

cells as a source of power have been constructed. Fuel cells using<br />

hydrogen as a fuel <strong>and</strong> oxygen to burn the fuel have been used in a<br />

van built by General Motors Corporation. Thirty-two fuel cells are<br />

installed below the floorboards, <strong>and</strong> tanks of liquid oxygen are carried<br />

in the back of the van. A power plant built in New York City<br />

contains stacks of hydrogen-oxygen fuel cells, which can be put on<br />

line quickly in response to power needs. The Sanyo Electric Company<br />

has developed an electric car that is partially powered by a<br />

fuel cell.<br />

These tremendous technical advances are the result of the singleminded<br />

dedication of Francis Thomas Bacon, who struggled all of<br />

his life with an experiment he was convinced would be successful.


See also Alkaline storage battery; Breeder reactor; Compressedair-accumulating<br />

power plant; Fluorescent lighting; Geothermal<br />

power; Heat pump; Photoelectric cell; Photovoltaic cell; Solar thermal<br />

engine; Tidal power plant.<br />

Further Reading<br />

Fuel cell / 359<br />

Eisenberg, Anne. “Fuel Cell May Be the Future ‘Battery.’” New York<br />

Times (October 21, 1999).<br />

Hoverstein, Paul. “Century-Old Invention Finding a Niche Today.<br />

USA Today (June 3, 1994).<br />

Kufahl, Pam. “Electric: Lighting Up the Twentieth Century.” Unity<br />

Business 3, no. 7 (June, 2000).<br />

Stobart, Richard. Fuel Cell Technology for Vehicles. Warrendale, Pa.:<br />

Society of Automotive Engineers, 2001.


360<br />

Gas-electric car<br />

Gas-electric car<br />

The invention: A hybrid automobile with both an internal combustion<br />

engine <strong>and</strong> an electric motor.<br />

The people behind the invention:<br />

Victor Wouk (1919- ), an American engineer<br />

Tom Elliott, executive vice president of American Honda Motor<br />

Company<br />

Hiroyuki Yoshino, president <strong>and</strong> chief executive officer of<br />

Honda Motor Company<br />

Fujio Cho, president of Toyota Motor Corporation<br />

Announcing Hybrid Vehicles<br />

At the 2000 North American International Auto Show in Detroit,<br />

not only did the Honda Motor Company show off its new Insight<br />

model, it also announced exp<strong>and</strong>ed use of its new technology. Hiroyuki<br />

Yoshino, president <strong>and</strong> chief executive officer, said that Honda’s integrated<br />

motor assist (IMA) system would be exp<strong>and</strong>ed to other massmarket<br />

models. The system basically fits a small electric motor directly<br />

on a one-liter, three-cylinder internal combustion engine. The two<br />

share the workload of powering the car, but the gasoline engine does<br />

not start up until it is needed. The electric motor is powered by a<br />

nickel-metal hydride (Ni-MH) battery pack, with the IMA system automatically<br />

recharging the energy pack during braking.<br />

Tom Elliott, Honda’s executive vice-president, said the vehicle<br />

was a continuation of the company’s philosophy of making the latest<br />

environmental technology accessible to consumers. The $18,000<br />

Insight was a two-seat sporty car that used many innovations to reduce<br />

its weight <strong>and</strong> improve its performance.<br />

Fujio Cho, president of Toyota, also spoke at the Detroit show,<br />

where his company showed off its new $20,000 hybrid Prius. The Toyota<br />

Prius relied more on the electric motor <strong>and</strong> had more energystorage<br />

capacity than the Insight, but was a four-door, five-seat model.<br />

The Toyota Hybrid System divided the power from its 1.5-liter gasoline<br />

engine <strong>and</strong> directed it to drive the wheels <strong>and</strong> a generator. The


generator alternately powered the motor <strong>and</strong> recharged the batteries.<br />

The electric motor was coupled with the gasoline engine to<br />

power the wheels under normal driving. The gasoline engine supplied<br />

average power needs, with the electric motor helping the<br />

peaks; at low speeds, it was all electric. A variable transmission<br />

seamlessly switched back <strong>and</strong> forth between the gasoline engine<br />

<strong>and</strong> electric motor or applied both of them.<br />

Variations on an Idea<br />

Gas-electric car / 361<br />

Automobiles generally use gasoline or diesel engines for driving,<br />

electric motors that start the main motors, <strong>and</strong> a means of recharging<br />

the batteries that power starter motors <strong>and</strong> other devices. In<br />

solely electric cars, gasoline engines are eliminated entirely, <strong>and</strong> the<br />

batteries that power the vehicles are recharged from stationary<br />

sources. In hybrid cars, the relationship between gasoline engines<br />

<strong>and</strong> electric motors is changed so that electric motors h<strong>and</strong>le some<br />

or all of the driving. This is at the expense of an increased number of<br />

batteries or other energy-storage devices.<br />

Possible in many combinations, “hybrids” couple the low-end<br />

torque <strong>and</strong> regenerative braking potential of electric motors with<br />

the range <strong>and</strong> efficient packaging of gasoline, natural gas, or even<br />

hydrogen fuel power plants. The return is greater energy efficiency<br />

<strong>and</strong> reduced pollution.<br />

With sufficient energy-storage capacity, an electric motor can<br />

actually propel a car from a st<strong>and</strong>ing start to a moving speed. In<br />

hybrid vehicles, the gasoline engines—which are more energyefficient<br />

at higher speeds, then kick in. However, the gasoline engines<br />

in these vehicles are smaller, lighter, <strong>and</strong> more efficient than<br />

ordinary gas engines. Designed for average—not peak—driving<br />

conditions, they reduce air pollution <strong>and</strong> considerably improve<br />

fuel economy.<br />

Batteries in hybrid vehicles are recharged partly by the gas engines<br />

<strong>and</strong> partly by regenerative braking; a third of the energy from<br />

slowing the car is turned into electricity. What has finally made hybrids<br />

feasible at reasonable cost are the new developments in computer<br />

technology, allowing sophisticated controls to coordinate electrical<br />

<strong>and</strong> mechanical power.


362 / Gas-electric car<br />

Victor Wouk<br />

H. Piper, an American engineer, filed the first patent for a<br />

hybrid gas-electric powered car in 1905, <strong>and</strong> from then until<br />

1915 they were popular, although not common, because they<br />

could accelerate faster than plain gas-powered cars. Then the<br />

gas-only models became as swift. Their hybrid cousins fells by<br />

the wayside.<br />

Interest in hybrids revived with the unheard-of gasoline<br />

prices during the 1973 oil crisis. The champion of their comeback—the<br />

father of the modern hybrid electric vehicle (HEV)—<br />

was Victor Wouk. Born in 1919 in New York City, Wouk earned<br />

a math <strong>and</strong> physics degree from Columbia University in 1939<br />

<strong>and</strong> a doctorate in electrical engineering from the California Institute<br />

of Technology in 1942. In 1946 he founded Beta Electric<br />

Corporation, which he led until 1959, when he founded <strong>and</strong><br />

was president of another company, Electronic Energy Conversion<br />

Corporation. After 1970, he became an independent consultant,<br />

hoping to build an HEV that people would prefer to<br />

gas-guzzlers.<br />

With his partner, Charles Rosen, Wouk gutted the engine<br />

compartment of a Buick Skylark <strong>and</strong> installed batteries designed<br />

for police cars, a 20-watt direct-current electric motor,<br />

<strong>and</strong> an RX-2 Mazda rotary engine. Only a test vehicle, it still got<br />

better gas mileage (thirty miles per gallon) than the original<br />

Skylark <strong>and</strong> met the requirements for emissions control set by<br />

the Clean Air Act of 1970, unlike all American automobiles of<br />

the era. Moreover, Wouk designed an HEV that would get fifty<br />

miles per gallon <strong>and</strong> pollute one-eighth as much as gas-powered<br />

automobiles. However, the oil crisis ended, gas prices went<br />

down, <strong>and</strong> consumers <strong>and</strong> the government lost interest. Wouk<br />

continued to publish, lecture, <strong>and</strong> design; still, it was not until<br />

the 1990’s that high gas prices <strong>and</strong> concerns over pollution<br />

made HEV’s attractive yet again.<br />

Wouk holds twelve patents, mostly for speed <strong>and</strong> braking<br />

controls in electric vehicles but also for air conditioning, high<br />

voltage direct-current power sources, <strong>and</strong> life extenders for inc<strong>and</strong>escent<br />

lamps.


One way to describe hybrids is to separate them into two types:<br />

parallel, in which either of the two power plants can propel the vehicle,<br />

<strong>and</strong> series, in which the auxiliary power plant is used to<br />

charge the battery, rather than propel the vehicle.<br />

Honda’s Insight is a simplified parallel hybrid that uses a small<br />

but efficient gasoline engine. The electric motor assists the engine,<br />

providing extra power for acceleration or hill climbing, helps provide<br />

regenerative braking, <strong>and</strong> starts the engine. However, it cannot<br />

run the car by itself.<br />

Toyota’s Prius is a parallel hybrid whose power train allows<br />

some series features. Its engine runs only at an efficient speed <strong>and</strong><br />

load <strong>and</strong> is combined with a unique power splitting device. It allows<br />

the car to operate like a parallel hybrid, motor alone, engine<br />

alone, or both. It can act as a series hybrid with the engine charging<br />

the batteries rather than powering the vehicle. It also provides a<br />

continually variable transmission using a planetary gear set that allows<br />

interaction between the engine, the motor, <strong>and</strong> the differential<br />

which drives the wheels.<br />

Impact<br />

Gas-electric car / 363<br />

In 2001 Honda <strong>and</strong> Toyota marketed gas-electric hybrids that offered<br />

better than 60-mile-per-gallon fuel economy <strong>and</strong> met California’s<br />

stringent st<strong>and</strong>ards for “super ultra-low emissions” vehicles.<br />

Both comparnies achieved these st<strong>and</strong>ards without the inconvenience<br />

of fully electric cars which could go only about a hundred<br />

miles on a single battery charge <strong>and</strong> required such gimmicks as kerosene-powered<br />

heaters. As a result, other manufacturers were beginning<br />

to follow suit. Ford, for example, promised a hybrid sport<br />

utility vehicle (SUV) by 2003. Other automakers, including General<br />

Motors <strong>and</strong> DaimlerChrysler, also have announced development of<br />

alternative fuel <strong>and</strong> low emission vehicles. An example is the ESX3<br />

concept car using a 1.5-liter, direct injection diesel combined with a<br />

electric motor <strong>and</strong> a lithium-ion battery<br />

While American automakers were planning to offer some “full<br />

hybrids”—cars capable of running on battery power alone at low<br />

speeds—they were focusing more enthusiastically on electrically<br />

assisted gasoline engines called “mild hybrids.” Full hybrids typi-


364 / Gas-electric car<br />

cally increase gas mileage by up to 60 percent; mild hybrids by only<br />

10 or 20 percent. The “mild hybrid” approach uses regenerative<br />

braking with electrical systems of a much lower voltage <strong>and</strong> storage<br />

capacity than for full hybrids, a much cheaper approach. But there<br />

still is enough energy available to allow the gasoline engine to turn<br />

off automatically when a vehicle stops <strong>and</strong> turn on instantly when<br />

the accelerator is touched. Because the “mild hybrid” approach<br />

adds only $1000 to $1500 to a vehicle’s price, it is likely to be used in<br />

many models. Full hybrids cost much more, but achieve more benefits.<br />

See also Airplane; Diesel locomotive; Hovercraft; Internal combustion<br />

engine; Supersonic passenger plane; Turbojet.<br />

Further Reading<br />

Morton, Ian. “Honda Insight Hybrid Makes Heavy Use of Light<br />

Metal.” Automotive News 74, no. 5853 (December 20, 1999).<br />

Peters, Eric. “Hybrid Cars: The Hope, Hype, <strong>and</strong> Future.” Consumers’<br />

Research Magazine 83, no. 6 (June, 2000).<br />

Reynolds, Kim. “Burt Rutan Ponders the Hybrid Car.” Road <strong>and</strong><br />

Track 51, no. 11 (July, 2000).<br />

Swoboda, Frank. “’Hybrid’ Cars Draw Waiting List of Buyers.”<br />

Washington Post (May 3, 2001).<br />

Yamaguchi, Jack. “Toyota Prius IC/Electric Hybrid Update.” Automotive<br />

Engineering International 108, no. 12 (December, 2000).


Geiger counter<br />

Geiger counter<br />

The invention: the first electronic device able to detect <strong>and</strong> measure<br />

radioactivity in atomic particles.<br />

The people behind the invention:<br />

Hans Geiger (1882-1945), a German physicist<br />

Ernest Rutherford (1871-1937), a British physicist<br />

Sir John Sealy Edward Townsend (1868-1957), an Irish physicist<br />

Sir William Crookes (1832-1919), an English physicist<br />

Wilhelm Conrad Röntgen (1845-1923), a German physicist<br />

Antoine-Henri Becquerel (1852-1908), a French physicist<br />

Discovering Natural Radiation<br />

365<br />

When radioactivity was discovered <strong>and</strong> first studied, the work<br />

was done with rather simple devices. In the 1870’s, Sir William<br />

Crookes learned how to create a very good vacuum in a glass tube.<br />

He placed electrodes in each end of the tube <strong>and</strong> studied the passage<br />

of electricity through the tube. This simple device became known as<br />

the “Crookes tube.” In 1895, Wilhelm Conrad Röntgen was experimenting<br />

with a Crookes tube. It was known that when electricity<br />

went through a Crookes tube, one end of the glass tube might glow.<br />

Certain mineral salts placed near the tube would also glow. In order<br />

to observe carefully the glowing salts, Röntgen had darkened the<br />

room <strong>and</strong> covered most of the Crookes tube with dark paper. Suddenly,<br />

a flash of light caught his eye. It came from a mineral sample<br />

placed some distance from the tube <strong>and</strong> shielded by the dark paper;<br />

yet when the tube was switched off, the mineral sample went dark.<br />

Experimenting further, Röntgen became convinced that some ray<br />

from the Crookes tube had penetrated the mineral <strong>and</strong> caused it to<br />

glow. Since light rays were blocked by the black paper, he called the<br />

mystery ray an “X ray,” with “X” st<strong>and</strong>ing for unknown.<br />

Antoine-Henri Becquerel heard of the discovery of X rays <strong>and</strong>, in<br />

February, 1886, set out to discover if glowing minerals themselves<br />

emitted X rays. Some minerals, called “phosphorescent,” begin to<br />

glow when activated by sunlight. Becquerel’s experiment involved


366 / Geiger counter<br />

wrapping photographic film in black paper <strong>and</strong> setting various<br />

phosphorescent minerals on top <strong>and</strong> leaving them in the sun. He<br />

soon learned that phosphorescent minerals containing uranium<br />

would expose the film.<br />

A series of cloudy days, however, brought a great surprise. Anxious<br />

to continue his experiments, Becquerel decided to develop film<br />

that had not been exposed to sunlight. He was astonished to discover<br />

that the film was deeply exposed. Some emanations must be<br />

coming from the uranium, he realized, <strong>and</strong> they had nothing to do<br />

with sunlight. Thus, natural radioactivity was discovered by accident<br />

with a simple piece of photographic film.<br />

Rutherford <strong>and</strong> Geiger<br />

Ernest Rutherford joined the world of international physics at<br />

about the same time that radioactivity was discovered. Studying the<br />

“Becquerel rays” emitted by uranium, Rutherford eventually distinguished<br />

three different types of radiation, which he named “alpha,”<br />

“beta,” <strong>and</strong> “gamma” after the first three letters of the Greek alphabet.<br />

He showed that alpha particles, the least penetrating of the three, are<br />

the nuclei of helium atoms (a group of two neutrons <strong>and</strong> a proton<br />

tightly bound together). It was later shown that beta particles are electrons.<br />

Gamma rays, which are far more penetrating than either alpha<br />

or beta particles, were shown to be similar to X rays, but with higher<br />

energies.<br />

Rutherford became director of the associated research laboratory<br />

at Manchester University in 1907. Hans Geiger became an assistant.<br />

At this time, Rutherford was trying to prove that alpha particles<br />

carry a double positive charge. The best way to do this was to measure<br />

the electric charge that a stream of alpha particles would bring<br />

to a target. By dividing that charge by the total number of alpha particles<br />

that fell on the target, one could calculate the charge of a single<br />

alpha particle. The problem lay in counting the particles <strong>and</strong> in<br />

proving that every particle had been counted.<br />

Basing their design upon work done by Sir John Sealy Edward<br />

Townsend, a former colleague of Rutherford, Geiger <strong>and</strong> Rutherford<br />

constructed an electronic counter. It consisted of a long brass<br />

tube sealed at both ends from which most of the air had been


Hans Geiger<br />

Geiger counter / 367<br />

Atomic radiation was the first physical phenomenon that<br />

humans discovered that they could not detect with any of their<br />

five natural senses. Hans Geiger found a way to make radiation<br />

observable.<br />

Born into a family with an academic tradition in 1882, Geiger<br />

became an academician himself. His father was a professor<br />

of linguistics at the University of Erlangen, where Geiger completed<br />

his own doctorate in physics in 1906. One of the world’s<br />

centers for experimental physics at the time was Engl<strong>and</strong>, <strong>and</strong><br />

there Geiger went in 1907. He became an assistant to Ernest<br />

Rutherford at the University of Manchester <strong>and</strong> thereby began<br />

the first of a series of successful collaborations during his career—all<br />

devoted to detecting or explaining types of radiation.<br />

Rutherford had distinguished three types of radiation. In<br />

1908, he <strong>and</strong> Geiger built a device to sense the first alpha particles.<br />

It gave them evidence for Rutherford’s conjecture that the<br />

atom was structured like a miniature solar system. Geiger also<br />

worked closely with Ernest Marsden, James Chadwick, <strong>and</strong><br />

Walter Bothe on aspects of radiation physics.<br />

Geiger’s stay in Engl<strong>and</strong> ended with the outbreak of World<br />

War I in 1914. He returned to Germany <strong>and</strong> served as an artillery<br />

officer. Immediately after the war he took up university<br />

posts again, first in Berlin, then in Kiel, Tubingen, <strong>and</strong> back to<br />

Berlin. With Walther Müller he perfected a compact version of<br />

the radiation detector in 1925, the Geiger- Müller counter. It became<br />

the st<strong>and</strong>ard radiation sensor for scientists thereafter, <strong>and</strong>,<br />

during the rush to locate uranium deposits during the 1950’s,<br />

for prospectors.<br />

Geiger used it to prove the existence of the Compton effect,<br />

which concerned the scattering of X rays, <strong>and</strong> his experiments<br />

further proved beyond doubt that light can take the form of<br />

quanta. He also discovered cosmic-ray showers with his detector.<br />

Geiger remained in German during World War II, although<br />

he vigorously opposed the Nazi party’s treatment of scientists.<br />

He died in Potsdam in 1945, after losing his home <strong>and</strong> possessions<br />

during the Allied occupation of Berlin.


368 / Geiger counter<br />

pumped. A thin wire, insulated from the brass, was suspended<br />

down the middle of the tube. This wire was connected to batteries<br />

producing about thirteen hundred volts <strong>and</strong> to an electrometer, a<br />

device that could measure the voltage of the wire. This voltage<br />

could be increased until a spark jumped between the wire <strong>and</strong> the<br />

tube. If the voltage was turned down a little, the tube was ready to<br />

operate. An alpha particle entering the tube would ionize (knock<br />

some electrons away from) at least a few atoms. These electrons<br />

would be accelerated by the high voltage <strong>and</strong>, in turn, would ionize<br />

more atoms, freeing more electrons. This process would continue<br />

until an avalanche of electrons struck the central wire <strong>and</strong> the electrometer<br />

registered the voltage change. Since the tube was nearly<br />

ready to arc because of the high voltage, every alpha particle, even if<br />

it had very little energy, would initiate a discharge. The most complex<br />

of the early radiation detection devices—the forerunner of the<br />

Geiger counter—had just been developed. The two physicists reported<br />

their findings in February, 1908.<br />

Impact<br />

Their first measurements showed that one gram of radium<br />

emitted 34 thous<strong>and</strong> million alpha particles per second. Soon, the<br />

number was refined to 32.8 thous<strong>and</strong> million per second. Next,<br />

Geiger <strong>and</strong> Rutherford measured the amount of charge emitted<br />

by radium each second. Dividing this number by the previous<br />

number gave them the charge on a single alpha particle. Just as<br />

Rutherford had anticipated, the charge was double that of a hydrogen<br />

ion (a proton). This proved to be the most accurate determination<br />

of the fundamental charge until the American physicist<br />

Robert Andrews Millikan conducted his classic oil-drop experiment<br />

in 1911.<br />

Another fundamental result came from a careful measurement of<br />

the volume of helium emitted by radium each second. Using that<br />

value, other properties of gases, <strong>and</strong> the number of helium nuclei<br />

emitted each second, they were able to calculate Avogadro’s number<br />

more directly <strong>and</strong> accurately than had previously been possible.<br />

(Avogadro’s number enables one to calculate the number of atoms<br />

in a given amount of material.)


The true Geiger counter evolved when Geiger replaced the central<br />

wire of the tube with a needle whose point lay just inside a thin<br />

entrance window. This counter was much more sensitive to alpha<br />

<strong>and</strong> beta particles <strong>and</strong> also to gamma rays. By 1928, with the assistance<br />

of Walther Müller, Geiger made his counter much more efficient,<br />

responsive, durable, <strong>and</strong> portable. There are probably few radiation<br />

facilities in the world that do not have at least one Geiger<br />

counter or one of its compact modern relatives.<br />

See also Carbon dating; Gyrocompass; Radar; Sonar; Richter<br />

scale.<br />

Further Reading<br />

Geiger counter / 369<br />

Campbell, John. Rutherford: Scientist Supreme. Christchurch, New<br />

Zeal<strong>and</strong>: AAS <strong>Public</strong>ations, 1999.<br />

Halacy, D. S. They Gave Their Names to Science. New York: Putnam,<br />

1967.<br />

Krebs, A. T. “Hans Geiger: Fiftieth Anniversary of the <strong>Public</strong>ation of<br />

His Doctoral Thesis, 23 July 1906.” Science 124 (1956).<br />

Weir, Fred. “Muscovites Check Radishes for Radiation; a $50 Personal<br />

Geiger Counter Gives Russians a Sense of Confidence at<br />

the Market.” Christian Science Monitor (November 4, 1999).


370<br />

Genetic “fingerprinting”<br />

Genetic “fingerprinting”<br />

The invention: A technique for using the unique characteristics of<br />

each human being’s DNA to identify individuals, establish connections<br />

among relatives, <strong>and</strong> identify criminals.<br />

The people behind the invention:<br />

Alec Jeffreys (1950- ), an English geneticist<br />

Victoria Wilson (1950- ), an English geneticist<br />

Swee Lay Thein (1951- ), a biochemical geneticist<br />

Microscopic Fingerprints<br />

In 1985, Alec Jeffreys, a geneticist at the University of Leicester in<br />

Engl<strong>and</strong>, developed a method of deoxyribonucleic acid (DNA)<br />

analysis that provides a visual representation of the human genetic<br />

structure. Jeffreys’s discovery had an immediate, revolutionary impact<br />

on problems of human identification, especially the identification<br />

of criminals. Whereas earlier techniques, such as conventional<br />

blood typing, provide evidence that is merely exclusionary (indicating<br />

only whether a suspect could or could not be the perpetrator of a<br />

crime), DNA fingerprinting provides positive identification.<br />

For example, under favorable conditions, the technique can establish<br />

with virtual certainty whether a given individual is a murderer<br />

or rapist. The applications are not limited to forensic science;<br />

DNA fingerprinting can also establish definitive proof of parenthood<br />

(paternity or maternity), <strong>and</strong> it is invaluable in providing<br />

markers for mapping disease-causing genes on chromosomes. In<br />

addition, the technique is utilized by animal geneticists to establish<br />

paternity <strong>and</strong> to detect genetic relatedness between social groups.<br />

DNA fingerprinting (also referred to as “genetic fingerprinting”)<br />

is a sophisticated technique that must be executed carefully to produce<br />

valid results. The technical difficulties arise partly from the<br />

complex nature of DNA. DNA, the genetic material responsible for<br />

heredity in all higher forms of life, is an enormously long, doublestr<strong>and</strong>ed<br />

molecule composed of four different units called “bases.”<br />

The bases on one str<strong>and</strong> of DNA pair with complementary bases on


the other str<strong>and</strong>. A human being contains twenty-three pairs of<br />

chromosomes; one member of each chromosome pair is inherited<br />

from the mother, the other from the father. The order, or sequence, of<br />

bases forms the genetic message, which is called the “genome.” Scientists<br />

did not know the sequence of bases in any sizable stretch of<br />

DNA prior to the 1970’s because they lacked the molecular tools to<br />

split DNA into fragments that could be analyzed. This situation<br />

changed with the advent of biotechnology in the mid-1970’s.<br />

The door to DNA analysis was opened with the discovery of bacterial<br />

enzymes called “DNA restriction enzymes.” A restriction enzyme<br />

binds to DNA whenever it finds a specific short sequence of<br />

base pairs (analogous to a code word), <strong>and</strong> it splits the DNA at a defined<br />

site within that sequence. A single enzyme finds millions of<br />

cutting sites in human DNA, <strong>and</strong> the resulting fragments range in<br />

size from tens of base pairs to hundreds or thous<strong>and</strong>s. The fragments<br />

are exposed to a radioactive DNA probe, which can bind to<br />

specific complementary DNA sequences in the fragments. X-ray<br />

film detects the radioactive pattern. The developed film, called an<br />

“autoradiograph,” shows a pattern of DNA fragments, which is<br />

similar to a bar code <strong>and</strong> can be compared with patterns from<br />

known subjects.<br />

The Presence of Minisatellites<br />

Genetic “fingerprinting” / 371<br />

The uniqueness of a DNA fingerprint depends on the fact that,<br />

with the exception of identical twins, no two human beings have<br />

identical DNA sequences. Of the three billion base pairs in human<br />

DNA, many will differ from one person to another.<br />

In 1985, Jeffreys <strong>and</strong> his coworkers, Victoria Wilson at the University<br />

of Leicester <strong>and</strong> Swee Lay Thein at the John Radcliffe Hospital<br />

in Oxford, discovered a way to produce a DNA fingerprint.<br />

Jeffreys had found previously that human DNA contains many repeated<br />

minisequences called “minisatellites.” Minisatellites consist<br />

of sequences of base pairs repeated in t<strong>and</strong>em, <strong>and</strong> the number of<br />

repeated units varies widely from one individual to another. Every<br />

person, with the exception of identical twins, has a different number<br />

of t<strong>and</strong>em repeats <strong>and</strong>, hence, different lengths of minisatellite<br />

DNA. By using two labeled DNA probes to detect two different


372 / Genetic “fingerprinting”<br />

minisatellite sequences, Jeffreys obtained a unique fragment b<strong>and</strong><br />

pattern that was completely specific for an individual.<br />

The power of the technique derives from the law of chance,<br />

which indicates that the probability (chance) that two or more unrelated<br />

events will occur simultaneously is calculated as the multiplication<br />

product of the two separate probabilities. As Jeffreys discovered,<br />

the likelihood of two unrelated people having completely<br />

identical DNA fingerprints is extremely small—less than one in ten<br />

trillion. Given the population of the world, it is clear that the technique<br />

can distinguish any one person from everyone else. Jeffreys<br />

called his b<strong>and</strong> patterns “DNA fingerprints” because of their ability<br />

to individualize. As he stated in his l<strong>and</strong>mark research paper, published<br />

in the English scientific journal Nature in 1985, probes to<br />

minisatellite regions of human DNA produce “DNA ‘fingerprints’<br />

which are completely specific to an individual (or to his or her identical<br />

twin) <strong>and</strong> can be applied directly to problems of human identification,<br />

including parenthood testing.”<br />

Consequences<br />

In addition to being used in human identification, DNA fingerprinting<br />

has found applications in medical genetics. In the search<br />

for a cause, a diagnostic test for, <strong>and</strong> ultimately the treatment of an<br />

inherited disease, it is necessary to locate the defective gene on a human<br />

chromosome. Gene location is accomplished by a technique<br />

called “linkage analysis,” in which geneticists use marker sections<br />

of DNA as reference points to pinpoint the position of a defective<br />

gene on a chromosome. The minisatellite DNA probes developed<br />

by Jeffreys provide a potent <strong>and</strong> valuable set of markers that are of<br />

great value in locating disease-causing genes. Soon after its discovery,<br />

DNA fingerprinting was used to locate the defective genes responsible<br />

for several diseases, including fetal hemoglobin abnormality<br />

<strong>and</strong> Huntington’s disease.<br />

Genetic fingerprinting also has had a major impact on genetic<br />

studies of higher animals. Because DNA sequences are conserved in<br />

evolution, humans <strong>and</strong> other vertebrates have many sequences in<br />

common. This commonality enabled Jeffreys to use his probes to<br />

human minisatellites to bind to the DNA of many different verte-


ates, ranging from mammals to birds, reptiles, amphibians, <strong>and</strong><br />

fish; this made it possible for him to produce DNA fingerprints of<br />

these vertebrates. In addition, the technique has been used to discern<br />

the mating behavior of birds, to determine paternity in zoo primates,<br />

<strong>and</strong> to detect inbreeding in imperiled wildlife. DNA fingerprinting<br />

can also be applied to animal breeding problems, such as<br />

the identification of stolen animals, the verification of semen samples<br />

for artificial insemination, <strong>and</strong> the determination of pedigree.<br />

The technique is not foolproof, however, <strong>and</strong> results may be far<br />

from ideal. Especially in the area of forensic science, there was a<br />

rush to use the tremendous power of DNA fingerprinting to identify<br />

a purported murderer or rapist, <strong>and</strong> the need for scientific st<strong>and</strong>ards<br />

was often neglected. Some problems arose because forensic<br />

DNA fingerprinting in the United States is generally conducted in<br />

private, unregulated laboratories. In the absence of rigorous scientific<br />

controls, the DNA fingerprint b<strong>and</strong>s of two completely unknown<br />

samples cannot be matched precisely, <strong>and</strong> the results may be<br />

unreliable.<br />

See also Amniocentesis; Artificial chromosome; Cloning; In vitro<br />

plant culture; Rice <strong>and</strong> wheat strains; Synthetic amino acid; Synthetic<br />

DNA; Synthetic RNA.<br />

Further Reading<br />

Genetic “fingerprinting” / 373<br />

Bodmer, Walter, <strong>and</strong> Robin McKie. “Probing the Present.” In The<br />

Book of Man: The Human Genome Project. New York: Scribner, 1985.<br />

Caetano-Anolles, Gustavo, <strong>and</strong> Peter M. Gresshoff. DNA Markers:<br />

Protocols, Applications, <strong>and</strong> Overviews. New York: Wiley-VCH,<br />

1997.<br />

Krawezak, Michael, <strong>and</strong> Jorg Schmidtke. DNA Fingerprinting.2ded.<br />

New York: Springer-Verlag, 1998.<br />

Schacter, Bernice Zeldin. Issues <strong>and</strong> Dilemmas of Biotechnology: A Reference<br />

Guide. Westport, Conn.: Greenwood Press, 1999.


374<br />

Genetically engineered insulin<br />

Genetically engineered insulin<br />

The invention: Artificially manufactured human insulin (Humulin)<br />

as a medication for people suffering from diabetes.<br />

The people behind the invention:<br />

Irving S. Johnson (1925- ), an American zoologist who was<br />

vice president of research at Eli Lilly Research Laboratories<br />

Ronald E. Chance (1934- ), an American biochemist at Eli<br />

Lilly Research Laboratories<br />

What Is Diabetes?<br />

Carbohydrates (sugars <strong>and</strong> related chemicals) are the main food<br />

<strong>and</strong> energy source for humans. In wealthy countries such as the<br />

United States, more than 50 percent of the food people eat is made<br />

up of carbohydrates, while in poorer countries the carbohydrate<br />

content of diets is higher, from 70 to 90 percent.<br />

Normally, most carbohydrates that a person eats are used (or metabolized)<br />

quickly to produce energy. Carbohydrates not needed for<br />

energy are either converted to fat or stored as a glucose polymer<br />

called “glycogen.” Most adult humans carry about a pound of body<br />

glycogen; this substance is broken down to produce energy when it<br />

is needed.<br />

Certain diseases prevent the proper metabolism <strong>and</strong> storage of<br />

carbohydrates. The most common of these diseases is diabetes mellitus,<br />

usually called simply “diabetes.” It is found in more than seventy<br />

million people worldwide. Diabetic people cannot produce or<br />

use enough insulin, a hormone secreted by the pancreas. When their<br />

condition is not treated, the eyes may deteriorate to the point of<br />

blindness. The kidneys may stop working properly, blood vessels<br />

may be damaged, <strong>and</strong> the person may fall into a coma <strong>and</strong> die. In<br />

fact, diabetes is the third most common killer in the United States.<br />

Most of the problems surrounding diabetes are caused by high levels<br />

of glucose in the blood. Cataracts often form in diabetics, as excess<br />

glucose is deposited in the lens of the eye.<br />

Important symptoms of diabetes include constant thirst, exces-


sive urination, <strong>and</strong> large amounts of sugar in the blood <strong>and</strong> in the<br />

urine. The glucose tolerance test (GTT) is the best way to find out<br />

whether a person is suffering from diabetes. People given a GTT are<br />

first told to fast overnight. In the morning their blood glucose level<br />

is measured; then they are asked to drink about a fourth of a pound<br />

of glucose dissolved in water. During the next four to six hours, the<br />

blood glucose level is measured repeatedly. In nondiabetics, glucose<br />

levels do not rise above a certain amount during a GTT, <strong>and</strong> the<br />

level drops quickly as the glucose is assimilated by the body. In diabetics,<br />

the blood glucose levels rise much higher <strong>and</strong> do not drop as<br />

quickly. The extra glucose then shows up in the urine.<br />

Treating Diabetes<br />

Genetically engineered insulin / 375<br />

Until the 1920’s, diabetes could be controlled only through a diet<br />

very low in carbohydrates, <strong>and</strong> this treatment was not always successful.<br />

Then Sir Frederick G. Banting <strong>and</strong> Charles H. Best found a<br />

way to prepare purified insulin from animal pancreases <strong>and</strong> gave it<br />

to patients. This gave diabetics their first chance to live a fairly normal<br />

life. Banting <strong>and</strong> his coworkers won the 1923 Nobel Prize in<br />

Physiology or Medicine for their work.<br />

The usual treatment for diabetics became regular shots of insulin.<br />

Drug companies took the insulin from the pancreases of cattle <strong>and</strong><br />

pigs slaughtered by the meat-packing industry. Unfortunately, animal<br />

insulin has two disadvantages. First, about 5 percent of diabetics<br />

are allergic to it <strong>and</strong> can have severe reactions. Second, the world<br />

supply of animal pancreases goes up <strong>and</strong> down depending on how<br />

much meat is being bought. Between 1970 <strong>and</strong> 1975, the supply of<br />

insulin fell sharply as people began to eat less red meat, yet the<br />

numbers of diabetics continued to increase. So researchers began to<br />

look for a better way to supply insulin.<br />

Studying pancreases of people who had donated their bodies to<br />

science, researchers found that human insulin did not cause allergic<br />

reactions. Scientists realized that it would be best to find a chemical<br />

or biological way to prepare human insulin, <strong>and</strong> pharmaceutical<br />

companies worked hard toward this goal. Eli Lilly <strong>and</strong> Company<br />

was the first to succeed, <strong>and</strong> on May 14, 1982, it filed a new drug application<br />

with the Food <strong>and</strong> Drug Administration (FDA) for the hu-


376 / Genetically engineered insulin<br />

man insulin preparation it named “Humulin.”<br />

Humulin is made by genetic engineering. Irving S. Johnson, who<br />

worked on the development of Humulin, described Eli Lilly’s method<br />

for producing Humulin. The common bacterium Escherichia coli<br />

is used. Two strains of the bacterium are produced by genetic engineering:<br />

The first strain is used to make a protein called an “A<br />

chain,” <strong>and</strong> the second strain is used to make a “B chain.” After the<br />

bacteria are harvested, the A <strong>and</strong> B chains are removed <strong>and</strong> purified<br />

separately. Then the two chains are combined chemically. When<br />

they are purified once more, the result is Humulin, which has been<br />

proved by Ronald E. Chance <strong>and</strong> his Eli Lilly coworkers to be chemically,<br />

biologically, <strong>and</strong> physically identical to human insulin.<br />

Consequences<br />

The FDA <strong>and</strong> other regulatory agencies around the world approved<br />

genetically engineered human insulin in 1982. Humulin<br />

does not trigger allergic reactions, <strong>and</strong> its supply does not fluctuate.<br />

It has brought an end to the fear that there would be a worldwide<br />

shortage of insulin.<br />

Humulin is important as well in being the first genetically engineered<br />

industrial chemical. It began an era in which such advanced<br />

technology could be a source for medical drugs, chemicals used in<br />

farming, <strong>and</strong> other important industrial products. Researchers hope<br />

that genetic engineering will help in the underst<strong>and</strong>ing of cancer<br />

<strong>and</strong> other diseases, <strong>and</strong> that it will lead to ways to grow enough<br />

food for a world whose population continues to rise.<br />

See also Artificial chromosome; Artificial insemination; Cloning;<br />

Genetic “fingerprinting”; Synthetic amino acid; Synthetic DNA;<br />

Synthetic RNA.<br />

Further Reading<br />

Berger, Abi. “Gut Cells Engineered to Produce Insulin.” British Medical<br />

Journal 321, no. 7275 (December 16, 2000).<br />

“Genetically Engineered Duckweed to Produce Insulin.” Resource 6,<br />

no. 3 (March, 1999).


Genetically engineered insulin / 377<br />

“Lilly Gets FDA Approval for New Insulin Formula.” Wall Street<br />

Journal (October 3, 1985).<br />

Williams, Linda. “UC Regents Sue Lilly in Dispute Over Biotech<br />

Patent for Insulin.” Los Angeles Times (February 8, 1990).


378<br />

Geothermal power<br />

Geothermal power<br />

The invention: Energy generated from the earth’s natural hot<br />

springs.<br />

The people behind the invention:<br />

Prince Piero Ginori Conti (1865-1939), an Italian nobleman <strong>and</strong><br />

industrialist<br />

Sir Charles Parsons (1854-1931), an English engineer<br />

B. C. McCabe, an American businessman<br />

Developing a Practical System<br />

The first successful use of geothermal energy was at Larderello in<br />

northern Italy. The Larderello geothermal field, located near the city<br />

of Pisa about 240 kilometers northwest of Rome, contains many hot<br />

springs <strong>and</strong> fumaroles (steam vents). In 1777, these springs were<br />

found to be rich in boron, <strong>and</strong> in 1818, Francesco de Larderel began<br />

extracting the useful mineral borax from them. Shortly after 1900,<br />

Prince Piero Ginori Conti, director of the Larderello borax works,<br />

conceived the idea of using the steam for power production. An experimental<br />

electrical power plant was constructed at Larderello in<br />

1904 to provide electric power to the borax plant. After this initial<br />

experiment proved successful, a 250-kilowatt generating station<br />

was installed in 1913 <strong>and</strong> commercial power production began.<br />

As the Larderello field grew, additional geothermal sites throughout<br />

the region were prospected <strong>and</strong> tapped for power. Power production<br />

grew steadily until the 1940’s, when production reached<br />

130 megawatts; however, the Larderello power plants were destroyed<br />

late in World War II (1939-1945). After the war, the generating<br />

plants were rebuilt, <strong>and</strong> they were producing more than 400<br />

megawatts by 1980.<br />

The Larderello power plants encountered many of the technical<br />

problems that were later to concern other geothermal facilities. For<br />

example, hydrogen sulfide in the steam was highly corrosive to copper,<br />

so the Larderello power plant used aluminum for electrical connections<br />

much more than did conventional power plants of the


time. Also, the low pressure of the steam in early wells at Larderello<br />

presented problems. The first generators simply used steam to drive<br />

a generator <strong>and</strong> vented the spent steam into the atmosphere. A system<br />

of this sort, called a “noncondensing system,” is useful for small<br />

generators but not efficient to produce large amounts of power.<br />

Most steam engines derive power not only from the pressure of<br />

the steam but also from the vacuum created when the steam is condensed<br />

back to water. Geothermal systems that generate power<br />

from condensation, as well as direct steam pressure, are called “condensing<br />

systems.” Most large geothermal generators are of this<br />

type. Condensation of geothermal steam presents special problems<br />

not present in ordinary steam engines: There are other gases present<br />

that do not condense. Instead of a vacuum, condensation of steam<br />

contaminated with other gases would result in only a limited drop<br />

in pressure <strong>and</strong>, consequently, very low efficiency.<br />

Initially, the operators of Larderello tried to use the steam to heat<br />

boilers that would, in turn, generate pure steam. Eventually, a device<br />

was developed that removed most of the contaminating gases from<br />

the steam. Although later wells at Larderello <strong>and</strong> other geothermal<br />

fields produced steam at greater pressure, these engineering innovations<br />

improved the efficiency of any geothermal power plant.<br />

Exp<strong>and</strong>ing the Idea<br />

Geothermal power / 379<br />

In 1913, the English engineer Sir Charles Parsons proposed drilling<br />

an extremely deep (12-kilometer) hole to tap the earth’s deep<br />

heat. Power from such a deep hole would not come from natural<br />

steam as at Larderello but would be generated by pumping fluid<br />

into the hole <strong>and</strong> generating steam (as hot as 500 degrees Celsius) at<br />

the bottom. In modern terms, Parsons proposed tapping “hot dryrock”<br />

geothermal energy. (No such plant has been commercially operated<br />

yet, but research is being actively pursued in several countries.)<br />

The first use of geothermal energy in the United States was for direct<br />

heating. In 1890, the municipal water company of Boise, Idaho,<br />

began supplying hot water from a geothermal well. Water was<br />

piped from the well to homes <strong>and</strong> businesses along appropriately<br />

named Warm Springs Avenue. At its peak, the system served more


380 / Geothermal power<br />

than four hundred customers, but as cheap natural gas became<br />

available, the number declined.<br />

Although Larderello was the first successful geothermal electric<br />

power plant, the modern era of geothermal electric power began<br />

with the opening of the Geysers Geothermal Field in California.<br />

Early attempts began in the 1920’s, but it was not until 1955 that B.<br />

C. McCabe, a Los Angeles businessman, leased 14.6 square kilometers<br />

in the Geysers area <strong>and</strong> founded the Magma Power Company.<br />

The first 12.5-megawatt generator was installed at the Geysers in<br />

1960, <strong>and</strong> production increased steadily from then on. The Geysers<br />

surpassed Larderello as the largest producing geothermal field in<br />

the 1970’s, <strong>and</strong> more than 1,000 megawatts were being generated by<br />

1980. By the end of 1980, geothermal plants had been installed in<br />

thirteen countries, with a total capacity of almost 2,600 megawatts,<br />

<strong>and</strong> projects with a total capacity of more than 15,000 megawatts<br />

were being planned in more than twenty countries.<br />

Impact<br />

Geothermal power has many attractive features. Because the<br />

steam is naturally heated <strong>and</strong> under pressure, generating equipment<br />

can be simple, inexpensive, <strong>and</strong> quickly installed. Equipment<br />

<strong>and</strong> installation costs are offset by savings in fuel. It is economically<br />

practical to install small generators, a fact that makes geothermal<br />

plants attractive in remote or underdeveloped areas. Most important<br />

to a world faced with a variety of technical <strong>and</strong> environmental<br />

problems connected with fossil fuels, geothermal power does not<br />

deplete fossil fuel reserves, produces little pollution, <strong>and</strong> contributes<br />

little to the greenhouse effect.<br />

Despite its attractive features, geothermal power has some limitations.<br />

Geologic settings suitable for easy geothermal power production<br />

are rare; there must be a hot rock or magma body close to<br />

the surface. Although it is technically possible to pump water from<br />

an external source into a geothermal well to generate steam, most<br />

geothermal sites require a plentiful supply of natural underground<br />

water that can be tapped as a source of steam. In contrast, fossil-fuel<br />

generating plants can be at any convenient location.


See also Breeder reactor; Compressed-air-accumulating power<br />

plant; Fuel cell; Heat pump; Nuclear power plant; Solar thermal engine;<br />

Thermal cracking process; Tidal power plant.<br />

Further Reading<br />

Geothermal power / 381<br />

Appleyard, Rollo. Charles Parsons: His Life <strong>and</strong> Work. London: Constable,<br />

1933.<br />

Boyle, Godfrey. Renewable Energy: Power for a Sustainable Future. Oxford:<br />

Oxford University Press, 1998.<br />

Cassedy, Edward S. Prospects for Sustainable Energy: A Critical Assessment.<br />

New York: Cambridge University Press, 2000.<br />

Parsons, Robert Hodson. The Steam Turbine <strong>and</strong> Other <strong>Inventions</strong> of<br />

Sir Charles Parsons, O.M. New York: Longmans Green, 1946.


382<br />

Gyrocompass<br />

Gyrocompass<br />

The invention: The first practical navigational device that enabled<br />

ships <strong>and</strong> submarines to stay on course without relying on the<br />

earth’s unreliable magnetic poles.<br />

The people behind the invention:<br />

Hermann Anschütz-Kaempfe (1872-1931), a German inventor<br />

<strong>and</strong> manufacturer<br />

Jean-Bernard-Léon Foucault (1819-1868), a French experimental<br />

physicist <strong>and</strong> inventor<br />

Elmer Ambrose Sperry (1860-1930), an American engineer <strong>and</strong><br />

inventor<br />

From Toys to Tools<br />

A gyroscope consists of a rapidly spinning wheel mounted in a<br />

frame that enables the wheel to tilt freely in any direction. The<br />

amount of momentum allows the wheel to maintain its “attitude”<br />

even when the whole device is turned or rotated.<br />

These devices have been used to solve problems arising in such<br />

areas as sailing <strong>and</strong> navigation. For example, a gyroscope aboard a<br />

ship maintains its orientation even while the ship is rolling. Among<br />

other things, this allows the extent of the roll to be measured accurately.<br />

Moreover, the spin axis of a free gyroscope can be adjusted to<br />

point toward true north. It will (with some exceptions) stay that<br />

way despite changes in the direction of a vehicle in which it is<br />

mounted. Gyroscopic effects were employed in the design of various<br />

objects long before the theory behind them was formally<br />

known. A classic example is a child’s top, which balances, seemingly<br />

in defiance of gravity, as long as it continues to spin. Boomerangs<br />

<strong>and</strong> flying disks derive stability <strong>and</strong> accuracy from the spin<br />

imparted by the thrower. Likewise, the accuracy of rifles improved<br />

when barrels were manufactured with internal spiral grooves that<br />

caused the emerging bullet to spin.<br />

In 1852, the French inventor Jean-Bernard-Léon Foucault built<br />

the first gyroscope, a measuring device consisting of a rapidly spinning<br />

wheel mounted within concentric rings that allowed the wheel


to move freely about two axes. This device, like the Foucault pendulum,<br />

was used to demonstrate the rotation of the earth around its<br />

axis, since the spinning wheel, which is not fixed, retains its orientation<br />

in space while the earth turns under it. The gyroscope had a related<br />

interesting property: As it continued to spin, the force of the<br />

earth’s rotation caused its axis to rotate gradually until it was oriented<br />

parallel to the earth’s axis, that is, in a north-south direction. It<br />

is this property that enables the gyroscope to be used as a compass.<br />

When Magnets Fail<br />

Gyrocompass / 383<br />

In 1904, Hermann Anschütz-Kaempfe, a German manufacturer<br />

working in the Kiel shipyards, became interested in the navigation<br />

problems of submarines used in exploration under the polar ice cap.<br />

By 1905, efficient working submarines were a reality, <strong>and</strong> it was evident<br />

to all major naval powers that submarines would play an increasingly<br />

important role in naval strategy.<br />

Submarine navigation posed problems, however, that could not<br />

be solved by instruments designed for surface vessels. A submarine<br />

needs to orient itself under water in three dimensions; it has no automatic<br />

horizon with respect to which it can level itself. Navigation<br />

by means of stars or l<strong>and</strong>marks is impossible when the submarine is<br />

submerged. Furthermore, in an enclosed metal hull containing machinery<br />

run by electricity, a magnetic compass is worthless. To a<br />

lesser extent, increasing use of metal, massive moving parts, <strong>and</strong><br />

electrical equipment had also rendered the magnetic compass unreliable<br />

in conventional surface battleships.<br />

It made sense for Anschütz-Kaempfe to use the gyroscopic effect<br />

to design an instrument that would enable a ship to maintain its<br />

course while under water. Yet producing such a device would not be<br />

easy. First, it needed to be suspended in such a way that it was free to<br />

turn in any direction with as little mechanical resistance as possible.<br />

At the same time, it had to be able to resist the inevitable pitching <strong>and</strong><br />

rolling of a vessel at sea. Finally, a continuous power supply was required<br />

to keep the gyroscopic wheels spinning at high speed.<br />

The original Anschütz-Kaempfe gyrocompass consisted of a pair<br />

of spinning wheels driven by an electric motor. The device was connected<br />

to a compass card visible to the ship’s navigator. Motor, gyro-


384 / Gyrocompass<br />

Elmer Sperry<br />

Although Elmer Ambrose Sperry, born in 1860, had only a<br />

grade school education as a child in rural New York, the equipment<br />

used on local farms piqued his interest in machinery <strong>and</strong><br />

he learned about technology on his own. He attended a local<br />

teachers’ college, <strong>and</strong> graduating in 1880, he was determined to<br />

become an inventor.<br />

He was especially interested in the application of electricity.<br />

He designed his own arc lighting system <strong>and</strong> opened the Sperry<br />

Electric Light, Motor, <strong>and</strong> Car Brake Company to sell it, changing<br />

its name to Sperry Electric Company in 1887. He made such<br />

progress in devising electric mining equipment, electric brakes<br />

for automobiles <strong>and</strong> streetcars, <strong>and</strong> his own electric car that<br />

General Electric bought him out.<br />

In 1900 Sperry opened a laboratory in Washington, D.C., <strong>and</strong><br />

continued research on a gyroscope that he began in 1896. After<br />

more than a decade he patented his device, <strong>and</strong> after successful<br />

trials aboard the USS Worden, he established the Sperry Gyroscope<br />

Company in 1910, later supplying the American, British,<br />

<strong>and</strong> Russian navies as well as commercial ships. In 1914 he<br />

successfully demonstrated a gyrostabilizer for aircraft <strong>and</strong> exp<strong>and</strong>ed<br />

his company to manufacture aeronautical technology.<br />

Before he sold the company in 1926 he had registered more than<br />

four hundred patents. Sperry died in Brooklyn in 1930.<br />

scope, <strong>and</strong> suspension system were mounted in a frame that allowed<br />

the apparatus to remain stable despite the pitch <strong>and</strong> roll of the ship.<br />

In 1906, the German navy installed a prototype of the Anschütz-<br />

Kaempfe gyrocompass on the battleship Undine <strong>and</strong> subjected it to<br />

exhaustive tests under simulated battle conditions, sailing the ship<br />

under forced draft <strong>and</strong> suddenly reversing the engines, changing the<br />

position of heavy turrets <strong>and</strong> other mechanisms, <strong>and</strong> firing heavy<br />

guns. In conditions under which a magnetic compass would have<br />

been worthless, the gyrocompass proved a satisfactory navigational<br />

tool, <strong>and</strong> the results were impressive enough to convince the German<br />

navy to undertake installation of gyrocompasses in submarines <strong>and</strong><br />

heavy battleships, including the battleship Deutschl<strong>and</strong>.<br />

Elmer Ambrose Sperry, a New York inventor intimately associated<br />

with pioneer electrical development, was independently work-


ing on a design for a gyroscopic compass at about the same time.<br />

In 1907, he patented a gyrocompass consisting of a single rotor<br />

mounted within two concentric shells, suspended by fine piano<br />

wire from a frame mounted on gimbals. The rotor of the Sperry<br />

compass operated in a vacuum, which enabled it to rotate more<br />

rapidly. The Sperry gyrocompass was in use on larger American<br />

battleships <strong>and</strong> submarines on the eve of World War I (1914-1918).<br />

Impact<br />

The ability to navigate submerged submarines was of critical<br />

strategic importance in World War I. Initially, the German navy<br />

had an advantage both in the number of submarines at its disposal<br />

<strong>and</strong> in their design <strong>and</strong> maneuverability. The German U-boat fleet<br />

declared all-out war on Allied shipping, <strong>and</strong>, although their efforts<br />

to blockade Engl<strong>and</strong> <strong>and</strong> France were ultimately unsuccessful, the<br />

tremendous toll they inflicted helped maintain the German position<br />

<strong>and</strong> prolong the war. To a submarine fleet operating throughout<br />

the Atlantic <strong>and</strong> in the Caribbean, as well as in near-shore European<br />

waters, effective long-distance navigation was critical.<br />

Gyrocompasses were st<strong>and</strong>ard equipment on submarines <strong>and</strong><br />

battleships <strong>and</strong>, increasingly, on larger commercial vessels during<br />

World War I, World War II (1939-1945), <strong>and</strong> the period between the<br />

wars. The devices also found their way into aircraft, rockets, <strong>and</strong><br />

guided missiles. Although the compasses were made more accurate<br />

<strong>and</strong> easier to use, the fundamental design differed little from that invented<br />

by Anschütz-Kaempfe.<br />

See also Atomic-powered ship; Dirigible; Hovercraft; Radar; Sonar.<br />

Further Reading<br />

Gyrocompass / 385<br />

Hughes, Thomas Parke. Elmer Sperry: Inventor <strong>and</strong> Engineer. Baltimore:<br />

Johns Hopkins University Press, 1993.<br />

_____. Science <strong>and</strong> the Instrument-Maker: Michelson, Sperry, <strong>and</strong> the<br />

Speed of Light. Washington: Smithsonian Institution Press, 1976.<br />

Sorg, H. W. “From Serson to Draper: Two Centuries of Gyroscopic<br />

Development.” Journal of the Institute of Navigation 23, no. 4 (Winter,<br />

1976-1977).


386<br />

Hard disk<br />

Hard disk<br />

The invention: A large-capacity, permanent magnetic storage device<br />

built into most personal computers.<br />

The people behind the invention:<br />

Alan Shugart (1930- ), an engineer who first developed the<br />

floppy disk<br />

Philip D. Estridge (1938?-1985), the director of IBM’s product<br />

development facility<br />

Thomas J. Watson, Jr. (1914-1993), the chief executive officer of<br />

IBM<br />

The Personal Oddity<br />

When the International Business Machines (IBM) Corporation<br />

introduced its first microcomputer, called simply the IBM PC (for<br />

“personal computer”), the occasion was less a dramatic invention<br />

than the confirmation of a trend begun some years before. A number<br />

of companies had introduced microcomputers before IBM; one<br />

of the best known at that time was Apple Corporation’s Apple II, for<br />

which software for business <strong>and</strong> scientific use was quickly developed.<br />

Nevertheless, the microcomputer was quite expensive <strong>and</strong><br />

was often looked upon as an oddity, not as a useful tool.<br />

Under the leadership of Thomas J. Watson, Jr., IBM, which had<br />

previously focused on giant mainframe computers, decided to develop<br />

the PC. A design team headed by Philip D. Estridge was assembled<br />

in Boca Raton, Florida, <strong>and</strong> it quickly developed its first,<br />

pacesetting product. It is an irony of history that IBM anticipated<br />

selling only one hundred thous<strong>and</strong> or so of these machines, mostly<br />

to scientists <strong>and</strong> technically inclined hobbyists. Instead, IBM’s product<br />

sold exceedingly well, <strong>and</strong> its design parameters, as well as its<br />

operating system, became st<strong>and</strong>ards.<br />

The earliest microcomputers used a cassette recorder as a means<br />

of mass storage; a floppy disk drive capable of storing approximately<br />

160 kilobytes of data was initially offered only as an option.<br />

While home hobbyists were accustomed to using a cassette recorder


for storage purposes, such a system was far too slow <strong>and</strong> awkward<br />

for use in business <strong>and</strong> science. As a result, virtually every IBM PC<br />

sold was equipped with at least one 5.25-inch floppy disk drive.<br />

Memory Requirements<br />

Hard disk / 387<br />

All computers require memory of two sorts in order to carry out<br />

their tasks. One type of memory is main memory, or r<strong>and</strong>om access<br />

memory (RAM), which is used by the computer’s central processor<br />

to store data it is using while operating. The type of memory used<br />

for this function is built typically of silicon-based integrated circuits<br />

that have the advantage of speed (to allow the processor to fetch or<br />

store the data quickly), but the disadvantage of possibly losing or<br />

“forgetting” data when the electric current is turned off. Further,<br />

such memory generally is relatively expensive.<br />

To reduce costs, another type of memory—long-term storage<br />

memory, known also as “mass storage”—was developed. Mass<br />

storage devices include magnetic media (tape or disk drives) <strong>and</strong><br />

optical media (such as the compact disc, read-only memory, or CD-<br />

ROM). While the speed with which data may be retrieved from or<br />

stored in such devices is rather slow compared to the central processor’s<br />

speed, a disk drive—the most common form of mass storage<br />

used in PCs—can store relatively large amounts of data quite inexpensively.<br />

Early floppy disk drives (so called because the magnetically<br />

treated material on which data are recorded is made of a very flexible<br />

plastic) held 160 kilobytes of data using only one side of the<br />

magnetically coated disk (about eighty pages of normal, doublespaced,<br />

typewritten information). Later developments increased<br />

storage capacities to 360 kilobytes by using both sides of the disk<br />

<strong>and</strong> later, with increasing technological ability, 1.44 megabytes (millions<br />

of bytes). In contrast, mainframe computers, which are typically<br />

connected to large <strong>and</strong> expensive tape drive storage systems,<br />

could store gigabytes (millions of megabytes) of information.<br />

While such capacities seem large, the needs of business <strong>and</strong> scientific<br />

users soon outstripped available space. Since even the mailing<br />

list of a small business or a scientist’s mathematical model of a<br />

chemical reaction easily could require greater storage potential than


388 / Hard disk<br />

early PCs allowed, the need arose for a mass storage device that<br />

could accommodate very large files of data.<br />

The answer was the hard disk drive, also known as a “fixed disk<br />

drive,” reflecting the fact that the disk itself is not only rigid but also<br />

permanently installed inside the machine. In 1955, IBM had envisioned<br />

the notion of a fixed, hard magnetic disk as a means of storing<br />

computer data, <strong>and</strong>, under the direction of Alan Shugart in the<br />

1960’s, the floppy disk was developed as well.<br />

As the engineers of IBM’s facility in Boca Raton refined the idea<br />

of the original PC to design the new IBM PC XT, it became clear that<br />

chief among the needs of users was the availability of large-capability<br />

storage devices. The decision was made to add a 10-megabyte<br />

hard disk drive to the PC. On March 8, 1983, less than two years after<br />

the introduction of its first PC, IBM introduced the PC XT. Like<br />

the original, it was an evolutionary design, not a revolutionary one.<br />

The inclusion of a hard disk drive, however, signaled that mass storage<br />

devices in personal computers had arrived.<br />

Consequences<br />

Above all else, any computer provides a means for storing, ordering,<br />

analyzing, <strong>and</strong> presenting information. If the personal computer<br />

is to become the information appliance some have suggested<br />

it will be, the ability to manipulate very large amounts of data will<br />

be of paramount concern. Hard disk technology was greeted enthusiastically<br />

in the marketplace, <strong>and</strong> the dem<strong>and</strong> for hard drives has<br />

seen their numbers increase as their quality increases <strong>and</strong> their<br />

prices drop.<br />

It is easy to underst<strong>and</strong> one reason for such eager acceptance:<br />

convenience. Floppy-bound computer users find themselves frequently<br />

changing (or “swapping”) their disks in order to allow programs<br />

to find the data they need. Moreover, there is a limit to how<br />

much data a single floppy disk can hold. The advantage of a hard<br />

drive is that it allows users to keep seemingly unlimited amounts of<br />

data <strong>and</strong> programs stored in their machines <strong>and</strong> readily available.<br />

Also, hard disk drives are capable of finding files <strong>and</strong> transferring<br />

their contents to the processor much more quickly than a<br />

floppy drive. A user may thus create exceedingly large files, keep


them on h<strong>and</strong> at all times, <strong>and</strong> manipulate data more quickly than<br />

with a floppy. Finally, while a hard drive is a slow substitute for<br />

main memory, it allows users to enjoy the benefits of larger memories<br />

at significantly lower cost.<br />

The introduction of the PC XT with its 10-megabyte hard drive<br />

was a milestone in the development of the PC. Over the next two decades,<br />

the size of computer hard drives increased dramatically. By<br />

2001, few personal computers were sold with hard drives with less<br />

than three gigabytes of storage capacity, <strong>and</strong> hard drives with more<br />

than thirty gigabytes were becoming the st<strong>and</strong>ard. Indeed, for less<br />

money than a PC XT cost in the mid-1980’s, one could buy a fully<br />

equipped computer with a hard drive holding sixty gigabytes—a<br />

storage capacity equivalent to six thous<strong>and</strong> 10-megabyte hard drives.<br />

See also Bubble memory; Compact disc; Computer chips;<br />

Floppy disk; Optical disk; Personal computer.<br />

Further Reading<br />

Hard disk / 389<br />

Chposky, James, <strong>and</strong> Ted Leonsis. Blue Magic: The People, Power, <strong>and</strong><br />

Politics Behind the IBM Personal Computer. New York: Facts on File,<br />

1988.<br />

Freiberger, Paul, <strong>and</strong> Michael Swaine. Fire in the Valley: The Making of<br />

the Personal Computer. 2d ed. New York: McGraw-Hill, 2000.<br />

Grossman, Wendy. Remembering the Future: Interviews from Personal<br />

Computer World. New York: Springer, 1997.<br />

Watson, Thomas J., <strong>and</strong> Peter Petre. Father, Son <strong>and</strong> Co.: My Life at<br />

IBM <strong>and</strong> Beyond. New York: Bantam Books, 2000.


390<br />

Hearing aid<br />

Hearing aid<br />

The invention: Miniaturized electronic amplifier worn inside the<br />

ears of hearing-impaired persons.<br />

The organization behind the invention:<br />

Bell Labs, the research <strong>and</strong> development arm of the American<br />

Telephone <strong>and</strong> Telegraph Company<br />

Trapped in Silence<br />

Until the middle of the twentieth century, people who experienced<br />

hearing loss had little hope of being able to hear sounds without the<br />

use of large, awkward, heavy appliances. For many years, the only<br />

hearing aids available were devices known as ear trumpets. The ear<br />

trumpet tried to compensate for hearing loss by increasing the number<br />

of sound waves funneled into the ear canal. A wide, bell-like<br />

mouth similar to the bell of a musical trumpet narrowed to a tube that<br />

the user placed in his or her ear. Ear trumpets helped a little, but they<br />

could not truly increase the volume of the sounds heard.<br />

Beginning in the nineteenth century, inventors tried to develop<br />

electrical devices that would serve as hearing aids. The telephone<br />

was actually a by-product of Alex<strong>and</strong>er Graham Bell’s efforts to<br />

make a hearing aid. Following the invention of the telephone, electrical<br />

engineers designed hearing aids that employed telephone<br />

technology, but those hearing aids were only a slight improvement<br />

over the old ear trumpets. They required large, heavy battery packs<br />

<strong>and</strong> used a carbon microphone similar to the receiver in a telephone.<br />

More sensitive than purely physical devices such as the ear trumpet,<br />

they could transmit a wider range of sounds but could not amplify<br />

them as effectively as electronic hearing aids now do.<br />

Transistors Make Miniaturization Possible<br />

Two types of hearing aids exist: body-worn <strong>and</strong> head-worn.<br />

Body-worn hearing aids permit the widest range of sounds to be<br />

heard, but because of the devices’ larger size, many hearing-


Hearing aid / 391<br />

impaired persons do not like to wear them. Head-worn hearing<br />

aids, especially those worn completely in the ear, are much less conspicuous.<br />

In addition to in-ear aids, the category of head-worn hearing<br />

aids includes both hearing aids mounted in eyeglass frames <strong>and</strong><br />

those worn behind the ear.<br />

All hearing aids, whether head-worn or body-worn, consist of<br />

four parts: a microphone to pick up sounds, an amplifier, a receiver,<br />

<strong>and</strong> a power source. The microphone gathers sound waves <strong>and</strong> converts<br />

them to electrical signals; the amplifier boosts, or increases,<br />

those signals; <strong>and</strong> the receiver then converts the signals back into<br />

sound waves. In effect, the hearing aid is a miniature radio. After<br />

the receiver converts the signals back to sound waves, those waves<br />

are directed into the ear canal through an earpiece or ear mold. The<br />

ear mold generally is made of plastic <strong>and</strong> is custom fitted from an<br />

impression taken from the prospective user’s ear.<br />

Effective head-worn hearing aids could not be built until the<br />

electronic circuit was developed in the early 1950’s. The same invention—the<br />

transistor—that led to small portable radios <strong>and</strong> tape<br />

players allowed engineers to create miniaturized, inconspicuous<br />

hearing aids. Depending on the degree of amplification required,<br />

the amplifier in a hearing aid contains three or more transistors.<br />

Transistors first replaced vacuum tubes in devices such as radios<br />

<strong>and</strong> phonographs, <strong>and</strong> then engineers realized that they could be<br />

used in devices for the hearing-impaired.<br />

The research at Bell Labs that led to the invention of the transistor<br />

rose out of military research during World War II. The vacuum tubes<br />

used in, for example, radar installations to amplify the strength of electronic<br />

signals were big, were fragile because they were made of<br />

blown glass, <strong>and</strong> gave off high levels of heat when they were used.<br />

Transistors, however, made it possible to build solid-state, integrated<br />

circuits. These are made from crystals of metals such as germanium<br />

or arsenic alloys <strong>and</strong> therefore are much less fragile than glass. They<br />

are also extremely small (in fact, some integrated circuits are barely<br />

visible to the naked eye) <strong>and</strong> give off no heat during use.<br />

The number of transistors in a hearing aid varies depending upon<br />

the amount of amplification required. The first transistor is the most<br />

important for the listener in terms of the quality of sound heard. If the<br />

frequency response is set too high—that is, if the device is too sensi-


392 / Hearing aid<br />

tive—the listener will be bothered by distracting background noise.<br />

Theoretically, there is no limit on the amount of amplification that a<br />

hearing aid can be designed to provide, but there are practical limits.<br />

The higher the amplification, the more power is required to operate<br />

the hearing aid. This is why body-worn hearing aids can convey a<br />

wider range of sounds than head-worn devices can. It is the power<br />

source—not the electronic components—that is the limiting factor. A<br />

body-worn hearing aid includes a larger battery pack than can be<br />

used with a head-worn device. Indeed, despite advances in battery<br />

technology, the power requirements of a head-worn hearing aid are<br />

such that a 1.4-volt battery that could power a wristwatch for several<br />

years will last only a few days in a hearing aid.<br />

Consequences<br />

The invention of the electronic hearing aid made it possible for<br />

many hearing-impaired persons to participate in a hearing world.<br />

Prior to the invention of the hearing aid, hearing-impaired children<br />

often were unable to participate in routine school activities or function<br />

effectively in mainstream society. Instead of being able to live at<br />

home with their families <strong>and</strong> enjoy the same experiences that were<br />

available to other children their age, often they were forced to attend<br />

special schools operated by the state or by charities.<br />

Hearing-impaired people were singled out as being different <strong>and</strong><br />

were limited in their choice of occupations. Although not every<br />

hearing-impaired person can be helped to hear with a hearing aid—<br />

particularly in cases of total hearing loss—the electronic hearing aid<br />

has ended restrictions for many hearing-impaired people. Hearingimpaired<br />

children are now included in public school classes, <strong>and</strong><br />

hearing-impaired adults can now pursue occupations from which<br />

they were once excluded.<br />

Today, many deaf <strong>and</strong> hearing-impaired persons have chosen to<br />

live without the help of a hearing aid. They believe that they are not<br />

disabled but simply different, <strong>and</strong> they point out that their “disability”<br />

often allows them to appreciate <strong>and</strong> participate in life in unique<br />

<strong>and</strong> positive ways. For them, the use of hearing aids is a choice, not a<br />

necessity. For those who choose, hearing aids make it possible to<br />

participate in the hearing world.


See also Artificial heart; Artificial kidney; Cell phone; Contact<br />

lenses; Heart-lung machine; Pacemaker.<br />

Further Reading<br />

Hearing aid / 393<br />

Alex<strong>and</strong>er, Howard. “Hearing Aids: Smaller <strong>and</strong> Smarter.” New<br />

York Times (November 26, 1998).<br />

Fong, Petti. “Guess What’s the New Buzz in Hearing Aids.” Business<br />

Week, no. 3730 (April 30, 2001).<br />

Levitt, Harry. “Noise Reduction in Hearing Aids: AReview.” Journal<br />

of Rehabilitation Research <strong>and</strong> Development 38, no. 1 (January/February,<br />

2001).


394<br />

Heart-lung machine<br />

Heart-lung machine<br />

The invention: The first artificial device to oxygenate <strong>and</strong> circulate<br />

blood during surgery, the heart-lung machine began the era of<br />

open-heart surgery.<br />

The people behind the invention:<br />

John H. Gibbon, Jr. (1903-1974), a cardiovascular surgeon<br />

Mary Hopkinson Gibbon (1905- ), a research technician<br />

Thomas J. Watson (1874-1956), chairman of the board of IBM<br />

T. L. Stokes <strong>and</strong> J. B. Flick, researchers in Gibbon’s laboratory<br />

Bernard J. Miller (1918- ), a cardiovascular surgeon <strong>and</strong><br />

research associate<br />

Cecelia Bavolek, the first human to undergo open-heart surgery<br />

successfully using the heart-lung machine<br />

A Young Woman’s Death<br />

In the first half of the twentieth century, cardiovascular medicine<br />

had many triumphs. Effective anesthesia, antiseptic conditions, <strong>and</strong><br />

antibiotics made surgery safer. Blood-typing, anti-clotting agents,<br />

<strong>and</strong> blood preservatives made blood transfusion practical. Cardiac<br />

catheterization (feeding a tube into the heart), electrocardiography,<br />

<strong>and</strong> fluoroscopy (visualizing living tissues with an X-ray machine)<br />

made the nonsurgical diagnosis of cardiovascular problems possible.<br />

As of 1950, however, there was no safe way to treat damage or defects<br />

within the heart. To make such a correction, this vital organ’s<br />

function had to be interrupted. The problem was to keep the body’s<br />

tissues alive while working on the heart. While some surgeons practiced<br />

so-called blind surgery, in which they inserted a finger into the<br />

heart through a small incision without observing what they were attempting<br />

to correct, others tried to reduce the body’s need for circulation<br />

by slowly chilling the patient until the heart stopped. Still other<br />

surgeons used “cross-circulation,” in which the patient’s circulation<br />

was connected to a donor’s circulation. All these approaches carried<br />

profound risks of hemorrhage, tissue damage, <strong>and</strong> death.<br />

In February of 1931, Gibbon witnessed the death of a young


woman whose lung circulation was blocked by a blood clot. Because<br />

her blood could not pass through her lungs, she slowly lost<br />

consciousness from lack of oxygen. As he monitored her pulse <strong>and</strong><br />

breathing, Gibbon thought about ways to circumvent the obstructed<br />

lungs <strong>and</strong> straining heart <strong>and</strong> provide the oxygen required. Because<br />

surgery to remove such a blood clot was often fatal, the woman’s<br />

surgeons operated only as a last resort. Though the surgery took<br />

only six <strong>and</strong> one-half minutes, she never regained consciousness.<br />

This experience prompted Gibbon to pursue what few people then<br />

considered a practical line of research: a way to circulate <strong>and</strong> oxygenate<br />

blood outside the body.<br />

A Woman’s Life Restored<br />

Heart-lung machine / 395<br />

Gibbon began the project in earnest in 1934, when he returned to<br />

the laboratory of Edward D. Churchill at Massachusetts General<br />

Hospital for his second surgical research fellowship. He was assisted<br />

by Mary Hopkinson Gibbon. Together, they developed, using<br />

cats, a surgical technique for removing blood from a vein, supplying<br />

the blood with oxygen, <strong>and</strong> returning it to an artery using tubes inserted<br />

into the blood vessels. Their objective was to create a device<br />

that would keep the blood moving, spread it over a very thin layer<br />

to pick up oxygen efficiently <strong>and</strong> remove carbon dioxide, <strong>and</strong> avoid<br />

both clotting <strong>and</strong> damaging blood cells. In 1939, they reported that<br />

prolonged survival after heart-lung bypass was possible in experimental<br />

animals.<br />

World War II (1939-1945) interrupted the progress of this work; it<br />

was resumed by Gibbon at Jefferson Medical College in 1944. Shortly<br />

thereafter, he attracted the interest of Thomas J. Watson, chairman of<br />

the board of the International Business Machines (IBM) Corporation,<br />

who provided the services of IBM’s experimental physics laboratory<br />

<strong>and</strong> model machine shop as well as the assistance of staff engineers.<br />

IBM constructed <strong>and</strong> modified two experimental machines<br />

over the next seven years, <strong>and</strong> IBM engineers contributed significantly<br />

to the evolution of a machine that would be practical in humans.<br />

Gibbon’s first attempt to use the pump-oxygenator in a human<br />

being was in a fifteen-month-old baby. This attempt failed, not be-


396 / Heart-lung machine<br />

cause of a malfunction or a surgical mistake but because of a misdiagnosis.<br />

The child died following surgery because the real problem<br />

had not been corrected by the surgery.<br />

On May 6, 1953, the heart-lung machine was first used successfully<br />

on Cecelia Bavolek. In the six months before surgery, Bavolek<br />

had been hospitalized three times for symptoms of heart failure<br />

when she tried to engage in normal activity. While her circulation<br />

was connected to the heart-lung machine for forty-five minutes, the<br />

surgical team headed by Gibbon was able to close an opening between<br />

her atria <strong>and</strong> establish normal heart function. Two months<br />

later, an examination of the defect revealed that it was fully closed;<br />

Bavolek resumed a normal life. The age of open-heart surgery had<br />

begun.<br />

Consequences<br />

The heart-lung bypass technique alone could not make openheart<br />

surgery truly practical. When it was possible to keep tissues<br />

alive by diverting blood around the heart <strong>and</strong> oxygenating it, other<br />

questions already under investigation became even more critical:<br />

how to prolong the survival of bloodless organs, how to measure<br />

oxygen <strong>and</strong> carbon dioxide levels in the blood, <strong>and</strong> how to prolong<br />

anesthesia during complicated surgery. Thus, following the first<br />

successful use of the heart-lung machine, surgeons continued to refine<br />

the methods of open-heart surgery.<br />

The heart-lung apparatus set the stage for the advent of “replacement<br />

parts” for many types of cardiovascular problems. Cardiac<br />

valve replacement was first successfully accomplished in 1960 by<br />

placing an artificial ball valve between the left atrium <strong>and</strong> ventricle.<br />

In 1957, doctors performed the first coronary bypass surgery, grafting<br />

sections of a leg vein into the heart’s circulation system to divert<br />

blood around clogged coronary arteries. Likewise, the first successful<br />

heart transplant (1967) <strong>and</strong> the controversial Jarvik-7 artificial<br />

heart implantation (1982) required the ability to stop the heart <strong>and</strong><br />

keep the body’s tissues alive during time-consuming <strong>and</strong> delicate<br />

surgical procedures. Gibbon’s heart-lung machine paved the way<br />

for all these developments.


See also Artificial heart; Blood transfusion; CAT scanner; Coronary<br />

artery bypass surgery; Electrocardiogram; Iron lung; Mammography;<br />

Nuclear magnetic resonance; Pacemaker; X-ray image<br />

intensifier.<br />

Further Reading<br />

Heart-lung machine / 397<br />

DeJauregui, Ruth. One Hundred Medical Milestones That Shaped World<br />

History. San Mateo, Calif.: Bluewood Books, 1998.<br />

Romaine-Davis, Ada. John Gibbon <strong>and</strong> his Heart-Lung Machine. Philadelphia:<br />

University of Pennsylvania Press, 1991.<br />

Shumacker, Harris B. A Dream of the Heart: The Life of John H. Gibbon,<br />

Jr., Father of the Heart-Lung Machine. Santa Barbara, Calif.: Fithian<br />

Press, 1999.<br />

Watson, Thomas J., <strong>and</strong> Peter Petre. Father, Son <strong>and</strong> Co.: My Life at<br />

IBM <strong>and</strong> Beyond. New York: Bantam Books, 2000.


398<br />

Heat pump<br />

Heat pump<br />

The invention: A device that warms <strong>and</strong> cools buildings efficiently<br />

<strong>and</strong> cheaply by moving heat from one area to another.<br />

The people behind the invention:<br />

T. G. N. Haldane, a British engineer<br />

Lord Kelvin (William Thomson, 1824-1907), a British<br />

mathematician, scientist, <strong>and</strong> engineer<br />

Sadi Carnot (1796-1832), a French physicist <strong>and</strong><br />

thermodynamicist<br />

The Heat Pump<br />

A heat pump is a device that takes in heat at one temperature <strong>and</strong><br />

releases it at a higher temperature. When operated to provide heat (for<br />

example, for space heating), the heat pump is said to operate in the<br />

heating mode; when operated to remove heat (for example, for air conditioning),<br />

it is said to operate in the cooling mode. Some type of work<br />

must be done to drive the pump, no matter which mode is being used.<br />

There are two general types of heat pumps: vapor compression<br />

pumps <strong>and</strong> absorption pumps. The basic principle of vapor compression<br />

cycle heat pumps is derived from the work of Sadi Carnot<br />

in the early nineteenth century. Carnot’s work was published in<br />

1824. It was William Thomson (later to become known as Lord Kelvin),<br />

however, who first proposed a practical heat pump system, or<br />

“heat multiplier,” as it was known then, <strong>and</strong> he also indicated that a<br />

refrigerating machine could be used for heating.<br />

Thomson’s heat pump used air as its working fluid. Thomson<br />

claimed that his heat pump was able to produce heat by using only<br />

3 percent of the energy that would be required for direct heating.<br />

Absorption cycle machines have an even longer history. Refrigerators<br />

based on the use of sulfuric acid <strong>and</strong> water date back to 1777.<br />

Systems using this fluid combination, improved <strong>and</strong> modified by<br />

Edmond Carré, were used extensively in Paris cafés in the late<br />

1800’s. In 1849, a patent was filed by Ferdin<strong>and</strong> Carré for the working-fluid<br />

pair of ammonia <strong>and</strong> water in absorption cycle machines.


Refrigerator or Heater<br />

Heat pump / 399<br />

In the early nineteenth century, many people (including some<br />

electrical engineers) believed that electrical energy could never be<br />

used economically to produce large quantities of heat under ordinary<br />

conditions. A few researchers, however, believed that it was<br />

possible to produce heat by using electrical energy if that energy<br />

was first converted to mechanical energy <strong>and</strong> if the Carnot principle<br />

was then used to pump heat from a lower to a higher temperature.<br />

In 1927, T. G. N. Haldane carried out detailed experiments showing<br />

that the heat pump can be made to operate in either the heating<br />

mode or the cooling mode. A heat pump in the cooling mode works<br />

like a refrigerator; a heat pump in the heating mode supplies heat<br />

for heating. Haldane demonstrated that a refrigerator could be<br />

modified to work as a heating unit. He used a vapor compression<br />

cycle refrigerator for his demonstration.<br />

In the design of a refrigerating device, the primary objective is<br />

the production of cold rather than heat, but the two operations are<br />

complementary. The process of producing cold is simply that of<br />

pumping heat from a relatively cold to a relatively hot source, but in<br />

the refrigeration process particular attention is paid to the prevention<br />

of the leakage of heat into the cold source, whereas no attempt<br />

is made to prevent the escape of heat from the hot source. If a refrigerating<br />

device were treated as a heat pump in which the primary<br />

product is the heat rejected to the hot source, the order of importance<br />

would be reversed, <strong>and</strong> every opportunity would be taken to<br />

allow heat to leak into the cold source <strong>and</strong> every precaution would<br />

be taken against allowing heat to leak out of the hot source.<br />

The components of a heat pump that operates on the principle of<br />

vapor compression include an electric motor, a compressor, an evaporator,<br />

<strong>and</strong> a condenser. The compressor sucks in gas from the evaporator<br />

<strong>and</strong> compresses it to a pressure that corresponds to a saturation<br />

temperature that is slightly higher than that of the required heat. From<br />

the compressor, the compressed gas passes to the condenser, where it is<br />

cooled <strong>and</strong> condensed, thereby giving up a large quantity of heat to the<br />

water or other substance that it is intended to heat. The condensed gas<br />

then passes through the expansion valve, where a sudden reduction of<br />

pressure takes place. This reduction of pressure lowers the boiling


400 / Heat pump<br />

point of the liquid, which therefore vaporizes <strong>and</strong> takes in heat from<br />

the medium surrounding the evaporator. After evaporation, the gas<br />

passes on to the compressor, <strong>and</strong> the cycle is complete.<br />

Haldane was the first person in the United Kingdom to install a<br />

heat pump. He was also the first person to install a domestic heat<br />

pump to provide hot water <strong>and</strong> space heating.<br />

Heat In<br />

Impact<br />

Low-Pressure<br />

Vapor<br />

Electric Power<br />

High-Pressure<br />

Vapor<br />

Compressor<br />

Evaporator Condenser<br />

Low-Pressure<br />

Liquid<br />

Expansion Valve<br />

Components of a heat pump.<br />

High-Pressure<br />

Liquid<br />

Heat Out<br />

Since Haldane’s demonstration of the use of the heat pump, the<br />

device has been highly successful in people’s homes, especially in<br />

those regions where both heating <strong>and</strong> cooling are required for single-<br />

<strong>and</strong> multifamily residences (for example, Australia, Japan, <strong>and</strong><br />

the United States). This is the case because the heat pump can provide<br />

both heating <strong>and</strong> cooling; therefore, the cost of a heat pump<br />

system can be spread over both heating <strong>and</strong> cooling seasons. Total<br />

annual sales of heat pumps worldwide have risen to the millions,<br />

with most sales being made in Japan <strong>and</strong> the United States.<br />

The use of heat pumps can save energy. In addition, because they<br />

are electric, they can save significant quantities of oil, especially in<br />

the residential retrofit <strong>and</strong> replacement markets <strong>and</strong> when used as<br />

add-on devices for existing heating systems. Some heat pumps are<br />

now available that may compete cost-effectively with other heating<br />

systems in meeting the heating dem<strong>and</strong>s of cooler regions.


Technological developments by heat pump manufacturers are<br />

continually improving the performance <strong>and</strong> cost-effectiveness of<br />

heat pumps. The electric heat pump will continue to dominate the<br />

residential market, although engine-driven systems are likely to<br />

have a greater impact on the multifamily market.<br />

See also Breeder reactor; Compressed-air-accumulating power<br />

plant; Fuel cell; Geothermal power; Nuclear power plant; Solar<br />

thermal engine; Tidal power plant.<br />

Further Reading<br />

Heat pump / 401<br />

Kavanaugh, Stephen P., <strong>and</strong> Kevin D. Rafferty. Ground-Source Heat<br />

Pumps: Design of Geothermal Systems for Commercial <strong>and</strong> Institutional<br />

Buildings. Atlanta: American Society of Heating, Refrigerating<br />

<strong>and</strong> Air-Conditioning Engineers, 1997.<br />

Nisson, Ned. “Efficient <strong>and</strong> Affordable.” Popular Science 247, no. 2<br />

(August, 1995).<br />

Using the Earth to Heat <strong>and</strong> Cool Homes. Washington, D.C.: U.S. Department<br />

of Energy, 1983.


402<br />

Holography<br />

Holography<br />

The invention: A lensless system of three-dimensional photography<br />

that was one of the most important developments in twentieth<br />

century optical science.<br />

The people behind the invention:<br />

Dennis Gabor (1900-1979), a Hungarian-born inventor <strong>and</strong><br />

physicist who was awarded the 1971 Nobel Prize in Physics<br />

Emmett Leith (1927- ), a radar researcher who, with Juris<br />

Upatnieks, produced the first laser holograms<br />

Juris Upatnieks (1936- ), a radar researcher who, with<br />

Emmett Leith, produced the first laser holograms<br />

Easter Inspiration<br />

The development of photography in the early 1900’s made possible<br />

the recording of events <strong>and</strong> information in ways unknown before<br />

the twentieth century: the photographing of star clusters, the<br />

recording of the emission spectra of heated elements, the storing of<br />

data in the form of small recorded images (for example, microfilm),<br />

<strong>and</strong> the photographing of microscopic specimens, among other<br />

things. Because of its vast importance to the scientist, the science of<br />

photography has developed steadily.<br />

An underst<strong>and</strong>ing of the photographic <strong>and</strong> holographic processes<br />

requires some knowledge of the wave behavior of light. Light is an<br />

electromagnetic wave that, like a water wave, has an amplitude <strong>and</strong> a<br />

phase. The amplitude corresponds to the wave height, while the<br />

phase indicates which part of the wave is passing a given point at a<br />

given time. A cork floating in a pond bobs up <strong>and</strong> down as waves<br />

pass under it. The position of the cork at any time depends on both<br />

amplitude <strong>and</strong> phase: The phase determines on which part of the<br />

wave the cork is floating at any given time, <strong>and</strong> the amplitude determines<br />

how high or low the cork can be moved. Waves from more<br />

than one source arriving at the cork combine in ways that depend on<br />

their relative phases. If the waves meet in the same phase, they add<br />

<strong>and</strong> produce a large amplitude; if they arrive out of phase, they sub-


tract <strong>and</strong> produce a small amplitude. The total amplitude, or intensity,<br />

depends on the phases of the combining waves.<br />

Dennis Gabor, the inventor of holography, was intrigued by the<br />

way in which the photographic image of an object was stored by a<br />

photographic plate but was unable to devote any consistent research<br />

effort to the question until the 1940’s. At that time, Gabor was involved<br />

in the development of the electron microscope. On Easter<br />

morning in 1947, as Gabor was pondering the problem of how to<br />

improve the electron microscope, the solution came to him. He<br />

would attempt to take a poor electron picture <strong>and</strong> then correct it optically.<br />

The process would require coherent electron beams—that is,<br />

electron waves with a definite phase.<br />

This two-stage method was inspired by the work of Lawrence<br />

Bragg. Bragg had formed the image of a crystal lattice by diffracting<br />

the photographic X-ray diffraction pattern of the original lattice.<br />

This double diffraction process is the basis of the holographic process.<br />

Bragg’s method was limited because of his inability to record<br />

the phase information of the X-ray photograph. Therefore, he could<br />

study only those crystals for which the phase relationship of the reflected<br />

waves could be predicted.<br />

Waiting for the Laser<br />

Holography / 403<br />

Gabor devised a way of capturing the phase information after he<br />

realized that adding coherent background to the wave reflected from<br />

an object would make it possible to produce an interference pattern<br />

on the photographic plate. When the phases of the two waves are<br />

identical, a maximum intensity will be recorded; when they are out of<br />

phase, a minimum intensity is recorded. Therefore, what is recorded<br />

in a hologram is not an image of the object but rather the interference<br />

pattern of the two coherent waves. This pattern looks like a collection<br />

of swirls <strong>and</strong> blank spots. The hologram (or photograph) is then illuminated<br />

by the reference beam, <strong>and</strong> part of the transmitted light is a<br />

replica of the original object wave. When viewing this object wave,<br />

one sees an exact replica of the original object.<br />

The major impediment at the time in making holograms using<br />

any form of radiation was a lack of coherent sources. For example,<br />

the coherence of the mercury lamp used by Gabor <strong>and</strong> his assistant


404 / Holography<br />

Dennis Gabor<br />

The eldest son of a mine director, Dennis Gabor was born in<br />

1900 in Budapest, Hungary. At fifteen, suddenly developing<br />

an intense interest in optics <strong>and</strong> photography, Gabor <strong>and</strong> his<br />

brother sent up their own home laboratory <strong>and</strong> experimented<br />

in those fields as well as with X rays <strong>and</strong> radioactivity. The love<br />

of physics never left him.<br />

Gabor graduated from the Berlin Technische Hochschule in<br />

1924 <strong>and</strong> earned a doctorate of engineering in 1927 after developing<br />

a high-speed cathode ray oscillograph <strong>and</strong> a new kind of<br />

magnetic lens for controlling electrons. After graduate school<br />

he joined Siemens <strong>and</strong> Halske Limited <strong>and</strong> invented a highpressure<br />

mercury lamp, which was later used widely in street<br />

lamps. In 1933, Gabor left Germany because of the rise of Nazism<br />

<strong>and</strong> moved to Engl<strong>and</strong>. He worked in industrial research<br />

until 1948, improving gas-discharge tubes <strong>and</strong> stereoscopic cinematography,<br />

but he also published scientific papers on his<br />

own, including the first of many on communications theory. At<br />

the beginning of 1949, Gabor became a faculty member of the<br />

Imperial College of Science <strong>and</strong> Technology in London, first as a<br />

reader in electronics <strong>and</strong> later as a professor of applied physics.<br />

During his academic years came more inventions, including<br />

the hologram, an electron-velocity spectroscope, an analog computer,<br />

a flat color television tube, <strong>and</strong> a new type of thermionic<br />

converter. He also build a cloud chamber for detecting subatomic<br />

particles <strong>and</strong> used it to study electron interactions. As<br />

interested in theory as he was in applied physics, Gabor<br />

published papers on theoretical aspects of communications,<br />

plasma, magnetrons, <strong>and</strong> fusion. In his later years he worried<br />

deeply about the modern tendency for technology to advance<br />

out of step with social institutions <strong>and</strong> wrote popular books<br />

outlining his belief that social reform should be given priority.<br />

Gabor became a member of Britain’s Royal Society in 1956<br />

<strong>and</strong> was awarded its Rumsford Medal in 1968. In 1971 he received<br />

the Nobel Prize in Physics for inventing holography. He<br />

died in London in 1979.


Ivor Williams was so short that they were able to make holograms of<br />

only about a centimeter in diameter. The early results were rather<br />

poor in terms of image quality <strong>and</strong> also had a double image. For this<br />

reason, there was little interest in holography, <strong>and</strong> the subject lay almost<br />

untouched for more than ten years.<br />

Interest in the field was rekindled after the laser (light amplification<br />

by stimulated emission of radiation) was developed in 1962.<br />

Emmett Leith <strong>and</strong> Juris Upatnieks, who were conducting radar research<br />

at the University of Michigan, published the first laser holographs<br />

in 1963. The laser was an intense light source with a very<br />

long coherence length. Its monochromatic nature improved the resolution<br />

of the images greatly. Also, there was no longer any restriction<br />

on the size of the object to be photographed.<br />

The availability of the laser allowed Leith <strong>and</strong> Upatnieks to propose<br />

another improvement in holographic technique. Before 1964,<br />

holograms were made of only thin transparent objects. A small region<br />

of the hologram bore a one-to-one correspondence to a region<br />

of the object. Only a small portion of the image could be viewed at<br />

one time without the aid of additional optical components. Illuminating<br />

the transparency diffusely allowed the whole image to be<br />

seen at one time. This development also made it possible to record<br />

holograms of diffusely reflected three-dimensional objects. Gabor<br />

had seen from the beginning that this should make it possible to create<br />

three-dimensional images.<br />

After the early 1960’s, the field of holography developed very<br />

quickly. Because holography is different from conventional photography,<br />

the two techniques often complement each other. Gabor saw<br />

his idea blossom into a very important technique in optical science.<br />

Impact<br />

Holography / 405<br />

The development of the laser <strong>and</strong> the publication of the first laser<br />

holograms in 1963 caused a blossoming of the new technique in<br />

many fields. Soon, techniques were developed that allowed holograms<br />

to be viewed with white light. It also became possible for holograms<br />

to reconstruct multicolored images. Holographic methods<br />

have been used to map terrain with radar waves <strong>and</strong> to conduct surveillance<br />

in the fields of forestry, agriculture, <strong>and</strong> meteorology.


406 / Holography<br />

By the 1990’s, holography had become a multimillion-dollar industry,<br />

finding applications in advertising, as an art form, <strong>and</strong> in security<br />

devices on credit cards, as well as in scientific fields. An alternate<br />

form of holography, also suggested by Gabor, uses sound<br />

waves. Acoustical imaging is useful whenever the medium around<br />

the object to be viewed is opaque to light rays—for example, in<br />

medical diagnosis. Holography has affected many areas of science,<br />

technology, <strong>and</strong> culture.<br />

See also Color film; Electron microscope; Infrared photography;<br />

Laser; Mammography; Mass spectrograph; X-ray crystallography.<br />

Further Reading<br />

Greguss, Pál, Tung H. Jeong, <strong>and</strong> Dennis Gabor. Holography: Commemorating<br />

the Ninetieth Anniversary of the Birth of Dennis Gabor.<br />

Bellingham, Wash.: SPIE Optical Engineering Press, 1991.<br />

Kasper, Joseph Emil, <strong>and</strong> Steven A. Feller. The Complete Book of Holograms:<br />

How They Work <strong>and</strong> How to Make Them. Mineola, N.Y.: Dover,<br />

2001.<br />

McNair, Don. How to Make Holograms. Blue Ridge Summit, Pa.: Tab<br />

Books, 1983.<br />

Saxby, Graham. Holograms: How to Make <strong>and</strong> Display Them. New<br />

York: Focal Press, 1980.


Hovercraft<br />

Hovercraft<br />

The invention: A vehicle requiring no surface contact for traction<br />

that moves freely over a variety of surfaces—particularly<br />

water—while supported on a self-generated cushion of air.<br />

The people behind the invention:<br />

Christopher Sydney Cockerell (1910- ), a British engineer<br />

who built the first hovercraft<br />

Ronald A. Shaw (1910- ), an early pioneer in aerodynamics<br />

who experimented with hovercraft<br />

Sir John Isaac Thornycroft (1843-1928), a Royal Navy architect<br />

who was the first to experiment with air-cushion theory<br />

Air-Cushion Travel<br />

407<br />

The air-cushion vehicle was first conceived by Sir John Isaac<br />

Thornycroft of Great Britain in the 1870’s. He theorized that if a<br />

ship had a plenum chamber (a box open at the bottom) for a hull<br />

<strong>and</strong> it were pumped full of air, the ship would rise out of the water<br />

<strong>and</strong> move faster, because there would be less drag. The main problem<br />

was keeping the air from escaping from under the craft.<br />

In the early 1950’s, Christopher Sydney Cockerell was experimenting<br />

with ways to reduce both the wave-making <strong>and</strong> frictional<br />

resistance that craft had to water. In 1953, he constructed a punt<br />

with a fan that supplied air to the bottom of the craft, which could<br />

thus glide over the surface with very little friction. The air was contained<br />

under the craft by specially constructed side walls. In 1955,<br />

the first true “hovercraft,” as Cockerell called it, was constructed of<br />

balsa wood. It weighed only 127 grams <strong>and</strong> traveled over water at a<br />

speed of 13 kilometers per hour.<br />

On November 16, 1956, Cockerell successfully demonstrated<br />

his model hovercraft at the patent agent’s office in London. It was<br />

immediately placed on the “secret” list, <strong>and</strong> Saunders-Roe Ltd.<br />

was given the first contract to build hovercraft in 1957. The first experimental<br />

piloted hovercraft, the SR.N1, which had a weight of<br />

3,400 kilograms <strong>and</strong> could carry three people at the speed of 25


408 / Hovercraft<br />

knots, was completed on May 28, 1959, <strong>and</strong> publicly demonstrated<br />

on June 11, 1959.<br />

Ground Effect Phenomenon<br />

In a hovercraft, a jet airstream is directed downward through a<br />

hole in a metal disk, which forces the disk to rise. The jet of air has a<br />

reverse effect of its own that forces the disk away from the surface.<br />

Some of the air hitting the ground bounces back against the disk to<br />

add further lift. This is called the “ground effect.” The ground effect<br />

is such that the greater the under-surface area of the hovercraft, the<br />

greater the reverse thrust of the air that bounces back. This makes<br />

the hovercraft a mechanically efficient machine because it provides<br />

three functions.<br />

First, the ground effect reduces friction between the craft <strong>and</strong> the<br />

earth’s surface. Second, it acts as a spring suspension to reduce<br />

some of the vertical acceleration effects that arise from travel over<br />

an uneven surface. Third, it provides a safe <strong>and</strong> comfortable ride at<br />

high speed, whatever the operating environment. The air cushion<br />

can distribute the weight of the hovercraft over almost its entire area<br />

so that the cushion pressure is low.<br />

The basic elements of the air-cushion vehicle are a hull, a propulsion<br />

system, <strong>and</strong> a lift system. The hull, which accommodates the<br />

crew, passengers, <strong>and</strong> freight, contains both the propulsion <strong>and</strong> lift<br />

systems. The propulsion <strong>and</strong> lift systems can be driven by the same<br />

power plant or by separate power plants. Early designs used only<br />

one unit, but this proved to be a problem when adequate power was<br />

not achieved for movement <strong>and</strong> lift. Better results are achieved<br />

when two units are used, since far more power is used to lift the vehicle<br />

than to propel it.<br />

For lift, high-speed centrifugal fans are used to drive the air<br />

through jets that are located under the craft. A redesigned aircraft<br />

propeller is used for propulsion. Rudderlike fins <strong>and</strong> an air fan that<br />

can be swiveled to provide direction are placed at the rear of the<br />

craft.<br />

Several different air systems can be used, depending on whether<br />

a skirt system is used in the lift process. The plenum chamber system,<br />

the peripheral jet system, <strong>and</strong> several types of recirculating air


Hovercraft / 409<br />

systems have all been successfully tried without skirting. A variety<br />

of rigid <strong>and</strong> flexible skirts have also proved to be satisfactory, depending<br />

on the use of the vehicle.<br />

Skirts are used to hold the air for lift. Skirts were once hung like cur-<br />

Sir John Isaac Thornycroft<br />

To be truly ahead of one’s time as an inventor, one must simply<br />

know everything there is to know about a specialty <strong>and</strong><br />

then imagine something useful that contemporary technology<br />

is not quite ready for.<br />

John Isaac Thornycroft was such an inventor. Born in 1843 in<br />

what were then the Papal States (Rome, Italy), he trained as an<br />

engineer <strong>and</strong> became a naval architect. He opened a boatbuilding<br />

<strong>and</strong> engineering company at Chiswick in London in<br />

1866 <strong>and</strong> began looking for ways to improve the performance of<br />

small seacraft. In 1877 he delivered the HMS Lightning, Engl<strong>and</strong>’s<br />

first torpedo boat, to the Royal Navy. He continued to<br />

make torpedo boats for coastal waters, nicknamed “scooters,”<br />

<strong>and</strong> made himself a leading expert on boat design. He introduced<br />

stabilizers <strong>and</strong> modified hull <strong>and</strong> propeller shapes in order<br />

to reduce drag from the hull’s contact with water <strong>and</strong><br />

thereby increase a boat’s speed.<br />

One of his best ideas was to have the boat ride on a cushion<br />

of air, so that air acted as a lubricant between the hull <strong>and</strong> water.<br />

He even filed patents for the concept <strong>and</strong> built models, but the<br />

power-source technology of the day was simply too inefficient.<br />

Engines were too heavy for the amount of power they put out.<br />

None could lift a full-size boat off the water <strong>and</strong> keep it on an air<br />

cushion. So the hovercraft had to wait until the 1950’s <strong>and</strong> incorporation<br />

of sophisticated internal combustion engines into<br />

the design.<br />

Meanwhile, Thornycroft <strong>and</strong> the company named after him<br />

continued to make innovative transports <strong>and</strong> engines: a steampowered<br />

van in 1896, a gas engine in 1902, <strong>and</strong> heavy trucks in<br />

1912 that the British government used during World War I. By<br />

the time Thornycroft died in 1928, on the Isle of Wight, he had<br />

been knighted by a grateful government, which would benefit<br />

from his company’s products <strong>and</strong> his advanced ideas for the<br />

rest of the twentieth century.


410 / Hovercraft<br />

tains around hovercraft. Instead of simple curtains to contain the air,<br />

there are now complicated designs that contain the cushion, duct the<br />

air, <strong>and</strong> even provide a secondary suspension. The materials used in<br />

the skirting have also changed from a rubberized fabric to pure rubber<br />

<strong>and</strong> nylon <strong>and</strong>, finally, to neoprene, a lamination of nylon <strong>and</strong> plastic.<br />

The three basic types of hovercraft are the amphibious, nonamphibious,<br />

<strong>and</strong> semiamphibious models. The amphibious type can<br />

travel over water <strong>and</strong> l<strong>and</strong>, whereas the nonamphibious type is restricted<br />

to water travel. The semiamphibious model is also restricted<br />

to water travel but may terminate travel by nosing up on a prepared<br />

ramp or beach. All hovercraft contain built-in buoyancy tanks in the<br />

side skirting as a safety measure in the event that a hovercraft must<br />

settle on the water. Most hovercraft are equipped with gas turbines<br />

<strong>and</strong> use either propellers or water-jet propulsion.<br />

Impact<br />

Hovercraft are used primarily for short passenger ferry services.<br />

Great Britain was the only nation to produce a large number of hovercraft.<br />

The British built larger <strong>and</strong> faster craft <strong>and</strong> pioneered their<br />

successful use as ferries across the English Channel, where they<br />

could reach speeds of 111 kilometers per hour (160 knots) <strong>and</strong> carry<br />

more than four hundred passengers <strong>and</strong> almost one hundred vehicles.<br />

France <strong>and</strong> the former Soviet Union have also effectively demonstrated<br />

hovercraft river travel, <strong>and</strong> the Soviets have experimented<br />

with military applications as well.<br />

The military adaptations of hovercraft have been more diversified.<br />

Beach l<strong>and</strong>ings have been performed effectively, <strong>and</strong> the United<br />

States used hovercraft for river patrols during the Vietnam War.<br />

Other uses also exist for hovercraft. They can be used as harbor pilot<br />

vessels <strong>and</strong> for patrolling shores in a variety of police-<strong>and</strong> customs-related<br />

duties. Hovercraft can also serve as flood-rescue craft<br />

<strong>and</strong> fire-fighting vehicles. Even a hoverfreighter is being considered.<br />

The air-cushion theory in transport systems is rapidly developing.<br />

It has spread to trains <strong>and</strong> smaller people movers in many<br />

countries. Their smooth, rapid, clean, <strong>and</strong> efficient operation makes<br />

hovercraft attractive to transportation designers around the world.


See also Airplane; Atomic-powered ship; Bullet train; Gyrocompass.<br />

Further Reading<br />

Hovercraft / 411<br />

Amyot, Joseph R. Hovercraft Technology, Economics, <strong>and</strong> Applications.<br />

Amsterdam: Elsevier, 1989.<br />

Croome, Angela. Hover Craft. 4th ed. London: Hodder <strong>and</strong> Stoughton,<br />

1984.<br />

Gromer, Cliff. “Flying Low.” Popular Mechanics 176, no. 9 (September,<br />

1999).<br />

McLeavy, Roy. Hovercraft <strong>and</strong> Hydrofoils. London: Jane’s Publishing,<br />

1980.<br />

Pengelley, Rupert. “Hovercraft Cushion the Blow of Amphibious<br />

Operations.” Jane’s Navy International 104, no. 008 (October 1,<br />

1999).<br />

Robertson, Don. A Restless Spirit. New Port, Isle of Wight: Cross<br />

Publishing, 1994.


412<br />

Hydrogen bomb<br />

Hydrogen bomb<br />

The invention: Popularly known as the “H-Bomb,” the hydrogen<br />

bomb differs from the original atomic bomb in using fusion,<br />

rather than fission, to create a thermonuclear explosion almost a<br />

thous<strong>and</strong> times more powerful.<br />

The people behind the invention:<br />

Edward Teller (1908- ), a Hungarian-born theoretical<br />

physicist<br />

Stanislaw Ulam (1909-1984), a Polish-born mathematician<br />

Crash Development<br />

A few months before the 1942 creation of the Manhattan Project,<br />

the United States-led effort to build the atomic (fission) bomb, physicist<br />

Enrico Fermi suggested to Edward Teller that such a bomb<br />

could release more energy by the process of heating a mass of the<br />

hydrogen isotope deuterium <strong>and</strong> igniting the fusion of hydrogen<br />

into helium. Fusion is the process whereby two atoms come together<br />

to form a larger atom, <strong>and</strong> this process usually occurs only in stars,<br />

such as the Sun. Physicists Hans Bethe, George Gamow, <strong>and</strong> Teller<br />

had been studying fusion since 1934 <strong>and</strong> knew of the tremendous<br />

energy than could be released by this process—even more energy<br />

than the fission (atom-splitting) process that would create the atomic<br />

bomb. Initially, Teller dismissed Fermi’s idea, but later in 1942, in<br />

collaboration with Emil Konopinski, he concluded that a hydrogen<br />

bomb, or superbomb, could be made.<br />

For practical considerations, it was decided that the design of the<br />

superbomb would have to wait until after the war. In 1946, a secret<br />

conference on the superbomb was held in Los Alamos, New Mexico,<br />

that was attended by, among other Manhattan Project veterans,<br />

Stanislaw Ulam <strong>and</strong> Klaus Emil Julius Fuchs. Supporting the investigation<br />

of Teller’s concept, the conferees requested a more complete<br />

mathematical analysis of his own admittedly crude calculations<br />

on the dynamics of the fusion reaction. In 1947, Teller believed<br />

that these calculations might take years. Two years later, however,


the Soviet explosion of an atomic bomb convinced Teller that America’s<br />

Cold War adversary was hard at work on its own superbomb.<br />

Even when new calculations cast further doubt on his designs,<br />

Teller began a vigorous campaign for crash development of the hydrogen<br />

bomb, or H-bomb.<br />

The Superbomb<br />

Hydrogen bomb / 413<br />

Scientists knew that fusion reactions could be induced by the explosion<br />

of an atomic bomb. The basic problem was simple <strong>and</strong> formidable:<br />

How could fusion fuel be heated <strong>and</strong> compressed long<br />

enough to achieve significant thermonuclear burning before the<br />

atomic fission explosion blew the assembly apart? A major part of<br />

the solution came from Ulam in 1951. He proposed using the energy<br />

from an exploding atomic bomb to induce significant thermonuclear<br />

reactions in adjacent fusion fuel components.<br />

This arrangement, in which the A-bomb (the primary) is physically<br />

separated from the H-bomb’s (the secondary’s) fusion fuel, became<br />

known as the “Teller-Ulam configuration.” All H-bombs are<br />

cylindrical, with an atomic device at one end <strong>and</strong> the other components<br />

filling the remaining space. Energy from the exploding primary<br />

could be transported by X rays <strong>and</strong> would therefore affect the<br />

fusion fuel at near light speed—before the arrival of the explosion.<br />

Frederick de Hoffman’s work verified <strong>and</strong> enriched the new concept.<br />

In the revised method, moderated X rays from the primary irradiate<br />

a reactive plastic medium surrounding concentric <strong>and</strong> generally<br />

cylindrical layers of fusion <strong>and</strong> fission fuel in the secondary.<br />

Instantly, the plastic becomes a hot plasma that compresses <strong>and</strong><br />

heats the inner layer of fusion fuel, which in turn compresses a central<br />

core of fissile plutonium to supercriticality. Thus compressed,<br />

<strong>and</strong> bombarded by fusion-produced, high-energy neutrons, the fission<br />

element exp<strong>and</strong>s rapidly in a chain reaction from the inside<br />

out, further compressing <strong>and</strong> heating the surrounding fusion fuel,<br />

releasing more energy <strong>and</strong> more neutrons that induce fission in a<br />

fuel casing-tamper made of normally stable uranium 238.<br />

With its equipment to refrigerate the hydrogen isotopes, the device<br />

created to test Teller’s new concept weighed more than sixty<br />

tons. During Operation Ivy, it was tested at Elugelab in the Marshall


414 / Hydrogen bomb<br />

Edward Teller<br />

To call Edward Teller “controversial” is equivalent to saying<br />

that the hydrogen bomb is “destructive”—an enormous understatement.<br />

His forceful support for nuclear arms prompted<br />

some to label him a war criminal while others consider him to<br />

be one of the most thoughtful statesmen among scientists.<br />

Teller was born into a Jewish family in Budapest, Hungary,<br />

in 1908. He left his homel<strong>and</strong> to flee the anti-Semitic fascist government<br />

of the late 1920’s <strong>and</strong> attended the University of Leipzig<br />

in Germany. In 1930 he completed his doctorate <strong>and</strong> hoped<br />

to settle into an academic career there, but he fled Germany<br />

when Adolf Hitler came to power. Teller migrated to the United<br />

States in 1935 <strong>and</strong> taught at George Washington University,<br />

where with George Gamow he studied aspects of quantum mechanics<br />

<strong>and</strong> nuclear physics. He became a U.S. citizen in 1941.<br />

Teller was among the first physicists to realize the possibility<br />

of an atomic (fission) bomb, <strong>and</strong> he became a central figure in<br />

the Manhattan Project that built it during World War II. However,<br />

he was already exploring the idea of a “superbomb” that<br />

explodes because of a fusion reaction. He helped persuade<br />

President Harry Truman to finance a project to build it <strong>and</strong> continued<br />

to influence the politics of nuclear weapons <strong>and</strong> power<br />

afterward. Teller developed the theoretical basis for the hydrogen<br />

bomb <strong>and</strong> its rough design—<strong>and</strong> so is know as its father.<br />

However, controversy later erupted over credit. Mathematician<br />

Stanislaw Ulam claimed he contributed key insights <strong>and</strong> calculations,<br />

a claim Teller vehemently denied. Teller, however, did<br />

credit a young physicist, Richard L. Garwin, with creating the<br />

successful working design for the first bomb.<br />

Fiercely anticommunist, Teller argued for a strong nuclear<br />

arsenal to make the Soviet Union afraid of attacking the United<br />

States <strong>and</strong> supported space-based missile defense systems. He<br />

served as director of the Lawrence Livermore National Laboratory,<br />

professor at the University of California at Berkeley, <strong>and</strong><br />

senior fellow at the nearby Hoover Institution. In his nineties he<br />

outraged environmentalists by suggesting that the atmosphere<br />

could be manipulated with technology to offset the effects of<br />

global warming.


Isl<strong>and</strong>s on November 1, 1952. Exceeding the expectations of all concerned<br />

<strong>and</strong> vaporizing the isl<strong>and</strong>, the explosion equaled 10.4 million<br />

tons of trinitrotoluene (TNT), which meant that it was about<br />

seven hundred times more powerful than the atomic bomb dropped<br />

on Hiroshima, Japan, in 1945. A version of this device weighing<br />

about 20 tons was prepared for delivery by specially modified Air<br />

Force B-36 bombers in the event of an emergency during wartime.<br />

In development at Los Alamos before the 1952 test was a device<br />

weighing only about 4 tons, a “dry bomb” that did not require refrigeration<br />

equipment or liquid fusion fuel; when sufficiently compressed<br />

<strong>and</strong> heated in its molded-powder form, the new fusion fuel<br />

component, lithium-6 deutride, instantly produced tritium, an isotope<br />

of hydrogen. This concept was tested during Operation Castle<br />

at Bikini atoll in 1954 <strong>and</strong> produced a yield of 15 million tons of TNT,<br />

the largest-ever nuclear explosion created by the United States.<br />

Consequences<br />

Hydrogen bomb / 415<br />

Teller was not alone in believing that the world could produce<br />

thermonuclear devices capable of causing great destruction. Months<br />

before Fermi suggested to Teller the possibility of explosive thermonuclear<br />

reactions on Earth, Japanese physicist Tokutaro Hagiwara<br />

had proposed that a uranium 235 bomb could ignite significant fusion<br />

reactions in hydrogen. The Soviet Union successfully tested an<br />

H-bomb dropped from an airplane in 1955, one year before the<br />

United States did so.<br />

Teller became the scientific adviser on nuclear affairs of many<br />

presidents, from Dwight D. Eisenhower to Ronald Reagan. The<br />

widespread blast <strong>and</strong> fallout effects of H-bombs assured the mutual<br />

destruction of the users of such weapons. During the Cold War<br />

(from about 1947 to 1981), both the United States <strong>and</strong> the Soviet<br />

Union possessed H-bombs. “Testing” these bombs made each side<br />

aware of how powerful the other side was. Everyone wanted to<br />

avoid nuclear war. It was thought that no one would try to start a<br />

war that would end in the world’s destruction. This theory was<br />

called deterrence: The United States wanted to let the Soviet Union<br />

know that it had just as many bombs, or more, than it did, so that the<br />

leaders of the Sovet Union would be deterred from starting a war.


416 / Hydrogen bomb<br />

Teller knew that the availability of H-bombs on both sides was<br />

not enough to guarantee that such weapons would never be used. It<br />

was also necessary to make the Soviet Union aware of the existence<br />

of the bombs through testing. He consistently advised against U.S.<br />

participation with the Soviet Union in a moratorium (period of<br />

waiting) on nuclear weapons testing. Largely based on Teller’s urging<br />

that underground testing be continued, the United States rejected<br />

a total moratorium in favor of the 1963 Atmospheric Test Ban<br />

Treaty.<br />

During the 1980’s, Teller, among others, convinced President<br />

Reagan to embrace the Strategic Defense Initiative (SDI). Teller argued<br />

that SDI components, such as the space-based “Excalibur,” a<br />

nuclear bomb-powered X-ray laser weapon proposed by the Lawrence-Livermore<br />

National Laboratory, would make thermonuclear<br />

war not unimaginable, but theoretically impossible.<br />

See also Airplane; Atomic bomb; Cruise missile; Rocket; Stealth<br />

aircraft; V-2 rocket.<br />

Further Reading<br />

Blumberg, Stanley A., <strong>and</strong> Louis G. Panos. Edward Teller, Giant of the<br />

Golden Age of Physics: A Biography. New York: Scribner’s, 1990.<br />

Clash, James M. “Teller Tells It.” Forbes (May 17, 1999).<br />

Teller, Edward, Wendy Teller, <strong>and</strong> Wilson Talley. Conversations on the<br />

Dark Secrets of Physics. New York: Plenum Press, 1991.<br />

York, Herbert E. The Advisors: Oppenheimer, Teller, <strong>and</strong> the Superbomb.<br />

Stanford, Calif.: Stanford University Press, 1989.


IBM Model 1401 computer<br />

IBM Model 1401 computer<br />

The invention: A relatively small, simple, <strong>and</strong> inexpensive computer<br />

that is often credited with having launched the personal<br />

computer age.<br />

The people behind the invention:<br />

Howard H. Aiken (1900-1973), an American mathematician<br />

Charles Babbage (1792-1871), an English mathematician <strong>and</strong><br />

inventor<br />

Herman Hollerith (1860-1929), an American inventor<br />

Computers: From the Beginning<br />

417<br />

Computers evolved into their modern form over a period of<br />

thous<strong>and</strong>s of years as a result of humanity’s efforts to simplify the<br />

process of counting. Two counting devices that are considered to be<br />

very simple, early computers are the abacus <strong>and</strong> the slide rule.<br />

These calculating devices are representative of digital <strong>and</strong> analog<br />

computers, respectively, because an abacus counts numbers of things,<br />

while the slide rule calculates length measurements.<br />

The first modern computer, which was planned by Charles Babbage<br />

in 1833, was never built. It was intended to perform complex<br />

calculations with a data processing/memory unit that was controlled<br />

by punched cards. In 1944, Harvard University’s Howard H.<br />

Aiken <strong>and</strong> the International Business Machines (IBM) Corporation<br />

built such a computer—the huge, punched-tape-controlled Automatic<br />

Sequence Controlled Calculator, or Mark I ASCC, which<br />

could perform complex mathematical operations in seconds. During<br />

the next fifteen years, computer advances produced digital computers<br />

that used binary arithmetic for calculation, incorporated<br />

simplified components that decreased the sizes of computers, had<br />

much faster calculating speeds, <strong>and</strong> were transistorized.<br />

Although practical computers had become much faster than<br />

they had been only a few years earlier, they were still huge <strong>and</strong> extremely<br />

expensive. In 1959, however, IBM introduced the Model<br />

1401 computer. Smaller, simpler, <strong>and</strong> much cheaper than the multi-


418 / IBM Model 1401 computer<br />

million-dollar computers that were available, the IBM Model 1401<br />

computer was also relatively easy to program <strong>and</strong> use. Its low cost,<br />

simplicity of operation, <strong>and</strong> very wide use have led many experts<br />

to view the IBM Model 1401 computer as beginning the age of the<br />

personal computer.<br />

Computer Operation <strong>and</strong> IBM’s Model 1401<br />

Modern computers are essentially very fast calculating machines<br />

that are capable of sorting, comparing, analyzing, <strong>and</strong> outputting information,<br />

as well as storing it for future use. Many sources credit<br />

Aiken’s Mark I ASCC as being the first modern computer to be built.<br />

This huge, five-ton machine used thous<strong>and</strong>s of relays to perform complex<br />

mathematical calculations in seconds. Soon after its introduction,<br />

other companies produced computers that were faster <strong>and</strong> more versatile<br />

than the Mark I. The computer development race was on.<br />

All these early computers utilized the decimal system for calculations<br />

until it was found that binary arithmetic, whose numbers are<br />

combinations of the binary digits 1 <strong>and</strong> 0, was much more suitable<br />

for the purpose. The advantage of the binary system is that the electronic<br />

switches that make up a computer (tubes, transistors, or<br />

chips) can be either on or off; in the binary system, the on state can<br />

be represented by the digit 1, the off state by the digit 0. Strung together<br />

correctly, binary numbers, or digits, can be inputted rapidly<br />

<strong>and</strong> used for high-speed computations. In fact, the computer term<br />

bit is a contraction of the phrase “binary digit.”<br />

A computer consists of input <strong>and</strong> output devices, a storage device<br />

(memory), arithmetic <strong>and</strong> logic units, <strong>and</strong> a control unit. In<br />

most cases, a central processing unit (CPU) combines the logic,<br />

arithmetic, memory, <strong>and</strong> control aspects. Instructions are loaded<br />

into the memory via an input device, processed, <strong>and</strong> stored. Then,<br />

the CPU issues comm<strong>and</strong>s to the other parts of the system to carry<br />

out computations or other functions <strong>and</strong> output the data as needed.<br />

Most output is printed as hard copy or displayed on cathode-ray<br />

tube monitors, or screens.<br />

The early modern computers—such as the Mark I ASCC—were<br />

huge because their information circuits were large relays or tubes.<br />

Computers became smaller <strong>and</strong> smaller as the tubes were replaced—


first with transistors, then with simple integrated circuits, <strong>and</strong> then<br />

with silicon chips. Each technological changeover also produced<br />

more powerful, more cost-effective computers.<br />

In the 1950’s, with reliable transistors available, IBM began the<br />

development of two types of computers that were completed by<br />

about 1959. The larger version was the Stretch computer, which was<br />

advertised as the most powerful computer of its day. Customized<br />

for each individual purchaser (for example, the Atomic Energy<br />

Commission), a Stretch computer cost $10 million or more. Some innovations<br />

in Stretch computers included semiconductor circuits,<br />

new switching systems that quickly converted various kinds of data<br />

into one language that was understood by the CPU, rapid data readers,<br />

<strong>and</strong> devices that seemed to anticipate future operations.<br />

Consequences<br />

IBM Model 1401 computer / 419<br />

The IBM Model 1401 was the first computer sold in very large<br />

numbers. It led IBM <strong>and</strong> other companies to seek to develop less expensive,<br />

more versatile, smaller computers that would be sold to<br />

small businesses <strong>and</strong> to individuals. Six years after the development<br />

of the Model 1401, other IBM models—<strong>and</strong> those made by<br />

other companies—became available that were more compact <strong>and</strong><br />

had larger memories. The search for compactness <strong>and</strong> versatility<br />

continued. A major development was the invention of integrated<br />

circuits by Jack S. Kilby of Texas Instruments; these integrated circuits<br />

became available by the mid-1960’s. They were followed by<br />

even smaller “microprocessors” (computer chips) that became available<br />

in the 1970’s. Computers continued to become smaller <strong>and</strong> more<br />

powerful.<br />

Input <strong>and</strong> storage devices also decreased rapidly in size. At first,<br />

the punched cards invented by Herman Hollerith, founder of the<br />

Tabulation Machine Company (which later became IBM), were read<br />

by bulky readers. In time, less bulky magnetic tapes <strong>and</strong> more compact<br />

readers were developed, after which magnetic disks <strong>and</strong> compact<br />

disc drives were introduced.<br />

Many other advances have been made. Modern computers can<br />

talk, create art <strong>and</strong> graphics, compose music, play games, <strong>and</strong> operate<br />

robots. Further advancement is expected as societal needs


420 / IBM Model 1401 computer<br />

change. Many experts believe that it was the sale of large numbers<br />

of IBM Model 1401 computers that began the trend.<br />

See also Apple II computer; BINAC computer; Colossus computer;<br />

ENIAC computer; Personal computer; Supercomputer;<br />

UNIVAC computer.<br />

Further Reading<br />

Carroll, Paul. Big Blues: The Unmaking of IBM. New York: Crown,<br />

1993.<br />

Chposky, James, <strong>and</strong> Ted Leonsis. Blue Magic: The People, Power, <strong>and</strong><br />

Politics Behind the IBM Personal Computer. New York: Facts on File,<br />

1988.<br />

Manes, Stephen, <strong>and</strong> Paul Andrews. Gates: How Microsoft’s Mogul<br />

Reinvented an Industry. New York: Doubleday, 1993.


In vitro plant culture<br />

In vitro plant culture<br />

The invention: Method for propagating plants in artificial media<br />

that has revolutionized agriculture.<br />

The people behind the invention:<br />

Georges Michel Morel (1916-1973), a French physiologist<br />

Philip Cleaver White (1913- ), an American chemist<br />

Plant Tissue Grows “In Glass”<br />

In the mid-1800’s, biologists began pondering whether a cell isolated<br />

from a multicellular organism could live separately if it were<br />

provided with the proper environment. In 1902, with this question in<br />

mind, the German plant physiologist Gottlieb Haberl<strong>and</strong>t attempted<br />

to culture (grow) isolated plant cells under sterile conditions on an artificial<br />

growth medium. Although his cultured cells never underwent<br />

cell division under these “in vitro” (in glass) conditions, Haberl<strong>and</strong>t<br />

is credited with originating the concept of cell culture.<br />

Subsequently, scientists attempted to culture plant tissues <strong>and</strong><br />

organs rather than individual cells <strong>and</strong> tried to determine the medium<br />

components necessary for the growth of plant tissue in vitro.<br />

In 1934, Philip White grew the first organ culture, using tomato<br />

roots. The discovery of plant hormones, which are compounds that<br />

regulate growth <strong>and</strong> development, was crucial to the successful culture<br />

of plant tissues; in 1939, Roger Gautheret, P. Nobécourt, <strong>and</strong><br />

White independently reported the successful culture of plant callus<br />

tissue. “Callus” is an irregular mass of dividing cells that often results<br />

from the wounding of plant tissue. Plant scientists were fascinated<br />

by the perpetual growth of such tissue in culture <strong>and</strong> spent<br />

years establishing optimal growth conditions <strong>and</strong> exploring the nutritional<br />

<strong>and</strong> hormonal requirements of plant tissue.<br />

Plants by the Millions<br />

421<br />

A lull in botanical research occurred during World War II, but<br />

immediately afterward there was a resurgence of interest in applying<br />

tissue culture techniques to plant research. Georges Morel, a


422 / In vitro plant culture<br />

plant physiologist at the National Institute for Agronomic Research<br />

in France, was one of many scientists during this time who<br />

had become interested in the formation of tumors in plants as well<br />

as in studying various pathogens such as fungi <strong>and</strong> viruses that<br />

cause plant disease.<br />

To further these studies, Morel adapted existing techniques in order<br />

to grow tissue from a wider variety of plant types in culture, <strong>and</strong><br />

he continued to try to identify factors that affected the normal<br />

growth <strong>and</strong> development of plants. Morel was successful in culturing<br />

tissue from ferns <strong>and</strong> was the first to culture monocot plants.<br />

Monocots have certain features that distinguish them from the other<br />

classes of seed-bearing plants, especially with respect to seed structure.<br />

More important, the monocots include the economically important<br />

species of grasses (the major plants of range <strong>and</strong> pasture)<br />

<strong>and</strong> cereals.<br />

For these cultures, Morel utilized a small piece of the growing tip<br />

of a plant shoot (the shoot apex) as the starting tissue material. This<br />

tissue was placed in a glass tube, supplied with a medium containing<br />

specific nutrients, vitamins, <strong>and</strong> plant hormones, <strong>and</strong> allowed<br />

to grow in the light. Under these conditions, the apex tissue grew<br />

roots <strong>and</strong> buds <strong>and</strong> eventually developed into a complete plant.<br />

Morel was able to generate whole plants from pieces of the shoot<br />

apex that were only 100 to 250 micrometers in length.<br />

Morel also investigated the growth of parasites such as fungi <strong>and</strong><br />

viruses in dual culture with host-plant tissue. Using results from<br />

these studies <strong>and</strong> culture techniques that he had mastered, Morel<br />

<strong>and</strong> his colleague Claude Martin regenerated virus-free plants from<br />

tissue that had been taken from virally infected plants. Tissues from<br />

certain tropical species, dahlias, <strong>and</strong> potato plants were used for the<br />

original experiments, but after Morel adapted the methods for the<br />

generation of virus-free orchids, plants that had previously been<br />

difficult to propagate by any means, the true significance of his<br />

work was recognized.<br />

Morel was the first to recognize the potential of the in vitro culture<br />

methods for the mass propagation of plants. He estimated that several<br />

million plants could be obtained in one year from a single small<br />

piece of shoot-apex tissue. Plants generated in this manner were<br />

clonal (genetically identical organisms prepared from a single plant).


With other methods of plant propagation, there is often a great variation<br />

in the traits of the plants produced, but as a result of Morel’s<br />

ideas, breeders could select for some desirable trait in a particular<br />

plant <strong>and</strong> then produce multiple clonal plants, all of which expressed<br />

the desired trait. The methodology also allowed for the production of<br />

virus-free plant material, which minimized both the spread of potential<br />

pathogens during shipping <strong>and</strong> losses caused by disease.<br />

Consequences<br />

In vitro plant culture / 423<br />

In vitro plant culture has been especially useful for species such as palm trees that cannot be<br />

propagated by other methods, such as by sowing seeds or grafting. (PhotoDisc)<br />

Variations on Morel’s methods are used to propagate plants used<br />

for human food consumption; plants that are sources of fiber, oil,<br />

<strong>and</strong> livestock feed; forest trees; <strong>and</strong> plants used in l<strong>and</strong>scaping <strong>and</strong><br />

in the floral industry. In vitro stocks are preserved under deepfreeze<br />

conditions, <strong>and</strong> disease-free plants can be proliferated quickly<br />

at any time of the year after shipping or storage.<br />

The in vitro multiplication of plants has been especially useful<br />

for species such as coconut <strong>and</strong> certain palms that cannot be propagated<br />

by other methods, such as by sowing seeds or grafting, <strong>and</strong><br />

has also become important in the preservation <strong>and</strong> propagation of


424 / In vitro plant culture<br />

rare plant species that might otherwise have become extinct. Many<br />

of these plants are sources of pharmaceuticals, oils, fragrances, <strong>and</strong><br />

other valuable products.<br />

The capability of regenerating plants from tissue culture has also<br />

been crucial in basic scientific research. Plant cells grown in culture<br />

can be studied more easily than can intact plants, <strong>and</strong> scientists have<br />

gained an in-depth underst<strong>and</strong>ing of plant physiology <strong>and</strong> biochemistry<br />

by using this method. This information <strong>and</strong> the methods<br />

of Morel <strong>and</strong> others have made possible the genetic engineering <strong>and</strong><br />

propagation of crop plants that are resistant to disease or disastrous<br />

environmental conditions such as drought <strong>and</strong> freezing. In vitro<br />

techniques have truly revolutionized agriculture.<br />

See also Artificial insemination; Cloning; Genetically engineered<br />

insulin; Rice <strong>and</strong> wheat strains.<br />

Further Reading<br />

Arbury, Jim, Richard Bird, Mike Honour, Clive Innes, <strong>and</strong> Mike Salmon.<br />

The Complete Book of Plant Propagation. Newtown, Conn.:<br />

Taunton Press, 1997.<br />

Clarke, Graham. The Complete Book of Plant Propagation. London:<br />

Seven Dials, 2001.<br />

Hartmann, Hudson T. Plant Propagation: Principles <strong>and</strong> Practices. 6th<br />

ed. London: Prentice-Hall, 1997.<br />

Heuser, Charles. The Complete Book of Plant Propagation. Newtown,<br />

Conn.: Taunton Press, 1997.


Infrared photography<br />

Infrared photography<br />

The invention: The first application of color to infrared photography,<br />

which performs tasks not possible for ordinary photography.<br />

The person behind the invention:<br />

Sir William Herschel (1738-1822), a pioneering English<br />

astronomer<br />

Invisible Light<br />

425<br />

Photography developed rapidly in the nineteenth century when it<br />

became possible to record the colors <strong>and</strong> shades of visible light on<br />

sensitive materials. Visible light is a form of radiation that consists of<br />

electromagnetic waves, which also make up other forms of radiation<br />

such as X rays <strong>and</strong> radio waves. Visible light occupies the range of<br />

wavelengths from about 400 nanometers (1 nanometer is 1 billionth<br />

of a meter) to about 700 nanometers in the electromagnetic spectrum.<br />

Infrared radiation occupies the range from about 700 nanometers<br />

to about 1,350 nanometers in the electromagnetic spectrum. Infrared<br />

rays cannot be seen by the human eye, but they behave in the<br />

same way that rays of visible light behave; they can be reflected, diffracted<br />

(broken), <strong>and</strong> refracted (bent).<br />

Sir William Herschel, a British astronomer, discovered infrared<br />

rays in 1800 by calculating the temperature of the heat that they produced.<br />

The term “infrared,” which was probably first used in 1800,<br />

was used to indicate rays that had wavelengths that were longer than<br />

those on the red end (the high end) of the spectrum of visible light but<br />

shorter than those of the microwaves, which appear higher on the<br />

electromagnetic spectrum. Infrared film is therefore sensitive to the<br />

infrared radiation that the human eye cannot see or record. Dyes that<br />

were sensitive to infrared radiation were discovered early in the<br />

twentieth century, but they were not widely used until the 1930’s. Because<br />

these dyes produced only black-<strong>and</strong>-white images, their usefulness<br />

to artists <strong>and</strong> researchers was limited. After 1930, however, a<br />

tidal wave of infrared photographic applications appeared.


426 / Infrared photography<br />

The Development of Color-Sensitive Infrared Film<br />

In the early 1940’s, military intelligence used infrared viewers for<br />

night operations <strong>and</strong> for gathering information about the enemy. One<br />

device that was commonly used for such purposes was called a<br />

“snooper scope.” Aerial photography with black-<strong>and</strong>-white infrared<br />

film was used to locate enemy hiding places <strong>and</strong> equipment. The images<br />

that were produced, however, often lacked clear definition.<br />

The development in 1942 of the first color-sensitive infrared film,<br />

Ektachrome Aero Film, became possible when researchers at the<br />

Eastman Kodak Company’s laboratories solved some complex chemical<br />

<strong>and</strong> physical problems that had hampered the development of<br />

color infrared film up to that point. Regular color film is sensitive to<br />

all visible colors of the spectrum; infrared color film is sensitive to<br />

violet, blue, <strong>and</strong> red light as well as to infrared radiation. Typical<br />

color film has three layers of emulsion, which are sensitized to blue,<br />

green, <strong>and</strong> red. Infrared color film, however, has its three emulsion<br />

layers sensitized to green, red, <strong>and</strong> infrared. Infrared wavelengths<br />

are recorded as reds of varying densities, depending on the intensity<br />

of the infrared radiation. The more infrared radiation there is,<br />

the darker the color of the red that is recorded.<br />

In infrared photography, a filter is placed over the camera lens to<br />

block the unwanted rays of visible light. The filter blocks visible <strong>and</strong><br />

ultraviolet rays but allows infrared radiation to pass. All three layers<br />

of infrared film are sensitive to blue, so a yellow filter is used. All<br />

blue radiation is absorbed by this filter.<br />

In regular photography, color film consists of three basic layers:<br />

the top layer is sensitive to blue light, the middle layer is sensitive to<br />

green, <strong>and</strong> the third layer is sensitive to red. Exposing the film to<br />

light causes a latent image to be formed in the silver halide crystals<br />

that make up each of the three layers. In infrared photography, color<br />

film consists of a top layer that is sensitive to infrared radiation, a<br />

middle layer sensitive to green, <strong>and</strong> a bottom layer sensitive to red.<br />

“Reversal processing” produces blue in the infrared-sensitive layer,<br />

yellow in the green-sensitive layer, <strong>and</strong> magenta in the red-sensitive<br />

layer. The blue, yellow, <strong>and</strong> magenta layers of the film produce the<br />

“false colors” that accentuate the various levels of infrared radiation<br />

shown as red in a color transparency, slide, or print.


Sir William Herschel<br />

Infrared photography / 427<br />

During his long career Sir William Herschel passed from human<br />

music to the music of the spheres, <strong>and</strong> in doing so revealed<br />

the invisible unlike any astronomer before him.<br />

He was born Friedrich Wilhelm Herschel in Hannover, Germany,<br />

in 1738. Like his brothers, he trained to be a musician in a<br />

local regimental b<strong>and</strong>. In 1757 he had to flee to Engl<strong>and</strong><br />

because his regiment was on the losing side of a<br />

war. Settling in the town of Bath, he supported himself<br />

with music, eventually becoming the organist for<br />

the city’s celebrated Octagon Chapel. He studied the<br />

music theory in Robert Smith’s book on harmonics<br />

<strong>and</strong>, discovering another book by Smith about optics<br />

<strong>and</strong> astronomy, read that too. He was immediately<br />

hooked. By 1773 he was assembling his own telescopes,<br />

<strong>and</strong> within ten years he had built the most powerful instruments<br />

in the l<strong>and</strong>. He interested King George III in astronomy<br />

<strong>and</strong> was rewarded with a royal pension that gave him the<br />

leisure to survey the heavens.<br />

Herschel looked deeper into space than anyone before him.<br />

He discovered thous<strong>and</strong>s of double stars <strong>and</strong> nebulae that had<br />

been invisible to astronomers with less powerful telescopes<br />

than his. He was the first person in recorded history to discover<br />

a planet—Uranus. While trying to learn the construction of the<br />

sun, he conducted hundreds of experiments with light. He<br />

found, unexpectedly, that he could feel heat from the sun even<br />

when visible light was filtered out, <strong>and</strong> concluded that some solar<br />

radiation—in this case infrared—was invisible to human<br />

eyes.<br />

Late in his career Herschel addressed the gr<strong>and</strong>est of all invisible<br />

aspects of the nature: the structure of the universe. His<br />

investigations led him to conclude that the nebulae he had so<br />

often observed were in themselves vast clouds of stars, very far<br />

away—they were galaxies. It was a key conceptual step in the<br />

development of modern cosmology.<br />

By the time Herschel died in 1822, he had trained his sister<br />

Caroline <strong>and</strong> his son John to carry on his work. Both became celebrated<br />

astronomers in their own right.<br />

(Library of Congess)


428 / Infrared photography<br />

The color of the dye that is formed in a particular layer bears no<br />

relationship to the color of light to which the layer is sensitive. If the<br />

relationship is not complementary, the resulting colors will be false.<br />

This means that objects whose colors appear to be similar to the<br />

human eye will not necessarily be recorded as similar colors on infrared<br />

film. A red rose with healthy green leaves will appear on infrared<br />

color film as being yellow with red leaves, because the chlorophyll<br />

contained in the plant leaf reflects infrared radiation <strong>and</strong><br />

causes the green leaves to be recorded as red. Infrared radiation<br />

from about 700 nanometers to about 900 nanometers on the electromagnetic<br />

spectrum can be recorded by infrared color film. Above<br />

900 nanometers, infrared radiation exists as heat patterns that must<br />

be recorded by nonphotographic means.<br />

Impact<br />

Infrared photography has proved to be valuable in many of the<br />

sciences <strong>and</strong> the arts. It has been used to create artistic images that<br />

are often unexpected visual explosions of everyday views. Because<br />

infrared radiation penetrates haze easily, infrared films are often<br />

used in mapping areas or determining vegetation types. Many<br />

cloud-covered tropical areas would be impossible to map without<br />

infrared photography. False-color infrared film can differentiate between<br />

healthy <strong>and</strong> unhealthy plants, so it is widely used to study insect<br />

<strong>and</strong> disease problems in plants. Medical research uses infrared<br />

photography to trace blood flow, detect <strong>and</strong> monitor tumor growth,<br />

<strong>and</strong> to study many other physiological functions that are invisible<br />

to the human eye.<br />

Some forms of cancer can be detected by infrared analysis before<br />

any other tests are able to perceive them. Infrared film is used in<br />

criminology to photograph illegal activities in the dark <strong>and</strong> to study<br />

evidence at crime scenes. Powder burns around a bullet hole, which<br />

are often invisible to the eye, show clearly on infrared film. In addition,<br />

forgeries in documents <strong>and</strong> works of art can often be seen<br />

clearly when photographed on infrared film. Archaeologists have<br />

used infrared film to locate ancient sites that are invisible in daylight.<br />

Wildlife biologists also document the behavior of animals at<br />

night with infrared equipment.


See also Autochrome plate; Color film; Fax machine; Instant<br />

photography.<br />

Further Reading<br />

Infrared photography / 429<br />

Collins, Douglas. The Story of Kodak. New York: Harry N. Abrams,<br />

1990.<br />

Cummins, Richard. “Infrared Revisited.” Petersen’s Photographic<br />

Magazine 23 (February, 1995).<br />

Paduano, Joseph. The Art of Infrared Photography. 4th ed. Buffalo,<br />

N.Y: Amherst Media, 1998.<br />

Richards, Dan. “The Strange Otherworld of Infrared.” Popular Photography<br />

62, no. 6 (June, 1998).<br />

White, Laurie. Infrared Photography H<strong>and</strong>book. Amherst, N.Y.: Amherst<br />

Media, 1995.


430<br />

Instant photography<br />

Instant photography<br />

The invention: Popularly known by its Polaroid tradename, a camera<br />

capable of producing finished photographs immediately after<br />

its film was exposed.<br />

The people behind the invention:<br />

Edwin Herbert L<strong>and</strong> (1909-1991), an American physicist <strong>and</strong><br />

chemist<br />

Howard G. Rogers (1915- ), a senior researcher at Polaroid<br />

<strong>and</strong> L<strong>and</strong>’s collaborator<br />

William J. McCune (1915- ), an engineer <strong>and</strong> head of the<br />

Polaroid team<br />

Ansel Adams (1902-1984), an American photographer <strong>and</strong><br />

L<strong>and</strong>’s technical consultant<br />

The Daughter of Invention<br />

Because he was a chemist <strong>and</strong> physicist interested primarily in<br />

research relating to light <strong>and</strong> vision, <strong>and</strong> to the materials that affect<br />

them, it was inevitable that Edwin Herbert L<strong>and</strong> should be drawn<br />

into the field of photography. L<strong>and</strong> founded the Polaroid Corporation<br />

in 1929. During the summer of 1943, while L<strong>and</strong> <strong>and</strong> his wife<br />

were vacationing in Santa Fe, New Mexico, with their three-yearold<br />

daughter, L<strong>and</strong> stopped to take a picture of the child. After the<br />

picture was taken, his daughter asked to see it. When she was told<br />

she could not see the picture immediately, she asked how long it<br />

would be. Within an hour after his daughter’s question, L<strong>and</strong> had<br />

conceived a preliminary plan for designing the camera, the film,<br />

<strong>and</strong> the physical chemistry of what would become the instant camera.<br />

Such a device would, he hoped, produce a picture immediately<br />

after exposure.<br />

Within six months, L<strong>and</strong> had solved most of the essential problems<br />

of the instant photography system. He <strong>and</strong> a small group of associates<br />

at Polaroid secretly worked on the project. Howard G. Rogers<br />

was L<strong>and</strong>’s collaborator in the laboratory. L<strong>and</strong> conferred the<br />

responsibility for the engineering <strong>and</strong> mechanical phase of the project<br />

on William J. McCune, who led the team that eventually de-


signed the original camera <strong>and</strong> the machinery that produced both<br />

the camera <strong>and</strong> L<strong>and</strong>’s new film.<br />

The first Polaroid L<strong>and</strong> camera—the Model 95—produced photographs<br />

measuring 8.25 by 10.8 centimeters; there were eight pictures<br />

to a roll. Rather than being black-<strong>and</strong>-white, the original Polaroid<br />

prints were sepia-toned (producing a warm, reddish-brown color).<br />

The reasons for the sepia coloration were chemical rather than aesthetic;<br />

as soon as L<strong>and</strong>’s researchers could devise a workable formula<br />

for sharp black-<strong>and</strong>-white prints (about ten months after the camera<br />

was introduced commercially), they replaced the sepia film.<br />

A Sophisticated Chemical Reaction<br />

Instant photography / 431<br />

Although the mechanical process involved in the first demonstration<br />

camera was relatively simple, this process was merely<br />

the means by which a highly sophisticated chemical reaction—<br />

the diffusion transfer process—was produced.<br />

In the basic diffusion transfer process, when an exposed negative<br />

image is developed, the undeveloped portion corresponds<br />

to the opposite aspect of the image, the positive. Almost all selfprocessing<br />

instant photography materials operate according to<br />

three phases—negative development, diffusion transfer, <strong>and</strong><br />

positive development. These occur simultaneously, so that positive<br />

image formation begins instantly. With black-<strong>and</strong>-white materials,<br />

the positive was originally completed in about sixty seconds; with<br />

color materials (introduced later), the process took somewhat longer.<br />

The basic phenomenon of silver in solution diffusing from one<br />

emulsion to another was first observed in the 1850’s, but no practical<br />

use of this action was made until 1939. The photographic use of<br />

diffusion transfer for producing normal-continuous-tone images<br />

was investigated actively from the early 1940’s by L<strong>and</strong> <strong>and</strong> his associates.<br />

The instant camera using this method was demonstrated<br />

in 1947 <strong>and</strong> marketed in 1948.<br />

The fundamentals of photographic diffusion transfer are simplest<br />

in a black-<strong>and</strong>-white peel-apart film. The negative sheet is exposed<br />

in the camera in the normal way. It is then pulled out of the<br />

camera, or film pack holder, by a paper tab. Next, it passes through a<br />

set of rollers, which press it face-to-face with a sheet of receiving ma-


432 / Instant photography<br />

Edwin H. L<strong>and</strong><br />

Born in Bridgeport, Connecticut in 1909, Edwin Herbert<br />

L<strong>and</strong> developed an obsession with color vision. As a boy, he<br />

slept with a copy of an optics textbook under his pillow. When<br />

he went to Harvard to study physics, he found the instruction<br />

too elementary <strong>and</strong> spent much of the time educating himself at<br />

the New York <strong>Public</strong> Library. While there, he thought of the first<br />

of his many sight-related inventions.<br />

He realized that by lining up tiny crystals <strong>and</strong> embedding<br />

them in clear plastic he could make a large, inexpensive light polarizer.<br />

He patented the idea for this “Polaroid” lens in 1929 (the<br />

first of more than five hundred patents) <strong>and</strong> in 1932 set up a commercial<br />

laboratory with his Harvard physics professor, George<br />

Wheelwright III. Five years later he opened the Polaroid Corporation<br />

in Boston to exploit the commercial potential of the lenses.<br />

They were to be used most famously as sunglasses, camera filters,<br />

eyeglasses for producing three-dimensional effects in movies,<br />

<strong>and</strong> glare-reduction screens for visual display terminals.<br />

In 1937, with Joseph Mallory, L<strong>and</strong> invented the vectograph—<br />

a device that superimposed two photographs in order to create<br />

a three-dimensional image. The invention dramatically improved<br />

the aerial photography during World War II <strong>and</strong> the Cold War.<br />

In fact, L<strong>and</strong> had a h<strong>and</strong> in designing both the camera carried<br />

aboard Lockheed’s U2 spyplane <strong>and</strong> the plane itself.<br />

While not busy running the Polaroid Corporation <strong>and</strong> overseeing<br />

development of its cameras, L<strong>and</strong> pursued his passion for<br />

experimenting with color <strong>and</strong> developed a widely respected theory<br />

of color vision. When he retired in 1982, he launched the<br />

Rowl<strong>and</strong> Institute for Science in Boston, once described as a cross<br />

between a private laboratory <strong>and</strong> a private art gallery. (L<strong>and</strong> had<br />

a deep interest in modern art.) He <strong>and</strong> other scientists there conducted<br />

research on artificial intelligence, genetics, microscopy,<br />

holography, protein dynamics, <strong>and</strong> color vision. L<strong>and</strong> died in<br />

1991 in Cambridge, Massachusetts, but the institute carries forward<br />

his legacy of scientific curiosity <strong>and</strong> practical application.<br />

terial included in the film pack. Simultaneously, the rollers rupture<br />

a pod of reagent chemicals that are spread evenly by the rollers<br />

between the two layers. The reagent contains a strong alkali <strong>and</strong> a<br />

silver halide solvent, both of which diffuse into the negative emul-


sion. There the alkali activates the developing agent, which immediately<br />

reduces the exposed halides to a negative image. At the<br />

same time, the solvent dissolves the unexposed halides. The silver<br />

in the dissolved halides forms the positive image.<br />

Impact<br />

The Polaroid L<strong>and</strong> camera had a tremendous impact on the photographic<br />

industry as well as on the amateur <strong>and</strong> professional photographer.<br />

Ansel Adams, who was known for his monumental,<br />

ultrasharp black-<strong>and</strong>-white panoramas of the American West, suggested<br />

to L<strong>and</strong> ways in which the tonal value of Polaroid film could<br />

be enhanced, as well as new applications for Polaroid photographic<br />

technology.<br />

Soon after it was introduced, Polaroid photography became part<br />

of the American way of life <strong>and</strong> changed the face of amateur photography<br />

forever. By the 1950’s, Americans had become accustomed<br />

to the world of recorded visual information through films, magazines,<br />

<strong>and</strong> newspapers; they also had become enthusiastic picturetakers<br />

as a result of the growing trend for simpler <strong>and</strong> more convenient<br />

cameras. By allowing these photographers not only to record<br />

their perceptions but also to see the results almost immediately, Polaroid<br />

brought people closer to the creative process.<br />

See also Autochrome plate; Brownie camera; Color film; Fax machine;<br />

Xerography.<br />

Further Reading<br />

Instant photography / 433<br />

Adams, Ansel. Polaroid L<strong>and</strong> Photography Manual. New York: Morgan<br />

& Morgan, 1963.<br />

Innovation/Imagination: Fifty Years of Polaroid Photography. New York:<br />

H. N. Abrams in association with the Friends of Photography,<br />

1999.<br />

McElheny, Victor K. Insisting on the Impossible: The Life of Edwin L<strong>and</strong>.<br />

Cambridge, Mass.: Perseus Books, 1998.<br />

Olshaker, Mark. The Instant Image. New York: Stein & Day, 1978.<br />

Wensberg, Peter C. L<strong>and</strong>’s Polaroid. Boston: Houghton Mifflin, 1987.


434<br />

Interchangeable parts<br />

Interchangeable parts<br />

The invention: A key idea in the late Industrial Revolution, the<br />

interchangeability of parts made possible mass production of<br />

identical products.<br />

The people behind the invention:<br />

Henry M. Lel<strong>and</strong> (1843-1932), president of Cadillac Motor Car<br />

Company in 1908, known as a master of precision<br />

Frederick Bennett, the British agent for Cadillac Motor Car<br />

Company who convinced the Royal Automobile Club to run<br />

the st<strong>and</strong>ardization test at Brookl<strong>and</strong>s, Engl<strong>and</strong><br />

Henry Ford (1863-1947), founder of Ford Motor Company who<br />

introduced the moving assembly line into the automobile<br />

industry in 1913<br />

An American Idea<br />

Mass production is a twentieth century methodology that for the<br />

most part is a result of nineteenth century ideas. It is a phenomenon<br />

that, although its origins were mostly American, has consequently<br />

changed the entire world. The use of interchangeable parts, the feasibility<br />

of which was demonstrated by the Cadillac Motor Car Company<br />

in 1908, was instrumental in making mass production possible.<br />

The British phase of the Industrial Revolution saw the application<br />

of division of labor, the first principle of industrialization, to capitalistdirected<br />

manufacturing processes. Centralized power sources were<br />

connected through shafts, pulleys, <strong>and</strong> belts to machines housed in<br />

factories. Even after these dramatic changes, the British preferred to<br />

produce unique, h<strong>and</strong>crafted products formed one step at a time using<br />

general-purpose machine tools. Seldom did they make separate components<br />

to be assembled into st<strong>and</strong>ardized products.<br />

Stories about American products that were assembled from fully<br />

interchangeable parts began to reach Great Britain. In 1851, the British<br />

public saw a few of these products on display at an exhibition in<br />

London’s Crystal Palace. In 1854, they were informed by one of their<br />

own investigative commissions that American manufacturers were


uilding military weapons <strong>and</strong> a number of consumer products<br />

with separately made parts that could be easily assembled, with little<br />

filing <strong>and</strong> fitting, by semiskilled workers.<br />

English industrialists had probably heard as much as they ever<br />

wanted to about this so-called “American system of manufacturing”<br />

by the first decade of the twentieth century, when word came<br />

that American companies were building automobiles with parts<br />

manufactured so precisely that they were interchangeable.<br />

The Cadillac<br />

Interchangeable parts / 435<br />

During the fall of 1907, Frederick Bennett, an Englishman who<br />

served as the British agent for the Cadillac Motor Car Company, paid<br />

a visit to the company’s Detroit, Michigan, factory <strong>and</strong> was amazed<br />

at what he saw. He later described the assembling of the relatively inexpensive<br />

Cadillac vehicles as a demonstration of the beauty <strong>and</strong><br />

practicality of precision. He was convinced that if his countrymen<br />

could see what he had seen they would also be impressed.<br />

Most automobile builders at the time claimed that their vehicles<br />

were built with h<strong>and</strong>crafted quality, yet at the same time they advertised<br />

that they could supply repair parts that would fit perfectly.<br />

In actuality, machining <strong>and</strong> filing were almost always required<br />

when parts were replaced, <strong>and</strong> only shops with proper equipment<br />

could do the job.<br />

Upon his return to London, Bennett convinced the Royal Automobile<br />

Club to sponsor a test of the precision of automobile parts. A<br />

st<strong>and</strong>ardization test was set to begin on February 29, 1908, <strong>and</strong> all of<br />

the companies then selling automobiles were invited to participate.<br />

Only the company that Bennett represented, Cadillac, was willing<br />

to enter the contest.<br />

Three one-cylinder Cadillacs, each painted a different color, were<br />

taken from stock at the company’s warehouse in London to a garage<br />

near the Brookl<strong>and</strong>s race track. The cars were first driven around<br />

the track ten times to prove that they were operable. British mechanics<br />

then dismantled the vehicles, placing their parts in piles in the<br />

center of the garage, making sure that there was no way of identifying<br />

from which car each internal piece came. Then, as a further test,<br />

eighty-nine r<strong>and</strong>omly selected parts were removed from the piles


436 / Interchangeable parts<br />

<strong>and</strong> replaced with new ones straight from Cadillac’s storeroom in<br />

London. The mechanics then proceeded to reassemble the automobiles,<br />

using only screwdrivers <strong>and</strong> wrenches.<br />

After the reconstruction, which took two weeks, the cars were<br />

driven from the garage. They were a motley looking trio, with fenders,<br />

doors, hoods, <strong>and</strong> wheels of mixed colors. All three were then<br />

driven five hundred miles around the Brookl<strong>and</strong>s track. The British<br />

were amazed. Cadillac was awarded the club’s prestigious Dewar<br />

Trophy, considered in the young automobile industry to be almost<br />

the equivalent of a Nobel Prize. A number of European <strong>and</strong> American<br />

automobile manufacturers began to consider the promise of interchangeable<br />

parts <strong>and</strong> the assembly line system.<br />

Henry M. Lel<strong>and</strong><br />

Cadillac’s precision-built automobiles were the result of a lifetime<br />

of experience of Henry M. Lel<strong>and</strong>, an American engineer.<br />

Known in Detroit at the turn of the century as a master of precision,<br />

Lel<strong>and</strong> became the primary connection between a series of nineteenth<br />

century attempts to make interchangeable parts <strong>and</strong> the<br />

large-scale use of precision parts in mass production manufacturing<br />

during the twentieth century.<br />

The first American use of truly interchangeable parts had occurred<br />

in the military, nearly three-quarters of a century before the<br />

test at Brookl<strong>and</strong>s. Thomas Jefferson had written from France about<br />

a demonstration of uniform parts for musket locks in 1785. A few<br />

years later, Eli Whitney attempted to make muskets for the American<br />

military by producing separate parts for assembly using specialized<br />

machines. He was never able to produce the precision necessary<br />

for truly interchangeable parts, but he promoted the idea<br />

intensely. It was in 1822 at the Harpers Ferry Armory in Virginia,<br />

<strong>and</strong> then a few years later at the Springfield Armory in Massachusetts,<br />

that the necessary accuracy in machining was finally achieved<br />

on a relatively large scale.<br />

Lel<strong>and</strong> began his career at the Springfield Armory in 1863, at the<br />

age of nineteen. He worked as a tool builder during the Civil War<br />

years <strong>and</strong> soon became an advocate of precision manufacturing. In<br />

1890, Lel<strong>and</strong> moved to Detroit, where he began a firm, Lel<strong>and</strong> &


Henry Martyn Lel<strong>and</strong><br />

Interchangeable parts / 437<br />

Henry Martyn Lel<strong>and</strong> (1843-1932) is the unsung giant of<br />

early automobile manufacturers, launching two of the bestknown<br />

American car companies, Cadillac <strong>and</strong> Lincoln, <strong>and</strong> influenced<br />

the success of General Motors, as well as introducing<br />

the use of interchangeable parts. Had he allowed a model to be<br />

named after him, as did Henry Ford <strong>and</strong> Ransom Olds, he<br />

might have become a household name too, but he refused any<br />

such suggestion.<br />

Lel<strong>and</strong> worked in factories during his youth. During the<br />

Civil War he honed his skills as a machinist at the U.S. Armory<br />

in Springfield, Massachusetts, helping build rifles with interchangeable<br />

parts. After the war, he learned how to machine<br />

parts to within one-thous<strong>and</strong>th of an inch, fabricated the first<br />

mechanical barber’s clippers, <strong>and</strong> refined the workings of air<br />

brakes for locomotives.<br />

This was all warm-up. In 1890 he moved to Detroit <strong>and</strong><br />

opened his own business, Lel<strong>and</strong> <strong>and</strong> Faulconer Manufacturing<br />

Company, specializing in automobile engines. The 10.25-horsepower<br />

engine he built for Olds in 1901 was rejected, but the single-cylinder<br />

(“one-lunger”) design that powered the first Cadillacs<br />

set him on the high road in the automotive industry. More<br />

innovations followed. He developed the electric starter, electric<br />

lights, <strong>and</strong> dimmable headlights. During World War I he built<br />

airplane engines for the U.S. government, <strong>and</strong> afterward converted<br />

the design for use in his new creation, the Lincoln.<br />

Throughout, he dem<strong>and</strong>ed precision from himself <strong>and</strong> those<br />

working for him. Once, for example, he complained to Alfred P.<br />

Sloan that a lot of ball bearings that Sloan had sold him varied<br />

from the required engineering tolerances <strong>and</strong> showed Sloan a<br />

few misshapen bearings to prove the claim. “Even though you<br />

make thous<strong>and</strong>s,” Lel<strong>and</strong> admonished Sloan, “the first <strong>and</strong> last<br />

should be precisely the same.” Sloan took the lesson very seriously.<br />

When he later led General Motors to the top of the industry,<br />

he credited Lel<strong>and</strong> with teaching him what mass production<br />

was all about.<br />

Faulconer, that would become internationally known for precision<br />

machining. His company did well supplying parts to the bicycle industry<br />

<strong>and</strong> internal combustion engines <strong>and</strong> transmissions to early


438 / Interchangeable parts<br />

automobile makers. In 1899, Lel<strong>and</strong> & Faulconer became the primary<br />

supplier of engines to the first of the major automobile producers,<br />

the Olds Motor Works.<br />

In 1902, the directors of another Detroit firm, the Henry Ford<br />

Company, found themselves in a desperate situation. Henry Ford,<br />

the company founder <strong>and</strong> chief engineer, had resigned after a disagreement<br />

with the firm’s key owner, William Murphy. Lel<strong>and</strong> was<br />

asked to take over the reorganization of the company. Because it<br />

could no longer use Ford’s name, the business was renamed in<br />

memory of the French explorer who had founded Detroit two hundred<br />

years earlier, Antoine de la Mothe Cadillac.<br />

Lel<strong>and</strong> was appointed president of the Cadillac Motor Car Company.<br />

The company, under his influence, soon became known for its<br />

precision manufacturing. He disciplined its suppliers, rejecting anything<br />

that did not meet his specifications, <strong>and</strong> insisted on precision<br />

machining for all parts. By 1906, Cadillac was outselling all of its<br />

competitors, including Oldsmobile <strong>and</strong> Ford’s new venture, the<br />

Ford Motor Company. After the Brookl<strong>and</strong>s demonstration in 1908,<br />

Cadillac became recognized worldwide for quality <strong>and</strong> interchangeability<br />

at a reasonable price.<br />

Impact<br />

The Brookl<strong>and</strong>s demonstration went a long way in proving that<br />

mass-produced goods could be durable <strong>and</strong> of relatively high quality.<br />

It showed that st<strong>and</strong>ardized products, although often less costly<br />

to make, were not necessarily cheap substitutes for h<strong>and</strong>crafted <strong>and</strong><br />

painstakingly fitted products. It also demonstrated that, through<br />

the use of interchangeable parts, the job of repairing such complex<br />

machines as automobiles could be made comparatively simple,<br />

moving maintenance <strong>and</strong> repair work from the well-equipped machine<br />

shop to the neighborhood garage or even to the home.<br />

Because of the international publicity Cadillac received, Lel<strong>and</strong>’s<br />

methods began to be emulated by others in the automobile industry.<br />

His precision manufacturing, as his daughter-in-law would later<br />

write in his biography, “laid the foundation for the future American<br />

[automobile] industry.” The successes of automobile manufacturers<br />

quickly led to the introduction of mass production methods, <strong>and</strong>


strategies designed to promote their necessary corollary mass consumption,<br />

in many other American businesses.<br />

In 1909, Cadillac was acquired by William Crapo Durant as the<br />

flagship company of his new holding company, which he labeled<br />

General Motors. Lel<strong>and</strong> continued to improve his production methods,<br />

while also influencing his colleagues in the other General Motors<br />

companies to implement many of his techniques. By the mid-<br />

1920’s, General Motors had become the world’s largest manufacturer<br />

of automobiles. Much of its success resulted from extensions<br />

of Lel<strong>and</strong>’s ideas. The company began offering a number of br<strong>and</strong><br />

name vehicles in a variety of price ranges for marketing purposes,<br />

while still keeping the costs of production down by including in<br />

each design a large number of commonly used, highly st<strong>and</strong>ardized<br />

components.<br />

Henry Lel<strong>and</strong> resigned from Cadillac during World War I after<br />

trying to convince Durant that General Motors should play an important<br />

part in the war effort by contracting to build Liberty aircraft<br />

engines for the military. He formed his own firm, named after his favorite<br />

president, Abraham Lincoln, <strong>and</strong> went on to build about four<br />

thous<strong>and</strong> aircraft engines in 1917 <strong>and</strong> 1918. In 1919, ready to make<br />

automobiles again, Lel<strong>and</strong> converted the Lincoln Motor Company<br />

into a car manufacturer. Again he influenced the industry by setting<br />

high st<strong>and</strong>ards for precision, but in 1921 an economic recession<br />

forced his new venture into receivership. Ironically, Lincoln was<br />

purchased at auction by Henry Ford. Lel<strong>and</strong> retired, his name overshadowed<br />

by those of individuals to whom he had taught the importance<br />

of precision <strong>and</strong> interchangeable parts. Ford, as one example,<br />

went on to become one of America’s industrial legends by<br />

applying the st<strong>and</strong>ardized parts concept.<br />

Ford <strong>and</strong> the Assembly Line<br />

Interchangeable parts / 439<br />

In 1913, Henry Ford, relying on the ease of fit made possible<br />

through the use of machined <strong>and</strong> stamped interchangeable parts,<br />

introduced the moving assembly line to the automobile industry.<br />

He had begun production of the Model T in 1908 using stationary<br />

assembly methods, bringing parts to assemblers. After having learned<br />

how to increase component production significantly, through experi-


440 / Interchangeable parts<br />

ments with interchangeable parts <strong>and</strong> moving assembly methods in<br />

the magneto department, he began to apply this same concept to final<br />

assembly. In the spring of 1913, Ford workers began dragging car<br />

frames past stockpiles of parts for assembly. Soon a power source<br />

was attached to the cars through a chain drive, <strong>and</strong> the vehicles<br />

were pulled past the stockpiles at a constant rate.<br />

From this time on, the pace of tasks performed by assemblers<br />

would be controlled by the rhythm of the moving line. As dem<strong>and</strong><br />

for the Model T increased, the number of employees along the line<br />

was increased <strong>and</strong> the jobs were broken into smaller <strong>and</strong> simpler<br />

tasks. With stationary assembly methods, the time required to assemble<br />

a Model T had averaged twelve <strong>and</strong> one-half person-hours.<br />

Dragging the chassis to the parts cut the time to six hours per vehicle,<br />

<strong>and</strong> the power-driven, constant-rate line produced a Model T<br />

with only ninety-three minutes of labor time. Because of these<br />

amazing increases in productivity, Ford was able to lower the selling<br />

price of the basic model from $900 in 1910 to $260 in 1925. He<br />

had revolutionized automobile manufacturing: The average family<br />

could now afford an automobile.<br />

Soon the average family would also be able to afford many of the<br />

other new products they had seen in magazines <strong>and</strong> newspapers.<br />

At the turn of the century, there were many new household appliances,<br />

farm machines, ready-made fashions, <strong>and</strong> prepackaged food<br />

products on the market, but only the wealthier class could afford<br />

most of these items. Major consumer goods retailers such as Sears,<br />

Roebuck <strong>and</strong> Company, Montgomery Ward, <strong>and</strong> the Great Atlantic<br />

<strong>and</strong> Pacific Tea Company were anxious to find lower-priced versions<br />

of these products to sell to a growing middle-class constituency.<br />

The methods of mass production that Henry Ford had popularized<br />

seemed to carry promise for these products as well. During<br />

the 1920’s, by working with such key manufacturers as Whirlpool,<br />

Hoover, General Electric, <strong>and</strong> Westinghouse, these large distributors<br />

helped introduce mass production methods into a large number<br />

of consumer product industries. They changed class markets<br />

into mass markets.<br />

The movement toward precision also led to the birth of a separate<br />

industry based on the manufacture of machine tools. A general<br />

purpose lathe, milling machine, or grinder could be used for a num-


er of operations, but mass production industries called for narrowpurpose<br />

machines designed for high-speed use in performing one<br />

specialized step in the production process. Many more machines<br />

were now required, one at each step in the production process. Each<br />

machine had to be simpler to operate, with more automatic features,<br />

because of an increased dependence on unskilled workers. The machine<br />

tool industry became the foundation of modern production.<br />

The miracle of mass production that followed, in products as<br />

diverse as airplanes, communication systems, <strong>and</strong> hamburgers,<br />

would not have been possible without the precision insisted upon<br />

by Henry Lel<strong>and</strong> in the first decade of the twentieth century. It<br />

would not have come about without the lessons learned by Henry<br />

Ford in the use of specialized machines <strong>and</strong> assembly methods, <strong>and</strong><br />

it would not have occurred without the growth of the machine tool<br />

industry. Cadillac’s demonstration at Brookl<strong>and</strong>s in 1908 proved<br />

the practicality of precision manufacturing <strong>and</strong> interchangeable<br />

parts to the world. It inspired American manufacturers to continue<br />

to develop these ideas; it convinced Europeans that such production<br />

was possible; <strong>and</strong>, for better or for worse, it played a major part<br />

in changing the world.<br />

See also CAD/CAM; Assembly line; Internal combustion engine.<br />

Further Reading<br />

Interchangeable parts / 441<br />

Hill, Frank Ernest. The Automobile: How It Came, Grew, <strong>and</strong> Has<br />

Changed Our Lives. New York: Dodd, Mead, 1967.<br />

Hounshell, David A. From the American System to Mass Production,<br />

1800-1932. Baltimore: Johns Hopkins University Press, 1984.<br />

Lel<strong>and</strong>, Ottilie M., <strong>and</strong> Minnie Dubbs Millbrook. Master of Precision:<br />

Henry M. Lel<strong>and</strong>. 1966. Reprint. Detroit: Wayne State University<br />

Press, 1996.<br />

Marcus, Alan I., <strong>and</strong> Howard P. Segal. Technology in America: A Brief<br />

History. Fort Worth, Texas: Harcourt Brace College, 1999.<br />

Nevins, Allan, <strong>and</strong> Frank Ernest Hill. The Times, the Man, the Company.<br />

Vol. 1 in Ford. New York: Charles Scribner’s Sons, 1954.


442<br />

Internal combustion engine<br />

Internal combustion engine<br />

The invention: The most common type of engine in automobiles<br />

<strong>and</strong> many other vehicles, the internal combusion engine is characterized<br />

by the fact that it burns its liquid fuelly internally—in<br />

contrast to engines, such as the steam engine, that burn fuel in external<br />

furnaces.<br />

The people behind the invention:<br />

Sir Harry Ralph Ricardo (1885-1974), an English engineer<br />

Oliver Thornycroft (1885-1956), an engineer <strong>and</strong> works manager<br />

Sir David R<strong>and</strong>all Pye (1886-1960), an engineer <strong>and</strong><br />

administrator<br />

Sir Robert Waley Cohen (1877-1952), a scientist <strong>and</strong> industrialist<br />

The Internal Combustion Engine: 1900-1916<br />

By the beginning of the twentieth century, internal combustion<br />

engines were almost everywhere. City streets in Berlin, London,<br />

<strong>and</strong> New York were filled with automobile <strong>and</strong> truck traffic; gasoline-<br />

<strong>and</strong> diesel-powered boat engines were replacing sails; stationary<br />

steam engines for electrical generation were being edged out by<br />

internal combustion engines. Even aircraft use was at h<strong>and</strong>: To<br />

progress from the Wright brothers’ first manned flight in 1903 to the<br />

fighting planes of World War I took only a little more than a decade.<br />

The internal combustion engines of the time, however, were<br />

primitive in design. They were heavy (10 to 15 pounds per output<br />

horsepower, as opposed to 1 to 2 pounds today), slow (typically<br />

1,000 or fewer revolutions per minute or less, as opposed to 2,000 to<br />

5,000 today), <strong>and</strong> extremely inefficient in extracting the energy content<br />

of their fuel. These were not major drawbacks for stationary applications,<br />

or even for road traffic that rarely went faster than 30 or<br />

40 miles per hour, but the advent of military aircraft <strong>and</strong> tanks dem<strong>and</strong>ed<br />

that engines be made more efficient.


Engine <strong>and</strong> Fuel Design<br />

Internal combustion engine / 443<br />

Harry Ricardo, son of an architect <strong>and</strong> gr<strong>and</strong>son (on his mother’s<br />

side) of an engineer, was a central figure in the necessary redesign of<br />

internal combustion engines. As a schoolboy, he built a coal-fired<br />

steam engine for his bicycle, <strong>and</strong> at Cambridge University he produced<br />

a single-cylinder gasoline motorcycle, incorporating many of<br />

his own ideas, which won a fuel-economy competition when it traveled<br />

almost 40 miles on a quart of gasoline. He also began development<br />

of a two-cycle engine called the “Dolphin,” which later was<br />

produced for use in fishing boats <strong>and</strong> automobiles. In fact, in 1911,<br />

Ricardo took his new bride on their honeymoon trip in a Dolphinpowered<br />

car.<br />

The impetus that led to major engine research came in 1916<br />

when Ricardo was an engineer in his family’s firm. The British<br />

government asked for newly designed tank engines, which had to<br />

operate in the dirt <strong>and</strong> mud of battle, at a tilt of up to 35 degrees,<br />

<strong>and</strong> could not give off telltale clouds of blue oil smoke. Ricardo<br />

solved the problem with a special piston design <strong>and</strong> with air circulation<br />

around the carburetor <strong>and</strong> within the engine to keep the oil<br />

cool.<br />

Design work on the tank engines turned Ricardo into a fullfledged<br />

research engineer. In 1917, he founded his own company,<br />

<strong>and</strong> a remarkable series of discoveries quickly followed. He investigated<br />

the problem of detonation of the fuel-air mixture in the internal<br />

combustion cylinder. The mixture is supposed to be ignited<br />

by the spark plug at the top of the compression stroke, with a controlled<br />

flame front spreading at a rate about equal to the speed of<br />

the piston head as it moves downward in the power stroke. Some<br />

fuels, however, detonated (ignited spontaneously throughout the<br />

entire fuel-air mixture) as a result of the compression itself, causing<br />

loss of fuel efficiency <strong>and</strong> damage to the engine.<br />

With the cooperation of Robert Waley Cohen of Shell Petroleum,<br />

Ricardo evaluated chemical mixtures of fuels <strong>and</strong> found that paraffins<br />

(such as n-heptane, the current low-octane st<strong>and</strong>ard) detonated<br />

readily, but aromatics such as toluene were nearly immune to detonation.<br />

He established a “toluene number” rating to describe the<br />

tendency of various fuels to detonate; this number was replaced in


444 / Internal combustion engine<br />

St<strong>and</strong>ard Four-Stroke Internal Combustion Engine<br />

Intake port Port<br />

Spark plug Plug Exhaust Exhaust Port port<br />

Intake Compression<br />

Ignition Expansion<br />

<strong>and</strong> Exhaust<br />

Intake Compression Power Exhaust<br />

The four cycles of a st<strong>and</strong>ard internal combustion engine (left to right): (1) intake, when air<br />

enters the cylinder <strong>and</strong> mixes with gasoline vapor; (2) compression, when the cylinder is<br />

sealed <strong>and</strong> the piston moves up to compress the air-fuel mixture; (3) power, when the spark<br />

plug ignites the mixture, creating more pressure that propels the piston downward; <strong>and</strong> (4)<br />

exhaust, when the burned gases exit the cylinder through the exhaust port.<br />

the 1920’s by the “octane number” devised by Thomas Midgley at<br />

the Delco laboratories in Dayton, Ohio.<br />

The fuel work was carried out in an experimental engine designed<br />

by Ricardo that allowed direct observation of the flame front<br />

as it spread <strong>and</strong> permitted changes in compression ratio while the<br />

engine was running. Three principles emerged from the investigation:<br />

the fuel-air mixture should be admitted with as much turbulence<br />

as possible, for thorough mixing <strong>and</strong> efficient combustion; the<br />

spark plug should be centrally located to prevent distant pockets of<br />

the mixture from detonating before the flame front reaches them;<br />

<strong>and</strong> the mixture should be kept as cool as possible to prevent detonation.<br />

These principles were then applied in the first truly efficient sidevalve<br />

(“L-head”) engine—that is, an engine with the valves in a<br />

chamber at the side of the cylinder, in the engine block, rather than<br />

overhead, in the engine head. Ricardo patented this design, <strong>and</strong> after<br />

winning a patent dispute in court in 1932, he received royalties<br />

or consulting fees for it from engine manufacturers all over the<br />

world.


Impact<br />

The side-valve engine was the workhorse design for automobile<br />

<strong>and</strong> marine engines until after World War II. With its valves actuated<br />

directly by a camshaft in the crankcase, it is simple, rugged,<br />

<strong>and</strong> easy to manufacture. Overhead valves with overhead camshafts<br />

are the st<strong>and</strong>ard in automobile engines today, but the sidevalve<br />

engine is still found in marine applications <strong>and</strong> in small engines<br />

for lawn mowers, home generator systems, <strong>and</strong> the like. In its<br />

widespread use <strong>and</strong> its decades of employment, the side-valve engine<br />

represents a scientific <strong>and</strong> technological breakthrough in the<br />

twentieth century.<br />

Ricardo <strong>and</strong> his colleagues, Oliver Thornycroft <strong>and</strong> D. R. Pye,<br />

went on to create other engine designs—notably, the sleeve-valve<br />

aircraft engine that was the basic pattern for most of the great British<br />

planes of World War II <strong>and</strong> early versions of the aircraft jet engine.<br />

For his technical advances <strong>and</strong> service to the government, Ricardo<br />

was elected a Fellow of the Royal Society in 1929, <strong>and</strong> he was<br />

knighted in 1948.<br />

See also Alkaline storage battery; Assembly line; Diesel locomotive;<br />

Dirigible; Gas-electric car; Interchangeable parts; Thermal cracking<br />

process.<br />

Further Reading<br />

Internal combustion engine / 445<br />

A History of the Automotive Internal Combustion Engine. Warrendale,<br />

Pa.: Society of Automotive Engineers, 1976.<br />

Mowery, David C., <strong>and</strong> Nathan Rosenberg. Paths of Innovation: Technological<br />

Change in Twentieth Century America. New York: Cambridge<br />

University Press, 1999.<br />

Ricardo, Harry R. Memories <strong>and</strong> Machines: The Pattern of My Life. London:<br />

Constable, 1968.


446<br />

The Internet<br />

The Internet<br />

The invention: A worldwide network of interlocking computer<br />

systems, developed out of a U.S. government project to improve<br />

military preparedness.<br />

The people behind the invention:<br />

Paul Baran, a researcher for the RAND corporation<br />

Vinton G. Cerf (1943- ), an American computer scientist<br />

regarded as the “father of the Internet”<br />

Cold War Computer Systems<br />

In 1957, the world was stunned by the launching of the satellite<br />

Sputnik I by the Soviet Union. The international image of the United<br />

States as the world’s technology superpower <strong>and</strong> its perceived edge<br />

in the Cold War were instantly brought into question. As part of the<br />

U.S. response, the Defense Department quickly created the Advanced<br />

Research Projects Agency (ARPA) to conduct research into<br />

“comm<strong>and</strong>, control, <strong>and</strong> communications” systems. Military planners<br />

in the Pentagon ordered ARPA to develop a communications<br />

network that would remain usable in the wake of a nuclear attack.<br />

The solution, proposed by Paul Baran, a scientist at the RAND Corporation,<br />

was the creation of a network of linked computers that<br />

could route communications around damage to any part of the system.<br />

Because the centralized control of data flow by major “hub”<br />

computers would make such a system vulnerable, the system could<br />

not have any central comm<strong>and</strong>, <strong>and</strong> all surviving points had to be<br />

able to reestablish contact following an attack on any single point.<br />

This redundancy of connectivity (later known as “packet switching”)<br />

would not monopolize a single circuit for communications, as<br />

telephones do, but would automatically break up computer messages<br />

into smaller packets, each of which could reach a destination<br />

by rerouting along different paths.<br />

ARPA then began attempting to link university computers over<br />

telephone lines. The historic connecting of four sites conducting<br />

ARPA research was accomplished in 1969 at a computer laboratory


at the University of California at Los Angeles (UCLA), which was<br />

connected to computers at the University of California at Santa<br />

Barbara, the Stanford Research Institute, <strong>and</strong> the University of Utah.<br />

UCLA graduate student Vinton Cerf played a major role in establishing<br />

the connection, which was first known as “ARPAnet.” By<br />

1971, more than twenty sites had been connected to the network, including<br />

supercomputers at the Massachusetts Institute of Technology<br />

<strong>and</strong> Harvard University; by 1981, there were more than two<br />

hundred computers on the system.<br />

The Development of the Internet<br />

The Internet / 447<br />

Because factors such as equipment failure, overtaxed telecommunications<br />

lines, <strong>and</strong> power outages can quickly reduce or abort<br />

(“crash”) computer network performance, the ARPAnet managers<br />

<strong>and</strong> others quickly sought to build still larger “internetting” projects.<br />

In the late 1980’s, the National Science Foundation built its<br />

own network of five supercomputer centers to give academic researchers<br />

access to high-power computers that had previously been<br />

available only to military contractors. The “NSFnet” connected university<br />

networks by linking them to the closest regional center; its<br />

development put ARPAnet out of commission in 1990. The economic<br />

savings that could be gained from the use of electronic mail<br />

(“e-mail”), which reduced postage <strong>and</strong> telephone costs, were motivation<br />

enough for many businesses <strong>and</strong> institutions to invest in<br />

hardware <strong>and</strong> network connections.<br />

The evolution of ARPAnet <strong>and</strong> NSFnet eventually led to the creation<br />

of the “Internet,” an international web of interconnected government,<br />

education, <strong>and</strong> business computer networks that has been<br />

called “the largest machine ever constructed.” Using appropriate<br />

software, a computer terminal or personal computer can send <strong>and</strong><br />

receive data via an “Internet Protocol” packet (an electronic envelope<br />

with an address). Communications programs on the intervening<br />

networks “read” the addresses on packets moving through the<br />

Internet <strong>and</strong> forward the packets toward their destinations. From<br />

approximately one thous<strong>and</strong> networks in the mid-1980’s, the Internet<br />

grew to an estimated thirty thous<strong>and</strong> connected networks by<br />

1994, with an estimated 25 million users accessing it regularly. The


448 / The Internet<br />

Vinton Cerf<br />

Although Vinton Cerf is widely hailed as the “father of the<br />

Internet,” he himself disavows that honor. He has repeatedly<br />

emphasized that the Internet was built on the work of countless<br />

others, <strong>and</strong> that he <strong>and</strong> his partner merely happened to make a<br />

crucial contribution at a turning point in Internet development.<br />

The path leading Cerf to the Internet began early. He was<br />

born in New Haven, Connecticut, in 1943. He read widely, devouring<br />

L. Frank Baum’s Oz books <strong>and</strong> science fiction novels—<br />

especially those dealing with real-science themes. When he was<br />

ten, a book called The Boy Scientist fired his interest in science.<br />

After starting high school in Los Angeles in 1958, he got his first<br />

glimpse of computers, which were very different devices in<br />

those days. During a visit to a Santa Monica lab, he inspected a<br />

computer filling three rooms with wires <strong>and</strong> vacuum tubes that<br />

analyzed data from a Canadian radar system built to detect<br />

sneak missile attacks from the Soviet Union. Two years later he<br />

<strong>and</strong> a friend began programming a paper-tape computer at<br />

UCLA while they were still in high school.<br />

After graduating from Stanford University in 1965 with a<br />

degree in computer science, Cerf worked for IBM for two years,<br />

then entered graduate school at UCLA. His work on multiprocessing<br />

computer systems got sidetracked when a Defense<br />

Department request came in asking for help on a packet-switching<br />

project. This new project drew him into the br<strong>and</strong>-new field<br />

of computer networking on a system that became known as the<br />

ARPAnet. In 1972 Cerf returned to Stanford as an assistant professor.<br />

There he <strong>and</strong> a colleague, Robert Kahn, developed the<br />

concepts <strong>and</strong> protocols that became the basis of the modern Internet—a<br />

term they coined in a paper they delivered in 1974.<br />

Afterward Cerf made development of the Internet the focus<br />

of his distinguished career, <strong>and</strong> he later moved back into the<br />

business world. In 1994 he returned to MCI as senior vice president<br />

of Internet architecture. Meanwhile, he founded the Internet<br />

Society in 1992 <strong>and</strong> the Internet Societal Task Force in 1999.<br />

majority of Internet users live in the United States <strong>and</strong> Europe, but<br />

the Internet has continued to exp<strong>and</strong> internationally as telecommunications<br />

lines are improved in other countries.


Impact<br />

Most individual users access the Internet through modems attached<br />

to their home personal computers by subscribing to local area<br />

networks. These services make information sources available such as<br />

on-line encyclopedias <strong>and</strong> magazines <strong>and</strong> embrace electronic discussion<br />

groups <strong>and</strong> bulletin boards on nearly every specialized interest<br />

area imaginable. Many universities converted large libraries to electronic<br />

form for Internet distribution, with an ambitious example being<br />

Cornell University’s conversion to electronic form of more than<br />

100,000 books on the development of America’s infrastructure.<br />

Numerous corporations <strong>and</strong> small businesses soon began to<br />

market their products <strong>and</strong> services over the Internet. Problems soon<br />

became apparent with the commercial use of the new medium,<br />

however, as the protection of copyrighted material proved to be difficult;<br />

data <strong>and</strong> other text available on the system can be “downloaded,”<br />

or electronically copied. To protect their resources from<br />

unauthorized use via the Internet, therefore, most companies set up<br />

a “firewall” computer to screen incoming communications.<br />

The economic policies of the Bill Clinton administration highlighted<br />

the development of the “information superhighway” for<br />

improving the delivery of social services <strong>and</strong> encouraging new<br />

businesses; however, many governmental agencies <strong>and</strong> offices, including<br />

the U.S. Senate <strong>and</strong> House of Representative, have been<br />

slow to install high-speed fiber-optic network links. Nevertheless,<br />

the Internet soon came to contain numerous information sites to improve<br />

public access to the institutions of government.<br />

See also Cell phone; Communications satellite; Fax machine;<br />

Personal computer.<br />

Further Reading<br />

The Internet / 449<br />

Abbate, Janet. Inventing the Internet. Cambridge, Mass.: MIT Press,<br />

2000.<br />

Brody, Herb. “Net Cerfing.” Technology Review (Cambridge, Mass.)<br />

101, no. 3 (May-June, 1998).<br />

Bryant, Stephen. The Story of the Internet. London: Peason Education,<br />

2000.


450 / The Internet<br />

Rodriguez, Karen. “Plenty Deserve Credit as ‘Father’ of the Internet.”<br />

Business Journal 17, no. 27 (October 22, 1999).<br />

Stefik, Mark J., <strong>and</strong> Vinton Cerf. Internet Dreams: Archetypes, Myths,<br />

<strong>and</strong> Metaphors. Cambridge, Mass.: MIT Press, 1997.<br />

“Vint Cerf.” Forbes 160, no. 7 (October 6, 1997).<br />

Wollinsky, Art. The History of the Internet <strong>and</strong> the World Wide Web.<br />

Berkeley Heights, N.J.: Enslow, 1999.


Iron lung<br />

Iron lung<br />

The invention: A mechanical respirator that saved the lives of victims<br />

of poliomyelitis.<br />

The people behind the invention:<br />

Philip Drinker (1894-1972), an engineer who made many<br />

contributions to medicine<br />

Louis Shaw (1886-1940), a respiratory physiologist who<br />

assisted Drinker<br />

Charles F. McKhann III (1898-1988), a pediatrician <strong>and</strong><br />

founding member of the American Board of Pediatrics<br />

A Terrifying Disease<br />

451<br />

Poliomyelitis (polio, or infantile paralysis) is an infectious viral<br />

disease that damages the central nervous system, causing paralysis<br />

in many cases. Its effect results from the destruction of neurons<br />

(nerve cells) in the spinal cord. In many cases, the disease produces<br />

crippled limbs <strong>and</strong> the wasting away of muscles. In others, polio results<br />

in the fatal paralysis of the respiratory muscles. It is fortunate<br />

that use of the Salk <strong>and</strong> Sabin vaccines beginning in the 1950’s has<br />

virtually eradicated the disease.<br />

In the 1920’s, poliomyelitis was a terrifying disease. Paralysis of<br />

the respiratory muscles caused rapid death by suffocation, often<br />

within only a few hours after the first signs of respiratory distress<br />

had appeared. In 1929, Philip Drinker <strong>and</strong> Louis Shaw, both of Harvard<br />

University, reported the development of a mechanical respirator<br />

that would keep those afflicted with the disease alive for indefinite<br />

periods of time. This device, soon nicknamed the “iron lung,”<br />

helped thous<strong>and</strong>s of people who suffered from respiratory paralysis<br />

as a result of poliomyelitis or other diseases.<br />

Development of the iron lung arose after Drinker, then an assistant<br />

professor in Harvard’s Department of Industrial Hygiene, was<br />

appointed to a Rockefeller Institute commission formed to improve<br />

methods for resuscitating victims of electric shock. The best-known<br />

use of the iron lung—treatment of poliomyelitis—was a result of<br />

numerous epidemics of the disease that occurred from 1898 until


452 / Iron lung<br />

the 1920’s, each leaving thous<strong>and</strong>s of Americans paralyzed.<br />

The concept of the iron lung reportedly arose from Drinker’s observation<br />

of physiological experiments carried out by Shaw <strong>and</strong><br />

Drinker’s brother, Cecil. The experiments involved the placement<br />

of a cat inside an airtight box—a body plethysmograph—with the<br />

cat’s head protruding from an airtight collar. Shaw <strong>and</strong> Cecil Drinker<br />

then measured the volume changes in the plethysmograph to identify<br />

normal breathing patterns. Philip Drinker then placed cats paralyzed<br />

by curare inside plethysmographies <strong>and</strong> showed that they<br />

could be kept breathing artificially by use of air from a hypodermic<br />

syringe connected to the device.<br />

Next, they proceeded to build a human-sized plethysmographlike<br />

machine, with a five-hundred-dollar grant from the New York<br />

Consolidated Gas Company. This was done by a tinsmith <strong>and</strong> the<br />

Harvard Medical School machine shop.<br />

Breath for Paralyzed Lungs<br />

The first machine was tested on Drinker <strong>and</strong> Shaw, <strong>and</strong> after several<br />

modifications were made, a workable iron lung was made<br />

available for clinical use. This machine consisted of a metal cylinder<br />

large enough to hold a human being. One end of the cylinder, which<br />

contained a rubber collar, slid out on casters along with a stretcher<br />

on which the patient was placed. Once the patient was in position<br />

<strong>and</strong> the collar was fitted around the patient’s neck, the stretcher was<br />

pushed back into the cylinder <strong>and</strong> the iron lung was made airtight.<br />

The iron lung then “breathed” for the patient by using an electric<br />

blower to remove <strong>and</strong> replace air alternatively inside the machine.<br />

In the human chest, inhalation occurs when the diaphragm contracts<br />

<strong>and</strong> powerful muscles (which are paralyzed in poliomyelitis<br />

sufferers) exp<strong>and</strong> the rib cage. This lowers the air pressure in the<br />

lungs <strong>and</strong> allows inhalation to occur. In exhalation, the diaphragm<br />

<strong>and</strong> chest muscles relax, <strong>and</strong> air is expelled as the chest cavity returns<br />

to its normal size. In cases of respiratory paralysis treated with<br />

an iron lung, the air coming into or leaving the iron lung alternately<br />

compressed the patient’s chest, producing artificial exhalation, <strong>and</strong><br />

the allowed it to exp<strong>and</strong> to so that the chest could fill with air. In this<br />

way, iron lungs “breathed” for the patients using them.


Careful examination of each patient was required to allow technicians<br />

to adjust the rate of operation of the machine. A cooling system<br />

<strong>and</strong> ports for drainage lines, intravenous lines, <strong>and</strong> the other<br />

apparatus needed to maintain a wide variety of patients were included<br />

in the machine.<br />

The first person treated in an iron lung was an eight-year-old girl<br />

afflicted with respiratory paralysis resulting from poliomyelitis. The<br />

iron lung kept her alive for five days. Unfortunately, she died from<br />

heart failure as a result of pneumonia. The next iron lung patient, a<br />

Harvard University student, was confined to the machine for several<br />

weeks <strong>and</strong> later recovered enough to resume a normal life.<br />

Impact<br />

The Drinker respirator, or iron lung, came into use in 1929 <strong>and</strong><br />

soon was considered indispensable, saving lives of poliomyelitis<br />

victims until the development of the Salk vaccine in the 1950’s.<br />

Although the iron lung is no longer used, it played a critical role<br />

in the development of modern respiratory care, proving that large<br />

numbers of patients could be kept alive with mechanical support.<br />

The iron lung <strong>and</strong> polio treatment began an entirely new era in<br />

treatment of respiratory conditions.<br />

In addition to receiving a number of awards <strong>and</strong> honorary degrees<br />

for his work, Drinker was elected president of the American<br />

Industrial Hygiene Association in 1942 <strong>and</strong> became chairman of<br />

Harvard’s Department of Industrial Hygiene.<br />

See also Electrocardiogram; Electroencephalogram; Heart-lung<br />

machine; Pacemaker; Polio vaccine (Sabin); Polio vaccine (Salk).<br />

Further Reading<br />

Iron lung / 453<br />

DeJauregui, Ruth. One Hundred Medical Milestones That Shaped World<br />

History. San Mateo, Calif.: Bluewood Books, 1998.<br />

Hawkins, Leonard C. The Man in the Iron Lung: The Frederick B. Snite,<br />

Jr., Story. Garden City, N.Y.: Doubleday, 1956.<br />

Rudulph, Mimi. Inside the Iron Lung. Buckinghamshire: Kensal<br />

Press, 1984.


454<br />

Laminated glass<br />

Laminated glass<br />

The invention: Double sheets of glass separated by a thin layer of<br />

plastic s<strong>and</strong>wiched between them.<br />

The people behind the invention:<br />

Edouard Benedictus (1879-1930), a French artist<br />

Katherine Burr Blodgett (1898-1979), an American physicist<br />

The Quest for Unbreakable Glass<br />

People have been fascinated for centuries by the delicate transparency<br />

of glass <strong>and</strong> the glitter of crystals. They have also been frustrated<br />

by the brittleness <strong>and</strong> fragility of glass. When glass breaks, it<br />

forms sharp pieces that can cut people severely. During the 1800’s<br />

<strong>and</strong> early 1900’s, a number of people demonstrated ways to make<br />

“unbreakable” glass. In 1855 in Engl<strong>and</strong>, the first “unbreakable”<br />

glass panes were made by embedding thin wires in the glass. The<br />

embedded wire grid held the glass together when it was struck or<br />

subjected to the intense heat of a fire. Wire glass is still used in windows<br />

that must be fire resistant. The concept of embedding the wire<br />

within a glass sheet so that the glass would not shatter was a predecessor<br />

of the concept of laminated glass.<br />

A series of inventors in Europe <strong>and</strong> the United States worked on<br />

the idea of using a durable, transparent inner layer of plastic between<br />

two sheets of glass to prevent the glass from shattering when it was<br />

dropped or struck by an impact. In 1899, Charles E. Wade of Scranton,<br />

Pennsylvania, obtained a patent for a kind of glass that had a sheet or<br />

netting of mica fused within it to bind it. In 1902, Earnest E. G. Street<br />

of Paris, France, proposed coating glass battery jars with pyroxylin<br />

plastic (celluloid) so that they would hold together if they cracked. In<br />

Swindon, Engl<strong>and</strong>, in 1905, John Crewe Wood applied for a patent<br />

for a material that would prevent automobile windshields from shattering<br />

<strong>and</strong> injuring people when they broke. He proposed cementing<br />

a sheet of material such as celluloid between two sheets of glass.<br />

When the window was broken, the inner material would hold the<br />

glass splinters together so that they would not cut anyone.


Katharine Burr Blodgett<br />

Remembering a Fortuitous Fall<br />

Laminated glass / 455<br />

Besides the danger of shattering, glass poses another problem.<br />

It reflects light, as much as 10 percent of the rays hitting it,<br />

<strong>and</strong> that is bad for many precision instruments. Katharine Burr<br />

Blodgett cleared away that problem.<br />

Blodgett was born in 1898 in Schenectady, New York, just<br />

months after her father died. Her widowed mother, intent upon<br />

giving her <strong>and</strong> her brother the best upbringing possible, devoted<br />

herself to their education <strong>and</strong> took them abroad to live for<br />

extended periods. She succeeded. Blodgett attended Bryn Mawr<br />

<strong>and</strong> then earned a master’s degree in physics from the University<br />

of Chicago. With the help of a family friend, Irving Langmuir,<br />

who later won a Nobel Prize in Chemistry, she was promised<br />

a job at the General Electric (GE) research laboratory.<br />

However, Langmuir first wanted her to study more physics.<br />

Blodgett went to Cambridge University <strong>and</strong> under the guidance<br />

of Ernest Rutherford became the first women to receive a<br />

doctorate in physics there. Then she went to work at GE.<br />

Collaborating with Langmuir, Blodgett found that she could<br />

coat glass with a film one layer of molecules at a time, a feat<br />

never accomplished before. Moreover, the color of light reflected<br />

differed with the number of layers of film. She discovered<br />

that by adjusting the number of layers she could cancel out<br />

the light reflected by the glass beneath, so as much as 99 percent<br />

of natural light would pass through the glass. Producing almost<br />

no reflection, this treated glass was “invisible.” It was perfect<br />

for lenses, such as those in cameras <strong>and</strong> microscopes. Blodgett<br />

also devised a way to measure the thickness of films based on<br />

the wavelengths of light they reflect—a color gauge—that became<br />

a st<strong>and</strong>ard laboratory technique.<br />

Blodgett died in the town of her birth in 1979.<br />

In his patent application, Edouard Benedictus described himself<br />

as an artist <strong>and</strong> painter. He was also a poet, musician, <strong>and</strong><br />

philosopher who was descended from the philosopher Baruch<br />

Benedictus Spinoza; he seemed an unlikely contributor to the<br />

progress of glass manufacture. In 1903, Benedictus was cleaning


456 / Laminated glass<br />

his laboratory when he dropped a glass bottle that held a nitrocellulose<br />

solution. The solvents, which had evaporated during the<br />

years that the bottle had sat on a shelf, had left a strong celluloid<br />

coating on the glass. When Benedictus picked up the bottle, he was<br />

surprised to see that it had not shattered: It was starred, but all the<br />

glass fragments had been held together by the internal celluloid<br />

coating. He looked at the bottle closely, labeled it with the date<br />

(November, 1903) <strong>and</strong> the height from which it had fallen, <strong>and</strong> put<br />

it back on the shelf.<br />

One day some years later (the date is uncertain), Benedictus became<br />

aware of vehicular collisions in which two young women received<br />

serious lacerations from broken glass. He wrote a poetic account<br />

of a daydream he had while he was thinking intently about<br />

the two women. He described a vision in which the faintly illuminated<br />

bottle that had fallen some years before but had not shattered<br />

appeared to float down to him from the shelf. He got up, went into<br />

his laboratory, <strong>and</strong> began to work on an idea that originated with his<br />

thoughts of the bottle that would not splinter.<br />

Benedictus found the old bottle <strong>and</strong> devised a series of experiments<br />

that he carried out until the next evening. By the time he had<br />

finished, he had made the first sheet of Triplex glass, for which he<br />

applied for a patent in 1909. He also founded the Société du Verre<br />

Triplex (The Triplex Glass Society) in that year. In 1912, the Triplex<br />

Safety Glass Company was established in Engl<strong>and</strong>. The company<br />

sold its products for military equipment in World War I, which began<br />

two years later.<br />

Triplex glass was the predecessor of laminated glass. Laminated<br />

glass is composed of two or more sheets of glass with a thin<br />

layer of plastic (usually polyvinyl butyral, although Benedictus<br />

used pyroxylin) laminated between the glass sheets using pressure<br />

<strong>and</strong> heat. The plastic layer will yield rather than rupture when subjected<br />

to loads <strong>and</strong> stresses. This prevents the glass from shattering<br />

into sharp pieces. Because of this property, laminated glass is also<br />

known as “safety glass.”<br />

Impact<br />

Even after the protective value of laminated glass was known,


the product was not widely used for some years. There were a number<br />

of technical difficulties that had to be solved, such as the discoloring<br />

of the plastic layer when it was exposed to sunlight; the relatively<br />

high cost; <strong>and</strong> the cloudiness of the plastic layer, which<br />

obscured vision—especially at night. Nevertheless, the exp<strong>and</strong>ing<br />

automobile industry <strong>and</strong> the corresponding increase in the number<br />

of accidents provided the impetus for improving the qualities <strong>and</strong><br />

manufacturing processes of laminated glass. In the early part of the<br />

century, almost two-thirds of all injuries suffered in automobile accidents<br />

involved broken glass.<br />

Laminated glass is used in many applications in which safety is<br />

important. It is typically used in all windows in cars, trucks, ships,<br />

<strong>and</strong> aircraft. Thick sheets of bullet-resistant laminated glass are<br />

used in banks, jewelry displays, <strong>and</strong> military installations. Thinner<br />

sheets of laminated glass are used as security glass in museums, libraries,<br />

<strong>and</strong> other areas where resistance to break-in attempts is<br />

needed. Many buildings have large ceiling skylights that are made<br />

of laminated glass; if the glass is damaged, it will not shatter, fall,<br />

<strong>and</strong> hurt people below. Laminated glass is used in airports, hotels,<br />

<strong>and</strong> apartments in noisy areas <strong>and</strong> in recording studios to reduce<br />

the amount of noise that is transmitted. It is also used in safety goggles<br />

<strong>and</strong> in viewing ports at industrial plants <strong>and</strong> test chambers.<br />

Edouard Benedictus’s recollection of the bottle that fell but did not<br />

shatter has thus helped make many situations in which glass is used<br />

safer for everyone.<br />

See also Buna rubber; Contact lenses; Neoprene; Plastic; Pyrex<br />

glass; Silicones.<br />

Further Reading<br />

Laminated glass / 457<br />

Eastman, Joel W. Styling vs. Safety: The American Automobile Industry<br />

<strong>and</strong> the Development of Automotive Safety, 1900-1966. Lanham: University<br />

Press of America, 1984.<br />

Fariss, Robert H. “Fifty Years of Safer Windshields.” CHEMTECH<br />

23, no. 9 (September, 1993).<br />

Miel, Rhoda. “New Process Promises Safer Glass.” Automotive News<br />

74, no. 5863 (February 28, 2000).


458 / Laminated glass<br />

Polak, James L. “Eighty Years Plus of Automotive Glass Development:<br />

Windshields Were Once an Option, Today They Are an Integral<br />

Part of the Automobile.” Automotive Engineering 98, no. 6<br />

(June, 1990).


Laser<br />

Laser<br />

The invention: Taking its name from the acronym for light amplification<br />

by the stimulated emission of radiation, a laser is a<br />

beam of electromagnetic radiation that is monochromatic, highly<br />

directional, <strong>and</strong> coherent. Lasers have found multiple applications<br />

in electronics, medicine, <strong>and</strong> other fields.<br />

The people behind the invention:<br />

Theodore Harold Maiman (1927- ), an American physicist<br />

Charles Hard Townes (1915- ), an American physicist who<br />

was a cowinner of the 1964 Nobel Prize in Physics<br />

Arthur L. Schawlow (1921-1999), an American physicist,<br />

cowinner of the 1981 Nobel Prize in Physics<br />

Mary Spaeth (1938- ), the American inventor of the tunable<br />

laser<br />

Coherent Light<br />

459<br />

Laser beams differ from other forms of electromagnetic radiation<br />

in being consisting of a single wavelength, being highly directional,<br />

<strong>and</strong> having waves whose crests <strong>and</strong> troughs are aligned. A laser<br />

beam launched from Earth has produced a spot a few kilometers<br />

wide on the Moon, nearly 400,000 kilometers away. Ordinary light<br />

would have spread much more <strong>and</strong> produced a spot several times<br />

wider than the Moon. Laser light can also be concentrated so as to<br />

yield an enormous intensity of energy, more than that of the surface<br />

of the Sun, an impossibility with ordinary light.<br />

In order to appreciate the difference between laser light <strong>and</strong> ordinary<br />

light, one must examine how light of any kind is produced. An<br />

ordinary light bulb contains atoms of gas. For the bulb to light up,<br />

these atoms must be excited to a state of energy higher then their<br />

normal, or ground, state. This is accomplished by sending a current<br />

of electricity through the bulb; the current jolts the atoms into the<br />

higher-energy state. This excited state is unstable, however, <strong>and</strong> the<br />

atoms will spontaneously return to their ground state by ridding<br />

themselves of excess energy.


460 / Laser<br />

Scanner device using a laser beam to read shelf labels. (PhotoDisc)<br />

As these atoms emit energy, light is produced. The light emitted<br />

by a lamp full of atoms is disorganized <strong>and</strong> emitted in all directions<br />

r<strong>and</strong>omly. This type of light, common to all ordinary sources, from<br />

fluorescent lamps to the Sun, is called “incoherent light.”<br />

Laser light is different. The excited atoms in a laser emit their excess<br />

energy in a unified, controlled manner. The atoms remain in the<br />

excited state until there are a great many excited atoms. Then, they<br />

are stimulated to emit energy, not independently, but in an organized<br />

fashion, with all their light waves traveling in the same direction,<br />

crests <strong>and</strong> troughs perfectly aligned. This type of light is called<br />

“coherent light.”<br />

Theory to Reality<br />

In 1958, Charles Hard Townes of Columbia University, together<br />

with Arthur L. Schawlow, explored the requirements of the laser in<br />

a theoretical paper. In the Soviet Union, F. A. Butayeva <strong>and</strong> V. A.<br />

Fabrikant had amplified light in 1957 using mercury; however, their<br />

work was not published for two years <strong>and</strong> was not published in a<br />

scientific journal. The work of the Soviet scientists, therefore, re-


Mary Spaeth<br />

Born in 1938, Mary Dietrich Spaeth, inventor of the tunable<br />

laser, learned to put things together early. When she was just<br />

three years old, her father began giving her tools to play with.<br />

She learned to use them well <strong>and</strong> got interested in science along<br />

the way. She studied mathematics <strong>and</strong> physics at Valparaiso<br />

University, graduating in 1960, <strong>and</strong> earned a master’s degree in<br />

nuclear physics from Wayne State University in 1962.<br />

The same year she joined Hughes Aircraft Company as a researcher.<br />

While waiting for supplies for her regular research in<br />

1966, she examined the lasers in her laboratory. She wondered<br />

if, by adding dyes, she could cause the beams to change colors.<br />

Cobbling together two lasers—one to boost the power of the<br />

test laser—with Duco cement, she added dyes <strong>and</strong> succeeded at<br />

once. She found that she could produce light in a wide range of<br />

colors with different dyes. The tunable dye laser afterward was<br />

used to separate isotopes in nuclear reactor fuel, to purify plutonium<br />

for weapons, <strong>and</strong> to boost the power of ground-based<br />

astronomical telescopes. She also invented a resonant reflector<br />

for ruby range finders <strong>and</strong> performed basic research on passive<br />

Q switches used in lasers.<br />

Because Spaeth considered Hughes’s promotion policies to<br />

discriminate against women scientists, she moved to the Lawrence<br />

Livermore National Laboratory in 1974. In 1986 she became<br />

the deputy associate director of its Laser Isotope Separation<br />

program.<br />

Laser / 461<br />

ceived virtually no attention in the Western world.<br />

In 1960, Theodore Harold Maiman constructed the first laser in<br />

the United States using a single crystal of synthetic pink ruby,<br />

shaped into a cylindrical rod about 4 centimeters long <strong>and</strong> 0.5 centimeter<br />

across. The ends, polished flat <strong>and</strong> made parallel to within<br />

about a millionth of a centimeter, were coated with silver to make<br />

them mirrors.<br />

It is a property of stimulated emission that stimulated light<br />

waves will be aligned exactly (crest to crest, trough to trough, <strong>and</strong><br />

with respect to direction) with the radiation that does the stimulating.<br />

From the group of excited atoms, one atom returns to its ground


462 / Laser<br />

state, emitting light. That light hits one of the other exited atoms <strong>and</strong><br />

stimulates it to fall to its ground state <strong>and</strong> emit light. The two light<br />

waves are exactly in step. The light from these two atoms hits other<br />

excited atoms, which respond in the same way, “amplifying” the total<br />

sum of light.<br />

If the first atom emits light in a direction parallel to the length of<br />

the crystal cylinder, the mirrors at both ends bounce the light waves<br />

back <strong>and</strong> forth, stimulating more light <strong>and</strong> steadily building up an<br />

increasing intensity of light. The mirror at one end of the cylinder is<br />

constructed to let through a fraction of the light, enabling the light to<br />

emerge as a straight, intense, narrow beam.<br />

Consequences<br />

When the laser was introduced, it was an immediate sensation. In<br />

the eighteen months following Maiman’s announcement that he had<br />

succeeded in producing a working laser, about four hundred companies<br />

<strong>and</strong> several government agencies embarked on work involving<br />

lasers. Activity centered on improving lasers, as well as on exploring<br />

their applications. At the same time, there was equal activity in publicizing<br />

the near-miraculous promise of the device, in applications covering<br />

the spectrum from “death” rays to sight-saving operations. A<br />

popular film in the James Bond series, Goldfinger (1964), showed the<br />

hero under threat of being sliced in half by a laser beam—an impossibility<br />

at the time the film was made because of the low power-output<br />

of the early lasers.<br />

In the first decade after Maiman’s laser, there was some disappointment.<br />

Successful use of lasers was limited to certain areas of<br />

medicine, such as repairing detached retinas, <strong>and</strong> to scientific applications,<br />

particularly in connection with st<strong>and</strong>ards: The speed of<br />

light was measured with great accuracy, as was the distance to the<br />

Moon. By 1990, partly because of advances in other fields, essentially<br />

all the laser’s promise had been fulfilled, including the death<br />

ray <strong>and</strong> James Bond’s slicer. Yet the laser continued to find its place<br />

in technologies not envisioned at the time of the first laser. For example,<br />

lasers are now used in computer printers, in compact disc<br />

players, <strong>and</strong> even in arterial surgery.


See also Atomic clock; Compact disc; Fiber-optics; Holography;<br />

Laser-diode recording process; Laser vaporization; Optical disk.<br />

Further Reading<br />

Laser / 463<br />

Townes, Charles H. How the Laser Happened: Adventures of a Scientist.<br />

New York: Oxford University Press, 1999.<br />

Weber, Robert L. Pioneers of Science: Nobel Prize Winners in Physics.2d<br />

ed. Philadelphia: A. Hilger, 1988.<br />

Yen, W. M., Marc D. Levenson, <strong>and</strong> Arthur L. Schawlow. Lasers,<br />

Spectroscopy, <strong>and</strong> New Ideas: A Tribute to Arthur L. Schawlow. New<br />

York: Springer-Verlag, 1987.


464<br />

Laser-diode recording process<br />

Laser-diode recording process<br />

The invention: Video <strong>and</strong> audio playback system that uses a lowpower<br />

laser to decode information digitally stored on reflective<br />

disks.<br />

The organization behind the invention:<br />

The Philips Corporation, a Dutch electronics firm<br />

The Development of Digital Systems<br />

Since the advent of the computer age, it has been the goal of<br />

many equipment manufacturers to provide reliable digital systems<br />

for the storage <strong>and</strong> retrieval of video <strong>and</strong> audio programs. A need<br />

for such devices was perceived for several reasons. Existing storage<br />

media (movie film <strong>and</strong> 12-inch, vinyl, long-playing records) were<br />

relatively large <strong>and</strong> cumbersome to manipulate <strong>and</strong> were prone to<br />

degradation, breakage, <strong>and</strong> unwanted noise. Thus, during the late<br />

1960’s, two different methods for storing video programs on disc<br />

were invented. A mechanical system was demonstrated by the<br />

Telefunken Company, while the Radio Corporation of America<br />

(RCA) introduced an electrostatic device (a device that used static<br />

electricity). The first commercially successful system, however, was<br />

developed during the mid-1970’s by the Philips Corporation.<br />

Philips devoted considerable resources to creating a digital video<br />

system, read by light beams, which could reproduce an entire feature-length<br />

film from one 12-inch videodisc. An integral part of this<br />

innovation was the fabrication of a device small enough <strong>and</strong> fast<br />

enough to read the vast amounts of greatly compacted data stored<br />

on the 12-inch disc without introducing unwanted noise. Although<br />

Philips was aware of the other formats, the company opted to use an<br />

optical scanner with a small “semiconductor laser diode” to retrieve<br />

the digital information. The laser diode is only a fraction of a millimeter<br />

in size, operates quite efficiently with high amplitude <strong>and</strong> relatively<br />

low power (0.1 watt), <strong>and</strong> can be used continuously. Because<br />

this configuration operates at a high frequency, its informationcarrying<br />

capacity is quite large.


Although the digital videodisc system (called “laservision”) works<br />

well, the low level of noise <strong>and</strong> the clear images offered by this system<br />

were masked by the low quality of the conventional television<br />

monitors on which they were viewed. Furthermore, the high price<br />

of the playback systems <strong>and</strong> the discs made them noncompetitive<br />

with the videocassette recorders (VCRs) that were then capturing<br />

the market for home systems. VCRs had the additional advantage<br />

that programs could be recorded or copied easily. The Philips Corporation<br />

turned its attention to utilizing this technology in an area<br />

where low noise levels <strong>and</strong> high quality would be more readily apparent—audio<br />

disc systems. By 1979, they had perfected the basic<br />

compact disc (CD) system, which soon revolutionized the world of<br />

stereophonic home systems.<br />

Reading Digital Discs with Laser Light<br />

Laser-diode recording process / 465<br />

Digital signals (signals composed of numbers) are stored on<br />

discs as “pits” impressed into the plastic disc <strong>and</strong> then coated with a<br />

thin reflective layer of aluminum. A laser beam, manipulated by<br />

delicate, fast-moving mirrors, tracks <strong>and</strong> reads the digital information<br />

as changes in light intensity. These data are then converted to a<br />

varying electrical signal that contains the video or audio information.<br />

The data are then recovered by means of a sophisticated<br />

pickup that consists of the semiconductor laser diode, a polarizing<br />

beam splitter, an objective lens, a collective lens system, <strong>and</strong> a<br />

photodiode receiver. The beam from the laser diode is focused by a<br />

collimator lens (a lens that collects <strong>and</strong> focuses light) <strong>and</strong> then<br />

passes through the polarizing beam splitter (PBS). This device acts<br />

like a one-way mirror mounted at 45 degrees to the light path. Light<br />

from the laser passes through the PBS as if it were a window, but the<br />

light emerges in a polarized state (which means that the vibration of<br />

the light takes place in only one plane). For the beam reflected from<br />

the CD surface, however, the PBS acts like a mirror, since the reflected<br />

beam has an opposite polarization. The light is thus deflected<br />

toward the photodiode detector. The objective lens is needed<br />

to focus the light onto the disc surface. On the outer surface of the<br />

transparent disc, the main spot of light has a diameter of 0.8 millimeter,<br />

which narrows to only 0.0017 millimeter at the reflective sur-


466 / Laser-diode recording process<br />

face. At the surface, the spot is about three times the size of the microscopic<br />

pits (0.0005 millimeter).<br />

The data encoded on the disc determine the relative intensity of<br />

the reflected light, on the basis of the presence or absence of pits.<br />

When the reflected laser beam enters the photodiode, a modulated<br />

light beam is changed into a digital signal that becomes an analog<br />

(continuous) audio signal after several stages of signal processing<br />

<strong>and</strong> error correction.<br />

Consequences<br />

The development of the semiconductor laser diode <strong>and</strong> associated<br />

circuitry for reading stored information has made CD audio<br />

systems practical <strong>and</strong> affordable. These systems can offer the quality<br />

of a live musical performance with a clarity that is undisturbed<br />

by noise <strong>and</strong> distortion. Digital systems also offer several other significant<br />

advantages over analog devices. The dynamic range (the<br />

difference between the softest <strong>and</strong> the loudest signals that can be<br />

stored <strong>and</strong> reproduced) is considerably greater in digital systems. In<br />

addition, digital systems can be copied precisely; the signal is not<br />

degraded by copying, as is the case with analog systems. Finally,<br />

error-correcting codes can be used to detect <strong>and</strong> correct errors in<br />

transmitted or reproduced digital signals, allowing greater precision<br />

<strong>and</strong> a higher-quality output sound.<br />

Besides laser video systems, there are many other applications<br />

for laser-read CDs. Compact disc read-only memory (CD-ROM) is<br />

used to store computer text. One st<strong>and</strong>ard CD can store 500 megabytes<br />

of information, which is about twenty times the storage of a<br />

hard-disk drive on a typical home computer. Compact disc systems<br />

can also be integrated with conventional televisions (called CD-V)<br />

to present twenty minutes of sound <strong>and</strong> five minutes of sound with<br />

picture. Finally, CD systems connected with a computer (CD-I) mix<br />

audio, video, <strong>and</strong> computer programming. These devices allow the<br />

user to stop at any point in the program, request more information,<br />

<strong>and</strong> receive that information as sound with graphics, film clips, or<br />

as text on the screen.<br />

See also Compact disc; Laser; Videocassette recorder; Walkman<br />

cassette player.


Further Reading<br />

Laser-diode recording process / 467<br />

Atkinson, Terry. “Picture This: CD’s with Video, By Christmas ‘87.”<br />

Los Angeles Times (February 20, 1987).<br />

Botez, Dan, <strong>and</strong> Luis Figueroa. Laser-Diode Technology <strong>and</strong> Applications<br />

II: 16-19 January 1990, Los Angeles, California. Bellingham,<br />

Wash.: SPIE, 1990.<br />

Clemens, Jon K. “Video Disks: Three Choices.” IEEE Spectrum 19,<br />

no. 3 (March, 1982).<br />

“Self-Pulsating Laser for DVD.” Electronics Now 67, no. 5 (May,<br />

1996).


468<br />

Laser eye surgery<br />

Laser eye surgery<br />

The invention: The first significant clinical ophthalmic application<br />

of any laser system was the treatment of retinal tears with a<br />

pulsed ruby laser.<br />

The people behind the invention:<br />

Charles J. Campbell (1926- ), an ophthalmologist<br />

H. Christian Zweng (1925- ), an ophthalmologist<br />

Milton M. Zaret (1927- ), an ophthalmologist<br />

Theodore Harold Maiman (1927- ), the physicist who<br />

developed the first laser<br />

Monkeys <strong>and</strong> Rabbits<br />

The term “laser” is an acronym for light amplification by the<br />

stimulated emission of radiation. The development of the laser for<br />

ophthalmic (eye surgery) surgery arose from the initial concentration<br />

of conventional light by magnifying lenses.<br />

Within a laser, atoms are highly energized. When one of these atoms<br />

loses its energy in the form of light, it stimulates other atoms to<br />

emit light of the same frequency <strong>and</strong> in the same direction. A cascade<br />

of these identical light waves is soon produced, which then oscillate<br />

back <strong>and</strong> forth between the mirrors in the laser cavity. One<br />

mirror is only partially reflective, allowing some of the laser light to<br />

pass through. This light can be concentrated further into a small<br />

burst of high intensity.<br />

On July 7, 1960, Theodore Harold Maiman made public his discovery<br />

of the first laser—a ruby laser. Shortly thereafter, ophthalmologists<br />

began using ruby lasers for medical purposes.<br />

The first significant medical uses of the ruby laser occurred in<br />

1961, with experiments on animals conducted by Charles J. Campbell<br />

in New York, H. Christian Zweng, <strong>and</strong> Milton M. Zaret. Zaret <strong>and</strong> his<br />

colleagues produced photocoagulation (a thickening or drawing together<br />

of substances by use of light) of the eyes of rabbits by flashes<br />

from a ruby laser. Sufficient energy was delivered to cause immediate<br />

thermal injury to the retina <strong>and</strong> iris of the rabbit. The beam also was


directed to the interior of the rabbit eye, resulting in retinal coagulations.<br />

The team examined the retinal lesions <strong>and</strong> pointed out both<br />

the possible advantages of laser as a tool for therapeutic photocoagulation<br />

<strong>and</strong> the potential applications in medical research.<br />

In 1962, Zweng, along with several of his associates, began experimenting<br />

with laser photocoagulation on the eyes of monkeys<br />

<strong>and</strong> rabbits in order to establish parameters for the use of lasers on<br />

the human eye.<br />

Reflected by Blood<br />

Laser eye surgery / 469<br />

The vitreous humor, a transparent jelly that usually fills the vitreous<br />

cavity of the eyes of younger individuals, commonly shrinks with age,<br />

with myopia, or with certain pathologic conditions. As these conditions<br />

occur, the vitreous humor begins to separate from the adjacent<br />

retina. In some patients, the separating vitreous humor produces a<br />

traction (pulling), causing a retinal tear to form. Through this opening in<br />

the retina, liquefied vitreous humor can pass to a site underneath the<br />

retina, producing retinal detachment <strong>and</strong> loss of vision.<br />

Alaser can be used to cause photocoagulation of a retinal tear. As a<br />

result, an adhesive scar forms between the retina surrounding the<br />

tear <strong>and</strong> the underlying layers so that, despite traction, the retina<br />

does not detach. If more than a small area of retina has detached, the<br />

laser often is ineffective <strong>and</strong> major retinal detachment surgery must<br />

be performed. Thus, in the experiments of Campbell <strong>and</strong> Zweng, the<br />

ruby laser was used to prevent, rather than treat, retinal detachment.<br />

In subsequent experiments with humans, all patients were treated<br />

with the experimental laser photocoagulator without anesthesia.<br />

Although usually no attempt was made to seal holes or tears, the<br />

diseased portions of the retina were walled off satisfactorily so that<br />

no detachments occurred. One problem that arose involved microaneurysms.<br />

A “microaneurysm” is a tiny aneurysm, or blood-filled<br />

bubble extending from the wall of a blood vessel. When attempts to<br />

obliterate microaneurysms were unsuccessful, the researchers postulated<br />

that the color of the ruby pulse so resembled the red of blood<br />

that the light was reflected rather than absorbed. They believed that<br />

another lasing material emitting light in another part of the spectrum<br />

might have performed more successfully.


470 / Laser eye surgery<br />

Previously, xenon-arc lamp photocoagulators had been used to<br />

treat retinal tears. The long exposure time required of these systems,<br />

combined with their broad spectral range emission (versus<br />

the single wavelength output of a laser), however, made the retinal<br />

spot on which the xenon-arc could be focused too large for many<br />

applications. Focused laser spots on the retina could be as small as<br />

50 microns.<br />

Consequences<br />

The first laser in ophthalmic use by Campbell, Zweng, <strong>and</strong> Zaret,<br />

among others, was a solid laser—Maiman’s ruby laser. While the results<br />

they achieved with this laser were more impressive than with<br />

the previously used xenon-arc, in the decades following these experiments,<br />

argon gas replaced ruby as the most frequently used material<br />

in treating retinal tears.<br />

Argon laser energy is delivered to the area around the retinal tear<br />

through a slit lamp or by using an intraocular probe introduced directly<br />

into the eye. The argon wavelength is transmitted through the<br />

clear structures of the eye, such as the cornea, lens, <strong>and</strong> vitreous.<br />

This beam is composed of blue-green light that can be effectively<br />

aimed at the desired portion of the eye. Nevertheless, the beam can<br />

be absorbed by cataracts <strong>and</strong> by vitreous or retinal blood, decreasing<br />

its effectiveness.<br />

Moreover, while the ruby laser was found to be highly effective<br />

in producing an adhesive scar, it was not useful in the treatment of<br />

vascular diseases of the eye. A series of laser sources, each with different<br />

characteristics, was considered, investigated, <strong>and</strong> used clinically<br />

for various durations during the period that followed Campbell<br />

<strong>and</strong> Zweng’s experiments.<br />

Other laser types that are being adapted for use in ophthalmology<br />

are carbon dioxide lasers for scleral surgery (surgery on the<br />

tough, white, fibrous membrane covering the entire eyeball except<br />

the area covered by the cornea) <strong>and</strong> eye wall resection, dye lasers to<br />

kill or slow the growth of tumors, eximer lasers for their ability to<br />

break down corneal tissue without heating, <strong>and</strong> pulsed erbium lasers<br />

used to cut intraocular membranes.


See also Contact lenses; Coronary artery bypass surgery; Laser;<br />

Laser vaporization.<br />

Further Reading<br />

Laser eye surgery / 471<br />

Constable, Ian J., <strong>and</strong> Arthur Siew Ming Lin. Laser: Its Clinical Uses<br />

in Eye Diseases. Edinburgh: Churchill Livingstone, 1981.<br />

Guyer, David R. Retina, Vitreous, Macula. Philadelphia: Saunders,<br />

1999.<br />

Hecht, Jeff. Laser Pioneers. Rev. ed. Boston: Academic Press, 1992.<br />

Smiddy, William E., Lawrence P. Chong, <strong>and</strong> Donald A. Frambach.<br />

Retinal Surgery <strong>and</strong> Ocular Trauma. Philadelphia: Lippincott, 1995.


472<br />

Laser vaporization<br />

Laser vaporization<br />

The invention: Technique using laser light beams to vaporize the<br />

plaque that clogs arteries.<br />

The people behind the invention:<br />

Albert Einstein (1879-1955), a theoretical American physicist<br />

Theodore Harold Maiman (1927- ), inventor of the laser<br />

Light, Lasers, <strong>and</strong> Coronary Arteries<br />

Visible light, a type of electromagnetic radiation, is actually a<br />

form of energy. The fact that the light beams produced by a light<br />

bulb can warm an object demonstrates that this is the case. Light<br />

beams are radiated in all directions by a light bulb. In contrast, the<br />

device called the “laser” produces light that travels in the form of a<br />

“coherent” unidirectional beam. Coherent light beams can be focused<br />

on very small areas, generating sufficient heat to melt steel.<br />

The term “laser” was invented in 1957 by R. Gordon Gould of<br />

Columbia University. It st<strong>and</strong>s for light amplification by stimulated<br />

emission of radiation, the means by which laser light beams are<br />

made. Many different materials—including solid ruby gemstones,<br />

liquid dye solutions, <strong>and</strong> mixtures of gases—can produce such<br />

beams in a process called “lasing.” The different types of lasers yield<br />

light beams of different colors that have many uses in science, industry,<br />

<strong>and</strong> medicine. For example, ruby lasers, which were developed<br />

in 1960, are widely used in eye surgery. In 1983, a group of<br />

physicians in Toulouse, France, used a laser for cardiovascular treatment.<br />

They used the laser to vaporize the “atheroma” material that<br />

clogs the arteries in the condition called “atherosclerosis.” The technique<br />

that they used is known as “laser vaporization surgery.”<br />

Laser Operation, Welding, <strong>and</strong> Surgery<br />

Lasers are electronic devices that emit intense beams of light<br />

when a process called “stimulated emission” occurs. The principles<br />

of laser operation, including stimulated emission, were established<br />

by Albert Einstein <strong>and</strong> other scientists in the first third of the twenti-


Laser vaporization / 473<br />

eth century. In 1960, Theodore H. Maiman of the Hughes Research<br />

Center in Malibu, California, built the first laser, using a ruby crystal<br />

to produce a laser beam composed of red light.<br />

All lasers are made up of three main components. The first of<br />

these, the laser’s “active medium,” is a solid (like Maiman’s ruby<br />

crystal), a liquid, or a gas that can be made to lase. The second component<br />

is a flash lamp or some other light energy source that puts<br />

light into the active medium. The third component is a pair of mirrors<br />

that are situated on both sides of the active medium <strong>and</strong> are designed<br />

in such a way that one mirror transmits part of the energy<br />

that strikes it, yielding the light beam that leaves the laser.<br />

Lasers can produce energy because light is one of many forms of<br />

energy that are called, collectively, electromagnetic radiation (among<br />

the other forms of electromagnetic radiation are X rays <strong>and</strong> radio<br />

waves). These forms of electromagnetic radiation have different wavelengths;<br />

the smaller the wavelength, the higher the energy level. The<br />

energy level is measured in units called “quanta.” The emission of<br />

light quanta from atoms that are said to be in the “excited state” produces<br />

energy, <strong>and</strong> the absorption of quanta by unexcited atoms—<br />

atoms said to be in the “ground state”—excites those atoms.<br />

The familiar light bulb spontaneously <strong>and</strong> haphazardly emits<br />

light of many wavelengths from excited atoms. This emission occurs<br />

in all directions <strong>and</strong> at widely varying times. In contrast, the<br />

light reflection between the mirrors at the ends of a laser causes all<br />

of the many excited atoms present in the active medium simultaneously<br />

to emit light waves of the same wavelength. This process is<br />

called “stimulated emission.”<br />

Stimulated emission ultimately causes a laser to yield a beam of<br />

coherent light, which means that the wavelength, emission time,<br />

<strong>and</strong> direction of all the waves in the laser beam are the same. The<br />

use of focusing devices makes it possible to convert an emitted laser<br />

beam into a point source that can be as small as a few thous<strong>and</strong>ths of<br />

an inch in diameter. Such focused beams are very hot, <strong>and</strong> they can<br />

be used for such diverse functions as cutting or welding metal objects<br />

<strong>and</strong> performing delicate surgery. The nature of the active medium<br />

used in a laser determines the wavelength of its emitted light<br />

beam; this in turn dictates both the energy of the emitted quanta <strong>and</strong><br />

the appropriate uses for the laser.


474 / Laser vaporization<br />

A blocked artery (top) can be threaded with a flexible fiber-optic fiber or bundle of fibers until<br />

it reaches the blockage; the fiber then emits laser light, vaporizing the plaque (bottom) <strong>and</strong><br />

restoring circulation.<br />

Maiman’s ruby laser, for example, has been used since the 1960’s<br />

in eye surgery to reattach detached retinas. This is done by focusing<br />

the laser on the tiny retinal tear that causes a retina to become detached.<br />

The very hot, high-intensity light beam then “welds” the<br />

retina back into place, bloodlessly, by burning it to produce scar tissue.<br />

The burning process has no effect on nearby tissues. Other<br />

types of lasers have been used in surgeries on the digestive tract <strong>and</strong><br />

the uterus since the 1970’s.<br />

In 1983, a group of physicians began using lasers to treat cardiovascular<br />

disease. The original work, which was carried out by a<br />

number of physicians in Toulouse, France, involved the vaporization<br />

of atheroma deposits (atherosclerotic plaque) in a human ar-


tery. This very exciting event added a new method to medical science’s<br />

arsenal of life-saving techniques.<br />

Consequences<br />

Since their discovery, lasers have been used for many purposes<br />

in science <strong>and</strong> industry. Such uses include the study of the laws of<br />

chemistry <strong>and</strong> physics, photography, communications, <strong>and</strong> surveying.<br />

Lasers have been utilized in surgery since the mid-1960’s, <strong>and</strong><br />

their use has had a tremendous impact on medicine. The first type<br />

of laser surgery to be conducted was the repair of detached retinas<br />

via ruby lasers. This technique has become the method of choice for<br />

such eye surgery because it takes only minutes to perform rather<br />

than the hours required for conventional surgical methods. It is also<br />

beneficial because the lasing of the surgical site cauterizes that site,<br />

preventing bleeding.<br />

In the late 1970’s, the use of other lasers for abdominal cancer<br />

surgery <strong>and</strong> uterine surgery began <strong>and</strong> flourished. In these<br />

forms of surgery, more powerful lasers are used. In the 1980’s,<br />

laser vaporization surgery (LVS) began to be used to clear atherosclerotic<br />

plaque (atheromas) from clogged arteries. This methodology<br />

gives cardiologists a useful new tool. Before LVS was<br />

available, surgeons dislodged atheromas by means of “transluminal<br />

angioplasty,” which involved pushing small, fluoroscopeguided<br />

inflatable balloons through clogged arteries.<br />

See also Blood transfusion; CAT scanner; Coronary artery bypass<br />

surgery; Electrocardiogram; Laser; Laser eye surgery; Ultrasound.<br />

Further Reading<br />

Laser vaporization / 475<br />

Fackelmann, Kathleen. “Internal Laser Blast Might Ease Heart<br />

Pain.” USA Today (March 8, 1999).<br />

Hecht, Jeff. Laser Pioneers. Rev. ed. Boston: Academic Press, 1992.<br />

“Is Cervical Laser Therapy Painful?” Lancet no. 8629 (January 14,<br />

1989).


476 / Laser vaporization<br />

Lothian, Cheri L. “Laser Angioplasty: Vaporizing Coronary Artery<br />

Plaque.” Nursing 22, no. 1 (January, 1992).<br />

“New Cool Laser Procedure Has Promise for Treating Blocked Coronary<br />

Arteries.” Wall Street Journal (May 15, 1989).<br />

Rundle, Rhonda L. “FDA Approves Laser Systems for Angioplasty.”<br />

Wall Street Journal (February 3, 1992).<br />

Sutton, C. J. G., <strong>and</strong> Michael P. Diamond. Endoscopic Surgery for Gynecologists.<br />

Philadelphia: W. B. Saunders, 1993.


Long-distance radiotelephony<br />

Long-distance radiotelephony<br />

The invention: The first radio transmissions from the United States<br />

to Europe opened a new era in telecommunications.<br />

The people behind the invention:<br />

Guglielmo Marconi (1874-1937), Italian inventor of transatlantic<br />

telegraphy<br />

Reginald Aubrey Fessenden (1866-1932), an American radio<br />

engineer<br />

Lee de Forest (1873-1961), an American inventor<br />

Harold D. Arnold (1883-1933), an American physicist<br />

John J. Carty (1861-1932), an American electrical engineer<br />

An Accidental Broadcast<br />

477<br />

The idea of commercial transatlantic communication was first<br />

conceived by Italian physicist <strong>and</strong> inventor Guglielmo Marconi, the<br />

pioneer of wireless telegraphy. Marconi used a spark transmitter to<br />

generate radio waves that were interrupted, or modulated, to form<br />

the dots <strong>and</strong> dashes of Morse code. The rapid generation of sparks<br />

created an electromagnetic disturbance that sent radio waves of different<br />

frequencies into the air—a broad, noisy transmission that was<br />

difficult to tune <strong>and</strong> detect.<br />

The inventor Reginald Aubrey Fessenden produced an alternative<br />

method that became the basis of radio technology in the twentieth<br />

century. His continuous radio waves kept to one frequency,<br />

making them much easier to detect at long distances. Furthermore,<br />

the continuous waves could be modulated by an audio signal, making<br />

it possible to transmit the sound of speech.<br />

Fessenden used an alternator to generate electromagnetic waves<br />

at the high frequencies required in radio transmission. It was specially<br />

constructed at the laboratories of the General Electric Company.<br />

The machine was shipped to Brant Rock, Massachusetts, in<br />

1906 for testing. Radio messages were sent to a boat cruising offshore,<br />

<strong>and</strong> the feasibility of radiotelephony was thus demonstrated.<br />

Fessenden followed this success with a broadcast of messages <strong>and</strong>


478 / Long-distance radiotelephony<br />

music between Brant Rock <strong>and</strong> a receiving station constructed at<br />

Plymouth, Massachusetts.<br />

The equipment installed at Brant Rock had a range of about 160<br />

kilometers. The transmission distance was determined by the strength<br />

of the electric power delivered by the alternator, which was measured<br />

in watts. Fessenden’s alternator was rated at 500 watts, but it<br />

usually delivered much less power.<br />

Yet this was sufficient to send a radio message across the Atlantic.<br />

Fessenden had built a receiving station at Machrihanish, Scotl<strong>and</strong>,<br />

to test the operation of a large rotary spark transmitter that he<br />

had constructed. An operator at this station picked up the voice of<br />

an engineer at Brant Rock who was sending instructions to Plymouth.<br />

Thus, the first radiotelephone message had been sent across<br />

the Atlantic by accident. Fessenden, however, decided not to make<br />

this startling development public. The station at Machrihanish was<br />

destroyed in a storm, making it impossible to carry out further tests.<br />

The successful transmission undoubtedly had been the result of exceptionally<br />

clear atmospheric conditions that might never again favor<br />

the inventor.<br />

One of the parties following the development of the experiments<br />

in radio telephony was the American Telephone <strong>and</strong> Telegraph<br />

(AT&T) Company. Fessenden entered into negotiations to sell his<br />

system to the telephone company, but, because of the financial panic<br />

of 1907, the sale was never made.<br />

Virginia to Paris <strong>and</strong> Hawaii<br />

The English physicist John Ambrose Fleming had invented a twoelement<br />

(diode) vacuum tube in 1904 that could be used to generate<br />

<strong>and</strong> detect radio waves. Two years later, the American inventor Lee<br />

de Forest added a third element to the diode to produce his “audion”<br />

(triode), which was a more sensitive detector. John J. Carty, head of a<br />

research <strong>and</strong> development effort at AT&T, examined these new devices<br />

carefully. He became convinced that an electronic amplifier, incorporating<br />

the triode into its design, could be used to increase the<br />

strength of telephone signals <strong>and</strong> to long distances.<br />

On Carty’s advice, AT&T purchased the rights to de Forest’s<br />

audion. A team of about twenty-five researchers, under the leader-


Reginald Aubrey Fessenden<br />

Long-distance radiotelephony / 479<br />

Reginald Aubrey Fessenden was born in Canada in 1866 to<br />

a small-town minister <strong>and</strong> his wife. After graduating from<br />

Bishop’s College in Lennoxville, Quebec, he took a job as head<br />

of Whitney Institute in Bermuda. However, he was brilliant <strong>and</strong><br />

volatile <strong>and</strong> had greater ambitions. After two years, he l<strong>and</strong>ed a<br />

job as a tester for his idol, Thomas Edison. Soon he was working<br />

as an engineer <strong>and</strong> chemist.<br />

Fessenden became a professor of electrical engineering at<br />

Purdue University in 1892 <strong>and</strong> then a year later at the University<br />

of Pittsburgh. His ideas were often advanced, so far advanced<br />

that some were not developed until much later, <strong>and</strong> by<br />

others. His first patented invention, an electrolyte detector in<br />

1900, was far more sensitive than others in use <strong>and</strong> made it possible<br />

to pick up radio signals carrying complex sound. To transmit<br />

such signals, he pioneered the use of carrier waves. During<br />

his career he registered more than three hundred patents.<br />

Suspicious <strong>and</strong> feisty, he also spent a lot of time in disputes,<br />

<strong>and</strong> frequently in court, over his inventions. He sued his backers<br />

at the National Electric Signaling Company over rights to<br />

operate a connection to Great Britain, <strong>and</strong> won a $406,000 settlement,<br />

which bankrupted the company. He sued Radio Corporation<br />

of America (RCA) claiming it prevented him from exploiting<br />

his own patents commercially. RCA settled out of court but<br />

was enriched by Fessenden’s invention.<br />

Having returned to Bermuda, Fessenden died in 1932. He<br />

never succeeded in winning the fame <strong>and</strong> wealth for the radio<br />

that he felt was due to him.<br />

ship of physicist Harold D. Arnold, were assigned the job of perfecting<br />

the triode <strong>and</strong> turning it into a reliable amplifier. The improved<br />

triode was responsible for the success of transcontinental cable telephone<br />

service, which was introduced in January, 1915. The triode<br />

was also the basis of AT&T’s foray into radio telephony.<br />

Carty’s research plan called for a system with three components:<br />

an oscillator to generate the radio waves, a modulator to add the<br />

audio signals to the waves, <strong>and</strong> an amplifier to transmit the radio<br />

waves. The total power output of the system was 7,500 watts,<br />

enough to send the radio waves over thous<strong>and</strong>s of kilometers.


480 / Long-distance radiotelephony<br />

The apparatus was installed in the U.S. Navy’s radio tower in<br />

Arlington, Virginia, in 1915. Radio messages from Arlington were<br />

picked up at a receiving station in California, a distance of 4,000 kilometers,<br />

then at a station in Pearl Harbor, Hawaii, which was 7,200<br />

kilometers from Arlington. AT&T’s engineers had succeeded in<br />

joining the company telephone lines with the radio transmitter at<br />

Arlington; therefore, the president of AT&T, Theodore Vail, could<br />

pick up his telephone <strong>and</strong> talk directly with someone in California.<br />

The next experiment was to send a radio message from Arlington<br />

to a receiving station set up in the Eiffel Tower in Paris. After several<br />

unsuccessful attempts, the telephone engineers in the Eiffel Tower<br />

finally heard Arlington’s messages on October 21, 1915. The AT&T<br />

receiving station in Hawaii also picked up the messages. The two receiving<br />

stations had to send their reply by telegraph to the United<br />

States because both stations were set up to receive only. Two-way<br />

radio communication was still years in the future.<br />

Impact<br />

The announcement that messages had been received in Paris was<br />

front-page news <strong>and</strong> brought about an outburst of national pride in<br />

the United States. The demonstration of transatlantic radio telephony<br />

was more important as publicity for AT&T than as a scientific<br />

advance. All the credit went to AT&T <strong>and</strong> to Carty’s laboratory.<br />

Both Fessenden <strong>and</strong> de Forest attempted to draw attention to their<br />

contributions to long-distance radio telephony, but to no avail. The<br />

Arlington-to-Paris transmission was a triumph for corporate public<br />

relations <strong>and</strong> corporate research.<br />

The development of the triode had been achieved with large<br />

teams of highly trained scientists—in contrast to the small-scale efforts<br />

of Fessenden <strong>and</strong> de Forest, who had little formal scientific<br />

training. Carty’s laboratory was an example of the new type of industrial<br />

research that was to dominate the twentieth century. The<br />

golden days of the lone inventor, in the mold of Thomas Edison or<br />

Alex<strong>and</strong>er Graham Bell, were gone.<br />

In the years that followed the first transatlantic radio telephone<br />

messages, little was done by AT&T to advance the technology or to<br />

develop a commercial service. The equipment used in the 1915 dem-


onstration was more a makeshift laboratory apparatus than a prototype<br />

for a new radio technology. The messages sent were short <strong>and</strong><br />

faint. There was a great gulf between hearing “hello” <strong>and</strong> “goodbye”<br />

amid the static. The many predictions of a direct telephone<br />

connection between New York <strong>and</strong> other major cities overseas were<br />

premature. It was not until 1927 that a transatlantic radio circuit was<br />

opened for public use. By that time, a new technological direction<br />

had been taken, <strong>and</strong> the method used in 1915 had been superseded<br />

by shortwave radio communication.<br />

See also Communications satellite; Internet; Long-distance telephone;<br />

Radio; Radio crystal sets; Radiotelephony; Television.<br />

Further Reading<br />

Long-distance radiotelephony / 481<br />

Marconi, Degna. My Father: Marconi. Toronto: Guernica Editions,<br />

1996.<br />

Masini, Giancarlo. Marconi. New York: Marsilio, 1995.<br />

Seitz, Frederick. The Cosmic Inventor: Reginald Aubrey Fessenden.<br />

Philadelphia: American Philosophical Society, 1999.<br />

Streissguth, Thomas. Communications: Sending the Message. Minneapolis,<br />

Minn.: Oliver Press, 1997.


482<br />

Long-distance telephone<br />

Long-distance telephone<br />

The invention: System for conveying voice signals via wires over<br />

long distances.<br />

The people behind the invention:<br />

Alex<strong>and</strong>er Graham Bell (1847-1922), a Scottish American<br />

inventor<br />

Thomas A. Watson (1854-1934), an American electrical engineer<br />

The Problem of Distance<br />

The telephone may be the most important invention of the nineteenth<br />

century. The device developed by Alex<strong>and</strong>er Graham Bell<br />

<strong>and</strong> Thomas A. Watson opened a new era in communication <strong>and</strong><br />

made it possible for people to converse over long distances for the<br />

first time. During the last two decades of the nineteenth century <strong>and</strong><br />

the first decade of the twentieth century, the American Telephone<br />

<strong>and</strong> Telegraph (AT&T) Company continued to refine <strong>and</strong> upgrade<br />

telephone facilities, introducing such innovations as automatic dialing<br />

<strong>and</strong> long-distance service.<br />

One of the greatest challenges faced by Bell engineers was to<br />

develop a way of maintaining signal quality over long distances.<br />

Telephone wires were susceptible to interference from electrical<br />

storms <strong>and</strong> other natural phenomena, <strong>and</strong> electrical resistance<br />

<strong>and</strong> radiation caused a fairly rapid drop-off in signal strength,<br />

which made long-distance conversations barely audible or unintelligible.<br />

By 1900, Bell engineers had discovered that signal strength could<br />

be improved somewhat by wrapping the main wire conductor with<br />

thinner wires called “loading coils” at prescribed intervals along<br />

the length of the cable. Using this procedure, Bell extended longdistance<br />

service from New York to Denver, Colorado, which was<br />

then considered the farthest point that could be reached with acceptable<br />

quality. The result, however, was still unsatisfactory, <strong>and</strong><br />

Bell engineers realized that some form of signal amplification would<br />

be necessary to improve the quality of the signal.


A breakthrough came in 1906, when Lee de Forest invented the<br />

“audion tube,” which could send <strong>and</strong> amplify radio waves. Bell scientists<br />

immediately recognized the potential of the new device for<br />

long-distance telephony <strong>and</strong> began building amplifiers that would<br />

be placed strategically along the long-distance wire network.<br />

Work progressed so quickly that by 1909, Bell officials were predicting<br />

that the first transcontinental long-distance telephone service,<br />

between New York <strong>and</strong> San Francisco, was imminent. In that<br />

year, Bell president Theodore N. Vail went so far as to promise the<br />

organizers of the Panama-Pacific Exposition, scheduled to open in<br />

San Francisco in 1914, that Bell would offer a demonstration at<br />

the exposition. The promise was risky, because certain technical<br />

problems associated with sending a telephone signal over a 4,800kilometer<br />

wire had not yet been solved. De Forest’s audion tube was<br />

a crude device, but progress was being made.<br />

Two more breakthroughs came in 1912, when de Forest improved<br />

on his original concept <strong>and</strong> Bell engineer Harold D. Arnold<br />

improved it further. Bell bought the rights to de Forest’s vacuumtube<br />

patents in 1913 <strong>and</strong> completed the construction of the New<br />

York-San Francisco circuit. The last connection was made at the<br />

Utah-Nevada border on June 17, 1914.<br />

Success Leads to Further Improvements<br />

Long-distance telephone / 483<br />

Bell’s long-distance network was tested successfully on June 29,<br />

1914, but the official demonstration was postponed until January<br />

25, 1915, to accommodate the Panama-Pacific Exposition, which<br />

had also been postponed. On that date, a connection was established<br />

between Jekyll Isl<strong>and</strong>, Georgia, where Theodore Vail was recuperating<br />

from an illness, <strong>and</strong> New York City, where Alex<strong>and</strong>er<br />

Graham Bell was st<strong>and</strong>ing by to talk to his former associate Thomas<br />

Watson, who was in San Francisco. When everything was in place,<br />

the following conversation took place. Bell: “Hoy! Hoy! Mr. Watson?<br />

Are you there? Do you hear me?” Watson: “Yes, Dr. Bell, I hear<br />

you perfectly. Do you hear me well?” Bell: “Yes, your voice is perfectly<br />

distinct. It is as clear as if you were here in New York.”<br />

The first transcontinental telephone conversation transmitted<br />

by wire was followed quickly by another that was transmitted via


484 / Long-distance telephone<br />

radio. Although the Bell company was slow to recognize the potential<br />

of radio wave amplification for the “wireless” transmission<br />

of telephone conversations, by 1909 the company had made a significant<br />

commitment to conduct research in radio telephony. On<br />

April 4, 1915, a wireless signal was transmitted by Bell technicians<br />

from Montauk Point on Long Isl<strong>and</strong>, New York, to Wilmington,<br />

Delaware, a distance of more than 320 kilometers. Shortly thereafter,<br />

a similar test was conducted between New York City <strong>and</strong><br />

Brunswick, Georgia, via a relay station at Montauk Point. The total<br />

distance of the transmission was more than 1,600 kilometers. Finally,<br />

in September, 1915, Vail placed a successful transcontinental radiotelephone<br />

call from his office in New York to Bell engineering chief<br />

J. J. Carty in San Francisco.<br />

Only a month later, the first telephone transmission across the<br />

Atlantic Ocean was accomplished via radio from Arlington, Virginia,<br />

to the Eiffel Tower in Paris, France. The signal was detectable,<br />

although its quality was poor. It would be ten years before true<br />

transatlantic radio-telephone service would begin.<br />

The Bell company recognized that creating a nationwide longdistance<br />

network would increase the volume of telephone calls simply<br />

by increasing the number of destinations that could be reached<br />

from any single telephone station. As the network exp<strong>and</strong>ed, each<br />

subscriber would have more reason to use the telephone more often,<br />

thereby increasing Bell’s revenues. Thus, the company’s strategy<br />

became one of tying local <strong>and</strong> regional networks together to create<br />

one large system.<br />

Impact<br />

Just as the railroads had interconnected centers of commerce, industry,<br />

<strong>and</strong> agriculture all across the continental United States in the<br />

nineteenth century, the telephone promised to bring a new kind of<br />

interconnection to the country in the twentieth century: instantaneous<br />

voice communication. During the first quarter century after<br />

the invention of the telephone <strong>and</strong> during its subsequent commercialization,<br />

the emphasis of telephone companies was to set up central<br />

office switches that would provide interconnections among<br />

subscribers within a fairly limited geographical area. Large cities


were wired quickly, <strong>and</strong> by the beginning of the twentieth century<br />

most were served by telephone switches that could accommodate<br />

thous<strong>and</strong>s of subscribers.<br />

The development of intercontinental telephone service was a<br />

milestone in the history of telephony for two reasons. First, it was a<br />

practical demonstration of the almost limitless applications of this<br />

innovative technology. Second, for the first time in its brief history,<br />

the telephone network took on a national character. It became clear<br />

that large central office networks, even in large cities such as New<br />

York, Chicago, <strong>and</strong> Baltimore, were merely small parts of a much<br />

larger, universally accessible communication network that spanned<br />

a continent. The next step would be to look abroad, to Europe <strong>and</strong><br />

beyond.<br />

See also Cell phone; Fax machine; Internet; Long-distance radiotelephony;<br />

Rotary dial telephone; Telephone switching; Touch-tone<br />

telephone.<br />

Further Reading<br />

Long-distance telephone / 485<br />

Coe, Lewis. The Telephone <strong>and</strong> Its Several <strong>Inventors</strong>: A History. Jefferson,<br />

N.C.: McFarl<strong>and</strong>, 1995.<br />

Mackay, James A. Alex<strong>and</strong>er Graham Bell: A Life. New York: J. Wiley,<br />

1997.<br />

Young, Peter. Person to Person: The International Impact of the Telephone.<br />

Cambridge: Granta Editions, 1991.


486<br />

Mammography<br />

Mammography<br />

The invention: The first X-ray procedure for detecting <strong>and</strong> diagnosing<br />

breast cancer.<br />

The people behind the invention:<br />

Albert Salomon, the first researcher to use X-ray technology<br />

instead of surgery to identify breast cancer<br />

Jacob Gershon-Cohen (1899-1971), a breast cancer researcher<br />

Studying Breast Cancer<br />

Medical researchers have been studying breast cancer for more<br />

than a century. At the end of the nineteenth century, however, no one<br />

knew how to detect breast cancer until it was quite advanced. Often,<br />

by the time it was detected, it was too late for surgery; many patients<br />

who did have surgery died. So after X-ray technology first appeared<br />

in 1896, cancer researchers were eager to experiment with it.<br />

The first scientist to use X-ray techniques in breast cancer experiments<br />

was Albert Salomon, a German surgeon. Trying to develop a<br />

biopsy technique that could tell which tumors were cancerous <strong>and</strong><br />

thereby avoid unnecessary surgery, he X-rayed more than three<br />

thous<strong>and</strong> breasts that had been removed from patients during breast<br />

cancer surgery. In 1913, he published the results of his experiments,<br />

showing that X rays could detect breast cancer. Different types of Xray<br />

images, he said, showed different types of cancer.<br />

Though Salomon is recognized as the inventor of breast radiology,<br />

he never actually used his technique to diagnose breast cancer.<br />

In fact, breast cancer radiology, which came to be known as “mammography,”<br />

was not taken up quickly by other medical researchers.<br />

Those who did try to reproduce his research often found that their<br />

results were not conclusive.<br />

During the 1920’s, however, more research was conducted in Leipzig,<br />

Germany, <strong>and</strong> in South America. Eventually, the Leipzig researchers,<br />

led by Erwin Payr, began to use mammography to diagnose<br />

cancer. In the 1930’s, a Leipzig researcher named W. Vogel<br />

published a paper that accurately described differences between<br />

cancerous <strong>and</strong> noncancerous tumors as they appeared on X-ray pho-


tographs. Researchers in the United States paid little attention to<br />

mammography until 1926. That year, a physician in Rochester, New<br />

York, was using a fluoroscope to examine heart muscle in a patient<br />

<strong>and</strong> discovered that the fluoroscope could be used to make images of<br />

breast tissue as well. The physician, Stafford L. Warren, then developed<br />

a stereoscopic technique that he used in examinations before<br />

surgery. Warren published his findings in 1930; his article also described<br />

changes in breast tissue that occurred because of pregnancy,<br />

lactation (milk production), menstruation, <strong>and</strong> breast disease. Yet<br />

Stafford’s technique was complicated <strong>and</strong> required equipment that<br />

most physicians of the time did not have. Eventually, he lost interest<br />

in mammography <strong>and</strong> went on to other research.<br />

Using the Technique<br />

Mammography / 487<br />

In the late 1930’s, Jacob Gershon-Cohen became the first clinician<br />

to advocate regular mammography for all women to detect breast<br />

cancer before it became a major problem. Mammography was not<br />

very expensive, he pointed out, <strong>and</strong> it was already quite accurate. A<br />

milestone in breast cancer research came in 1956, when Gershon-<br />

Cohen <strong>and</strong> others began a five-year study of more than 1,300 women<br />

to test the accuracy of mammography for detecting breast cancer.<br />

Each woman studied was screened once every six months. Of the<br />

1,055 women who finished the study, 92 were diagnosed with benign<br />

tumors <strong>and</strong> 23 with malignant tumors. Remarkably, out of all<br />

these, only one diagnosis turned out to be wrong.<br />

During the same period, Robert Egan of Houston began tracking<br />

breast cancer X rays. Over a span of three years, one thous<strong>and</strong> X-ray<br />

photographs were used to make diagnoses. When these diagnoses<br />

were compared to the results of surgical biopsies, it was confirmed<br />

that mammography had produced 238 correct diagnoses of cancer,<br />

out of 240 cases. Egan therefore joined the crusade for regular breast<br />

cancer screening.<br />

Once mammography was finally accepted by doctors in the late<br />

1950’s <strong>and</strong> early 1960’s, researchers realized that they needed a way<br />

to teach mammography quickly <strong>and</strong> effectively to those who would<br />

use it. A study was done, <strong>and</strong> it showed that any radiologist could<br />

conduct the procedure with only five days of training.


488 / Mammography<br />

In the early 1970’s, the American Cancer Society <strong>and</strong> the National<br />

Cancer Institute joined forces on a nationwide breast cancer<br />

screening program called the “Breast Cancer Detection Demonstration<br />

Project.” Its goal in 1971 was to screen more than 250,000<br />

women over the age of thirty-five.<br />

Since the 1960’s, however, some people had argued that mammography<br />

was dangerous because it used radiation on patients. In<br />

1976, Ralph Nader, a consumer advocate, stated that women who<br />

were to undergo mammography should be given consent forms<br />

that would list the dangers of radiation. In the years that followed,<br />

mammography was refined to reduced the amount of radiation<br />

needed to detect cancer. It became a st<strong>and</strong>ard tool for diagnosis, <strong>and</strong><br />

doctors recommended that women have a mammogram every two<br />

or three years after the age of forty.<br />

Impact<br />

Radiology is not a science that concerns only breast cancer screening.<br />

While it does provide the technical facilities necessary to practice<br />

mammography, the photographic images obtained must be interpreted<br />

by general practitioners, as well as by specialists. Once<br />

Physicians recommend that women have a mammogram every two or three years after the<br />

age of forty. (Digital Stock)


Gershon-Cohen had demonstrated the viability of the technique, a<br />

means of training was devised that made it fairly easy for clinicians<br />

to learn how to practice mammography successfully. Once all these<br />

factors—accuracy, safety, simplicity—were in place, mammography<br />

became an important factor in the fight against breast cancer.<br />

The progress made in mammography during the twentieth century<br />

was a major improvement in the effort to keep more women<br />

from dying of breast cancer. The disease has always been one of the<br />

primary contributors to the number of female cancer deaths that occur<br />

annually in the United States <strong>and</strong> around the world. This high<br />

figure stems from the fact that women had no way of detecting the<br />

disease until tumors were in an advanced state.<br />

Once Salomon’s procedure was utilized, physicians had a means<br />

by which they could look inside breast tissue without engaging in<br />

exploratory surgery, thus giving women a screening technique that<br />

was simple <strong>and</strong> inexpensive. By 1971, a quarter million women over<br />

age thirty-five had been screened. Twenty years later, that number<br />

was in the millions.<br />

See also Amniocentesis; CAT scanner; Electrocardiogram; Electroencephalogram;<br />

Holography; Nuclear magnetic resonance; Pap<br />

test; Syphilis test; Ultrasound.<br />

Further Reading<br />

Mammography / 489<br />

“First Digital Mammography System Approved by FDA.” FDA<br />

Consumer 34, no. 3 (May/June, 2000).<br />

Hindle, William H. Breast Care: A Clinical Guidebook for Women’s Primary<br />

Health Care Providers. New York: Springer, 1999.<br />

Okie, Susan. “More Women Are Getting Mammograms: Experts<br />

Agree That the Test Has Played Big Role in Reducing Deaths<br />

from Breast Cancer.” Washington Post (January 21, 1997).<br />

Wolbarst, Anthony B. Looking Within: How X-ray, CT, MRI, Ultrasound,<br />

<strong>and</strong> Other Medical Images Are Created, <strong>and</strong> How They Help<br />

Physicians Save Lives. Berkeley: University of California Press,<br />

1999.


490<br />

Mark I calculator<br />

Mark I calculator<br />

The invention: Early digital calculator designed to solve differential<br />

equations that was a forerunner of modern computers.<br />

The people behind the invention:<br />

Howard H. Aiken (1900-1973), Harvard University professor<br />

<strong>and</strong> architect of the Mark I<br />

Clair D. Lake (1888-1958), a senior engineer at IBM<br />

Francis E. Hamilton (1898-1972), an IBM engineer<br />

Benjamin M. Durfee (1897-1980), an IBM engineer<br />

The Human Computer<br />

The physical world can be described by means of mathematics.<br />

In principle, one can accurately describe nature down to the smallest<br />

detail. In practice, however, this is impossible except for the simplest<br />

of atoms. Over the years, physicists have had great success in<br />

creating simplified models of real physical processes whose behavior<br />

can be described by the branch of mathematics called “calculus.”<br />

Calculus relates quantities that change over a period of time. The<br />

equations that relate such quantities are called “differential equations,”<br />

<strong>and</strong> they can be solved precisely in order to yield information<br />

about those quantities. Most natural phenomena, however, can<br />

be described only by differential equations that can be solved only<br />

approximately. These equations are solved by numerical means that<br />

involve performing a tremendous number of simple arithmetic operations<br />

(repeated additions <strong>and</strong> multiplications). It has been the<br />

dream of many scientists since the late 1700’s to find a way to automate<br />

the process of solving these equations.<br />

In the early 1900’s, people who spent day after day performing the<br />

tedious operations that were required to solve differential equations<br />

were known as “computers.” During the two world wars, these human<br />

computers created ballistics tables by solving the differential<br />

equations that described the hurling of projectiles <strong>and</strong> the dropping<br />

of bombs from aircraft. The war effort was largely responsible for accelerating<br />

the push to automate the solution to these problems.


A Computational Behemoth<br />

Mark I calculator / 491<br />

The ten-year period from 1935 to 1945 can be considered the<br />

prehistory of the development of the digital computer. (In a digital<br />

computer, digits represent magnitudes of physical quantities.<br />

These digits can have only certain values.) Before this time, all<br />

machines for automatic calculation were either analog in nature<br />

(in which case, physical quantities such as current or voltage represent<br />

the numerical values of the equation <strong>and</strong> can vary in a continuous<br />

fashion) or were simplistic mechanical or electromechanical<br />

adding machines.<br />

This was the situation that faced Howard Aiken. At the time, he<br />

was a graduate student working on his doctorate in physics. His<br />

dislike for the tremendous effort required to solve the differential<br />

equations used in his thesis drove him to propose, in the fall of 1937,<br />

constructing a machine that would automate the process. He proposed<br />

taking existing business machines that were commonly used<br />

in accounting firms <strong>and</strong> combining them into one machine that<br />

would be controlled by a series of instructions. One goal was to<br />

eliminate all manual intervention in the process in order to maximize<br />

the speed of the calculation.<br />

Aiken’s proposal came to the attention of Thomas Watson, who<br />

was then the president of International Business Machines Corporation<br />

(IBM). At that time, IBM was a major supplier of business machines<br />

<strong>and</strong> did not see much of a future in such “specialized” machines.<br />

It was the pressure provided by the computational needs of<br />

the military in World War II that led IBM to invest in building automated<br />

calculators. In 1939, a contract was signed in which IBM<br />

agreed to use its resources (personnel, equipment, <strong>and</strong> finances) to<br />

build a machine for Howard Aiken <strong>and</strong> Harvard University.<br />

IBM brought together a team of seasoned engineers to fashion a<br />

working device from Aiken’s sketchy ideas. Clair D. Lake, who was<br />

selected to manage the project, called on two talented engineers—<br />

Francis E. Hamilton <strong>and</strong> Benjamin M. Durfee—to assist him.<br />

After four years of effort, which was interrupted at times by the<br />

dem<strong>and</strong>s of the war, a machine was constructed that worked remarkably<br />

well. Completed in January, 1943, at Endicott, New York,<br />

it was then disassembled <strong>and</strong> moved to Harvard University in Cam-


492 / Mark I calculator<br />

bridge, Massachusetts, where it was reassembled. Known as the IBM<br />

automatic sequence controlled calculator (ASCC), it began operation<br />

in the spring of 1944 <strong>and</strong> was formally dedicated <strong>and</strong> revealed to the<br />

public on August 7, 1944. Its name indicates the machine’s distinguishing<br />

feature: the ability to load automatically the instructions<br />

that control the sequence of the calculation. This capability was provided<br />

by punching holes, representing the instructions, in a long,<br />

ribbonlike paper tape that could be read by the machine.<br />

Computers of that era were big, <strong>and</strong> the ASCC I was particularly<br />

impressive. It was 51 feet long by 8 feet tall, <strong>and</strong> it weighed 5 tons. It<br />

contained more than 750,000 parts, <strong>and</strong> when it was running, it<br />

sounded like a room filled with sewing machines. The ASCC later<br />

became known as the Harvard Mark I.<br />

Impact<br />

Although this machine represented a significant technological<br />

achievement at the time <strong>and</strong> contributed ideas that would be used<br />

in subsequent machines, it was almost obsolete from the start. It was<br />

electromechanical, since it relied on relays, but it was built at the<br />

dawn of the electronic age. Fully electronic computers offered better<br />

reliability <strong>and</strong> faster speeds. Howard Aiken continued, without the<br />

help of IBM, to develop successors to the Mark I. Because he resisted<br />

using electronics, however, his machines did not significantly affect<br />

the direction of computer development.<br />

For all its complexity, the Mark I operated reasonably well, first<br />

solving problems related to the war effort <strong>and</strong> then turning its attention<br />

to the more mundane tasks of producing specialized mathematical<br />

tables. It remained in operation at the Harvard Computational<br />

Laboratory until 1959, when it was retired <strong>and</strong> disassembled.<br />

Parts of this l<strong>and</strong>mark computational tool are now kept at the<br />

Smithsonian Institute.<br />

See also BASIC programming language; Differential analyzer;<br />

Personal computer; Pocket calculator; UNIVAC computer.


Further Reading<br />

Mark I calculator / 493<br />

Cohen, I. Bernard. Howard Aiken: Portrait of a Computer Pioneer. Cambridge,<br />

Mass.: MIT Press, 1999.<br />

Ritchie, David. The Computer Pioneers: The Making of the Modern Computer.<br />

New York: Simon <strong>and</strong> Schuster, 1986.<br />

Slater, Robert. Portraits in Silicon. Cambridge, Mass.: MIT Press,<br />

1987.


494<br />

Mass spectrograph<br />

Mass spectrograph<br />

The invention: The first device used to measure the mass of atoms,<br />

which was found to be the result of the combination of isotopes.<br />

The people behind the invention:<br />

Francis William Aston (1877-1945), an English physicist who<br />

was awarded the 1922 Nobel Prize in Chemistry<br />

Sir Joseph John Thomson (1856-1940), an English physicist<br />

William Prout (1785-1850), an English biochemist<br />

Ernest Rutherford (1871-1937), an English physicist<br />

Same Element, Different Weights<br />

Isotopes are different forms of a chemical element that act similarly<br />

in chemical or physical reactions. Isotopes differ in two ways:<br />

They possess different atomic weights <strong>and</strong> different radioactive transformations.<br />

In 1803, John Dalton proposed a new atomic theory of<br />

chemistry that claimed that chemical elements in a compound combine<br />

by weight in whole number proportions to one another. By 1815,<br />

William Prout had taken Dalton’s hypothesis one step further <strong>and</strong><br />

claimed that the atomic weights of elements were integral (the integers<br />

are the positive <strong>and</strong> negative whole numbers <strong>and</strong> zero) multiples<br />

of the hydrogen atom. For example, if the weight of hydrogen<br />

was 1, then the weight of carbon was 12, <strong>and</strong> that of oxygen 16. Over<br />

the next decade, several carefully controlled experiments were conducted<br />

to determine the atomic weights of a number of elements. Unfortunately,<br />

the results of these experiments did not support Prout’s<br />

hypothesis. For example, the atomic weight of chlorine was found to<br />

be 35.5. It took a theory of isotopes, developed in the early part of the<br />

twentieth century, to verify Prout’s original theory.<br />

After his discovery of the electron, Sir Joseph John Thomson, the<br />

leading physicist at the Cavendish Laboratory in Cambridge, Engl<strong>and</strong>,<br />

devoted much of his remaining research years to determining<br />

the nature of “positive electricity.” (Since electrons are negatively<br />

charged, most electricity is negative.) While developing an instrument<br />

sensitive enough to analyze the positive electron, Thomson in-


vited Francis William Aston to work with him at the Cavendish Laboratory.<br />

Recommended by J. H. Poynting, who had taught Aston<br />

physics at Mason College, Aston began a lifelong association at<br />

Cavendish, <strong>and</strong> Trinity College became his home.<br />

When electrons are stripped from an atom, the atom becomes positively<br />

charged. Through the use of magnetic <strong>and</strong> electrical fields, it is<br />

possible to channel the resulting positive rays into parabolic tracks.<br />

By examining photographic plates of these tracks, Thomson was able<br />

to identify the atoms of different elements. Aston’s first contribution<br />

at Cavendish was to improve the instrument used to photograph<br />

the parabolic tracks. He developed a more efficient pump to<br />

create the required vacuum <strong>and</strong> devised a camera that would provide<br />

sharper photographs. By 1912, the improved apparatus had<br />

provided proof that the individual molecules of a substance have<br />

the same mass. While working on the element neon, however,<br />

Thomson obtained two parabolas, one with a mass of 20 <strong>and</strong> the<br />

other with a mass of 22, which seemed to contradict the previous<br />

findings that molecules of any substance have the same mass. Aston<br />

was given the task of resolving this mystery.<br />

Treating Particles Like Light<br />

Mass spectrograph / 495<br />

In 1919, Aston began to build a device called a “mass spectrograph.”<br />

The idea was to treat ionized or positive atoms like light. He<br />

reasoned that, because light can be dispersed into a rainbowlike<br />

spectrum <strong>and</strong> analyzed by means of its different colors, the same<br />

procedure could be used with atoms of an element such as neon. By<br />

creating a device that used magnetic fields to focus the stream of<br />

particles emitted by neon, he was able to create a mass spectrum<br />

<strong>and</strong> record it on a photographic plate. The heavier mass of neon (the<br />

first neon isotope) was collected on one part of a spectrum <strong>and</strong> the<br />

lighter neon (the second neon isotope) showed up on another. This<br />

mass spectrograph was a magnificent apparatus: The masses could<br />

be analyzed without reference to the velocity of the particles, which<br />

was a problem with the parabola method devised by Thomson.<br />

Neon possessed two isotopes: one with a mass of 20 <strong>and</strong> the other<br />

with a mass of 22, in a ratio of 10:1. When combined, this gave the<br />

atomic weight 20.20, which was the accepted weight of neon.


496 / Mass spectrograph<br />

Francis William Aston<br />

Francis W. Aston was born near Birmingham, Engl<strong>and</strong>, in<br />

1877 to William Aston, a farmer <strong>and</strong> metals dealer, <strong>and</strong> Fanny<br />

Charlotte Hollis, a gunmaker’s daughter. As a boy he loved to<br />

perform experiments by himself in his own small laboratory at<br />

home. His diligence helped him earn top marks in school, <strong>and</strong><br />

he attended Mason College (later the University of Birmingham).<br />

However, he failed to win a scholarship to continue his<br />

studies after graduation in 1901.<br />

He did not give up on experiments, however, even while<br />

holding a job as the chemist for a local brewery. He built his own<br />

equipment <strong>and</strong> investigated the nature of electricity. This work<br />

attracted the attention of the most famous researchers of the<br />

day. He finally got a scholarship in 1903 to the University of Birmingham<br />

<strong>and</strong> then joined the staff of Joseph John Thomson at<br />

the Royal Institution in London <strong>and</strong> Cambridge University,<br />

which remained his home until his death in 1945.<br />

Aston liked to work alone as much as possible. Given his<br />

unflagging attention to the details of measurement <strong>and</strong> his inventiveness<br />

with experimental equipment, his colleagues respected<br />

his lone-dog approach. Their trust was rewarded. After<br />

refining the mass spectrograph, Aston was able to explain a<br />

thorny problem in chemistry by showing that elements are<br />

composed of differing percentages of isotopes <strong>and</strong> that atomic<br />

weight varied slightly depending on the density of their atoms’<br />

nuclei. The research earned him the Nobel Prize in Chemistry in<br />

1922.<br />

Aston’s solitude extended into his private life. He never<br />

married, lavishing his affection instead on animals, outdoor<br />

sports, photography, travel, <strong>and</strong> music.<br />

Aston’s accomplishment in developing the mass spectrograph<br />

was recognized immediately by the scientific community. His was a<br />

simple device that was capable of accomplishing a large amount of<br />

research quickly. The field of isotope research, which had been<br />

opened up by Aston’s research, ultimately played an important part<br />

in other areas of physics.


Impact<br />

The years following 1919 were highly charged with excitement,<br />

since month after month new isotopes were announced. Chlorine<br />

had two; bromine had isotopes of 79 <strong>and</strong> 81, which gave an almost<br />

exact atomic weight of 80; krypton had six isotopes; <strong>and</strong> xenon had<br />

even more. In addition to the discovery of nonradioactive isotopes,<br />

the “whole-number rule” for chemistry was verified: Protons were<br />

the basic building blocks for different atoms, <strong>and</strong> they occurred exclusively<br />

in whole numbers.<br />

Aston’s original mass spectrograph had an accuracy of 1 in 1,000.<br />

In 1927, he built an even more accurate instrument, which was ten<br />

times more accurate. The new apparatus was sensitive enough to<br />

measure Albert Einstein’s law of mass energy conversion during a<br />

nuclear reaction. Between 1927 <strong>and</strong> 1935, Aston reviewed all the elements<br />

that he had worked on earlier <strong>and</strong> published updated results.<br />

He also began to build a still more accurate instrument, which<br />

proved to be of great value to nuclear chemistry.<br />

The discovery of isotopes opened the way to further research in<br />

nuclear physics <strong>and</strong> completed the speculations begun by Prout<br />

during the previous century. Although radioactivity was discovered<br />

separately, isotopes played a central role in the field of nuclear<br />

physics <strong>and</strong> chain reactions.<br />

See also Cyclotron; Electron microscope; Neutrino detector;<br />

Scanning tunneling microscope; Synchrocyclotron; Tevatron accelerator;<br />

Ultramicroscope.<br />

Further Reading<br />

Mass spectrograph / 497<br />

Aston, Francis William. “Mass Spectra <strong>and</strong> Isotopes” [Nobel lecture].<br />

In Chemistry, 1922-1941. River Edge, N.J.: World Scientific,<br />

1999.<br />

Squires, Gordon. “Francis Aston <strong>and</strong> the Mass Spectrograph.” Journal<br />

of the Chemical Society. Dalton Transactions no. 23 (1998).<br />

Thackray, Arnold. Atoms <strong>and</strong> Powers: An Essay on Newtonian Matter-<br />

Theory <strong>and</strong> the Development of Chemistry. Cambridge, Mass.: Harvard<br />

University Press, 1970.


498<br />

Memory metal<br />

Memory metal<br />

The invention: Known as nitinol, a metal alloy that returns to its<br />

original shape, after being deformed, when it is heated to the<br />

proper temperature.<br />

The person behind the invention:<br />

William Buehler (1923- ), an American metallurgist<br />

The Alloy with a Memory<br />

In 1960, William Buehler developed an alloy that consisted of 53<br />

to 57 percent nickel (by weight) <strong>and</strong> the balance titanium. This alloy,<br />

which is called nitinol, turned out to have remarkable properties.<br />

Nitinol is a “memory metal,” which means that, given the proper<br />

conditions, objects made of nitinol can be restored to their original<br />

shapes even after they have been radically deformed. The return to<br />

the original shape is triggered by heating the alloy to a moderate<br />

temperature. As the metal “snaps back” to its original shape, considerable<br />

force is exerted <strong>and</strong> mechanical work can be done.<br />

Alloys made of nickel <strong>and</strong> titanium have great potential in a<br />

wide variety of industrial <strong>and</strong> government applications. These include:<br />

for the computer market, a series of high-performance electronic<br />

connectors; for the medical market, intravenous fluid devices<br />

that feature precise fluid control; for the consumer market, eyeglass<br />

frame components; <strong>and</strong>, for the industrial market, power cable couplings<br />

that provide durability at welded joints.<br />

The Uncoiling Spring<br />

At one time, the “uncoiling spring experiment” was used to<br />

amuse audiences, <strong>and</strong> a number of scientists have had fun with<br />

nitinol in front of unsuspecting viewers. It is now generally recognized<br />

that the shape memory effect involves a thermoelastic transformation<br />

at the atomic level. This process is unique in that the<br />

transformation back to the original shape occurs as a result of stored<br />

elastic energy that assists the chemical driving force that is unleashed<br />

by heating the metal.


Memory metal / 499<br />

The mechanism, simply stated, is that shape memory alloys<br />

are rather easily deformed below their “critical temperature.”<br />

Provided that the extent of the deformation is not too great, the<br />

original, undeformed state can be recovered by heating the alloy<br />

to a temperature just below the critical temperature. It is also<br />

significant that substantial stresses are generated when a deformed<br />

specimen “springs back” to its original shape. This phenomenon<br />

is very peculiar compared to the ordinary behavior of<br />

most materials.<br />

Researchers at the Naval Ordnance Laboratory discovered nitinol<br />

by accident in the process of trying to learn how to make titanium<br />

less brittle. They tried adding nickel, <strong>and</strong> when they were showing a<br />

wire of the alloy to some administrators, someone smoking a cigar<br />

held his match too close to the sample, causing the nitinol to spring<br />

back into shape. One of the first applications of the discovery was a<br />

new way to link hydraulic lines on the Navy’s F-14 fighter jets. The<br />

nitinol “sleeve” was cooled with liquid nitrogen, which enlarged<br />

the sample. Then it was slipped into place between two pipes.<br />

When the sleeve was warmed up, it contracted, clamping the pipes<br />

together <strong>and</strong> keeping them clamped with a force of nearly 50,000<br />

pounds per square inch.<br />

Nitinol is not an easy alloy with which to work. When it is drilled<br />

or passed through a lathe, it becomes hardened <strong>and</strong> resists change.<br />

Welding nitinol <strong>and</strong> electroplating it have become manufacturing<br />

nightmares. It also resists taking on a desired shape. The frictional<br />

forces of many processes heat the nitinol, which activates its memory.<br />

Its fantastic elasticity also causes difficulties. If it is placed in a press<br />

with too little force, the spring comes out of the die unchanged. With<br />

too much force, the metal breaks into fragments. Using oil as a cooling<br />

lubricant <strong>and</strong> taking a step-wise approach to altering the alloy,<br />

however, allows it to be fashioned into particular shapes.<br />

One unique use of nitinol occurs in cardiac surgery. Surgical<br />

tools made of nitinol can be bent up to 90 degrees, allowing them<br />

to be passed into narrow vessels <strong>and</strong> then retrieved. The tools are<br />

then straightened out in an autoclave so that they can be reused.


500 / Memory metal<br />

Consequences<br />

Many of the technical problems of working with nitinol have<br />

been solved, <strong>and</strong> manufacturers of the alloy are selling more than<br />

twenty different nitinol products to countless companies in the<br />

fields of medicine, transportation, consumer products, <strong>and</strong> toys.<br />

Nitinol toys include blinking movie posters, butterflies with<br />

flapping wings, <strong>and</strong> dinosaurs whose tails move; all these applications<br />

are driven by a contracting bit of wire that is connected to a<br />

watch battery. The “Thermobile” <strong>and</strong> the “Icemobile” are toys whose<br />

wheels are set in motion by hot water or by ice cubes.<br />

Orthodontists sometimes use nitinol wires <strong>and</strong> springs in braces<br />

because the alloy pulls with a force that is more gentle <strong>and</strong> even<br />

than that of stainless steel, thus causing less pain. Nitinol does not<br />

react with organic materials, <strong>and</strong> it is also useful as a new type of<br />

blood-clot filter. Best of all, however, is the use of nitinol for eyeglass<br />

frames. If the wearer deforms the frames by sitting on them (<strong>and</strong><br />

people do so frequently), the optometrist simply dips the crumpled<br />

frames in hot water <strong>and</strong> the frames regain their original shape.<br />

From its beginnings as an “accidental” discovery, nitinol has<br />

gone on to affect various fields of science <strong>and</strong> technology, from the<br />

“Cryofit” couplings used in the hydraulic tubing of aircraft to the<br />

pin-<strong>and</strong>-socket contacts used in electrical circuits. Nitinol has also<br />

found its way into integrated circuit packages.<br />

In an age of energy conservation, the unique phase transformation<br />

of nickel-titanium alloys allows them to be used in lowtemperature<br />

heat engines. The world has abundant resources of<br />

low-grade thermal energy, <strong>and</strong> the recovery of this energy can be<br />

accomplished by the use of materials such as nitinol. Despite the<br />

limitations imposed on heat engines working at low temperatures<br />

across a small temperature change, sources of low-grade heat are<br />

so widespread that the economical conversion of a fractional percentage<br />

of that energy could have a significant impact on the<br />

world’s energy supply.<br />

Nitinol has also become useful as a material capable of absorbing<br />

internal vibrations in structural materials, <strong>and</strong> it has been used as<br />

“Harrington rods” to treat scoliosis (curvature of the spine).


See also Disposable razor; Neoprene; Plastic; Steelmaking process;<br />

Teflon; Tungsten filament.<br />

Further Reading<br />

Memory metal / 501<br />

Gisser, Kathleen R. C., et al. “Nickel-Titanium Memory Metal.” Journal<br />

of Chemical Education 71, no. 4 (April, 1994).<br />

Iovine, John. “The World’s ‘Smartest’ Metal.” Poptronics 1, no. 12<br />

(December, 2000).<br />

Jackson, Curtis M., H. J. Wagner, <strong>and</strong> Roman Jerzy Wasilewski. 55-<br />

Nitinol: The Alloy with a Memory: Its Physical Metallurgy, Properties,<br />

<strong>and</strong> Applications. Washington: Technology Utilization Office,<br />

1972.<br />

Walker, Jearl. “The Amateur Scientist.” Scientific American 254, no. 5<br />

(May, 1986).


502<br />

Microwave cooking<br />

Microwave cooking<br />

The invention: System of high-speed cooking that uses microwave<br />

radition to agitate liquid molecules to raise temperatures by friction.<br />

The people behind the invention:<br />

Percy L. Spencer (1894-1970), an American engineer<br />

Heinrich Hertz (1857-1894), a German physicist<br />

James Clerk Maxwell (1831-1879), a Scottish physicist<br />

The Nature of Microwaves<br />

Microwaves are electromagnetic waves, as are radio waves, X<br />

rays, <strong>and</strong> visible light. Water waves <strong>and</strong> sound waves are waveshaped<br />

disturbances of particles in the media—water in the case of<br />

water waves <strong>and</strong> air or water in the case of sound waves—through<br />

which they travel. Electromagnetic waves, however, are wavelike<br />

variations of intensity in electric <strong>and</strong> magnetic fields.<br />

Electromagnetic waves were first studied in 1864 by James Clerk<br />

Maxwell, who explained mathematically their behavior <strong>and</strong> velocity.<br />

Electromagnetic waves are described in terms of their “wavelength”<br />

<strong>and</strong> “frequency.” The wavelength is the length of one cycle,<br />

which is the distance from the highest point of one wave to the highest<br />

point of the next wave, <strong>and</strong> the frequency is the number of cycles<br />

that occur in one second. Frequency is measured in units called<br />

“hertz,” named for the German physicist Heinrich Hertz. The frequencies<br />

of microwaves run from 300 to 3,000 megahertz (1 megahertz<br />

equals 1 million hertz, or 1 million cycles per second), corresponding<br />

to wavelengths of 100 to 10 centimeters.<br />

Microwaves travel in the same way that light waves do; they are<br />

reflected by metallic objects, absorbed by some materials, <strong>and</strong> transmitted<br />

by other materials. When food is subjected to microwaves, it<br />

heats up because the microwaves make the water molecules in foods<br />

(water is the most common compound in foods) vibrate. Water is a<br />

“dipole molecule,” which means that it contains both positive <strong>and</strong><br />

negative charges. When the food is subjected to microwaves, the di-


pole water molecules try to align themselves with the alternating<br />

electromagnetic field of the microwaves. This causes the water molecules<br />

to collide with one another <strong>and</strong> with other molecules in the<br />

food. Consequently, heat is produced as a result of friction.<br />

Development of the Microwave Oven<br />

Microwave cooking / 503<br />

Percy L. Spencer apparently discovered the principle of microwave<br />

cooking while he was experimenting with a radar device at<br />

the Raytheon Company. A c<strong>and</strong>y bar in his pocket melted after being<br />

exposed to microwaves. After realizing what had happened,<br />

Spencer made the first microwave oven from a milk can <strong>and</strong> applied<br />

for two patents, “Method of Treating Foodstuffs” <strong>and</strong> “Means for<br />

Treating Foodstuffs,” on October 8, 1945, giving birth to microwaveoven<br />

technology.<br />

Spencer wrote that his invention “relates to the treatment of<br />

foodstuffs <strong>and</strong>, more particularly, to the cooking thereof through<br />

the use of electromagnetic energy.” Though the use of electromagnetic<br />

energy for heating was recognized at that time, the frequencies<br />

that were used were lower than 50 megahertz. Spencer discovered<br />

that heating at such low frequencies takes a long time. He eliminated<br />

the time disadvantage by using shorter wavelengths in the<br />

microwave region. Wavelengths of 10 centimeters or shorter were<br />

comparable to the average dimensions of foods. When these wavelengths<br />

were used, the heat that was generated became intense, the<br />

energy that was required was minimal, <strong>and</strong> the process became efficient<br />

enough to be exploited commercially.<br />

Although Spencer’s patents refer to the cooking of foods with<br />

microwave energy, neither deals directly with a microwave oven.<br />

The actual basis for a microwave oven may be patents filed by other<br />

researchers at Raytheon. A patent by Karl Stiefel in 1949 may be the<br />

forerunner of the microwave oven, <strong>and</strong> in 1950, Fritz Gross received<br />

a patent entitled “Cooking Apparatus,” which specifically describes<br />

an oven that is very similar to modern microwave ovens.<br />

Perhaps the first mention of a commercial microwave oven was<br />

made in the November, 1946, issue of Electronics magazine. This article<br />

described the newly developed Radarange as a device that<br />

could bake biscuits in 29 seconds, cook hamburgers in 35 seconds,


504 / Microwave cooking<br />

Percy L. Spencer<br />

Percy L. Spencer (1894-1970) had an unpromising background<br />

for the inventor of the twentieth century’s principal innovation<br />

in the technology of cooking. He was orphaned while<br />

still a young boy <strong>and</strong> never completed grade school. However,<br />

he possessed a keen curiosity <strong>and</strong> the imaginative intelligence<br />

to educate himself <strong>and</strong> recognize how to make things better.<br />

In 1941 the magnetron, which produces microwaves, was so<br />

complex <strong>and</strong> difficult to make that fewer than two dozen were<br />

produced in a day. This pace delayed the campaign to improve<br />

radar, which used magnetrons, so Spencer, while working for<br />

Raytheon Corporation, set out to speed things along. He simplified<br />

the design <strong>and</strong> made it more efficient at the same time. Production<br />

of magnetrons soon increased more than a thous<strong>and</strong>fold.<br />

In 1945 he discovered by accident that microwaves could<br />

heat chocolate past the melting point. He immediately tried an<br />

experiment by training microwaves on popcorn kernels <strong>and</strong><br />

was delighted to see them puff up straight away.<br />

The first microwave oven based on his discovery stood five<br />

feet, six inches tall <strong>and</strong> weighed 750 pounds, suitable only for<br />

restaurants. However, it soon got smaller, thanks to researchers<br />

at Raytheon. And after some initial hostility from cooks, it became<br />

popular. Raytheon bought Amana Refrigeration in 1965<br />

to manufacture the home models <strong>and</strong> marketed them worldwide.<br />

Meanwhile, Spencer had become a senior vice president<br />

at the company <strong>and</strong> a member of its board of directors. Raytheon<br />

named one of its buildings after him, the U.S. Navy presented<br />

him with the Distinguished Service Medal for his contributions,<br />

<strong>and</strong> in 1999 he entered the <strong>Inventors</strong> Hall of Fame.<br />

<strong>and</strong> grill a hot dog in 8 to 10 seconds. Another article that appeared a<br />

month later mentioned a unit that had been developed specifically<br />

for airline use. The frequency used in this oven was 3,000 megahertz.<br />

Within a year, a practical model 13 inches wide, 14 inches<br />

deep, <strong>and</strong> 15 inches high appeared, <strong>and</strong> several new models were<br />

operating in <strong>and</strong> around Boston. In June, 1947, Electronics magazine<br />

reported the installation of a Radarange in a restaurant, signaling<br />

the commercial use of microwave cooking. It was reported that this


method more than tripled the speed of service. The Radarange became<br />

an important addition to a number of restaurants, <strong>and</strong> in 1948,<br />

Bernard Proctor <strong>and</strong> Samuel Goldblith used it for the first time to<br />

conduct research into microwave cooking.<br />

In the United States, the radio frequencies that can be used for<br />

heating are allocated by the Federal Communications Commission<br />

(FCC). The two most popular frequencies for microwave cooking<br />

are 915 <strong>and</strong> 2,450 megahertz, <strong>and</strong> the 2,450 frequency is used in<br />

home microwave ovens. It is interesting that patents filed by Spencer<br />

in 1947 mention a frequency on the order of 2,450 megahertz. This<br />

fact is another example of Spencer’s vision in the development of<br />

microwave cooking principles. The Raytheon Company concentrated<br />

on using 2,450 megahertz, <strong>and</strong> in 1955, the first domestic microwave<br />

oven was introduced. It was not until the late 1960’s, however,<br />

that the price of the microwave oven decreased sufficiently for<br />

the device to become popular. The first patent describing a microwave<br />

heating system being used in conjunction with a conveyor<br />

was issued to Spencer in 1952. Later, based on this development,<br />

continuous industrial applications of microwaves were developed.<br />

Impact<br />

Microwave cooking / 505<br />

Initially, microwaves were viewed as simply an efficient means<br />

of rapidly converting electric energy to heat. Since that time, however,<br />

they have become an integral part of many applications. Because<br />

of the pioneering efforts of Percy L. Spencer, microwave applications<br />

in the food industry for cooking <strong>and</strong> for other processing<br />

operations have flourished. In the early 1970’s, there were eleven<br />

microwave oven companies worldwide, two of which specialized<br />

in food processing operations, but the growth of the microwave<br />

oven industry has paralleled the growth in the radio <strong>and</strong> television<br />

industries. In 1984, microwave ovens accounted for more shipments<br />

than had ever been achieved by any appliance—9.1 million units.<br />

By 1989, more than 75 percent of the homes in the United States<br />

had microwave ovens, <strong>and</strong> in the 1990’s, microwavable foods were<br />

among the fastest-growing products in the food industry. Microwave<br />

energy facilitates reductions in operating costs <strong>and</strong> required<br />

energy, higher-quality <strong>and</strong> more reliable products, <strong>and</strong> positive en-


506 / Microwave cooking<br />

vironmental effects. To some degree, the use of industrial microwave<br />

energy remains in its infancy. New <strong>and</strong> improved applications<br />

of microwaves will continue to appear.<br />

See also Electric refrigerator; Fluorescent lighting; Food freezing;<br />

Robot (household); Television; Tupperware; Vacuum cleaner;<br />

Washing machine.<br />

Further Reading<br />

Baird, Davis, R. I. G. Hughes, <strong>and</strong> Alfred Nordmann. Heinrich Hertz:<br />

Classical Physicist, Modern Philosopher. Boston: Kluwer Academic,<br />

1998.<br />

Roman, Mark. “That Marvelous Machine in Your Kitchen.” Reader’s<br />

Digest (February, 1990).<br />

Scott, Otto. The Creative Ordeal: The Story of Raytheon. New York:<br />

Atheneum, 1974.<br />

Simpson, Thomas K. Maxwell on the Electromagnetic Field: A Guided<br />

Study. New Brunswick, N.J.: Rutgers University Press, 1997.<br />

Tolstoy, Ivan. James Clerk Maxwell: A Biography. Chicago: University<br />

of Chicago Press, 1982.


Neoprene<br />

Neoprene<br />

The invention: The first commercially practical synthetic rubber,<br />

Neoprene gave a boost to polymer chemistry <strong>and</strong> the search for<br />

new materials.<br />

The people behind the invention:<br />

Wallace Hume Carothers (1896-1937), an American chemist<br />

Arnold Miller Collins (1899- ), an American chemist<br />

Elmer Keiser Bolton (1886-1968), an American chemist<br />

Julius Arthur Nieuwl<strong>and</strong> (1879-1936), a Belgian American<br />

priest, botanist, <strong>and</strong> chemist<br />

Synthetic Rubber: A Mirage?<br />

507<br />

The growing dependence of the industrialized nations upon<br />

elastomers (elastic substances) <strong>and</strong> the shortcomings of natural<br />

rubber motivated the twentieth century quest for rubber substitutes.<br />

By 1914, rubber had become nearly as indispensable as coal<br />

or iron. The rise of the automobile industry, in particular, had created<br />

a strong dem<strong>and</strong> for rubber. Unfortunately, the availability of<br />

rubber was limited by periodic shortages <strong>and</strong> spiraling prices. Furthermore,<br />

the particular properties of natural rubber, such as its<br />

lack of resistance to oxygen, oils, <strong>and</strong> extreme temperatures, restrict<br />

its usefulness in certain applications. These limitations stimulated<br />

a search for special-purpose rubber substitutes.<br />

Interest in synthetic rubber dates back to the 1860 discovery by<br />

the English chemist Greville Williams that the main constituent<br />

of rubber is isoprene, a liquid hydrocarbon. Nineteenth century<br />

chemists attempted unsuccessfully to transform isoprene into<br />

rubber. The first large-scale production of a rubber substitute occurred<br />

during World War I. A British blockade forced Germany to<br />

begin to manufacture methyl rubber in 1916, but methyl rubber<br />

turned out to be a poor substitute for natural rubber. When the<br />

war ended in 1918, a practical synthetic rubber was still only a mirage.<br />

Nevertheless, a breakthrough was on the horizon.


508 / Neoprene<br />

Mirage Becomes Reality<br />

In 1930, chemists at E. I. Du Pont de Nemours discovered the<br />

elastomer known as neoprene. Of the more than twenty chemists<br />

who helped to make this discovery possible, four st<strong>and</strong> out: Elmer<br />

Bolton, Julius Nieuwl<strong>and</strong>, Wallace Carothers, <strong>and</strong> Arnold Collins.<br />

Bolton directed Du Pont’s drystuffs department in the mid-<br />

1920’s. Largely because of the rapidly increasing price of rubber, he<br />

initiated a project to synthesize an elastomer from acetylene, a gaseous<br />

hydrocarbon. In December, 1925, Bolton attended the American<br />

Chemical Society’s convention in Rochester, New York, <strong>and</strong><br />

heard a presentation dealing with acetylene reactions. The presenter<br />

was Julius Nieuwl<strong>and</strong>, the foremost authority on the chemistry<br />

of acetylene.<br />

Nieuwl<strong>and</strong> was a professor of organic chemistry at the University<br />

of Notre Dame. (One of his students was the legendary football<br />

coach Knute Rockne.) The priest-scientist had been investigating<br />

acetylene reactions for more than twenty years. Using a copper<br />

chloride catalyst he had discovered, he isolated a new compound,<br />

divinylacetylene (DVA). He later treated DVA with a vulcanizing<br />

(hardening) agent <strong>and</strong> succeeded in producing a rubberlike substance,<br />

but the substance proved to be too soft for practical use.<br />

Bolton immediately recognized the importance of Nieuwl<strong>and</strong>’s<br />

discoveries <strong>and</strong> discussed with him the possibility of using DVA as<br />

a raw material for a synthetic rubber. Seven months later, an alliance<br />

was formed that permitted Du Pont researchers to use Nieuwl<strong>and</strong>’s<br />

copper catalyst. Bolton hoped that the catalyst would be the key to<br />

making an elastomer from acetylene. As it turned out, Nieuwl<strong>and</strong>’s<br />

catalyst was indispensable for manufacturing neoprene.<br />

Over the next several years, Du Pont scientists tried unsuccessfully<br />

to produce rubberlike materials. Using Nieuwl<strong>and</strong>’s catalyst,<br />

they managed to prepare DVA <strong>and</strong> also to isolate monovinylacetylene<br />

(MVA), a new compound that eventually proved to be the vital<br />

intermediate chemical in the making of neoprene. Reactions of<br />

MVA <strong>and</strong> DVA, however, produced only hard, brittle materials.<br />

In 1928, Du Pont hired a thirty-one-year-old Harvard instructor,<br />

Wallace Carothers, to direct the organic chemicals group. He began<br />

a systematic exploration of polymers (complex molecules). In early


1930, he accepted an assignment to investigate the chemistry of<br />

DVA. He appointed one of his assistants, Arnold Collins, to conduct<br />

the laboratory experiments. Carothers suggested that Collins<br />

should explore the reaction between MVA <strong>and</strong> hydrogen chloride.<br />

His suggestion would lead to the discovery of neoprene.<br />

One of Collins’s experiments yielded a new liquid, <strong>and</strong> on April<br />

17, 1930, he recorded in his laboratory notebook that the liquid had<br />

solidified into a rubbery substance. When he dropped it on a bench,<br />

it bounced. This was the first batch of neoprene. Carothers named<br />

Collins’s liquid “chloroprene.” Chloroprene is analogous structurally<br />

to isoprene, but it polymerizes much more rapidly. Carothers<br />

conducted extensive investigations of the chemistry of chloroprene<br />

<strong>and</strong> related compounds. His studies were the foundation for Du<br />

Pont’s development of an elastomer that was superior to all previously<br />

known synthetic rubbers.<br />

Du Pont chemists, including Carothers <strong>and</strong> Collins, formally introduced<br />

neoprene—originally called “DuPrene”—on November 3,<br />

1931, at the meeting of the American Chemical Society in Akron,<br />

Ohio. Nine months later, the new elastomer began to be sold.<br />

Impact<br />

Neoprene / 509<br />

The introduction of neoprene was a milestone in humankind’s development<br />

of new materials. It was the first synthetic rubber worthy<br />

of the name. Neoprene possessed higher tensile strength than rubber<br />

<strong>and</strong> much better resistance to abrasion, oxygen, heat, oils, <strong>and</strong> chemicals.<br />

Its main applications included jacketing for electric wires <strong>and</strong><br />

cables, work-shoe soles, gasoline hoses, <strong>and</strong> conveyor <strong>and</strong> powertransmission<br />

belting. By 1939, when Adolf Hitler’s troops invaded Pol<strong>and</strong>,<br />

nearly every major industry in America was using neoprene.<br />

After the Japanese bombing of Pearl Harbor, in 1941, the elastomer<br />

became even more valuable to the United States. It helped the United<br />

States <strong>and</strong> its allies survive the critical shortage of natural rubber that<br />

resulted when Japan seized Malayan rubber plantations.<br />

A scientifically <strong>and</strong> technologically significant side effect of the<br />

introduction of neoprene was the stimulus that the breakthrough<br />

gave to polymer research. Chemists had long debated whether<br />

polymers were mysterious aggregates of smaller units or were gen-


510 / Neoprene<br />

uine molecules. Carothers ended the debate by demonstrating in a<br />

series of now-classic papers that polymers were indeed ordinary—<br />

but very large—molecules. In the 1930’s, he put polymer studies on<br />

a firm footing. The advance of polymer science led, in turn, to the<br />

development of additional elastomers <strong>and</strong> synthetic fibers, including<br />

nylon, which was invented by Carothers himself in 1935.<br />

See also Buna rubber; Nylon; Orlon; Plastic; Polyester; Polyethylene;<br />

Polystyrene; Silicones; Teflon.<br />

Further Reading<br />

Furukawa, Yasu. Inventing Polymer Science: Staudinger, Carothers, <strong>and</strong><br />

the Emergence of Macromolecular Chemistry. Philadelphia: University<br />

of Pennsylvania Press, 1998.<br />

Hermes, Matthew E. Enough for One Lifetime: Wallace Carothers, Inventor<br />

of Rayon. Washington, D.C.: American Chemical Society<br />

<strong>and</strong> the Chemical Heritage Foundation, 1996.<br />

Taylor, Graham D., <strong>and</strong> Patricia E. Sudnik. Du Pont <strong>and</strong> the International<br />

Chemical Industry. Boston, Mass.: Twayne, 1984.


Neutrino detector<br />

Neutrino detector<br />

The invention: A device that provided the first direct evidence that<br />

the Sun runs on thermonuclear power <strong>and</strong> challenged existing<br />

models of the Sun.<br />

The people behind the invention:<br />

Raymond Davis, Jr. (1914- ), an American chemist<br />

John Norris Bahcall (1934- ), an American astrophysicist<br />

Missing Energy<br />

511<br />

In 1871, Hermann von Helmholtz, the German physicist, anatomist,<br />

<strong>and</strong> physiologist, suggested that no ordinary chemical reaction<br />

could be responsible for the enormous energy output of the<br />

Sun. By the 1920’s, astrophysicists had realized that the energy radiated<br />

by the Sun must come from nuclear fusion, in which protons or<br />

nuclei combine to form larger nuclei <strong>and</strong> release energy. These reactions<br />

were assumed to be taking place deep in the interior of the<br />

Sun, in an immense thermonuclear furnace, where the pressures<br />

<strong>and</strong> temperatures were high enough to allow fusion to proceed.<br />

Conventional astronomical observations could record only the<br />

particles of light emitted by the much cooler outer layers of the Sun<br />

<strong>and</strong> could not provide evidence for the existence of a thermonuclear<br />

furnace in the interior. Then scientists realized that the neutrino<br />

might be used to prove that this huge furnace existed. Of all the particles<br />

released in the fusion process, only one type—the neutrino—<br />

interacts so infrequently with matter that it can pass through the<br />

Sun <strong>and</strong> reach the earth. These neutrinos provide a way to verify directly<br />

the hypothesis of thermonuclear energy generated in stars.<br />

The neutrino was “invented” in 1930 by the American physicist<br />

Wolfgang Pauli to account for the apparent missing energy in the<br />

beta decay, or emission of an electron, from radioactive nuclei. He<br />

proposed that an unseen nuclear particle, which he called a neutrino,<br />

was also emitted in beta decay, <strong>and</strong> that it carried off the<br />

“missing” energy. To balance the energy but not be observed in the<br />

decay process, Pauli’s hypothetical particle had to have no electrical


512 / Neutrino detector<br />

charge, have little or no mass, <strong>and</strong> interact only very weakly with<br />

ordinary matter. Typical neutrinos would have to be able to pass<br />

through millions of miles of ordinary matter in order to reach the<br />

earth. Scientists’ detectors, <strong>and</strong> even the whole earth or Sun, were<br />

essentially transparent as far as Pauli’s neutrinos were concerned.<br />

Because the neutrino is so difficult to detect, it took more than<br />

twenty-five years to confirm its existence. In 1956, Clyde Cowan<br />

<strong>and</strong> Frederick Reines, both physicists at the Los Alamos National<br />

Laboratory, built the world’s largest scintillation counter, a device to<br />

detect the small flash of light given off when the neutrino strikes<br />

(“interacts” with) a certain substance in the apparatus. They placed<br />

this scintillation counter near the Savannah River Nuclear Reactor,<br />

which was producing about 1 trillion neutrinos every second. Although<br />

only one neutrino interaction was observed in their detector<br />

every twenty minutes, Cowan <strong>and</strong> Reines were able to confirm the<br />

existence of Pauli’s elusive particle.<br />

The task of detecting the solar neutrinos was even more formidable.<br />

If an apparatus similar to the Cowan <strong>and</strong> Reines detector were<br />

employed to search for the neutrinos from the Sun, only one interaction<br />

could be expected every few thous<strong>and</strong> years.<br />

Missing Neutrinos<br />

At about the same time that Cowan <strong>and</strong> Reines performed their<br />

experiment, another type of neutrino detector was under development<br />

by Raymond Davis, Jr., a chemist at the Brookhaven National<br />

Laboratory. Davis employed an idea, originally suggested in 1948<br />

by the nuclear physicist Bruno Pontecorvo, that when a neutrino interacts<br />

with a chlorine-37 nucleus, it produces a nucleus of argon 37.<br />

Any argon so produced could then be extracted from large volumes<br />

of chlorine-rich liquid by passing helium gas through the liquid.<br />

Since argon 37 is radioactive, it is relatively easy to detect.<br />

Davis tested a version of this neutrino detector, containing about<br />

3,785 liters of carbon tetrachloride liquid, near a nuclear reactor at<br />

the Brookhaven National Laboratory from 1954 to 1956. In the scientific<br />

paper describing his results, Davis suggested that this type of<br />

neutrino detector could be made large enough to permit detection<br />

of solar neutrinos.


Neutrino detector / 513<br />

Patients undergoing nuclear magnetic resonance image (MRI) examinations are placed inside<br />

cylindrical chambers in which their bodies are held rigidly in place. (Digital Stock)<br />

Although Davis’s first attempt to detect solar neutrinos from a<br />

limestone mine at Barberton, Ohio, failed, he continued his search<br />

with a much larger detector 1,478 meters underground in the<br />

Homestake Gold Mine in Lead, South Dakota. The cylindrical tank<br />

(6.1 meters in diameter, 16 meters long, <strong>and</strong> containing 378,540 liters<br />

of perchloroethylene) was surrounded by water to shield the<br />

detector from neutrons emitted by trace quantities of uranium <strong>and</strong><br />

thorium in the walls of the mine. The experiment was conducted<br />

underground to shield it from cosmic radiation.<br />

To describe his results, Davis coined a new unit, the “solar<br />

neutrino unit” (SNU), with 1 SNU indicating the production of<br />

one atom of argon 37 every six days. Astrophysicist John Norris<br />

Bahcall, using the best available astronomical models of the nuclear<br />

reactions going on in the sun’s interior, as well as the physical<br />

properties of the neutrinos, had predicted a capture rate of 50<br />

SNUs in 1963. The 1967 results from Davis’s detector, however,<br />

had an upper limit of only 3 SNUs.


514 / Neutrino detector<br />

Consequences<br />

The main significance of the detection of solar neutrinos by Davis<br />

was the direct confirmation that thermonuclear fusion must be occurring<br />

at the center of the Sun. The low number of solar neutrinos<br />

Davis detected, however, has called into question some of the fundamental<br />

beliefs of astrophysics. As Bahcall explained: “We know<br />

more about the Sun than about any other star....TheSunisalso in<br />

what is believed to be the best-understood stage of stellar evolution....Ifwearetohave<br />

confidence in the many astronomical <strong>and</strong><br />

cosmological applications of the theory of stellar evolution, it ought<br />

at least to give the right answers about the Sun.”<br />

Many solutions to the problem of the “missing” solar neutrinos<br />

have been proposed. Most of these solutions can be divided into<br />

two broad classes: those that challenge the model of the sun’s interior<br />

<strong>and</strong> those that challenge the underst<strong>and</strong>ing of the behavior of<br />

the neutrino. Since the number of neutrinos produced is very sensitive<br />

to the temperature of the sun’s interior, some astrophysicists<br />

have suggested that the true solar temperature may be lower than<br />

expected. Others suggest that the sun’s outer layer may absorb<br />

more neutrinos than expected. Some physicists, however, believe<br />

neutrinos may occur in several different forms, only one of which<br />

can be detected by the chlorine detectors.<br />

Alpha Rays (charged helium nuclei)<br />

Beta Rays (charged electrons)<br />

Gamma Rays, X Rays (photons)<br />

Neutrinos<br />

(chargeless, nearly massless subatomic particles)<br />

Skin<br />

Metal Metal<br />

(0.12 cm aluminum) (lead)<br />

Neutrinos can pass through most forms of matter without interacting with other nuclear<br />

particles.


Davis’s discovery of the low number of neutrinos reaching Earth<br />

has focused years of attention on a better underst<strong>and</strong>ing of how the<br />

Sun generates its energy <strong>and</strong> how the neutrino behaves. New <strong>and</strong><br />

more elaborate solar neutrino detectors have been built with the<br />

aim of underst<strong>and</strong>ing stars, including the Sun, as well as the physics<br />

<strong>and</strong> behavior of the elusive neutrino.<br />

See also Radio interferometer; Weather satellite.<br />

Further Reading<br />

Neutrino detector / 515<br />

Bartusiak, Marcia. “Underground Astronomer.” Astronomy 28, no. 1<br />

(January, 2000).<br />

“Neutrino Test to Probe Sun.” New Scientist 140, no. 1898 (November<br />

6, 1993).<br />

“Pioneering Neutrino Astronomers to Share 2000 Wolf Prize in<br />

Physics.” Physics Today 53, no. 3 (March, 2000).<br />

Schwarzschild, Bertram. “Can Helium Mixing Explain the Solar<br />

Neutrino Shortages?” Physics Today 50, no. 3 (March, 1997).<br />

Zimmerman, Robert. “The Shadow Boxer.” The Sciences 36, no. 1<br />

(January/February, 1996).


516<br />

Nuclear magnetic resonance<br />

Nuclear magnetic resonance<br />

The invention: Procedure that uses hydrogen atoms in the human<br />

body, strong electromagnets, radio waves, <strong>and</strong> detection equipment<br />

to produce images of sections of the brain.<br />

The people behind the invention:<br />

Raymond Damadian (1936- ), an American physicist <strong>and</strong><br />

inventor<br />

Paul C. Lauterbur (1929- ), an American chemist<br />

Peter Mansfield (1933- ), a scientist at the University of<br />

Nottingham, Engl<strong>and</strong><br />

Peering into the Brain<br />

Doctors have always wanted the ability to look into the skull <strong>and</strong><br />

see the human brain without harming the patient who is being examined.<br />

Over the years, various attempts were made to achieve this<br />

ability. At one time, the use of X rays, which were first used by Wilhelm<br />

Conrad Röntgen in 1895, seemed to be an option, but it was<br />

found that X rays are absorbed by bone, so the skull made it impossible<br />

to use X-ray technology to view the brain. The relatively recent<br />

use of computed tomography (CT) scanning, a computer-assisted<br />

imaging technology, made it possible to view sections of the head<br />

<strong>and</strong> other areas of the body, but the technique requires that the part<br />

of the body being “imaged,” or viewed, be subjected to a small<br />

amount of radiation, thereby putting the patient at risk. Positron<br />

emission tomography (PET) could also be used, but it requires that<br />

small amounts of radiation be injected into the patient, which also<br />

puts the patient at risk. Since the early 1940’s, however, a new technology<br />

had been developing.<br />

This technology, which appears to pose no risk to patients, is<br />

called “nuclear magnetic resonance spectroscopy.” It was first used<br />

to study the molecular structures of pure samples of chemicals. This<br />

method developed until it could be used to follow one chemical as it<br />

changed into another, <strong>and</strong> then another, in a living cell. By 1971,<br />

Raymond Damadian had proposed that body images that were


more vivid <strong>and</strong> more useful than X rays could be produced by<br />

means of nuclear magnetic resonance spectroscopy. In 1978, he<br />

founded his own company, FONAR, which manufactured the scanners<br />

that are necessary for the technique.<br />

Magnetic Resonance Images<br />

Nuclear magnetic resonance / 517<br />

The first nuclear magnetic resonance images (MRIs) were published<br />

by Paul Lauterbur in 1973. Although there seemed to be no<br />

possibility that MRI could be harmful to patients, everyone involved<br />

in MRI research was very cautious. In 1976, Peter Mansfield,<br />

at the University of Nottingham, Engl<strong>and</strong>, obtained an MRI of his<br />

partner’s finger. The next year, Paul Bottomley, a member of Waldo<br />

Hinshaw’s research group at the same university, put his left wrist<br />

into an experimental machine that the group had developed. A<br />

vivid cross section that showed layers of skin, muscle, bone, muscle,<br />

<strong>and</strong> skin, in that order, appeared on the machine’s monitor. Studies<br />

with animals showed no apparent memory or other brain problems.<br />

In 1978, Electrical <strong>and</strong> Musical Industries (EMI), a British corporate<br />

pioneer in electronics that merged with Thorn in 1980, obtained the<br />

first MRI of the human head. It took six minutes.<br />

An MRI of the brain, or any other part of the body, is made possible<br />

by the water content of the body. The gray matter of the brain<br />

contains more water than the white matter does. The blood vessels<br />

<strong>and</strong> the blood itself also have water contents that are different from<br />

those of other parts of the brain. Therefore, the different structures<br />

<strong>and</strong> areas of the brain can be seen clearly in an MRI. Bone contains<br />

very little water, so it does not appear on the monitor. This is why<br />

the skull <strong>and</strong> the backbone cause no interference when the brain or<br />

the spinal cord is viewed.<br />

Every water molecule contains two hydrogen atoms <strong>and</strong> one<br />

oxygen atom. A strong electromagnetic field causes the hydrogen<br />

molecules to line up like marchers in a parade. Radio waves can be<br />

used to change the position of these parallel hydrogen molecules.<br />

When the radio waves are discontinued, a small radio signal is<br />

produced as the molecules return to their marching position. This<br />

distinct radio signal is the basis for the production of the image on<br />

a computer screen.


518 / Nuclear magnetic resonance<br />

Hydrogen was selected for use in MRI work because it is very<br />

abundant in the human body, it is part of the water molecule, <strong>and</strong> it<br />

has the proper magnetic qualities. The nucleus of the hydrogen<br />

atom consists of a single proton, a particle with a positive charge.<br />

The signal from the hydrogen’s proton is comparatively strong.<br />

There are several methods by which the radio signal from the<br />

hydrogen atom can be converted into an image. Each method<br />

uses a computer to create first a two-dimensional, then a threedimensional,<br />

image.<br />

Peter Mansfield’s team at the University of Nottingham holds<br />

the patent for the slice-selection technique that makes it possible to<br />

excite <strong>and</strong> image selectively a specific cross section of the brain or<br />

any other part of the body. This is the key patent in MRI technology.<br />

Damadian was granted a patent that described the use of two coils,<br />

one to drive <strong>and</strong> one to pick up signals across selected portions of<br />

the human body. EMI, the company that introduced the X-ray scanner<br />

for CT images, developed a commercial prototype for the MRI.<br />

The British Technology Group, a state-owned company that helps to<br />

bring innovations to the marketplace, has sixteen separate MRIrelated<br />

patents. Ten years after EMI produced the first image of the<br />

human brain, patents <strong>and</strong> royalties were still being sorted out.<br />

Consequences<br />

MRI technology has revolutionized medical diagnosis, especially<br />

in regard to the brain <strong>and</strong> the spinal cord. For example, in multiple<br />

sclerosis, the loss of the covering on nerve cells can be detected. Tumors<br />

can be identified accurately. The painless <strong>and</strong> noninvasive use<br />

of MRI has almost completely replaced the myelogram, which involves<br />

using a needle to inject dye into the spine.<br />

Although there is every indication that the use of MRI is very<br />

safe, there are some people who cannot benefit from this valuable<br />

tool. Those whose bodies contain metal cannot be placed into the<br />

MRI machine. No one instrument can meet everyone’s needs.<br />

The development of MRI st<strong>and</strong>s as an example of the interaction<br />

of achievements in various fields of science. Fundamental physics,<br />

biochemistry, physiology, electronic image reconstruction, advances<br />

in superconducting wires, the development of computers, <strong>and</strong> ad-


vancements in anatomy all contributed to the development of MRI.<br />

Its development is also the result of international efforts. Scientists<br />

<strong>and</strong> laboratories in Engl<strong>and</strong> <strong>and</strong> the United States pioneered the<br />

technology, but contributions were also made by scientists in France,<br />

Switzerl<strong>and</strong>, <strong>and</strong> Scotl<strong>and</strong>. This kind of interaction <strong>and</strong> cooperation<br />

can only lead to greater underst<strong>and</strong>ing of the human brain.<br />

See also Amniocentesis; CAT scanner; Electrocardiogram; Electroencephalogram;<br />

Mammography; Ultrasound; X-ray image intensifier.<br />

Further Reading<br />

Nuclear magnetic resonance / 519<br />

Elster, Allen D., <strong>and</strong> Jonathan H. Burdette. Questions <strong>and</strong> Answers in<br />

Magnetic Resonance Imaging. 2d ed. St. Louis, Mo.: Mosby, 2001.<br />

Mackay, R. Stuart. Medical Images <strong>and</strong> Displays: Comparisons of Nuclear<br />

Magnetic Resonance, Ultrasound, X-rays, <strong>and</strong> Other Modalities.<br />

New York: Wiley, 1984.<br />

Mattson, James, <strong>and</strong> Merrill Simon. The Story of MRI: The Pioneers of<br />

NMR <strong>and</strong> Magnetic Resonance in Medicine. Jericho, N.Y.: Dean<br />

Books, 1996.<br />

Wakefield, Julie. “The ‘Indomitable’ MRI.” Smithsonian 31, no. 3<br />

(June, 2000).<br />

Wolbarst, Anthony B. Looking Within: How X-ray, CT, MRI, Ultrasound,<br />

<strong>and</strong> Other Medical Images Are Created, <strong>and</strong> How they Help<br />

Physicians Save Lives. Berkeley: University of California Press,<br />

1999.


520<br />

Nuclear power plant<br />

Nuclear power plant<br />

The invention: The first full-scale commercial nuclear power plant,<br />

which gave birth to the nuclear power industry.<br />

The people behind the invention:<br />

Enrico Fermi (1901-1954), an Italian American physicist who<br />

won the 1938 Nobel Prize in Physics<br />

Otto Hahn (1879-1968), a German physical chemist who won the<br />

1944 Nobel Prize in Chemistry<br />

Lise Meitner (1878-1968), an Austrian Swedish physicist<br />

Hyman G. Rickover (1898-1986), a Polish American naval officer<br />

Discovering Fission<br />

Nuclear fission involves the splitting of an atomic nucleus, leading<br />

to the release of large amounts of energy. Nuclear fission was<br />

discovered in Germany in 1938 by Otto Hahn after he had bombarded<br />

uranium with neutrons <strong>and</strong> observed traces of radioactive<br />

barium. When Hahn’s former associate, Lise Meitner, heard of this,<br />

she realized that the neutrons may have split the uranium nuclei<br />

(each of which holds 92 protons) into two smaller nuclei to produce<br />

barium (56 protons) <strong>and</strong> krypton (36 protons). Meitner <strong>and</strong> her<br />

nephew, Otto Robert Frisch, were able to calculate the enormous energy<br />

that would be released in this type of reaction. They published<br />

their results early in 1939.<br />

Nuclear fission was quickly verified in several laboratories, <strong>and</strong><br />

the Danish physicist Niels Bohr soon demonstrated that the rare uranium<br />

235 (U-235) isotope is much more likely to fission than the common<br />

uranium 238 (U-238) isotope, which makes up 99.3 percent of<br />

natural uranium. It was also recognized that fission would produce<br />

additional neutrons that could cause new fissions, producing even<br />

more neutrons <strong>and</strong> thus creating a self-sustaining chain reaction. In<br />

this process, the fissioning of one gram of U-235 would release about<br />

as much energy as the burning of three million tons of coal.<br />

The first controlled chain reaction was demonstrated on December<br />

2, 1942, in a nuclear reactor at the University of Chicago, under


the leadership of Enrico Fermi. He used a graphite moderator to<br />

slow the neutrons by collisions with carbon atoms. “Critical mass”<br />

was achieved when the mass of graphite <strong>and</strong> uranium assembled<br />

was large enough that the number of neutrons not escaping from<br />

the pile would be sufficient to sustain a U-235 chain reaction. Cadmium<br />

control rods could be inserted to absorb neutrons <strong>and</strong> slow<br />

the reaction.<br />

It was also recognized that the U-238 in the reactor would absorb<br />

accelerated neutrons to produce the new element plutonium, which<br />

is also fissionable. During World War II (1939-1945), large reactors<br />

were built to “breed” plutonium, which was easier to separate than<br />

U-235. An experimental breeder reactor at Arco, Idaho, was the first<br />

to use the energy of nuclear fission to produce a small amount of<br />

electricity (about 100 watts) on December 20, 1951.<br />

Nuclear Electricity<br />

Nuclear power plant / 521<br />

Power reactors designed to produce substantial amounts of<br />

electricity use the heat generated by fission to produce steam or<br />

hot gas to drive a turbine connected to an ordinary electric generator.<br />

The first power reactor design to be developed in the United<br />

States was the pressurized water reactor (PWR). In the PWR, water<br />

under high pressure is used both as the moderator <strong>and</strong> as the coolant.<br />

After circulating through the reactor core, the hot pressurized<br />

water flows through a heat exchanger to produce steam. Reactors<br />

moderated by “heavy water” (in which the hydrogen in the water<br />

is replaced with deuterium, which contains an extra neutron) can<br />

operate with natural uranium.<br />

The pressurized water system was used in the first reactor to<br />

produce substantial amounts of power, the experimental Mark I<br />

reactor. It was started up on May 31, 1953, at the Idaho National<br />

Engineering Laboratory. The Mark I became the prototype for the<br />

reactor used in the first nuclear-powered submarine. Under the<br />

leadership of Hyman G. Rickover, who was head of the Division of<br />

Naval Reactors of the Atomic Energy Commission (AEC), Westinghouse<br />

Electric Corporation was engaged to build a PWR system<br />

to power the submarine USS Nautilus. It began sea trials in January<br />

of 1955 <strong>and</strong> ran for two years before refueling.


522 / Nuclear power plant<br />

Cooling towers of a nuclear power plant. (PhotoDisc)<br />

In the meantime, the first experimental nuclear power plant for<br />

generating electricity was completed in the Soviet Union in June of<br />

1954, under the direction of the Soviet physicist Igor Kurchatov. It<br />

produced 5 megawatts of electric power. The first full-scale nuclear<br />

power plant was built in Engl<strong>and</strong> under the direction of the British<br />

nuclear engineer Sir Christopher Hinton. It began producing about<br />

90 megawatts of electric power in October, 1956.


On December 2, 1957, on the fifteenth anniversary of the first controlled<br />

nuclear chain reaction, the Shippingport Atomic Power Station<br />

in Shippingport, Pennsylvania, became the first full-scale commercial<br />

nuclear power plant in the United States. It produced about<br />

60 megawatts of electric power for the Duquesne Light Company until<br />

1964, when its reactor core was replaced, increasing its power to<br />

100 megawatts with a maximum capacity of 150 megawatts.<br />

Consequences<br />

Nuclear power plant / 523<br />

The opening of the Shippingport Atomic Power Station marked<br />

the beginning of the nuclear power industry in the United States,<br />

with all of its glowing promise <strong>and</strong> eventual problems. It was predicted<br />

that electrical energy would become too cheap to meter. The<br />

AEC hoped to encourage the participation of industry, with government<br />

support limited to research <strong>and</strong> development. They encouraged<br />

a variety of reactor types in the hope of extending technical<br />

knowledge.<br />

The Dresden Nuclear Power Station, completed by Commonwealth<br />

Edison in September, 1959, at Morris, Illinois, near Chicago,<br />

was the first full-scale privately financed nuclear power station in<br />

the United States. By 1973, forty-two plants were in operation producing<br />

26,000 megawatts, fifty more were under construction, <strong>and</strong><br />

about one hundred were on order. Industry officials predicted that<br />

50 percent of the nation’s electric power would be nuclear by the<br />

end of the twentieth century.<br />

The promise of nuclear energy has not been completely fulfilled.<br />

Growing concerns about safety <strong>and</strong> waste disposal have led to increased<br />

efforts to delay or block the construction of new plants. The<br />

cost of nuclear plants rose as legal delays <strong>and</strong> inflation pushed costs<br />

higher, so that many in the planning stages could no longer be competitive.<br />

The 1979 Three Mile Isl<strong>and</strong> accident in Pennsylvania <strong>and</strong><br />

the much more serious 1986 Chernobyl accident in the Soviet Union<br />

increased concerns about the safety of nuclear power. Nevertheless,<br />

by 1986, more than one hundred nuclear power plants were operating<br />

in the United States, producing about 60,000 megawatts of<br />

power. More than three hundred reactors in twenty-five countries<br />

provide about 200,000 megawatts of electric power worldwide.


524 / Nuclear power plant<br />

Many believe that, properly controlled, nuclear energy offers a<br />

clean-energy solution to the problem of environmental pollution.<br />

See also Breeder reactor; Compressed-air-accumulating power<br />

plant; Fuel cell; Geothermal power; Nuclear reactor; Solar thermal<br />

engine; Tidal power plant.<br />

Further Reading<br />

Henderson, Harry. Nuclear Power: A Reference H<strong>and</strong>book. Santa<br />

Barbara, Calif.: ABC-CLIO, 2000.<br />

Rockwell, Theodore. The Rickover Effect: The Inside Story of How Admiral<br />

Hyman Rickover Built the Nuclear Navy. New York: J. Wiley,<br />

1995.<br />

Shea, William R. Otto Hahn <strong>and</strong> the Rise of Nuclear Physics. Boston: D.<br />

Reidel, 1983.<br />

Sime, Ruth Lewin. Lise Meitner: A Life in Physics. Berkeley: University<br />

of California Press, 1996.


Nuclear reactor<br />

Nuclear reactor<br />

The invention: The first nuclear reactor to produce substantial<br />

quantities of plutonium, making it practical to produce usable<br />

amounts of energy from a chain reaction.<br />

The people behind the invention:<br />

Enrico Fermi (1901-1954), an American physicist<br />

Martin D. Whitaker (1902-1960), the first director of Oak Ridge<br />

National Laboratory<br />

Eugene Paul Wigner (1902-1995), the director of research <strong>and</strong><br />

development at Oak Ridge<br />

The Technology to End a War<br />

525<br />

The construction of the nuclear reactor at Oak Ridge National<br />

Laboratory in 1943 was a vital part of the Manhattan Project, the effort<br />

by the United States during World War II (1939-1945) to develop<br />

an atomic bomb. The successful operation of that reactor<br />

was a major achievement not only for the project itself but also for<br />

the general development <strong>and</strong> application of nuclear technology.<br />

The first director of the Oak Ridge National Laboratory was Martin<br />

D. Whitaker; the director of research <strong>and</strong> development was Eugene<br />

Paul Wigner.<br />

The nucleus of an atom is made up of protons <strong>and</strong> neutrons. “Fission”<br />

is the process by which the nucleus of certain elements is split<br />

in two by a neutron from some material that emits an occasional<br />

neutron naturally. When an atom splits, two things happen: A tremendous<br />

amount of thermal energy is released, <strong>and</strong> two or three<br />

neutrons, on the average, escape from the nucleus. If all the atoms in<br />

a kilogram of “uranium 235” were to fission, they would produce as<br />

much heat energy as the burning of 3 million kilograms of coal. The<br />

neutrons that are released are important, because if at least one of<br />

them hits another atom <strong>and</strong> causes it to fission (<strong>and</strong> thus to release<br />

more energy <strong>and</strong> more neutrons), the process will continue. It will<br />

become a self-sustaining chain reaction that will produce a continuing<br />

supply of heat.


526 / Nuclear reactor<br />

Inside a reactor, a nuclear chain reaction is controlled so that it<br />

proceeds relatively slowly. The most familiar use for the heat thus<br />

released is to boil water <strong>and</strong> make steam to turn the turbine generators<br />

that produce electricity to serve industrial, commercial, <strong>and</strong><br />

residential needs. The fissioning process in a weapon, however, proceeds<br />

very rapidly, so that all the energy in the atoms is produced<br />

<strong>and</strong> released virtually at once. The first application of nuclear technology,<br />

which used a rapid chain reaction, was to produce the two<br />

atomic bombs that ended World War II.<br />

Breeding Bomb Fuel<br />

The work that began at Oak Ridge in 1943 was made possible by a<br />

major event that took place in 1942. At the University of Chicago,<br />

Enrico Fermi had demonstrated for the first time that it was possible to<br />

achieve a self-sustaining atomic chain reaction. More important, the reaction<br />

could be controlled: It could be started up, it could generate heat<br />

<strong>and</strong> sufficient neutrons to keep itself going, <strong>and</strong> it could be turned off.<br />

That first chain reaction was very slow, <strong>and</strong> it generated very little heat;<br />

but it demonstrated that controlled fission was possible.<br />

Any heat-producing nuclear reaction is an energy conversion<br />

process that requires fuel. There is only one readily fissionable element<br />

that occurs naturally <strong>and</strong> can be used as fuel. It is a form of<br />

uranium called uranium 235. It makes up less than 1 percent of all<br />

naturally occurring uranium. The remainder is uranium 238, which<br />

does not fission readily. Even uranium 235, however, must be enriched<br />

before it can be used as fuel.<br />

The process of enrichment increases the concentration of uranium<br />

235 sufficiently for a chain reaction to occur. Enriched uranium is used<br />

to fuel the reactors used by electric utilities. Also, the much more plentiful<br />

uranium 238 can be converted into plutonium 239, a form of the<br />

human-made element plutonium, which does fission readily. That<br />

conversion process is the way fuel is produced for a nuclear weapon.<br />

Therefore, the major objective of the Oak Ridge effort was to develop a<br />

pilot operation for separating plutonium from the uranium in which it<br />

was produced. Large-scale plutonium production, which had never<br />

been attempted before, eventually would be done at the Hanford Engineer<br />

Works in Washington. First, however, plutonium had to be pro-


duced successfully on a small scale at Oak Ridge.<br />

The reactor was started up on November 4, 1943. By March 1,<br />

1944, the Oak Ridge laboratory had produced several grams of plutonium.<br />

The material was sent to the Los Alamos laboratory in New<br />

Mexico for testing. By July, 1944, the reactor operated at four times<br />

its original power level. By the end of that year, however, plutonium<br />

production at Oak Ridge had ceased, <strong>and</strong> the reactor thereafter was<br />

used principally to produce radioisotopes for physical <strong>and</strong> biological<br />

research <strong>and</strong> for medical treatment. Ultimately, the Hanford Engineer<br />

Works’ reactors produced the plutonium for the bomb that<br />

was dropped on Nagasaki, Japan, on August 9, 1945.<br />

The original objectives for which Oak Ridge had been built had<br />

been achieved, <strong>and</strong> subsequent activity at the facility was directed<br />

toward peacetime missions that included basic studies of the structure<br />

of matter.<br />

Impact<br />

Nuclear reactor / 527<br />

Part of the Oak Ridge National Laboratory, where plutonium was separated to create the<br />

first atomic bomb. (Martin Marietta)<br />

The most immediate impact of the work done at Oak Ridge was<br />

its contribution to ending World War II. When the atomic bombs<br />

were dropped, the war ended, <strong>and</strong> the United States emerged intact.<br />

The immediate <strong>and</strong> long-range devastation to the people of Japan,


528 / Nuclear reactor<br />

however, opened the public’s eyes to the almost unimaginable<br />

death <strong>and</strong> destruction that could be caused by a nuclear war. Fears<br />

of such a war remain to this day, especially as more <strong>and</strong> more nations<br />

develop the technology to build nuclear weapons.<br />

On the other h<strong>and</strong>, great contributions to human civilization<br />

have resulted from the development of nuclear energy. Electric<br />

power generation, nuclear medicine, spacecraft power, <strong>and</strong> ship<br />

propulsion have all profited from the pioneering efforts at the Oak<br />

Ridge National Laboratory. Currently, the primary use of nuclear<br />

energy is to produce electric power. H<strong>and</strong>led properly, nuclear energy<br />

may help to solve the pollution problems caused by the burning<br />

of fossil fuels.<br />

See also Breeder reactor; Compressed-air-accumulating power<br />

plant; Fuel cell; Geothermal power; Heat pump; Nuclear power<br />

plant; Solar thermal engine; Tidal power plant.<br />

Further Reading<br />

Epstein, Sam, Beryl Epstein, <strong>and</strong> Raymond Burns. Enrico Fermi: Father<br />

of Atomic Power. Champaign, Ill.: Garrard, 1970.<br />

Johnson, Lel<strong>and</strong>, <strong>and</strong> Daniel Schaffer. Oak Ridge National Laboratory:<br />

The First Fifty Years. Knoxville: University of Tennessee Press,<br />

1994.<br />

Morgan, K. Z., <strong>and</strong> Ken M. Peterson. The Angry Genie: One Man’s<br />

Walk Through the Nuclear Age. Norman: University of Oklahoma<br />

Press, 1999.<br />

Wagner, Francis S. Eugene P. Wigner, An Architect of the Atomic Age.<br />

Toronto: Rákóczi Foundation, 1981.


Nylon<br />

Nylon<br />

The invention: A resilient, high-strength polymer with applications<br />

ranging from women’s hose to safety nets used in space<br />

flights.<br />

The people behind the invention:<br />

Wallace Hume Carothers (1896-1937), an American organic<br />

chemist<br />

Charles M. A. Stine (1882-1954), an American chemist <strong>and</strong><br />

director of chemical research at Du Pont<br />

Elmer Keiser Bolton (1886-1968), an American industrial<br />

chemist<br />

Pure Research<br />

529<br />

In the twentieth century, American corporations created industrial<br />

research laboratories. Their directors became the organizers of<br />

inventions, <strong>and</strong> their scientists served as the sources of creativity.<br />

The research program of E. I. Du Pont de Nemours <strong>and</strong> Company<br />

(Du Pont), through its most famous invention—nylon—became the<br />

model for scientifically based industrial research in the chemical<br />

industry.<br />

During World War I (1914-1918), Du Pont tried to diversify,<br />

concerned that after the war it would not be able to exp<strong>and</strong> with<br />

only explosives as a product. Charles M. A. Stine, Du Pont’s director<br />

of chemical research, proposed that Du Pont should move<br />

into fundamental research by hiring first-rate academic scientists<br />

<strong>and</strong> giving them freedom to work on important problems in<br />

organic chemistry. He convinced company executives that a program<br />

to explore the fundamental science underlying Du Pont’s<br />

technology would ultimately result in discoveries of value to the<br />

company. In 1927, Du Pont gave him a new laboratory for research.<br />

Stine visited universities in search of brilliant, but not-yetestablished,<br />

young scientists. He hired Wallace Hume Carothers.<br />

Stine suggested that Carothers do fundamental research in polymer<br />

chemistry.


530 / Nylon<br />

Before the 1920’s, polymers were a mystery to chemists. Polymeric<br />

materials were the result of ingenious laboratory practice,<br />

<strong>and</strong> this practice ran far ahead of theory <strong>and</strong> underst<strong>and</strong>ing. German<br />

chemists debated whether polymers were aggregates of smaller<br />

units held together by some unknown special force or genuine molecules<br />

held together by ordinary chemical bonds.<br />

German chemist Hermann Staudinger asserted that they were<br />

large molecules with endlessly repeating units. Carothers shared<br />

this view, <strong>and</strong> he devised a scheme to prove it by synthesizing very<br />

large molecules by simple reactions in such a way as to leave no<br />

doubt about their structure. Carothers’s synthesis of polymers revealed<br />

that they were ordinary molecules but giant in size.<br />

The Longest Molecule<br />

In April, 1930, Carothers’s research group produced two major<br />

innovations: neoprene synthetic rubber <strong>and</strong> the first laboratorysynthesized<br />

fiber. Neither result was the goal of their research. Neoprene<br />

was an incidental discovery during a project to study short<br />

polymers of acetylene. During experimentation, an unexpected substance<br />

appeared that polymerized spontaneously. Carothers studied<br />

its chemistry <strong>and</strong> developed the process into the first successful synthetic<br />

rubber made in the United States.<br />

The other discovery was an unexpected outcome of the group’s<br />

project to synthesize polyesters by the reaction of acids <strong>and</strong> alcohols.<br />

Their goal was to create a polyester that could react indefinitely<br />

to form a substance with high molecular weight. The scientists<br />

encountered a molecular weight limit of about 5,000 units to the<br />

size of the polyesters, until Carothers realized that the reaction also<br />

produced water, which was decomposing polyesters back into acid<br />

<strong>and</strong> alcohol. Carothers <strong>and</strong> his associate Julian Hill devised an apparatus<br />

to remove the water as it formed. The result was a polyester<br />

with a molecular weight of more than 12,000, far higher than any<br />

previous polymer.<br />

Hill, while removing a sample from the apparatus, found that he<br />

could draw it out into filaments that on cooling could be stretched to<br />

form very strong fibers. This procedure, called “cold-drawing,” oriented<br />

the molecules from a r<strong>and</strong>om arrangement into a long, linear


one of great strength. The polyester fiber, however, was unsuitable<br />

for textiles because of its low melting point.<br />

In June, 1930, Du Pont promoted Stine; his replacement as research<br />

director was Elmer Keiser Bolton. Bolton wanted to control<br />

fundamental research more closely, relating it to projects that would<br />

pay off <strong>and</strong> not allowing the research group freedom to pursue<br />

purely theoretical questions.<br />

Despite their differences, Carothers <strong>and</strong> Bolton shared an interest<br />

in fiber research. On May 24, 1934, Bolton’s assistant Donald<br />

Coffman “drew” a strong fiber from a new polyamide. This was the<br />

first nylon fiber, although not the one commercialized by Du Pont.<br />

The nylon fiber was high-melting <strong>and</strong> tough, <strong>and</strong> it seemed that a<br />

practical synthetic fiber might be feasible.<br />

By summer of 1934, the fiber project was the heart of the research<br />

group’s activity. The one that had the best fiber properties was nylon<br />

5-10, the number referring to the number of carbon atoms in the<br />

amine <strong>and</strong> acid chains. Yet the nylon 6-6 prepared on February 28,<br />

1935, became Du Pont’s nylon. Nylon 5-10 had some advantages,<br />

but Bolton realized that its components would be unsuitable for<br />

commercial production, whereas those of nylon 6-6 could be obtained<br />

from chemicals in coal.<br />

A determined Bolton pursued nylon’s practical development,<br />

a process that required nearly four years. Finally, in April, 1937,<br />

Du Pont filed a patent for synthetic fibers, which included a statement<br />

by Carothers that there was no previous work on polyamides;<br />

this was a major breakthrough. After Carothers’s death<br />

on April 29, 1937, the patent was issued posthumously <strong>and</strong> assigned<br />

to Du Pont. Du Pont made the first public announcement<br />

of nylon on October 27, 1938.<br />

Impact<br />

Nylon / 531<br />

Nylon was a generic term for polyamides, <strong>and</strong> several types of<br />

nylon became commercially important in addition to nylon 6-6.<br />

These nylons found widespread use as both a fiber <strong>and</strong> a moldable<br />

plastic. Since it resisted abrasion <strong>and</strong> crushing, was nonabsorbent,<br />

was stronger than steel on a weight-for-weight basis, <strong>and</strong> was almost<br />

nonflammable, it embraced an astonishing range of uses: in


532 / Nylon<br />

laces, screens, surgical sutures, paint, toothbrushes, violin strings,<br />

coatings for electrical wires, lingerie, evening gowns, leotards, athletic<br />

equipment, outdoor furniture, shower curtains, h<strong>and</strong>bags, sails,<br />

luggage, fish nets, carpets, slip covers, bus seats, <strong>and</strong> even safety<br />

nets on the space shuttle.<br />

The invention of nylon stimulated notable advances in the chemistry<br />

<strong>and</strong> technology of polymers. Some historians of technology<br />

have even dubbed the postwar period as the “age of plastics,” the<br />

age of synthetic products based on the chemistry of giant molecules<br />

made by ingenious chemists <strong>and</strong> engineers.<br />

The success of nylon <strong>and</strong> other synthetics, however, has come at<br />

a cost. Several environmental problems have surfaced, such as those<br />

created by the nondegradable feature of some plastics, <strong>and</strong> there is<br />

the problem of the increasing utilization of valuable, vanishing resources,<br />

such as petroleum, which contains the essential chemicals<br />

needed to make polymers. The challenge to reuse <strong>and</strong> recycle these<br />

polymers is being addressed by both scientists <strong>and</strong> policymakers.<br />

See also Buna rubber; Neoprene; Orlon; Plastic; Polyester; Polyethylene;<br />

Polystyrene.<br />

Further Reading<br />

Furukawa, Yasu. Inventing Polymer Science: Staudinger, Carothers, <strong>and</strong><br />

the Emergence of Macromolecular Chemistry. Philadelphia: University<br />

of Pennsylvania Press, 1998.<br />

H<strong>and</strong>ley, Susannah. Nylon: The Story of a Fashion Revolution: A Celebration<br />

of Design from Art Silk to Nylon <strong>and</strong> Thinking Fibres. Baltimore:<br />

Johns Hopkins University Press, 1999.<br />

Hermes, Matthew E. Enough for One Lifetime: Wallace Carothers, Inventor<br />

of Rayon. Washington, D.C.: American Chemical Society<br />

<strong>and</strong> the Chemical Heritage Foundation, 1996.<br />

Joyce, Robert M. Elmer Keiser Bolton: June 23, 1886-July 30, 1968.<br />

Washington, D.C.: National Academy Press, 1983.


Oil-well drill bit<br />

Oil-well drill bit<br />

The invention: A rotary cone drill bit that enabled oil-well drillers<br />

to penetrate hard rock formations.<br />

The people behind the invention:<br />

Howard R. Hughes (1869-1924), an American lawyer, drilling<br />

engineer, <strong>and</strong> inventor<br />

Walter B. Sharp (1860-1912), an American drilling engineer,<br />

inventor, <strong>and</strong> partner to Hughes<br />

Digging for Oil<br />

533<br />

A rotary drill rig of the 1990’s is basically unchanged in its essential<br />

components from its earlier versions of the 1900’s. A drill bit is<br />

attached to a line of hollow drill pipe. The latter passes through a<br />

hole on a rotary table, which acts essentially as a horizontal gear<br />

wheel <strong>and</strong> is driven by an engine. As the rotary table turns, so do the<br />

pipe <strong>and</strong> drill bit.<br />

During drilling operations, mud-laden water is pumped under<br />

high pressure down the sides of the drill pipe <strong>and</strong> jets out with great<br />

force through the small holes in the rotary drill bit against the bottom<br />

of the borehole. This fluid then returns outside the drill pipe to<br />

the surface, carrying with it rock material cuttings from below. Circulated<br />

rock cuttings <strong>and</strong> fluids are regularly examined at the surface<br />

to determine the precise type <strong>and</strong> age of rock formation <strong>and</strong> for<br />

signs of oil <strong>and</strong> gas.<br />

A key part of the total rotary drilling system is the drill bit, which<br />

has sharp cutting edges that make direct contact with the geologic<br />

formations to be drilled. The first bits used in rotary drilling were<br />

paddlelike “fishtail” bits, fairly successful for softer formations, <strong>and</strong><br />

tubular coring bits for harder surfaces. In 1893, M. C. Baker <strong>and</strong> C. E.<br />

Baker brought a rotary water-well drill rig to Corsicana, Texas, for<br />

modification to deeper oil drilling. This rig led to the discovery of<br />

the large Corsicana-Powell oil field in Navarro County, Texas. This<br />

success also motivated its operators, the American Well <strong>and</strong> Prospecting<br />

Company, to begin the first large-scale manufacture of rotary<br />

drilling rigs for commercial sale.


534 / Oil-well drill bit<br />

In the earliest rotary drilling for oil, short fishtail bits were the<br />

tool of choice, insofar as they were at that time the best at being able<br />

to bore through a wide range of geologic strata without needing frequent<br />

replacement. Even so, in the course of any given oil well,<br />

many bits were required typically in coastal drilling in the Gulf of<br />

Mexico. Especially when encountering locally harder rock units<br />

such as limestone, dolomite, or gravel beds, fishtail bits would typically<br />

either curl backward or break off in the hole, requiring the<br />

time-consuming work of pulling out all drill pipe <strong>and</strong> “fishing” to<br />

retrieve fragments <strong>and</strong> clear the hole.<br />

Because of the frequent bit wear <strong>and</strong> damage, numerous small<br />

blacksmith shops established themselves near drill rigs, dressing or<br />

sharpening bits with a h<strong>and</strong> forge <strong>and</strong> hammer. Each bit-forging<br />

shop had its own particular way of shaping bits, producing a wide<br />

variety of designs. Nonst<strong>and</strong>ard bit designs were frequently modified<br />

further as experiments to meet the specific requests of local drillers<br />

encountering specific drilling difficulties in given rock layers.<br />

Speeding the Process<br />

In 1907 <strong>and</strong> 1908, patents were obtained in New Jersey <strong>and</strong><br />

Texas for steel, cone-shaped drill bits incorporating a roller-type<br />

coring device with many serrated teeth. Later in 1908, both patents<br />

were bought by lawyer Howard R. Hughes.<br />

Although comparatively weak rocks such as s<strong>and</strong>s, clays, <strong>and</strong><br />

soft shales could be drilled rapidly (at rates exceeding 30 meters per<br />

hour), in harder shales, lime-dolostones, <strong>and</strong> gravels, drill rates of 1<br />

meter per hour or less were not uncommon. Conventional drill bits<br />

of the time had average operating lives of three to twelve hours.<br />

Economic drilling m<strong>and</strong>ated increases in both bit life <strong>and</strong> drilling<br />

rate. Directly motivated by his petroleum prospecting interests,<br />

Hughes <strong>and</strong> his partner, Walter B. Sharp, undertook what were<br />

probably the first recorded systematic studies of drill bit performance<br />

while matched against specific rock layers.<br />

Although many improvements in detail <strong>and</strong> materials have been<br />

made to the Hughes cone bit since its inception in 1908, its basic design<br />

is still used in rotary drilling. One of Hughes’s major innovations<br />

was the much larger size of the cutters, symmetrically distrib-


Howard R. Hughes<br />

Oil-well drill bit / 535<br />

Howard Hughes (1905-1976) is famous for having been one<br />

of the most dashing, innovative, quirky tycoons of the twentieth<br />

century. It all started with his father, Howard R. Hughes. In<br />

fact it was the father’s enterprise, Hughes Tool Company, that<br />

the son took over at age eighteen <strong>and</strong> built into an immense financial<br />

empire based on high-tech products.<br />

The senior Hughes was born in Lancaster, Missouri, in 1869.<br />

He spent his boyhood in Keokuk, Iowa, where his own father<br />

practiced law. He himself studied law at Harvard University<br />

<strong>and</strong> the University of Iowa <strong>and</strong> then joined his father’s practice,<br />

but not for long. In 1901 news came of a big oil strike near Beaumont,<br />

Texas. Like hundreds of other ambitious men, Hughes<br />

headed there. By 1906 he had immersed himself in the technical<br />

problems of drilling <strong>and</strong> began experimenting to improve drill<br />

bits. He produced a wooden model of the roller-type drill two<br />

years later while in Oil City, Louisiana. With business associate<br />

Walter Sharp he successfully tested a prototype in an oil well in<br />

the Goose Creek field near Houston. It drilled faster <strong>and</strong> more<br />

efficiently than those then in use.<br />

Hughes <strong>and</strong> Sharp opened the Sharp-Hughes Tool Company<br />

to manufacture the drills <strong>and</strong> related equipment, <strong>and</strong> their<br />

products quickly became the industry st<strong>and</strong>ard. A shrewd business<br />

strategist, Hughes leased, rather than sold, his drill bits<br />

for $30,000 per well, retaining his patents to preserve his monopoly<br />

over the rotary drill technology. After Sharp died in<br />

1912, Hughes changed the company to the Hughes Tool Company.<br />

When Hughes himself died in 1924, he left his son, then a<br />

student at Rice Institute (later Rice University), the company<br />

<strong>and</strong> a million-dollar fortune, which Hughes junior would eventually<br />

multiply hundreds of times over.<br />

uted as a large number of small individual teeth on the outer face of<br />

two or more cantilevered bearing pins. In addition, “hard facing”<br />

was employed to drill bit teeth to increase usable life. Hard facing is<br />

a metallurgical process basically consisting of wedding a thin layer<br />

of a hard metal or alloy of special composition to a metal surface to<br />

increase its resistance to abrasion <strong>and</strong> heat. A less noticeable but<br />

equally essential innovation, not included in other drill bit patents,


536 / Oil-well drill bit<br />

was an ingeniously designed gauge surface that provided strong<br />

uniform support for all the drill teeth. The force-fed oil lubrication<br />

was another new feature included in Hughes’s patent <strong>and</strong> prototypes,<br />

reducing the power necessary to rotate the bit by 50 percent<br />

over that of prior mud or water lubricant designs.<br />

Impact<br />

In 1925, the first superhard facing was used on cone drill bits. In<br />

addition, the first so-called self-cleaning rock bits appeared from<br />

Hughes, with significant advances in roller bearings <strong>and</strong> bit tooth<br />

shape translating into increased drilling efficiency. The much larger<br />

teeth were more adaptable to drilling in a wider variety of geological<br />

formations than earlier models. In 1928, tungsten carbide was<br />

introduced as an additional bit facing hardener by Hughes metallurgists.<br />

This, together with other improvements, resulted in the<br />

Hughes ACME tooth form, which has been in almost continuous<br />

use since 1926.<br />

Many other drilling support technologies, such as drilling mud,<br />

mud circulation pumps, blowout detectors <strong>and</strong> preventers, <strong>and</strong><br />

pipe properties <strong>and</strong> connectors have enabled rotary drilling rigs to<br />

reach new depths (exceeding 5 kilometers in 1990). The successful<br />

experiments by Hughes in 1908 were critical initiators of these developments.<br />

See also Geothermal power; Steelmaking process; Thermal<br />

cracking process.<br />

Further Reading<br />

Brantly, John Edward. History of Oil Well Drilling. Houston: Gulf<br />

Publishing, 1971.<br />

Charlez, Philippe A. Rock Mechanics. Vol. 2: Petroleum Applications.<br />

Paris: Editions Technip, 1997.<br />

Rao, Karanam Umamaheshwar, <strong>and</strong> Misra Banabihari. Principles of<br />

Rock Drilling. Brookfield, Vt.: Balkema, 1998.


Optical disk<br />

Optical disk<br />

The invention: A nonmagnetic storage medium for computers that<br />

can hold much greater quantities of data than similar size magnetic<br />

media, such as hard <strong>and</strong> floppy disks.<br />

The people behind the invention:<br />

Klaas Compaan, a Dutch physicist<br />

Piet Kramer, head of Philips’ optical research laboratory<br />

Lou F. Ottens, director of product development for Philips’<br />

musical equipment division<br />

George T. de Kruiff, manager of Philips’ audio-product<br />

development department<br />

Joop Sinjou, a Philips project leader<br />

Holograms Can Be Copied Inexpensively<br />

537<br />

Holography is a lensless photographic method that uses laser<br />

light to produce three-dimensional images. This is done by splitting<br />

a laser beam into two beams. One of the beams is aimed at the object<br />

whose image is being reproduced so that the laser light will reflect<br />

from the object <strong>and</strong> strike a photographic plate or film. The second<br />

beam of light is reflected from a mirror near the object <strong>and</strong> also<br />

strikes the photographic plate or film. The “interference pattern,”<br />

which is simply the pattern created by the differences between the<br />

two reflected beams of light, is recorded on the photographic surface.<br />

The recording that is made in this way is called a “hologram.”<br />

When laser light or white light strikes the hologram, an image is created<br />

that appears to be a three-dimensional object.<br />

Early in 1969, Radio Corporation of America (RCA) engineers<br />

found a way to copy holograms inexpensively by impressing interference<br />

patterns on a nickel sheet that then became a mold from<br />

which copies could be made. Klaas Compaan, a Dutch physicist,<br />

learned of this method <strong>and</strong> had the idea that images could be recorded<br />

in a similar way <strong>and</strong> reproduced on a disk the size of a phonograph<br />

record. Once the images were on the disk, they could be<br />

projected onto a screen in any sequence. Compaan saw the possibilities<br />

of such a technology in the fields of training <strong>and</strong> education.


538 / Optical disk<br />

Computer Data Storage Breakthrough<br />

In 1969, Compaan shared his idea with Piet Kramer, who was the<br />

head of Philips’ optical research laboratory. The idea intrigued<br />

Kramer. Between 1969 <strong>and</strong> 1971, Compaan spent much of his time<br />

working on the development of a prototype.<br />

By September, 1971, Compaan <strong>and</strong> Kramer, together with a h<strong>and</strong>ful<br />

of others, had assembled a prototype that could read a black<strong>and</strong>-white<br />

video signal from a spinning glass disk. Three months<br />

later, they demonstrated it for senior managers at Philips. In July,<br />

1972, a color prototype was demonstrated publicly. After the demonstration,<br />

Philips began to consider putting sound, rather than images,<br />

on the disks. The main attraction of that idea was that the 12inch<br />

(305-millimeter) disks would hold up to forty-eight hours of<br />

music. Very quickly, however, Lou F. Ottens, director of product development<br />

for Philips’ musical equipment division, put an end to<br />

any talk of a long-playing audio disk.<br />

Ottens had developed the cassette-tape cartridge in the 1960’s.<br />

He had plenty of experience with the recording industry, <strong>and</strong> he had<br />

no illusions that the industry would embrace that new medium. He<br />

was convinced that the recording companies would consider fortyeight<br />

hours of music unmarketable. He also knew that any new<br />

medium would have to offer a dramatic improvement over existing<br />

vinyl records.<br />

In 1974, only three years after the first microprocessor (the basic<br />

element of computers) was invented, designing a digital consumer<br />

product—rather than an analog product such as those that were already<br />

commonly accepted—was risky. (Digital technology uses<br />

numbers to represent information, whereas analog technology represents<br />

information by mechanical or physical means.) When<br />

George T. de Kruiff became Ottens’s manager of audio-product<br />

development in June, 1974, he was amazed that there were no<br />

digital circuit specialists in the audio department. De Kruiff recruited<br />

new digital engineers, bought computer-aided design<br />

tools, <strong>and</strong> decided that the project should go digital.<br />

Within a few months, Ottens’s engineers had rigged up a digital<br />

system. They used an audio signal that was representative of an<br />

acoustical wave, sampled it to change it to digital form, <strong>and</strong> en-


Optical disk / 539<br />

coded it as a series of pulses. On the disk itself, they varied the<br />

length of the “dimples” that were used to represent the sound so<br />

that the rising <strong>and</strong> falling edges of the series of pulses corresponded<br />

to the dimples’ walls. A helium-neon laser was reflected from<br />

the dimples to photodetectors that were connected to a digital-toanalog<br />

converter.<br />

In 1978, Philips demonstrated a prototype for Polygram (a West<br />

German company) <strong>and</strong> persuaded Polygram to develop an inexpensive<br />

disk material with the appropriate optical qualities. Most<br />

important was that the material could not warp. Polygram spent<br />

about $150,000 <strong>and</strong> three months to develop the disk. In addition, it<br />

was determined that the gallium-arsenide (GaAs) laser would be<br />

used in the project. Sharp Corporation agreed to manufacture a<br />

long-life GaAs diode laser to Philips’ specifications.<br />

The optical-system designers wanted to reduce the number<br />

of parts in order to decrease manufacturing costs <strong>and</strong> improve<br />

reliability. Therefore, the lenses were simplified <strong>and</strong> considerable<br />

work was devoted to developing an error-correction code.<br />

Philips <strong>and</strong> Sony engineers also worked together to create a st<strong>and</strong>ard<br />

format. In 1983, Philips made almost 100,000 units of optical<br />

disks.<br />

Optical Disk<br />

An optical memory.<br />

Laser Beam<br />

Direction of<br />

Rotation<br />

Direction of Laser


540 / Optical disk<br />

Consequences<br />

In 1983, one of the most successful consumer products of all time<br />

was introduced: the optical-disk system. The overwhelming success<br />

of optical-disk reproduction led to the growth of a multibillion-dollar<br />

industry around optical information <strong>and</strong> laid the groundwork<br />

for a whole crop of technologies that promise to revolutionize computer<br />

data storage. Common optical-disk products are the compact<br />

disc (CD), the compact disc read-only memory (CD-ROM), the<br />

write-once, read-many (WORM) erasable disk, <strong>and</strong> CD-I (interactive<br />

CD).<br />

The CD-ROM, the WORM, <strong>and</strong> the erasable optical disk, all of<br />

which are used in computer applications, can hold more than 550<br />

megabytes, from 200 to 800 megabytes, <strong>and</strong> 650 megabytes of data,<br />

respectively.<br />

The CD-ROM is a nonerasable disc that is used to store computer<br />

data. After the write-once operation is performed, a WORM becomes<br />

a read-only optical disk. An erasable optical disk can be<br />

erased <strong>and</strong> rewritten easily. CD-ROMs, coupled with expert-system<br />

technology, are expected to make data retrieval easier. The CD-ROM,<br />

the WORM, <strong>and</strong> the erasable optical disk may replace magnetic<br />

hard <strong>and</strong> floppy disks as computer data storage devices.<br />

See also Bubble memory; Compact disc; Computer chips;<br />

Floppy disk; Hard disk; Holography.<br />

Further Reading<br />

Fox, Barry. “Head to Head in the Recording Wars.” New Scientist<br />

136, no. 1843 (October 17, 1992).<br />

Goff, Leslie. “Philips’ Eye on the Future.” Computerworld 33, no. 32<br />

(August 9, 1999).<br />

Kolodziej, Stan. “Optical Discs: The Dawn of a New Era in Mass<br />

Storage.” Canadian Datasystems 14, no. 9 (September, 1982). 36-39.<br />

Savage, Maria. “Beyond Film.” Bulletin of the American Society for Information<br />

Science 7, no. 1 (October, 1980).


Orlon<br />

Orlon<br />

The invention: A synthetic fiber made from polyacrylonitrile that<br />

has become widely used in textiles <strong>and</strong> in the preparation of<br />

high-strength carbon fibers.<br />

The people behind the invention:<br />

Herbert Rein (1899-1955), a German chemist<br />

Ray C. Houtz (1907- ), an American chemist<br />

A Difficult Plastic<br />

541<br />

“Polymers” are large molecules that are made up of chains of<br />

many smaller molecules, called “monomers.” Materials that are<br />

made of polymers are also called polymers, <strong>and</strong> some polymers,<br />

such as proteins, cellulose, <strong>and</strong> starch, occur in nature. Most polymers,<br />

however, are synthetic materials, which means that they were<br />

created by scientists.<br />

The twenty-year period beginning in 1930 was the age of great<br />

discoveries in polymers by both chemists <strong>and</strong> engineers. During<br />

this time, many of the synthetic polymers, which are also known as<br />

plastics, were first made <strong>and</strong> their uses found. Among these polymers<br />

were nylon, polyester, <strong>and</strong> polyacrylonitrile. The last of these<br />

materials, polyacrylonitrile (PAN), was first synthesized by German<br />

chemists in the late 1920’s. They linked more than one thous<strong>and</strong><br />

of the small, organic molecules of acrylonitrile to make a polymer.<br />

The polymer chains of this material had the properties that<br />

were needed to form strong fibers, but there was one problem. Instead<br />

of melting when heated to a high temperature, PAN simply<br />

decomposed. This made it impossible, with the technology that existed<br />

then, to make fibers.<br />

The best method available to industry at that time was the process<br />

of melt spinning, in which fibers were made by forcing molten<br />

polymer through small holes <strong>and</strong> allowing it to cool. Researchers realized<br />

that, if PAN could be put into a solution, the same apparatus<br />

could be used to spin PAN fibers. Scientists in Germany <strong>and</strong> the<br />

United States tried to find a solvent or liquid that would dissolve<br />

PAN, but they were unsuccessful until World War II began.


542 / Orlon<br />

Fibers for War<br />

In 1938, the German chemist Walter Reppe developed a new<br />

class of organic solvents called “amides.” These new liquids were<br />

able to dissolve many materials, including some of the recently discovered<br />

polymers. When World War II began in 1940, both the Germans<br />

<strong>and</strong> the Allies needed to develop new materials for the war effort.<br />

Materials such as rubber <strong>and</strong> fibers were in short supply. Thus,<br />

there was increased governmental support for chemical <strong>and</strong> industrial<br />

research on both sides of the war. This support was to result in<br />

two independent solutions to the PAN problem.<br />

In 1942, Herbert Rein, while working for I. G. Farben in Germany,<br />

discovered that PAN fibers could be produced from a solution of<br />

polyacrylonitrile dissolved in the newly synthesized solvent dimethylformamide.<br />

At the same time Ray C. Houtz, who was working for E.<br />

I. Du Pont de Nemours in Wilmington, Delaware, found that the related<br />

solvent dimethylacetamide would also form excellent PAN fibers.<br />

His work was patented, <strong>and</strong> some fibers were produced for use<br />

by the military during the war. In 1950, Du Pont began commercial<br />

production of a form of polyacrylonitrile fibers called Orlon. The<br />

Monsanto Company followed with a fiber called Acrilon in 1952, <strong>and</strong><br />

other companies began to make similar products in 1958.<br />

There are two ways to produce PAN fibers. In both methods,<br />

polyacrylonitrile is first dissolved in a suitable solvent. The solution<br />

is next forced through small holes in a device called a “spinneret.”<br />

The solution emerges from the spinneret as thin streams of a thick,<br />

gooey liquid. In the “wet spinning method,” the streams then enter<br />

another liquid (usually water or alcohol), which extracts the solvent<br />

from the solution, leaving behind the pure PAN fiber. After air drying,<br />

the fiber can be treated like any other fiber. The “dry spinning<br />

method” uses no liquid. Instead, the solvent is evaporated from the<br />

emerging streams by means of hot air, <strong>and</strong> again the PAN fiber is left<br />

behind.<br />

In 1944, another discovery was made that is an important part of<br />

the polyacrylonitrile fiber story. W. P. Coxe of Du Pont <strong>and</strong> L. L.<br />

Winter at Union Carbide Corporation found that, when PAN fibers<br />

are heated under certain conditions, the polymer decomposes <strong>and</strong><br />

changes into graphite (one of the elemental forms of carbon) but still


keeps its fiber form. In contrast to most forms of graphite, these fibers<br />

were exceptionally strong. These were the first carbon fibers<br />

ever made. Originally known as “black Orlon,” they were first produced<br />

commercially by the Japanese in 1964, but they were too<br />

weak to find many uses. After new methods of graphitization were<br />

developed jointly by labs in Japan, Great Britain, <strong>and</strong> the United<br />

States, the strength of the carbon fibers was increased, <strong>and</strong> the fibers<br />

began to be used in many fields.<br />

Impact<br />

Orlon / 543<br />

As had been predicted earlier, PAN fibers were found to have<br />

some very useful properties. Their discovery <strong>and</strong> commercialization<br />

helped pave the way for the acceptance <strong>and</strong> wide use of polymers.<br />

The fibers derive their properties from the stiff, rodlike structure<br />

of polyacrylonitrile. Known as acrylics, these fibers are more<br />

durable than cotton, <strong>and</strong> they are the best alternative to wool for<br />

sweaters. Acrylics are resistant to heat <strong>and</strong> chemicals, can be dyed<br />

easily, resist fading or wrinkling, <strong>and</strong> are mildew-resistant. Thus, after<br />

their introduction, PAN fibers were very quickly made into<br />

yarns, blankets, draperies, carpets, rugs, sportswear, <strong>and</strong> various<br />

items of clothing. Often, the fibers contain small amounts of other<br />

polymers that give them additional useful properties.<br />

A significant amount of PAN fiber is used in making carbon fibers.<br />

These lightweight fibers are stronger for their weight than any<br />

known material, <strong>and</strong> they are used to make high-strength composites<br />

for applications in aerospace, the military, <strong>and</strong> sports. A “fiber<br />

composite” is a material made from two parts: a fiber, such as carbon<br />

or glass, <strong>and</strong> something to hold the fibers together, which is<br />

usually a plastic called an “epoxy.” Fiber composites are used in<br />

products that require great strength <strong>and</strong> light weight. Their applications<br />

can be as ordinary as a tennis racket or fishing pole or as exotic<br />

as an airplane tail or the body of a spacecraft.<br />

See also Buna rubber; Neoprene; Nylon; Plastic; Polyester; Polyethylene;<br />

Polystyrene.


544 / Orlon<br />

Further Reading<br />

H<strong>and</strong>ley, Susannah. Nylon: The Story of a Fashion Revolution: A Celebration<br />

of Design from Art Silk to Nylon <strong>and</strong> Thinking Fibres. Baltimore:<br />

Johns Hopkins University Press, 1999.<br />

Hunter, David. “Du Pont Bids Adieu to Acrylic Fibers.” Chemical<br />

Week 146, no. 24 (June 20, 1990).<br />

Kornheiser, Tony. “So Long, Orlon.” Washington Post (June 13, 1990).<br />

Seymour, Raymond Benedict, <strong>and</strong> Roger Stephen Porter. Manmade<br />

Fibers: Their Origin <strong>and</strong> Development. New York: Elsevier Applied<br />

Science, 1993.


Pacemaker<br />

Pacemaker<br />

The invention: A small device using transistor circuitry that regulates<br />

the heartbeat of the patient in whom it is surgically emplaced.<br />

The people behind the invention:<br />

Ake Senning (1915- ), a Swedish physician<br />

Rune Elmquist, co-inventor of the first pacemaker<br />

Paul Maurice Zoll (1911- ), an American cardiologist<br />

Cardiac Pacing<br />

545<br />

The fundamentals of cardiac electrophysiology (the electrical activity<br />

of the heart) were determined during the eighteenth century;<br />

the first successful cardiac resuscitation by electrical stimulation occurred<br />

in 1774. The use of artificial pacemakers for resuscitation was<br />

demonstrated in 1929 by Mark Lidwell. Lidwell <strong>and</strong> his coworkers<br />

developed a portable apparatus that could be connected to a power<br />

source. The pacemaker was used successfully on several stillborn<br />

infants after other methods of resuscitation failed. Nevertheless,<br />

these early machines were unreliable.<br />

Ake Senning’s first experience with the effect of electrical stimulation<br />

on cardiac physiology was memorable; grasping a radio<br />

ground wire, Senning felt a brief episode of ventricular arrhythmia<br />

(irregular heartbeat). Later, he was able to apply a similar electrical<br />

stimulation to control a heartbeat during surgery.<br />

The principle of electrical regulation of the heart was valid. It was<br />

shown that pacemakers introduced intravenously into the sinus<br />

node area of a dog’s heart could be used to control the heartbeat<br />

rate. Although Paul Maurice Zoll utilized a similar apparatus in<br />

several patients with cardiac arrhythmia, it was not appropriate for<br />

extensive clinical use; it was large <strong>and</strong> often caused unpleasant sensations<br />

or burns. In 1957, however, Ake Senning observed that attaching<br />

stainless steel electrodes to a child’s heart made it possible<br />

to regulate the heart’s rate of contraction. Senning considered this to<br />

represent the beginning of the era of clinical pacing.


546 / Pacemaker<br />

Development of Cardiac Pacemakers<br />

Senning’s observations of the successful use of the cardiac pacemaker<br />

had allowed him to identify the problems inherent in the device.<br />

He realized that the attachment of the device to the lower, ventricular<br />

region of the heart made possible more reliable control, but<br />

other problems remained unsolved. It was inconvenient, for example,<br />

to carry the machine externally; a cord was wrapped around the<br />

patient that allowed the pacemaker to be recharged, which had to be<br />

done frequently. Also, for unknown reasons, heart resistance would<br />

increase with use of the pacemaker, which meant that increasingly<br />

large voltages had to be used to stimulate the heart. Levels as high<br />

as 20 volts could cause quite a “start” in the patient. Furthermore,<br />

there was a continuous threat of infection.<br />

In 1957, Senning <strong>and</strong> his colleague Rune Elmquist developed a<br />

pacemaker that was powered by rechargeable nickel-cadmium batteries,<br />

which had to be recharged once a month. Although Senning<br />

<strong>and</strong> Elmquist did not yet consider the pacemaker ready for human<br />

testing, fate intervened. Aforty-three-year-old man was admitted to<br />

the hospital suffering from an atrioventricular block, an inability of<br />

the electrical stimulus to travel along the conductive fibers of the<br />

“bundle of His” (a b<strong>and</strong> of cardiac muscle fibers). As a result of this<br />

condition, the patient required repeated cardiac resuscitation. Similar<br />

types of heart block were associated with a mortality rate higher<br />

than 50 percent per year <strong>and</strong> nearly 95 percent over five years.<br />

Senning implanted two pacemakers (one failed) into the myocardium<br />

of the patient’s heart, one of which provided a regulatory<br />

rate of 64 beats per minute. Although the pacemakers required periodic<br />

replacement, the patient remained alive <strong>and</strong> active for twenty<br />

years. (He later became president of the Swedish Association for<br />

Heart <strong>and</strong> Lung Disease.)<br />

During the next five years, the development of more reliable <strong>and</strong><br />

more complex pacemakers continued, <strong>and</strong> implanting the pacemaker<br />

through the vein rather than through the thorax made it simpler<br />

to use the procedure. The first pacemakers were of the “asynchronous”<br />

type, which generated a regular charge that overrode the<br />

natural pacemaker in the heart. The rate could be set by the physician<br />

but could not be altered if the need arose. In 1963, an atrial-


triggered synchronous pacemaker was installed by a Swedish team.<br />

The advantage of this apparatus lay in its ability to trigger a heart<br />

contraction only when the normal heart rhythm was interrupted.<br />

Most of these pacemakers contained a sensing device that detected<br />

the atrial impulse <strong>and</strong> generated an electrical discharge only when<br />

the heart rate fell below 68 to 72 beats per minute.<br />

The biggest problems during this period lay in the size of the<br />

pacemaker <strong>and</strong> the short life of the battery. The expiration of the<br />

electrical impulse sometimes caused the death of the patient. In addition,<br />

the most reliable method of checking the energy level of the<br />

battery was to watch for a decreased pulse rate. As improvements<br />

were made in electronics, the pacemaker became smaller, <strong>and</strong> in<br />

1972, the more reliable lithium-iodine batteries were introduced.<br />

These batteries made it possible to store more energy <strong>and</strong> to monitor<br />

the energy level more effectively. The use of this type of power<br />

source essentially eliminated the battery as the limiting factor in the<br />

longevity of the pacemaker. The period of time that a pacemaker<br />

could operate continuously in the body increased from a period of<br />

days in 1958 to five to ten years by the 1970’s.<br />

Consequences<br />

Pacemaker / 547<br />

The development of electronic heart pacemakers revolutionized<br />

cardiology. Although the initial machines were used primarily to<br />

control cardiac bradycardia, the often life-threatening slowing of<br />

the heartbeat, a wide variety of arrhythmias <strong>and</strong> problems with cardiac<br />

output can now be controlled through the use of these devices.<br />

The success associated with the surgical implantation of pacemakers<br />

is attested by the frequency of its use. Prior to 1960, only three<br />

pacemakers had been implanted. During the 1990’s, however, some<br />

300,000 were implanted each year throughout the world. In the<br />

United States, the prevalence of implants is on the order of 1 per<br />

1,000 persons in the population.<br />

Pacemaker technology continues to improve. Newer models can<br />

sense pH <strong>and</strong> oxygen levels in the blood, as well as respiratory rate.<br />

They have become further sensitized to minor electrical disturbances<br />

<strong>and</strong> can adjust accordingly. The use of easily sterilized circuitry<br />

has eliminated the danger of infection. Once the pacemaker


548 / Pacemaker<br />

has been installed in the patient, the basic electronics require no additional<br />

attention. With the use of modern pacemakers, many forms<br />

of electrical arrhythmias need no longer be life-threatening.<br />

See also Artificial heart; Contact lenses; Coronary artery bypass<br />

surgery; Electrocardiogram; Hearing aid; Heart-lung machine.<br />

Further Reading<br />

Bigelow, W. G. Cold Hearts: The Story of Hypothermia <strong>and</strong> the Pacemaker<br />

in Heart Surgery. Toronto: McClell<strong>and</strong> <strong>and</strong> Stewart, 1984.<br />

Greatbatch, Wilson. The Making of the Pacemaker: Celebrating a Lifesaving<br />

Invention. Amherst, N.Y.: Prometheus Books, 2000.<br />

“The Pacemaker.” Newsweek 130, no. 24A (Winter, 1997/1998).<br />

Thalen, H. J. The Artificial Cardiac Pacemaker: Its History, Development<br />

<strong>and</strong> Clinical Application. London: Heinemann Medical, 1969.


Pap test<br />

Pap test<br />

The invention: A cytologic technique the diagnosing uterine cancer,<br />

the second most common fatal cancer in American women.<br />

The people behind the invention:<br />

George N. Papanicolaou (1883-1962), a Greek-born American<br />

physician <strong>and</strong> anatomist<br />

Charles Stockard (1879-1939), an American anatomist<br />

Herbert Traut (1894-1972), an American gynecologist<br />

Cancer in History<br />

549<br />

Cancer, first named by the ancient Greek physician Hippocrates<br />

of Cos, is one of the most painful <strong>and</strong> dreaded forms of human disease.<br />

It occurs when body cells run wild <strong>and</strong> interfere with the normal<br />

activities of the body. The early diagnosis of cancer is extremely<br />

important because early detection often makes it possible to effect<br />

successful cures. The modern detection of cancer is usually done by<br />

the microscopic examination of the cancer cells, using the techniques<br />

of the area of biology called “cytology, ” or cell biology.<br />

Development of cancer cytology began in 1867, after L. S. Beale<br />

reported tumor cells in the saliva from a patient who was afflicted<br />

with cancer of the pharynx. Beale recommended the use in cancer<br />

detection of microscopic examination of cells shed or removed (exfoliated)<br />

from organs including the digestive, the urinary, <strong>and</strong> the<br />

reproductive tracts. Soon, other scientists identified numerous striking<br />

differences, including cell size <strong>and</strong> shape, the size of cell nuclei,<br />

<strong>and</strong> the complexity of cell nuclei.<br />

Modern cytologic detection of cancer evolved from the work of<br />

George N. Papanicolaou, a Greek physician who trained at the University<br />

of Athens Medical School. In 1913, he emigrated to the<br />

United States.<br />

In 1917, he began studying sex determination of guinea pigs with<br />

Charles Stockard at New York’s Cornell Medical College. Papanicolaou’s<br />

efforts required him to obtain ova (egg cells) at a precise<br />

period in their maturation cycle, a process that required an indicator


550 / Pap test<br />

of the time at which the animals ovulated. In search of this indicator,<br />

Papanicolaou designed a method that involved microscopic examination<br />

of the vaginal discharges from female guinea pigs.<br />

Initially, Papanicolaou sought traces of blood, such as those<br />

seen in the menstrual discharges from both primates <strong>and</strong> humans.<br />

Papanicolaou found no blood in the guinea pig vaginal discharges.<br />

Instead, he noticed changes in the size <strong>and</strong> the shape of the uterine<br />

cells shed in these discharges. These changes recurred in a fifteento-sixteen-day<br />

cycle that correlated well with the guinea pig menstrual<br />

cycle.<br />

“New Cancer Detection Method”<br />

Papanicolaou next extended his efforts to the study of humans.<br />

This endeavor was designed originally to identify whether comparable<br />

changes in the exfoliated cells of the human vagina occurred<br />

in women. Its goal was to gain an underst<strong>and</strong>ing of the human menstrual<br />

cycle. In the course of this work, Papanicolaou observed distinctive<br />

abnormal cells in the vaginal fluid from a woman afflicted<br />

with cancer of the cervix. This led him to begin to attempt to develop<br />

a cytologic method for the detection of uterine cancer, the second<br />

most common type of fatal cancer in American women of the<br />

time.<br />

In 1928, Papanicolaou published his cytologic method of cancer<br />

detection in the Proceedings of the Third Race Betterment Conference,<br />

held in Battle Creek, Michigan. The work was received well by the<br />

news media (for example, the January 5, 1928, New York World credited<br />

him with a “new cancer detection method”). Nevertheless, the<br />

publication—<strong>and</strong> others he produced over the next ten years—was<br />

not very interesting to gynecologists of the time. Rather, they preferred<br />

use of the st<strong>and</strong>ard methodology of uterine cancer diagnosis<br />

(cervical biopsy <strong>and</strong> curettage).<br />

Consequently, in 1932, Papanicolaou turned his energy toward<br />

studying human reproductive endocrinology problems related to<br />

the effects of hormones on cells of the reproductive system. One example<br />

of this work was published in a 1933 issue of The American<br />

Journal of Anatomy, where he described “the sexual cycle in the human<br />

female.” Other such efforts resulted in better underst<strong>and</strong>ing of


eproductive problems that include amenorrhea <strong>and</strong> menopause.<br />

It was not until Papanicolaou’s collaboration with gynecologist<br />

Herbert Traut (beginning in 1939), which led to the publication of<br />

Diagnosis of Uterine Cancer by the Vaginal Smear (1943), that clinical<br />

acceptance of the method began to develop. Their monograph documented<br />

an impressive, irrefutable group of studies of both normal<br />

<strong>and</strong> disease states that included nearly two hundred cases of cancer<br />

of the uterus.<br />

Soon, many other researchers began to confirm these findings;<br />

by 1948, the newly named American Cancer Society noted that the<br />

“Pap” smear seemed to be a very valuable tool for detecting vaginal<br />

cancer. Wide acceptance of the Pap test followed, <strong>and</strong>, beginning<br />

in 1947, hundreds of physicians from all over the world<br />

flocked to Papanicolaou’s course on the subject. They learned his<br />

smear/diagnosis techniques <strong>and</strong> disseminated them around the<br />

world.<br />

Impact<br />

Pap test / 551<br />

The Pap test has been cited by many physicians as being the most<br />

significant <strong>and</strong> useful modern discovery in the field of cancer research.<br />

One way of measuring its impact is the realization that the<br />

test allows the identification of uterine cancer in the earliest stages,<br />

long before other detection methods can be used. Moreover, because<br />

of resultant early diagnosis, the disease can be cured in more<br />

than 80 percent of all cases identified by the test. In addition, Pap<br />

testing allows the identification of cancer of the uterine cervix so<br />

early that its cure rate can be nearly 100 percent.<br />

Papanicolaou extended the use of the smear technique from<br />

examination of vaginal discharges to diagnosis of cancer in many<br />

other organs from which scrapings, washings, <strong>and</strong> discharges<br />

can be obtained. These tissues include the colon, the kidney, the<br />

bladder, the prostate, the lung, the breast, <strong>and</strong> the sinuses. In<br />

most cases, such examination of these tissues has made it possible<br />

to diagnose cancer much sooner than is possible by using<br />

other existing methods. As a result, the smear method has become<br />

a basis of cancer control in national health programs throughout the<br />

world.


552 / Pap test<br />

See also Amniocentesis; Birth control pill; Mammography;<br />

Syphilis test; Ultrasound.<br />

Further Reading<br />

Apgar, Barbara, Lawrence L. Gabel, <strong>and</strong> Robert T. Brown. Oncology.<br />

Philadelphia: W. B. Saunders, 1998.<br />

Entman, Stephen S., <strong>and</strong> Charles B. Rush. Office Gynecology. Philadelphia:<br />

Saunders, 1995.<br />

Glass, Robert H., Michèle G. Curtis, <strong>and</strong> Michael P. Hopkins. Glass’s<br />

Office Gynecology. 5th ed. Baltimore: Williams & Wilkins, 1999.<br />

Rushing, Lynda, <strong>and</strong> Nancy Joste. Abnormal Pap Smears: What Every<br />

Woman Needs to Know. Amherst, N.Y.: Prometheus Books, 2001.


Penicillin<br />

Penicillin<br />

The invention: The first successful <strong>and</strong> widely used antibiotic<br />

drug, penicillin has been called the twentieth century’s greatest<br />

“wonder drug.”<br />

The people behind the invention:<br />

Sir Alex<strong>and</strong>er Fleming (1881-1955), a Scottish bacteriologist,<br />

cowinner of the 1945 Nobel Prize in Physiology or Medicine<br />

Baron Florey (1898-1968), an Australian pathologist, cowinner<br />

of the 1945 Nobel Prize in Physiology or Medicine<br />

Ernst Boris Chain (1906-1979), an émigré German biochemist,<br />

cowinner of the 1945 Nobel Prize in Physiology or Medicine<br />

The Search for the Perfect Antibiotic<br />

553<br />

During the early twentieth century, scientists were aware of antibacterial<br />

substances but did not know how to make full use of them<br />

in the treatment of diseases. Sir Alex<strong>and</strong>er Fleming discovered penicillin<br />

in 1928, but he was unable to duplicate his laboratory results<br />

of its antibiotic properties in clinical tests; as a result, he did not recognize<br />

the medical potential of penicillin. Between 1935 <strong>and</strong> 1940,<br />

penicillin was purified, concentrated, <strong>and</strong> clinically tested by pathologist<br />

Baron Florey, biochemist Ernst Boris Chain, <strong>and</strong> members<br />

of their Oxford research group. Their achievement has since been regarded<br />

as one of the greatest medical discoveries of the twentieth<br />

century.<br />

Florey was a professor at Oxford University in charge of the Sir<br />

William Dunn School of Pathology. Chain had worked for two years<br />

at Cambridge University in the laboratory of Frederick Gowl<strong>and</strong><br />

Hopkins, an eminent chemist <strong>and</strong> discoverer of vitamins. Hopkins<br />

recommended Chain to Florey, who was searching for a c<strong>and</strong>idate<br />

to lead a new biochemical unit in the Dunn School of Pathology.<br />

In 1938, Florey <strong>and</strong> Chain formed a research group to investigate<br />

the phenomenon of antibiosis, or the antagonistic association between<br />

different forms of life. The union of Florey’s medical knowledge<br />

<strong>and</strong> Chain’s biochemical expertise proved to be an ideal com-


554 / Penicillin<br />

bination for exploring the antibiosis potential of penicillin. Florey<br />

<strong>and</strong> Chain began their investigation with a literature search in<br />

which Chain came across Fleming’s work <strong>and</strong> added penicillin to<br />

their list of potential antibiotics.<br />

Their first task was to isolate pure penicillin from a crude liquid<br />

extract. A culture of Fleming’s original Penicillium notatum was<br />

maintained at Oxford <strong>and</strong> was used by the Oxford group for penicillin<br />

production. Extracting large quantities of penicillin from the<br />

medium was a painstaking task, as the solution contained only one<br />

part of the antibiotic in ten million. When enough of the raw juice<br />

was collected, the Oxford group focused on eliminating impurities<br />

<strong>and</strong> concentrating the penicillin. The concentrated liquid was then<br />

freeze-dried, leaving a soluble brown powder.<br />

Spectacular Results<br />

In May, 1940, Florey’s clinical tests of the crude penicillin proved<br />

its value as an antibiotic. Following extensive controlled experiments<br />

with mice, the Oxford group concluded that they had discovered<br />

an antibiotic that was nontoxic <strong>and</strong> far more effective against<br />

pathogenic bacteria than any of the known sulfa drugs. Furthermore,<br />

penicillin was not inactivated after injection into the bloodstream<br />

but was excreted unchanged in the urine. Continued tests<br />

showed that penicillin did not interfere with white blood cells <strong>and</strong><br />

had no adverse effect on living cells. Bacteria susceptible to the antibiotic<br />

included those responsible for gas gangrene, pneumonia,<br />

meningitis, diphtheria, <strong>and</strong> gonorrhea. American researchers later<br />

proved that penicillin was also effective against syphilis.<br />

In January, 1941, Florey injected a volunteer with penicillin<br />

<strong>and</strong> found that there were no side effects to treatment with the<br />

antibiotic. In February, the group began treatment of Albert Alex<strong>and</strong>er,<br />

a forty-three-year-old policeman with a serious staphylococci<br />

<strong>and</strong> streptococci infection that was resisting massive doses of<br />

sulfa drugs. Alex<strong>and</strong>er had been hospitalized for two months after<br />

an infection in the corner of his mouth had spread to his face,<br />

shoulder, <strong>and</strong> lungs. After receiving an injection of 200 milligrams<br />

of penicillin, Alex<strong>and</strong>er showed remarkable progress, <strong>and</strong> for the<br />

next ten days his condition improved. Unfortunately, the Oxford


Sir Alex<strong>and</strong>er Fleming<br />

Penicillin / 555<br />

In 1900 Alex<strong>and</strong>er Fleming (1881-1955) enlisted in the London<br />

Scottish Regiment, hoping to see action in the South African<br />

(Boer) War then underway between Great Britain <strong>and</strong> South<br />

Africa’s independent Afrikaner republics. However, the war<br />

ended too soon for him. So, having come into a small inheritance,<br />

he decided to become a physician instead. Accumulating<br />

honors <strong>and</strong> prizes along the way, he succeeded <strong>and</strong> became a<br />

fellow of the Royal College of Surgeons of Engl<strong>and</strong> in 1909.<br />

His mentor was Sir Almroth Wright. Fleming assisted him at<br />

St. Mary’s Hospital in Paddington, <strong>and</strong> they were at the forefront<br />

of the burgeoning field of bacteriology. They were, for example,<br />

among the first to treat syphilis with the newly discovered<br />

Salvarsan, <strong>and</strong> they championed immunization through<br />

vaccination. With the outbreak of World War I, Fleming followed<br />

Wright into the Royal Army Medical Corps, conducting<br />

research on battlefield wounds at a laboratory near Boulogne.<br />

The infections Fleming inspected horrified him. After the war,<br />

again at St. Mary’s Hospital, he dedicated himself to finding<br />

anti-bacterial agents.<br />

He succeed twice: “lysozyme” in 1921 <strong>and</strong> penicillin in 1928.<br />

To his great disappointment, he was unable to produce pure,<br />

potent concentrations of the drug. That had to await the work of<br />

Ernst Chain <strong>and</strong> Howard Florey in 1940. Meanwhile, Fleming<br />

studied the antibacterial properties of sulfa drugs. He was overjoyed<br />

that Chain <strong>and</strong> Florey succeeded where he had failed <strong>and</strong><br />

that penicillin saved lives during World War II <strong>and</strong> afterward,<br />

but he was taken aback when with them he began to receive a<br />

stream of tributes, awards, decorations, honorary degrees, <strong>and</strong><br />

fellowships, including the Nobel Prize in Physiology or Medicine<br />

in 1945. He was by nature a reserved man.<br />

However, he adjusted to his role as one of the most lionized<br />

medical researchers of his generation <strong>and</strong> continued his work,<br />

both as a professor of medicine at the University of London from<br />

1928 until 1948 <strong>and</strong> as director of the same St. Mary’s Hospital<br />

laboratory where he had started his career (renamed the Wright-<br />

Fleming Institute in 1948). He died soon after he retired in 1955.


556 / Penicillin<br />

production facility was unable to generate enough penicillin to<br />

overcome Alex<strong>and</strong>er’s advanced infection completely, <strong>and</strong> he died<br />

on March 15. A later case involving a fourteen-year-old boy with<br />

staphylococcal septicemia <strong>and</strong> osteomyelitis had a more spectacular<br />

result: The patient made a complete recovery in two months. In<br />

all the early clinical treatments, patients showed vast improvement,<br />

<strong>and</strong> most recovered completely from infections that resisted<br />

all other treatment.<br />

Impact<br />

Penicillin is among the greatest medical discoveries of the twentieth<br />

century. Florey <strong>and</strong> Chain’s chemical <strong>and</strong> clinical research<br />

brought about a revolution in the treatment of infectious disease.<br />

Almost every organ in the body is vulnerable to bacteria. Before<br />

penicillin, the only antimicrobial drugs available were quinine, arsenic,<br />

<strong>and</strong> sulfa drugs. Of these, only the sulfa drugs were useful for<br />

treatment of bacterial infection, but their high toxicity often limited<br />

their use. With this small arsenal, doctors were helpless to treat<br />

thous<strong>and</strong>s of patients with bacterial infections.<br />

The work of Florey <strong>and</strong> Chain achieved particular attention because<br />

of World War II <strong>and</strong> the need for treatments of such scourges<br />

as gas gangrene, which had infected the wounds of numerous<br />

World War I soldiers. With the help of Florey <strong>and</strong> Chain’s Oxford<br />

group, scientists at the U.S. Department of Agriculture’s Northern<br />

Regional Research Laboratory developed a highly efficient method<br />

for producing penicillin using fermentation. After an extended search,<br />

scientists were also able to isolate a more productive penicillin<br />

strain, Penicillium chrysogenum. By 1945, a strain was developed that<br />

produced five hundred times more penicillin than Fleming’s original<br />

mold had.<br />

Penicillin, the first of the “wonder drugs,” remains one of the<br />

most powerful antibiotic in existence. Diseases such as pneumonia,<br />

meningitis, <strong>and</strong> syphilis are still treated with penicillin. Penicillin<br />

<strong>and</strong> other antibiotics also had a broad impact on other fields of medicine,<br />

as major operations such as heart surgery, organ transplants,<br />

<strong>and</strong> management of severe burns became possible once the threat of<br />

bacterial infection was minimized.


Florey <strong>and</strong> Chain received numerous awards for their achievement,<br />

the greatest of which was the 1945 Nobel Prize in Physiology<br />

or Medicine, which they shared with Fleming for his original discovery.<br />

Florey was among the most effective medical scientists of<br />

his generation, <strong>and</strong> Chain earned similar accolades in the science of<br />

biochemistry. This combination of outst<strong>and</strong>ing medical <strong>and</strong> chemical<br />

expertise made possible one of the greatest discoveries in human<br />

history.<br />

See also Antibacterial drugs; Artificial hormone; Genetically engineered<br />

insulin; Polio vaccine (Sabin); Polio vaccine (Salk); Reserpine;<br />

Salvarsan; Tuberculosis vaccine; Typhus vaccine; Yellow fever<br />

vaccine.<br />

Further Reading<br />

Penicillin / 557<br />

Bickel, Lennard. Florey, The Man Who Made Penicillin. Carlton South,<br />

Victoria, Australia: Melbourne University Press, 1995.<br />

Clark, Ronald William. The Life of Ernst Chain: Penicillin <strong>and</strong> Beyond.<br />

New York: St. Martin’s Press, 1985.<br />

Hughes, William Howard. Alex<strong>and</strong>er Fleming <strong>and</strong> Penicillin. Hove:<br />

Wayl<strong>and</strong>, 1979.<br />

Mateles, Richard I. Penicillin: A Paradigm for Biotechnology. Chicago:<br />

Canadida Corporation, 1998.


558<br />

Personal computer<br />

Personal computer<br />

The invention: Originally a tradename of the IBM Corporation,<br />

“personal computer” has become a generic term for increasingly<br />

powerful desktop computing systems using microprocessors.<br />

The people behind the invention:<br />

Tom J. Watson, (1874-1956), the founder of IBM, who set<br />

corporate philosophy <strong>and</strong> marketing principles<br />

Frank Cary (1920- ), the chief executive officer of IBM at the<br />

time of the decision to market a personal computer<br />

John Opel (1925- ), a member of the Corporate Management<br />

Committee<br />

George Belzel, a member of the Corporate Management<br />

Committee<br />

Paul Rizzo, a member of the Corporate Management Committee<br />

Dean McKay (1921- ), a member of the Corporate<br />

Management Committee<br />

William L. Sydnes, the leader of the original twelve-member<br />

design team<br />

Shaking up the System<br />

For many years, the International Business Machines (IBM) Corporation<br />

had been set in its ways, sticking to traditions established<br />

by its founder, Tom Watson, Sr. If it hoped to enter the new microcomputer<br />

market, however, it was clear that only nontraditional<br />

methods would be useful. Apple Computer was already beginning<br />

to make inroads into large IBM accounts, <strong>and</strong> IBM stock was starting<br />

to stagnate on Wall Street. A 1979 Business Week article asked: “Is<br />

IBM just another stodgy, mature company?” The microcomputer<br />

market was expected to grow more than 40 percent in the early<br />

1980’s, but IBM would have to make some changes in order to bring<br />

a competitive personal computer (PC) to the market.<br />

The decision to build <strong>and</strong> market the PC was made by the company’s<br />

Corporate Management Committee (CMC). CMC members<br />

included chief executive officer Frank Cary, John Opel, George


Belzel, Paul Rizzo, Dean McKay, <strong>and</strong> three senior vice presidents. In<br />

July of 1980, Cary gave the order to proceed. He wanted the PC to be<br />

designed <strong>and</strong> built within a year. The CMC approved the initial design<br />

of the PC one month later. Twelve engineers, with William L.<br />

Sydnes as their leader, were appointed as the design team. At the<br />

end of 1980, the team had grown to 150.<br />

Most parts of the PC had to be produced outside IBM. Microsoft<br />

Corporation won the contract to produce the PC’s disk operating system<br />

(DOS) <strong>and</strong> the BASIC (Beginner’s All-purpose Symbolic Instruction<br />

Code) language that is built into the PC’s read-only memory<br />

(ROM). Intel Corporation was chosen to make the PC’s central processing<br />

unit (CPU) chip, the “brains” of the machine. Outside programmers<br />

wrote software for the PC. Ten years earlier, this strategy<br />

would have been unheard of within IBM since all aspects of manufacturing,<br />

service, <strong>and</strong> repair were traditionally taken care of in-house.<br />

Marketing the System<br />

Personal computer / 559<br />

IBM hired a New York firm to design a media campaign for the<br />

new PC. Readers of magazines <strong>and</strong> newspapers saw the character<br />

of Charlie Chaplin advertising the new PC. The machine was delivered<br />

on schedule on August 12, 1981. The price of the basic “system<br />

unit” was $1,565. A system with 64 kilobytes of r<strong>and</strong>om access<br />

memory (RAM), a 13-centimeter single-sided disk drive holding<br />

160 kilobytes, <strong>and</strong> a monitor was priced at about $3,000. A system<br />

with color graphics, a second disk drive, <strong>and</strong> a dot matrix printer<br />

cost about $4,500.<br />

Many useful computer programs had been adapted to the PC<br />

<strong>and</strong> were available when it was introduced. VisiCalc from Personal<br />

Software—the program that is credited with “making” the microcomputer<br />

revolution—was one of the first available. Other packages<br />

included a comprehensive accounting system by Peachtree<br />

Software <strong>and</strong> a word processing package called Easywriter by Information<br />

Unlimited Software.<br />

As the selection of software grew, so did sales. In the first year after<br />

its introduction, the IBM PC went from a zero market share to 28<br />

percent of the market. Yet the credit for the success of the PC does<br />

not go to IBM alone. Many hundreds of companies were able to pro-


560 / Personal computer<br />

duce software <strong>and</strong> hardware for the PC. Within two years, powerful<br />

products such as Lotus Corporation’s 1-2-3 business spreadsheet<br />

had come to the market. Many believed that Lotus 1-2-3 was the<br />

program that caused the PC to become so phenomenally successful.<br />

Other companies produced hardware features (expansion boards)<br />

that increased the PC’s memory storage or enabled the machine to<br />

“drive” audiovisual presentations such as slide shows. Business especially<br />

found the PC to be a powerful tool. The PC has survived because<br />

of its expansion capability.<br />

IBM has continued to upgrade the PC. In 1983, the PC/XT was<br />

introduced. It had more expansion slots <strong>and</strong> a fixed disk offering 10<br />

million bytes of storage for programs <strong>and</strong> data. Many of the companies<br />

that made expansion boards found themselves able to make<br />

whole PCs. An entire range of PC-compatible systems was introduced<br />

to the market, many offering features that IBM did not include<br />

in the original PC. The original PC has become a whole family<br />

of computers, sold by both IBM <strong>and</strong> other companies. The hardware<br />

<strong>and</strong> software continue to evolve; each generation offers more computing<br />

power <strong>and</strong> storage with a lower price tag.<br />

Consequences<br />

IBM’s entry into the microcomputer market gave microcomputers<br />

credibility. Apple Computer’s earlier introduction of its computer<br />

did not win wide acceptance with the corporate world. Apple<br />

did, however, thrive within the educational marketplace. IBM’s<br />

name already carried with it much clout, because IBM was a successful<br />

company. Apple Computer represented all that was great<br />

about the “new” microcomputer, but the IBM PC benefited from<br />

IBM’s image of stability <strong>and</strong> success.<br />

IBM coined the term personal computer <strong>and</strong> its acronym PC. The<br />

acronym PC is now used almost universally to refer to the microcomputer.<br />

It also had great significance with users who had previously<br />

used a large mainframe computer that had to be shared with<br />

the whole company. This was their personal computer. That was important<br />

to many PC buyers, since the company mainframe was perceived<br />

as being complicated <strong>and</strong> slow. The PC owner now had complete<br />

control.


See also Apple II computer; BINAC computer; Colossus computer;<br />

ENIAC computer; Floppy disk; Hard disk; IBM Model 1401<br />

computer; Internet; Supercomputer; UNIVAC computer.<br />

Further Reading<br />

Personal computer / 561<br />

Cerruzi, Paul E. A History of Modern Computing. Cambridge, Mass.:<br />

MIT Press, 2000.<br />

Chposky, James, <strong>and</strong> Ted Leonsis. Blue Magic: The People, Power, <strong>and</strong><br />

Politics Behind the IBM Personal Computer. New York: Facts on File,<br />

1988.<br />

Freiberger, Paul, <strong>and</strong> Michael Swaine. Fire in the Valley: The Making of<br />

the Personal Computer. New York: McGraw-Hill, 2000.<br />

Grossman. Wendy. Remembering the Future: Interviews from Personal<br />

Computer World. New York: Springer, 1997.


562<br />

Photoelectric cell<br />

Photoelectric cell<br />

The invention: The first devices to make practical use of the photoelectric<br />

effect, photoelectric cells were of decisive importance in<br />

the electron theory of metals.<br />

The people behind the invention:<br />

Julius Elster (1854-1920), a German experimental physicist<br />

Hans Friedrich Geitel (1855-1923), a German physicist<br />

Wilhelm Hallwachs (1859-1922), a German physicist<br />

Early Photoelectric Cells<br />

The photoelectric effect was known to science in the early<br />

nineteenth century when the French physicist Alex<strong>and</strong>re-Edmond<br />

Becquerel wrote of it in connection with his work on glass-enclosed<br />

primary batteries. He discovered that the voltage of his batteries increased<br />

with intensified illumination <strong>and</strong> that green light produced<br />

the highest voltage. Since Becquerel researched batteries exclusively,<br />

however, the liquid-type photocell was not discovered until<br />

1929, when the Wein <strong>and</strong> Arcturus cells were introduced commercially.<br />

These cells were miniature voltaic cells arranged so that light<br />

falling on one side of the front plate generated a considerable<br />

amount of electrical energy. The cells had short lives, unfortunately;<br />

when subjected to cold, the electrolyte froze, <strong>and</strong> when subjected to<br />

heat, the gas generated would exp<strong>and</strong> <strong>and</strong> explode the cells.<br />

What came to be known as the photoelectric cell, a device connecting<br />

light <strong>and</strong> electricity, had its beginnings in the 1880’s. At<br />

that time, scientists noticed that a negatively charged metal plate<br />

lost its charge much more quickly in the light (especially ultraviolet<br />

light) than in the dark. Several years later, researchers demonstrated<br />

that this phenomenon was not an “ionization” effect because<br />

of the air’s increased conductivity, since the phenomenon<br />

took place in a vacuum but did not take place if the plate were positively<br />

charged. Instead, the phenomenon had to be attributed to<br />

the light that excited the electrons of the metal <strong>and</strong> caused them to<br />

fly off: A neutral plate even acquired a slight positive charge under


the influence of strong light. Study of this effect not only contributed<br />

evidence to an electronic theory of matter—<strong>and</strong>, as a result of<br />

some brilliant mathematical work by the physicist Albert Einstein,<br />

later increased knowledge of the nature of radiant energy—but<br />

also further linked the studies of light <strong>and</strong> electricity. It even explained<br />

certain chemical phenomena, such as the process of photography.<br />

It is important to note that all the experimental work on<br />

photoelectricity accomplished prior to the work of Julius Elster<br />

<strong>and</strong> Hans Friedrich Geitel was carried out before the existence of<br />

the electron was known.<br />

Explaining Photoelectric Emission<br />

Photoelectric cell / 563<br />

After the English physicist Sir Joseph John Thomson’s discovery<br />

of the electron in 1897, investigators soon realized that the photoelectric<br />

effect was caused by the emission of electrons under the influence<br />

of radiation. The fundamental theory of photoelectric emission<br />

was put forward by Einstein in 1905 on the basis of the German<br />

physicist Max Planck’s quantum theory (1900). Thus, it was not surprising<br />

that light was found to have an electronic effect. Since it was<br />

known that the longer radio waves could shake electrons into resonant<br />

oscillations <strong>and</strong> the shorter X rays could detach electrons from<br />

the atoms of gases, the intermediate waves of visual light would<br />

have been expected to have some effect upon electrons—such as detaching<br />

them from metal plates <strong>and</strong> therefore setting up a difference<br />

of potential. The photoelectric cell, developed by Elster <strong>and</strong> Geitel<br />

in 1904, was a practical device that made use of this effect.<br />

In 1888, Wilhelm Hallwachs observed that an electrically charged<br />

zinc electrode loses its charge when exposed to ultraviolet radiation<br />

if the charge is negative, but is able to retain a positive charge under<br />

the same conditions. The following year, Elster <strong>and</strong> Geitel discovered<br />

a photoelectric effect caused by visible light; however, they<br />

used the alkali metals potassium <strong>and</strong> sodium for their experiments<br />

instead of zinc.<br />

The Elster-Geitel photocell (a vacuum emission cell, as opposed to<br />

a gas-filled cell) consisted of an evacuated glass bulb containing two<br />

electrodes. The cathode consisted of a thin film of a rare, chemically<br />

active metal (such as potassium) that lost its electrons fairly readily;


564 / Photoelectric cell<br />

Julius Elster <strong>and</strong> Hans Geitel<br />

Nicknamed the Castor <strong>and</strong> Pollux of physics after the twins<br />

of Greek mythology, Johann Philipp Ludwig Julius Elster <strong>and</strong><br />

Hans Friedrich Geitel were among the most productive teams<br />

in the history of science. Elster, born in 1854, <strong>and</strong> Geitel, born in<br />

1855, met in 1875 while attending university in Heidelberg,<br />

Germany. Graduate studies took them to separate cities, but<br />

then in 1881 they were together again as mathematics <strong>and</strong> physics<br />

teachers at Herzoglich Gymnasium in Wolfenbüttel. In 1884<br />

they began their scientific collaboration, which lasted more<br />

than thirty years <strong>and</strong> produced more than 150 reports.<br />

Essentially experimentalists, they investigated phenomena<br />

that were among the greatest mysteries of the times. Their first<br />

works concerned the electrification of flames <strong>and</strong> the electrical<br />

properties of thunderstorms. They went on to study the photoelectric<br />

effect, thermal electron emission, practical uses for photocells,<br />

<strong>and</strong> Becquerel rays in the earth <strong>and</strong> air. They developed a<br />

method for measuring electrical phenomena in gases that remained<br />

the st<strong>and</strong>ard for the following forty years.<br />

Their greatest achievements, however, lay with radioactivity<br />

<strong>and</strong> radiation. Their demonstration that inc<strong>and</strong>escent filaments<br />

emitted “negative electricity” proved beyond doubt that<br />

electrons, which J. J. Thomson had recently claimed to have detected,<br />

did in fact exist. They also proved that radioactivity,<br />

such as that from uranium, came wholly from within the atom,<br />

not from environmental influences. Ernest Rutherford, the great<br />

English physicist, said in 1913 that Elster <strong>and</strong> Geitel had contributed<br />

more to the underst<strong>and</strong>ing of terrestrial <strong>and</strong> atmospheric<br />

radioactivity than anyone else.<br />

The pair were practically inseparable until Elster died in<br />

1920. Geitel died three years later.<br />

the anode was simply a wire sealed in to complete the circuit. This anode<br />

was maintained at a positive potential in order to collect the negative<br />

charges released by light from the cathode. The Elster-Geitel<br />

photocell resembled two other types of vacuum tubes in existence at<br />

the time: the cathode-ray tube, in which the cathode emitted electrons<br />

under the influence of a high potential, <strong>and</strong> the thermionic<br />

valve (a valve that permits the passage of current in one direction


only), in which it emitted electrons under the influence of heat. Like<br />

both of these vacuum tubes, the photoelectric cell could be classified<br />

as an “electronic” device.<br />

The new cell, then, emitted electrons when stimulated by light, <strong>and</strong><br />

at a rate proportional to the intensity of the light. Hence, a current<br />

could be obtained from the cell. Yet Elster <strong>and</strong> Geitel found that their<br />

photoelectric currents fell off gradually; they therefore spoke of “fatigue”<br />

(instability). It was discovered later that most of this change was<br />

not a direct effect of a photoelectric current’s passage; it was not even<br />

an indirect effect but was caused by oxidation of the cathode by the air.<br />

Since all modern cathodes are enclosed in sealed vessels, that source of<br />

change has been completely abolished. Nevertheless, the changes that<br />

persist in modern cathodes often are indirect effects of light that can be<br />

produced independently of any photoelectric current.<br />

Impact<br />

Photoelectric cell / 565<br />

The Elster-Geitel photocell was, for some twenty years, used in<br />

all emission cells adapted for the visible spectrum, <strong>and</strong> throughout<br />

the twentieth century, the photoelectric cell has had a wide variety<br />

of applications in numerous fields. For example, if products leaving<br />

a factory on a conveyor belt were passed between a light <strong>and</strong> a cell,<br />

they could be counted as they interrupted the beam. Persons entering<br />

a building could be counted also, <strong>and</strong> if invisible ultraviolet rays<br />

were used, those persons could be detected without their knowledge.<br />

Simple relay circuits could be arranged that would automatically<br />

switch on street lamps when it grew dark. The sensitivity of<br />

the cell with an amplifying circuit enabled it to “see” objects too<br />

faint for the human eye, such as minor stars or certain lines in the<br />

spectra of elements excited by a flame or discharge. The fact that the<br />

current depended on the intensity of the light made it possible to<br />

construct photoelectric meters that could judge the strength of illumination<br />

without risking human error—for example, to determine<br />

the right exposure for a photograph.<br />

A further use for the cell was to make talking films possible. The<br />

early “talkies” had depended on gramophone records, but it was very<br />

difficult to keep the records in time with the film. Now, the waves of<br />

speech <strong>and</strong> music could be recorded in a “sound track” by turning the


566 / Photoelectric cell<br />

sound first into current through a microphone <strong>and</strong> then into light with<br />

a neon tube or magnetic shutter; next, the variations in the intensity of<br />

this light on the side of the film were photographed. By reversing the<br />

process <strong>and</strong> running the film between a light <strong>and</strong> a photoelectric cell,<br />

the visual signals could be converted back to sound.<br />

See also Alkaline storage battery; Photovoltaic cell; Solar thermal<br />

engine.<br />

Further Reading<br />

Hoberman, Stuart. Solar Cell <strong>and</strong> Photocell Experimenters Guide. Indianapolis,<br />

Ind.: H. W. Sams, 1965.<br />

Perlin, John. From Space to Earth: The Story of Solar Electricity. Ann Arbor,<br />

Mich.: Aatec <strong>Public</strong>ations, 1999.<br />

Walker, R. C., <strong>and</strong> T. M. C. Lance. Photoelectric Cell Applications: A<br />

Practical Book Describing the Uses of Photoelectric Cells in Television,<br />

Talking Pictures, Electrical Alarms, Counting Devices, Etc. 3ded.<br />

London: Sir I. Pitman & Sons, 1938.


Photovoltaic cell<br />

Photovoltaic cell<br />

The invention: Drawing their energy directly from the Sun, the<br />

first photovoltaic cells powered instruments on early space vehicles<br />

<strong>and</strong> held out hope for future uses of solar energy.<br />

The people behind the invention:<br />

Daryl M. Chapin (1906-1995), an American physicist<br />

Calvin S. Fuller (1902-1994), an American chemist<br />

Gerald L. Pearson (1905- ), an American physicist<br />

Unlimited Energy Source<br />

All the energy that the world has at its disposal ultimately comes<br />

from the Sun. Some of this solar energy was trapped millions of years<br />

ago in the form of vegetable <strong>and</strong> animal matter that became the coal,<br />

oil, <strong>and</strong> natural gas that the world relies upon for energy. Some of this<br />

fuel is used directly to heat homes <strong>and</strong> to power factories <strong>and</strong> gasoline<br />

vehicles. Much of this fossil fuel, however, is burned to produce<br />

the electricity on which modern society depends.<br />

The amount of energy available from the Sun is difficult to imagine,<br />

but some comparisons may be helpful. During each forty-hour<br />

period, the Sun provides the earth with as much energy as the<br />

earth’s total reserves of coal, oil, <strong>and</strong> natural gas. It has been estimated<br />

that the amount of energy provided by the sun’s radiation<br />

matches the earth’s reserves of nuclear fuel every forty days. The<br />

annual solar radiation that falls on about twelve hundred square<br />

miles of l<strong>and</strong> in Arizona matched the world’s estimated total annual<br />

energy requirement for 1960. Scientists have been searching for<br />

many decades for inexpensive, efficient means of converting this<br />

vast supply of solar radiation directly into electricity.<br />

The Bell Solar Cell<br />

567<br />

Throughout its history, Bell Systems has needed to be able to<br />

transmit, modulate, <strong>and</strong> amplify electrical signals. Until the 1930’s,<br />

these tasks were accomplished by using insulators <strong>and</strong> metallic con-


568 / Photovoltaic cell<br />

ductors. At that time, semiconductors, which have electrical properties<br />

that are between those of insulators <strong>and</strong> those of conductors,<br />

were developed. One of the most important semiconductor materials<br />

is silicon, which is one of the most common elements on the<br />

earth. Unfortunately, silicon is usually found in the form of compounds<br />

such as s<strong>and</strong> or quartz, <strong>and</strong> it must be refined <strong>and</strong> purified<br />

before it can be used in electrical circuits. This process required<br />

much initial research, <strong>and</strong> very pure silicon was not available until<br />

the early 1950’s.<br />

Electric conduction in silicon is the result of the movement of<br />

negative charges (electrons) or positive charges (holes). One way of<br />

accomplishing this is by deliberately adding to the silicon phosphorus<br />

or arsenic atoms, which have five outer electrons. This addition<br />

creates a type of semiconductor that has excess negative charges (an<br />

n-type semiconductor). Adding boron atoms, which have three<br />

outer electrons, creates a semiconductor that has excess positive<br />

charges (a p-type semiconductor). Calvin Fuller made an important<br />

study of the formation of p-n junctions, which are the points at<br />

which p-type <strong>and</strong> n-type semiconductors meet, by using the process<br />

of diffusing impurity atoms—that is, adding atoms of materials that<br />

would increase the level of positive or negative charges, as described<br />

above. Fuller’s work stimulated interested in using the process<br />

of impurity diffusion to create cells that would turn solar energy<br />

into electricity. Fuller <strong>and</strong> Gerald Pearson made the first largearea<br />

p-n junction by using the diffusion process. Daryl Chapin,<br />

Fuller, <strong>and</strong> Pearson made a similar p-n junction very close to the<br />

surface of a silicon crystal, which was then exposed to sunlight.<br />

The cell was constructed by first making an ingot of arsenicdoped<br />

silicon that was then cut into very thin slices. Then a very<br />

thin layer of p-type silicon was formed over the surface of the n-type<br />

wafer, providing a p-n junction close to the surface of the cell. Once<br />

the cell cooled, the p-type layer was removed from the back of the<br />

cell <strong>and</strong> lead wires were attached to the two surfaces. When light<br />

was absorbed at the p-n junction, electron-hole pairs were produced,<br />

<strong>and</strong> the electric field that was present at the junction forced<br />

the electrons to the n side <strong>and</strong> the holes to the p side.<br />

The recombination of the electrons <strong>and</strong> holes takes place after the<br />

electrons have traveled through the external wires, where they do


useful work. Chapin, Fuller, <strong>and</strong> Pearson announced in 1954 that<br />

the resulting photovoltaic cell was the most efficient (6 percent)<br />

means then available for converting sunlight into electricity.<br />

The first experimental use of the silicon solar battery was in amplifiers<br />

for electrical telephone signals in rural areas. An array of 432<br />

silicon cells, capable of supplying 9 watts of power in bright sunlight,<br />

was used to charge a nickel-cadmium storage battery. This, in<br />

turn, powered the amplifier for the telephone signal. The electrical<br />

energy derived from sunlight during the day was sufficient to keep<br />

the storage battery charged for continuous operation. The system<br />

was successfully tested for six months of continuous use in Americus,<br />

Georgia, in 1956. Although it was a technical success, the silicon solar<br />

cell was not ready to compete economically with conventional<br />

means of producing electrical power.<br />

Consequences<br />

Parabolic mirrors at a solar power plant. (PhotoDisc)<br />

Photovoltaic cell / 569<br />

One of the immediate applications of the solar cell was to supply<br />

electrical energy for Telstar satellites. These cells are used extensively<br />

on all satellites to generate power. The success of the U.S. sat-


570 / Photovoltaic cell<br />

ellite program prompted serious suggestions in 1965 for the use of<br />

an orbiting power satellite. A large satellite could be placed into a<br />

synchronous orbit of the earth. It would collect sunlight, convert it<br />

to microwave radiation, <strong>and</strong> beam the energy to an Earth-based receiving<br />

station. Many technical problems must be solved, however,<br />

before this dream can become a reality.<br />

Solar cells are used in small-scale applications such as power<br />

sources for calculators. Large-scale applications are still not economically<br />

competitive with more traditional means of generating<br />

electric power. The development of the Third World countries, however,<br />

may provide the incentive to search for less-expensive solar<br />

cells that can be used, for example, to provide energy in remote villages.<br />

As the st<strong>and</strong>ards of living in such areas improve, the need for<br />

electric power will grow. Solar cells may be able to provide the necessary<br />

energy while safeguarding the environment for future generations.<br />

See also Alkaline storage battery; Fluorescent lighting; Fuel cell;<br />

Photoelectric cell; Solar thermal engine.<br />

Further Reading<br />

Green, Martin A. Power to the People: Sunlight to Electricity Using Solar<br />

Cells. Sydney, Australia: University of South Wales Press, 2000.<br />

_____. “Photovoltaics: Technology Overview.” Energy Policy 28, no.<br />

14 (November, 2000).<br />

Perlin, John. From Space to Earth: The Story of Solar Electricity. Ann Arbor,<br />

Mich.: Aatec <strong>Public</strong>ations, 1999.


Plastic<br />

Plastic<br />

The invention: The first totally synthetic thermosetting plastic,<br />

which paved the way for modern materials science.<br />

The people behind the invention:<br />

John Wesley Hyatt (1837-1920), an American inventor<br />

Leo Hendrik Baekel<strong>and</strong> (1863-1944), a Belgian-born chemist,<br />

consultant, <strong>and</strong> inventor<br />

Christian Friedrich Schönbein (1799-1868), a German chemist<br />

who produced guncotton, the first artificial polymer<br />

Adolf von Baeyer (1835-1917), a German chemist<br />

Exploding Billiard Balls<br />

571<br />

In the 1860’s, the firm of Phelan <strong>and</strong> Collender offered a prize of<br />

ten thous<strong>and</strong> dollars to anyone producing a substance that could<br />

serve as an inexpensive substitute for ivory, which was somewhat<br />

difficult to obtain in large quantities at reasonable prices. Earlier,<br />

Christian Friedrich Schönbein had laid the groundwork for a breakthrough<br />

in the quest for a new material in 1846 by the serendipitous<br />

discovery of nitrocellulose, more commonly known as “guncotton,”<br />

which was produced by the reaction of nitric acid with cotton.<br />

An American inventor, John Wesley Hyatt, while looking for a<br />

substitute for ivory as a material for making billiard balls, discovered<br />

that the addition of camphor to nitrocellulose under certain<br />

conditions led to the formation of a white material that could be<br />

molded <strong>and</strong> machined. He dubbed this substance “celluloid,” <strong>and</strong><br />

this product is now acknowledged as the first synthetic plastic. Celluloid<br />

won the prize for Hyatt, <strong>and</strong> he promptly set out to exploit his<br />

product. Celluloid was used to make baby rattles, collars, dentures,<br />

<strong>and</strong> other manufactured goods.<br />

As a billiard ball substitute, however, it was not really adequate,<br />

for various reasons. First, it is thermoplastic—in other words, a material<br />

that softens when heated <strong>and</strong> can then be easily deformed or<br />

molded. It was thus too soft for billiard ball use. Second, it was<br />

highly flammable, hardly a desirable characteristic. A widely circu-


572 / Plastic<br />

lated, perhaps apocryphal, story claimed that celluloid billiard balls<br />

detonated when they collided.<br />

Truly Artificial<br />

Since celluloid can be viewed as a derivative of a natural product,<br />

it is not a completely synthetic substance. Leo Hendrik Baekel<strong>and</strong><br />

has the distinction of being the first to produce a completely artificial<br />

plastic. Born in Ghent, Belgium, Baekel<strong>and</strong> emigrated to the<br />

United States in 1889 to pursue applied research, a pursuit not encouraged<br />

in Europe at the time. One area in which Baekel<strong>and</strong> hoped<br />

to make an inroad was in the development of an artificial shellac.<br />

Shellac at the time was a natural <strong>and</strong> therefore expensive product,<br />

<strong>and</strong> there would be a wide market for any reasonably priced substitute.<br />

Baekel<strong>and</strong>’s research scheme, begun in 1905, focused on finding<br />

a solvent that could dissolve the resinous products from a certain<br />

class of organic chemical reaction.<br />

The particular resins he used had been reported in the mid-<br />

1800’s by the German chemist Adolf von Baeyer. These resins were<br />

produced by the condensation reaction of formaldehyde with a<br />

class of chemicals called “phenols.” Baeyer found that frequently<br />

the major product of such a reaction was a gummy residue that was<br />

virtually impossible to remove from glassware. Baekel<strong>and</strong> focused<br />

on finding a material that could dissolve these resinous products.<br />

Such a substance would prove to be the shellac substitute he sought.<br />

These efforts proved frustrating, as an adequate solvent for these<br />

resins could not be found. After repeated attempts to dissolve these<br />

residues, Baekel<strong>and</strong> shifted the orientation of his work. Ab<strong>and</strong>oning<br />

the quest to dissolve the resin, he set about trying to develop a resin<br />

that would be impervious to any solvent, reasoning that such a material<br />

would have useful applications.<br />

Baekel<strong>and</strong>’s experiments involved the manipulation of phenolformaldehyde<br />

reactions through precise control of the temperature<br />

<strong>and</strong> pressure at which the reactions were performed. Many of these<br />

experiments were performed in a 1.5-meter-tall reactor vessel, which<br />

he called a “Bakelizer.” In 1907, these meticulous experiments paid<br />

off when Baekel<strong>and</strong> opened the reactor to reveal a clear solid that<br />

was heat resistant, nonconducting, <strong>and</strong> machinable. Experimenta-


Plastic / 573<br />

tion proved that the material could be dyed practically any color in<br />

the manufacturing process, with no effect on the physical properties<br />

of the solid.<br />

Baekel<strong>and</strong> filed a patent for this new material in 1907. (This patent<br />

was filed one day before that filed by James Swinburne, a British<br />

John Wesley Hyatt<br />

John Wesley Hyatt’s parents wanted him to be a minister, a<br />

step up in status from his father’s job as a blacksmith. Born in<br />

1837 in Starkey, New York, Hyatt received the st<strong>and</strong>ard primary<br />

education <strong>and</strong> then obediently went to a seminary as a teenager.<br />

However, his mind was on making things rather than spirituality;<br />

he was especially ingenious with machinery. The seminary<br />

held him only a year. He became a printer’s apprentice at<br />

sixteen <strong>and</strong> later set up shop in Albany.<br />

His mind ranged beyond printing too. He invented a method<br />

to make emery wheels for sharpening cutlery, which brought<br />

him his first patent at twenty-four. In an attempt to win the<br />

Phelan <strong>and</strong> Collender Company contest for artificial billiard<br />

balls, he developed several moldable compounds from wood<br />

pulp. He started the Embossing Company in Albany to make<br />

chess <strong>and</strong> checker pieces from the compounds <strong>and</strong> put his<br />

youngest brother in charge. With another brother he experimented<br />

with guncotton until he invented celluloid. In 1872, he<br />

<strong>and</strong> his brothers started the Celluloid Manufacturing Company.<br />

They designed new milling machinery for the new substance<br />

<strong>and</strong> turned out billiard balls, bowling balls, golf club<br />

heads <strong>and</strong> other sporting goods but then branched out into domestic<br />

items, such as boxes, h<strong>and</strong>les, combs, <strong>and</strong> even collars.<br />

Celluloid became the basic material of photographic film <strong>and</strong>,<br />

later, motion picture film.<br />

Meanwhile, Hyatt continued to invent—machinery for cutting<br />

<strong>and</strong> molding plastic <strong>and</strong> rolling steel, a water purification<br />

system, a method for squeezing juice from sugar cane, an industrial<br />

sewing machine, roller bearings for heavy machinery—registering<br />

more than 250 patents, which is impressive for<br />

a person with no formal scientific or technical training. The Society<br />

of Chemical Industry awarded Hyatt its prestigious Perkin<br />

Medal in 1914. Hyatt died in 1920.


574 / Plastic<br />

electrical engineer who had developed a similar material in his<br />

quest to produce an insulating material.) Baekel<strong>and</strong> dubbed his new<br />

creation “Bakelite” <strong>and</strong> announced its existence to the scientific<br />

community on February 15, 1909, at the annual meeting of the American<br />

Chemical Society. Among its first uses was in the manufacture<br />

of ignition parts for the rapidly growing automobile industry.<br />

Impact<br />

Bakelite proved to be the first of a class of compounds called<br />

“synthetic polymers.” Polymers are long chains of molecules chemically<br />

linked together. There are many natural polymers, such as cotton.<br />

The discovery of synthetic polymers led to vigorous research<br />

into the field <strong>and</strong> attempts to produce other useful artificial materials.<br />

These efforts met with a fair amount of success; by 1940, a multitude<br />

of new products unlike anything found in nature had been discovered.<br />

These included such items as polystyrene <strong>and</strong> low-density<br />

polyethylene. In addition, artificial substitutes for natural polymers,<br />

such as rubber, were a goal of polymer chemists. One of the results<br />

of this research was the development of neoprene.<br />

Industries also were interested in developing synthetic polymers<br />

to produce materials that could be used in place of natural fibers<br />

such as cotton. The most dramatic success in this area was achieved<br />

by Du Pont chemist Wallace Carothers, who had also developed<br />

neoprene. Carothers focused his energies on forming a synthetic fiber<br />

similar to silk, resulting in the synthesis of nylon.<br />

Synthetic polymers constitute one branch of a broad area known<br />

as “materials science.” Novel, useful materials produced synthetically<br />

from a variety of natural materials have allowed for tremendous<br />

progress in many areas. Examples of these new materials include<br />

high-temperature superconductors, composites, ceramics, <strong>and</strong><br />

plastics. These materials are used to make the structural components<br />

of aircraft, artificial limbs <strong>and</strong> implants, tennis rackets, garbage<br />

bags, <strong>and</strong> many other common objects.<br />

See also Buna rubber; Contact lenses; Laminated glass; Neoprene;<br />

Nylon; Orlon; Polyester; Polyethylene; Polystyrene; Pyrex<br />

glass; Silicones; Teflon; Velcro.


Further Reading<br />

Plastic / 575<br />

Amato, Ivan. “Chemist: Leo Baekel<strong>and</strong>.” Time 153, no. 12 (March 29,<br />

1999).<br />

Clark, Tessa. Bakelite Style. Edison, N.J.: Chartwell Books, 1997.<br />

Fenichell, Stephen. Plastic: The Making of a Synthetic Century. New<br />

York: HarperBusiness, 1997.<br />

Sparke, Penny. The Plastics Age: From Bakelite to Beanbags <strong>and</strong> Beyond.<br />

Woodstock, N.Y.: Overlook Press, 1990.


576<br />

Pocket calculator<br />

Pocket calculator<br />

The invention: The first portable <strong>and</strong> reliable h<strong>and</strong>-held calculator<br />

capable of performing a wide range of mathematical computations.<br />

The people behind the invention:<br />

Jack St. Clair Kilby (1923- ), the inventor of the<br />

semiconductor microchip<br />

Jerry D. Merryman (1932- ), the first project manager of the<br />

team that invented the first portable calculator<br />

James Van Tassel (1929- ), an inventor <strong>and</strong> expert on<br />

semiconductor components<br />

An Ancient Dream<br />

In the earliest accounts of civilizations that developed number<br />

systems to perform mathematical calculations, evidence has been<br />

found of efforts to fashion a device that would permit people to perform<br />

these calculations with reduced effort <strong>and</strong> increased accuracy.<br />

The ancient Babylonians are regarded as the inventors of the first<br />

abacus (or counting board, from the Greek abakos, meaning “board”<br />

or “tablet”). It was originally little more than a row of shallow<br />

grooves with pebbles or bone fragments as counters.<br />

The next step in mechanical calculation did not occur until the<br />

early seventeenth century. John Napier, a Scottish baron <strong>and</strong> mathematician,<br />

originated the concept of “logarithms” as a mathematical<br />

device to make calculating easier. This concept led to the first slide<br />

rule, created by the English mathematician William Oughtred of<br />

Cambridge. Oughtred’s invention consisted of two identical, circular<br />

logarithmic scales held together <strong>and</strong> adjusted by h<strong>and</strong>. The slide<br />

rule made it possible to perform rough but rapid multiplication <strong>and</strong><br />

division. Oughtred’s invention in 1623 was paralleled by the work<br />

of a German professor, Wilhelm Schickard, who built a “calculating<br />

clock” the same year. Because the record of Schickard’s work was<br />

lost until 1935, however, the French mathematician Blaise Pascal<br />

was generally thought to have built the first mechanical calculator,<br />

the “Pascaline,” in 1645.


Other versions of mechanical calculators were built in later centuries,<br />

but none was rapid or compact enough to be useful beyond specific<br />

laboratory or mercantile situations. Meanwhile, the dream of<br />

such a machine continued to fascinate scientists <strong>and</strong> mathematicians.<br />

The development that made a fast, small calculator possible did<br />

not occur until the middle of the twentieth century, when Jack St.<br />

Clair Kilby of Texas Instruments invented the silicon microchip (or<br />

integrated circuit) in 1958. An integrated circuit is a tiny complex of<br />

electronic components <strong>and</strong> their connections that is produced in or<br />

on a small slice of semiconductor material such as silicon. Patrick<br />

Haggerty, then president of Texas Instruments, wrote in 1964 that<br />

“integrated electronics” would “remove limitations” that determined<br />

the size of instruments, <strong>and</strong> he recognized that Kilby’s invention<br />

of the microchip made possible the creation of a portable,<br />

h<strong>and</strong>-held calculator. He challenged Kilby to put together a team to<br />

design a calculator that would be as powerful as the large, electromechanical<br />

models in use at the time but small enough to fit into a<br />

coat pocket. Working with Jerry D. Merryman <strong>and</strong> James Van Tassel,<br />

Kilby began to work on the project in October, 1965.<br />

An Amazing Reality<br />

Pocket calculator / 577<br />

At the outset, there were basically five elements that had to be designed.<br />

These were the logic designs that enabled the machine to<br />

perform the actual calculations, the keyboard or keypad, the power<br />

supply, the readout display, <strong>and</strong> the outer case. Kilby recalls that<br />

once a particular size for the unit had been determined (something<br />

that could be easily held in the h<strong>and</strong>), project manager Merryman<br />

was able to develop the initial logic designs in three days. Van Tassel<br />

contributed his experience with semiconductor components to solve<br />

the problems of packaging the integrated circuit. The display required<br />

a thermal printer that would work on a low power source.<br />

The machine also had to include a microencapsulated ink source so<br />

that the paper readouts could be imprinted clearly. Then the paper<br />

had to be advanced for the next calculation. Kilby, Merryman, <strong>and</strong><br />

Van Tassel filed for a patent on their work in 1967.<br />

Although this relatively small, working prototype of the minicalculator<br />

made obsolete the transistor-operated design of the much


578 / Pocket calculator<br />

Jerry D. Merryman<br />

In 1965 Texas Instruments assigned two engineers to join<br />

Jack St. Clair Kilby, inventor of the integrated circuit, in an effort<br />

to produce a pocket-sized calculator: James H. Van Tassel, a<br />

specialist in semiconductor components, <strong>and</strong> Jerry D. Merryman,<br />

a versatile engineer who became the project manager. It<br />

took Merryman only seventy-two hours to work out the logic<br />

design for the calculator, <strong>and</strong> the team set about designing, fabricating,<br />

<strong>and</strong> testing its components. After two years, it had a<br />

prototype, the first pocket calculator. However, it required a<br />

large, strong pocket. It measured 4.25 inches by 6.12 inches by<br />

1.76 inches <strong>and</strong> weighed 2.8 pounds. Kilby, Van Tassel, <strong>and</strong><br />

Merry filed for a patent <strong>and</strong> received it in 1975. In 1989 the team<br />

was jointly presented the Holley Medical for the achievement<br />

by the American Society of Mechanical Engineers. By then<br />

Merryman held sixty other patents, foreign <strong>and</strong> domestic.<br />

Born in 1932, Merryman grew up in Hearne, Texas, <strong>and</strong> after<br />

high school went to Texas A&M University. He never graduated,<br />

but he did become extraordinarily adept at electrical engineering,<br />

teaching himself what he needed to know while doing<br />

small jobs on his own. He was said to have almost an intuitive<br />

sense for circuitry. After he joined Texas Instruments in 1963<br />

he quickly earned a reputation for solving complex problems,<br />

one of the reasons he was made part of the h<strong>and</strong> calculator<br />

team. He became a Texas Instruments Fellow in 1975 <strong>and</strong> helped<br />

design semiconductor manufacturing equipment, particularly<br />

by adapting high-speed lasers for use in extremely fine optical<br />

lithography. He also invented thermal data systems.<br />

Along with Kilby <strong>and</strong> Van Tassel, Merryman received the<br />

George R. Stibitz Computer Pioneer Award in 1997.<br />

larger desk calculators, the cost of setting up new production lines<br />

<strong>and</strong> the need to develop a market made it impractical to begin production<br />

immediately. Instead, Texas Instruments <strong>and</strong> Canon of Tokyo<br />

formed a joint venture, which led to the introduction of the<br />

Canon Pocketronic Printing Calculator in Japan in April, 1970, <strong>and</strong><br />

in the United States that fall. Built entirely of Texas Instruments<br />

parts, this four-function machine with three metal oxide semicon-


Pocket calculator / 579<br />

True pocket calculators fit as easily in shirt pockets as pencils <strong>and</strong> pens. (PhotoDisc)<br />

ductor (MOS) circuits was similar to the prototype designed in 1967.<br />

The calculator was priced at $400, weighed 740 grams, <strong>and</strong> measured<br />

101 millimeters wide by 208 millimeters long by 49 millimeters<br />

high. It could perform twelve-digit calculations <strong>and</strong> worked up<br />

to four decimal places.<br />

In September, 1972, Texas Instruments put the Datamath, its first<br />

commercial h<strong>and</strong>-held calculator using a single MOS chip, on the<br />

retail market. It weighed 340 grams <strong>and</strong> measured 75 millimeters<br />

wide by 137 millimeters long by 42 millimeters high. The Datamath<br />

was priced at $120 <strong>and</strong> included a full-floating decimal point that<br />

could appear anywhere among the numbers on its eight-digit, lightemitting<br />

diode (LED) display. It came with a rechargeable battery<br />

that could also be connected to a st<strong>and</strong>ard alternating current (AC)<br />

outlet. The Datamath also had the ability to conserve power while<br />

awaiting the next keyboard entry. Finally, the machine had a built-in<br />

limited amount of memory storage.


580 / Pocket calculator<br />

Consequences<br />

Prior to 1970, most calculating machines were of such dimensions<br />

that professional mathematicians <strong>and</strong> engineers were either tied to<br />

their desks or else carried slide rules whenever they had to be away<br />

from their offices. By 1975, Keuffel & Esser, the largest slide rule manufacturer<br />

in the world, was producing its last model, <strong>and</strong> mechanical<br />

engineers found that problems that had previously taken a week<br />

could now be solved in an hour using the new machines.<br />

That year, the Smithsonian Institution accepted the world’s first<br />

miniature electronic calculator for its permanent collection, noting<br />

that it was the forerunner of more than one hundred million pocket<br />

calculators then in use. By the 1990’s, more than fifty million portable<br />

units were being sold each year in the United States. In general,<br />

the electronic pocket calculator revolutionized the way in which<br />

people related to the world of numbers.<br />

Moreover, the portability of the h<strong>and</strong>-held calculator made it<br />

ideal for use in remote locations, such as those a petroleum engineer<br />

might have to explore. Its rapidity <strong>and</strong> reliability made it an indispensable<br />

instrument for construction engineers, architects, <strong>and</strong> real<br />

estate agents, who could figure the volume of a room <strong>and</strong> other<br />

building dimensions almost instantly <strong>and</strong> then produce cost estimates<br />

almost on the spot.<br />

See also Cell phone; Differential analyzer; Mark I calculator; Personal<br />

computer; Transistor radio; Walkman cassette player.<br />

Further Reading<br />

Ball, Guy. Collector’s Guide to Pocket Calculators. Tustin, Calif.: Wilson/Barnett<br />

Publishing, 1996.<br />

Clayton, Mark. “Calculators in Class: Freedom from Scratch Paper<br />

or ‘Crutch’?” Christian Science Monitor (May 23, 2000).<br />

Lederer, Victor. “Calculators: The Applications Are Unlimited. Administrative<br />

Management 38 (July, 1977).<br />

Lee, Jennifer. “Throw Teachers a New Curve.” New York Times (September<br />

2, 1999).<br />

“The Semiconductor Becomes a New Marketing Force.” Business<br />

Week (August 24, 1974).


Polio vaccine (Sabin)<br />

Polio vaccine (Sabin)<br />

The invention: Albert Bruce Sabin’s vaccine was the first to stimulate<br />

long-lasting immunity against polio without the risk of causing<br />

paralytic disease.<br />

The people behind the invention:<br />

Albert Bruce Sabin (1906-1993), a Russian-born American<br />

virologist<br />

Jonas Edward Salk (1914-1995), an American physician,<br />

immunologist, <strong>and</strong> virologist<br />

Renato Dulbecco (1914- ), an Italian-born American<br />

virologist who shared the 1975 Nobel Prize in Physiology or<br />

Medicine<br />

The Search for a Living Vaccine<br />

581<br />

Almost a century ago, the first major poliomyelitis (polio) epidemic<br />

was recorded. Thereafter, epidemics of increasing frequency<br />

<strong>and</strong> severity struck the industrialized world. By the 1950’s, as many<br />

as sixteen thous<strong>and</strong> individuals, most of them children, were being<br />

paralyzed by the disease each year.<br />

Poliovirus enters the body through ingestion by the mouth. It<br />

replicates in the throat <strong>and</strong> the intestines <strong>and</strong> establishes an infection<br />

that normally is harmless. From there, the virus can enter the<br />

bloodstream. In some individuals it makes its way to the nervous<br />

system, where it attacks <strong>and</strong> destroys nerve cells crucial for muscle<br />

movement. The presence of antibodies in the bloodstream will prevent<br />

the virus from reaching the nervous system <strong>and</strong> causing paralysis.<br />

Thus, the goal of vaccination is to administer poliovirus that<br />

has been altered so that it cannot cause disease but nevertheless will<br />

stimulate the production of antibodies to fight the disease.<br />

Albert Bruce Sabin received his medical degree from New York<br />

University College of Medicine in 1931. Polio was epidemic in 1931,<br />

<strong>and</strong> for Sabin polio research became a lifelong interest. In 1936,<br />

while working at the Rockefeller Institute, Sabin <strong>and</strong> Peter Olinsky<br />

successfully grew poliovirus using tissues cultured in vitro. Tissue<br />

culture proved to be an excellent source of virus. Jonas Edward Salk


582 / Polio vaccine (Sabin)<br />

soon developed an inactive polio vaccine consisting of virus grown<br />

from tissue culture that had been inactivated (killed) by chemical<br />

treatment. This vaccine became available for general use in 1955, almost<br />

fifty years after poliovirus had first been identified.<br />

Sabin, however, was not convinced that an inactivated virus vaccine<br />

was adequate. He believed that it would provide only temporary<br />

protection <strong>and</strong> that individuals would have to be vaccinated<br />

repeatedly in order to maintain protective levels of antibodies.<br />

Knowing that natural infection with poliovirus induced lifelong immunity,<br />

Sabin believed that a vaccine consisting of a living virus<br />

was necessary to produce long-lasting immunity. Also, unlike the<br />

inactive vaccine, which is injected, a living virus (weakened so that<br />

it would not cause disease) could be taken orally <strong>and</strong> would invade<br />

the body <strong>and</strong> replicate of its own accord.<br />

Sabin was not alone in his beliefs. Hilary Koprowski <strong>and</strong> Harold<br />

Cox also favored a living virus vaccine <strong>and</strong> had, in fact, begun<br />

searching for weakened strains of poliovirus as early as 1946 by repeatedly<br />

growing the virus in rodents. When Sabin began his search<br />

for weakened virus strains in 1953, a fiercely competitive contest ensued<br />

to achieve an acceptable live virus vaccine.<br />

Rare, Mutant Polioviruses<br />

Sabin’s approach was based on the principle that, as viruses acquire<br />

the ability to replicate in a foreign species or tissue (for example,<br />

in mice), they become less able to replicate in humans <strong>and</strong> thus<br />

less able to cause disease. Sabin used tissue culture techniques to<br />

isolate those polioviruses that grew most rapidly in monkey kidney<br />

cells. He then employed a technique developed by Renato Dulbecco<br />

that allowed him to recover individual virus particles. The recovered<br />

viruses were injected directly into the brains or spinal cords of<br />

monkeys in order to identify those viruses that did not damage the<br />

nervous system. These meticulously performed experiments, which<br />

involved approximately nine thous<strong>and</strong> monkeys <strong>and</strong> more than<br />

one hundred chimpanzees, finally enabled Sabin to isolate rare mutant<br />

polioviruses that would replicate in the intestinal tract but not<br />

in the nervous systems of chimpanzees or, it was hoped, of humans.<br />

In addition, the weakened virus strains were shown to stimulate an-


Polio vaccine (Sabin) / 583<br />

tibodies when they were fed to chimpanzees; this was a critical attribute<br />

for a vaccine strain.<br />

By 1957, Sabin had identified three strains of attenuated viruses that<br />

were ready for small experimental trials in humans. A small group of<br />

volunteers, including Sabin’s own wife <strong>and</strong> children, were fed the vaccine<br />

with promising results. Sabin then gave his vaccine to virologists<br />

in the Soviet Union, Eastern Europe, Mexico, <strong>and</strong> Holl<strong>and</strong> for further<br />

testing. Combined with smaller studies in the United States, these trials<br />

established the effectiveness <strong>and</strong> safety of his oral vaccine.<br />

During this period, the strains developed by Cox <strong>and</strong> by Koprowski<br />

were being tested also in millions of persons in field trials<br />

around the world. In 1958, two laboratories independently compared<br />

the vaccine strains <strong>and</strong> concluded that the Sabin strains were<br />

superior. In 1962, after four years of deliberation by the U.S. <strong>Public</strong><br />

Health Service, all three of Sabin’s vaccine strains were licensed for<br />

general use.<br />

Albert Sabin<br />

Born in Bialystok, Pol<strong>and</strong>, in 1906, Albert Bruce Sabin emigrated<br />

with his family to the United States in 1921. Like Jonas<br />

Salk—the other great inventor of a polio vaccine—Sabin earned<br />

his medical degree at New York University (1931), where he began<br />

his research on polio.<br />

While in the U.S. Army Medical Corps during World War II,<br />

he helped produce vaccines for dengue fever <strong>and</strong> Japanese encephalitis.<br />

After the war he returned to his professorship at the<br />

University of Cincinnati College of Medicine <strong>and</strong> Children’s<br />

Hospital Research Foundation. The polio vaccine he developed<br />

there saved millions of children worldwide from paralytic polio.<br />

Many of these lives were doubtless saved because of his refusal<br />

to patent the vaccine, thereby making it simpler to produce<br />

<strong>and</strong> distribute <strong>and</strong> less expensive to administer<br />

Sabin’s work brought him more than forty honorary degrees<br />

from American <strong>and</strong> foreign universities <strong>and</strong> medals from the<br />

governments of the United States <strong>and</strong> Soviet Union. He was<br />

president of the Weizmann Institute of Science after 1970 <strong>and</strong><br />

later became a professor of biomedicine at the Medical University<br />

of South Carolina. He died in 1993.


584 / Polio vaccine (Sabin)<br />

Consequences<br />

The development of polio vaccines ranks as one of the triumphs of<br />

modern medicine. In the early 1950’s, paralytic polio struck 13,500<br />

out of every 100 million Americans. The use of the Salk vaccine<br />

greatly reduced the incidence of polio, but outbreaks of paralytic disease<br />

continued to occur: Fifty-seven hundred cases were reported in<br />

1959 <strong>and</strong> twenty-five hundred cases in 1960. In 1962, the oral Sabin<br />

vaccine became the vaccine of choice in the United States. Since its<br />

widespread use, the number of paralytic cases in the United States<br />

has dropped precipitously, eventually averaging fewer than ten per<br />

year. Worldwide, the oral vaccine prevented an estimated 5 million<br />

cases of paralytic poliomyelitis between 1970 <strong>and</strong> 1990.<br />

The oral vaccine is not without problems. Occasionally, the living<br />

virus mutates to a disease-causing (virulent) form as it multiplies in<br />

the vaccinated person. When this occurs, the person may develop<br />

paralytic poliomyelitis. The inactive vaccine, in contrast, cannot<br />

mutate to a virulent form. Ironically, nearly every incidence of polio<br />

in the United States is caused by the vaccine itself.<br />

In the developing countries of the world, the issue of vaccination is<br />

more pressing. Millions receive neither form of polio vaccine; as a result,<br />

at least 250,000 individuals are paralyzed or die each year. The World<br />

Health Organization <strong>and</strong> other health providers continue to work toward<br />

the very practical goal of completely eradicating this disease.<br />

See also Antibacterial drugs; Birth control pill; Iron lung; Penicillin;<br />

Polio vaccine (Salk); Reserpine; Salvarsan; Tuberculosis vaccine;<br />

Typhus vaccine; Yellow fever vaccine.<br />

Further Reading<br />

DeJauregui, Ruth. 100 Medical Milestones That Shaped World History.<br />

San Mateo, Calif.: Bluewood Books, 1998.<br />

Grady, Denise. “As Polio Fades, Dr. Salk’s Vaccine Re-emerges.”<br />

New York Times (December 14, 1999).<br />

Plotkin, Stanley A., <strong>and</strong> Edward A. Mortimer. Vaccines. 2d ed. Philadelphia:<br />

W. B. Saunders, 1994.<br />

Seavey, Nina Gilden, Jane S. Smith, <strong>and</strong> Paul Wagner. A Paralyzing<br />

Fear: The Triumph over Polio in America. New York: TV Books, 1998.


Polio vaccine (Salk)<br />

Polio vaccine (Salk)<br />

The invention: Jonas Salk’s vaccine was the first that prevented polio,<br />

resulting in the virtual eradication of crippling polio epidemics.<br />

The people behind the invention:<br />

Jonas Edward Salk (1914-1995), an American physician,<br />

immunologist, <strong>and</strong> virologist<br />

Thomas Francis, Jr. (1900-1969), an American microbiologist<br />

Cause for Celebration<br />

585<br />

Poliomyelitis (polio) is an infectious disease that can adversely<br />

affect the central nervous system, causing paralysis <strong>and</strong> great muscle<br />

wasting due to the destruction of motor neurons (nerve cells) in<br />

the spinal cord. Epidemiologists believe that polio has existed since<br />

ancient times, <strong>and</strong> evidence of its presence in Egypt, circa 1400 b.c.e.,<br />

has been presented. Fortunately, the Salk vaccine <strong>and</strong> the later vaccine<br />

developed by the American virologist Albert Bruce Sabin can<br />

prevent the disease. Consequently, except in underdeveloped nations,<br />

polio is rare. Moreover, although once a person develops polio,<br />

there is still no cure for it, a large number of polio cases end without<br />

paralysis or any observable effect.<br />

Polio is often called “infantile paralysis.” This results from the<br />

fact that it is seen most often in children. It is caused by a virus <strong>and</strong><br />

begins with body aches, a stiff neck, <strong>and</strong> other symptoms that are<br />

very similar to those of a severe case of influenza. In some cases,<br />

within two weeks after its onset, the course of polio begins to lead to<br />

muscle wasting <strong>and</strong> paralysis.<br />

On April 12, 1955, the world was thrilled with the announcement<br />

that Jonas Edward Salk’s poliomyelitis vaccine could prevent the<br />

disease. It was reported that schools were closed in celebration of<br />

this event. Salk, the son of a New York City garment worker, has<br />

since become one of the most well-known <strong>and</strong> publicly venerated<br />

medical scientists in the world.<br />

Vaccination is a method of disease prevention by immunization,<br />

whereby a small amount of virus is injected into the body to prevent


586 / Polio vaccine (Salk)<br />

a viral disease. The process depends on the production of antibodies<br />

(body proteins that are specifically coded to prevent the disease<br />

spread by the virus) in response to the vaccination. Vaccines are<br />

made of weakened or killed virus preparations.<br />

Electrifying Results<br />

Jonas Salk<br />

The son of a garment industry worker, Jonas Edward Salk<br />

was born in New York City in 1914. He worked his way through<br />

school, graduating from New York University School of Medicine<br />

in 1938. Afterward he joined microbiologist Thomas Francis,<br />

Jr., in developing a vaccine for influenza.<br />

In 1942, Salk began a research fellowship at the University of<br />

Michigan <strong>and</strong> subsequently joined the epidemiology faculty.<br />

He moved to the University of Pittsburgh in 1947, directing its<br />

Viral Research Lab, <strong>and</strong> while there developed his vaccine for<br />

poliomyelitis. The discovery catapulted Salk into worldwide<br />

fame, but he was a controversial figure among scientists.<br />

Although Salk received the Presidential Medal of Freedom,<br />

a Congressional gold medal, <strong>and</strong> the Nehru Award for International<br />

Underst<strong>and</strong>ing, he was turned down for membership in<br />

the National Academy of Sciences. In 1963 he opened the Salk<br />

Institute for Biological Sciences in La Jolla, California. Well<br />

aware of his reputation among medical researchers, he once<br />

joked, “I couldn’t possibly have become a member of this institute<br />

if I hadn’t founded it myself.” He died in 1995.<br />

The Salk vaccine was produced in two steps. First, polio viruses<br />

were grown in monkey kidney tissue cultures. These polio viruses<br />

were then killed by treatment with the right amount of formaldehyde<br />

to produce an effective vaccine. The killed-virus polio vaccine<br />

was found to be safe <strong>and</strong> to cause the production of antibodies<br />

against the disease, a sign that it should prevent polio.<br />

In early 1952, Salk tested a prototype vaccine against Type I polio virus<br />

on children who were afflicted with the disease <strong>and</strong> were thus<br />

deemed safe from reinfection. This test showed that the vaccination


greatly elevated the concentration of polio antibodies in these children.<br />

On July 2, 1952, encouraged by these results, Salk vaccinated fortythree<br />

children who had never had polio with vaccines against each of<br />

the three virus types (Type I, Type II, <strong>and</strong> Type III). All inoculated children<br />

produced high levels of polio antibodies, <strong>and</strong> none of them developed<br />

the disease. Consequently, the vaccine appeared to be both safe in<br />

humans <strong>and</strong> likely to become an effective public health tool.<br />

In 1953, Salk reported these findings in the Journal of the American<br />

Medical Association. In April, 1954, nationwide testing of the Salk<br />

vaccine began, via the mass vaccination of American schoolchildren.<br />

The results of the trial were electrifying. The vaccine was safe,<br />

<strong>and</strong> it greatly reduced the incidence of the disease. In fact, it was estimated<br />

that Salk’s vaccine gave schoolchildren 60 to 90 percent protection<br />

against polio.<br />

Salk was instantly praised. Then, however, several cases of polio<br />

occurred as a consequence of the vaccine. Its use was immediately<br />

suspended by the U.S. surgeon general, pending a complete examination.<br />

Soon, it was evident that all the cases of vaccine-derived polio<br />

were attributable to faulty batches of vaccine made by one<br />

pharmaceutical company. Salk <strong>and</strong> his associates were in no way responsible<br />

for the problem. Appropriate steps were taken to ensure<br />

that such an error would not be repeated, <strong>and</strong> the Salk vaccine was<br />

again released for use by the public.<br />

Consequences<br />

Polio vaccine (Salk) / 587<br />

The first reports on the polio epidemic in the United States had<br />

occurred on June 27, 1916, when one hundred residents of Brooklyn,<br />

New York, were afflicted. Soon, the disease had spread. By August,<br />

twenty-seven thous<strong>and</strong> people had developed polio. Nearly seven<br />

thous<strong>and</strong> afflicted people died, <strong>and</strong> many survivors of the epidemic<br />

were permanently paralyzed to varying extents. In New York City<br />

alone, nine thous<strong>and</strong> people developed polio <strong>and</strong> two thous<strong>and</strong><br />

died. Chaos reigned as large numbers of terrified people attempted<br />

to leave <strong>and</strong> were turned back by police. Smaller polio epidemics<br />

occurred throughout the nation in the years that followed (for example,<br />

the Catawba County, North Carolina, epidemic of 1944). A<br />

particularly horrible aspect of polio was the fact that more than 70


588 / Polio vaccine (Salk)<br />

percent of polio victims were small children. Adults caught it too;<br />

the most famous of these adult polio victims was U.S. President<br />

Franklin D. Roosevelt. There was no cure for the disease. The best<br />

available treatment was physical therapy.<br />

As of August, 1955, more than four million polio vaccines had<br />

been given. The Salk vaccine appeared to work very well. There were<br />

only half as many reported cases of polio in 1956 as there had been in<br />

1955. It appeared that polio was being conquered. By 1957, the number<br />

of cases reported nationwide had fallen below six thous<strong>and</strong>.<br />

Thus, in two years, its incidence had dropped by about 80 percent.<br />

This was very exciting, <strong>and</strong> soon other countries clamored for the<br />

vaccine. By 1959, ninety other countries had been supplied with the<br />

Salk vaccine. Worldwide, the disease was being eradicated. The introduction<br />

of an oral polio vaccine by Albert Bruce Sabin supported<br />

this progress.<br />

Salk received many honors, including honorary degrees from<br />

American <strong>and</strong> foreign universities, the Lasker Award, a Congressional<br />

Medal for Distinguished Civilian Service, <strong>and</strong> membership in<br />

the French Legion of Honor, yet he received neither the Nobel Prize<br />

nor membership in the American National Academy of Sciences. It<br />

is believed by many that this neglect was a result of the personal antagonism<br />

of some of the members of the scientific community who<br />

strongly disagreed with his theories of viral inactivation.<br />

See also Antibacterial drugs; Birth control pill; Iron lung; Penicillin;<br />

Polio vaccine (Sabin); Reserpine; Salvarsan; Tuberculosis vaccine;<br />

Typhus vaccine; Yellow fever vaccine.<br />

Further Reading<br />

DeJauregui, Ruth. 100 Medical Milestones That Shaped World History.<br />

San Mateo, Calif.: Bluewood Books, 1998.<br />

Plotkin, Stanley A., <strong>and</strong> Edward A. Mortimer. Vaccines. 2d ed. Philadelphia:<br />

W. B. Saunders, 1994.<br />

Seavey, Nina Gilden, Jane S. Smith, <strong>and</strong> Paul Wagner. A Paralyzing<br />

Fear: The Triumph over Polio in America. New York: TV Books, 1998.<br />

Smith, Jane S. Patenting the Sun: Polio <strong>and</strong> the Salk Vaccine. New York:<br />

Anchor/Doubleday, 1991.


Polyester<br />

Polyester<br />

The invention: A synthetic fibrous polymer used especially in fabrics.<br />

The people behind the invention:<br />

Wallace H. Carothers (1896-1937), an American polymer<br />

chemist<br />

Hilaire de Chardonnet (1839-1924), a French polymer chemist<br />

John R. Whinfield (1901-1966), a British polymer chemist<br />

A Story About Threads<br />

589<br />

Human beings have worn clothing since prehistoric times. At<br />

first, clothing consisted of animal skins sewed together. Later, people<br />

learned to spin threads from the fibers in plant or animal materials<br />

<strong>and</strong> to weave fabrics from the threads (for example, wool, silk,<br />

<strong>and</strong> cotton). By the end of the nineteenth century, efforts were begun<br />

to produce synthetic fibers for use in fabrics. These efforts were<br />

motivated by two concerns. First, it seemed likely that natural materials<br />

would become too scarce to meet the needs of a rapidly increasing<br />

world population. Second, a series of natural disasters—<br />

affecting the silk industry in particular—had demonstrated the<br />

problems of relying solely on natural fibers for fabrics.<br />

The first efforts to develop synthetic fabric focused on artificial<br />

silk, because of the high cost of silk, its beauty, <strong>and</strong> the fact that silk<br />

production had been interrupted by natural disasters more often<br />

than the production of any other material. The first synthetic silk<br />

was rayon, which was originally patented by a French count,<br />

Hilaire de Chardonnet, <strong>and</strong> was later much improved by other<br />

polymer chemists. Rayon is a semisynthetic material that is made<br />

from wood pulp or cotton.<br />

Because there was a need for synthetic fabrics whose manufacture<br />

did not require natural materials, other avenues were explored. One<br />

of these avenues led to the development of totally synthetic polyester<br />

fibers. In the United States, the best-known of these is Dacron, which<br />

is manufactured by E. I. Du Pont de Nemours. Easily made into


590 / Polyester<br />

threads, Dacron is widely used in clothing. It is also used to make audiotapes<br />

<strong>and</strong> videotapes <strong>and</strong> in automobile <strong>and</strong> boat bodies.<br />

From Polymers to Polyester<br />

Dacron belongs to a group of chemicals known as “synthetic<br />

polymers.” All polymers are made of giant molecules, each of<br />

which is composed of a large number of simpler molecules (“monomers”)<br />

that have been linked, chemically, to form long strings. Efforts<br />

by industrial chemists to prepare synthetic polymers developed<br />

in the twentieth century after it was discovered that many<br />

natural building materials <strong>and</strong> fabrics (such as rubber, wood, wool,<br />

silk, <strong>and</strong> cotton) were polymers, <strong>and</strong> as the ways in which monomers<br />

could be joined to make polymers became better understood.<br />

One group of chemists who studied polymers sought to make inexpensive<br />

synthetic fibers to replace expensive silk <strong>and</strong> wool. Their efforts<br />

led to the development of well-known synthetic fibers such as<br />

nylon <strong>and</strong> Dacron.<br />

Wallace H. Carothers of Du Pont pioneered the development of<br />

polyamide polymers, collectively called “nylon,” <strong>and</strong> was the first<br />

researcher to attempt to make polyester. It was British polymer<br />

chemists John R. Whinfield <strong>and</strong> J. T. Dickson of Calico Printers Association<br />

(CPA) Limited, however, who in 1941 perfected <strong>and</strong> patented<br />

polyester that could be used to manufacture clothing. The<br />

first polyester fiber products were produced in 1950 in Great Britain<br />

by London’s British Imperial Chemical Industries, which had secured<br />

the British patent rights from CPA. This polyester, which was<br />

made of two monomers, terphthalic acid <strong>and</strong> ethylene glycol, was<br />

called Terylene. In 1951, Du Pont, which had acquired Terylene patent<br />

rights for the Western Hemisphere, began to market its own version<br />

of this polyester, which was called Dacron. Soon, other companies<br />

around the world were selling polyester materials of similar<br />

composition.<br />

Dacron <strong>and</strong> other polyesters are used in many items in the<br />

United States. Made into fibers <strong>and</strong> woven, Dacron becomes cloth.<br />

When pressed into thin sheets, it becomes Mylar, which is used in<br />

videotapes <strong>and</strong> audiotapes. Dacron polyester, mixed with other materials,<br />

is also used in many industrial items, including motor vehi-


cle <strong>and</strong> boat bodies. Terylene <strong>and</strong> similar polyester preparations<br />

serve the same purposes in other countries.<br />

The production of polyester begins when monomers are mixed<br />

in huge reactor tanks <strong>and</strong> heated, which causes them to form giant<br />

polymer chains composed of thous<strong>and</strong>s of alternating monomer<br />

units. If T represents terphthalic acid <strong>and</strong> E represents ethylene glycol,<br />

a small part of a necklace-like polymer can be shown in the following<br />

way: (TETETETETE). Once each batch of polyester polymer<br />

has the desired composition, it is processed for storage until it is<br />

needed. In this procedure, the material, in liquid form in the hightemperature<br />

reactor, is passed through a device that cools it <strong>and</strong><br />

forms solid strips. These strips are then diced, dried, <strong>and</strong> stored.<br />

When polyester fiber is desired, the diced polyester is melted <strong>and</strong><br />

then forced through tiny holes in a “spinneret” device; this process<br />

is called “extruding.” The extruded polyester cools again, while<br />

passing through the spinneret holes, <strong>and</strong> becomes fine fibers called<br />

“filaments.” The filaments are immediately wound into threads that<br />

are collected in rolls. These rolls of thread are then dyed <strong>and</strong> used to<br />

weave various fabrics. If polyester sheets or other forms of polyester<br />

are desired, the melted, diced polyester is processed in other ways.<br />

Polyester preparations are often mixed with cotton, glass fibers, or<br />

other synthetic polymers to produce various products.<br />

Impact<br />

Polyester / 591<br />

The development of polyester was a natural consequence of the<br />

search for synthetic fibers that developed from work on rayon. Once<br />

polyester had been developed, its great utility led to its widespread<br />

use in industry. In addition, the profitability of the material spurred<br />

efforts to produce better synthetic fibers for specific uses. One example<br />

is that of stretchy polymers such as Helance, which is a form<br />

of nylon. In addition, new chemical types of polymer fibers were developed,<br />

including the polyurethane materials known collectively<br />

as “sp<strong>and</strong>ex” (for example, Lycra <strong>and</strong> Vyrenet).<br />

The wide variety of uses for polyester is amazing. Mixed with<br />

cotton, it becomes wash-<strong>and</strong>-wear clothing; mixed with glass, it is<br />

used to make boat <strong>and</strong> motor vehicle bodies; combined with other<br />

materials, it is used to make roofing materials, conveyor belts,


592 / Polyester<br />

hoses, <strong>and</strong> tire cords. In Europe, polyester has become the main<br />

packaging material for consumer goods, <strong>and</strong> the United States does<br />

not lag far behind in this area.<br />

The future is sure to hold more uses for polyester <strong>and</strong> the invention<br />

of new polymers. These spinoffs of polyester will be essential in<br />

the development of high technology.<br />

See also Buna rubber; Neoprene; Nylon; Orlon; Plastic; Polyethylene;<br />

Polystyrene.<br />

Further Reading<br />

Furukawa, Yasu. Inventing Polymer Science: Staudinger, Carothers, <strong>and</strong><br />

the Emergence of Macromolecular Chemistry. Philadelphia: University<br />

of Pennsylvania Press, 1998.<br />

H<strong>and</strong>ley, Susannah. Nylon: The Story of a Fashion Revolution, A Celebration<br />

of Design from Art Silk to Nylon <strong>and</strong> Thinking Fibres. Baltimore:<br />

Johns Hopkins University Press, 1999.<br />

Hermes, Matthew E. Enough for One Lifetime: Wallace Carothers, Inventor<br />

of Nylon. Washington, D.C.: American Chemical Society<br />

<strong>and</strong> the Chemical Heritage Foundation, 1996.<br />

Smith, Matthew Boyd. Polyester: The Indestructible Fashion. Atglen,<br />

Pa.: Schiffer, 1998.


Polyethylene<br />

Polyethylene<br />

The invention: An artificial polymer with strong insulating properties<br />

<strong>and</strong> many other applications.<br />

The people behind the invention:<br />

Karl Ziegler (1898-1973), a German chemist<br />

Giulio Natta (1903-1979), an Italian chemist<br />

August Wilhelm von Hofmann (1818-1892), a German chemist<br />

The Development of Synthetic Polymers<br />

593<br />

In 1841, August Hofmann completed his Ph.D. with Justus von<br />

Liebig, a German chemist <strong>and</strong> founding father of organic chemistry.<br />

One of Hofmann’s students, William Henry Perkin, discovered that<br />

coal tars could be used to produce brilliant dyes. The German chemical<br />

industry, under Hofmann’s leadership, soon took the lead in<br />

this field, primarily because the discipline of organic chemistry was<br />

much more developed in Germany than elsewhere.<br />

The realities of the early twentieth century found the chemical<br />

industry struggling to produce synthetic substitutes for natural<br />

materials that were in short supply, particularly rubber. Rubber is<br />

a natural polymer, a material composed of a long chain of small<br />

molecules that are linked chemically. An early synthetic rubber,<br />

neoprene, was one of many synthetic polymers (some others were<br />

Bakelite, polyvinyl chloride, <strong>and</strong> polystyrene) developed in the<br />

1920’s <strong>and</strong> 1930’s. Another polymer, polyethylene, was developed<br />

in 1936 by Imperial Chemical Industries. Polyethylene was a<br />

tough, waxy material that was produced at high temperature <strong>and</strong><br />

at pressures of about one thous<strong>and</strong> atmospheres. Its method of<br />

production made the material expensive, but it was useful as an insulating<br />

material.<br />

World War II <strong>and</strong> the material shortages associated with it brought<br />

synthetic materials into the limelight. Many new uses for polymers<br />

were discovered, <strong>and</strong> after the war they were in dem<strong>and</strong> for the production<br />

of a variety of consumer goods, although polyethylene was<br />

still too expensive to be used widely.


594 / Polyethylene<br />

Organometallics Provide the Key<br />

Karl Ziegler, an organic chemist with an excellent international<br />

reputation, spent most of his career in Germany. With his international<br />

reputation <strong>and</strong> lack of political connections, he was a natural<br />

c<strong>and</strong>idate to take charge of the Kaiser Wilhelm Institute for Coal Research<br />

(later renamed the Max Planck Institute) in 1943. Wise planners<br />

saw him as a director who would be favored by the conquering<br />

Allies. His appointment was a shrewd one, since he was allowed to<br />

retain his position after World War II ended. Ziegler thus played a<br />

key role in the resurgence of German chemical research after the war.<br />

Before accepting the position at the Kaiser Wilhelm Institute,<br />

Ziegler made it clear that he would take the job only if he could pursue<br />

his own research interests in addition to conducting coal research.<br />

The location of the institute in the Ruhr Valley meant that<br />

abundant supplies of ethylene were available from the local coal industry,<br />

so it is not surprising that Ziegler began experimenting with<br />

that material.<br />

Although Ziegler’s placement as head of the institute was an important<br />

factor in his scientific breakthrough, his previous research<br />

was no less significant. Ziegler devoted much time to the field of<br />

organometallic compounds, which are compounds that contain a<br />

metal atom that is bonded to one or more carbon atoms. Ziegler was<br />

interested in organoaluminum compounds, which are compounds<br />

that contain aluminum-carbon bonds.<br />

Ziegler was also interested in polymerization reactions, which<br />

involve the linking of thous<strong>and</strong>s of smaller molecules into the single<br />

long chain of a polymer. Several synthetic polymers were known,<br />

but chemists could exert little control on the actual process. It was<br />

impossible to regulate the length of the polymer chain, <strong>and</strong> the extent<br />

of branching in the chain was unpredictable. It was as a result of<br />

studying the effect of organoaluminum compounds on these chain<br />

formation reactions that the key discovery was made.<br />

Ziegler <strong>and</strong> his coworkers already knew that ethylene would react<br />

with organoaluminum compounds to produce hydrocarbons,<br />

which are compounds that contain only carbon <strong>and</strong> hydrogen <strong>and</strong><br />

that have varying chain lengths. Regulating the product chain length<br />

continued to be a problem.


At this point, fate intervened in the form of a trace of nickel left in a<br />

reactor from a previous experiment. The nickel caused the chain<br />

lengthening to stop after two ethylene molecules had been linked.<br />

Ziegler <strong>and</strong> his colleagues then tried to determine whether metals<br />

other than nickel caused a similar effect with a longer polymeric<br />

chain. Several metals were tested, <strong>and</strong> the most important finding<br />

was that a trace of titanium chloride in the reactor caused the deposition<br />

of large quantities of high-density polyethylene at low pressures.<br />

Ziegler licensed the procedure, <strong>and</strong> within a year, Giulio Natta<br />

had modified the catalysts to give high yields of polymers with<br />

highly ordered side chains branching from the main chain. This<br />

opened the door for the easy production of synthetic rubber. For<br />

their discovery of Ziegler-Natta catalysts, Ziegler <strong>and</strong> Natta shared<br />

the 1963 Nobel Prize in Chemistry.<br />

Consequences<br />

Polyethylene / 595<br />

Ziegler’s process produced polyethylene that was much more<br />

rigid than the material produced at high pressure. His product also<br />

had a higher density <strong>and</strong> a higher softening temperature. Industrial<br />

exploitation of the process was unusually rapid, <strong>and</strong> within ten years<br />

more than twenty plants utilizing the process had been built throughout<br />

Europe, producing more than 120,000 metric tons of polyethylene.<br />

This rapid exploitation was one reason Ziegler <strong>and</strong> Natta were<br />

awarded the Nobel Prize after such a relatively short time.<br />

By the late 1980’s, total production stood at roughly 18 billion<br />

pounds worldwide. Other polymeric materials, including polypropylene,<br />

can be produced by similar means. The ready availability<br />

<strong>and</strong> low cost of these versatile materials have radically transformed<br />

the packaging industry. Polyethylene bottles are far lighter<br />

than their glass counterparts; in addition, gases <strong>and</strong> liquids do not<br />

diffuse into polyethylene very easily, <strong>and</strong> it does not break easily.<br />

As a result, more <strong>and</strong> more products are bottled in containers<br />

made of polyethylene or other polymers. Other novel materials<br />

possessing properties unparalleled by any naturally occurring material<br />

(Kevlar, for example, which is used to make bullet-resistant<br />

vests) have also been an outgrowth of the availability of low-cost<br />

polymeric materials.


596 / Polyethylene<br />

See also Buna rubber; Neoprene; Nylon; Orlon; Plastic; Polyester;<br />

Polystyrene.<br />

Further Reading<br />

Boor, John. Ziegler-Natta Catalysts <strong>and</strong> Polymerizations. New York:<br />

Academic Press, 1979.<br />

Clarke, Alison J. Tupperware: The Promise of Plastic in 1950s America.<br />

Washington, D.C.: Smithsonian Institution Press, 1999.<br />

Natta, Giulio. “From Stereospecific Polymerization to Asymmetric<br />

Autocatalytic Synthesis of Macromolecules.” In Chemistry, 1963-<br />

1970. River Edge, N.J.: World Scientific, 1999.<br />

Ziegler, Karl. “Consequences <strong>and</strong> Development of an Invention.” In<br />

Chemistry, 1963-1970. River Edge, N.J.: World Scientific, 1999.


Polystyrene<br />

Polystyrene<br />

The invention: A clear, moldable polymer with many industrial<br />

uses whose overuse has also threatened the environment.<br />

The people behind the invention:<br />

Edward Simon, an American chemist<br />

Charles Gerhardt (1816-1856), a French chemist<br />

Marcellin Pierre Berthelot (1827-1907), a French chemist<br />

Polystyrene Is Characterized<br />

597<br />

In the late eighteenth century, a scientist by the name of Casper<br />

Neuman described the isolation of a chemical called “storax” from a<br />

balsam tree that grew in Asia Minor. This isolation led to the first report<br />

on the physical properties of the substance later known as “styrene.”<br />

The work of Neuman was confirmed <strong>and</strong> exp<strong>and</strong>ed upon<br />

years later, first in 1839 by Edward Simon, who evaluated the temperature<br />

dependence of styrene, <strong>and</strong> later by Charles Gerhardt,<br />

who proposed its molecular formula. The work of these two men<br />

sparked an interest in styrene <strong>and</strong> its derivatives.<br />

Polystyrene belongs to a special class of molecules known as<br />

polymers. Apolymer (the name means “many parts”) is a giant molecule<br />

formed by combining small molecular units, called “monomers.”<br />

This combination results in a macromolecule whose physical<br />

properties—especially its strength <strong>and</strong> flexibility—are significantly<br />

different from those of its monomer components. Such polymers are<br />

often simply called “plastics.”<br />

Polystyrene has become an important material in modern society<br />

because it exhibits a variety of physical characteristics that can be<br />

manipulated for the production of consumer products. Polystyrene<br />

is a “thermoplastic,” which means that it can be softened by heat<br />

<strong>and</strong> then reformed, after which it can be cooled to form a durable<br />

<strong>and</strong> resilient product.<br />

At 94 degrees Celsius, polystyrene softens; at room temperature,<br />

however, it rings like a metal when struck. Because of the glasslike<br />

nature <strong>and</strong> high refractive index of polystyrene, products made


598 / Polystyrene<br />

from it are known for their shine <strong>and</strong> attractive texture. In addition,<br />

the material is characterized by a high level of water resistance <strong>and</strong><br />

by electrical insulating qualities. It is also flammable, can by dissolved<br />

or softened by many solvents, <strong>and</strong> is sensitive to light. These<br />

qualities make polystyrene a valuable material in the manufacture<br />

of consumer products.<br />

Plastics on the Market<br />

In 1866, Marcellin Pierre Berthelot prepared styrene from ethylene<br />

<strong>and</strong> benzene mixtures in a heated reaction flask. This was the<br />

first synthetic preparation of polystyrene. In 1925, the Naugatuck<br />

Chemical Company began to operate the first commercial styrene/<br />

polystyrene manufacturing plant. In the 1930’s, the Dow Chemical<br />

Company became involved in the manufacturing <strong>and</strong> marketing of<br />

styrene/polystyrene products. Dow’s Styron 666 was first marketed<br />

as a general-purpose polystyrene in 1938. This material was<br />

the first plastic product to demonstrate polystyrene’s excellent mechanical<br />

properties <strong>and</strong> ease of fabrication.<br />

The advent of World War II increased the need for plastics. When<br />

the Allies’ supply of natural rubber was interrupted, chemists sought<br />

to develop synthetic substitutes. The use of additives with polymer<br />

species was found to alter some of the physical properties of those<br />

species. Adding substances called “elastomers” during the polymerization<br />

process was shown to give a rubberlike quality to a normally<br />

brittle species. An example of this is Dow’s Styron 475, which<br />

was marketed in 1948 as the first “impact” polystyrene. It is called<br />

an impact polystyrene because it also contains butadiene, which increases<br />

the product’s resistance to breakage. The continued characterization<br />

of polystyrene products has led to the development of a<br />

worldwide industry that fills a wide range of consumer needs.<br />

Following World War II, the plastics industry revolutionized<br />

many aspects of modern society. Polystyrene is only one of the<br />

many plastics involved in this process, but it has found its way into<br />

a multitude of consumer products. Disposable kitchen utensils,<br />

trays <strong>and</strong> packages, cups, videocassettes, insulating foams, egg cartons,<br />

food wrappings, paints, <strong>and</strong> appliance parts are only a few of<br />

the typical applications of polystyrenes. In fact, the production of


polystyrene has grown to exceed 5 billion pounds per year.<br />

The tremendous growth of this industry in the postwar era has<br />

been fueled by a variety of factors. Having studied the physical<br />

<strong>and</strong> chemical properties of polystyrene, chemists <strong>and</strong> engineers<br />

were able to envision particular uses <strong>and</strong> to tailor the manufacture<br />

of the product to fit those uses precisely. Because of its low cost of<br />

production, superior performance, <strong>and</strong> light weight, polystyrene<br />

has become the material of choice for the packaging industry. The<br />

automobile industry also enjoys its benefits. Polystyrene’s lower<br />

density compared to those of glass <strong>and</strong> steel makes it appropriate<br />

for use in automobiles, since its light weight means that using<br />

it can reduce the weight of automobiles, thereby increasing gas<br />

efficiency.<br />

Impact<br />

Polystyrene / 599<br />

There is no doubt that the marketing of polystyrene has greatly<br />

affected almost every aspect of modern society. From computer keyboards<br />

to food packaging, the use of polystyrene has had a powerful<br />

impact on both the quality <strong>and</strong> the prices of products. Its use is not,<br />

however, without drawbacks; it has also presented humankind<br />

with a dilemma. The wholesale use of polystyrene has created an<br />

environmental problem that represents a danger to wildlife, adds to<br />

roadside pollution, <strong>and</strong> greatly contributes to the volume of solid<br />

waste in l<strong>and</strong>fills.<br />

Polystyrene has become a household commodity because it lasts.<br />

The reciprocal effect of this fact is that it may last forever. Unlike natural<br />

products, which decompose upon burial, polystyrene is very<br />

difficult to convert into degradable forms. The newest challenge facing<br />

engineers <strong>and</strong> chemists is to provide for the safe <strong>and</strong> efficient<br />

disposal of plastic products. Thermoplastics such as polystyrene<br />

can be melted down <strong>and</strong> remolded into new products, which makes<br />

recycling <strong>and</strong> reuse of polystyrene a viable option, but this option<br />

requires the cooperation of the same consumers who have benefited<br />

from the production of polystyrene products.<br />

See also Food freezing; Nylon; Orlon; Plastic; Polyester; Polyethylene;<br />

Pyrex glass; Teflon; Tupperware.


600 / Polystyrene<br />

Further Reading<br />

Fenichell, Stephen. Plastic: The Making of a Synthetic Century. New<br />

York: HarperBusiness, 1997.<br />

Mossman, S. T. I. Early Plastics: Perspectives, 1850-1950. London: Science<br />

Museum, 1997.<br />

Wünsch, J. R. Polystyrene: Synthesis, Production <strong>and</strong> Applications.<br />

Shropshire, Engl<strong>and</strong>: Rapra Technology, 2000.


Propeller-coordinated<br />

machine gun<br />

Propeller-coordinated machine gun<br />

The invention: A mechanism that synchronized machine gun fire<br />

with propeller movement to prevent World War I fighter plane<br />

pilots from shooting off their own propellers during combat.<br />

The people behind the invention:<br />

Anthony Herman Gerard Fokker (1890-1939), a Dutch-born<br />

American entrepreneur, pilot, aircraft designer, <strong>and</strong><br />

manufacturer<br />

Rol<strong>and</strong> Garros (1888-1918), a French aviator<br />

Max Immelmann (1890-1916), a German aviator<br />

Raymond Saulnier (1881-1964), a French aircraft designer <strong>and</strong><br />

manufacturer<br />

French Innovation<br />

601<br />

The first true aerial combat of World War I took place in 1915. Before<br />

then, weapons attached to airplanes were inadequate for any<br />

real combat work. H<strong>and</strong>-held weapons <strong>and</strong> clumsily mounted machine<br />

guns were used by pilots <strong>and</strong> crew members in attempts to<br />

convert their observation planes into fighters. On April 1, 1915, this<br />

situation changed. From an airfield near Dunkerque, France, a<br />

French airman, Lieutenant Rol<strong>and</strong> Garros, took off in an airplane<br />

equipped with a device that would make his plane the most feared<br />

weapon in the air at that time.<br />

During a visit to Paris, Garros met with Raymond Saulnier, a French<br />

aircraft designer. In April of 1914, Saulnier had applied for a patent on<br />

a device that mechanically linked the trigger of a machine gun to a cam<br />

on the engine shaft. Theoretically, such an assembly would allow the<br />

gun to fire between the moving blades of the propeller. Unfortunately,<br />

the available machine gun Saulnier used to test his device was a<br />

Hotchkiss gun, which tended to fire at an uneven rate. On Garros’s arrival,<br />

Saulnier showed him a new invention: a steel deflector shield<br />

that, when fastened to the propeller, would deflect the small percentage<br />

of mistimed bullets that would otherwise destroy the blade.


602 / Propeller-coordinated machine gun<br />

The first test-firing was a disaster, shooting the propeller off <strong>and</strong><br />

destroying the fuselage. Modifications were made to the deflector<br />

braces, streamlining its form into a wedge shape with gutterchannels<br />

for deflected bullets. The invention was attached to a<br />

Morane-Saulnier monoplane, <strong>and</strong> on April 1, Garros took off alone<br />

toward the German lines. Success was immediate. Garros shot<br />

down a German observation plane that morning. During the next<br />

two weeks, Garros shot down five more German aircraft.<br />

German Luck<br />

The German high comm<strong>and</strong>, frantic over the effectiveness of the<br />

French “secret weapon,” sent out spies to try to steal the secret <strong>and</strong><br />

also ordered engineers to develop a similar weapon. Luck was with<br />

them. On April 18, 1915, despite warnings by his superiors not to fly<br />

over enemy-held territory, Garros was forced to crash-l<strong>and</strong> behind<br />

German lines with engine trouble. Before he could destroy his aircraft,<br />

Garros <strong>and</strong> his plane were captured by German troops. The secret<br />

weapon was revealed.<br />

The Germans were ecstatic about the opportunity to examine<br />

the new French weapon. Unlike the French, the Germans had the<br />

first air-cooled machine gun, the Parabellum, which shot continuous<br />

b<strong>and</strong>s of one hundred bullets <strong>and</strong> was reliable enough to be<br />

adapted to a timing mechanism.<br />

In May of 1915, Anthony Herman Gerard Fokker was shown<br />

Garros’s captured plane <strong>and</strong> was ordered to copy the idea. Instead,<br />

Fokker <strong>and</strong> his assistant designed a new firing system. It is unclear<br />

whether Fokker <strong>and</strong> his team were already working on a synchronizer<br />

or to what extent they knew of Saulnier’s previous work in<br />

France. Within several days, however, they had constructed a working<br />

prototype <strong>and</strong> attached it to a Fokker Eindecker 1 airplane. The<br />

design consisted of a simple linkage of cams <strong>and</strong> push-rods connected<br />

to the oil-pump drive of an Oberursel engine <strong>and</strong> the trigger<br />

of a Parabellum machine gun. The firing of the gun had to be timed<br />

precisely to fire its six hundred rounds per minute between the<br />

twelve-hundred-revolutions-per-minute propeller blades.<br />

Fokker took his invention to Doberitz air base, <strong>and</strong> after a series


Propeller-coordinated machine gun / 603<br />

Anthony Herman Gerard Fokker<br />

Anthony Fokker was born on the isl<strong>and</strong> of Java in the Dutch<br />

East Indies (now Indonesia) in 1890. He returned to his parent’s<br />

home country, the Netherl<strong>and</strong>s, to attend school <strong>and</strong> then studied<br />

aeronautics in Germany. He built his first plane in 1910 <strong>and</strong><br />

established Fokker Aeroplanbau near Berlin in 1912.<br />

His monoplanes were highly esteemed when World War I<br />

erupted in 1914, <strong>and</strong> he offered his designs to both the German<br />

<strong>and</strong> the French governments. The Germans hired him. By the<br />

end of the war his fighters, especially the Dr I triplane <strong>and</strong> D VII<br />

biplane, were practically synonymous with German air warfare<br />

because they had been the scourge of Allied pilots.<br />

In 1922 Fokker moved to the United States <strong>and</strong> opened the<br />

Atlantic Aircraft Corporation in New Jersey. He had lost enthusiasm<br />

for military aircraft <strong>and</strong> turned his skills toward producing<br />

advanced designs for civilian use. The planes his company<br />

turned out established one first after another. His T-2 monoplane<br />

became the first to fly nonstop from coast to coast, New<br />

York to San Diego. His ten-seat airliner, the F VII/3m, carried<br />

Lieutenant Comm<strong>and</strong>er Richard Byrd over the North Pole in<br />

1926 <strong>and</strong> Charles Kingsford-Smith across the Pacific Ocean in<br />

1928.<br />

By the time Fokker died in New York in 1939, he had become<br />

a visionary. He foresaw passenger planes as the means to knit<br />

together the far-flung nations of the world into a network of<br />

rapid travel <strong>and</strong> communications.<br />

of exhausting trials before the German high comm<strong>and</strong>, both on the<br />

ground <strong>and</strong> in the air, he was allowed to take two prototypes of the<br />

machine-gun-mounted airplanes to Douai in German-held France.<br />

At Douai, two German pilots crowded into the cockpit with Fokker<br />

<strong>and</strong> were given demonstrations of the plane’s capabilities. The airmen<br />

were Oswald Boelcke, a test pilot <strong>and</strong> veteran of forty reconnaissance<br />

missions, <strong>and</strong> Max Immelmann, a young, skillful aviator<br />

who was assigned to the front.<br />

When the first combat-ready versions of Fokker’s Eindecker 1<br />

were delivered to the front lines, one was assigned to Boelcke, the<br />

other to Immelmann. On August 1, 1915, with their aerodrome un-


604 / Propeller-coordinated machine gun<br />

der attack from nine English bombers, Boelcke <strong>and</strong> Immelmann<br />

manned their aircraft <strong>and</strong> attacked. Boelcke’s gun jammed, <strong>and</strong> he<br />

was forced to cut off his attack <strong>and</strong> return to the aerodrome. Immelmann,<br />

however, succeeded in shooting down one of the bombers<br />

with his synchronized machine gun. It was the first victory credited<br />

to the Fokker-designed weapon system.<br />

Impact<br />

At the outbreak of World War I, military strategists <strong>and</strong> comm<strong>and</strong>ers<br />

on both sides saw the wartime function of airplanes as a<br />

means to supply intelligence information behind enemy lines or as<br />

airborne artillery spotting platforms. As the war progressed <strong>and</strong> aircraft<br />

flew more or less freely across the trenches, providing vital information<br />

to both armies, it became apparent to ground comm<strong>and</strong>ers<br />

that while it was important to obtain intelligence on enemy<br />

movements, it was important also to deny the enemy similar information.<br />

Early in the war, the French used airplanes as strategic bombing<br />

platforms. As both armies began to use their air forces for strategic<br />

bombing of troops, railways, ports, <strong>and</strong> airfields, it became evident<br />

that aircraft would have to be employed against enemy aircraft to<br />

prevent reconnaissance <strong>and</strong> bombing raids.<br />

With the invention of the synchronized forward-firing machine<br />

gun, pilots could use their aircraft as attack weapons. A pilot finally<br />

could coordinate control of his aircraft <strong>and</strong> his armaments with<br />

maximum efficiency. This conversion of aircraft from nearly passive<br />

observation platforms to attack fighters is the single greatest innovation<br />

in the history of aerial warfare. The development of fighter<br />

aircraft forced a change in military strategy, tactics, <strong>and</strong> logistics <strong>and</strong><br />

ushered in the era of modern warfare. Fighter planes are responsible<br />

for the battle-tested military adage: Whoever controls the sky controls<br />

the battlefield.<br />

See also Airplane; Radar; Stealth aircraft.


Further Reading<br />

Propeller-coordinated machine gun / 605<br />

Dierikx, M. L. J. Fokker: A Transatlantic Biography. Washington:<br />

Smithsonian Institution Press, 1997.<br />

Franks, Norman L. R. Aircraft Versus Aircraft: The Illustrated Story of<br />

Fighter Pilot Combat from 1914 to the Present Day. New York:<br />

Barnes & Noble Books, 1999.<br />

Guttman, Jon. Fighting Firsts: Fighter Aircraft Combat Debuts from<br />

1914 to 1944. London: Cassell, 2000.


606<br />

Pyrex glass<br />

Pyrex glass<br />

The invention: A superhard <strong>and</strong> durable glass product with widespread<br />

uses in industry <strong>and</strong> home products.<br />

The people behind the invention:<br />

Jesse T. Littleton (1888-1966), the chief physicist of Corning<br />

Glass Works’ research department<br />

Eugene G. Sullivan (1872-1962), the founder of Corning’s<br />

research laboratories<br />

William C. Taylor (1886-1958), an assistant to Sullivan<br />

Cooperating with Science<br />

By the twentieth century, Corning Glass Works had a reputation<br />

as a corporation that cooperated with the world of science to improve<br />

existing products <strong>and</strong> develop new ones. In the 1870’s, the<br />

company had hired university scientists to advise on improving the<br />

optical quality of glasses, an early example of today’s common practice<br />

of academics consulting for industry.<br />

When Eugene G. Sullivan established Corning’s research laboratory<br />

in 1908 (the first of its kind devoted to glass research), the task<br />

that he undertook with William C. Taylor was that of making a heatresistant<br />

glass for railroad lantern lenses. The problem was that ordinary<br />

flint glass (the kind in bottles <strong>and</strong> windows, made by melting<br />

together silica s<strong>and</strong>, soda, <strong>and</strong> lime) has a fairly high thermal expansion,<br />

but a poor heat conductivity. The glass thus exp<strong>and</strong>s<br />

unevenly when exposed to heat. This condition can cause the glass<br />

to break, sometimes violently. Colored lenses for oil or gas railroad<br />

signal lanterns sometimes shattered if they were heated too much<br />

by the flame that produced the light <strong>and</strong> were then sprayed by rain<br />

or wet snow. This changed a red “stop” light to a clear “proceed”<br />

signal <strong>and</strong> caused many accidents or near misses in railroading in<br />

the late nineteenth century.<br />

Two solutions were possible: to improve the thermal conductivity<br />

or reduce the thermal expansion. The first is what metals do:<br />

When exposed to heat, most metals have an expansion much greater


than that of glass, but they conduct heat so quickly that they exp<strong>and</strong><br />

nearly equally throughout <strong>and</strong> seldom lose structural integrity from<br />

uneven expansion. Glass, however, is an inherently poor heat conductor,<br />

so this approach was not possible.<br />

Therefore, a formulation had to be found that had little or no<br />

thermal expansivity. Pure silica (one example is quartz) fits this description,<br />

but it is expensive <strong>and</strong>, with its high melting point, very<br />

difficult to work.<br />

The formulation that Sullivan <strong>and</strong> Taylor devised was a borosilicate<br />

glass—essentially a soda-lime glass with the lime replaced by<br />

borax, with a small amount of alumina added. This gave the low thermal<br />

expansion needed for signal lenses. It also turned out to have<br />

good acid-resistance, which led to its being used for the battery jars<br />

required for railway telegraph systems <strong>and</strong> other applications. The<br />

glass was marketed as “Nonex” (for “nonexpansion glass”).<br />

From the Railroad to the Kitchen<br />

Pyrex glass / 607<br />

Jesse T. Littleton joined Corning’s research laboratory in 1913.<br />

The company had a very successful lens <strong>and</strong> battery jar material,<br />

but no one had even considered it for cooking or other heat-transfer<br />

applications, because the prevailing opinion was that glass absorbed<br />

<strong>and</strong> conducted heat poorly. This meant that, in glass pans,<br />

cakes, pies, <strong>and</strong> the like would cook on the top, where they were exposed<br />

to hot air, but would remain cold <strong>and</strong> wet (or at least undercooked)<br />

next to the glass surface. As a physicist, Littleton knew that<br />

glass absorbed radiant energy very well. He thought that the heatconduction<br />

problem could be solved by using the glass vessel itself<br />

to absorb <strong>and</strong> distribute heat. Glass also had a significant advantage<br />

over metal in baking. Metal bakeware mostly reflects radiant energy<br />

to the walls of the oven, where it is lost ultimately to the surroundings.<br />

Glass would absorb this radiation energy <strong>and</strong> conduct it evenly to<br />

the cake or pie, giving a better result than that of the metal bakeware.<br />

Moreover, glass would not absorb <strong>and</strong> carry over flavors from<br />

one baking effort to the next, as some metals do.<br />

Littleton took a cut-off battery jar home <strong>and</strong> asked his wife to<br />

bake a cake in it. He took it to the laboratory the next day, h<strong>and</strong>ing<br />

pieces around <strong>and</strong> not disclosing the method of baking until all had


608 / Pyrex glass<br />

Jesse T. Littleton<br />

To prove that glass is good for baking, place an uncooked pie<br />

in a pie tin <strong>and</strong> place another pie pan under it, made half of tin<br />

<strong>and</strong> half of non-exp<strong>and</strong>ing glass. Place it in all the oven. That is<br />

the experiment Jesse Talbot Littleton, Jr., used at Corning Glass<br />

Works soon after he hired on in 1913. The story behind it began<br />

with a ceramic dish that cracked when his wife baked a cake.<br />

That would not happen, he realized, with the right kind of<br />

glass. Although his wife baked a cake successfully in a glass<br />

battery jar bottom at his request, Littleton had to demonstrate<br />

the feat for his superiors scientifically. The half of the pie over<br />

the glass, it turned out, cooked faster <strong>and</strong> more evenly. Kitchen<br />

glassware was born.<br />

Littleton was born in Belle Haven, Virginia, in 1888. After<br />

taking degrees from Southern University <strong>and</strong> Tulane University,<br />

he earned a doctorate in physics from the University of<br />

Wisconsin in 1911. He briefly vowed to remain a bachelor <strong>and</strong><br />

dedicate his life to physics, but Besse Cook, a pretty Mississippi<br />

school teacher, turned his head, <strong>and</strong> so he got married instead.<br />

He was the first physicist added to the newly organized research<br />

laboratories at Corning in New York. There he studied<br />

practical problems involved in the industrial applications of<br />

glass, including tempering, <strong>and</strong> helped invent a gas pressure<br />

meter to measure the flow of air in blowing glass <strong>and</strong> a sensitive,<br />

faster thermometer. He rose rapidly in the organization. In<br />

1920 he became chief of the physical lab, assistant director of research<br />

in 1940, vice president in 1943, director of all Corning research<br />

<strong>and</strong> development in 1946, <strong>and</strong> general technical adviser<br />

in 1951.<br />

Littleton retired a year later <strong>and</strong>, a passionate outdoorsman,<br />

devoted himself to hunting <strong>and</strong> fishing. A leading figure in the<br />

ceramics industry, he belonged to the American Academy for<br />

the Advancement of Science, American Physical Society, <strong>and</strong><br />

the American Institute of Engineers <strong>and</strong> was an editor for the<br />

Journal of Applied Physics. He died in 1966.


agreed that the results were excellent. With this agreement, he was<br />

able to commit laboratory time to developing variations on the<br />

Nonex formula that were more suitable for cooking. The result was<br />

Pyrex, patented <strong>and</strong> trademarked in May of 1915.<br />

Impact<br />

Pyrex glass / 609<br />

In the 1930’s, Pyrex “Flameware” was introduced, with a new<br />

glass formulation that could resist the increased heat of stovetop<br />

cooking. In the half century since Flameware was introduced,<br />

Corning went on to produce a variety of other products <strong>and</strong> materials:<br />

tableware in tempered opal glass; cookware in Pyroceram, a<br />

glass product that during heat treatment gained such mechanical<br />

strength as to be virtually unbreakable; even hot plates <strong>and</strong> stoves<br />

topped with Pyroceram.<br />

In the same year that Pyrex was marketed for cooking, it was<br />

also introduced for laboratory apparatus. Laboratory glassware<br />

had been coming from Germany at the beginning of the twentieth<br />

century; World War I cut off the supply. Corning filled the gap<br />

with Pyrex beakers, flasks, <strong>and</strong> other items. The delicate blownglass<br />

equipment that came from Germany was completely displaced<br />

by the more rugged <strong>and</strong> heat-resistant machine-made Pyrex<br />

ware.<br />

Any number of operations are possible with Pyrex that cannot<br />

be performed safely in flint glass: Test tubes can be thrust directly<br />

into burner flames, with no preliminary warming; beakers <strong>and</strong><br />

flasks can be heated on hot plates; <strong>and</strong> materials that dissolve<br />

when exposed to heat can be made into solutions directly in Pyrex<br />

storage bottles, a process that cannot be performed in regular<br />

glass. The list of such applications is almost endless.<br />

Pyrex has also proved to be the material of choice for lenses in<br />

the great reflector telescopes, beginning in 1934 with that at Mount<br />

Palomar. By its nature, astronomical observation must be done<br />

with the scope open to the weather. This means that the mirror<br />

must not change shape with temperature variations, which rules<br />

out metal mirrors. Silvered (or aluminized) Pyrex serves very well,<br />

<strong>and</strong> Corning has developed great expertise in casting <strong>and</strong> machining<br />

Pyrex blanks for mirrors of all sizes.


610 / Pyrex glass<br />

See also Laminated glass; Microwave cooking; Plastic; Polystyrene;<br />

Teflon; Tupperware.<br />

Further Reading<br />

Blaszczyk, Regina Lee. Imagining Consumers: Design <strong>and</strong> Innovation<br />

from Wedgwood to Corning. Baltimore: Johns Hopkins University<br />

Press, 2000.<br />

Graham, Margaret B. W., <strong>and</strong> Alec T. Shuldiner. Corning <strong>and</strong> the Craft<br />

of Innovation. New York: Oxford University Press, 2001.<br />

Stage, Sarah, <strong>and</strong> Virginia Bramble Vincenti. Rethinking Home Economics:<br />

Women <strong>and</strong> the History of a Profession. Ithaca, N.Y.: Cornell<br />

University Press, 1997.<br />

Rogove, Susan Tobier, <strong>and</strong> Marcia B. Steinhauer. Pyrex by Corning: A<br />

Collector’s Guide. Marietta, Ohio: Antique <strong>Public</strong>ations, 1993.


Radar<br />

Radar<br />

The invention: An electronic system for detecting objects at great<br />

distances, radar was a major factor in the Allied victory of World<br />

War II <strong>and</strong> now pervades modern life, including scientific research.<br />

The people behind the invention:<br />

Sir Robert Watson-Watt (1892-1973), the father of radar who<br />

proposed the chain air-warning system<br />

Arnold F. Wilkins, the person who first calculated the intensity<br />

of a radio wave<br />

William C. Curtis (1914-1976), an American engineer<br />

Looking for Thunder<br />

611<br />

Sir Robert Watson-Watt, a scientist with twenty years of experience<br />

in government, led the development of the first radar, an acronym<br />

for radio detection <strong>and</strong> ranging. “Radar” refers to any instrument<br />

that uses the reflection of radio waves to determine the<br />

distance, direction, <strong>and</strong> speed of an object.<br />

In 1915, during World War I (1914-1918), Watson-Watt joined<br />

Great Britain’s Meteorological Office. He began work on the detection<br />

<strong>and</strong> location of thunderstorms at the Royal Aircraft Establishment<br />

in Farnborough <strong>and</strong> remained there throughout the<br />

war. Thunderstorms were known to be a prolific source of “atmospherics”<br />

(audible disturbances produced in radio receiving apparatus<br />

by atmospheric electrical phenomena), <strong>and</strong> Watson-Watt<br />

began the design of an elementary radio direction finder that<br />

gave the general position of such storms. Research continued after<br />

the war <strong>and</strong> reached a high point in 1922 when sealed-off<br />

cathode-ray tubes first became available. With assistance from<br />

J. F. Herd, a fellow Scot who had joined him at Farnborough, he<br />

constructed an instantaneous direction finder, using the new<br />

cathode-ray tubes, that gave the direction of thunderstorm activity.<br />

It was admittedly of low sensitivity, but it worked, <strong>and</strong> it was<br />

the first of its kind.


612 / Radar<br />

William C. Curtis<br />

In addition to radar’s applications in navigation, civil aviation,<br />

<strong>and</strong> science, it rapidly became an integral part of military<br />

aircraft by guiding weaponry <strong>and</strong> detecting enemy aircraft <strong>and</strong><br />

missiles. The research <strong>and</strong> development industry that grew to<br />

provide offensive <strong>and</strong> defensive systems greatly exp<strong>and</strong>ed the<br />

opportunities for young scientists during the Cold War. Among<br />

them was William C. Curtis (1914-1976), one of the most influential<br />

African Americans in defense research.<br />

Curtis graduated from the Tuskegee Institute (later Tuskegee<br />

University), where he later served as its first dean of engineering.<br />

While there, he helped form <strong>and</strong> train the Tuskegee<br />

Airmen, a famous squadron of African American fighter pilots<br />

during World War II. He also worked for the Radio Corporation<br />

of American (RCA) for twenty-three years. It was while at RCA<br />

that he contributed innovations to military radar. These include<br />

the Black Cat weapons system, MG-3 fire control system, 300-A<br />

weapon radar system, <strong>and</strong> Airborne Interceptor Data Link.<br />

Watson-Watt did much of this work at a new site at Ditton Park,<br />

near Slough, where the National Physical Laboratory had a field<br />

station devoted to radio research. In 1927, the two endeavors were<br />

combined as the Radio Research Station; it came under the general<br />

supervision of the National Physical Laboratory, with Watson-Watt<br />

as the first superintendent. This became a center with unrivaled expertise<br />

in direction finding using the cathode-ray tube <strong>and</strong> in studying<br />

the ionosphere using radio waves. No doubt these facilities<br />

were a factor when Watson-Watt invented radar in 1935.<br />

As radar developed, its practical uses exp<strong>and</strong>ed. Meteorological<br />

services around the world, using ground-based radar, gave warning<br />

of approaching rainstorms. Airborne radars proved to be a great<br />

help to aircraft by allowing them to recognize potentially hazardous<br />

storm areas. This type of radar was used also to assist research into<br />

cloud <strong>and</strong> rain physics. In this type of research, radar-equipped research<br />

aircraft observe the radar echoes inside a cloud as rain develops,<br />

<strong>and</strong> then fly through the cloud, using on-board instruments to<br />

measure the water content.


Technician at a modern radar display. (PhotoDisc)


614 / Radar<br />

Aiming Radar at the Moon<br />

The principles of radar were further developed through the discipline<br />

of radio astronomy. This field began with certain observations<br />

made by the American electrical engineer Karl Jansky in 1933<br />

at the Bell Laboratories at Holmdell, New Jersey. Radio astronomers<br />

learn about objects in space by intercepting the radio waves that<br />

these objects emit.<br />

Jansky found that radio signals were coming to Earth from space.<br />

He called these mysterious pulses “cosmic noise.” In particular, there<br />

was an unusual amount of radar noise when the radio antennas were<br />

pointed at the Sun, which increased at the time of sun-spot activity.<br />

All this information lay dormant until after World War II (1939-<br />

1945), at which time many investigators turned their attention to interpreting<br />

the cosmic noise. The pioneers were Sir Bernard Lovell at<br />

Manchester, Engl<strong>and</strong>, Sir Martin Ryle at Cambridge, Engl<strong>and</strong>, <strong>and</strong><br />

Joseph Pawsey of the Commonwealth of Science Industrial Research<br />

Organization, in Australia. The intensity of these radio waves was<br />

first calculated by Arnold F. Wilkins.<br />

As more powerful tools became available toward the end of<br />

World War II, curiosity caused experimenters to try to detect radio<br />

signals from the Moon. This was accomplished successfully in the<br />

late 1940’s <strong>and</strong> led to experiments on other objects in the solar system:<br />

planets, satellites, comets, <strong>and</strong> asteroids.<br />

Impact<br />

Radar introduced some new <strong>and</strong> revolutionary concepts into warfare,<br />

<strong>and</strong> in doing so gave birth to entirely new branches of technology.<br />

In the application of radar to marine navigation, the long-range<br />

navigation system developed during the war was taken up at once<br />

by the merchant fleets that used military-style radar equipment<br />

without modification. In addition, radar systems that could detect<br />

buoys <strong>and</strong> other ships <strong>and</strong> obstructions in closed waters, particularly<br />

under conditions of low visibility, proved particularly useful<br />

to peacetime marine navigation.<br />

In the same way, radar was adopted to assist in the navigation of<br />

civil aircraft. The various types of track guidance systems devel-


oped after the war were aimed at guiding aircraft in the critical last<br />

hundred kilometers or so of their run into an airport. Subsequent<br />

improvements in the system meant that an aircraft could place itself<br />

on an approach or l<strong>and</strong>ing path with great accuracy.<br />

The ability of radar to measure distance to an extraordinary degree<br />

of accuracy resulted in the development of an instrument that<br />

provided pilots with a direct measurement of the distances between<br />

airports. Along with these aids, ground-based radars were developed<br />

for the control of aircraft along the air routes or in the airport<br />

control area.<br />

The development of electronic computers can be traced back to<br />

the enormous advances in circuit design, which were an integral part<br />

of radar research during the war. During that time, some elements<br />

of electronic computing had been built into bombsights <strong>and</strong> other<br />

weaponry; later, it was realized that a whole range of computing operations<br />

could be performed electronically. By the end of the war,<br />

many pulse-forming networks, pulse-counting circuits, <strong>and</strong> memory<br />

circuits existed in the form needed for an electronic computer.<br />

Finally, the developing radio technology has continued to help<br />

astronomers explore the universe. Large radio telescopes exist in almost<br />

every country <strong>and</strong> enable scientists to study the solar system<br />

in great detail. Radar-assisted cosmic background radiation studies<br />

have been a building block for the big bang theory of the origin of<br />

the universe.<br />

See also Airplane; Cruise missile; Radio interferometer; Sonar;<br />

Stealth aircraft.<br />

Further Reading<br />

Radar / 615<br />

Brown, Louis. A Radar History of World War II: Technical <strong>and</strong> Military<br />

Imperatives. Philadelphia: Institute of Physics, 1999.<br />

Latham, Colin, <strong>and</strong> Anne Stobbs. Pioneers of Radar. Gloucestershire:<br />

Sutton, 1999.<br />

Rowl<strong>and</strong>, John. The Radar Man: The Story of Sir Robert Watson-Watt.<br />

New York: Roy Publishers, 1964.<br />

Watson-Watt, Robert Alex<strong>and</strong>er. The Pulse of Radar: The Autobiography<br />

of Sir Robert Watson-Watt. New York: Dial Press, 1959.


616<br />

Radio<br />

Radio<br />

The invention: The first radio transmissions of music <strong>and</strong> voice<br />

laid the basis for the modern radio <strong>and</strong> television industries.<br />

The people behind the invention:<br />

Guglielmo Marconi (1874-1937), an Italian physicist <strong>and</strong><br />

inventor<br />

Reginald Aubrey Fessenden (1866-1932), an American radio<br />

pioneer<br />

True Radio<br />

The first major experimenter in the United States to work with<br />

wireless radio was Reginald Aubrey Fessenden. This transplanted<br />

Canadian was a skilled, self-made scientist, but unlike American inventor<br />

Thomas Alva Edison, he lacked the business skills to gain the<br />

full credit <strong>and</strong> wealth that such pathbreaking work might have merited.<br />

Guglielmo Marconi, in contrast, is most often remembered as<br />

the person who invented wireless (as opposed to telegraphic) radio.<br />

There was a great difference between the contributions of Marconi<br />

<strong>and</strong> Fessenden. Marconi limited himself to experiments with<br />

radio telegraphy; that is, he sought to send through the air messages<br />

that were currently being sent by wire—signals consisting of dots<br />

<strong>and</strong> dashes. Fessenden sought to perfect radio telephony, or voice<br />

communication by wireless transmission. Fessenden thus pioneered<br />

the essential precursor of modern radio broadcasting. At the beginning<br />

of the twentieth century, Fessenden spent much time <strong>and</strong> energy<br />

publicizing his experiments, thus promoting interest in the<br />

new science of radio broadcasting.<br />

Fessenden began his career as an inventor while working for the<br />

U.S. Weather Bureau. He set out to invent a radio system by which<br />

to broadcast weather forecasts to users on l<strong>and</strong> <strong>and</strong> at sea. Fessenden<br />

believed that his technique of using continuous waves in the<br />

radio frequency range (rather than interrupted waves Marconi had<br />

used to produce the dots <strong>and</strong> dashes of Morse code) would provide<br />

the power necessary to carry Morse telegraph code yet be effective<br />

enough to h<strong>and</strong>le voice communication. He would turn out to be


correct. He conducted experiments as early as 1900 at Rock Point,<br />

Maryl<strong>and</strong>, about 80 kilometers south of Washington, D.C., <strong>and</strong> registered<br />

his first patent in the area of radio research in 1902.<br />

Fame <strong>and</strong> Glory<br />

Radio / 617<br />

In 1900, Fessenden asked the General Electric Company to produce<br />

a high-speed generator of alternating current—or alternator—<br />

to use as the basis of his radio transmitter. This proved to be the first<br />

major request for wireless radio apparatus that could project voices<br />

<strong>and</strong> music. It took the engineers three years to design <strong>and</strong> deliver<br />

the alternator. Meanwhile, Fessenden worked on an improved radio<br />

receiver. To fund his experiments, Fessenden aroused the interest<br />

of financial backers, who put up one million dollars to create the<br />

National Electric Signalling Company in 1902.<br />

Fessenden, along with a small group of h<strong>and</strong>picked scientists,<br />

worked at Brant Rock on the Massachusetts coast south of Boston.<br />

Working outside the corporate system, Fessenden sought fame <strong>and</strong><br />

glory based on his own work, rather than on something owned by a<br />

corporate patron.<br />

Fessenden’s moment of glory came on December 24, 1906, with<br />

the first announced broadcast of his radio telephone. Using an ordinary<br />

telephone microphone <strong>and</strong> his special alternator to generate<br />

the necessary radio energy, Fessenden alerted ships up <strong>and</strong> down<br />

the Atlantic coast with his wireless telegraph <strong>and</strong> arranged for<br />

newspaper reporters to listen in from New York City. Fessenden<br />

made himself the center of the show. He played the violin, sang,<br />

<strong>and</strong> read from the Bible. Anticipating what would become st<strong>and</strong>ard<br />

practice fifty years later, Fessenden also transmitted the sounds of a<br />

phonograph recording. He ended his first broadcast by wishing those<br />

listening “a Merry Christmas.” A similar, equally well-publicized<br />

demonstration came on December 31.<br />

Although Fessenden was skilled at drawing attention to his invention<br />

<strong>and</strong> must be credited, among others, as one of the engineering<br />

founders of the principles of radio, he was far less skilled at<br />

making money with his experiments, <strong>and</strong> thus his long-term impact<br />

was limited. The National Electric Signalling Company had a fine<br />

beginning <strong>and</strong> for a time was a supplier of equipment to the United


618 / Radio<br />

Fruit Company. The financial panic of 1907, however, wiped out an<br />

opportunity to sell the Fessenden patents—at a vast profit—to a corporate<br />

giant, the American Telephone <strong>and</strong> Telegraph Corporation.<br />

Impact<br />

Had there been more receiving equipment available <strong>and</strong> in place,<br />

a massive audience could have heard Fessenden’s first broadcast.<br />

He had the correct idea, even to the point of playing a crude phonograph<br />

record. Yet Fessenden, Marconi, <strong>and</strong> their rivals were unable<br />

to establish a regular series of broadcasts. Their “stations” were experimental<br />

<strong>and</strong> promotional.<br />

It took the stresses of World War I to encourage broader use of<br />

wireless radio based on Fessenden’s experiments. Suddenly, communicating<br />

from ship to ship or from a ship to shore became a frequent<br />

matter of life or death. Generating publicity was no longer<br />

necessary. Governments fought over crucial patent rights. The Radio<br />

Corporation of America (RCA) pooled vital knowledge. Ultimately,<br />

RCA came to acquire the Fessenden patents. Radio broadcasting<br />

commenced, <strong>and</strong> the radio industry, with its multiple uses<br />

for mass communication, was off <strong>and</strong> running.<br />

Antique tabletop radio. (PhotoDisc)


Guglielmo Marconi<br />

Guglielmo Marconi failed his entrance examinations to the<br />

University of Bologna in 1894. He had a weak educational background,<br />

particularly in science, but he was not about to let<br />

that—or his father’s disapproval—stop him after he conceived<br />

a deep interest in wireless telegraphy during his teenage years.<br />

Marconi was born in 1874 to a wealthy Italian l<strong>and</strong>owner<br />

<strong>and</strong> an Irish whiskey distiller’s daughter <strong>and</strong> grew up both in<br />

Italy <strong>and</strong> Engl<strong>and</strong>. His parents provided tutors for<br />

him, but he <strong>and</strong> his brother often accompanied their<br />

mother, a socialite, on extensive travels. He acquired<br />

considerable social skills, easy self-confidence, <strong>and</strong><br />

determination from the experience.<br />

Thus, when he failed his exams, he simply tried another<br />

route for his ambitions. He <strong>and</strong> his mother persuaded<br />

a science professor to let Marconi use a university<br />

laboratory unofficially. His father thought it a<br />

waste of time. However, he changed his mind when<br />

his son succeeded in building equipment that could<br />

transmit electronic signals around their house without wires, an<br />

achievement right at the vanguard of technology.<br />

Now supported by his father’s money, Marconi <strong>and</strong> his<br />

brother built an elaborate set of equipment—including an oscillator,<br />

coherer, galvanometer, <strong>and</strong> antennas—that they hoped<br />

would send a signal outside over a long distance. His brother<br />

walked off a mile <strong>and</strong> a half, out of sight, with the galvanometer<br />

<strong>and</strong> a rifle. When the galvanometer moved, indicating a signal<br />

had arrived from the oscillator, he fired the rifle to let Marconi<br />

know he had succeeded. The incident is widely cited as the first<br />

radio transmission.<br />

Marconi went on to send signals over greater <strong>and</strong> greater<br />

distances. He patented a tuner to permit transmissions at specific<br />

frequencies, <strong>and</strong> he started the Wireless Telegraph <strong>and</strong> Signal<br />

Company to bring his inventions to the public; its American<br />

branch was the Radio Corporation of America (RCA). He not<br />

only grew wealthy at a young age; he also was awarded half of<br />

the 1909 Nobel Prize in Physics for his work. He died in Rome<br />

in 1937, one of the most famous inventors in the world.<br />

Radio / 619<br />

(Library of Congress)


620 / Radio<br />

See also Communications satellite; Compact disc; Dolby noise<br />

reduction; FM radio; Long-distance radiotelephony; Radio crystal<br />

sets; Television; Transistor; Transistor radio.<br />

Further Reading<br />

Fessenden, <strong>and</strong> Helen May Trott. Fessenden: Builder of Tomorrows.<br />

New York: Arno Press, 1974.<br />

Lewis, Tom. Empire of the Air: The Men Who Made Radio. New York:<br />

HarperPerennial, 1993.<br />

Masini, Giancarlo. Marconi. New York: Marsilio Publishers, 1995.<br />

Seitz. Frederick. The Cosmic Inventor: Reginald Aubrey Fessenden,<br />

1866-1932. Philadelphia: American Philosophical Society, 1999.


Radio crystal sets<br />

Radio crystal sets<br />

The invention: The first primitive radio receivers, crystal sets led<br />

to the development of the modern radio.<br />

The people behind the invention:<br />

H. H. Dunwoody (1842-1933), an American inventor<br />

Sir John A. Fleming (1849-1945), a British scientist-inventor<br />

Heinrich Rudolph Hertz (1857-1894), a German physicist<br />

Guglielmo Marconi (1874-1937), an Italian engineer-inventor<br />

James Clerk Maxwell (1831-1879), a Scottish physicist<br />

Greenleaf W. Pickard (1877-1956), an American inventor<br />

From Morse Code to Music<br />

621<br />

In the 1860’s, James Clerk Maxwell demonstrated that electricity<br />

<strong>and</strong> light had electromagnetic <strong>and</strong> wave properties. The conceptualization<br />

of electromagnetic waves led Maxwell to propose that<br />

such waves, made by an electrical discharge, would eventually be<br />

sent long distances through space <strong>and</strong> used for communication<br />

purposes. Then, near the end of the nineteenth century, the technology<br />

that produced <strong>and</strong> transmitted the needed Hertzian (or radio)<br />

waves was devised by Heinrich Rudolph Hertz, Guglielmo Marconi<br />

(inventor of the wireless telegraph), <strong>and</strong> many others. The resultant<br />

radio broadcasts, however, were limited to the dots <strong>and</strong><br />

dashes of the Morse code.<br />

Then, in 1901, H. H. Dunwoody <strong>and</strong> Greenleaf W. Pickard invented<br />

the crystal set. Crystal sets were the first radio receivers<br />

that made it possible to hear music <strong>and</strong> the many other types of<br />

now-familiar radio programs. In addition, the simple construction<br />

of the crystal set enabled countless amateur radio enthusiasts<br />

to build “wireless receivers” (the name for early radios) <strong>and</strong><br />

to modify them. Although, except as curiosities, crystal sets were<br />

long ago replaced by more effective radios, they are where it all<br />

began.


622 / Radio crystal sets<br />

Crystals, Diodes, Transistors, <strong>and</strong> Chips<br />

Radio broadcasting works by means of electromagnetic radio<br />

waves, which are low-energy cousins of light waves. All electromagnetic<br />

waves have characteristic vibration frequencies <strong>and</strong> wavelengths.<br />

This article will deal mostly with long radio waves of frequencies<br />

from 550 to 1,600 kilocycles (kilohertz), which can be seen<br />

on amplitude-modulation (AM) radio dials. Frequency-modulation<br />

(FM), shortwave, <strong>and</strong> microwave radio transmission use higherenergy<br />

radio frequencies.<br />

The broadcasting of radio programs begins with the conversion<br />

of sound to electrical impulses by means of microphones. Then, radio<br />

transmitters turn the electrical impulses into radio waves that<br />

are broadcast together with higher-energy carrier waves. The combined<br />

waves travel at the speed of light to listeners. Listeners hear<br />

radio programs by using radio receivers that pick up broadcast<br />

waves through antenna wires <strong>and</strong> reverse the steps used in broadcasting.<br />

This is done by converting those waves to electrical impulses<br />

<strong>and</strong> then into sound waves. The two main types of radio<br />

broadcasting are AM <strong>and</strong> FM, which allow the selection (modulation)<br />

of the power (amplitude) or energy (frequency) of the broadcast<br />

waves.<br />

The crystal set radio receiver of Dunwoody <strong>and</strong> Pickard had<br />

many shortcomings. These led to the major modifications that produced<br />

modern radios. Crystal sets, however, began the radio industry<br />

<strong>and</strong> fostered its development. Today, it is possible to purchase<br />

somewhat modified forms of crystal sets, as curiosity items. All<br />

crystal sets, original or modern versions, are crude AM radio receivers<br />

that are composed of four components: an antenna wire, a crystal<br />

detector, a tuning circuit, <strong>and</strong> a headphone or loudspeaker.<br />

Antenna wires (aerials) pick up radio waves broadcast by external<br />

sources. Originally simple wires, today’s aerials are made to<br />

work better by means of insulation <strong>and</strong> grounding. The crystal detector<br />

of a crystal set is a mineral crystal that allows radio waves to<br />

be selected (tuned). The original detectors were crystals of a leadsulfur<br />

mineral, galena. Later, other minerals (such as silicon <strong>and</strong> carborundum)<br />

were also found to work. The tuning circuit is composed<br />

of 80 to 100 turns of insulated wire, wound on a 0.33-inch


support. Some surprising supports used in homemade tuning circuits<br />

include cardboard toilet-paper-roll centers <strong>and</strong> Quaker Oats<br />

cereal boxes. When realism is desired in collector crystal sets, the<br />

coil is usually connected to a wire probe selector called a “cat’s<br />

whisker.” In some such crystal sets, a condenser (capacitor) <strong>and</strong> additional<br />

components are used to extend the range of tunable signals.<br />

Headphones convert chosen radio signals to sound waves that are<br />

heard by only one listener. If desired, loudspeakers can be used to<br />

enable a roomful of listeners to hear chosen programs.<br />

An interesting characteristic of the crystal set is the fact that its<br />

operation does not require an external power supply. Offsetting<br />

this are its short reception range <strong>and</strong> a great difficulty in tuning or<br />

maintaining tuned-in radio signals. The short range of these radio<br />

receivers led to, among other things, the use of power supplies<br />

(house current or batteries) in more sophisticated radios. Modern<br />

solutions to tuning problems include using manufactured diode<br />

vacuum tubes to replace crystal detectors, which are a kind of natural<br />

diode. The first manufactured diodes, used in later crystal sets<br />

<strong>and</strong> other radios, were invented by John Ambrose Fleming, a colleague<br />

of Marconi’s. Other modifications of crystal sets that led to<br />

more sophisticated modern radios include more powerful aerials,<br />

better circuits, <strong>and</strong> vacuum tubes. Then came miniaturization,<br />

which was made possible by the use of transistors <strong>and</strong> silicon chips.<br />

Impact<br />

Radio crystal sets / 623<br />

The impact of the invention of crystal sets is almost incalculable,<br />

since they began the modern radio industry. These early radio receivers<br />

enabled countless radio enthusiasts to build radios, to receive radio<br />

messages, <strong>and</strong> to become interested in developing radio communication<br />

systems. Crystal sets can be viewed as having spawned all<br />

the variant modern radios. These include boom boxes <strong>and</strong> other portable<br />

radios; navigational radios used in ships <strong>and</strong> supersonic jet<br />

airplanes; <strong>and</strong> the shortwave, microwave, <strong>and</strong> satellite networks<br />

used in the various aspects of modern communication.<br />

The later miniaturization of radios <strong>and</strong> the development of sophisticated<br />

radio system components (for example, transistors<br />

<strong>and</strong> silicon chips) set the stage for both television <strong>and</strong> computers.


624 / Radio crystal sets<br />

Certainly, if one tried to assess the ultimate impact of crystal sets by<br />

simply counting the number of modern radios in the United States,<br />

one would find that few Americans more than ten years old own<br />

fewer than two radios. Typically, one of these is run by house electric<br />

current <strong>and</strong> the other is a portable set that is carried almost everywhere.<br />

See also FM radio; Long-distance radiotelephony; Radio; Television;<br />

Transistor radio.<br />

Further Reading<br />

Masini, Giancarlo. Marconi. New York: Marsilio, 1995.<br />

Sievers, Maurice L. Crystal Clear: Vintage American Crystal Sets, Crystal<br />

Detectors, <strong>and</strong> Crystals. Vestal, N.Y.: Vestal Press, 1991.<br />

Tolstoy, Ivan. James Clerk Maxwell: A Biography. Chicago: University<br />

of Chicago Press, 1982.


Radio interferometer<br />

Radio interferometer<br />

The invention: An astronomical instrument that combines multiple<br />

radio telescopes into a single system that makes possible the<br />

exploration of distant space.<br />

The people behind the invention:<br />

Sir Martin Ryle (1918-1984), an English astronomer<br />

Karl Jansky (1905-1950), an American radio engineer<br />

Hendrik Christoffel van de Hulst (1918- ), a Dutch radio<br />

astronomer<br />

Harold Irving Ewan (1922- ), an American astrophysicist<br />

Edward Mills Purcell (1912-1997), an American physicist<br />

Seeing with Radio<br />

625<br />

Since the early 1600’s, astronomers have relied on optical telescopes<br />

for viewing stellar objects. Optical telescopes detect the<br />

visible light from stars, galaxies, quasars, <strong>and</strong> other astronomical<br />

objects. Throughout the late twentieth century, astronomers developed<br />

more powerful optical telescopes for peering deeper into the<br />

cosmos <strong>and</strong> viewing objects located hundreds of millions of lightyears<br />

away from the earth.<br />

In 1933, Karl Jansky, an American radio engineer with Bell Telephone<br />

Laboratories, constructed a radio antenna receiver for locating<br />

sources of telephone interference. Jansky discovered a daily radio<br />

burst that he was able to trace to the center of the Milky Way<br />

galaxy. In 1935, Grote Reber, another American radio engineer, followed<br />

up Jansky’s work with the construction of the first dishshaped<br />

“radio” telescope. Reber used his 9-meter-diameter radio<br />

telescope to repeat Jansky’s experiments <strong>and</strong> to locate other radio<br />

sources in space. He was able to map precisely the locations of various<br />

radio sources in space, some of which later were identified as<br />

galaxies <strong>and</strong> quasars.<br />

Following World War II (that is, after 1945), radio astronomy<br />

blossomed with the help of surplus radar equipment. Radio astronomy<br />

tries to locate objects in space by picking up the radio waves


626 / Radio interferometer<br />

that they emit. In 1944, the Dutch astronomer Hendrik Christoffel<br />

van de Hulst had proposed that hydrogen atoms emit radio waves<br />

with a 21-centimeter wavelength. Because hydrogen is the most<br />

abundant element in the universe, van de Hulst’s discovery had explained<br />

the nature of extraterrestrial radio waves. His theory later<br />

was confirmed by the American radio astronomers Harold Irving<br />

Ewen <strong>and</strong> Edward Mills Purcell of Harvard University.<br />

By coupling the newly invented computer technology with radio<br />

telescopes, astronomers were able to generate a radio image of a star<br />

almost identical to the star’s optical image. A major advantage of radio<br />

telescopes over optical telescopes is the ability of radio telescopes<br />

to detect extraterrestrial radio emissions day or night, as well as their<br />

ability to bypass the cosmic dust that dims or blocks visible light.<br />

More with Less<br />

After 1945, major research groups were formed in Engl<strong>and</strong>, Australia,<br />

<strong>and</strong> The Netherl<strong>and</strong>s. Sir Martin Ryle was head of the Mullard<br />

Radio Astronomy Observatory of the Cavendish Laboratory,<br />

University of Cambridge. He had worked with radar for the Telecommunications<br />

Research Establishment during World War II.<br />

The radio telescopes developed by Ryle <strong>and</strong> other astronomers<br />

operate on the same basic principle as satellite television receivers.<br />

A constant stream of radio waves strikes the parabolic-shaped reflector<br />

dish, which aims all the radio waves at a focusing point<br />

above the dish. The focusing point directs the concentrated radio<br />

beam to the center of the dish, where it is sent to a radio receiver,<br />

then an amplifier, <strong>and</strong> finally to a chart recorder or computer.<br />

With large-diameter radio telescopes, astronomers can locate<br />

stars <strong>and</strong> galaxies that cannot be seen with optical telescopes. This<br />

ability to detect more distant objects is called “resolution.” Like<br />

optical telescopes, large-diameter radio telescopes have better resolution<br />

than smaller ones. Very large radio telescopes were constructed<br />

in the late 1950’s <strong>and</strong> early 1960’s (Jodrell Bank, Engl<strong>and</strong>;<br />

Green Bank, West Virginia; Arecibo, Puerto Rico). Instead of just<br />

building larger radio telescopes to achieve greater resolution, however,<br />

Ryle developed a method called “interferometry.” In Ryle’s<br />

method, a computer is used to combine the incoming radio waves


Moving<br />

Spacecraft<br />

Angular<br />

Separation Fixed<br />

Radio interferometer / 627<br />

Radio Star<br />

(Quasar)<br />

California Spain<br />

One use of VLBI is to navigate a spacecraft: By measuring the angular separation between a<br />

fixed radio star, such as a quasar, <strong>and</strong> a moving spacecraft, the craft’s location, orientation,<br />

<strong>and</strong> path can be precisely monitored <strong>and</strong> adjusted.<br />

of two or more movable radio telescopes pointed at the same stellar<br />

object.<br />

Suppose that one had a 30-meter-diameter radio telescope. Its radio<br />

wave-collecting area would be limited to its diameter. If a second<br />

identical 30-meter-diameter radio telescope was linked with<br />

the first, then one would have an interferometer. The two radio telescopes<br />

would point exactly at the same stellar object, <strong>and</strong> the radio<br />

emissions from this object captured by the two telescopes would be<br />

combined by computer to produce a higher-resolution image. If the<br />

two radio telescopes were located 1.6 kilometers apart, then their<br />

combined resolution would be equivalent to that of a single radio<br />

telescope dish 1.6 kilometers in diameter.<br />

Ryle constructed the first true radio telescope interferometer at<br />

the Mullard Radio Astronomy Observatory in 1955. He used combinations<br />

of radio telescopes to produce interferometers containing<br />

about twelve radio receivers. Ryle’s interferometer greatly improved<br />

radio telescope resolution for detecting stellar radio sources, mapping<br />

the locations of stars <strong>and</strong> galaxies, assisting in the discovery of


628 / Radio interferometer<br />

“quasars” (quasi-stellar radio sources), measuring the earth’s rotation<br />

around the Sun, <strong>and</strong> measuring the motion of the solar system<br />

through space.<br />

Consequences<br />

Following Ryle’s discovery, interferometers were constructed at<br />

radio astronomy observatories throughout the world. The United<br />

States established the National Radio Astronomy Observatory (NRAO)<br />

in rural Green Bank, West Virginia. The NRAO is operated by nine<br />

eastern universities <strong>and</strong> is funded by the National Science Foundation.<br />

At Green Bank, a three-telescope interferometer was constructed,<br />

with each radio telescope having a 26-meter-diameter<br />

dish. During the late 1970’s, the NRAO constructed the largest radio<br />

interferometer in the world, the Very Large Array (VLA). The VLA,<br />

located approximately 80 kilometers west of Socorro, New Mexico,<br />

consists of twenty-seven 25-meter-diameter radio telescopes linked<br />

by a supercomputer. The VLA has a resolution equivalent to that of<br />

a single radio telescope 32 kilometers in diameter.<br />

Even larger radio telescope interferometers can be created with<br />

a technique known as “very long baseline interferometry” (VLBI).<br />

VLBI has been used to construct a radio telescope having an effective<br />

diameter of several thous<strong>and</strong> kilometers. Such an arrangement<br />

involves the precise synchronization of radio telescopes located<br />

in several different parts of the world. Supernova 1987A in<br />

the Large Magellanic Cloud was studied using a VLBI arrangement<br />

between observatories located in Australia, South America,<br />

<strong>and</strong> South Africa.<br />

Launching radio telescopes into orbit <strong>and</strong> linking them with<br />

ground-based radio telescopes could produce a radio telescope<br />

whose effective diameter would be larger than that of the earth.<br />

Such instruments will enable astronomers to map the distribution<br />

of galaxies, quasars, <strong>and</strong> other cosmic objects, to underst<strong>and</strong> the<br />

origin <strong>and</strong> evolution of the universe, <strong>and</strong> possibly to detect meaningful<br />

radio signals from extraterrestrial civilizations.<br />

See also Neutrino detector; Weather satellite; Artificial satellite;<br />

Communications satellite; Radar; Rocket; Weather satellite.


Further Reading<br />

Radio interferometer / 629<br />

Graham-Smith, Francis. Sir Martin Ryle: A Biographical Memoir. London:<br />

Royal Society, 1987.<br />

Malphrus, Benjamin K. The History of Radio Astronomy <strong>and</strong> the National<br />

Radio Astronomy Observatory: Evolution Toward Big Science.<br />

Malabar, Fla.: Krieger, 1996.<br />

Pound, Robert V. Edward Mills Purcell: August 30, 1912-March 7,<br />

1997. Washington, D.C.: National Academy Press, 2000.


630<br />

Refrigerant gas<br />

Refrigerant gas<br />

The invention: A safe refrigerant gas for domestic refrigerators,<br />

dichlorodifluoromethane helped promote a rapid growth in the<br />

acceptance of electrical refrigerators in homes.<br />

The people behind the invention:<br />

Thomas Midgley, Jr. (1889-1944), an American engineer <strong>and</strong><br />

chemist<br />

Charles F. Kettering (1876-1958), an American engineer <strong>and</strong><br />

inventor who was the head of research for General Motors<br />

Albert Henne (1901-1967), an American chemist who was<br />

Midgley’s chief assistant<br />

Frédéric Swarts (1866-1940), a Belgian chemist<br />

Toxic Gases<br />

Refrigerators, freezers, <strong>and</strong> air conditioners have had a major impact<br />

on the way people live <strong>and</strong> work in the twentieth century. With<br />

them, people can live more comfortably in hot <strong>and</strong> humid areas,<br />

<strong>and</strong> a great variety of perishable foods can be transported <strong>and</strong><br />

stored for extended periods. As recently as the early nineteenth century,<br />

the foods most regularly available to Americans were bread<br />

<strong>and</strong> salted meats. Items now considered essential to a balanced diet,<br />

such as vegetables, fruits, <strong>and</strong> dairy products, were produced <strong>and</strong><br />

consumed only in small amounts.<br />

Through the early part of the twentieth century, the pattern of<br />

food storage <strong>and</strong> distribution evolved to make perishable foods<br />

more available. Farmers shipped dairy products <strong>and</strong> frozen meats<br />

to mechanically refrigerated warehouses. Smaller stores <strong>and</strong> most<br />

American households used iceboxes to keep perishable foods fresh.<br />

The iceman was a familiar figure on the streets of American towns,<br />

delivering large blocks of ice regularly.<br />

In 1930, domestic mechanical refrigerators were being produced<br />

in increasing numbers. Most of them were vapor compression machines,<br />

in which a gas was compressed in a closed system of pipes<br />

outside the refrigerator by a mechanical pump <strong>and</strong> condensed to a


liquid. The liquid was pumped into a sealed chamber in the refrigerator<br />

<strong>and</strong> allowed to evaporate to a gas. The process of evaporation<br />

removes heat from the environment, thus cooling the interior of the<br />

refrigerator.<br />

The major drawback of early home refrigerators involved the<br />

types of gases used. In 1930, these included ammonia, sulfur dioxide,<br />

<strong>and</strong> methyl chloride. These gases were acceptable if the refrigerator’s<br />

gas pipes never sprang a leak. Unfortunately, leaks sometimes<br />

occurred, <strong>and</strong> all these gases are toxic. Ammonia <strong>and</strong> sulfur<br />

dioxide both have unpleasant odors; if they leaked, at least they<br />

would be detected rapidly. Methyl chloride however, can form a<br />

dangerously explosive mixture with air, <strong>and</strong> it has only a very faint,<br />

<strong>and</strong> not unpleasant, odor. In a hospital in Clevel<strong>and</strong> during the<br />

1920’s, a refrigerator with methyl chloride leaked, <strong>and</strong> there was a<br />

disastrous explosion of the methyl chloride-air mixture. After that,<br />

methyl chloride for use in refrigerators was mixed with a small<br />

amount of a very bad-smelling compound to make leaks detectable.<br />

(The same tactic is used with natural gas.)<br />

Three-Day Success<br />

Refrigerant gas / 631<br />

General Motors, through its Frigidaire division, had a substantial<br />

interest in the domestic refrigerator market. Frigidaire refrigerators<br />

used sulfur dioxide as the refrigerant gas. Charles F. Kettering,<br />

director of research for General Motors, decided that Frigidaire<br />

needed a new refrigerant gas that would have good thermal properties<br />

but would be nontoxic <strong>and</strong> nonexplosive. In early 1930, he sent<br />

Lester S. Keilholtz, chief engineer of General Motors’ Frigidaire division,<br />

to Thomas Midgley, Jr., a mechanical engineer <strong>and</strong> selftaught<br />

chemist. He challenged them to develop such a new gas.<br />

Midgley’s associates, Albert Henne <strong>and</strong> Robert McNary, researched<br />

what types of compounds might already fit Kettering’s specifications.<br />

Working with research that had been done by the Belgian<br />

chemist Frédéric Swarts in the late nineteenth <strong>and</strong> early twentieth<br />

centuries, Midgley, Henne, <strong>and</strong> McNary realized that dichlorodifluoromethane<br />

would have ideal thermal properties <strong>and</strong> the right<br />

boiling point for a refrigerant gas. The only question left to be answered<br />

was whether the compound was toxic.


632 / Refrigerant gas<br />

The chemists prepared a few grams of dichlorodifluoromethane<br />

<strong>and</strong> put it, along with a guinea pig, into a closed chamber. They<br />

were delighted to see that the animal seemed to suffer no ill effects<br />

at all <strong>and</strong> was able to breathe <strong>and</strong> move normally. They were briefly<br />

puzzled when a second batch of the compound killed a guinea pig<br />

almost instantly. Soon, they discovered that an impurity in one of<br />

the ingredients had produced a potent poison in their refrigerant<br />

gas. A simple washing procedure completely removed the poisonous<br />

contaminant.<br />

This astonishingly successful research project was completed in<br />

three days. The boiling point of dichlorodifluoromethane is −5.6 degrees<br />

Celsius. It is nontoxic <strong>and</strong> nonflammable <strong>and</strong> possesses excellent<br />

thermal properties. When Midgley was awarded the Perkin<br />

Medal for industrial chemistry in 1937, he gave the audience a<br />

graphic demonstration of the properties of dichlorodifluoromethane:<br />

He inhaled deeply of its vapors <strong>and</strong> exhaled gently into a jar<br />

containing a burning c<strong>and</strong>le. The c<strong>and</strong>le flame promptly went out.<br />

This visual evidence proved that dichlorodifluoromethane was not<br />

poisonous <strong>and</strong> would not burn.<br />

Impact<br />

The availability of this safe refrigerant gas, which was renamed<br />

Freon, led to drastic changes in the United States. The current patterns<br />

of food production, distribution, <strong>and</strong> consumption are a direct<br />

result, as is air conditioning. Air conditioning was developed early<br />

in the twentieth century; by the late 1970’s, most American cars <strong>and</strong><br />

residences were equipped with air conditioning, <strong>and</strong> other countries<br />

with hot climates followed suit. Consequently, major relocations<br />

of populations <strong>and</strong> businesses have become possible. Since<br />

World War II, there have been steady migrations to the “Sun Belt,”<br />

the states spanning the United States from southeast to southwest,<br />

because air conditioners have made these areas much more livable.<br />

Freon is a member of a family of chemicals called “chlorofluorocarbons.”<br />

In addition to refrigeration, it is also used as a propellant<br />

in aerosols <strong>and</strong> in the production of polystyrene plastics. In 1974,<br />

scientists began to suspect that chlorofluorocarbons, when released<br />

into the air, might have a serious effect on the environment. They


speculated that the compounds might migrate into the stratosphere,<br />

where they could be decomposed by the intense ultraviolet light<br />

from the sunlight that is prevented from reaching the earth’s surface<br />

by the thin but vital layer of ozone in the stratosphere. In the process,<br />

large amounts of the ozone layer might also be destroyed—<br />

letting in the dangerous ultraviolet light. In addition to possible climatic<br />

effects, the resulting increase in ultraviolet light reaching the<br />

earth’s surface would raise the incidence of skin cancers. As a result,<br />

chemical manufacturers are trying to develop alternative refrigerant<br />

gases that will not harm the ozone layer.<br />

See also Electric refrigerator; Electric refrigerator; Food freezing;<br />

Microwave cooking.<br />

Further Reading<br />

Refrigerant gas / 633<br />

Leslie, Stuart W. Boss Kettering. New York: Columbia University<br />

Press, 1983.<br />

Mahoney, Thomas A. “The Seventy-one-year Saga of CFC’s.” Air<br />

Conditioning, Heating <strong>and</strong> Refrigeration News (March 15, 1999).<br />

Preville, Cherie R., <strong>and</strong> Chris King. “Cooling Takes Off in the Roaring<br />

Twenties.” Air Conditioning, Heating <strong>and</strong> Refrigeration News<br />

(April 30, 2001).


634<br />

Reserpine<br />

Reserpine<br />

The invention: A drug with unique hypertension-decreasing effects<br />

that provides clinical medicine with a versatile <strong>and</strong> effective<br />

tool.<br />

The people behind the invention:<br />

Robert Wallace Wilkins (1906- ), an American physician <strong>and</strong><br />

clinical researcher<br />

Walter E. Judson (1916- ) , an American clinical researcher<br />

Treating Hypertension<br />

Excessively elevated blood pressure, clinically known as “hypertension,”<br />

has long been recognized as a pervasive <strong>and</strong> serious human<br />

malady. In a few cases, hypertension is recognized as an effect<br />

brought about by particular pathologies (diseases or disorders). Often,<br />

however, hypertension occurs as the result of unknown causes.<br />

Despite the uncertainty about its origins, unattended hypertension<br />

leads to potentially dramatic health problems, including increased<br />

risk of kidney disease, heart disease, <strong>and</strong> stroke.<br />

Recognizing the need to treat hypertension in a relatively straightforward<br />

<strong>and</strong> effective way, Robert Wallace Wilkins, a clinical researcher<br />

at Boston University’s School of Medicine <strong>and</strong> the head of<br />

Massachusetts Memorial Hospital’s Hypertension Clinic, began to<br />

experiment with reserpine in the early 1950’s. Initially, the samples<br />

that were made available to Wilkins were crude <strong>and</strong> unpurified.<br />

Eventually, however, a purified version was used.<br />

Reserpine has a long <strong>and</strong> fascinating history of use—both clinically<br />

<strong>and</strong> in folk medicine—in India. The source of reserpine is the<br />

root of the shrub Rauwolfia serpentina, first mentioned in Western<br />

medical literature in the 1500’s but virtually unknown, or at least<br />

unaccepted, outside India until the mid-twentieth century. Crude<br />

preparations of the shrub had been used for a variety of ailments in<br />

India for centuries prior to its use in the West.<br />

Wilkins’s work with the drug did not begin on an encouraging<br />

note, because reserpine does not act rapidly—a fact that had been


noted in Indian medical literature. The st<strong>and</strong>ard observation in<br />

Western pharmacotherapy, however, was that most drugs work<br />

rapidly; if a week has elapsed without positive effects being shown<br />

by a drug, the conventional Western wisdom is that it is unlikely<br />

to work at all. Additionally, physicians <strong>and</strong> patients alike tend to<br />

look for rapid improvement or at least positive indications. Reserpine<br />

is deceptive in this temporal context, <strong>and</strong> Wilkins <strong>and</strong> his<br />

coworkers were nearly deceived. In working with crude preparations<br />

of Rauwolfia serpentina, they were becoming very pessimistic,<br />

when a patient who had been treated for many consecutive<br />

days began to show symptomatic relief. Nevertheless, only after<br />

months of treatment did Wilkins become a believer in the drug’s<br />

beneficial effects.<br />

The Action of Reserpine<br />

Reserpine / 635<br />

When preparations of pure reserpine became available in 1952,<br />

the drug did not at first appear to be the active ingredient in the<br />

crude preparations. When patients’ heart rate <strong>and</strong> blood pressure<br />

began to drop after weeks of treatment, however, the investigators<br />

saw that reserpine was indeed responsible for the improvements.<br />

Once reserpine’s activity began, Wilkins observed a number of<br />

important <strong>and</strong> unique consequences. Both the crude preparations<br />

<strong>and</strong> pure reserpine significantly reduced the two most meaningful<br />

measures of blood pressure. These two measures are systolic blood<br />

pressure <strong>and</strong> diastolic blood pressure. Systolic pressure represents<br />

the peak of pressure produced in the arteries following a contraction<br />

of the heart. Diastolic pressure is the low point that occurs<br />

when the heart is resting. To lower the mean blood pressure in the<br />

system significantly, both of these pressures must be reduced. The<br />

administration of low doses of reserpine produced an average drop<br />

in pressure of about 15 percent, a figure that was considered less<br />

than dramatic but still highly significant. The complex phenomenon<br />

of blood pressure is determined by a multitude of factors, including<br />

the resistance of the arteries, the force of contraction of the<br />

heart, <strong>and</strong> the heartbeat rate. In addition to lowering the blood pressure,<br />

reserpine reduced the heartbeat rate by about 15 percent, providing<br />

an important auxiliary action.


636 / Reserpine<br />

In the early 1950’s, various therapeutic drugs were used to treat<br />

hypertension. Wilkins recognized that reserpine’s major contribution<br />

would be as a drug that could be used in combination with<br />

drugs that were already in use. His studies established that reserpine,<br />

combined with at least one of the drugs already in use, produced<br />

an additive effect in lowering blood pressure. Indeed, at<br />

times, the drug combinations produced a “synergistic effect,” which<br />

means that the combination of drugs created an effect that was more<br />

effective than the sum of the effects of the drugs when they were administered<br />

alone. Wilkins also discovered that reserpine was most<br />

effective when administered in low dosages. Increasing the dosage<br />

did not increase the drug’s effect significantly, but it did increase the<br />

likelihood of unwanted side effects. This fact meant that reserpine<br />

was indeed most effective when administered in low dosages along<br />

with other drugs.<br />

Wilkins believed that reserpine’s most unique effects were not<br />

those found directly in the cardiovascular system but those produced<br />

indirectly by the brain. Hypertension is often accompanied<br />

by neurotic anxiety, which is both a consequence of the justifiable<br />

fears of future negative health changes brought on by<br />

prolonged hypertension <strong>and</strong> contributory to the hypertension itself.<br />

Wilkins’s patients invariably felt better mentally, were less<br />

anxious, <strong>and</strong> were sedated, but in an unusual way. Reserpine<br />

made patients drowsy but did not generally cause sleep, <strong>and</strong> if<br />

sleep did occur, patients could be awakened easily. Such effects<br />

are now recognized as characteristic of tranquilizing drugs, or<br />

antipsychotics. In effect, Wilkins had discovered a new <strong>and</strong> important<br />

category of drugs: tranquilizers.<br />

Impact<br />

Reserpine holds a vital position in the historical development of<br />

antihypertensive drugs for two reasons. First, it was the first drug<br />

that was discovered to block activity in areas of the nervous system<br />

that use norepinephrine or its close relative dopamine as transmitter<br />

substances. Second, it was the first hypertension drug to be<br />

widely accepted <strong>and</strong> used. Its unusual combination of characteristics<br />

made it effective in most patients.


Since the 1950’s, medical science has rigorously examined cardiovascular<br />

functioning <strong>and</strong> diseases such as hypertension. Many<br />

new factors, such as diet <strong>and</strong> stress, have been recognized as factors<br />

in hypertension. Controlling diet <strong>and</strong> life-style help tremendously<br />

in treating hypertension, but if the nervous system could not be partially<br />

controlled, many cases of hypertension would continue to be<br />

problematic. Reserpine has made that control possible.<br />

See also Abortion pill; Antibacterial drugs; Artificial kidney;<br />

Birth control pill; Salvarsan.<br />

Further Reading<br />

Reserpine / 637<br />

MacGregor, G. A., <strong>and</strong> Norman M. Kaplan. Hypertension. 2ded.<br />

Abingdon: Health Press, 2001.<br />

“Reconsidering Reserpine.” American Family Physician 45 (March,<br />

1992).<br />

Weber, Michael A. Hypertension Medicine. Totowa, N.J.: Humana,<br />

2001.


638<br />

Rice <strong>and</strong> wheat strains<br />

Rice <strong>and</strong> wheat strains<br />

The invention: Artificially created high-yielding wheat <strong>and</strong> rice<br />

varieties that are helping food producers in developing countries<br />

keep pace with population growth<br />

The people behind the invention:<br />

Orville A. Vogel (1907-1991), an agronomist who developed<br />

high-yielding semidwarf winter wheats <strong>and</strong> equipment for<br />

wheat research<br />

Norman E. Borlaug (1914- ), a distinguished agricultural<br />

scientist<br />

Robert F. Ch<strong>and</strong>ler, Jr. (1907-1999), an international agricultural<br />

consultant <strong>and</strong> director of the International Rice Research<br />

Institute, 1959-1972<br />

William S. Gaud (1907-1977), a lawyer <strong>and</strong> the administrator of<br />

the U.S. Agency for International Development, 1966-1969<br />

The Problem of Hunger<br />

In the 1960’s, agricultural scientists created new, high-yielding<br />

strains of rice <strong>and</strong> wheat designed to fight hunger in developing<br />

countries. Although the introduction of these new grains raised levels<br />

of food production in poor countries, population growth <strong>and</strong><br />

other factors limited the success of the so-called “Green Revolution.”<br />

Before World War II, many countries of Asia, Africa, <strong>and</strong> Latin<br />

America exported grain to Western Europe. After the war, however,<br />

these countries began importing food, especially from the United<br />

States. By 1960, they were importing about nineteen million tons of<br />

grain a year; that level nearly doubled to thirty-six million tons in<br />

1966. Rapidly growing populations forced the largest developing<br />

countries—China, India, <strong>and</strong> Brazil in particular—to import huge<br />

amounts of grain. Famine was averted on the Indian subcontinent<br />

in 1966 <strong>and</strong> 1967 only by the United States shipping wheat to the region.<br />

The United States then changed its food policy. Instead of contributing<br />

food aid directly to hungry countries, the U.S. began


working to help such countries feed themselves.<br />

The new rice <strong>and</strong> wheat strains were introduced just as countries<br />

in Africa <strong>and</strong> Asia were gaining their independence from the European<br />

nations that had colonized them. The Cold War was still going<br />

strong, <strong>and</strong> Washington <strong>and</strong> other Western capitals feared that the<br />

Soviet Union was gaining influence in the emerging countries. To<br />

help counter this threat, the U.S. Agency for International Development<br />

(USAID) was active in the Third World in the 1960’s, directing<br />

or contributing to dozens of agricultural projects, including building<br />

rural infrastructure (farm-to-market roads, irrigation projects,<br />

<strong>and</strong> rural electric systems), introducing modern agricultural techniques,<br />

<strong>and</strong> importing fertilizer or constructing fertilizer factories in<br />

other countries. By raising the st<strong>and</strong>ard of living of impoverished<br />

people in developing countries through applying technology to agriculture,<br />

policymakers hoped to eliminate the socioeconomic conditions<br />

that would support communism.<br />

The Green Revolution<br />

Rice <strong>and</strong> wheat strains / 639<br />

It was against this background that William S. Gaud, administrator<br />

of USAID from 1966 to 1969, first talked about a “green revolution”<br />

in a 1968 speech before the Society for International Development<br />

in Washington, D.C. The term “green revolution” has<br />

been used to refer to both the scientific development of highyielding<br />

food crops <strong>and</strong> the broader socioeconomic changes in a<br />

country’s agricultural sector stemming from farmers’ adoption of<br />

these crops.<br />

In 1947, S. C. Salmon, a United States Department of Agriculture<br />

(USDA) scientist, brought a wheat-dwarfing gene to the United<br />

States. Developed in Japan, the gene produced wheat on a short<br />

stalk that was strong enough to bear a heavy head of grain. Orville<br />

Vogel, another USDA scientist, then introduced the gene into local<br />

wheat strains, creating a successful dwarf variety known as Gaines<br />

wheat. Under irrigation, Gaines wheat produced record yields. After<br />

hearing about Vogel’s work, Norman Borlaug, who headed<br />

the Rockefeller Foundation’s wheat-breeding program in Mexico,<br />

adapted Gaines wheat, later called “miracle wheat,” to a variety of<br />

growing conditions in Mexico.


640 / Rice <strong>and</strong> wheat strains<br />

Workers in an Asian rice field. (PhotoDisc)<br />

Success with the development of high-yielding wheat varieties<br />

persuaded the Rockefeller <strong>and</strong> Ford foundations to pursue similar<br />

ends in rice culture. The foundations funded the International Rice<br />

Research Institute (IRRI) in Los Banos, Philippines, appointing as director<br />

Robert F. Ch<strong>and</strong>ler, Jr., an international agricultural consultant.<br />

Under his leadership, IRRI researchers cross-bred Peta, a tall variety<br />

of rice from Indonesia, with Deo-geo-woo-gen, a dwarf rice from Taiwan,<br />

to produce a new strain, IR-8. Released in 1966 <strong>and</strong> dubbed<br />

“miracle rice,” IR-8 produced yields double those of other Asian rice<br />

varieties <strong>and</strong> in a shorter time, 120 days in contrast to 150 to 180 days.<br />

Statistics from India illustrate the expansion of the new grain varieties.<br />

During the 1966-1967 growing season, Indian farmers planted<br />

improved rice strains on 900,000 hectares, or 2.5 percent of the total<br />

area planted in rice. By 1984-1985, the surface area planted in improved<br />

rice varieties stood at 23.4 million hectares, or 56.9 percent of<br />

the total. The rate of adoption was even faster for wheat. In 1966-<br />

1967, improved varieties covered 500,000 hectares, comprising 4.2<br />

percent of the total wheat crop. By the 1984-1985 growing season,<br />

the surface area had exp<strong>and</strong>ed to 19.6 million hectares, or 82.9 percent<br />

of the total wheat crop.


To produce such high yields, IR-8 <strong>and</strong> other genetically engineered<br />

varieties of rice <strong>and</strong> wheat required the use of irrigation, fertilizers,<br />

<strong>and</strong> pesticides. Irrigation further increased food production<br />

by allowing year-round farming <strong>and</strong> the planting of multiple crops<br />

on the same plot of l<strong>and</strong>, either two crops of high-yielding grain varieties<br />

or one grain crop <strong>and</strong> another food crop.<br />

Expectations<br />

Rice <strong>and</strong> wheat strains / 641<br />

The rationale behind the introduction of high-yielding grains in<br />

developing countries was that it would start a cycle of improvement<br />

in the lives of the rural poor. High-yielding grains would lead to<br />

bigger harvests <strong>and</strong> better-nourished <strong>and</strong> healthier families. If better<br />

nutrition enabled more children to survive, the need to have large<br />

families to ensure care for elderly parents would ease. A higher survival<br />

rate of children would lead couples to use family planning,<br />

slowing overall population growth <strong>and</strong> allowing per capita food intake<br />

to rise.<br />

The greatest impact of the Green Revolution has been seen in<br />

Asia, which experienced dramatic increases in rice production, <strong>and</strong><br />

on the Indian subcontinent, with increases in rice <strong>and</strong> wheat yields.<br />

Latin America, especially Mexico, enjoyed increases in wheat harvests.<br />

Subsaharan Africa initially was left out of the revolution, as<br />

scientists paid scant attention to increasing the yields of such staple<br />

food crops as yams, cassava, millet, <strong>and</strong> sorghum. By the 1980’s,<br />

however, this situation was being remedied with new research directed<br />

toward millet <strong>and</strong> sorghum.<br />

Research is conducted by a network of international agricultural<br />

research centers. Backed by both public <strong>and</strong> private funds, these<br />

centers cooperate with international assistance agencies, private<br />

foundations, universities, multinational corporations, <strong>and</strong> government<br />

agencies to pursue <strong>and</strong> disseminate research into improved<br />

crop varieties to farmers in the Third World. IRRI <strong>and</strong> the International<br />

Maize <strong>and</strong> Wheat Improvement Center (IMMYT) in Mexico<br />

City are two of these agencies.


642 / Rice <strong>and</strong> wheat strains<br />

Impact<br />

Expectations went unrealized in the first few decades following<br />

the green revolution. Despite the higher yields from millions of<br />

tons of improved grain seeds imported into the developing world,<br />

lower-yielding grains still accounted for much of the surface area<br />

planted in grain. The reasons for this explain the limits <strong>and</strong> impact<br />

of the Green Revolution.<br />

The subsistence mentality dies hard. The main targets of Green<br />

Revolution programs were small farmers, people whose crops provide<br />

barely enough to feed their families <strong>and</strong> provide seed for the<br />

next crop. If an experimental grain failed, they faced starvation.<br />

Such farmers hedged their bets when faced with a new proposition,<br />

for example, by intercropping, alternating rows of different grains<br />

in the same field. In this way, even if one crop failed, another might<br />

feed the family.<br />

Poor farmers in developing countries also were likely to be illiterate<br />

<strong>and</strong> not eager to try something they did not fully underst<strong>and</strong>.<br />

Also, by definition, poor farmers often did not have the means to<br />

purchase the inputs—irrigation, fertilizer, <strong>and</strong> pesticides—required<br />

to grow the improved varieties.<br />

In many developing countries, therefore, rich farmers tended to be<br />

the innovators. More likely than poor farmers to be literate, they also<br />

had the money to exploit fully the improved grain varieties. They<br />

also were more likely than subsistence-level farmers to be in touch<br />

with the monetary economy, making purchases from the agricultural<br />

supply industry <strong>and</strong> arranging sales through established marketing<br />

channels, rather than producing primarily for personal or family use.<br />

Once wealthy farmers adopted the new grains, it often became<br />

more difficult for poor farmers to do so. Increased dem<strong>and</strong> for limited<br />

supplies, such as pesticides <strong>and</strong> fertilizers, raised costs, while<br />

bigger-than-usual harvests depressed market prices. With high sales<br />

volumes, owners of large farms could withst<strong>and</strong> the higher costs <strong>and</strong><br />

lower-per-unit profits, but smaller farmers often could not.<br />

Often, the result of adopting improved grains was that small<br />

farmers could no longer make ends meet solely by farming. Instead,<br />

they were forced to hire themselves out as laborers on large farms.<br />

Surges of laborers into a limited market depressed rural wages,


making it even more difficult for small farmers to eke out a living.<br />

The result was that rich farmers got richer <strong>and</strong> poor farmers got<br />

poorer. Often, small farmers who could no longer support their<br />

families would leave rural areas <strong>and</strong> migrate to the cities, seeking<br />

work <strong>and</strong> swelling the ranks of the urban poor.<br />

Mixed Results<br />

Orville A. Vogel<br />

Rice <strong>and</strong> wheat strains / 643<br />

Born in 1907, Orville Vogel grew up on a farm in eastern Nebraska,<br />

<strong>and</strong> farming remained his passion for his entire life. He<br />

earned bachelor’s <strong>and</strong> master’s degrees in agriculture from the<br />

University of Nebraska, <strong>and</strong> then a doctorate in agronomy from<br />

Washington State University (WSU) in 1939.<br />

Eastern Washington agreed with him, <strong>and</strong> he stayed there.<br />

He began his career as a wheat breeder 1931 for the U.S. Department<br />

of Agriculture, stationed at WSU. During the next fortytwo<br />

years, he also took on the responsibilities of associate<br />

agronomist for the university’s Division of Agronomy <strong>and</strong> from<br />

1960 until his retirement in 1973 was professor of agronomy.<br />

At heart Vogel was an experimenter <strong>and</strong> tinkerer, renowned<br />

among his peers for his keen powers of observation <strong>and</strong> his unselfishness.<br />

In addition to the wheat strains he bred that helped<br />

launch the Green Revolution, he took part in the search for<br />

plant varieties resistant to snow mold <strong>and</strong> foot rot. However,<br />

according to the father of the Green Revolution, Nobel laureate<br />

Norman Borlaug, Vogel’s greatest contribution may not have<br />

been semi-dwarf wheat varieties but the many innovations in<br />

farming equipment he built as a sideline. These unheralded inventions<br />

automated the planting <strong>and</strong> harvesting of research<br />

plots, <strong>and</strong> so made research much easier to carry out <strong>and</strong> faster.<br />

In recognition of his achievements, Vogel received the U.S.<br />

National Medal of Science in 1975 <strong>and</strong> entered the Agricultural<br />

Research Service’s Science Hall of Fame in 1987. Vogel died in<br />

Washington in 1991.<br />

The effects of the Green Revolution were thus mixed. The dissemination<br />

of improved grain varieties unquestionably increased<br />

grain harvests in some of the poorest countries of the world. Seed


644 / Rice <strong>and</strong> wheat strains<br />

companies developed, produced, <strong>and</strong> sold commercial quantities of<br />

improved grains, <strong>and</strong> fertilizer <strong>and</strong> pesticide manufacturers logged<br />

sales to developing countries thanks to USAID-sponsored projects.<br />

Along with disrupting the rural social structure <strong>and</strong> encouraging<br />

rural flight to the cities, the Green Revolution has had other negative<br />

effects. For example, the millions of tube wells sunk in India to<br />

irrigate crops reduced groundwater levels in some regions faster<br />

than they could be recharged. In other areas, excessive use of pesticides<br />

created health hazards, <strong>and</strong> fertilizer use led to streams <strong>and</strong><br />

ponds being clogged by weeds. The scientific community became<br />

concerned that the use of improved varieties of grain, many of<br />

which were developed from the same mother variety, reduced the<br />

genetic diversity of the world’s food crops, making them especially<br />

vulnerable to attack by disease or pests.<br />

Perhaps the most significant impact of the Green Revolution is<br />

the change it wrought in the income <strong>and</strong> class structure of rural areas;<br />

often, malnutrition was not eliminated in either the countryside<br />

or the cities. Almost without exception, the relative position of peasants<br />

deteriorated. Many analysts admit that the Green Revolution<br />

did not end world hunger, but they argue that it did buy time. The<br />

poorest of the poor would be even worse off without it.<br />

See also Artificial chromosome; Cloning; Genetic “fingerprinting”;<br />

Genetically engineered insulin; In vitro plant culture.<br />

Further Reading<br />

Glaeser, Bernhard, ed. The Green Revolution Revisited: Critique <strong>and</strong> Alternatives.<br />

London: Allen & Unwin, 1987.<br />

Hayami, Yujiro, <strong>and</strong> Masao Kikuchi. A Rice Village Saga: Three Decades<br />

of Green Revolution in the Philippines. Lanham, Md.: Barnes<br />

<strong>and</strong> Noble, 2000.<br />

Karim, M. Bazlul. The Green Revolution: An International Bibliography.<br />

New York: Greenwood Press, 1986.<br />

Lipton, Michael, <strong>and</strong> Richard Longhurst. New Seeds <strong>and</strong> Poor People.<br />

Baltimore: Johns Hopkins University Press, 1989.<br />

Perkins, John H. Geopolitics <strong>and</strong> the Green Revolution: Wheat, Genes,<br />

<strong>and</strong> the Cold War. New York: Oxford University Press, 1997.


Richter scale<br />

Richter scale<br />

The invention: A scale for measuring the strength of earthquakes<br />

based on their seismograph recordings.<br />

The people behind the invention:<br />

Charles F. Richter (1900-1985), an American seismologist<br />

Beno Gutenberg (1889-1960), a German American seismologist<br />

Kiyoo Wadati (1902- ), a pioneering Japanese seismologist<br />

Giuseppe Mercalli (1850-1914), an Italian physicist,<br />

volcanologist, <strong>and</strong> meteorologist<br />

Earthquake Study by Eyewitness Report<br />

645<br />

Earthquakes range in strength from barely detectable tremors to<br />

catastrophes that devastate large regions <strong>and</strong> take hundreds of thous<strong>and</strong>s<br />

of lives. Yet the human impact of earthquakes is not an accurate<br />

measure of their power; minor earthquakes in heavily populated regions<br />

may cause great destruction, whereas powerful earthquakes in<br />

remote areas may go unnoticed. To study earthquakes, it is essential<br />

to have an accurate means of measuring their power.<br />

The first attempt to measure the power of earthquakes was the<br />

development of intensity scales, which relied on damage effects<br />

<strong>and</strong> reports by witnesses to measure the force of vibration. The<br />

first such scale was devised by geologists Michele Stefano de Rossi<br />

<strong>and</strong> François-Alphonse Forel in 1883. It ranked earthquakes on a<br />

scale of 1 to 10. The de Rossi-Forel scale proved to have two serious<br />

limitations: Its level 10 encompassed a great range of effects, <strong>and</strong> its<br />

description of effects on human-made <strong>and</strong> natural objects was so specifically<br />

European that it was difficult to apply the scale elsewhere.<br />

To remedy these problems, Giuseppe Mercalli published a revised<br />

intensity scale in 1902. The Mercalli scale, as it came to be<br />

called, added two levels to the high end of the de Rossi-Forel scale,<br />

making its highest level 12. It also was rewritten to make it more<br />

globally applicable. With later modifications by Charles F. Richter,<br />

the Mercalli scale is still in use.<br />

Intensity measurements, even though they are somewhat subjec-


646 / Richter scale<br />

Charles F. Richter<br />

Charles Francis Richter was born in Ohio in 1900. After his<br />

mother divorced his father, she moved the family to Los Angles<br />

in 1909. A precocious student, Richter entered the University of<br />

Southern California at sixteen <strong>and</strong> transferred to Stanford University<br />

a year later, majoring in physics. He graduated in 1920<br />

<strong>and</strong> finished a doctorate in theoretical physics at the California<br />

Institute of Technology in 1928.<br />

While Richter was a graduate student at Caltech, Noble laureate<br />

Robert A. Millikan lured him away from his original interest,<br />

astronomy, to become an assistant at the seismology laboratory.<br />

Richter realized that seismology was then a relatively new<br />

discipline <strong>and</strong> that he could help it mature. He stayed with it—<br />

<strong>and</strong> Caltech—for the rest of his university career, retiring as<br />

professor emeritus in 1970. In 1971 he opened a consulting<br />

firm—Lindvall, Richter <strong>and</strong> Associates—to assess the earthquake<br />

readiness of structures.<br />

Richter published more than two hundred articles about<br />

earthquakes <strong>and</strong> earthquake engineering <strong>and</strong> two influential<br />

books, Elementary Seismology <strong>and</strong> Seismicity of the Earth (with<br />

Beno Gutenberg). These works, together with his teaching,<br />

trained a generation of earthquake researchers <strong>and</strong> gave them a<br />

basic tool, the Richter scale, to work with. He died in California<br />

in 1985.<br />

tive, are very useful in mapping the extent of earthquake effects.<br />

Nevertheless, intensity measurements are still not ideal measuring<br />

techniques. Intensity varies from place to place <strong>and</strong> is strongly influenced<br />

by geologic features, <strong>and</strong> different observers frequently report<br />

different intensities. There is a need for an objective method of<br />

describing the strength of earthquakes with a single measurement.<br />

Measuring Earthquakes One Hundred Kilometers Away<br />

An objective technique for determining the power of earthquakes<br />

was devised in the early 1930’s by Richter at the California Institute<br />

of Technology in Pasadena, California. The eventual usefulness of<br />

the scale that came to be called the “Richter scale” was completely<br />

unforeseen at first.


Amplified Maximum Ground Motion (Microns)<br />

9<br />

10<br />

8<br />

10<br />

Richter scale / 647<br />

Alaska, 1964<br />

San Francisco, 1906<br />

Great devastation;<br />

many fatalities possible<br />

In 1931, the California Institute of Technology was preparing to<br />

issue a catalog of all earthquakes detected by its seismographs in the<br />

preceding three years. Several hundred earthquakes were listed,<br />

most of which had not been felt by humans, but detected only by instruments.<br />

Richter was concerned about the possible misinterpretations<br />

of the listing. With no indication of the strength of the earthquakes,<br />

the public might overestimate the risk of earthquakes in<br />

areas where seismographs were numerous <strong>and</strong> underestimate the<br />

risk in areas where seismographs were few.<br />

To remedy the lack of a measuring method, Richter devised the<br />

scale that now bears his name. On this scale, earthquake force is expressed<br />

in magnitudes, which in turn are expressed in whole numbers<br />

<strong>and</strong> decimals. Each increase of one magnitude indicates a tenfold jump<br />

in the earthquake’s force. These measurements were defined for a<br />

st<strong>and</strong>ard seismograph located one hundred kilometers from the earthquake.<br />

By comparing records for earthquakes recorded on different<br />

8<br />

8.9<br />

Great<br />

Major<br />

New Madrid, Missouri, 1812<br />

7<br />

10<br />

7<br />

6 10<br />

5<br />

10<br />

4<br />

10<br />

2<br />

10<br />

1<br />

10 -1 0<br />

3<br />

1<br />

2<br />

Not felt<br />

Strong<br />

6<br />

Moderate<br />

5<br />

4 Small<br />

Minor<br />

Damage begins;<br />

fatalities rare<br />

-1 0 1 2 3 4 5 6 7 8 9<br />

Magnitude<br />

Graphic representation of the Richter scale showing examples of historically important<br />

earthquakes.


648 / Richter scale<br />

devices at different distances, Richter was able to create conversion tables<br />

for measuring magnitudes for any instrument at any distance.<br />

Impact<br />

Richter had hoped to create a rough means of separating small,<br />

medium, <strong>and</strong> large earthquakes, but he found that the scale was capable<br />

of making much finer distinctions. Most magnitude estimates<br />

made with a variety of instruments at various distances from earthquakes<br />

agreed to within a few tenths of a magnitude. Richter formally<br />

published a description of his scale in January, 1935, in the<br />

Bulletin of the Seismological Society of America. Other systems of estimating<br />

magnitude had been attempted, notably that of Kiyoo Wadati,<br />

published in 1931, but Richter’s system proved to be the most workable<br />

scale yet devised <strong>and</strong> rapidly became the st<strong>and</strong>ard.<br />

Over the next few years, the scale was refined. One critical refinement<br />

was in the way seismic recordings were converted into magnitude.<br />

Earthquakes produce many types of waves, but it was not<br />

known which type should be the st<strong>and</strong>ard for magnitude. So-called<br />

surface waves travel along the surface of the earth. It is these waves<br />

that produce most of the damage in large earthquakes; therefore, it<br />

seemed logical to let these waves be the st<strong>and</strong>ard. Earthquakes deep<br />

within the earth, however, produce few surface waves. Magnitudes<br />

based on surface waves would therefore be too small for these earthquakes.<br />

Deep earthquakes produce mostly waves that travel through<br />

the solid body of the earth; these are the so-called body waves.<br />

It became apparent that two scales were needed: one based on<br />

surface waves <strong>and</strong> one on body waves. Richter <strong>and</strong> his colleague<br />

Beno Gutenberg developed scales for the two different types of<br />

waves, which are still in use. Magnitudes estimated from surface<br />

waves are symbolized by a capital M, <strong>and</strong> those based on body<br />

waves are denoted by a lowercase m.<br />

From a knowledge of Earth movements associated with seismic<br />

waves, Richter <strong>and</strong> Gutenberg succeeded in defining the energy<br />

output of an earthquake in measurements of magnitude. A magnitude<br />

6 earthquake releases about as much energy as a one-megaton<br />

nuclear explosion; a magnitude 0 earthquake releases about as<br />

much energy as a small car dropped off a two-story building.


See also Carbon dating; Geiger counter; Gyrocompass; Sonar;<br />

Scanning tunneling microscope.<br />

Further Reading<br />

Richter scale / 649<br />

Bates, Charles C., Thomas Frohock Gaskell, <strong>and</strong> Robert B. Rice. Geophysics<br />

in the Affairs of Man: A Personalized History of Exploration<br />

Geophysics <strong>and</strong> Its Allied Sciences of Seismology <strong>and</strong> Oceanography.<br />

New York: Pergamon Press, 1982.<br />

Davison, Charles. 1927. Reprint. The Founders of Seismology. New<br />

York: Arno Press, 1978.<br />

Howell, Benjamin F. An Introduction to Seismological Research: History<br />

<strong>and</strong> Development. Cambridge, N.Y.: Cambridge University Press,<br />

1990.


650<br />

Robot (household)<br />

Robot (household)<br />

The invention: The first available personal robot, Hero 1 could<br />

speak; carry small objects in a gripping arm, <strong>and</strong> sense light, motion,<br />

sound, <strong>and</strong> time.<br />

The people behind the invention:<br />

Karel Capek (1890-1938), a Czech playwright<br />

The Health Company, an American electronics manufacturer<br />

Personal Robots<br />

In 1920, the Czech playwright Karel Capek introduced the term<br />

robot, which he used to refer to intelligent, humanoid automatons<br />

that were subservient to humans. Robots such as those described<br />

by Capek have not yet been developed; their closest counterparts<br />

are the nonintelligent automatons used by industry <strong>and</strong> by private<br />

individuals. Most industrial robots are heavy-duty, immobile machines<br />

designed to replace humans in routine, undesirable, monotonous<br />

jobs. Most often, they use programmed gripping arms to<br />

carry out tasks such as spray painting cars, assembling watches,<br />

<strong>and</strong> shearing sheep.<br />

Modern personal robots are smaller, more mobile, less expensive<br />

models that serve mostly as toys or teaching tools. In some<br />

cases, they can be programmed to carry out activities such as walking<br />

dogs or serving mixed drinks. Usually, however, it takes more<br />

effort to program a robot to perform such activities than it does to<br />

do them oneself.<br />

The Hero 1, which was first manufactured by the Heath Company<br />

in 1982, has been a very popular personal robot. Conceived<br />

as a toy <strong>and</strong> a teaching tool, the Hero 1 can be programmed<br />

to speak; to sense light, sound, motion, <strong>and</strong> time; <strong>and</strong><br />

to carry small objects. The Hero 1 <strong>and</strong> other personal robots are<br />

often viewed as tools that will someday make it possible to produce<br />

intelligent robots.


Hero 1 Operation<br />

Robot (household) / 651<br />

The concept of artificial beings serving humanity has existed<br />

since antiquity (for example, it is found in Greek mythology). Such<br />

devices, which are now called robots, were first actualized, in a<br />

simple form, in the 1960’s. Then, in the mid-1970’s, the manufacture<br />

of personal robots began. One of the first personal robots was<br />

the Turtle, which was made by the Terrapin Company of Cambridge,<br />

Massachusetts. The Turtle was a toy that entertained owners<br />

via remote control, programmable motion, a beeper, <strong>and</strong> blinking<br />

displays. The Turtle was controlled by a computer to which it<br />

was linked by a cable.<br />

Among the first significant personal robots was the Hero 1. This<br />

robot, which was usually sold in the form of a $1,000 kit that had to<br />

be assembled, is a squat, thirty-nine-pound mobile unit containing a<br />

head, a body, <strong>and</strong> a base. The head contains control boards, sensors,<br />

<strong>and</strong> a manipulator arm. The body houses control boards <strong>and</strong> related<br />

electronics, while the base contains a three-wheel-drive unit that<br />

renders the robot mobile.<br />

The Heath Company, which produced the Hero 1, viewed it as<br />

providing entertainment for <strong>and</strong> teaching people who are interested<br />

in robot applications. To facilitate these uses, the following<br />

abilities were incorporated into the Hero 1: independent operation<br />

via rechargeable batteries; motion- <strong>and</strong> distance/position-sensing<br />

capability; light, sound, <strong>and</strong> language use/recognition; a manipulator<br />

arm to carry out simple tasks; <strong>and</strong> easy programmability.<br />

The Hero 1 is powered by four rechargeable batteries arranged as<br />

two 12-volt power supplies. Recharging is accomplished by means<br />

of a recharging box that is plugged into a home outlet. It takes six to<br />

eight hours to recharge depleted batteries, <strong>and</strong> complete charging is<br />

signaled by an indicator light. In the functioning robot, the power<br />

supplies provide 5-volt <strong>and</strong> 12-volt outputs to logic <strong>and</strong> motor circuits,<br />

respectively.<br />

The Hero 1 moves by means of a drive mechanism in its base. The<br />

mechanism contains three wheels, two of which are unpowered<br />

drones. The third wheel, which is powered for forward <strong>and</strong> reverse<br />

motion, is connected to a stepper motor that makes possible directional<br />

steering. Also included in the powered wheel is a metal disk


652 / Robot (household)<br />

with spaced reflective slots that helps Hero 1 to identify its position.<br />

As the robot moves, light is used to count the slots, <strong>and</strong> the slot<br />

count is used to measure the distance the robot has traveled, <strong>and</strong><br />

therefore its position.<br />

The robot’s “senses,” located in its head, consist of sound, light,<br />

<strong>and</strong> motion detectors as well as a phoneme synthesizer (phonemes<br />

are sounds, or units of speech). All these components are connected<br />

with the computer. The Hero 1 can detect sounds between 200 <strong>and</strong><br />

5,000 hertz. Its motion sensor detects all movement within a 15-foot<br />

radius. The phoneme synthesizer is capable of producing most<br />

words by using combinations of 64 phonemes. In addition, the robot<br />

keeps track of time by using an internal clock/calendar.<br />

The Hero 1 can carry out various tasks by using a gripper that<br />

serves as a h<strong>and</strong>. The arm on which the gripper is located is connected<br />

to the back of the robot’s head. The head (<strong>and</strong>, therefore, the<br />

arm) can rotate 350 degrees horizontally. In addition, the arm contains<br />

a shoulder motor that allows it to rise or drop 150 degrees vertically,<br />

<strong>and</strong> its forearm can be either extended or retracted. Finally, a<br />

wrist motor allows the gripper’s tip to rotate by 350 degrees, <strong>and</strong> the<br />

two-fingered gripper can open up to a maximum width of 3.5<br />

inches. The arm is not useful except as an educational tool, since its<br />

load-bearing capacity is only about a pound <strong>and</strong> its gripper can exert<br />

a force of only 6 ounces.<br />

The computational capabilities of the robot are much more impressive<br />

than its physical capabilities. Programming is accomplished<br />

by means of a simple keypad located on the robot’s head, which<br />

provides an inexpensive, easy-to-use method of operator-computer<br />

communication. To make things simpler for users who want entertainment<br />

without having to learn robotics, a manual mode is included<br />

for programming. In the manual mode, a h<strong>and</strong>-held teaching<br />

pendant is connected to Hero 1 <strong>and</strong> used to program all the<br />

motion capabilities of the robot. The programming of sensory <strong>and</strong><br />

language abilities, however, must be accomplished by using the<br />

keyboard. Using the keyboard <strong>and</strong> the various options that are<br />

available enables Hero 1 owners to program the robot to perform<br />

many interesting activities.


Consequences<br />

The Hero 1 had a huge impact on robotics; thous<strong>and</strong>s of people<br />

purchased it <strong>and</strong> used it for entertainment, study, <strong>and</strong> robot design.<br />

The Heath Company itself learned from the Hero 1 <strong>and</strong> later introduced<br />

an improved version: Heathkit 2000. This personal robot,<br />

which costs between $2,000 <strong>and</strong> $4,500, has ten times the capabilities<br />

of Hero 1, operates via radio-controlled keyboard, contains a<br />

voice synthesizer that can be programmed in any language, <strong>and</strong><br />

plugs itself in for recharging.<br />

Other companies, including the Androbot Company in California,<br />

have manufactured personal robots that sell for up to $10,000.<br />

One such robot is the Androbot BOB (brains on board). It can guard<br />

a home, call the police, walk at 2.5 kilometers per hour, <strong>and</strong> sing.<br />

Androbot has also designed Topo, a personal robot that can serve<br />

drinks. Still other robots can sort laundry <strong>and</strong>/or vacuum-clean<br />

houses. Although modern robots lack intelligence <strong>and</strong> merely have<br />

the ability to move when they are directed to by a program or by remote<br />

control, there is no doubt that intelligent robots will be developed<br />

in the future.<br />

See also Electric refrigerator; Microwave cooking; Robot (industrial);<br />

Vacuum cleaner; Washing machine.<br />

Further Reading<br />

Robot (household) / 653<br />

Aleks<strong>and</strong>er, Igor, <strong>and</strong> Piers Burnett. Reinventing Man: The Robot Becomes<br />

Reality. London: Kogan Page, 1983.<br />

Asimov, Isaac. Robots: Machines in Man’s Image. New York: Harmony<br />

Books, 1985.<br />

Bell, Trudy E. “Robots in the Home: Promises, Promises.” IEEE Spectrum<br />

22, no. 5 (May, 1985).<br />

Whalen, Bernie. “Upscale Consumers Adopt Home Robots, but<br />

Widespread Lifestyle Impact Is Years Away.” Marketing News 17,<br />

no. 24 (November 25, 1983).


654<br />

Robot (industrial)<br />

Robot (industrial)<br />

The invention: The first industrial robots, Unimates were designed to<br />

replace humans in undesirable, hazardous, <strong>and</strong> monotonous jobs.<br />

The people behind the invention:<br />

Karel Capek (1890-1938), a Czech playwright<br />

George C. Devol, Jr. (1912- ), an American inventor<br />

Joseph F. Engelberger (1925- ), an American entrepreneur<br />

Robots, from Concept to Reality<br />

The 1920 play Rossum’s Universal Robots, by Czech writer Karel<br />

Capek, introduced robots to the world. Capek’s humanoid robots—<br />

robot, a word created by Capek, essentially means slave—revolted<br />

<strong>and</strong> took over the world, which made the concept of robots somewhat<br />

frightening. The development of robots, which are now defined<br />

as machines that do work that would ordinarily be carried out<br />

by humans, has not yet advanced to the stage of being able to produce<br />

humanoid robots, however, much less robots capable of carrying<br />

out a revolt.<br />

Most modern robots are found in industry, where they perform<br />

dangerous or monotonous tasks that previously were done by humans.<br />

The first industrial robots were the Unimates (short for “universal<br />

automaton”), which were derived from a robot design invented<br />

by George C. Devol <strong>and</strong> patented in 1954. The first Unimate<br />

prototypes, developed by Devol <strong>and</strong> Joseph F. Engelberger, were<br />

completed in 1962 by Unimation Incorporated <strong>and</strong> tested in industry.<br />

They were so successful that the company, located in Danbury,<br />

Connecticut, manufactured <strong>and</strong> sold thous<strong>and</strong>s of Unimates to<br />

companies in the United States <strong>and</strong> abroad. Unimates are very versatile<br />

at performing routine industrial tasks <strong>and</strong> are easy to program<br />

<strong>and</strong> reprogram. The tasks they perform include various steps in automobile<br />

manufacturing, spray painting, <strong>and</strong> running lathes. The<br />

huge success of the Unimates led companies in other countries to<br />

produce their own industrial robots, <strong>and</strong> advancing technology has<br />

improved all industrial robots tremendously.


A New Industrial Revolution<br />

Robot (industrial) / 655<br />

Each of the first Unimate robots, which were priced at $25,000,<br />

was almost five feet tall <strong>and</strong> stood on a four-foot by five-foot base. It<br />

has often been said that a Unimate resembles the gun turret of a<br />

minitank, set atop a rectangular box. In operation, such a robot will<br />

swivel, swing, <strong>and</strong>/or dip <strong>and</strong> turn at the wrist of its hydraulically<br />

powered arm, which has a steel h<strong>and</strong>. The precisely articulated<br />

h<strong>and</strong> can pick up an egg without breaking it. At the same time, however,<br />

it is powerful enough to lift a hundred-pound weight.<br />

The Unimate is a robotic jack of all trades: It can be programmed,<br />

in about an hour, to carry out a complex operation, after which it can<br />

have its memory erased <strong>and</strong> be reprogrammed in another hour to<br />

do something entirely different. In addition, programming a Unimate<br />

requires no special training. The programmer simply uses a teachcable<br />

selector that allows the programmer to move the Unimate arm<br />

through the desired operation. This selector consists of a group of<br />

pushbutton control boxes, each of which is equipped with buttons<br />

in opposed pairs. Each button pair records the motion that will put a<br />

Unimate arm through one of five possible motions, in opposite directions.<br />

For example, pushing the correct buttons will record a motion<br />

in which the robot’s arm moves out to one side, aims upward,<br />

<strong>and</strong> angles appropriately to carry out the first portion of its intended<br />

job. If the Unimate overshoots, undershoots, or otherwise<br />

performs the function incorrectly, the activity can be fine-tuned<br />

with the buttons.<br />

Once the desired action has been performed correctly, pressing a<br />

“record” button on the robot’s main control panel enters the operation<br />

into its computer memory. In this fashion, Unimates can be programmed<br />

to carry out complex actions that require as many as two<br />

hundred comm<strong>and</strong>s. Each comm<strong>and</strong> tells the Unimate to move its<br />

arm or h<strong>and</strong> in a given way by combining the following five motions:<br />

sliding the arm forward, swinging the arm horizontally, tilting<br />

the arm up or down, bending the wrist up or down, <strong>and</strong> swiveling<br />

the h<strong>and</strong> in a half-circle clockwise or counterclockwise.<br />

Before pressing the “record” button on the Unimate’s control<br />

panel, the operator can also comm<strong>and</strong> the h<strong>and</strong> to grasp an item<br />

when in a particular position. Furthermore, the strength of the


656 / Robot (industrial)<br />

grasp can be controlled, as can the duration of time between each action.<br />

Finally, the Unimate can be instructed to start or stop another<br />

routine (such as operating a paint sprayer) at any point. Once the instructor<br />

is satisfied with the robot’s performance, pressing a “repeat<br />

continuous” control starts the Unimate working. The robot will stop<br />

repeating its program only when it is turned off.<br />

Inside the base of an original Unimate is a magnetic drum that<br />

contains its memory. The drum turns intermittently, moving each of<br />

two hundred long strips of metal beneath recording heads. This<br />

strip movement brings specific portions of each strip—dictated by<br />

particular motions—into position below the heads. When the “record”<br />

button is pressed after a motion is completed, the h<strong>and</strong> position<br />

is recorded as a series of numbers that tells the computer the<br />

complete h<strong>and</strong> position in each of the five permissible movement<br />

modes.<br />

Once “repeat continuous” is pressed, the computer begins the<br />

comm<strong>and</strong> series by turning the drum appropriately, carrying out<br />

each memorized comm<strong>and</strong> in the chosen sequence. When the sequence<br />

ends, the computer begins again, <strong>and</strong> the process repeats<br />

until the robot is turned off. If a Unimate user wishes to change the<br />

function of such a robot, its drum can be erased <strong>and</strong> reprogrammed.<br />

Users can also remove programmed drums, store them for future<br />

use, <strong>and</strong> replace them with new drums.<br />

Consequences<br />

The first Unimates had a huge impact on industrial manufacturing.<br />

In time, different sizes of robots became available so that additional<br />

tasks could be performed, <strong>and</strong> the robots’ circuitry was improved.<br />

Because they have no eyes <strong>and</strong> cannot make judgments,<br />

Unimates are limited to relatively simple tasks that are coordinated<br />

by means of timed operations <strong>and</strong> simple computer interactions.<br />

Most of the thous<strong>and</strong>s of modern Unimates <strong>and</strong> their multinational<br />

cousins in industry are very similar to the original Unimates<br />

in terms of general capabilities, although they can now assemble<br />

watches <strong>and</strong> perform other delicate tasks that the original Unimates<br />

could not perform. The crude magnetic drums <strong>and</strong> computer controls<br />

have given way to silicon chips <strong>and</strong> microcomputers, which


have made the robots more accurate <strong>and</strong> reliable. Some robots can<br />

even build other robots, <strong>and</strong> others can perform tasks such as mowing<br />

lawns <strong>and</strong> walking dogs.<br />

Various improvements have been planned that will ultimately<br />

lead to some very interesting <strong>and</strong> advanced modifications. It is<br />

likely that highly sophisticated humanoid robots like those predicted<br />

by Karel Capek will be produced at some future time. One<br />

can only hope that these robots will not rebel against their human<br />

creators.<br />

See also CAD/CAM; Robot (household); SAINT; Virtual machine.<br />

Further Reading<br />

Robot (industrial) / 657<br />

Aleks<strong>and</strong>er, Igor, <strong>and</strong> Piers Burnett. Reinventing Man: The Robot Becomes<br />

Reality. London: Kogan Page, 1983.<br />

Asimov, Isaac. Robots: Machines in Man’s Image. New York: Harmony<br />

Books, 1985.<br />

Chakravarty, Subrata N. “Springtime for an Ugly Duckling.” Forbes<br />

127, no. 9 (April, 1981).<br />

Hartley, J. “Robots Attack the Quiet World of Arc Welding.” Engineer<br />

246, no. 6376 (June, 1978).<br />

Lamb, W. G. Unimates at Work. Edited by C. W. Burckhardt. Basel,<br />

Switzerl<strong>and</strong>: Birkhauser Verlag, 1975.<br />

Tuttle, Howard C. “Robots’ Contribution: Faster Cycles, Better<br />

Quality.” Production 88, no. 5 (November, 1981).


658<br />

Rocket<br />

Rocket<br />

The invention: Liquid-fueled rockets developed by Robert H. Goddard<br />

made possible all later developments in modern rocketry,<br />

which in turn has made the exploration of space practical.<br />

The person behind the invention:<br />

Robert H. Goddard (1882-1945), an American physics professor<br />

History in a Cabbage Patch<br />

Just as the age of air travel began on an out-of-the-way shoreline<br />

at Kitty Hawk, North Carolina, with the Wright brothers’ airplane<br />

in 1903, so too the seemingly impossible dream of spaceflight<br />

began in a cabbage patch in Auburn, Massachusetts, with<br />

Robert H. Goddard’s launch of a liquid-fueled rocket on March 16,<br />

1926. On that clear, cold day, with snow still on the ground, Goddard<br />

launched a three-meter-long rocket using liquid oxygen <strong>and</strong><br />

gasoline. The flight lasted only about two <strong>and</strong> one-half seconds,<br />

during which the rocket rose 12 meters <strong>and</strong> l<strong>and</strong>ed about 56 meters<br />

away.<br />

Although the launch was successful, the rocket’s design was<br />

clumsy. At first, Goddard had thought that a rocket would be<br />

steadier if the motor <strong>and</strong> nozzles were ahead of the fuel tanks,<br />

rather like a horse <strong>and</strong> buggy. After this first launch, it was clear<br />

that the motor needed to be placed at the rear of the rocket. Although<br />

Goddard had spent several years working on different<br />

pumps to control the flow of fuel to the motor, the first rocket had<br />

no pumps or electrical system. Henry Sacks, a Clark University<br />

machinist, launched the rocket by turning a valve, placing an alcohol<br />

stove beneath the motor, <strong>and</strong> dashing for safety. Goddard <strong>and</strong><br />

his coworker Percy Roope watched the launch from behind an iron<br />

wall.<br />

Despite its humble setting, this simple event changed the course<br />

of history. Many people saw in Goddard’s launch the possibilities<br />

for high-altitude research, space travel, <strong>and</strong> modern weaponry. Although<br />

Goddard invented <strong>and</strong> experimented mostly in private,


others in the United States, the Soviet Union, <strong>and</strong> Germany quickly<br />

followed in his footsteps. The V-2 rockets used by Nazi Germany<br />

in World War II (1939-1945) included many of Goddard’s designs<br />

<strong>and</strong> ideas.<br />

A Lifelong Interest<br />

Rocket / 659<br />

Goddard’s success was no accident. He had first become interested<br />

in rockets <strong>and</strong> space travel when he was seventeen, no doubt<br />

because of reading books such as H. G. Wells’s The War of the Worlds<br />

(1898) <strong>and</strong> Garrett P. Serviss’s Edison’s Conquest of Mars (1898). In<br />

1907, he sent to several scientific journals a paper describing his ideas<br />

about traveling through a near vacuum. Although the essay was rejected,<br />

Goddard began thinking about liquid fuels in 1909. After finishing<br />

his doctorate in physics at Clark University <strong>and</strong> postdoctoral<br />

studies at Princeton University, he began to experiment.<br />

One of the things that made Goddard so successful was his ability<br />

to combine things he had learned from chemistry, physics, <strong>and</strong><br />

engineering into rocket design. More than anyone else at the time,<br />

Goddard had the ability to combine ideas with practice.<br />

Goddard was convinced that the key for moving about in space<br />

was the English physicist <strong>and</strong> mathematician Sir Isaac Newton’s<br />

third law of motion (for every action there is an equal <strong>and</strong> opposite<br />

reaction). To prove this, he showed that a gun recoiled when it was<br />

fired in a vacuum. During World War I (1914-1918), Goddard<br />

moved to the Mount Wilson Observatory in California, where he<br />

investigated the use of black powder <strong>and</strong> smokeless powder as<br />

rocket fuel. Goddard’s work led to the invention of the bazooka, a<br />

weapon that was much used during World War II, as well as bombardment<br />

<strong>and</strong> antiaircraft rockets.<br />

After World War I, Goddard returned to Clark University. By<br />

1920, mostly because of the experiments he had done during the<br />

war, he had decided that a liquid-fuel motor, with its smooth thrust,<br />

had the best chance of boosting a rocket into space. The most powerful<br />

fuel was hydrogen, but it is very difficult to h<strong>and</strong>le. Oxygen had<br />

many advantages, but it was hard to find <strong>and</strong> extremely dangerous,<br />

since it boils at −148 degrees Celsius <strong>and</strong> explodes when it comes in<br />

contact with oils, greases, <strong>and</strong> flames. Other possible fuels were pro-


(Library of Congress)<br />

660 / Rocket<br />

Robert H. Goddard<br />

In 1920 The New York Times made fun of Robert Hutchings<br />

Goddard (1882-1945) for claiming that rockets could travel<br />

through outer space to the Moon. It was impossible, the newspaper’s<br />

editorial writer confidently asserted, because in outer<br />

space the engine would have no air to push against <strong>and</strong> so<br />

could not move the rocket. A sensitive, quiet man, the Clark<br />

University physics professor was stung by the public rebuke,<br />

all the more so because it displayed ignorance of<br />

basic physics. “Every vision is a joke,” Goddard<br />

said, somewhat bitterly, “until the first man accomplishes<br />

it.”<br />

Goddard had already proved that a rocket could<br />

move in a vacuum, but he refrained from rebutting<br />

the Times article. In 1919 he had become the first<br />

American to describe mathematically the theory of<br />

rocket propulsion in his classic article “A Method of<br />

Reaching Extreme Altitude,” <strong>and</strong> during World War I<br />

he had acquired experience designing solid-fuel rockets.<br />

However, even though he was the world’s leading<br />

expert on rocketry, he decided to seek privacy for<br />

his experiments. His successful launch of a liquidfuel<br />

rocket in 1926, followed by new designs that reached ever<br />

higher altitudes, was a source of satisfaction, as were his 214<br />

patents, but real recognition of his achievements did not come<br />

his way until World War II. In 1942 he was named director of research<br />

at the U.S. Navy’s Bureau of Aeronautics, for which he<br />

worked on jet-assisted takeoff rockets <strong>and</strong> variable-thrust liquid-propellant<br />

rockets. In 1943 the Curtiss-Wright Corporation<br />

hired him as a consulting engineer, <strong>and</strong> in 1945 he became director<br />

of the American Rocket Society.<br />

The New York Times finally apologized to Goddard for its<br />

1920 article on the morning after Apollo 11 took off for the<br />

Moon in 1969. However, Goddard, who battled tuberculosis<br />

most of his life, had died twenty-four years earlier.<br />

pane, ether, kerosene, or gasoline, but they all had serious disadvantages.<br />

Finally, Goddard found a local source of oxygen <strong>and</strong> was able<br />

to begin testing its thrust.


Another problem was designing a fuel pump. Goddard <strong>and</strong> his<br />

assistant Nils Riffolt spent years on this problem before the historic<br />

test flight of March, 1926. In the end, because of pressure from the<br />

Smithsonian Institution <strong>and</strong> others who were funding his research,<br />

Goddard decided to do without a pump <strong>and</strong> use an inert gas to<br />

push the fuel into the explosion chamber.<br />

Goddard worked without much funding between 1920 <strong>and</strong> 1925.<br />

Riffolt helped him greatly in designing a pump, <strong>and</strong> Goddard’s<br />

wife, Esther, photographed some of the tests <strong>and</strong> helped in other<br />

ways. Clark University had granted him some research money in<br />

1923, but by 1925 money was in short supply, <strong>and</strong> the Smithsonian<br />

Institution did not seem willing to grant more. Goddard was convinced<br />

that his research would be taken seriously if he could show<br />

some serious results, so on March 16, 1926, he launched a rocket<br />

even though his design was not yet perfect. The success of that<br />

launch not only changed his career but also set the stage for rocketry<br />

experiments both in the United States <strong>and</strong> in Europe.<br />

Impact<br />

Rocket / 661<br />

Goddard was described as being secretive <strong>and</strong> a loner. He never<br />

tried to cash in on his invention but continued his research during<br />

the next three years. On July 17, 1929, Goddard launched a rocket<br />

carrying a camera <strong>and</strong> instruments for measuring temperature<br />

<strong>and</strong> air pressure. The New York Times published a story about the<br />

noisy crash of this rocket <strong>and</strong> local officials’ concerns about public<br />

safety. The article also mentioned Goddard’s idea that a similar<br />

rocket might someday strike the Moon. When American aviation<br />

hero Charles A. Lindbergh learned of Goddard’s work, Lindbergh<br />

helped him to get grants from the Carnegie Institution <strong>and</strong> the<br />

Guggenheim Foundation.<br />

By the middle of 1930, Goddard <strong>and</strong> a small group of assistants<br />

had established a full-time research program near Roswell, New<br />

Mexico. Now that money was not so much of a problem, Goddard<br />

began to make significant advances in almost every area of astronautics.<br />

In 1941, Goddard launched a rocket to a height of 2,700 meters.<br />

Flight stability was helped by a gyroscope, <strong>and</strong> he was finally<br />

able to use a fuel pump.


662 / Rocket<br />

During the 1920’s <strong>and</strong> 1930’s, members of the American Rocket<br />

Society <strong>and</strong> the German Society for Space Travel continued their<br />

own research. When World War II began, rocket research became a<br />

high priority for the American <strong>and</strong> German governments.<br />

Germany’s success with the V-2 rocket was a direct result of<br />

Goddard’s research <strong>and</strong> inventions, but the United States did not<br />

benefit fully from Goddard’s work until after his death. Nevertheless,<br />

Goddard remains modern rocketry’s foremost pioneer—a scientist<br />

with vision, underst<strong>and</strong>ing, <strong>and</strong> practical skill.<br />

See also Airplane; Artificial satellite; Communications satellite;<br />

Cruise missile; Hydrogen bomb; Stealth aircraft; Supersonic passenger<br />

plane; Turbojet; V-2 rocket; Weather satellite.<br />

Further Reading<br />

Alway, Peter. Retro Rockets: Experimental Rockets, 1926-1941. Ann Arbor,<br />

Mich.: Saturn Press, 1996.<br />

Goddard, Robert Hutchings. The Autobiography of Robert Hutchings<br />

Goddard, Father of the Space Age: Early Years to 1927. Worcester,<br />

Mass.: A. J. St. Onge, 1966.<br />

Lehman, Milton. Robert H. Goddard: Pioneer of Space Research. New<br />

York: Da Capo Press, 1988.


Rotary dial telephone<br />

Rotary dial telephone<br />

The invention: The first device allowing callers to connect their<br />

telephones to other parties without the aid of an operator, the rotary<br />

dial telephone preceded the touch-tone phone.<br />

The people behind the invention:<br />

Alex<strong>and</strong>er Graham Bell (1847-1922), an American inventor<br />

Antoine Barnay (1883-1945), a French engineer<br />

Elisha Gray (1835-1901), an American inventor<br />

Rotary Telephones Dials Make Phone Linkups Automatic<br />

663<br />

The telephone uses electricity to carry sound messages over long<br />

distances. When a call is made from a telephone set, the caller<br />

speaks into a telephone transmitter <strong>and</strong> the resultant sound waves<br />

are converted into electrical signals. The electrical signals are then<br />

transported over a telephone line to the receiver of a second telephone<br />

set that was designated when the call was initiated. This receiver<br />

reverses the process, converting the electrical signals into the<br />

sounds heard by the recipient of the call. The process continues as<br />

the parties talk to each other.<br />

The telephone was invented in the 1870’s <strong>and</strong> patented in 1876 by<br />

Alex<strong>and</strong>er Graham Bell. Bell’s patent application barely preceded<br />

an application submitted by his competitor Elisha Gray. After a<br />

heated patent battle between Bell <strong>and</strong> Gray, which Bell won, Bell<br />

founded the Bell Telephone Company, which later came to be called<br />

the American Telephone <strong>and</strong> Telegraph Company.<br />

At first, the transmission of phone calls between callers <strong>and</strong> recipients<br />

was carried out manually, by switchboard operators. In<br />

1923, however, automation began with Antoine Barnay’s development<br />

of the rotary telephone dial. This dial caused the emission of<br />

variable electrical impulses that could be decoded automatically<br />

<strong>and</strong> used to link the telephone sets of callers <strong>and</strong> call recipients. In<br />

time, the rotary dial system gave way to push-button dialing <strong>and</strong><br />

other more modern networking techniques.


664 / Rotary dial telephone<br />

Rotary-dial telephone. (Image Club Graphics)<br />

Telephones, Switchboards, <strong>and</strong> Automation<br />

The carbon transmitter, which is still used in many modern telephone<br />

sets, was the key to the development of the telephone by Alex<strong>and</strong>er<br />

Graham Bell. This type of transmitter—<strong>and</strong> its more modern<br />

replacements—operates like an electric version of the human<br />

ear. When a person talks into the telephone set in a carbon transmitter-equipped<br />

telephone, the sound waves that are produced strike<br />

an electrically connected metal diaphragm <strong>and</strong> cause it to vibrate.<br />

The speed of vibration of this electric eardrum varies in accordance<br />

with the changes in air pressure caused by the changing tones of the<br />

speaker’s voice.<br />

Behind the diaphragm of a carbon transmitter is a cup filled with<br />

powdered carbon. As the vibrations cause the diaphragm to press<br />

against the carbon, the electrical signals—electrical currents of varying<br />

strength—pass out of the instrument through a telephone wire.<br />

Once the electrical signals reach the receiver of the phone being<br />

called, they activate electromagnets in the receiver that make a second<br />

diaphragm vibrate. This vibration converts the electrical signals<br />

into sounds that are very similar to the sounds made by the person<br />

who is speaking. Therefore, a telephone receiver may be viewed<br />

as an electric mouth.<br />

In modern telephone systems, transportation of the electrical signals<br />

between any two phone sets requires the passage of those signals<br />

through vast telephone networks consisting of huge numbers<br />

of wires, radio systems, <strong>and</strong> other media. The linkup of any two


Alex<strong>and</strong>er Graham Bell<br />

Rotary dial telephone / 665<br />

During the funeral for Alex<strong>and</strong>er Graham Bell in 1922, telephone<br />

service throughout the United States stopped for one<br />

minute to honor him. To most people he was the inventor of the<br />

telephone. In fact, his genius ranged much further.<br />

Bell was born in Edinburgh, Scotl<strong>and</strong>, in 1847. His father,<br />

an elocutionist who invented a phonetic alphabet, <strong>and</strong> his<br />

mother, who was deaf, imbued him with deep curiosity, especially<br />

about sound. As a boy Bell became an exceptional pianist,<br />

<strong>and</strong> he produced his first invention, for cleaning wheat, at<br />

fourteen. After Edinburgh’s Royal High School, he attended<br />

classes at Edinburgh University <strong>and</strong> University College, London,<br />

but at the age of twenty-three, battling tuberculosis, he<br />

left school to move with his parents to Ontario, Canada, to<br />

convalesce. Meanwhile, he worked on his idea for a telegraph<br />

capable of sending multiple messages at once. From it grew<br />

the basic concept for the telephone. He developed it while<br />

teaching Visible Speech at the Boston School for Deaf Mutes<br />

after 1871. Assisted by Thomas Watson, he succeeded in sending<br />

speech over a wire <strong>and</strong> was issued a patent for his device,<br />

among the most valuable ever granted, in 1876. His demonstration<br />

of the telephone later that year at Philadelphia’s<br />

Centennial Exhibition <strong>and</strong> its subsequent development into a<br />

household appliance brought him wealth <strong>and</strong> fame.<br />

He moved to Nova Scotia, Canada, <strong>and</strong> continued inventing.<br />

He created a photophone, tetrahedron modules for construction,<br />

<strong>and</strong> an airplane, the Silver Dart, which flew in 1909.<br />

Even though existing technology made them impracticable,<br />

some of his ideas anticipated computers <strong>and</strong> magnetic sound<br />

recording. His last patented invention, tested three years before<br />

his death, was a hydrofoil. Capable of reaching seventy-one<br />

miles per hour <strong>and</strong> freighting fourteen thous<strong>and</strong> pounds, the<br />

HD-4 was then the fastest watercraft in the world.<br />

Bell also helped found the National Geographic Society in<br />

1888 <strong>and</strong> became its president in 1898. He hired Gilbert Grosvenor<br />

to edit the society’s famous magazine, National Geographic<br />

<strong>and</strong> together they planned the format—breathtaking<br />

photography <strong>and</strong> vivid writing—that made it one of the world’s<br />

best known magazines.


666 / Rotary dial telephone<br />

phone sets was originally, however, accomplished manually—on a<br />

relatively small scale—by a switchboard operator who made the<br />

necessary connections by h<strong>and</strong>. In such switchboard systems, each<br />

telephone set in the network was associated with a jack connector in<br />

the switchboard. The operator observed all incoming calls, identified<br />

the phone sets for which they were intended, <strong>and</strong> then used<br />

wires to connect the appropriate jacks. At the end of the call, the<br />

jacks were disconnected.<br />

This cumbersome methodology limited the size <strong>and</strong> efficiency of<br />

telephone networks <strong>and</strong> invaded the privacy of callers. The development<br />

of automated switching systems soon solved these problems<br />

<strong>and</strong> made switchboard operators obsolete. It was here that<br />

Antoine Barnay’s rotary dial was used, making possible an exchange<br />

that automatically linked the phone sets of callers <strong>and</strong> call<br />

recipients in the following way.<br />

First, a caller lifted a telephone “off the hook,” causing a switchhook,<br />

like those used in modern phones, to close the circuit that connected<br />

the telephone set to the telephone network. Immediately, a<br />

dial tone (still familiar to callers) came on to indicate that the automatic<br />

switching system could h<strong>and</strong>le the planned call. When the<br />

phone dial was used, each number or letter that was dialed produced<br />

a fixed number of clicks. Every click indicated that an electrical<br />

pulse had been sent to the network’s automatic switching system,<br />

causing switches to change position slightly. Immediately after<br />

a complete telephone number was dialed, the overall operation of<br />

the automatic switchers connected the two telephone sets. This connection<br />

was carried out much more quickly <strong>and</strong> accurately than had<br />

been possible when telephone operators at manual switchboards<br />

made the connection.<br />

Impact<br />

The telephone has become the world’s most important communication<br />

device. Most adults use it between six <strong>and</strong> eight times per<br />

day, for personal <strong>and</strong> business calls. This widespread use has developed<br />

because huge changes have occurred in telephones <strong>and</strong> telephone<br />

networks. For example, automatic switching <strong>and</strong> the rotary<br />

dial system were only the beginning of changes in phone calling.


Touch-tone dialing replaced Barnay’s electrical pulses with audio<br />

tones outside the frequency of human speech. This much-improved<br />

system can be used to send calls over much longer distances than<br />

was possible with the rotary dial system, <strong>and</strong> it also interacts well<br />

with both answering machines <strong>and</strong> computers.<br />

Another advance in modern telephoning is the use of radio<br />

transmission techniques in mobile phones, rendering telephone<br />

cords obsolete. The mobile phone communicates with base stations<br />

arranged in “cells” throughout the service area covered. As the user<br />

changes location, the phone link automatically moves from cell to<br />

cell in a cellular network.<br />

In addition, the use of microwave, laser, <strong>and</strong> fiber-optic technologies<br />

has helped to lengthen the distance over which phone calls can<br />

be transmitted. These technologies have also increased the number<br />

of messages that phone networks can h<strong>and</strong>le simultaneously <strong>and</strong><br />

have made it possible to send radio <strong>and</strong> television programs (such<br />

as cable television), scientific data (via modems), <strong>and</strong> written messages<br />

(via facsimile, or “fax,” machines) over phone lines. Many<br />

other advances in telephone technology are expected as society’s<br />

needs change <strong>and</strong> new technology is developed.<br />

See also Cell phone; Internet; Long-distance telephone; Telephone<br />

switching; Touch-tone telephone.<br />

Further Reading<br />

Rotary dial telephone / 667<br />

Aitken, William. Who Invented the Telephone? London: Blackie <strong>and</strong><br />

Son, 1939.<br />

Coe, Lewis. The Telephone <strong>and</strong> Its Several <strong>Inventors</strong>: A History. Jefferson,<br />

N.C.: McFarl<strong>and</strong>, 1995.<br />

Evenson, A. Edward. The Telephone Patent Conspiracy of 1876: The<br />

Elisha Gray-Alex<strong>and</strong>er Bell Controversy <strong>and</strong> Its Many Players. Jefferson,<br />

N.C.: McFarl<strong>and</strong>, 2000.<br />

Lisser, Eleena de. “Telecommunications: If You Have a Rotary<br />

Phone, Press 1: The Trials of Using the Old Apparatus.” Wall<br />

Street Journal (July 28, 1994).<br />

Mackay, James A. Alex<strong>and</strong>er Graham Bell: A Life. New York: J. Wiley,<br />

1997.


668<br />

SAINT<br />

SAINT<br />

The invention: Taking its name from the acronym for symbolic automatic<br />

integrator, SAINT is recognized as the first “expert system”—a<br />

computer program designed to perform mental tasks requiring<br />

human expertise.<br />

The person behind the invention:<br />

James R. Slagle (1934-1994), an American computer scientist<br />

The Advent of Artificial Intelligence<br />

In 1944, the Harvard-IBM Mark I was completed. This was an<br />

electromechanical (that is, not fully electronic) digital computer<br />

that was operated by means of coding instructions punched into<br />

paper tape. The machine took about six seconds to perform a multiplication<br />

operation, twelve for a division operation. In the following<br />

year, 1945, the world’s first fully electronic digital computer,<br />

the Electronic Numerical Integrator <strong>and</strong> Calculator (ENIAC),<br />

became operational. This machine, which was constructed at the<br />

University of Pennsylvania, was thirty meters long, three meters<br />

high, <strong>and</strong> one meter deep.<br />

At the same time that these machines were being built, a similar<br />

machine was being constructed in the United Kingdom: the automated<br />

computing engine (ACE). Akey figure in the British development<br />

was Alan Turing, a mathematician who had used computers<br />

to break German codes during World War II. After the war, Turing<br />

became interested in the area of “computing machinery <strong>and</strong> intelligence.”<br />

He posed the question “Can machines think?” <strong>and</strong> set the<br />

following problem, which is known as the “Turing test.” This test<br />

involves an interrogator who sits at a computer terminal <strong>and</strong> asks<br />

questions on the terminal about a subject for which he or she seeks intelligent<br />

answers. The interrogator does not know, however, whether<br />

the system is linked to a human or if the responses are, in fact, generated<br />

by a program that is acting intelligently. If the interrogator cannot<br />

tell the difference between the human operator <strong>and</strong> the computer<br />

system, then the system is said to have passed the Turing test<br />

<strong>and</strong> has exhibited intelligent behavior.


SAINT: An Expert System<br />

SAINT / 669<br />

In the attempt to answer Turing’s question <strong>and</strong> create machines<br />

that could pass the Turing test, researchers investigated techniques<br />

for performing tasks that were considered to require expert levels of<br />

knowledge. These tasks included games such as checkers, chess, <strong>and</strong><br />

poker. These games were chosen because the total possible number of<br />

variations in each game was very large. This led the researchers to<br />

several interesting questions for study. How do experts make a decision<br />

in a particular set of circumstances? How can a problem such as<br />

a game of chess be represented in terms of a computer program? Is it<br />

possible to know why the system chose a particular solution?<br />

One researcher, James R. Slagle at the Massachusetts Institute of<br />

Technology, chose to develop a program that would be able to solve<br />

elementary symbolic integration problems (involving the manipulation<br />

of integrals in calculus) at the level of a good college freshman.<br />

The program that Slagle constructed was known as SAINT, an<br />

acronym for symbolic automatic integrator, <strong>and</strong> it is acknowledged<br />

as the first “expert system.”<br />

An expert system is a system that performs at the level of a human<br />

expert. An expert system has three basic components: a knowledge<br />

base, in which domain-specific information is held (for example, rules<br />

on how best to perform certain types of integration problems); an inference<br />

engine, which decides how to break down a given problem utilizing<br />

the rules in the knowledge base; <strong>and</strong> a human-computer interface<br />

that inputs data—in this case, the integral to be solved—<strong>and</strong><br />

outputs the result of performing the integration. Another feature of expert<br />

systems is their ability to explain their reasoning.<br />

The integration problems that could be solved by SAINT were<br />

in the form of elementary integral functions. SAINT could perform<br />

indefinite integration (also called “antidifferentiation”) on these<br />

functions. In addition, it was capable of performing definite <strong>and</strong><br />

indefinite integration on trivial extensions of indefinite integration.<br />

SAINT was tested on a set of eighty-six problems, fifty-four of<br />

which were drawn from the MIT final examinations in freshman<br />

calculus; it succeeded in solving all but two. Slagle added more<br />

rules to the knowledge base so that problems of the type it encountered<br />

but could not solve could be solved in the future.


670 / SAINT<br />

The power of the SAINT system was, in part, based on its ability<br />

to perform integration through the adoption of a “heuristic” processing<br />

system. Aheuristic method is one that helps in discovering a<br />

problem’s solution by making plausible but feasible guesses about<br />

the best strategy to apply next to the current problem situation. A<br />

heuristic is a rule of thumb that makes it possible to take short cuts<br />

in reaching a solution, rather than having to go through every step<br />

in a solution path. These heuristic rules are contained in the knowledge<br />

base. SAINT was written in the LISP programming language<br />

<strong>and</strong> ran on an IBM 7090 computer. The program <strong>and</strong> research were<br />

Slagle’s doctoral dissertation.<br />

Consequences<br />

User<br />

Interface<br />

Global<br />

Database<br />

Control<br />

Mechanism<br />

Knowledge<br />

Base<br />

Basic structure of an expert system.<br />

Facts<br />

Search<br />

<strong>and</strong><br />

Resolve<br />

Rules<br />

The SAINT system that Slagle developed was significant for several<br />

reasons: First, it was the first serious attempt at producing a<br />

program that could come close to passing the Turing test. Second, it<br />

brought the idea of representing an expert’s knowledge in a computer<br />

program together with strategies for solving complex <strong>and</strong> difficult<br />

problems in an area that previously required human expertise.<br />

Third, it identified the area of knowledge-based systems <strong>and</strong>


James R. Slagle<br />

SAINT / 671<br />

James R. Slagle was born in 1934 in Brooklyn, New York, <strong>and</strong><br />

attended nearby St. John’s University. He majored in mathematics<br />

<strong>and</strong> graduated with a bachelor of science degree in 1955,<br />

also winning the highest scholastic average award. While earning<br />

his master’s degree (1957) <strong>and</strong> doctorate (1961) at the Massachusetts<br />

Institute of Technology (MIT), he was a staff mathematician<br />

in the university’s Lincoln Laboratory.<br />

Slagle taught in MIT’s electrical engineering department<br />

part-time after completing his dissertation on the first expert<br />

computer system <strong>and</strong> then moved to Lawrence-Livermore<br />

National Laboratory near Berkeley, California. While working<br />

there he also taught at the University of California. From 1967<br />

until 1974 he was an adjunct member of the computer science<br />

faculty of Johns Hopkins University in Baltimore, Maryl<strong>and</strong>,<br />

<strong>and</strong> then was appointed chief of the computer science laboratory<br />

at the Naval Research Laboratory (NRL) in Washington, D.C., receiving<br />

the Outst<strong>and</strong>ing H<strong>and</strong>icapped Federal Employee of the<br />

Year Award in 1979. In 1984 he was made a special assistant in<br />

the Navy Center for Applied Research in Artificial Intelligence<br />

at NRL but left in 1984 to become Distinguished Professor of<br />

Computer Science at the University of Minnesota.<br />

In these various positions Slagle helped mature the fledgling<br />

discipline of artificial intelligence, publishing the influential<br />

book Artificial Intelligence in 1971. He developed an expert system<br />

designed to set up other expert systems—A Generalized<br />

Network-based Expert System Shell, or AGNESS. He also worked<br />

on parallel expert systems, artificial neural networks, timebased<br />

logic, <strong>and</strong> methods for uncovering causal knowledge in<br />

large databases. He died in 1994.<br />

showed that computers could feasibly be used for programs that<br />

did not relate to business data processing. Fourth, the SAINT system<br />

showed how the use of heuristic rules <strong>and</strong> information could<br />

lead to the solution of problems that could not have been solved<br />

previously because of the amount of time needed to calculate a solution.<br />

SAINT’s major impact was in outlining the uses of these techniques,<br />

which led to continued research in the subfield of artificial<br />

intelligence that became known as expert systems.


672 / SAINT<br />

See also BASIC programming language; CAD/CAM; COBOL<br />

computer language; Differential analyzer; FORTRAN programming<br />

language; Robot (industrial).<br />

Further Reading<br />

Campbell-Kelly, Martin, <strong>and</strong> William Aspray. Computer: A History of<br />

the Information Machine. New York: Basic Books, 1996.<br />

Ceruzzi, Paul E. A History of Modern Computing. Cambridge, Mass.:<br />

MIT Press, 2000.<br />

Rojas, Paul. Encyclopedia of Computers <strong>and</strong> Computer History. London:<br />

Fitzroy Dearborn, 2001.


Salvarsan<br />

Salvarsan<br />

The invention: The first successful chemotherapeutic for the treatment<br />

of syphilis<br />

The people behind the invention:<br />

Paul Ehrlich (1854-1915), a German research physician <strong>and</strong><br />

chemist<br />

Wilhelm von Waldeyer (1836-1921), a German anatomist<br />

Friedrich von Frerichs (1819-1885), a German physician <strong>and</strong><br />

professor<br />

Sahachiro Hata (1872-1938), a Japanese physician <strong>and</strong><br />

bacteriologist<br />

Fritz Schaudinn (1871-1906), a German zoologist<br />

The Great Pox<br />

673<br />

The ravages of syphilis on humankind are seldom discussed<br />

openly. A disease that struck all varieties of people <strong>and</strong> was transmitted<br />

by direct <strong>and</strong> usually sexual contact, syphilis was both<br />

feared <strong>and</strong> reviled. Many segments of society across all national<br />

boundaries were secure in their belief that syphilis was divine punishment<br />

of the wicked for their evil ways.<br />

It was not until 1903 that bacteriologists Élie Metchnikoff <strong>and</strong><br />

Pierre-Paul-Émile Roux demonstrated the transmittal of syphilis to<br />

apes, ending the long-held belief that syphilis was exclusively a human<br />

disease. The disease destroyed families, careers, <strong>and</strong> lives,<br />

driving its infected victims mad, destroying the brain, or destroying<br />

the cardiovascular system. It was methodical <strong>and</strong> slow, but in every<br />

case, it killed with singular precision. There was no hope of a safe<br />

<strong>and</strong> effective cure prior to the discovery of Salvarsan.<br />

Prior to 1910, conventional treatment consisted principally of<br />

mercury or, later, potassium iodide. Mercury, however, administered<br />

in large doses, led to severe ulcerations of the tongue, jaws,<br />

<strong>and</strong> palate. Swelling of the gums <strong>and</strong> loosening of the teeth resulted.<br />

Dribbling saliva <strong>and</strong> the attending fetid odor also occurred. These<br />

side effects of mercury treatment were so severe that many pre-


674 / Salvarsan<br />

ferred to suffer the disease to the end rather than undergo the st<strong>and</strong>ard<br />

cure. About 1906, Metchnikoff <strong>and</strong> Roux demonstrated that<br />

mercurial ointments, applied very early, at the first appearance of<br />

the primary lesion, were effective.<br />

Once the spirochete-type bacteria invaded the bloodstream <strong>and</strong><br />

tissues, the infected person experienced symptoms of varying nature<br />

<strong>and</strong> degree—high fever, intense headaches, <strong>and</strong> excruciating<br />

pain. The patient’s skin often erupted in pustular lesions similar in<br />

appearance to smallpox. It was the distinguishing feature of these<br />

pustular lesions that gave syphilis its other name: the “Great Pox.”<br />

Death brought the only relief then available.<br />

Poison Dyes<br />

Paul Ehrlich became fascinated by the reactions of dyes with biological<br />

cells <strong>and</strong> tissues while a student at the University of Strasbourg<br />

under Wilhelm von Waldeyer. It was von Waldeyer who<br />

sparked Ehrlich’s interest in the chemical viewpoint of medicine.<br />

Thus, as a student, Ehrlich spent hours at this laboratory experimenting<br />

with different dyes on various tissues. In 1878, he published<br />

a book that detailed the discriminate staining of cells <strong>and</strong> cellular<br />

components by various dyes.<br />

Ehrlich joined Friedrich von Frerichs at the Charité Hospital in<br />

Berlin, where Frerichs allowed Ehrlich to do as much research as he<br />

wanted. Ehrlich began studying atoxyl in 1908, the year he won<br />

jointly with Metchnikoff the Nobel Prize in Physiology or Medicine<br />

for his work on immunity. Atoxyl was effective against trypanosome—a<br />

parasite responsible for a variety of infections, notably<br />

sleeping sickness—but also imposed serious side effects upon the<br />

patient, not the least of which was blindness. It was Ehrlich’s study<br />

of atoxyl, <strong>and</strong> several hundred derivatives sought as alternatives to<br />

atoxyl in trypanosome treatment, that led to the development of derivative<br />

606 (Salvarsan). Although compound 606 was the first<br />

chemotherapeutic to be used effectively against syphilis, it was discontinued<br />

as an atoxyl alternative <strong>and</strong> shelved as useless for five<br />

years.<br />

The discovery <strong>and</strong> development of compound 606 was enhanced<br />

by two critical events. First, the Germans Fritz Schaudinn <strong>and</strong> Erich


The wonder drug Salvarsan was often called “Ehrlich’s silver bullet,” after its developer,<br />

Paul Ehrlich (left). (Library of Congress)


676 / Salvarsan<br />

Hoffmann discovered that syphilis is a bacterially caused disease.<br />

The causative microorganism is a spirochete so frail <strong>and</strong> gossameric<br />

in substance that it is nearly impossible to detect by casual microscopic<br />

examination; Schaudinn chanced upon it one day in March,<br />

1905. This discovery led, in turn, to German bacteriologist August<br />

von Wassermann’s development of the now famous test for syphilis:<br />

the Wassermann test. Second, a Japanese bacteriologist, Sahachiro<br />

Hata, came to Frankfurt in 1909 to study syphilis with<br />

Ehrlich. Hata had studied syphilis in rabbits in Japan. Hata’s assignment<br />

was to test every atoxyl derivative ever developed under<br />

Ehrlich for its efficacy in syphilis treatment. After hundreds of tests<br />

<strong>and</strong> clinical trials, Ehrlich <strong>and</strong> Hata announced Salvarsan as a<br />

“magic bullet” that could cure syphilis, at the April, 1910, Congress<br />

of Internal Medicine in Wiesbaden, Germany.<br />

The announcement was electrifying. The remedy was immediately<br />

<strong>and</strong> widely sought, but it was not without its problems. A few deaths<br />

resulted from its use, <strong>and</strong> it was not safe for treatment of the gravely ill.<br />

Some of the difficulties inherent in Salvarsan were overcome by the development<br />

of neosalvarsan in 1912 <strong>and</strong> sodium salvarsan in 1913. Although<br />

Ehrlich achieved much, he fell short of his own assigned goal, a<br />

chemotherapeutic that would cure in one injection.<br />

Impact<br />

The significance of the development of Salvarsan as an antisyphilitic<br />

chemotherapeutic agent cannot be overstated. Syphilis at<br />

that time was as frightening <strong>and</strong> horrifying as leprosy <strong>and</strong> was a virtual<br />

sentence of slow, torturous death. Salvarsan was such a significant<br />

development that Ehrlich was recommended for a 1912 <strong>and</strong><br />

1913 Nobel Prize for his work in chemotherapy.<br />

It was several decades before any further significant advances in<br />

“wonder drugs” occurred, namely, the discovery of prontosil in 1932<br />

<strong>and</strong> its first clinical use in 1935. On the heels of prontosil—a sulfa<br />

drug—came other sulfa drugs. The sulfa drugs would remain supreme<br />

in the fight against bacterial infection until the antibiotics, the<br />

first being penicillin, were discovered in 1928; however, they were<br />

not clinically recognized until World War II (1939-1945). With the discovery<br />

of streptomycin in 1943 <strong>and</strong> Aureomycin in 1944, the assault


against bacteria was finally on a sound basis. Medicine possessed an<br />

arsenal with which to combat the pathogenic microbes that for centuries<br />

before had visited misery <strong>and</strong> death upon humankind.<br />

See also Abortion pill; Antibacterial drugs; Birth control pill;<br />

Penicillin; Reserpine; Syphilis test; Tuberculosis vaccine; Typhus<br />

vaccine; Yellow fever vaccine.<br />

Further Reading<br />

Salvarsan / 677<br />

Bäumler, Ernst. Paul Ehrlich: Scientist for Life. New York: Holmes &<br />

Meier, 1984.<br />

Leyden, John G. “From Nobel Prize to Courthouse Battle: Paul<br />

Ehrlich’s ‘Wonder Drug’ for Syphilis Won Him Acclaim but also<br />

Led Critics to Hound Him.” Washington Post (July 27, 1999).<br />

Quétel, Claude. History of Syphilis. Baltimore: Johns Hopkins University<br />

Press, 1992.


678<br />

Scanning tunneling microscope<br />

Scanning tunneling microscope<br />

The invention: A major advance on the field ion microscope, the<br />

scanning tunneling microscope has pointed toward new directions<br />

in the visualization <strong>and</strong> control of matter at the atomic<br />

level.<br />

The people behind the invention:<br />

Gerd Binnig (1947- ), a West German physicist who was a<br />

cowinner of the 1986 Nobel Prize in Physics<br />

Heinrich Rohrer (1933- ), a Swiss physicist who was a<br />

cowinner of the 1986 Nobel Prize in Physics<br />

Ernst Ruska (1906-1988), a West German engineer who was a<br />

cowinner of the 1986 Nobel Prize in Physics<br />

Antoni van Leeuwenhoek (1632-1723), a Dutch naturalist<br />

The Limit of Light<br />

The field of microscopy began at the end of the seventeenth century,<br />

when Antoni van Leeuwenhoek developed the first optical microscope.<br />

In this type of microscope, a magnified image of a sample<br />

is obtained by directing light onto it <strong>and</strong> then taking the light<br />

through a lens system. Van Leeuwenhoek’s microscope allowed<br />

him to observe the existence of life on a scale that is invisible to the<br />

naked eye. Since then, developments in the optical microscope have<br />

revealed the existence of single cells, pathogenic agents, <strong>and</strong> bacteria.<br />

There is a limit, however, to the resolving power of optical microscopes.<br />

Known as “Abbe’s barrier,” after the German physicist <strong>and</strong><br />

lens maker Ernst Abbe, this limit means that objects smaller than<br />

about 400 nanometers (about a millionth of a millimeter) cannot be<br />

viewed by conventional microscopes.<br />

In 1925, the physicist Louis de Broglie predicted that electrons<br />

would exhibit wave behavior as well as particle behavior. This prediction<br />

was confirmed by Clinton J. Davisson <strong>and</strong> Lester H. Germer<br />

of Bell Telephone Laboratories in 1927. It was found that highenergy<br />

electrons have shorter wavelengths than low-energy electrons<br />

<strong>and</strong> that electrons with sufficient energies exhibit wave-


lengths comparable to the diameter of the atom. In 1927, Hans<br />

Busch showed in a mathematical analysis that current-carrying<br />

coils behave like electron lenses <strong>and</strong> that they obey the same lens<br />

equation that governs optical lenses. Using these findings, Ernst<br />

Ruska developed the electron microscope in the early 1930’s.<br />

By 1944, the German corporation of Siemens <strong>and</strong> Halske had<br />

manufactured electron microscopes with a resolution of 7 nanometers;<br />

modern instruments are capable of resolving objects as<br />

small as 0.5 nanometer. This development made it possible to view<br />

structures as small as a few atoms across as well as large atoms <strong>and</strong><br />

large molecules.<br />

The electron beam used in this type of microscope limits the usefulness<br />

of the device. First, to avoid the scattering of the electrons,<br />

the samples must be put in a vacuum, which limits the applicability<br />

of the microscope to samples that can sustain such an environment.<br />

Most important, some fragile samples, such as organic molecules,<br />

are inevitably destroyed by the high-energy beams required for<br />

high resolutions.<br />

Viewing Atoms<br />

Scanning tunneling microscope / 679<br />

From 1936 to 1955, Erwin Wilhelm Müller developed the field ion<br />

microscope (FIM), which used an extremely sharp needle to hold the<br />

sample. This was the first microscope to make possible the direct<br />

viewing of atomic structures, but it was limited to samples capable of<br />

sustaining the high electric fields necessary for its operation.<br />

In the early 1970’s, Russel D. Young <strong>and</strong> Clayton Teague of the<br />

National Bureau of St<strong>and</strong>ards (NBS) developed the “topografiner,”<br />

a new kind of FIM. In this microscope, the sample is placed at a large<br />

distance from the tip of the needle. The tip is scanned across the surface<br />

of the sample with a precision of about a nanometer. The precision<br />

in the three-dimensional motion of the tip was obtained by using<br />

three legs made of piezoelectric crystals. These materials change<br />

shape in a reproducible manner when subjected to a voltage. The<br />

extent of expansion or contraction of the crystal depends on the<br />

amount of voltage that is applied. Thus, the operator can control the<br />

motion of the probe by varying the voltage acting on the three legs.<br />

The resolution of the topografiner is limited by the size of the probe.


680 / Scanning tunneling microscope<br />

Gerd Binnig <strong>and</strong> Heinrich Rohrer<br />

Both Gerd Binnig <strong>and</strong> Heinrich Rohrer believe an early <strong>and</strong><br />

pleasurable introduction to teamwork led to their later success<br />

in inventing the scanning tunneling microscope, for which they<br />

shared the 1986 Nobel Prize in Physics with Ernst Ruska.<br />

Binnig was born in Frankfurt, Germany, in 1947. He acquired<br />

an early interest in physics but was always deeply influenced<br />

by classical music, introduced to him by his mother, <strong>and</strong><br />

the rock music that his younger brother played for him. Binnig<br />

played in rock b<strong>and</strong>s as a teenager <strong>and</strong> learned to enjoy the creative<br />

interplay of teamwork. At J. W. Goethe University in<br />

Frankfurt he earned a bachelor’s degree (1973) <strong>and</strong> doctorate<br />

(1978) in physics <strong>and</strong> then took a position at International Business<br />

Machine’s Zurich Research Laboratory. There he recaptured<br />

the pleasures of working with a talented team after joining<br />

Rohrer in research.<br />

Rohrer had been at the Zurich facility since just after it<br />

opened in 1963. He was born in Buch, Switzerl<strong>and</strong>, in 1933, <strong>and</strong><br />

educated at the Swiss Federal Institute of Technology in Zurich,<br />

where he completed his doctorate in 1960. After post-doctoral<br />

work at Rutgers University, he joined the IBM research team, a<br />

time that he describes as among the most enjoyable passages of<br />

his career.<br />

In addition to the Nobel Prize, the pair also received the German<br />

Physics Prize, Otto Klung Prize, Hewlett Packard Prize,<br />

<strong>and</strong> King Faisal Prize. Rohrer became an IBM Fellow in 1986<br />

<strong>and</strong> was selected to manage the physical sciences department at<br />

the Zurich Research Laboratory. He retired from IBM in July<br />

1997. Binnig became an IBM Fellow in 1987.<br />

The idea for the scanning tunneling microscope (STM) arose<br />

when Heinrich Rohrer of the International Business Machines (IBM)<br />

Corporation’s Zurich research laboratory met Gerd Binnig in Frankfurt<br />

in 1978. The STM is very similar to the topografiner. In the STM,<br />

however, the tip is kept at a height of less than a nanometer away<br />

from the surface, <strong>and</strong> the voltage that is applied between the specimen<br />

<strong>and</strong> the probe is low. Under these conditions, the electron<br />

cloud of atoms at the end of the tip overlaps with the electron cloud<br />

of atoms at the surface of the specimen. This overlapping results in a


measurable electrical current flowing through the vacuum or insulating<br />

material existing between the tip <strong>and</strong> the sample. When the<br />

probe is moved across the surface <strong>and</strong> the voltage between the<br />

probe <strong>and</strong> sample is kept constant, the change in the distance between<br />

the probe <strong>and</strong> the surface (caused by surface irregularities)<br />

results in a change of the tunneling current.<br />

Two methods are used to translate these changes into an image of<br />

the surface. The first method involves changing the height of the<br />

probe to keep the tunneling current constant; the voltage used to<br />

change the height is translated by a computer into an image of the<br />

surface. The second method scans the probe at a constant height<br />

away from the sample; the voltage across the probe <strong>and</strong> sample is<br />

changed to keep the tunneling current constant. These changes in<br />

voltage are translated into the image of the surface. The main limitation<br />

of the technique is that it is applicable only to conducting samples<br />

or to samples with some surface treatment.<br />

Consequences<br />

Scanning tunneling microscope / 681<br />

In October, 1989, the STM was successfully used in the manipulation<br />

of matter at the atomic level. By letting the probe sink into the<br />

surface of a metal-oxide crystal, researchers at Rutgers University<br />

were able to dig a square hole about 250 atoms across <strong>and</strong> 10 atoms<br />

deep. Amore impressive feat was reported in the April 5, 1990, issue<br />

of Nature; M. Eigler <strong>and</strong> Erhard K. Schweiser of IBM’s Almaden Research<br />

Center spelled out their employer’s three-letter acronym using<br />

thirty-five atoms of xenon. This ability to move <strong>and</strong> place individual<br />

atoms precisely raises several possibilities, which include the<br />

creation of custom-made molecules, atomic-scale data storage, <strong>and</strong><br />

ultrasmall electrical logic circuits.<br />

The success of the STM has led to the development of several<br />

new microscopes that are designed to study other features of sample<br />

surfaces. Although they all use the scanning probe technique to<br />

make measurements, they use different techniques for the actual detection.<br />

The most popular of these new devices is the atomic force<br />

microscope (AFM). This device measures the tiny electric forces that<br />

exist between the electrons of the probe <strong>and</strong> the electrons of the<br />

sample without the need for electron flow, which makes the tech-


682 / Scanning tunneling microscope<br />

nique particularly useful in imaging nonconducting surfaces. Other<br />

scanned probe microscopes use physical properties such as temperature<br />

<strong>and</strong> magnetism to probe the surfaces.<br />

See also Cyclotron; Electron microscope; Ion field microscope;<br />

Mass spectrograph; Neutrino detector; Sonar; Synchrocyclotron;<br />

Tevatron accelerator; Ultramicroscope.<br />

Further Reading<br />

Morris, Michael D. Microscopic <strong>and</strong> Spectroscopic Imaging of the Chemical<br />

State. New York: M. Dekker, 1993.<br />

Wiesendanger, Robert. Scanning Probe Microscopy: Analytical Methods.<br />

New York: Springer-Verlag, 1998.<br />

_____, <strong>and</strong> Hans-Joachim Güntherodt. Scanning Tunneling Microscopy<br />

II: Further Applications <strong>and</strong> Related Scanning Techniques.2ded.<br />

New York: Springer, 1995.<br />

_____. Scanning Tunneling Microscopy III: Theory of STM <strong>and</strong> Related<br />

Scanning Probe Methods. 2d ed. New York: Springer, 1996.


Silicones<br />

Silicones<br />

The invention: Synthetic polymers characterized by lubricity, extreme<br />

water repellency, thermal stability, <strong>and</strong> inertness that are<br />

widely used in lubricants, protective coatings, paints, adhesives,<br />

electrical insulation, <strong>and</strong> prosthetic replacements for body parts.<br />

The people behind the invention:<br />

Eugene G. Rochow (1909- ), an American research chemist<br />

Frederic Stanley Kipping (1863-1949), a Scottish chemist <strong>and</strong><br />

professor<br />

James Franklin Hyde (1903- ), an American organic chemist<br />

Synthesizing Silicones<br />

683<br />

Frederic Stanley Kipping, in the first four decades of the twentieth<br />

century, made an extensive study of the organic (carbon-based)<br />

chemistry of the element silicon. He had a distinguished academic<br />

career <strong>and</strong> summarized his silicon work in a lecture in 1937 (“Organic<br />

Derivatives of Silicon”). Since Kipping did not have available<br />

any naturally occurring compounds with chemical bonds between<br />

carbon <strong>and</strong> silicon atoms (organosilicon compounds), it was necessary<br />

for him to find methods of establishing such bonds. The basic<br />

method involved replacing atoms in naturally occurring silicon<br />

compounds with carbon atoms from organic compounds.<br />

While Kipping was probably the first to prepare a silicone <strong>and</strong> was<br />

certainly the first to use the term silicone, he did not pursue the commercial<br />

possibilities of silicones. Yet his careful experimental work was<br />

a valuable starting point for all subsequent workers in organosilicon<br />

chemistry, including those who later developed the silicone industry.<br />

On May 10, 1940, chemist Eugene G. Rochow of the General<br />

Electric (GE) Company’s corporate research laboratory in<br />

Schenectady, New York, discovered that methyl chloride gas,<br />

passed over a heated mixture of elemental silicon <strong>and</strong> copper, reacted<br />

to form compounds with silicon-carbon bonds. Kipping<br />

had shown that these silicon compounds react with water to form<br />

silicones.


684 / Silicones<br />

The importance of Rochow’s discovery was that it opened the<br />

way to a continuous process that did not consume expensive metals,<br />

such as magnesium, or flammable ether solvents, such as those<br />

used by Kipping <strong>and</strong> other researchers. The copper acts as a catalyst,<br />

<strong>and</strong> the desired silicon compounds are formed with only minor<br />

quantities of by-products. This “direct synthesis,” as it came to be<br />

called, is now done commercially on a large scale.<br />

Silicone Structure<br />

Silicones are examples of what chemists call polymers. Basically, a<br />

polymer is a large molecule made up of many smaller molecules<br />

that are linked together. At the molecular level, silicones consist of<br />

long, repeating chains of atoms. In this molecular characteristic, silicones<br />

resemble plastics <strong>and</strong> rubber.<br />

Silicone molecules have a chain composed of alternate silicon <strong>and</strong><br />

oxygen atoms. Each silicon atom bears two organic groups as substituents,<br />

while the oxygen atoms serve to link the silicon atoms into a<br />

chain. The silicon-oxygen backbone of the silicones is responsible for<br />

their unique <strong>and</strong> useful properties, such as the ability of a silicone oil<br />

to remain liquid over an extremely broad temperature range <strong>and</strong> to<br />

resist oxidative <strong>and</strong> thermal breakdown at high temperatures.<br />

A fundamental scientific consideration with silicone, as with any<br />

polymer, is to obtain the desired physical <strong>and</strong> chemical properties in<br />

a product by closely controlling its chemical structure <strong>and</strong> molecular<br />

weight. Oily silicones with thous<strong>and</strong>s of alternating silicon <strong>and</strong><br />

oxygen atoms have been prepared. The average length of the molecular<br />

chain determines the flow characteristics (viscosity) of the oil.<br />

In samples with very long chains, rubber-like elasticity can be<br />

achieved by cross-linking the silicone chains in a controlled manner<br />

<strong>and</strong> adding a filler such as silica. High degrees of cross-linking<br />

could produce a hard, intractable material instead of rubber.<br />

The action of water on the compounds produced from Rochow’s<br />

direct synthesis is a rapid method of obtaining silicones, but does<br />

not provide much control of the molecular weight. Further development<br />

work at GE <strong>and</strong> at the Dow-Corning company showed that<br />

the best procedure for controlled formation of silicone polymers involved<br />

treating the crude silicones with acid to produce a mixture


Eugene G. Rochow<br />

Silicones / 685<br />

Eugene George Rochow was born in 1909 <strong>and</strong> grew up in the<br />

rural New Jersey town of Maplewood. There his father, who<br />

worked in the tanning industry, <strong>and</strong> his big brother maintained<br />

a small attic laboratory. They experimented with electricity, radio—Eugene<br />

put together his own crystal set in an oatmeal<br />

box—<strong>and</strong> chemistry.<br />

Rochow followed his brother to Cornell University in 1927.<br />

The Great Depression began during his junior year, <strong>and</strong> although<br />

he had to take jobs as lecture assistant to get by, he managed<br />

to earn his bachelor’s degree in chemistry in 1931 <strong>and</strong> his<br />

doctorate in 1935. Luck came his way in the extremely tight job<br />

market: General Electric (GE) hired him for his expertise in inorganic<br />

chemistry.<br />

In 1938 the automobile industry, among other manufacturers,<br />

had a growing need for a high-temperature-resistant insulators.<br />

Scientists at Corning were convinced that silicone would<br />

have the best properties for the purpose, but they could not find<br />

a way to synthesize it cheaply <strong>and</strong> in large volume. When word<br />

about their ideas got back to Rochow at GE, he reasoned that a<br />

flexible silicone able to withst<strong>and</strong> temperatures of 200 to 300 degrees<br />

Celsius could be made by bonding silicone to carbon. His<br />

research succeeded in producing methyl silicone in 1939, <strong>and</strong><br />

he devised a way to make it cheaply in 1941. It was the first<br />

commercially practical silicone. His process is still used.<br />

After World War II GE asked him to work on an effort to<br />

make aircraft carriers nuclear powered. However, Rochow was<br />

a Quaker <strong>and</strong> pacifist, <strong>and</strong> he refused. Instead, he moved to<br />

Harvard University as a chemistry professor in 1948 <strong>and</strong> remained<br />

there until his retirement in 1970. As the founder of a<br />

new branch of industrial chemistry, he received most of the discipline’s<br />

awards <strong>and</strong> medals, including the Perkin Award, <strong>and</strong><br />

honorary doctorates.<br />

from which high yields of an intermediate called “D4” could be obtained<br />

by distillation. The intermediate D4 could be polymerized in<br />

a controlled manner by use of acidic or basic catalysts. Wilton I.<br />

Patnode of GE <strong>and</strong> James F. Hyde of Dow-Corning made important<br />

advances in this area. Hyde’s discovery of the use of traces of potassium<br />

hydroxide as a polymerization catalyst for D4 made possible


686 / Silicones<br />

the manufacture of silicone rubber, which is the most commercially<br />

valuable of all the silicones.<br />

Impact<br />

Although Kipping’s discovery <strong>and</strong> naming of the silicones occurred<br />

from 1901 to 1904, the practical use <strong>and</strong> impact of silicones<br />

started in 1940, with Rochow’s discovery of direct synthesis.<br />

Production of silicones in the United States came rapidly enough<br />

to permit them to have some influence on military supplies for<br />

World War II (1939-1945). In aircraft communication equipment, extensive<br />

waterproofing of parts by silicones resulted in greater reliability<br />

of the radios under tropical conditions of humidity, where<br />

condensing water could be destructive. Silicone rubber, because<br />

of its ability to withst<strong>and</strong> heat, was used in gaskets under hightemperature<br />

conditions, in searchlights, <strong>and</strong> in the engines on B-29<br />

bombers. Silicone grease applied to aircraft engines also helped to<br />

protect spark plugs from moisture <strong>and</strong> promote easier starting.<br />

After World War II, the uses for silicones multiplied. Silicone rubber<br />

appeared in many products from caulking compounds to wire insulation<br />

to breast implants for cosmetic surgery. Silicone rubber boots were<br />

used on the moon walks where ordinary rubber would have failed.<br />

Silicones in their present form owe much to years of patient developmental<br />

work in industrial laboratories. Basic research, such as<br />

that conducted by Kipping <strong>and</strong> others, served to point the way <strong>and</strong><br />

catalyzed the process of commercialization.<br />

See also Buna rubber; Neoprene; Nylon; Plastic; Polystyrene; Teflon.<br />

Further Reading<br />

Clarson, Stephen J. Silicones <strong>and</strong> Silicone-Modified Materials. Washington,<br />

D.C.: American Chemical Society, 2000.<br />

Koerner, G. Silicones, Chemistry <strong>and</strong> Technology. Boca Raton, Fla.:<br />

CRC Press, 1991.<br />

Potter, Michael, <strong>and</strong> Noel R. Rose. Immunology of Silicones. New<br />

York: Springer, 1996.<br />

Smith, A. Lee. The Analytical Chemistry of Silicones. New York: Wiley,<br />

1991.


Solar thermal engine<br />

Solar thermal engine<br />

The invention: The first commercially practical plant for generating<br />

electricity from solar energy.<br />

The people behind the invention:<br />

Frank Shuman (1862-1918), an American inventor<br />

John Ericsson (1803-1889), an American engineer<br />

Augustin Mouchout (1825-1911), a French physics professor<br />

Power from the Sun<br />

687<br />

According to tradition, the Greek scholar Archimedes used<br />

reflective mirrors to concentrate the rays of the Sun <strong>and</strong> set afire<br />

the ships of an attacking Roman fleet in 212 b.c.e. The story illustrates<br />

the long tradition of using mirrors to concentrate solar energy<br />

from a large area onto a small one, producing very high<br />

temperatures.<br />

With the backing of Napoleon III, the Frenchman Augustin<br />

Mouchout built, between 1864 <strong>and</strong> 1872, several steam engines<br />

that were powered by the Sun. Mirrors concentrated the sun’s rays<br />

to a point, producing a temperature that would boil water. The<br />

steam drove an engine that operated a water pump. The largest engine<br />

had a cone-shaped collector, or “axicon,” lined with silverplated<br />

metal. The French government operated the engine for six<br />

months but decided it was too expensive to be practical.<br />

John Ericsson, the American famous for designing <strong>and</strong> building<br />

the Civil War ironclad ship Monitor, built seven steam-driven<br />

solar engines between 1871 <strong>and</strong> 1878. In Ericsson’s design,<br />

rays were focused onto a line rather than a point. Long mirrors,<br />

curved into a parabolic shape, tracked the Sun. The rays were focused<br />

onto a water-filled tube mounted above the reflectors to<br />

produce steam. The engineer’s largest engine, which used an 11- ×<br />

16-foot trough-shaped mirror, delivered nearly 2 horsepower. Because<br />

his solar engines were ten times more expensive than conventional<br />

steam engines, Ericsson converted them to run on coal to<br />

avoid financial loss.


688 / Solar thermal engine<br />

Frank Shuman, a well-known inventor in Philadelphia, Pennsylvania,<br />

entered the field of solar energy in 1906. The self-taught engineer<br />

believed that curved, movable mirrors were too expensive. His<br />

first large solar engine was a hot-box, or flat-plate, collector. It lay<br />

flat on the ground <strong>and</strong> had blackened pipes filled with a liquid that<br />

had a low boiling point. The solar-heated vapor ran a 3.5-horsepower<br />

engine.<br />

Shuman’s wealthy investors formed the Sun Power Company to<br />

develop <strong>and</strong> construct the largest solar plant ever built. The site chosen<br />

was in Egypt, but the plant was built near Shuman’s home for<br />

testing before it was sent to Egypt.<br />

When the inventor added ordinary flat mirrors to reflect more<br />

sunlight into each collector, he doubled the heat production of the<br />

collectors. The 572 trough-type collectors were assembled in twentysix<br />

rows. Water was piped through the troughs <strong>and</strong> converted to<br />

steam. A condenser converted the steam to water, which reentered<br />

the collectors. The engine pumped 3,000 gallons of water per minute<br />

<strong>and</strong> produced 14 horsepower per day; performance was expected to<br />

improve 25 percent in the sunny climate of Egypt.<br />

British investors requested that professor C. V. Boys review the<br />

solar plant before it was shipped to Egypt. Boys pointed out that the<br />

bottom of each collector was not receiving any direct solar energy;<br />

in fact, heat was being lost through the bottom. He suggested that<br />

each row of flat mirrors be replaced by a single parabolic reflector,<br />

<strong>and</strong> Shuman agreed. Shuman thought Boys’s idea was original, but<br />

he later realized it was based on Ericsson’s design.<br />

The company finally constructed the improved plant in Meadi,<br />

Egypt, a farming district on the Nile River. Five solar collectors,<br />

spaced 25 feet apart, were built in a north-south line. Each was<br />

about 200 feet long <strong>and</strong> 10 feet wide. Trough-shaped reflectors were<br />

made of mirrors held in place by brass springs that exp<strong>and</strong>ed<br />

<strong>and</strong> contracted with changing temperatures. The parabolic mirrors<br />

shifted automatically so that the rays were always focused on the<br />

boiler. Inside the 15-inch boiler that ran down the middle of the collector,<br />

water was heated <strong>and</strong> converted to steam. The engine produced<br />

more than 55 horsepower, which was enough to pump 6,000<br />

gallons of water per minute.<br />

The purchase price of Shuman’s solar plant was twice as high as


Solar thermal engine / 689<br />

Trough-shaped collectors with flat mirrors (above) produced enough solar thermal energy to<br />

pump 3,000 gallons of water per minute. Trough-shaped collectors with parabolic mirrors<br />

(below) produced enough solar thermal energy to pump 6,000 gallons of water per minute.<br />

that of a coal-fired plant, but its operating costs were far lower. In<br />

Egypt, where coal was expensive, the entire purchase price would<br />

be recouped in four years. Afterward, the plant would operate for<br />

practically nothing. The first practical solar engine was now in operation,<br />

providing enough energy to drive a large-scale irrigation system<br />

in the floodplain of the Nile River.<br />

By 1914, Shuman’s work was enthusiastically supported, <strong>and</strong> solar<br />

plants were planned for India <strong>and</strong> Africa. Shuman hoped to<br />

build 20,000 reflectors in the Sahara Desert <strong>and</strong> generate energy<br />

equal to all the coal mined in one year, but the outbreak of World


690 / Solar thermal engine<br />

War I ended his dreams of large-scale solar developments. The<br />

Meadi project was ab<strong>and</strong>oned in 1915, <strong>and</strong> Shuman died before the<br />

war ended. Powerful nations lost interest in solar power <strong>and</strong> began<br />

to replace coal with oil. Rich oil reserves were discovered in many<br />

desert zones that were ideal locations for solar power.<br />

Impact<br />

Although World War I ended Frank Shuman’s career, his breakthrough<br />

proved to the world that solar power held great promise for<br />

the future. His ideas were revived in 1957, when the Soviet Union<br />

planned a huge solar project for Siberia. A large boiler was fixed on<br />

a platform 140 feet high. Parabolic mirrors, mounted on 1,300 railroad<br />

cars, revolved on circular tracks to focus light on the boiler. The<br />

full-scale model was never built, but the design inspired the solar<br />

power tower.<br />

In the Mojave desert near Barstow, California, an experimental<br />

power tower, Solar One, began operation in 1982. The system collects<br />

solar energy to deliver steam to turbines that produce electric<br />

power. The 30-story tower is surrounded by more than 1,800 mirrors<br />

that adjust continually to track the Sun. Solar One generates<br />

about 10 megawatts per day, enough power for 5,000 people.<br />

Solar One was expensive, but future power towers will generate<br />

electricity as cheaply as fossil fuels can. If the costs of the air <strong>and</strong><br />

water pollution caused by coal burning were considered, solar power<br />

plants would already be recognized as cost effective. Meanwhile,<br />

Frank Shuman’s success in establishing <strong>and</strong> operating a thoroughly<br />

practical large-scale solar engine continues to inspire research <strong>and</strong><br />

development.<br />

See also Compressed-air-accumulating power plant; Fuel cell;<br />

Geothermal power; Nuclear power plant; Photoelectric cell; Photovoltaic<br />

cell; Tidal power plant.<br />

Further Reading<br />

De Kay, James T. Monitor: The Story of the Legendary Civil War Ironclad<br />

<strong>and</strong> the Man Whose Invention Changed the Course of History. New<br />

York: Ballantine, 1999.


Solar thermal engine / 691<br />

Mancini, Thomas R., James M. Chavez, <strong>and</strong> Gregory J. Kolb. “Solar<br />

Thermal Power Today <strong>and</strong> Tomorrow.” Mechanical Engineering<br />

116, no. 8 (August, 1994).<br />

Moore, Cameron M. “Cooking Up Electricity with Sunlight.” The<br />

World & I 12, no. 7 (July, 1997).<br />

Parrish, Michael. “Enron Makes Electrifying Proposal Energy: The<br />

Respected Developer Announces a Huge Solar Plant <strong>and</strong> a Breakthrough<br />

Price.” Los Angeles Times (November 5, 1994).


692<br />

Sonar<br />

Sonar<br />

The invention: A device that detects soundwaves transmitted<br />

through water, sonar was originally developed to detect enemy<br />

submarines but is also used in navigation, fish location, <strong>and</strong><br />

ocean mapping.<br />

The people behind the invention:<br />

Jacques Curie (1855-1941), a French physicist<br />

Pierre Curie (1859-1906), a French physicist<br />

Paul Langévin (1872-1946), a French physicist<br />

Active Sonar, Submarines, <strong>and</strong> Piezoelectricity<br />

Sonar, which st<strong>and</strong>s for sound navigation <strong>and</strong> ranging, is the<br />

American name for a device that the British call “asdic.” There are<br />

two types of sonar. Active sonar, the more widely used of the two<br />

types, detects <strong>and</strong> locates underwater objects when those objects reflect<br />

sound pulses sent out by the sonar. Passive sonar merely listens<br />

for sounds made by underwater objects. Passive sonar is used<br />

mostly when the loud signals produced by active sonar cannot be<br />

used (for example, in submarines).<br />

The invention of active sonar was the result of American, British,<br />

<strong>and</strong> French efforts, although it is often credited to Paul Langévin,<br />

who built the first working active sonar system by 1917. Langévin’s<br />

original reason for developing sonar was to locate icebergs, but the<br />

horrors of German submarine warfare in World War I led to the new<br />

goal of submarine detection. Both Langévin’s short-range system<br />

<strong>and</strong> long-range modern sonar depend on the phenomenon of “piezoelectricity,”<br />

which was discovered by Pierre <strong>and</strong> Jacques Curie in<br />

1880. (Piezoelectricity is electricity that is produced by certain materials,<br />

such as certain crystals, when they are subjected to pressure.)<br />

Since its invention, active sonar has been improved <strong>and</strong> its capabilities<br />

have been increased. Active sonar systems are used to detect<br />

submarines, to navigate safely, to locate schools of fish, <strong>and</strong> to map<br />

the oceans.


Sonar Theory, Development, <strong>and</strong> Use<br />

Sonar / 693<br />

Although active sonar had been developed by 1917, it was not<br />

available for military use until World War II. An interesting major<br />

use of sonar before that time was measuring the depth of the ocean.<br />

That use began when the 1922 German Meteor Oceanographic Expedition<br />

was equipped with an active sonar system. The system<br />

was to be used to help pay German World War I debts by aiding in<br />

the recovery of gold from wrecked vessels. It was not used successfully<br />

to recover treasure, but the expedition’s use of sonar to determine<br />

ocean depth led to the discovery of the Mid-Atlantic Ridge.<br />

This development revolutionized underwater geology.<br />

Active sonar operates by sending out sound pulses, often called<br />

“pings,” that travel through water <strong>and</strong> are reflected as echoes when<br />

they strike large objects. Echoes from these targets are received by<br />

the system, amplified, <strong>and</strong> interpreted. Sound is used instead of<br />

light or radar because its absorption by water is much lower. The<br />

time that passes between ping transmission <strong>and</strong> the return of an<br />

echo is used to identify the distance of a target from the system by<br />

means of a method called “echo ranging.” The basis for echo ranging<br />

is the normal speed of sound in seawater (5,000 feet per second).<br />

The distance of the target from the radar system is calculated by<br />

means of a simple equation: range = speed of sound × 0.5 elapsed<br />

time. The time is divided in half because it is made up of the time<br />

taken to reach the target <strong>and</strong> the time taken to return.<br />

The ability of active sonar to show detail increases as the energy<br />

of transmitted sound pulses is raised by decreasing the<br />

sound wavelength. Figuring out active sonar data is complicated<br />

by many factors. These include the roughness of the ocean, which<br />

scatters sound <strong>and</strong> causes the strength of echoes to vary, making<br />

it hard to estimate the size <strong>and</strong> identity of a target; the speed of<br />

the sound wave, which changes in accordance with variations in<br />

water temperature, pressure, <strong>and</strong> saltiness; <strong>and</strong> noise caused by<br />

waves, sea animals, <strong>and</strong> ships, which limits the range of active sonar<br />

systems.<br />

A simple active pulse sonar system produces a piezoelectric signal<br />

of a given frequency <strong>and</strong> time duration. Then, the signal is amplified<br />

<strong>and</strong> turned into sound, which enters the water. Any echo


694 / Sonar<br />

that is produced returns to the system to be amplified <strong>and</strong> used to<br />

determine the identity <strong>and</strong> distance of the target.<br />

Most active sonar systems are mounted near surface vessel keels<br />

or on submarine hulls in one of three ways. The first <strong>and</strong> most popular<br />

mounting method permits vertical rotation <strong>and</strong> scanning of a<br />

section of the ocean whose center is the system’s location. The second<br />

method, which is most often used in depth sounders, directs<br />

the beam downward in order to measure ocean depth. The third<br />

method, called wide scanning, involves the use of two sonar systems,<br />

one mounted on each side of the vessel, in such a way that the<br />

two beams that are produced scan the whole ocean at right angles to<br />

the direction of the vessel’s movement.<br />

Active single-beam sonar operation applies an alternating voltage<br />

to a piezoelectric crystal, making it part of an underwater loudspeaker<br />

(transducer) that creates a sound beam of a particular frequency.<br />

When an echo returns, the system becomes an underwater<br />

microphone (receiver) that identifies the target <strong>and</strong> determines its<br />

range. The sound frequency that is used is determined by the sonar’s<br />

Sonar<br />

Active sonar detects <strong>and</strong> locates underwater objects that reflect sound pulses sent out by the<br />

sonar.


purpose <strong>and</strong> the fact that the absorption of sound by water increases<br />

with frequency. For example, long-range submarine-seeking sonar<br />

systems (whose detection range is about ten miles) operate at 3 to 40<br />

kilohertz. In contrast, short-range systems that work at about 500 feet<br />

(in mine sweepers, for example) use 150 kilohertz to 2 megahertz.<br />

Impact<br />

Paul Langévin<br />

If he had not published the Special Theory of Relativity in<br />

1905, Albert Einstein once said, Paul Langévin would have<br />

done so not long afterward. Born in Paris in 1872, Langévin was<br />

among the foremost physicists of his generation. He studied in<br />

the best French schools of science—<strong>and</strong> with such teachers as<br />

Pierre Curie <strong>and</strong> Jean Perrin—<strong>and</strong> became a professor of physics<br />

at the College de France in 1904. He moved to the Sorbonne<br />

in 1909.<br />

Langévin’s research was always widely influential. In addition<br />

to his invention of active sonar, he was especially noted for<br />

his studies of the molecular structure of gases, analysis of secondary<br />

X rays from irradiated metals, his theory of magnetism,<br />

<strong>and</strong> work on piezoelectricity <strong>and</strong> piezoceramics. His suggestion<br />

that magnetic properties are linked to the valence electrons of atoms<br />

inspired Niels Bohr’s classic model of the atom. In his later<br />

career, a champion of Einstein’s theories of relativity, Langévin<br />

worked on the implications of the space-time continuum.<br />

During World War II, Langévin, a pacifist, publicly denounced<br />

the Nazis <strong>and</strong> their occupation of France. They jailed him for it.<br />

He escaped to Switzerl<strong>and</strong> in 1944, returning as soon as France<br />

was liberated. He died in late 1946.<br />

Sonar / 695<br />

Modern active sonar has affected military <strong>and</strong> nonmilitary activities<br />

ranging from submarine location to undersea mapping <strong>and</strong><br />

fish location. In all these uses, two very important goals have been<br />

to increase the ability of sonar to identify a target <strong>and</strong> to increase the<br />

effective range of sonar. Much work related to these two goals has<br />

involved the development of new piezoelectric materials <strong>and</strong> the replacement<br />

of natural minerals (such as quartz) with synthetic piezoelectric<br />

ceramics.


696 / Sonar<br />

Efforts have also been made to redesign the organization of sonar<br />

systems. One very useful development has been changing beammaking<br />

transducers from one-beam units to multibeam modules<br />

made of many small piezoelectric elements. Systems that incorporate<br />

these developments have many advantages, particularly the ability<br />

to search simultaneously in many directions. In addition, systems<br />

have been redesigned to be able to scan many echo beams simultaneously<br />

with electronic scanners that feed into a central receiver.<br />

These changes, along with computer-aided tracking <strong>and</strong> target<br />

classification, have led to the development of greatly improved active<br />

sonar systems. It is expected that sonar systems will become<br />

even more powerful in the future, finding uses that have not yet<br />

been imagined.<br />

See also Aqualung; Bathyscaphe; Bathysphere; Geiger counter;<br />

Gyrocompass; Radar; Richter scale; Ultrasound.<br />

Further Reading<br />

Curie, Marie. Pierre Curie. New York: Dover <strong>Public</strong>ations, 1923.<br />

Hackmann, Willem Dirk. Seek <strong>and</strong> Strike: Sonar, Anti-Submarine Warfare,<br />

<strong>and</strong> the Royal Navy, 1914-54. London: H.M.S.O., 1984.<br />

Segrè, Emilio. From X-Rays to Quarks: Modern Physicists <strong>and</strong> Their<br />

Discoveries. San Francisco: W. H. Freeman, 1980.<br />

Senior, John E. Marie <strong>and</strong> Pierre Curie. Gloucestershire: Sutton, 1998.


Stealth aircraft<br />

Stealth aircraft<br />

The invention: The first generation of “radar-invisible” aircraft,<br />

stealth planes were designed to elude enemy radar systems.<br />

The people behind the invention:<br />

Lockhead Corporation, an American research <strong>and</strong> development firm<br />

Northrop Corporation, an American aerospace firm<br />

Radar<br />

During World War II, two weapons were developed that radically<br />

altered the thinking of the U.S. military-industrial establishment<br />

<strong>and</strong> the composition of U.S. military forces. These weapons<br />

were the atomic bombs that were dropped on the Japanese cities of<br />

Hiroshima <strong>and</strong> Nagasaki by U.S. forces <strong>and</strong> “radio detection <strong>and</strong><br />

ranging,” or radar. Radar saved the English during the Battle of Britain,<br />

<strong>and</strong> it was radar that made it necessary to rethink aircraft design.<br />

With radar, attacking aircraft can be detected hundreds of<br />

miles from their intended targets, which makes it possible for those<br />

aircraft to be intercepted before they can attack. During World<br />

War II, radar, using microwaves, was able to relay the number, distance,<br />

direction, <strong>and</strong> speed of German aircraft to British fighter interceptors.<br />

This development allowed the fighter pilots of the Royal<br />

Air Force, “the few” who were so highly praised by Winston Churchill,<br />

to shoot down four times as many planes as they lost.<br />

Because of the development of radar, American airplane design<br />

strategy has been to reduce the planes’ cross sections, reduce or<br />

eliminate the use of metal by replacing it with composite materials,<br />

<strong>and</strong> eliminate the angles that are found on most aircraft control surfaces.<br />

These actions help make aircraft less visible—<strong>and</strong> in some<br />

cases, almost invisible—to radar. The Lockheed F-117A Nightrider<br />

<strong>and</strong> the Northrop B-2 Stealth Bomber are the results of these efforts.<br />

Airborne “Ninjas”<br />

697<br />

Hidden inside Lockheed Corporation is a research <strong>and</strong> development<br />

organization that is unique in the corporate world. This


698 / Stealth aircraft<br />

facility has provided the Air Force with the Sidewinder heatseeking<br />

missile; the SR-71, a titanium-skinned aircraft that can fly<br />

at four times the speed of sound; <strong>and</strong>, most recently, the F-117A<br />

Nightrider. The Nightrider eluded Iraqi radar so effectively during<br />

the 1991 Persian Gulf War that the Iraqis nicknamed it Shaba,<br />

which is an Arabic word that means ghost. In an unusual move<br />

for military projects, the Nightrider was delivered to the Air<br />

Force in 1982, before the plane had been perfected. This was done<br />

so that Air Force pilots could test fly the plane <strong>and</strong> provide input<br />

that could be used to improve the aircraft before it went into full<br />

production.<br />

The Northrop B-2 Stealth Bomber was the result of a design philosophy<br />

that was completely different from that of the F-117A<br />

Nightrider. The F-117A, for example, has a very angular appearance,<br />

but the angles are all greater than 180 degrees. This configuration<br />

spreads out radar waves rather than allowing them to be concentrated<br />

<strong>and</strong> sent back to their point of origin. The B-2, however,<br />

stays away from angles entirely, opting for a smooth surface that<br />

also acts to spread out the radar energy. (The B-2 so closely resembles<br />

the YB-49 Flying Wing, which was developed in the late 1940’s,<br />

that it even has the same wingspan.) The surface of the aircraft is<br />

covered with radar-absorbing material <strong>and</strong> carries its engines <strong>and</strong><br />

weapons inside to reduce the radar cross section. There are no vertical<br />

control surfaces, which has the disadvantage of making the aircraft<br />

unstable, so the stabilizing system uses computers to make<br />

small adjustments in the control elements on the trailing edges of<br />

the wings, thus increasing the craft’s stability.<br />

The F-117A Nightrider <strong>and</strong> the B-2 Stealth Bomber are the “ninjas”<br />

of military aviation. Capable of striking powerfully, rapidly,<br />

<strong>and</strong> invisibly, these aircraft added a dimension to the U.S. Air Force<br />

that did not exist previously. Before the advent of these aircraft, missions<br />

that required radar-avoidance tactics had to be flown below<br />

the horizon of ground-based radar, which is 30.5 meters above the<br />

ground. Such low-altitude flight is dangerous because of both the<br />

increased difficulty of maneuvering <strong>and</strong> vulnerability to ground<br />

fire. Additionally, such flying does not conceal the aircraft from the<br />

airborne radar carried by such craft as the American E-3A AWACS<br />

<strong>and</strong> the former Soviet Mainstay. In a major conflict, the only aircraft


that could effectively penetrate enemy airspace would be the Nightrider<br />

<strong>and</strong> the B-2.<br />

The purpose of the B-2 was to carry nuclear weapons into hostile<br />

airspace undetected. With the demise of the Soviet Union, mainl<strong>and</strong><br />

China seemed the only remaining major nuclear threat. For this reason,<br />

many defense experts believed that there was no longer a need<br />

for two radar-invisible planes, <strong>and</strong> cuts in U.S. military expenditures<br />

threatened the B-2 program during the early 1990’s.<br />

Consequences<br />

The development of the Nightrider <strong>and</strong> the B-2 meant that the<br />

former Soviet Union would have had to spend at least $60 billion to<br />

upgrade its air defense forces to meet the challenge offered by these<br />

aircraft. This fact, combined with the evolution of the Strategic Defense<br />

Initiative, commonly called “Star Wars,” led to the United<br />

States’ victory in the arms race. Additionally, stealth technology has<br />

found its way onto the conventional battlefield.<br />

As was shown in 1991 during the Desert Storm campaign in Iraq,<br />

targets that have strategic importance are often surrounded by a<br />

network of anti-air missiles <strong>and</strong> gun emplacements. During the<br />

Desert Storm air war, the F-117A was the only Allied aircraft to be<br />

assigned to targets in Baghdad. Nightriders destroyed more than 47<br />

percent of the strategic areas that were targeted, <strong>and</strong> every pilot <strong>and</strong><br />

plane returned to base unscathed.<br />

Since the world appears to be moving away from superpower<br />

conflicts <strong>and</strong> toward smaller regional conflicts, stealth aircraft may<br />

come to be used more for surveillance than for air attacks. This is<br />

particularly true because the SR-71, which previously played the<br />

primary role in surveillance, has been retired from service.<br />

See also Airplane; Cruise missile; Hydrogen bomb; Radar;<br />

Rocket; Turbojet; V-2 rocket.<br />

Further Reading<br />

Stealth aircraft / 699<br />

Chun, Clayton K. S. The Lockheed F-117A. Santa Monica, Calif.: R<strong>and</strong>,<br />

1991.


700 / Stealth aircraft<br />

Goodall, James C. America’s Stealth Fighters <strong>and</strong> Bombers. Osceola,<br />

Wis.: Motorbooks, 1992.<br />

Pape, Garry R., <strong>and</strong> John M. Campbell. Northrop Flying Wings: A History<br />

of Jack Northrop’s Visionary Aircraft. Atglen, Pa.: Schiffer, 1995.<br />

Thornborough, Anthony M. Stealth. London: Ian Allen, 1991.


Steelmaking process<br />

Steelmaking process<br />

The invention: Known as the basic oxygen, or L-D, process, a<br />

method for producing steel that worked about twelve times<br />

faster than earlier methods.<br />

The people behind the invention:<br />

Henry Bessemer (1813-1898), the English inventor of a process<br />

for making steel from iron<br />

Robert Durrer (1890-1978), a Swiss scientist who first proved<br />

the workability of the oxygen process in a laboratory<br />

F. A. Loosley (1891-1966), head of research <strong>and</strong> development at<br />

Dofasco Steel in Canada<br />

Theodor Suess (1894-1956), works manager at Voest<br />

Ferrous Metal<br />

701<br />

The modern industrial world is built on ferrous metal. Until<br />

1857, ferrous metal meant cast iron <strong>and</strong> wrought iron, though a few<br />

specialty uses of steel, especially for cutlery <strong>and</strong> swords, had existed<br />

for centuries. In 1857, Henry Bessemer developed the first largescale<br />

method of making steel, the Bessemer converter. By the 1880’s,<br />

modification of his concepts (particularly the development of a ‘’basic”<br />

process that could h<strong>and</strong>le ores high in phosphor) had made<br />

large-scale production of steel possible.<br />

Bessemer’s invention depended on the use of ordinary air, infused<br />

into the molten metal, to burn off excess carbon. Bessemer himself<br />

had recognized that if it had been possible to use pure oxygen instead<br />

of air, oxidation of the carbon would be far more efficient <strong>and</strong> rapid.<br />

Pure oxygen was not available in Bessemer’s day, except at very high<br />

prices, so steel producers settled for what was readily available, ordinary<br />

air. In 1929, however, the Linde-Frakl process for separating the<br />

oxygen in air from the other elements was discovered, <strong>and</strong> for the<br />

first time inexpensive oxygen became available.<br />

Nearly twenty years elapsed before the ready availability of pure<br />

oxygen was applied to refining the method of making steel. The first<br />

experiments were carried out in Switzerl<strong>and</strong> by Robert Durrer. In


702 / Steelmaking process<br />

1949, he succeeded in making steel expeditiously in a laboratory setting<br />

through the use of a blast of pure oxygen. Switzerl<strong>and</strong>, however,<br />

had no large-scale metallurgical industry, so the Swiss turned<br />

the idea over to the Austrians, who for centuries had exploited the<br />

large deposits of iron ore in a mountain in central Austria. Theodor<br />

Suess, the works manager of the state-owned Austrian steel complex,<br />

Voest, instituted some pilot projects. The results were sufficiently<br />

favorable to induce Voest to authorize construction of production<br />

converters. In 1952, the first ‘’heat” (as a batch of steel is<br />

called) was “blown in,” at the Voest works in Linz. The following<br />

year, another converter was put into production at the works in<br />

Donauwitz. These two initial locations led to the basic oxygen process<br />

sometimes being referred to as the L-D process.<br />

The L-D Process<br />

The basic oxygen, or L-D, process makes use of a converter similar<br />

to the Bessemer converter. Unlike the Bessemer, however, the L-<br />

D converter blows pure oxygen into the molten metal from above<br />

through a water-cooled injector known as a lance. The oxygen burns<br />

off the excess carbon rapidly, <strong>and</strong> the molten metal can then be<br />

poured off into ingots, which can later be reheated <strong>and</strong> formed into<br />

the ultimately desired shape. The great advantage of the process is<br />

the speed with which a “heat” reaches the desirable metallurgical<br />

composition for steel, with a carbon content between 0.1 percent<br />

<strong>and</strong> 2 percent. The basic oxygen process requires about forty minutes.<br />

In contrast, the prevailing method of making steel, using an<br />

open-hearth furnace (which transferred the technique from the<br />

closed Bessemer converter to an open-burning furnace to which the<br />

necessary additives could be introduced by h<strong>and</strong>) requires eight to<br />

eleven hours for a “heat” or batch.<br />

The L-D process was not without its drawbacks, however. It was<br />

adopted by the Austrians because, by carefully calibrating the timing<br />

<strong>and</strong> amount of oxygen introduced, they could turn their moderately<br />

phosphoric ore into steel without further intervention. The<br />

process required ore of a st<strong>and</strong>ardized metallurgical, or chemical,<br />

content, for which the lancing had been calculated. It produced a<br />

large amount of iron-oxide dust that polluted the surrounding at-


Steelmaking process / 703<br />

mosphere, <strong>and</strong> it required a lining in the converter of dolomitic<br />

brick. The specific chemical content of the brick contributed to the<br />

chemical mixture that produced the desired result.<br />

The Austrians quickly realized that the process was an improvement.<br />

In May, 1952, the patent specifications for the new process<br />

were turned over to a new company, Brassert Oxygen Technik, or<br />

BOT, which filed patent applications around the world. BOT embarked<br />

on an aggressive marketing campaign, bringing potential<br />

customers to Austria to observe the process in action. Despite BOT’s<br />

efforts, the new process was slow to catch on, even though in 1953<br />

BOT licensed a U.S. firm, Kaiser Engineers, to spread the process in<br />

the United States.<br />

Many factors serve to explain the reluctance of steel producers<br />

around the world to adopt the new process. One of these was the<br />

large investment most major steel producers had in their openhearth<br />

furnaces. Another was uncertainty about the pollution factor.<br />

Later, special pollution-control equipment would be developed<br />

to deal with this problem. A third concern was whether the necessary<br />

refractory liners for the new converters would be available. A<br />

fourth was the fact that the new process could h<strong>and</strong>le a load that<br />

contained no more than 30 percent scrap, preferably less. In practice,<br />

therefore, it would only work where a blast furnace smelting<br />

ore was already set up.<br />

One of the earliest firms to show serious interest in the new technology<br />

was Dofasco, a Canadian steel producer. Between 1952 <strong>and</strong><br />

1954, Dofasco, pushed by its head of research <strong>and</strong> development, F.<br />

A. Loosley, built pilot operations to test the methodology. The results<br />

were sufficiently promising that in 1954 Dofasco built the first<br />

basic oxygen furnace outside Austria. Dofasco had recently built its<br />

own blast furnace, so it had ore available on site. It was able to devise<br />

ways of dealing with the pollution problem, <strong>and</strong> it found refractory<br />

liners that would work. It became the first North American<br />

producer of basic oxygen steel.<br />

Having bought the licensing rights in 1953, Kaiser Engineers was<br />

looking for a U.S. steel producer adventuresome enough to invest in<br />

the new technology. It found that producer in McLouth Steel, a<br />

small steel plant in Detroit, Michigan. Kaiser Engineers supplied<br />

much of the technical advice that enabled McLouth to build the first


704 / Steelmaking process<br />

U.S. basic oxygen steel facility, though McLouth also sent one of its<br />

engineers to Europe to observe the Austrian operations. McLouth,<br />

which had backing from General Motors, also made use of technical<br />

descriptions in the literature.<br />

The Specifications Question<br />

Henry Bessemer<br />

Henry Bessemer was born in the small village of Charlton,<br />

Engl<strong>and</strong>, in 1813. His father was an early example of a technician,<br />

specializing in steam engines, <strong>and</strong> operated a business<br />

making metal type for printing presses. The elder Bessemer<br />

wanted his son to attend university, but Henry preferred to<br />

study under his father. During his apprenticeship, he learned<br />

the properties of alloys. At seventeen he moved to London to<br />

open his own business, which fabricated specialty metals.<br />

Three years later the Royal Academy held an exhibition of<br />

Bessemer’s work. His career, well begun, moved from one invention<br />

to another until at his death in 1898 he held 114 patents.<br />

Among them were processes for casting type <strong>and</strong> producing<br />

graphite for pencils; methods for manufacturing glass, sugar,<br />

bronze powder, <strong>and</strong> ships; <strong>and</strong> his best known creation, the Bessemer<br />

converter for making steel from iron. Bessemer built his<br />

first converter in 1855; fifteen years later Great Britain was producing<br />

half of the world’s steel.<br />

Bessemer’s life <strong>and</strong> career were models of early Industrial<br />

Age industry, prosperity, <strong>and</strong> longevity. A millionaire from patent<br />

royalties, he retired at fifty-nine, lived another twenty-six<br />

years, working on yet more inventions <strong>and</strong> cultivating astronomy<br />

as a hobby, <strong>and</strong> was married for sixty-four years. Among<br />

his many awards <strong>and</strong> honors was a knighthood, bestowed by<br />

Queen Victoria.<br />

One factor that held back adoption of basic oxygen steelmaking<br />

was the question of specifications. Many major engineering projects<br />

came with precise specifications detailing the type of steel to be<br />

used <strong>and</strong> even the method of its manufacture. Until basic oxygen<br />

steel was recognized as an acceptable form by the engineering fra-


Steelmaking process / 705<br />

ternity, so that job specifications included it as appropriate in specific<br />

applications, it could not find large-scale markets. It took a<br />

number of years for engineers to modify their specifications so that<br />

basic oxygen steel could be used.<br />

The next major conversion to the new steelmaking process occurred<br />

in Japan. The Japanese had learned of the process early,<br />

while Japanese metallurgical engineers were touring Europe in<br />

1951. Some of them stopped off at the Voest works to look at the pilot<br />

projects there, <strong>and</strong> they talked with the Swiss inventor, Robert<br />

Durrer. These engineers carried knowledge of the new technique<br />

back to Japan. In 1957 <strong>and</strong> 1958, Yawata Steel <strong>and</strong> Nippon Kokan,<br />

the largest <strong>and</strong> third-largest steel producers in Japan, decided to implement<br />

the basic oxygen process. An important contributor to this<br />

decision was the Ministry of International Trade <strong>and</strong> Industry, which<br />

brokered a licensing arrangement through Nippon Kokan, which in<br />

turn had signed a one-time payment arrangement with BOT. The<br />

licensing arrangement allowed other producers besides Nippon<br />

Kokan to use the technique in Japan.<br />

The Japanese made two important technical improvements in<br />

the basic oxygen technology. They developed a multiholed lance for<br />

blowing in oxygen, thus dispersing it more effectively in the molten<br />

metal <strong>and</strong> prolonging the life of the refractory lining of the converter<br />

vessel. They also pioneered the OG process for recovering<br />

some of the gases produced in the converter. This procedure reduced<br />

the pollution generated by the basic oxygen converter.<br />

The first large American steel producer to adopt the basic oxygen<br />

process was Jones <strong>and</strong> Laughlin, which decided to implement the<br />

new process for several reasons. It had some of the oldest equipment<br />

in the American steel industry, ripe for replacement. It also<br />

had experienced significant technical difficulties at its Aliquippa<br />

plant, difficulties it was unable to solve by modifying its openhearth<br />

procedures. It therefore signed an agreement with Kaiser Engineers<br />

to build some of the new converters for Aliquippa. These<br />

converters were constructed on license from Kaiser Engineers by<br />

Pennsylvania Engineering, with the exception of the lances, which<br />

were imported from Voest in Austria. Subsequent lances, however,<br />

were built in the United States. Some of Jones <strong>and</strong> Laughlin’s production<br />

managers were sent to Dofasco for training, <strong>and</strong> technical


706 / Steelmaking process<br />

advisers were brought to the Aliquippa plant both from Kaiser Engineers<br />

<strong>and</strong> from Austria.<br />

Other European countries were somewhat slower to adopt the<br />

new process. A major cause for the delay was the necessary modification<br />

of the process to fit the high phosphoric ores available in Germany<br />

<strong>and</strong> France. Europeans also experimented with modifications<br />

of the basic oxygen technique by developing converters that revolved.<br />

These converters, known as Kaldo in Sweden <strong>and</strong> Rotor in<br />

Germany, proved in the end to have sufficient technical difficulties<br />

that they were ab<strong>and</strong>oned in favor of the st<strong>and</strong>ard basic oxygen<br />

converter. The problems they had been designed to solve could be<br />

better dealt with through modifications of the lance <strong>and</strong> through<br />

adjustments in additives.<br />

By the mid-1980’s, the basic oxygen process had spread throughout<br />

the world. Neither Japan nor the European Community was<br />

producing any steel by the older, open-hearth method. In conjunction<br />

with the electric arc furnace, fed largely on scrap metal, the basic<br />

oxygen process had transformed the steel industry of the world.<br />

Impact<br />

The basic oxygen process has significant advantages over older<br />

procedures. It does not require additional heat, whereas the openhearth<br />

technique calls for the infusion of nine to twelve gallons of<br />

fuel oil to raise the temperature of the metal to the level necessary to<br />

burn off all the excess carbon. The investment cost of the converter<br />

is about half that of an open-hearth furnace. Fewer refractories are<br />

required, less than half those needed in an open-hearth furnace.<br />

Most important of all, however, a “heat” requires less than an hour,<br />

as compared with the eight or more hours needed for a “heat” in an<br />

open-hearth furnace.<br />

There were some disadvantages to the basic oxygen process. Perhaps<br />

the most important was the limited amount of scrap that could<br />

be included in a “heat,” a maximum of 30 percent. Because the process<br />

required at least 70 percent new ore, it could be operated most<br />

effectively only in conjunction with a blast furnace. Counterbalancing<br />

this last factor was the rapid development of the electric arc<br />

furnace, which could operate with 100 percent scrap. A firm with its


Steelmaking process / 707<br />

own blast furnace could, with both an oxygen converter <strong>and</strong> an electric<br />

arc furnace, h<strong>and</strong>le the available raw material.<br />

The advantages of the basic oxygen process overrode the disadvantages.<br />

Some other new technologies combined to produce this<br />

effect. The most important of these was continuous casting. Because<br />

of the short time required for a “heat,” it was possible, if a plant had<br />

two or three converters, to synchronize output with the fill needs of<br />

a continuous caster, thus largely canceling out some of the economic<br />

drawbacks of the batch process. Continuous production, always<br />

more economical, was now possible in the basic steel industry, particularly<br />

after development of computer-controlled rolling mills.<br />

These new technologies forced major changes in the world’s steel<br />

industry. Labor requirements for the basic oxygen converter were<br />

about half those for the open-hearth furnace. The high speed of the<br />

new technology required far less manual labor but much more technical<br />

expertise. Labor requirements were significantly reduced, producing<br />

major social dislocations in steel-producing regions. This effect<br />

was magnified by the fact that dem<strong>and</strong> for steel dropped<br />

sharply in the 1970’s, further reducing the need for steelworkers.<br />

The U.S. steel industry was slower than either the Japanese or the<br />

European to convert to the basic oxygen technique. The U.S. industry<br />

generally operated with larger quantities, <strong>and</strong> it took a number<br />

of years before the basic oxygen technique was adapted to converters<br />

with an output equivalent to that of the open-hearth furnace. By<br />

the time that had happened, world steel dem<strong>and</strong> had begun to<br />

drop. U.S. companies were less profitable, failing to generate internally<br />

the capital needed for the major investment involved in<br />

ab<strong>and</strong>oning open-hearth furnaces for oxygen converters. Although<br />

union contracts enabled companies to change work assignments<br />

when new technologies were introduced, there was stiff resistance<br />

to reducing employment of steelworkers, most of whom had lived<br />

all their lives in one-industry towns. Finally, engineers at the steel<br />

firms were wedded to the old methods <strong>and</strong> reluctant to change, as<br />

were the large bureaucracies of the big U.S. steel firms.<br />

The basic oxygen technology in steel is part of a spate of new<br />

technical developments that have revolutionized industrial production,<br />

drastically reducing the role of manual labor <strong>and</strong> dramatically<br />

increasing the need for highly skilled individuals with technical ex-


708 / Steelmaking process<br />

pertise. Because capital costs are significantly lower than for alternative<br />

processes, it has allowed a number of developing countries<br />

to enter a heavy industry <strong>and</strong> compete successfully with the old industrial<br />

giants. It has thus changed the face of the steel industry.<br />

See also Assembly line; Buna rubber; Disposable razor; Laminated<br />

glass; Memory metal; Neoprene; Oil-well drill bit; Pyrex<br />

glass.<br />

Further Reading<br />

Bain, Trevor. Banking the Furnace: Restructuring of the Steel Industry in<br />

Eight Countries. Kalamazoo, Mich.: W. E. Upjohn Institute for Employment<br />

Research, 1992.<br />

Gold, Bela, Gerhard Rosegger, <strong>and</strong> Myles G. Boylan, Jr. Evaluating<br />

Technological Innovations: Methods, Expectations, <strong>and</strong> Findings. Lexington,<br />

Mass.: Lexington Books, 1980.<br />

Hall, Christopher. Steel Phoenix: The Fall <strong>and</strong> Rise of the U.S. Steel Industry.<br />

New York: St. Martin’s Press, 1997.<br />

Hoerr, John P. And the Wolf Finally Came: The Decline of the American<br />

Steel Industry. Pittsburgh, Pa.: University of Pittsburgh Press,<br />

1988.<br />

Lynn, Leonard H. How Japan Innovates: A Comparison with the United<br />

States in the Case of Oxygen Steelmaking. Boulder, Colo.: Westview<br />

Press, 1982.<br />

Seely, Burce Edsall. Iron <strong>and</strong> Steel in the Twentieth Century. New York:<br />

Facts on File, 1994.


Supercomputer<br />

Supercomputer<br />

The invention: A computer that had the greatest computational<br />

power that then existed.<br />

The person behind the invention:<br />

Seymour R. Cray (1928-1996), American computer architect <strong>and</strong><br />

designer<br />

The Need for Computing Power<br />

Although modern computers have roots in concepts first proposed<br />

in the early nineteenth century, it was only around 1950 that they became<br />

practical. Early computers enabled their users to calculate equations<br />

quickly <strong>and</strong> precisely, but it soon became clear that even more<br />

powerful computers—machines capable of receiving, computing, <strong>and</strong><br />

sending out data with great precision <strong>and</strong> at the highest speeds—<br />

would enable researchers to use computer “models,” which are programs<br />

that simulate the conditions of complex experiments.<br />

Few computer manufacturers gave much thought to building the<br />

fastest machine possible, because such an undertaking is expensive<br />

<strong>and</strong> because the business use of computers rarely dem<strong>and</strong>s the<br />

greatest processing power. The first company to build computers<br />

specifically to meet scientific <strong>and</strong> governmental research needs was<br />

Control Data Corporation (CDC). The company had been founded<br />

in 1957 by William Norris, <strong>and</strong> its young vice president for engineering<br />

was the highly respected computer engineer Seymour R.<br />

Cray. When CDC decided to limit high-performance computer design,<br />

Cray struck out on his own, starting Cray Research in 1972. His<br />

goal was to design the most powerful computer possible. To that<br />

end, he needed to choose the principles by which his machine<br />

would operate; that is, he needed to determine its architecture.<br />

The Fastest Computer<br />

709<br />

All computers rely upon certain basic elements to process data.<br />

Chief among these elements are the central processing unit, or CPU


710 / Supercomputer<br />

(which h<strong>and</strong>les data), memory (where data are stored temporarily<br />

before <strong>and</strong> after processing), <strong>and</strong> the bus (the interconnection between<br />

memory <strong>and</strong> the processor, <strong>and</strong> the means by which data are<br />

transmitted to or from other devices, such as a disk drive or a monitor).<br />

The structure of early computers was based on ideas developed<br />

by the mathematician John von Neumann, who, in the 1940’s,<br />

conceived a computer architecture in which the CPU controls all<br />

events in a sequence: It fetches data from memory, performs calculations<br />

on those data, <strong>and</strong> then stores the results in memory. Because it<br />

functions in sequential fashion, the speed of this “scalar processor”<br />

is limited by the rate at which the processor is able to complete each<br />

cycle of tasks.<br />

Before Cray produced his first supercomputer, other designers<br />

tried different approaches. One alternative was to link a vector processor<br />

to a scalar unit. A vector processor achieves its speed by performing<br />

computations on a large series of numbers (called a vector)<br />

at one time rather than in sequential fashion, though specialized<br />

<strong>and</strong> complex programs were necessary to make use of this feature.<br />

In fact, vector processing computers spent most of their time operating<br />

as traditional scalar processors <strong>and</strong> were not always efficient at<br />

switching back <strong>and</strong> forth between the two processing types.<br />

Another option chosen by Cray’s competitors was the notion of<br />

“pipelining” the processor’s tasks. A scalar processor often must<br />

wait while data are retrieved or stored in memory. Pipelining techniques<br />

allow the processor to make use of idle time for calculations<br />

in other parts of the program being run, thus increasing the effective<br />

speed. A variation on this technique is “parallel processing,” in<br />

which multiple processors are linked. If each of, for example, eight<br />

central processors is given a portion of a computing task to perform,<br />

the task will be completed more quickly than the traditional von<br />

Neumann architecture, with its single processor, would allow.<br />

Ever the pragmatist, however, Cray decided to employ proved<br />

technology rather than use advanced techniques in his first supercomputer,<br />

the Cray 1, which was introduced in 1976. Although the<br />

Cray 1 did incorporate vector processing, Cray used a simple form<br />

of vector calculation that made the technique practical <strong>and</strong> easy to<br />

use. Most striking about this computer was its shape, which was far<br />

more modern than its internal design. The Cray 1 was shaped like a


Seymour R. Cray<br />

Supercomputer / 711<br />

Seymour R. Cray was born in 1928 in Chippewa Falls, Wisconsin.<br />

The son of a civil engineer, he became interested in radio<br />

<strong>and</strong> electronics as a boy. After graduating from high school in<br />

1943, he joined the U.S. Army, was posted to Europe in an infantry<br />

communications platoon, <strong>and</strong> fought in the Battle of the<br />

Bulge. Back from the war, he pursued his interest in electronics<br />

in college while majoring in mathematics at the University of<br />

Minnesota. Upon graduation in 1950, he took a job at Engineering<br />

Research Associates. It was there that he first learned<br />

about computers. In fact, he helped design the first digital computer,<br />

UNIVAC.<br />

Cray co-founded Control Data Corporation in 1957. Based<br />

on his ideas, the company built large-scale, high-speed computers.<br />

In 1972 he founded his own company, Cray Research Incorporated,<br />

with the intention of employing new processing methods<br />

<strong>and</strong> simplifying architecture <strong>and</strong> software to build the<br />

world’s fastest computers. He succeeded, <strong>and</strong> the series of computers<br />

that the company marketed made possible computer<br />

modeling as a central part of scientific research in areas as diverse<br />

as meteorology, oil exploration, <strong>and</strong> nuclear weapons design.<br />

Through the 1970’s <strong>and</strong> 1980’s Cray Research was at the<br />

forefront of supercomputer technology, which became one of<br />

the symbols of American technological leadership.<br />

In 1989 Cray left Cray Research to form still another company,<br />

Cray Computer Corporation. He planned to build the<br />

next generation supercomputer, the Cray 5, but advances in microprocessor<br />

technology undercut the dem<strong>and</strong> for supercomputers.<br />

Cray Computer entered bankruptcy in 1995. Ayear later<br />

he died from injuries sustained in an automobile accident near<br />

Colorado Springs, Colorado.<br />

cylinder with a small section missing <strong>and</strong> a hollow center, with<br />

what appeared to be a bench surrounding it. The shape of the machine<br />

was designed to minimize the length of the interconnecting<br />

wires that ran between circuit boards to allow electricity to move the<br />

shortest possible distance. The bench concealed an important part<br />

of the cooling system that kept the system at an appropriate operating<br />

temperature.


712 / Supercomputer<br />

The measurements that describe the performance of supercomputers<br />

are called MIPS (millions of instructions per second) for scalar<br />

processors <strong>and</strong> megaflops (millions of floating-point operations per<br />

second) for vector processors. (Floating-point numbers are numbers<br />

expressed in scientific notation; for example, 10 27 .) Whereas the fastest<br />

computer before the Cray 1 was capable of some 35 MIPS, the<br />

Cray 1 was capable of 80 MIPS. Moreover, the Cray 1 was theoretically<br />

capable of vector processing at the rate of 160 megaflops, a rate<br />

unheard of at the time.<br />

Consequences<br />

Seymour Cray first estimated that there would be few buyers for<br />

a machine as advanced as the Cray 1, but his estimate turned out to<br />

be incorrect. There were many scientists who wanted to perform<br />

computer modeling (in which scientific ideas are expressed in such<br />

a way that computer-based experiments can be conducted) <strong>and</strong><br />

who needed raw processing power.<br />

When dealing with natural phenomena such as the weather or<br />

geological structures, or in rocket design, researchers need to make<br />

calculations involving large amounts of data. Before computers,<br />

advanced experimental modeling was simply not possible, since<br />

even the modest calculations for the development of atomic energy,<br />

for example, consumed days <strong>and</strong> weeks of scientists’ time.<br />

With the advent of supercomputers, however, large-scale computation<br />

of vast amounts of information became possible. Weather<br />

researchers can design a detailed program that allows them to analyze<br />

complex <strong>and</strong> seemingly unpredictable weather events such<br />

as hurricanes; geologists searching for oil fields can gather data<br />

about successful finds to help identify new ones; <strong>and</strong> spacecraft<br />

designers can “describe” in computer terms experimental ideas<br />

that are too costly or too dangerous to carry out. As supercomputer<br />

performance evolves, there is little doubt that scientists will<br />

make ever greater use of its power.<br />

See also Apple II computer; BINAC computer; Colossus computer;<br />

ENIAC computer; IBM Model 1401 computer; Personal computer;<br />

UNIVAC computer.


Further Reading<br />

Supercomputer / 713<br />

Edwards, Owen. “Seymour Cray.” Forbes 154, no. 5 (August 29,<br />

1994).<br />

Lloyd, Therese, <strong>and</strong> Stanley N. Wellborn. “Computers’ Next Frontiers.”<br />

U.S. News & World Report 99 (August 26, 1985).<br />

Slater, Robert. Portraits in Silicon. Cambridge, Mass.: MIT Press,<br />

1987.<br />

Zipper, Stuart. “Chief Exec. Leaves Cray Computer.” Electronic<br />

News 38, no. 1908 (April, 1992).


714<br />

Supersonic passenger plane<br />

Supersonic passenger plane<br />

The invention: The first commercial airliner that flies passengers at<br />

speeds in excess of the speed of sound.<br />

The people behind the invention:<br />

Sir Archibald Russell (1904- ), a designer with the British<br />

Aircraft Corporation<br />

Pierre Satre (1909- ), technical director at Sud-Aviation<br />

Julian Amery (1919- ), British minister of aviation, 1962-1964<br />

Geoffroy de Cource (1912- ), French minister of aviation,<br />

1962<br />

William T. Coleman, Jr. (1920- ), U.S. secretary of<br />

transportation, 1975-1977<br />

Birth of Supersonic Transportations<br />

On January 21, 1976, the Anglo-French Concorde became the<br />

world’s first supersonic airliner to carry passengers on scheduled<br />

commercial flights. British Airways flew a Concorde from London’s<br />

Heathrow Airport to the Persian Gulf emirate of Bahrain in<br />

three hours <strong>and</strong> thirty-eight minutes. At about the same time, Air<br />

France flew a Concorde from Paris’s Charles de Gaulle Airport to<br />

Rio de Janeiro, Brazil, in seven hours <strong>and</strong> twenty-five minutes.<br />

The Concordes’ cruising speeds were about twice the speed of<br />

sound, or 1,350 miles per hour. On May 24, 1976, the United States<br />

<strong>and</strong> Europe became linked for the first time with commercial supersonic<br />

air transportation. British Airways inaugurated flights<br />

between Dulles International Airport in Washington, D.C., <strong>and</strong><br />

Heathrow Airport. Likewise, Air France inaugurated flights between<br />

Dulles International Airport <strong>and</strong> Charles de Gaulle Airport.<br />

The London-Washington, D.C., flight was flown in an unprecedented<br />

time of three hours <strong>and</strong> forty minutes. The Paris-<br />

Washington, D.C., flight was flown in a time of three hours <strong>and</strong><br />

fifty-five minutes.


The Decision to Build the SST<br />

Supersonic passenger plane / 715<br />

Events leading to the development <strong>and</strong> production of the Anglo-<br />

French Concorde went back almost twenty years <strong>and</strong> included approximately<br />

$3 billion in investment costs. Issues surrounding the<br />

development <strong>and</strong> final production of the supersonic transport (SST)<br />

were extremely complex <strong>and</strong> at times highly emotional. The concept<br />

of developing an SST brought with it environmental concerns<br />

<strong>and</strong> questions, safety issues both in the air <strong>and</strong> on the ground, political<br />

intrigue of international proportions, <strong>and</strong> enormous economic<br />

problems from costs of operations, research, <strong>and</strong> development.<br />

In Engl<strong>and</strong>, the decision to begin the SST project was made in October,<br />

1956. Under the promotion of Morien Morgan with the Royal<br />

Aircraft Establishment in Farnborough, Engl<strong>and</strong>, it was decided at<br />

the Aviation Ministry headquarters in London to begin development<br />

of a supersonic aircraft. This decision was based on the intense competition<br />

from the American Boeing 707 <strong>and</strong> Douglas DC-8 subsonic<br />

jets going into commercial service. There was little point in developing<br />

another subsonic plane; the alternative was to go above the speed<br />

of sound. In November, 1956, at Farnborough, the first meeting of the<br />

Supersonic Transport Aircraft Committee, known as STAC, was held.<br />

Members of the STAC proposed that development costs would be<br />

in the range of $165 million to $260 million, depending on the range,<br />

speed, <strong>and</strong> payload of the chosen SST. The committee also projected<br />

that by 1970, there would be a world market for at least 150 to 500 supersonic<br />

planes. Estimates were that the supersonic plane would recover<br />

its entire research <strong>and</strong> development cost through thirty sales.<br />

The British, in order to continue development of an SST, needed a European<br />

partner as a way of sharing the costs <strong>and</strong> preempting objections<br />

to proposed funding by Engl<strong>and</strong>’s Treasury.<br />

In 1960, the British government gave the newly organized British<br />

Aircraft Corporation (BAC) $1 million for an SST feasibility study.<br />

Sir Archibald Russell, BAC’s chief supersonic designer, visited Pierre<br />

Satre, the technical director at the French firm of Sud-Aviation.<br />

Satre’s suggestion was to evolve an SST from Sud-Aviation’s highly<br />

successful subsonic Caravelle transport. By September, 1962, an<br />

agreement was reached by Sud <strong>and</strong> BAC design teams on a new<br />

SST, the Super Caravelle.


716 / Supersonic passenger plane<br />

There was a bitter battle over the choice of engines with two British<br />

engine firms, Bristol-Siddeley <strong>and</strong> Rolls-Royce, as contenders.<br />

Sir Arnold Hall, the managing director of Bristol-Siddeley, in collaboration<br />

with the French aero-engine company SNECMA, was eventually<br />

awarded the contract for the engines. The engine chosen was<br />

a “civilianized” version of the Olympus, which Bristol had been developing<br />

for the multirole TRS-2 combat plane.<br />

The Concorde Consortium<br />

On November 29, 1962, the Concorde Consortium was created<br />

by an agreement between Engl<strong>and</strong> <strong>and</strong> the French Republic, signed<br />

by Ministers of Aviation Julian Amery <strong>and</strong> Geoffroy de Cource<br />

(1912- ). The first Concorde, Model 001, rolled out from Sud-<br />

Aviation’s St. Martin-du-Touch assembly plant on December 11,<br />

1968. The second, Model 002, was completed at the British Aircraft<br />

Corporation a few months later. Eight years later, on January 21,<br />

1976, the Concorde became the world’s first supersonic airliner to<br />

carry passengers on scheduled commercial flights.<br />

Development of the SST did not come easily for the Anglo-<br />

French consortium. The nature of supersonic flight created numerous<br />

problems <strong>and</strong> uncertainties not present for subsonic flight. The<br />

SST traveled faster than the speed of sound. Sound travels at 760<br />

miles per hour at sea level at a temperature of 59 degrees Fahrenheit.<br />

This speed drops to about 660 miles per hour at sixty-five thous<strong>and</strong><br />

feet, cruising altitude for the SST, where the air temperature<br />

drops to 70 degrees below zero.<br />

The Concorde was designed to fly at a maximum of 1,450 miles<br />

per hour. The European designers could use an aluminum alloy<br />

construction <strong>and</strong> stay below the critical skin-friction temperatures<br />

that required other airframe alloys, such as titanium. The Concorde<br />

was designed with a slender curved wing surface. The design incorporated<br />

widely separated engine nacelles, each housing two Olympus<br />

593 jet engines. The Concorde was also designed with a “droop<br />

snoot,” providing three positions: the supersonic configuration, a<br />

heat-visor retracted position for subsonic flight, <strong>and</strong> a nose-lowered<br />

position for l<strong>and</strong>ing patterns.


Impact<br />

Supersonic passenger plane / 717<br />

Early SST designers were faced with questions such as the intensity<br />

<strong>and</strong> ionization effect of cosmic rays at flight altitudes of sixty to<br />

seventy thous<strong>and</strong> feet. The “cascade effect” concerned the intensification<br />

of cosmic radiation when particles from outer space struck a<br />

metallic cover. Scientists looked for ways to shield passengers from<br />

this hazard inside the aluminum or titanium shell of an SST flying<br />

high above the protective blanket of the troposphere. Experts questioned<br />

whether the risk of being struck by meteorites was any<br />

greater for the SST than for subsonic jets <strong>and</strong> looked for evidence on<br />

wind shear of jet streams in the stratosphere.<br />

Other questions concerned the strength <strong>and</strong> frequency of clear air<br />

turbulence above forty-five thous<strong>and</strong> feet, whether the higher ozone<br />

content of the air at SST cruise altitude would affect the materials of<br />

the aircraft, whether SST flights would upset or destroy the protective<br />

nature of the earth’s ozone layer, the effect of aerodynamic heating<br />

on material strength, <strong>and</strong> the tolerable strength of sonic booms<br />

over populated areas. These <strong>and</strong> other questions consumed the designers<br />

<strong>and</strong> researchers involved in developing the Concorde.<br />

Through design research <strong>and</strong> flight tests, many of the questions<br />

were resolved or realized to be less significant than anticipated. Several<br />

issues did develop into environmental, economic, <strong>and</strong> international<br />

issues. In late 1975, the British <strong>and</strong> French governments requested<br />

permission to use the Concorde at New York’s John F.<br />

Kennedy International Airport <strong>and</strong> at Dulles International Airport<br />

for scheduled flights between the United States <strong>and</strong> Europe. In December,<br />

1975, as a result of strong opposition from anti-Concorde<br />

environmental groups, the U.S. House of Representatives approved<br />

a six-month ban on SSTs coming into the United States so that the<br />

impact of flights could be studied. Secretary of Transportation William<br />

T. Coleman, Jr., held hearings to prepare for a decision by February<br />

5, 1976, as to whether to allow the Concorde into U.S. airspace.<br />

The British <strong>and</strong> French, if denied l<strong>and</strong>ing rights, threatened<br />

to take the United States to an international court, claiming that<br />

treaties had been violated.<br />

The treaties in question were the Chicago Convention <strong>and</strong> Bermuda<br />

agreements of February 11, 1946, <strong>and</strong> March 27, 1946. These


718 / Supersonic passenger plane<br />

treaties prohibited the United States from banning aircraft that both<br />

France <strong>and</strong> Great Britain had certified to be safe. The Environmental<br />

Defense Fund contended that the United States had the right to ban<br />

SST aircraft on environmental grounds.<br />

Under pressure from both sides, Coleman decided to allow limited<br />

Concorde service at Dulles <strong>and</strong> John F. Kennedy airports for a<br />

sixteen-month trial period. Service into John F. Kennedy Airport,<br />

however, was delayed by a ban by the Port Authority of New York<br />

<strong>and</strong> New Jersey until a pending suit was pursued by the airlines.<br />

During the test period, detailed records were to be kept on the<br />

Concorde’s noise levels, vibration, <strong>and</strong> engine emission levels. Other<br />

provisions included that the plane would not fly at supersonic<br />

speeds over the continental United States; that all flights could be<br />

cancelled by the United States with four months notice, or immediately<br />

if they proved harmful to the health <strong>and</strong> safety of Americans;<br />

<strong>and</strong> that at the end of a year, four months of study would begin to<br />

determine if the trial period should be extended.<br />

The Concorde’s noise was one of the primary issues in determining<br />

whether the plane should be allowed into U.S. airports. The Federal<br />

Aviation Administration measured the effective perceived noise<br />

in decibels. After three months of monitoring the Concorde’s departure<br />

noise at 3.5 nautical miles was found to vary from 105 to 130<br />

decibels. The Concorde’s approach noise at one nautical mile from<br />

threshold varied from 115 to 130 decibels. These readings were approximately<br />

equal to noise levels of other four-engine jets, such as<br />

the Boeing 747, on l<strong>and</strong>ing but were twice as loud on takeoff.<br />

The Economics of Operation<br />

Another issue of significance was the economics of Concorde’s<br />

operation <strong>and</strong> its tremendous investment costs. In 1956, early predictions<br />

of Great Britain’s STAC were for a world market of 150 to<br />

500 supersonic planes. In November, 1976, Great Britain’s Gerald<br />

Kaufman <strong>and</strong> France’s Marcel Cavaille said that production of the<br />

Concorde would not continue beyond the sixteen vehicles then contracted<br />

for with BAC <strong>and</strong> Sud-Aviation. There was no dem<strong>and</strong> by<br />

U.S. airline corporations for the plane. Given that the planes could<br />

not fly at supersonic speeds over populated areas because of the


sonic boom phenomenon, markets for the SST had to be separated<br />

by at least three thous<strong>and</strong> miles, with flight paths over mostly water<br />

or desert. Studies indicated that there were only twelve to fifteen<br />

routes in the world for which the Concorde was suitable. The planes<br />

were expensive, at a price of approximately $74 million each <strong>and</strong><br />

had a limited seating capacity of one hundred passengers. The<br />

plane’s range was about four thous<strong>and</strong> miles.<br />

These statistics compared to a Boeing 747 with a cost of $35 million,<br />

seating capacity of 360, <strong>and</strong> a range of six thous<strong>and</strong> miles. In<br />

addition, the International Air Transport Association negotiated<br />

that the fares for the Concorde flights should be equivalent to current<br />

first-class fares plus 20 percent. The marketing promotion for<br />

the Anglo-French Concorde was thus limited to the elite business<br />

traveler who considered speed over cost of transportation. Given<br />

these factors, the recovery of research <strong>and</strong> development costs for<br />

Great Britain <strong>and</strong> France would never occur.<br />

See also Airplane; Bullet train; Dirigible; Rocket; Stealth aircraft;<br />

Turbojet; V-2 rocket.<br />

Further Reading<br />

Supersonic passenger plane / 719<br />

Ellingsworth, Rosalind K. “Concorde Stresses Time, Service.” Aviation<br />

Week <strong>and</strong> Space Technology 105 (August 16, 1976).<br />

Kozicharow, Eugene. “Concorde Legal Questions Raised.” Aviation<br />

Week <strong>and</strong> Space Technology 104 (January 12, 1976).<br />

Ropelewski, Robert. “Air France Poised for Concorde Service.” Aviation<br />

Week <strong>and</strong> Space Technology 104 (January 19, 1976).<br />

Sparaco, Pierre. “Official Optimism Grows for Concorde’s Return.”<br />

Aviation Week <strong>and</strong> Space Technology 154, no. 8 (February 19, 2001).<br />

Trubshaw, Brian. Concorde: The Inside Story. Thrupp, Stroud: Sutton,<br />

2000.


720<br />

Synchrocyclotron<br />

Synchrocyclotron<br />

The invention: A powerful particle accelerator that performed<br />

better than its predecessor, the cyclotron.<br />

The people behind the invention:<br />

Edwin Mattison McMillan (1907-1991), an American physicist<br />

who won the Nobel Prize in Chemistry in 1951<br />

Vladimir Iosifovich Veksler (1907-1966), a Soviet physicist<br />

Ernest Orl<strong>and</strong>o Lawrence (1901-1958), an American physicist<br />

Hans Albrecht Bethe (1906- ), a German American physicist<br />

The First Cyclotron<br />

The synchrocyclotron is a large electromagnetic apparatus designed<br />

to accelerate atomic <strong>and</strong> subatomic particles at high energies.<br />

Therefore, it falls under the broad class of scientific devices<br />

known as “particle accelerators.” By the early 1920’s, the experimental<br />

work of physicists such as Ernest Rutherford <strong>and</strong> George<br />

Gamow dem<strong>and</strong>ed that an artificial means be developed to generate<br />

streams of atomic <strong>and</strong> subatomic particles at energies much<br />

greater than those occurring naturally. This requirement led Ernest<br />

Orl<strong>and</strong>o Lawrence to develop the cyclotron, the prototype for most<br />

modern accelerators. The synchrocyclotron was developed in response<br />

to the limitations of the early cyclotron.<br />

In September, 1930, Lawrence announced the basic principles behind<br />

the cyclotron. Ionized—that is, electrically charged—particles<br />

are admitted into the central section of a circular metal drum. Once<br />

inside the drum, the particles are exposed to an electric field alternating<br />

within a constant magnetic field. The combined action of the<br />

electric <strong>and</strong> magnetic fields accelerates the particles into a circular<br />

path, or orbit. This increases the particles’ energy <strong>and</strong> orbital radii.<br />

This process continues until the particles reach the desired energy<br />

<strong>and</strong> velocity <strong>and</strong> are extracted from the machine for use in experiments<br />

ranging from particle-to-particle collisions to the synthesis of<br />

radioactive elements.


Although Lawrence was interested in the practical applications<br />

of his invention in medicine <strong>and</strong> biology, the cyclotron also was applied<br />

to a variety of experiments in a subfield of physics called<br />

“high-energy physics.” Among the earliest applications were studies<br />

of the subatomic, or nuclear, structure of matter. The energetic<br />

particles generated by the cyclotron made possible the very type of<br />

experiment that Rutherford <strong>and</strong> Gamow had attempted earlier.<br />

These experiments, which bombarded lithium targets with streams<br />

of highly energetic accelerated protons, attempted to probe the inner<br />

structure of matter.<br />

Although funding for scientific research on a large scale was<br />

scarce before World War II (1939-1945), Lawrence nevertheless conceived<br />

of a 467-centimeter cyclotron that would generate particles<br />

with energies approaching 100 million electronvolts. By the end of<br />

the war, increases in the public <strong>and</strong> private funding of scientific research<br />

<strong>and</strong> a dem<strong>and</strong> for higher-energy particles created a situation<br />

in which this plan looked as if it would become reality, were it not<br />

for an inherent limit in the physics of cyclotron operation.<br />

Overcoming the Problem of Mass<br />

Synchrocyclotron / 721<br />

In 1937, Hans Albrecht Bethe discovered a severe theoretical limitation<br />

to the energies that could be produced in a cyclotron. Physicist<br />

Albert Einstein’s special theory of relativity had demonstrated<br />

that as any mass particle gains velocity relative to the speed of light,<br />

its mass increases. Bethe showed that this increase in mass would<br />

eventually slow the rotation of each particle. Therefore, as the rotation<br />

of each particle slows <strong>and</strong> the frequency of the alternating electric<br />

field remains constant, particle velocity will decrease eventually.<br />

This factor set an upper limit on the energies that any cyclotron<br />

could produce.<br />

Edwin Mattison McMillan, a colleague of Lawrence at Berkeley,<br />

proposed a solution to Bethe’s problem in 1945. Simultaneously <strong>and</strong><br />

independently, Vladimir Iosifovich Veksler of the Soviet Union proposed<br />

the same solution. They suggested that the frequency of the<br />

alternating electric field be slowed to meet the decreasing rotational<br />

frequencies of the accelerating particles—in essence, “synchroniz-


722 / Synchrocyclotron<br />

ing” the electric field with the moving particles. The result was the<br />

synchrocyclotron.<br />

Prior to World War II, Lawrence <strong>and</strong> his colleagues had obtained<br />

the massive electromagnet for the new 100-million-electronvolt cyclotron.<br />

This 467-centimeter magnet would become the heart of the<br />

new Berkeley synchrocyclotron. After initial tests proved successful,<br />

the Berkeley team decided that it would be reasonable to convert<br />

the cyclotron magnet for use in a new synchrocyclotron. The<br />

apparatus was operational in November of 1946.<br />

These high energies combined with economic factors to make the<br />

synchrocyclotron a major achievement for the Berkeley Radiation<br />

Laboratory. The synchrocyclotron required less voltage to produce<br />

higher energies than the cyclotron because the obstacles cited by<br />

Bethe were virtually nonexistent. In essence, the energies produced<br />

by synchrocyclotrons are limited only by the economics of building<br />

them. These factors led to the planning <strong>and</strong> construction of other<br />

synchrocyclotrons in the United States <strong>and</strong> Europe. In 1957, the<br />

Berkeley apparatus was redesigned in order to achieve energies of<br />

720 million electronvolts, at that time the record for cyclotrons of<br />

any kind.<br />

Impact<br />

Previously, scientists had had to rely on natural sources for highly<br />

energetic subatomic <strong>and</strong> atomic particles with which to experiment.<br />

In the mid-1920’s, the American physicist Robert Andrews Millikan<br />

began his experimental work in cosmic rays, which are one natural<br />

source of energetic particles called “mesons.” Mesons are charged<br />

particles that have a mass more than two hundred times that of the<br />

electron <strong>and</strong> are therefore of great benefit in high-energy physics experiments.<br />

In February of 1949, McMillan announced the first synthetically<br />

produced mesons using the synchrocyclotron.<br />

McMillan’s theoretical development led not only to the development<br />

of the synchrocyclotron but also to the development of the<br />

electron synchrotron, the proton synchrotron, the microtron, <strong>and</strong><br />

the linear accelerator. Both proton <strong>and</strong> electron synchrotrons have<br />

been used successfully to produce precise beams of muons <strong>and</strong> pimesons,<br />

or pions (a type of meson).


The increased use of accelerator apparatus ushered in a new era<br />

of physics research, which has become dominated increasingly by<br />

large accelerators <strong>and</strong>, subsequently, larger teams of scientists <strong>and</strong><br />

engineers required to run individual experiments. More sophisticated<br />

machines have generated energies in excess of 2 trillion<br />

electronvolts at the United States’ Fermi National Accelerator Laboratory,<br />

or Fermilab, in Illinois. Part of the huge Tevatron apparatus<br />

at Fermilab, which generates these particles, is a proton synchrotron,<br />

a direct descendant of McMillan <strong>and</strong> Lawrence’s early<br />

efforts.<br />

See also Atomic bomb; Cyclotron; Electron microscope; Field ion<br />

microscope; Geiger counter; Hydrogen bomb; Mass spectrograph;<br />

Neutrino detector; Scanning tunneling microscope; Tevatron accelerator.<br />

Further Reading<br />

Synchrocyclotron / 723<br />

Bernstein, Jeremy. Hans Bethe: Prophet of Energy. New York: Basic<br />

Books, 1980.<br />

McMillan, Edwin. “The Synchrotron: A Proposed High-Energy Particle<br />

Accelerator.” Physical Review 68 (September, 1945).<br />

_____. “Vladimir Iosifovich Veksler.” Physics Today (November,<br />

1966).<br />

“Witness to a Century.” Discover 20 (December, 1999).


724<br />

Synthetic amino acid<br />

Synthetic amino acid<br />

The invention: A method for synthesizing amino acids by combining<br />

water, hydrogen, methane, <strong>and</strong> ammonia <strong>and</strong> exposing the<br />

mixture to an electric spark.<br />

The people behind the invention:<br />

Stanley Lloyd Miller (1930- ), an American professor of<br />

chemistry<br />

Harold Clayton Urey (1893-1981), an American chemist who<br />

won the 1934 Nobel Prize in Chemistry<br />

Aleks<strong>and</strong>r Ivanovich Oparin (1894-1980), a Russian biochemist<br />

John Burdon S<strong>and</strong>erson Haldane (1892-1964), a British scientist<br />

Prebiological Evolution<br />

The origin of life on Earth has long been a tough problem for scientists<br />

to solve. While most scientists can envision the development<br />

of life through geologic time from simple single-cell bacteria<br />

to complex mammals by the processes of mutation <strong>and</strong> natural selection,<br />

they have found it difficult to develop a theory to define<br />

how organic materials were first formed <strong>and</strong> organized into lifeforms.<br />

This stage in the development of life before biologic systems<br />

arose, which is called “chemical evolution,” occurred between<br />

4.5 <strong>and</strong> 3.5 billion years ago. Although great advances in<br />

genetics <strong>and</strong> biochemistry have shown the intricate workings of<br />

the cell, relatively little light has been shed on the origins of this intricate<br />

machinery of the cell. Some experiments, however, have<br />

provided important data from which to build a scientific theory of<br />

the origin of life. The first of these experiments was the classic<br />

work of Stanley Lloyd Miller.<br />

Miller worked with Harold Clayton Urey, a Nobel laureate, on the<br />

environments of the early earth. John Burdon S<strong>and</strong>erson Haldane, a<br />

British biochemist, had suggested in 1929 that the earth’s early atmosphere<br />

was a reducing one—that it contained no free oxygen. In<br />

1952, Urey published a seminal work in planetology, The Planets,in<br />

which he elaborated on Haldane’s suggestion, <strong>and</strong> he postulated


that the earth had formed from a cold stellar dust cloud. Urey<br />

thought that the earth’s primordial atmosphere probably contained<br />

elements in the approximate relative abundances found in the solar<br />

system <strong>and</strong> the universe.<br />

It had been discovered in 1929 that the Sun is approximately 87<br />

percent hydrogen, <strong>and</strong> by 1935 it was known that hydrogen encompassed<br />

the vast majority (92.8 percent) of atoms in the universe.<br />

Urey reasoned that the earth’s early atmosphere contained mostly<br />

hydrogen, with the oxygen, nitrogen, <strong>and</strong> carbon atoms chemically<br />

bonded to hydrogen to form water, ammonia, <strong>and</strong> methane. Most<br />

important, free oxygen could not exist in the presence of such an<br />

abundance of hydrogen.<br />

As early as the mid-1920’s, Aleks<strong>and</strong>r Ivanovich Oparin, a Russian<br />

biochemist, had argued that the organic compounds necessary<br />

for life had been built up on the early earth by chemical combinations<br />

in a reducing atmosphere. The energy from the Sun would<br />

have been sufficient to drive the reactions to produce life. Haldane<br />

later proposed that the organic compounds would accumulate in<br />

the oceans to produce a “dilute organic soup” <strong>and</strong> that life might<br />

have arisen by some unknown process from that mixture of organic<br />

compounds.<br />

Primordial Soup in a Bottle<br />

Synthetic amino acid / 725<br />

Miller combined the ideas of Oparin <strong>and</strong> Urey <strong>and</strong> designed a<br />

simple, but elegant, experiment. He decided to mix the gases presumed<br />

to exist in the early atmosphere (water vapor, hydrogen, ammonia,<br />

<strong>and</strong> methane) <strong>and</strong> expose them to an electrical spark to determine<br />

which, if any, organic compounds were formed. To do this,<br />

he constructed a relatively simple system, essentially consisting of<br />

two Pyrex flasks connected by tubing in a roughly circular pattern.<br />

The water <strong>and</strong> gases in the smaller flask were boiled <strong>and</strong> the resulting<br />

gas forced through the tubing into a larger flask that contained<br />

tungsten electrodes. As the gases passed the electrodes, an electrical<br />

spark was generated, <strong>and</strong> from this larger flask the gases <strong>and</strong> any<br />

other compounds were condensed. The gases were recycled through<br />

the system, whereas the organic compounds were trapped in the<br />

bottom of the system.


726 / Synthetic amino acid<br />

Miller was trying to simulate conditions that had prevailed on<br />

the early earth. During the one week of operation, Miller extracted<br />

<strong>and</strong> analyzed the residue of compounds at the bottom of the system.<br />

The results were truly astounding. He found that numerous organic<br />

compounds had, indeed, been formed in only that one week. As<br />

much as 15 percent of the carbon (originally in the gas methane) had<br />

been combined into organic compounds, <strong>and</strong> at least 5 percent of<br />

the carbon was incorporated into biochemically important compounds.<br />

The most important compounds produced were some of<br />

the twenty amino acids essential to life on Earth.<br />

The formation of amino acids is significant because they are the<br />

building blocks of proteins. Proteins consist of a specific sequence of<br />

amino acids assembled into a well-defined pattern. Proteins are necessary<br />

for life for two reasons. First, they are important structural<br />

Water in<br />

Vacuum<br />

Boiling Water<br />

Water Vapor<br />

Electrode<br />

Nh3<br />

CH4<br />

The Miller-Urey experiment.<br />

H2<br />

Condenser<br />

Cooled Water<br />

Containing<br />

Organic Compounds<br />

Sample for<br />

Chemical Analysis


materials used to build the cells of the body. Second, the enzymes<br />

that increase the rate of the multitude of biochemical reactions of life<br />

are also proteins. Miller not only had produced proteins in the laboratory<br />

but also had shown clearly that the precursors of proteins—<br />

the amino acids—were easily formed in a reducing environment<br />

with the appropriate energy.<br />

Perhaps the most important aspect of the experiment was the<br />

ease with which the amino acids were formed. Of all the thous<strong>and</strong>s<br />

of organic compounds that are known to chemists, amino acids<br />

were among those that were formed by this simple experiment. This<br />

strongly implied that one of the first steps in chemical evolution was<br />

not only possible but also highly probable. All that was necessary<br />

for the synthesis of amino acids were the common gases of the solar<br />

system, a reducing environment, <strong>and</strong> an appropriate energy source,<br />

all of which were present on early Earth.<br />

Consequences<br />

Synthetic amino acid / 727<br />

Miller opened an entirely new field of research with his pioneering<br />

experiments. His results showed that much about chemical<br />

evolution could be learned by experimentation in the laboratory.<br />

As a result, Miller <strong>and</strong> many others soon tried variations on<br />

his original experiment by altering the combination of gases, using<br />

other gases, <strong>and</strong> trying other types of energy sources. Almost all<br />

the essential amino acids have been produced in these laboratory<br />

experiments.<br />

Miller’s work was based on the presumed composition of the<br />

primordial atmosphere of Earth. The composition of this atmosphere<br />

was calculated on the basis of the abundance of elements<br />

in the universe. If this reasoning is correct, then it is highly likely<br />

that there are many other bodies in the universe that have similar<br />

atmospheres <strong>and</strong> are near energy sources similar to the Sun.<br />

Moreover, Miller’s experiment strongly suggests that amino acids,<br />

<strong>and</strong> perhaps life as well, should have formed on other planets.<br />

See also Artificial hormone; Artificial kidney; Synthetic DNA;<br />

Synthetic RNA.


728 / Synthetic amino acid<br />

Further Reading<br />

Dronamraju, Krishna R., <strong>and</strong> J. B. S. Haldane. Haldane’s Daedalus Revisited.<br />

New York: Oxford University Press, 1995.<br />

Lipkin, Richard. “Early Earth May Have Had Two Key RNA Bases.”<br />

Science News 148, no. 1 (July 1, 1995).<br />

Miller, Stanley L., <strong>and</strong> Leslie E. Orgel. The Origins of Life on the Earth.<br />

Englewood Cliffs, N.J.: Prentice-Hall, 1974.<br />

Nelson, Kevin E., Matthew Levy, <strong>and</strong> Stanley L. Miller. “Peptide<br />

Nucleic Acids Rather than RNA May Have Been the First Genetic<br />

Molecule.” Proceedings of the National Academy of Sciences of the<br />

United States of America 97, no. 8 (April 11, 2000).<br />

Yockey, Hubert P. “Walther Lob, Stanley L. Miller, <strong>and</strong> Prebiotic<br />

‘Building Blocks’ in the Silent Electrical Discharge.” Perspectives<br />

in Biology <strong>and</strong> Medicine 41, no. 1 (Autumn, 1997).


Synthetic DNA<br />

Synthetic DNA<br />

The invention: A method for replicating viral deoxyribonucleic<br />

acid (DNA) in a test tube that paved the way for genetic engineering.<br />

The people behind the invention:<br />

Arthur Kornberg (1918- ), an American physician <strong>and</strong><br />

biochemist<br />

Robert L. Sinsheimer (1920- ), an American biophysicist<br />

Mehran Goulian (1929- ), a physician <strong>and</strong> biochemist<br />

The Role of DNA<br />

729<br />

Until the mid-1940’s, it was believed that proteins were the<br />

carriers of genetic information, the source of heredity. Proteins<br />

appeared to be the only biological molecules that had the complexity<br />

necessary to encode the enormous amount of genetic information<br />

required to reproduce even the simplest organism.<br />

Nevertheless, proteins could not be shown to have genetic properties,<br />

<strong>and</strong> by 1944, it was demonstrated conclusively that deoxyribonucleic<br />

acid (DNA) was the material that transmitted hereditary<br />

information. It was discovered that DNA isolated from a<br />

strain of infective bacteria that can cause pneumonia was able to<br />

transform a strain of noninfective bacteria into an infective strain;<br />

in addition, the infectivity trait was transmitted to future generations.<br />

Subsequently, it was established that DNA is the genetic material<br />

in virtually all forms of life.<br />

Once DNA was known to be the transmitter of genetic information,<br />

scientists sought to discover how it performs its role. DNA is a<br />

polymeric molecule composed of four different units, called “deoxynucleotides.”<br />

The units consist of a sugar, a phosphate group, <strong>and</strong> a<br />

base; they differ only in the nature of the base, which is always one of<br />

four related compounds: adenine, guanine, cytosine, or thymine. The<br />

way in which such a polymer could transmit genetic information,<br />

however, was difficult to discern. In 1953, biophysicists James D. Watson<br />

<strong>and</strong> Francis Crick brilliantly determined the three-dimensional


730 / Synthetic DNA<br />

structure of DNA by analyzing X-ray diffraction photographs of DNA<br />

fibers. From their analysis of the structure of DNA, Watson <strong>and</strong> Crick<br />

inferred DNA’s mechanism of replication. Their work led to an underst<strong>and</strong>ing<br />

of gene function in molecular terms.<br />

Watson <strong>and</strong> Crick showed that DNA has a very long doublestr<strong>and</strong>ed<br />

(duplex) helical structure. DNA has a duplex structure because<br />

each base forms a link to a specific base on the opposite<br />

str<strong>and</strong>. The discovery of this complementary pairing of bases provided<br />

a model to explain the two essential functions of a hereditary<br />

molecule: It must preserve the genetic code from one generation to<br />

the next, <strong>and</strong> it must direct the development of the cell.<br />

Watson <strong>and</strong> Crick also proposed that DNA is able to serve as a<br />

mold (or template) for its own reproduction because the two str<strong>and</strong>s<br />

of DNA polymer can separate. Upon separation, each str<strong>and</strong> acts as a<br />

template for the formation of a new complementary str<strong>and</strong>. An adenine<br />

base in the existing str<strong>and</strong> gives rise to cytosine, <strong>and</strong> so on. In<br />

this manner, a new double-str<strong>and</strong>ed DNA is generated that is identical<br />

to the parent DNA.<br />

DNA in a Test Tube<br />

Watson <strong>and</strong> Crick’s theory provided a valuable model for the reproduction<br />

of DNA, but it did not explain the biological mechanism<br />

by which the process occurs. The biochemical pathway of DNA reproduction<br />

<strong>and</strong> the role of the enzymes required for catalyzing the<br />

reproduction process were discovered by Arthur Kornberg <strong>and</strong> his<br />

coworkers. For his success in achieving DNA synthesis in a test tube<br />

<strong>and</strong> for discovering <strong>and</strong> isolating an enzyme—DNA polymerase—<br />

that catalyzed DNA synthesis, Kornberg won the 1959 Nobel Prize<br />

in Physiology or Medicine.<br />

To achieve DNA replication in a test tube, Kornberg found that a<br />

small amount of preformed DNA must be present, in addition to<br />

DNA polymerase enzyme <strong>and</strong> all four of the deoxynucleotides that<br />

occur in DNA. Kornberg discovered that the base composition of<br />

the newly made DNA was determined solely by the base composition<br />

of the preformed DNA, which had been used as a template in<br />

the test-tube synthesis. This result showed that DNA polymerase<br />

obeys instructions dictated by the template DNA. It is thus said to


e “template-directed.” DNA polymerase was the first templatedirected<br />

enzyme to be discovered.<br />

Although test-tube synthesis was a most significant achievement,<br />

important questions about the precise character of the newly<br />

made DNA were still unanswered. Methods of analyzing the order,<br />

or sequence, of the bases in DNA were not available, <strong>and</strong> hence it<br />

could not be shown directly whether DNAmade in the test tube was<br />

an exact copy of the template of DNA or merely an approximate<br />

copy. In addition, some DNAs prepared by DNA polymerase appeared<br />

to be branched structures. Since chromosomes in living cells<br />

contain long, linear, unbranched str<strong>and</strong>s of DNA, this branching<br />

might have indicated that DNA synthesized in a test tube was not<br />

equivalent to DNA synthesized in the living cell.<br />

Kornberg realized that the best way to demonstrate that newly<br />

synthesized DNA is an exact copy of the original was to test the new<br />

DNA for biological activity in a suitable system. Kornberg reasoned<br />

that a demonstration of infectivity in viral DNA produced in a test<br />

tube would prove that polymerase-catalyzed synthesis was virtually<br />

error-free <strong>and</strong> equivalent to natural, biological synthesis. The<br />

experiment, carried out by Kornberg, Mehran Goulian at Stanford<br />

University, <strong>and</strong> Robert L. Sinsheimer at the California Institute of<br />

Technology, was a complete success. The viral DNAs produced in a<br />

test tube by the DNA polymerase enzyme, using a viral DNA template,<br />

were fully infective. This synthesis showed that DNA polymerase<br />

could copy not merely a single gene but also an entire chromosome<br />

of a small virus without error.<br />

Consequences<br />

Synthetic DNA / 731<br />

The purification of DNA polymerase <strong>and</strong> the preparation of biologically<br />

active DNA were major achievements that influenced<br />

biological research on DNA for decades. Kornberg’s methodology<br />

proved to be invaluable in the discovery of other enzymes that synthesize<br />

DNA. These enzymes have been isolated from Escherichia<br />

coli bacteria <strong>and</strong> from other bacteria, viruses, <strong>and</strong> higher organisms.<br />

The test-tube preparation of viral DNA also had significance in<br />

the studies of genes <strong>and</strong> chromosomes. In the mid-1960’s, it had not<br />

been established that a chromosome contains a continuous str<strong>and</strong> of


732 / Synthetic DNA<br />

DNA. Kornberg <strong>and</strong> Sinsheimer’s synthesis of a viral chromosome<br />

proved that it was, indeed, a very long str<strong>and</strong> of uninterrupted<br />

DNA.<br />

Kornberg <strong>and</strong> Sinsheimer’s work laid the foundation for subsequent<br />

recombinant DNA research <strong>and</strong> for genetic engineering technology.<br />

This technology promises to revolutionize both medicine<br />

<strong>and</strong> agriculture. The enhancement of food production <strong>and</strong> the generation<br />

of new drugs <strong>and</strong> therapies are only a few of the subsequent<br />

benefits that may be expected.<br />

See also Artificial chromosome; Artificial hormone; Cloning; Genetic<br />

“fingerprinting”; Genetically engineered insulin; In vitro plant<br />

culture; Synthetic amino acid; Synthetic RNA.<br />

Further Reading<br />

Baker, Tania A., <strong>and</strong> Arthur Kornberg. DNA Replication. 2d ed. New<br />

York: W. H. Freeman, 1991.<br />

Kornberg, Arthur. The Golden Helix: Inside Biotech Ventures. Sausalito,<br />

Calif.: University Science Books, 1995.<br />

_____. For the Love of Enzymes: The Odyssey of a Biochemist. Harvard<br />

University Press, 1991.<br />

Sinsheimer, Robert. The Str<strong>and</strong>s of a Life: The Science of DNA <strong>and</strong> the<br />

Art of Education. Berkeley: University of California Press, 1994.


Synthetic RNA<br />

Synthetic RNA<br />

The invention: A method for synthesizing the biological molecule<br />

RNA established that this process can occur outside the living<br />

cell.<br />

The people behind the invention:<br />

Severo Ochoa (1905-1993), a Spanish biochemist who shared<br />

the 1959 Nobel Prize in Physiology or Medicine<br />

Marianne Grunberg-Manago (1921- ), a French biochemist<br />

Marshall W. Nirenberg (1927- ), an American biochemist<br />

who won the 1968 Nobel Prize in Physiology or Medicine<br />

Peter Lengyel (1929- ), a Hungarian American biochemist<br />

RNA Outside the Cells<br />

733<br />

In the early decades of the twentieth century, genetics had not<br />

been experimentally united with biochemistry. This merging soon<br />

occurred, however, with work involving the mold Neurospora crassa.<br />

This Nobel award-winning work by biochemist Edward Lawrie<br />

Tatum <strong>and</strong> geneticist George Wells Beadle showed that genes control<br />

production of proteins, which are major functional molecules in<br />

cells. Yet no one knew the chemical composition of genes <strong>and</strong> chromosomes,<br />

or, rather, the molecules of heredity.<br />

The American bacteriologist Oswald T. Avery <strong>and</strong> his colleagues<br />

at New York’s Rockefeller Institute determined experimentally that<br />

the molecular basis of heredity was a large polymer known as deoxyribonucleic<br />

acid (DNA). Avery’s discovery triggered a furious<br />

worldwide search for the particular structural characteristics of<br />

DNA, which allow for the known biological characteristics of genes.<br />

One of the most famous studies in the history of science solved<br />

this problem in 1953. Scientists James D. Watson, Francis Crick, <strong>and</strong><br />

Maurice H. F. Wilkins postulated that DNA exists as a double helix.<br />

That is, two long str<strong>and</strong>s twist about each other in a predictable pattern,<br />

with each single str<strong>and</strong> held to the other by weak, reversible<br />

linkages known as “hydrogen bonds.” About this time, researchers<br />

recognized also that a molecule closely related to DNA, ribonucleic


734 / Synthetic RNA<br />

acid (RNA), plays an important role in transcribing the genetic information<br />

as well as in other biological functions.<br />

Severo Ochoa was born in Spain as the science of genetics was<br />

developing. In 1942, he moved to New York University, where he<br />

studied the bacterium Azobacter vinel<strong>and</strong>ii. Specifically, Ochoa was<br />

focusing on the question of how cells process energy in the form of<br />

organic molecules such as the sugar glucose to provide usable biological<br />

energy in the form of adenosine triphosphate (ATP). With<br />

postdoctoral fellow Marianne Grunberg-Manago, he studied enzymatic<br />

reactions capable of incorporating inorganic phosphate (a<br />

compound consisting of one atom of phosphorus <strong>and</strong> four atoms of<br />

oxygen) into adenosine diphosphate (ADP) to form ATP.<br />

One particularly interesting reaction was followed by monitoring<br />

the amount of radioactive phosphate reacting with ADP. Following<br />

separation of the reaction products, it was discovered that<br />

the main product was not ATP, but a much larger molecule. Chemical<br />

characterization demonstrated that this product was a polymer<br />

of adenosine monophosphate. When other nucleocide diphosphates,<br />

such as inosine diphosphate, were used in the reaction, the<br />

corresponding polymer of inosine monophosphate was formed.<br />

Thus, in each case, a polymer (a long string of building-block<br />

units) was formed. The polymers formed were synthetic RNAs, <strong>and</strong><br />

the enzyme responsible for the conversion became known as “polynucleotide<br />

phosphorylase.” This finding, once the early skepticism<br />

was resolved, was received by biochemists with great enthusiasm<br />

because no technique outside the cell had ever been discovered<br />

previously in which a nucleic acid similar to RNA could be<br />

synthesized.<br />

Learning the Language<br />

Ochoa, Peter Lengyel, <strong>and</strong> Marshall W. Nirenberg at the National<br />

Institute of Health took advantage of this breakthrough to synthesize<br />

different RNAs useful in cracking the genetic code. Crick had<br />

postulated that the flow of information in biological systems is from<br />

DNA to RNA to protein. In other words, genetic information contained<br />

in the DNA structure is transcribed into complementary<br />

RNA structures, which, in turn, are translated into the protein. Pro-


tein synthesis, an extremely complex process, involves bringing a<br />

type of RNA, known as messenger RNA, together with amino acids<br />

<strong>and</strong> huge cellular organelles known as ribosomes.<br />

Yet investigators did not know the nature of the nucleic acid alphabet—for<br />

example, how many single units of the RNA polymer<br />

code were needed for each amino acid, <strong>and</strong> the order that the units<br />

must be in to st<strong>and</strong> for a “word” in the nucleic acid language. In<br />

1961, Nirenberg demonstrated that the polymer of synthetic RNA<br />

with multiple units of uracil (poly U) would “code” only for a protein<br />

containing the amino acid phenylalanine. Each three units (U’s)<br />

gave one phenylalanine. Therefore, genetic words each contain<br />

three letters. UUU translates into phenylalanine. Poly A, the first<br />

polymer discovered with polynucleotide phosphorylase, was coded<br />

for a protein containing multiple lysines. That is, AAA translates<br />

into the amino acid lysine.<br />

The words, containing combinations of letters, such as AUG, were<br />

not as easily studied, but Nirenberg, Ochoa, <strong>and</strong> Gobind Khorana of<br />

the University of Wisconsin eventually uncovered the exact translation<br />

for each amino acid. In RNA, there are four possible letters (A, U,<br />

G, <strong>and</strong> C) <strong>and</strong> three letters in each word. Accordingly, there are sixtyfour<br />

possible words. With only twenty amino acids, it became clear<br />

that more than one RNA word can translate into a given amino acid.<br />

Yet, no given word st<strong>and</strong>s for any more than one amino acid. A few<br />

words do not translate into any amino acid; they are stop signals, telling<br />

the ribosome to cease translating RNA.<br />

The question of which direction an RNA is translated is critical.<br />

For example, CAA codes for the amino acid glutamine, but the reverse,<br />

AAC, translates to the amino acid asparagine. Such a difference<br />

is critical because the exact sequence of a protein determines its<br />

activity—that is, what it will do in the body <strong>and</strong> therefore what genetic<br />

trait it will express.<br />

Consequences<br />

Synthetic RNA / 735<br />

Synthetic RNAs provided the key to underst<strong>and</strong>ing the genetic<br />

code. The genetic code is universal; it operates in all organisms, simple<br />

or complex. It is used by viruses, which are nearly life but are not<br />

alive. Spelling out the genetic code was one of the top discoveries of


736 / Synthetic RNA<br />

the twentieth century. Nearly all work in molecular biology depends<br />

on this knowledge.<br />

The availability of synthetic RNAs has provided hybridization<br />

tools for molecular geneticists. Hybridization is a technique in which<br />

an RNA is allowed to bind in a complementary fashion to DNA under<br />

investigation. The greater the similarity between RNA <strong>and</strong> DNA,<br />

the greater the amount of binding. The differential binding allows for<br />

seeking, finding, <strong>and</strong> ultimately isolating a target DNA from a large,<br />

diverse pool of DNA—in short, finding a needle in a haystack. Hybridization<br />

has become an indispensable aid in experimental molecular<br />

genetics as well as in applied sciences, such as forensics.<br />

See also Artificial chromosome; Artificial hormone; Cloning; Genetic<br />

“fingerprinting”; Genetically engineered insulin; In vitro plant<br />

culture; Synthetic amino acid; Synthetic DNA.<br />

Further Reading<br />

“Biochemist Severo Ochoa Dies: Won Nobel Prize.” Washington Post<br />

(November 3, 1993).<br />

Santesmases, Maria Jesus. “Severo Ochoa <strong>and</strong> the Biomedical Sciences<br />

in Spain Under Franco, 1959-1975.” Isis 91, no. 4 (December,<br />

2000).<br />

“Severo Ochoa, 1905-1993.” Nature 366, no. 6454 (December, 1993).


Syphilis test<br />

Syphilis test<br />

The invention: The first simple test for detecting the presence of<br />

the venereal disease syphilis led to better syphilis control <strong>and</strong><br />

other advances in immunology.<br />

The people behind the invention:<br />

Reuben Leon Kahn (1887-1974), a Soviet-born American<br />

serologist <strong>and</strong> immunologist<br />

August von Wassermann (1866-1925), a German physician <strong>and</strong><br />

bacteriologist<br />

Columbus’s Discoveries<br />

737<br />

Syphilis is one of the chief venereal diseases, a group of diseases<br />

whose name derives from Venus, the Roman goddess of love. The<br />

term “venereal” arose from the idea that the diseases were transmitted<br />

solely by sexual contact with an infected individual. Although<br />

syphilis is almost always passed from one person to another in this<br />

way, it occasionally arises after contact with objects used by infected<br />

people in highly unclean surroundings, particularly in the underdeveloped<br />

countries of the world.<br />

It is believed by many that syphilis was introduced to Europe by<br />

the members of Spanish explorer Christopher Columbus’s crew—<br />

supposedly after they were infected by sexual contact with West Indian<br />

women—during their voyages of exploration. Columbus is reported<br />

to have died of heart <strong>and</strong> brain problems very similar to<br />

symptoms produced by advanced syphilis. At that time, according<br />

to many historians, syphilis spread rapidly over sixteenth century<br />

Europe. The name “syphilis” was coined by the Italian physician<br />

Girolamo Fracastoro in 1530 in an epic poem he wrote.<br />

Modern syphilis is much milder than the original disease <strong>and</strong> relatively<br />

uncommon. Yet, if it is not identified <strong>and</strong> treated appropriately,<br />

syphilis can be devastating <strong>and</strong> even fatal. It can also be passed from<br />

pregnant mothers to their unborn children. In these cases, the afflicted<br />

children will develop serious health problems that can include<br />

paralysis, insanity, <strong>and</strong> heart disease. Therefore, the underst<strong>and</strong>ing,


738 / Syphilis test<br />

detection, <strong>and</strong> cure of syphilis are important worldwide.<br />

Syphilis is caused by a spiral-shaped germ called a “spirochete.”<br />

Spirochetes enter the body through breaks in the skin or through the<br />

mucous membranes, regardless of how they are transmitted. Once<br />

spirochetes enter the body, they spread rapidly. During the first four<br />

to six weeks after infection, syphilis—said to be in its primary<br />

phase—is very contagious. During this time, it is identified by the<br />

appearance of a sore, or chancre, at the entry site of the infecting spirochetes.<br />

The chancre disappears quickly, <strong>and</strong> within six to twenty-four<br />

weeks, the disease shows itself as a skin rash, feelings of malaise,<br />

<strong>and</strong> other flulike symptoms (secondary-phase syphilis). These problems<br />

also disappear quickly in most cases, <strong>and</strong> here is where the real<br />

trouble—latent syphilis—begins. In latent syphilis, now totally without<br />

symptoms, spirochetes that have spread through the body may<br />

lodge in the brain or the heart. When this happens, paralysis, mental<br />

incapacitation, <strong>and</strong> death may follow.<br />

Testing Before Marriage<br />

Because of the danger to unborn children, Americans wishing to<br />

marry must be certified as being free of the disease before a marriage<br />

license is issued. The cure for syphilis is easily accomplished<br />

through the use of penicillin or other types of antibiotics, though no<br />

vaccine is yet available to prevent the disease. It is for this reason<br />

that syphilis detection is particularly important.<br />

The first viable test for syphilis was originated by August von<br />

Wassermann in 1906. In this test, blood samples are taken <strong>and</strong><br />

treated in a medical laboratory. The treatment of the samples is<br />

based on the fact that the blood of infected persons has formed antibodies<br />

to fight the syphilis spirochete, <strong>and</strong> that these antibodies will<br />

react with certain body chemicals to cause the blood sample to clot.<br />

This indicates the person has the disease. After the syphilis has been<br />

cured, the antibodies disappear, as does the clotting.<br />

Although the Wassermann test was effective in 95 percent of all<br />

infected persons, it was very time-consuming (requiring a two-day<br />

incubation period) <strong>and</strong> complex. In 1923, Reuben Leon Kahn developed<br />

a modified syphilis test, “the st<strong>and</strong>ard Kahn test,” that was


simpler <strong>and</strong> faster: The test was complete after only a few minutes.<br />

By 1925, Kahn’s test had become the st<strong>and</strong>ard syphilis test of the<br />

United States Navy <strong>and</strong> later became a worldwide test for the detection<br />

of the disease.<br />

Kahn soon realized that his test was not perfect <strong>and</strong> that in some<br />

cases, the results were incorrect. This led him to a broader study of<br />

the immune reactions at the center of the Kahn test. He investigated<br />

the role of various tissues in immunity, as compared to the role of<br />

white blood antibodies <strong>and</strong> white blood cells. Kahn showed, for example,<br />

that different tissues of immunized or nonimmunized animals<br />

possessed differing immunologic capabilities. Furthermore,<br />

the immunologic capabilities of test animals varied with their<br />

age, being very limited in newborns <strong>and</strong> increasing as they matured.<br />

This effort led, by 1951, to Kahn’s “universal serological reaction,”<br />

a precipitation reaction in which blood serum was tested<br />

against a reagent composed of tissue lipids. Kahn viewed it as a potentially<br />

helpful chemical indicator of how healthy or ill an individual<br />

was. This effort is viewed as an important l<strong>and</strong>mark in the development<br />

of the science of immunology.<br />

Impact<br />

Syphilis test / 739<br />

At the time that Kahn developed his st<strong>and</strong>ard Kahn test for syphilis,<br />

the Wassermann test was used all over the world for the diagnosis<br />

of syphilis. As has been noted, one of the great advantages of the<br />

st<strong>and</strong>ard Kahn test was its speed, minutes versus days. For example,<br />

in October, 1923, Kahn is reported to have tested forty serum<br />

samples in fifteen minutes.<br />

Kahn’s efforts have been important to immunology <strong>and</strong> to medicine.<br />

Among the consequences of his endeavors was the stimulation<br />

of other developments in the field, including the VDRL test (originated<br />

by the Venereal Disease Research Laboratory), which has replaced<br />

the Kahn test as one of the most often used screening tests for<br />

syphilis. Even more specific syphilis tests developed later include a<br />

fluorescent antibody test to detect the presence of the antibody to<br />

the syphilis spirochete.


740 / Syphilis test<br />

See also Abortion pill; Amniocentesis; Antibacterial drugs; Birth<br />

control pill; Mammography; Pap test; Penicillin; Ultrasound.<br />

Further Reading<br />

Cates, William, Jr., Richard B. Rothenberg, <strong>and</strong> Joseph H. Blount.<br />

“Syphilis Control.” Sexually Transmitted Diseases 23, no. 1 (January,<br />

1996).<br />

Cobb, W. Montague. “Reuben Leon Kahn.” Journal of the National<br />

Medical Association 63 (September, 1971).<br />

Quétel, Claude. History of Syphilis. Baltimore: Johns Hopkins University<br />

Press, 1992.<br />

St. Louis, Michael E., <strong>and</strong> Judith N. Wasserheit. “Elimination of<br />

Syphilis in the United States.” Science 281, no. 5375 (July, 1998).


Talking motion pictures<br />

Talking motion pictures<br />

The invention: The first practical system for linking sound with<br />

moving pictures.<br />

The people behind the invention:<br />

Harry Warner (1881-1958), the brother who used sound to<br />

fashion a major filmmaking company<br />

Albert Warner (1884-1967), the brother who persuaded theater<br />

owners to show Warner films<br />

Samuel Warner (1887-1927), the brother who adapted soundrecording<br />

technology to filmmaking<br />

Jack Warner (1892-1978), the brother who supervised the<br />

making of Warner films<br />

Taking the Lead<br />

741<br />

The silent films of the early twentieth century had live sound accompaniment<br />

featuring music <strong>and</strong> sound effects. Neighborhood<br />

theaters made do with a piano <strong>and</strong> violin; larger “picture palaces”<br />

in major cities maintained resident orchestras of more than seventy<br />

members. During the late 1920’s, Warner Bros. led the American<br />

film industry in producing motion pictures with their own soundtracks,<br />

which were first recorded on synchronized records <strong>and</strong> later<br />

added on to the film beside the images.<br />

The ideas that led to the addition of sound to film came from corporate-sponsored<br />

research by American Telephone <strong>and</strong> Telegraph<br />

Company (AT&T) <strong>and</strong> the Radio Corporation of America (RCA).<br />

Both companies worked to improve sound recording <strong>and</strong> playback,<br />

AT&T to help in the design of long-distance telephone equipment<br />

<strong>and</strong> RCA as part of the creation of better radio sets. Yet neither company<br />

could, or would, enter filmmaking. AT&T was willing to contract<br />

its equipment out to Paramount or one of the other major Hollywood<br />

studios of the day; such studios, however, did not want to<br />

risk their sizable profit positions by junking silent films. The giants<br />

of the film industry were doing fine with what they had <strong>and</strong> did not<br />

want to switch to something that had not been proved.


742 / Talking motion pictures<br />

In 1924, Warner Bros. was a prosperous, though small, corporation<br />

that produced films with the help of outside financial backing. That<br />

year, Harry Warner approached the important Wall Street investment<br />

banking house of Goldman, Sachs <strong>and</strong> secured the help he needed.<br />

As part of this initial wave of expansion, Warner Bros. acquired a<br />

Los Angeles radio station in order to publicize its films. Through<br />

this deal, the four Warner brothers learned of the new technology<br />

that the radio <strong>and</strong> telephone industries had developed to record<br />

sound, <strong>and</strong> they succeeded in securing the necessary equipment<br />

from AT&T. During the spring of 1925, the brothers devised a plan<br />

by which they could record the most popular musical artists on film<br />

<strong>and</strong> then offer these “shorts” as added attractions to theaters that<br />

booked its features. As a bonus, Warner Bros. could add recorded<br />

orchestral music to its feature films <strong>and</strong> offer this music to theaters<br />

that relied on small musical ensembles.<br />

“Vitaphone”<br />

On August 6, 1926, Warner Bros. premiered its new “Vitaphone”<br />

technology. The first package consisted of a traditional silent film<br />

(Don Juan) with a recorded musical accompaniment, plus six recordings<br />

of musical talent highlighted by a performance from Giovanni<br />

Martineli, the most famous opera tenor of the day.<br />

The first Vitaphone feature was The Jazz Singer, which premiered<br />

in October, 1927. The film was silent during much of the movie, but<br />

as soon as Al Jolson, the star, broke into song, the new technology<br />

would be implemented. The film was an immediate hit. The Jazz<br />

Singer package, which included accompanying shorts with sound,<br />

forced theaters in cities that rarely held films over for more than a<br />

single week to ask to have the package stay for two, three, <strong>and</strong><br />

sometimes four straight weeks.<br />

The Jazz Singer did well at the box office, but skeptics questioned<br />

the staying power of talkies. If sound was so important, they wondered,<br />

why hadn’t The Jazz Singer moved to the top of the all-time<br />

box-office list? Such success, though, would come a year later with<br />

The Singing Fool, also starring Jolson. From its opening day (September<br />

20, 1928), it was the financial success of its time; produced for an<br />

estimated $200,000, it took in $5 million. In New York City, The


Talking motion pictures / 743<br />

In the early days of sound films, cameras had to be soundproofed so their operating noises<br />

would not be picked up by the primitive sound-recording equipment. (Library of Congress)<br />

Singing Fool registered the heaviest business in Broadway history,<br />

with an advance sale that exceeded more than $100,000 (equivalent<br />

to more than half a million dollars in 1990’s currency).


744 / Talking motion pictures<br />

Impact<br />

The coming of sound transformed filmmaking, ushering in what<br />

became known as the golden age of Hollywood. By 1930, there were<br />

more reporters stationed in the filmmaking capital of the world<br />

than in any capital of Europe or Asia.<br />

The Warner Brothers<br />

Businessmen rather than inventors, the four Warner brothers<br />

were hustlers who knew a good thing when they saw it.<br />

They started out running theaters in 1903, evolved into film distributors,<br />

<strong>and</strong> began making their own films in 1909, in defiance<br />

of the Patents Company, a trust established by Thomas A. Edison<br />

to eliminate competition from independent filmmakers.<br />

Harry Warner was the president of the company, Sam <strong>and</strong> Jack<br />

were vice presidents in charge of production, <strong>and</strong> Abe (or Albert)<br />

was the treasurer.<br />

Theirs was a small concern. Their silent films <strong>and</strong> serials attracted<br />

few audiences, <strong>and</strong> during World War I they made<br />

training films for the government. In fact, their film about syphilis,<br />

Open Your Eyes, was their first real success. In 1918, however,<br />

they released My Four Years in Germany, a dramatized<br />

documentary, <strong>and</strong> it was their first blockbuster. Although considered<br />

gauche upstarts, they were suddenly taken seriously by<br />

the movie industry.<br />

When Sam first heard an actor talk on screen in an experimental<br />

film at the Bell lab in New York in 1925, he recognized a<br />

revolutionary opportunity. He soon convinced Jack that talking<br />

movies would be a gold mine. However, Harry <strong>and</strong> Abe were<br />

against the idea because of its costs—<strong>and</strong> because earlier attempts<br />

at “talkies” had been dismal failures. Sam <strong>and</strong> Jack<br />

tricked Harry into a seeing a experimental film of an orchestra,<br />

however, <strong>and</strong> he grew enthusiastic despite his misgivings. Within<br />

a year, the brothers released the all-music Don Juan. The rave<br />

notices from critics astounded Harry <strong>and</strong> Abe.<br />

Still, they thought sound in movies was simply a novelty.<br />

When Sam pointed out that they could make movies in which<br />

the actors talked, as on stage, Harry, who detested actors, snorted,<br />

“Who the hell wants to hear actors talk?” Sam <strong>and</strong> Jack pressed<br />

for dramatic talkies, nonetheless, <strong>and</strong> prevailed upon Harry to<br />

finance them. The silver screen has seldom been silent since.


As a result of its foresight, Warner Bros. was the sole small competitor<br />

of the early 1920’s to succeed in the Hollywood elite, producing<br />

successful films for consumption throughout the world.<br />

After Warner Bros.’ innovation, the soundtrack became one of<br />

the features that filmmakers controlled when making a film. Indeed,<br />

sound became a vital part of the filmmaker’s art; music, in<br />

particular, could make or break a film.<br />

Finally, the coming of sound helped make films a dominant medium<br />

of mass culture, both in the United States <strong>and</strong> throughout the<br />

world. Innumerable fashions, expressions, <strong>and</strong> designs were soon created<br />

or popularized by filmmakers. Many observers had not viewed<br />

the silent cinema as especially significant; with the coming of the talkies,<br />

however, there was no longer any question about the social <strong>and</strong><br />

cultural importance of films. As one clear consequence of the new<br />

power of the movie industry, within a few years of the coming of<br />

sound, the notorious Hays Code m<strong>and</strong>ating prior restraint of film content<br />

went into effect. The pairing of images <strong>and</strong> sound caused talking<br />

films to be deemed simply too powerful for uncensored presentation<br />

to audiences; although the Hays Code was gradually weakened <strong>and</strong><br />

eventually ab<strong>and</strong>oned, less onerous “rating systems” would continue<br />

to be imposed on filmmakers by various regulatory bodies.<br />

See also Autochrome plate; Dolby noise reduction; Electronic<br />

synthesizer; Television.<br />

Further Reading<br />

Talking motion pictures / 745<br />

Brayer, Elizabeth. George Eastman: A Biography. Baltimore: Johns<br />

Hopkins University Press, 1996.<br />

Crafton, Donald. The Talkies: American Cinema’s Transition to Sound,<br />

1926-1931. Berkeley: University of California Press, 1999.<br />

Geduld, Harry M. The Birth of the Talkies: From Edison to Jolson.<br />

Bloomington: Indiana University Press, 1975.<br />

Neale, Stephen. Cinema <strong>and</strong> Technology: Image, Sound, Colour. London:<br />

Macmillan Education, 1985.<br />

Wagner, A. F. Recollections of Thomas A. Edison: A Personal History of<br />

the Early Days of the Phonograph, the Silent <strong>and</strong> Sound Film, <strong>and</strong> Film<br />

Censorship. 2d ed. London: City of London Phonograph & Gramophone<br />

Society, 1996.


746<br />

Teflon<br />

Teflon<br />

The invention: A fluorocarbon polymer whose chemical inertness<br />

<strong>and</strong> physical properties have made it useful for many applications,<br />

from nonstick cookware coatings to suits for astronauts.<br />

The person behind the invention:<br />

Roy J. Plunkett (1910-1994), an American chemist<br />

Nontoxic Refrigerant Sought<br />

As the use of mechanical refrigeration increased in the late 1930’s,<br />

manufacturers recognized the need for a material to replace sulfur<br />

dioxide <strong>and</strong> ammonia, which, although they were the commonly<br />

used refrigerants of the time, were less than ideal for the purpose.<br />

The material sought had to be nontoxic, odorless, colorless, <strong>and</strong> not<br />

flammable. Thomas Midgley, Jr., <strong>and</strong> Albert Henne of General Motors<br />

Corporation’s Frigidaire Division concluded, from studying<br />

published reports listing properties of a wide variety of chemicals,<br />

that hydrocarbon-like materials with hydrogen atoms replaced by<br />

chlorine <strong>and</strong> fluorine atoms would be appropriate.<br />

Their conclusion led to the formation of a joint effort between the<br />

General Motors Corporation’s Frigidaire Division <strong>and</strong> E. I. Du Pont<br />

de Nemours to research <strong>and</strong> develop the chemistry of fluorocarbons.<br />

In this research effort, a number of scientists began making<br />

<strong>and</strong> studying the large number of individual chemicals in the general<br />

class of compounds being investigated. It fell to Roy J. Plunkett<br />

to do a detailed study of tetrafluoroethylene, a compound consisting<br />

of two carbon atoms, each of which is attached to the other as<br />

well as to two fluorine atoms.<br />

The “Empty” Tank<br />

Tetrafluoroethylene, at normal room temperature <strong>and</strong> pressure,<br />

is a gas that is supplied to users in small pressurized cylinders. On<br />

the morning of the day of the discovery, Plunkett attached such a<br />

tank to his experimental apparatus <strong>and</strong> opened the tank’s valve. To


Teflon / 747<br />

his great surprise, no gas flowed from the tank. Plunkett’s subsequent<br />

actions transformed this event from an experiment gone<br />

wrong into a historically significant discovery. Rather than replacing<br />

the tank with another <strong>and</strong> going on with the work planned for<br />

the day, Plunkett, who wanted to know what had happened, examined<br />

the “empty” tank. When he weighed the tank, he discovered<br />

that it was not empty; it did contain the chemical that was listed on<br />

the label. Opening the valve <strong>and</strong> running a wire through the opening<br />

proved that what had happened had not been caused by a malfunctioning<br />

valve. Finally, Plunkett sawed the cylinder in half <strong>and</strong><br />

discovered what had happened. The chemical in the tank was no<br />

longer a gas; instead, it was a waxy white powder.<br />

Plunkett immediately recognized the meaning of the presence of<br />

the solid. The six-atom molecules of the tetrafluoroethylene gas had<br />

somehow linked with one another to form much larger molecules.<br />

The gas had polymerized, becoming polytetrafluoroethylene, a solid<br />

with a high molecular weight. Capitalizing on this occurrence,<br />

Plunkett, along with other Du Pont chemists, performed a series of<br />

experiments <strong>and</strong> soon learned to control the polymerization reaction<br />

so that the product could be produced, its properties could be<br />

studied, <strong>and</strong> applications for it could be developed.<br />

The properties of the substance were remarkable indeed. It was<br />

unaffected by strong acids <strong>and</strong> bases, withstood high temperatures<br />

without reacting or melting, <strong>and</strong> was not dissolved by any solvent<br />

that the scientists tried. In addition to this highly unusual behavior,<br />

the polymer had surface properties that made it very slick. It was so<br />

slippery that other materials placed on its surface slid off in much<br />

the same way that beads of water slide off the surface of a newly<br />

waxed automobile.<br />

Although these properties were remarkable, no applications were<br />

suggested immediately for the new material. The polymer might<br />

have remained a laboratory curiosity if a conversation had not<br />

taken place between Leslie R. Groves, the head of the Manhattan<br />

Project (which engineered the construction of the first atomic bombs),<br />

<strong>and</strong> a Du Pont chemist who described the polymer to him. The<br />

Manhattan Project research team was hunting for an inert material<br />

to use for gaskets to seal pumps <strong>and</strong> piping. The gaskets had to be<br />

able to withst<strong>and</strong> the highly corrosive uranium hexafluoride with


748 / Teflon<br />

which the team was working. This uranium compound is fundamental<br />

to the process of upgrading uranium for use in explosive devices<br />

<strong>and</strong> power reactors. Polytetrafluoroethylene proved to be just<br />

the material that they needed, <strong>and</strong> Du Pont proceeded, throughout<br />

World War II <strong>and</strong> after, to manufacture gaskets for use in uranium<br />

enrichment plants.<br />

The high level of secrecy of the Manhattan Project in particular<br />

<strong>and</strong> atomic energy in general delayed the commercial introduction<br />

of the polymer, which was called Teflon, until the late 1950’s. At that<br />

time, the first Teflon-coated cooking utensils were introduced.<br />

Impact<br />

Roy J. Plunkett<br />

Roy J. Plunkett was born in 1910 in New Carlisle, Ohio. In<br />

1932 he received a bachelor’s degree in chemistry from Manchester<br />

College <strong>and</strong> transferred to Ohio State University for<br />

graduate school, earning a master’s degree in 1933 <strong>and</strong> a doctorate<br />

in 1936. The same year he went to work for E. I. Du Pont<br />

de Nemours <strong>and</strong> Company as a research chemist at the Jackson<br />

Laboratory in Deepwater, New Jersey. Less then two years later,<br />

when he was only twenty-seven years old, he found the strange<br />

polymer tetrafluoroethylene, whose trade name became Teflon.<br />

It would turn out to be among Du Pont’s most famous products.<br />

In 1938 Du Pont appointed Plunkett the chemical supervisor<br />

at its largest plant, the Chamber Works in Deepwater, which<br />

produced tetraethyl lead. He held the position until 1952 <strong>and</strong><br />

afterward directed the company’s Freon Products Division. He<br />

retired in 1975. In 1985 he was inducted into the Inventor’s Hall<br />

of Fame, <strong>and</strong> after his death in 1994, Du Pont created the<br />

Plunkett Award, presented to inventors who find new uses for<br />

Teflon <strong>and</strong> Tefzel, a related fluoropolymer, in aerospace, automotive,<br />

chemical, or electrical applications.<br />

Plunkett’s thoroughness in following up a chance observation<br />

gave the world a material that has found a wide variety of uses, ranging<br />

from home kitchens to outer space. Some applications make use


Teflon / 749<br />

of Teflon’s slipperiness, others<br />

make use of its inertness,<br />

<strong>and</strong> others take advantage<br />

of both properties.<br />

The best-known application<br />

of Teflon is as a nonstick<br />

coating for cookware.<br />

Teflon’s very slippery surface<br />

initially was troublesome,<br />

when it proved to be<br />

difficult to attach to other<br />

materials. Early versions of<br />

Teflon-coated cookware shed<br />

their surface coatings easily,<br />

even when care was taken<br />

to avoid scraping it off. A<br />

suitable bonding process was<br />

soon developed, however,<br />

<strong>and</strong> the present coated sur-<br />

An important space application for Teflon is its use<br />

faces are very rugged <strong>and</strong><br />

on the outer skins of suits worn by astronauts.<br />

(PhotoDisc)<br />

provide a noncontaminating<br />

coating that can be cleaned<br />

easily.<br />

Teflon has proved to be a useful material in making devices that<br />

are implanted in the human body. It is easily formed into various<br />

shapes <strong>and</strong> is one of the few materials that the human body does not<br />

reject. Teflon has been used to make heart valves, pacemakers, bone<br />

<strong>and</strong> tendon substitutes, artificial corneas, <strong>and</strong> dentures.<br />

Teflon’s space applications have included its use as the outer skin<br />

of the suits worn by astronauts, as insulating coating on wires <strong>and</strong><br />

cables in spacecraft that must resist high-energy cosmic radiation,<br />

<strong>and</strong> as heat-resistant nose cones <strong>and</strong> heat shields on spacecraft.<br />

See also Buna rubber; Neoprene; Nylon; Plastic; Polystyrene;<br />

Pyrex glass; Tupperware.


750 / Teflon<br />

Further Reading<br />

Friedel, Robert. “The Accidental Inventor.” Discover 17, no. 10 (October,<br />

1996).<br />

“Happy Birthday, Teflon.” Design News 44, no. 8 (April, 1988).<br />

“Teflon.” Newsweek 130, 24a (Winter, 1997/1998).


Telephone switching<br />

Telephone switching<br />

The invention: The first completely automatic electronic system<br />

for switching telephone calls.<br />

The people behind the invention:<br />

Almon B. Strowger (1839-1902), an American inventor<br />

Charles Wilson Hoover, Jr. (1925- ), supervisor of memory<br />

system development<br />

Wallace Andrew Depp (1914- ), director of Electronic<br />

Switching<br />

Merton Brown Purvis (1923- ), designer of switching<br />

matrices<br />

Electromechanical Switching Systems<br />

751<br />

The introduction of electronic switching technology into the telephone<br />

network was motivated by the desire to improve the quality<br />

of the telephone system, add new features, <strong>and</strong> reduce the cost of<br />

switching technology. Telephone switching systems have three features:<br />

signaling, control, <strong>and</strong> switching functions. There were several<br />

generations of telephone switching equipment before the first<br />

fully electronic switching “office” (device) was designed.<br />

The first automatic electromechanical (partly electronic <strong>and</strong> partly<br />

mechanical) switching office was the Strowger step-by-step switch.<br />

Strowger switches relied upon the dial pulses generated by rotary<br />

dial telephones to move their switching elements to the proper positions<br />

to connect one telephone with another. In the step-by-step process,<br />

the first digit dialed moved the first mechanical switch into position,<br />

the second digit moved the second mechanical switch into<br />

position, <strong>and</strong> so forth, until the proper telephone connection was established.<br />

These Strowger switching offices were quite large, <strong>and</strong><br />

they lacked flexibility <strong>and</strong> calling features.<br />

The second generation of automatic electromechanical telephone<br />

switching offices was of the “crossbar” type. Initially, crossbar<br />

switches relied upon a specialized electromechanical controller called<br />

a “marker” to establish call connections. Electromechanical telephone


752 / Telephone switching<br />

switching offices had difficulty implementing additional features<br />

<strong>and</strong> were unable to h<strong>and</strong>le large numbers of incoming calls.<br />

Electronic Switching Systems<br />

In the early 1940’s, research into the programmed control of<br />

switching offices began at the American Telephone <strong>and</strong> Telegraph<br />

Company’s Bell Labs. This early research resulted in a trial office being<br />

put into service in Morris, Illinois, in 1960. The Morris switch<br />

used a unique memory called the “flying spot store.” It used a photographic<br />

plate as a program memory, <strong>and</strong> the memory was accessed<br />

optically. In order to change the memory, one had to scratch<br />

out or cover parts of the photographic plate.<br />

Before the development of the Morris switch, gas tubes had been<br />

used to establish voice connections. This was accomplished by applying<br />

a voltage difference across the end points of the conversation.<br />

When this voltage difference was applied, the gas tubes would<br />

conduct electricity, thus establishing the voice connection. The Morris<br />

trial showed that gas tubes could not support the voltages that<br />

the new technology required to make telephones ring or to operate<br />

pay telephones.<br />

The knowledge gained from the Morris trial led to the development<br />

of the first full-scale, commercial, computer-controlled<br />

electronic switch, the electronic switching system 1 (ESS-1). The<br />

first ESS-1 went into service in New Jersey in 1965. In the ESS-1,<br />

electromechanical switching elements, or relays, were controlled<br />

by computer software. A centralized computer h<strong>and</strong>led call processing.<br />

Because the telephone service of an entire community<br />

depends on the reliability of the telephone switching office, the<br />

ESS-1 had two central processors, so that one would be available<br />

if the other broke down. The switching system of the ESS-1 was<br />

composed of electromechanical relays; the control of the switching<br />

system was electronic, but the switching itself remained mechanical.<br />

Bell Labs developed models to demonstrate the concept of integrating<br />

digital transmission <strong>and</strong> switching systems. Unfortunately,<br />

the solid state electronics necessary for such an undertaking had not<br />

developed sufficiently at that time, so the commercial development


Almon B. Strowger<br />

Telephone switching / 753<br />

Some people thought Almon B. Strowger was strange, perhaps<br />

even demented. Certainly, he was hot-tempered, restless,<br />

<strong>and</strong> argumentative. One thing he was not, however, was unimaginative.<br />

Born near Rochester, New York, in 1839, Strowger was old<br />

enough to fight for the Union at the second battle of Manassas<br />

during the American Civil War. The bloody battle apparently<br />

shattered <strong>and</strong> embittered him. He w<strong>and</strong>ered slowly west after<br />

the war, taught himself undertaking, <strong>and</strong> opened a funeral<br />

home in Topeka, Kansas, in 1882. There began his running war<br />

with telephone operators, which continued when he moved his<br />

business to Kansas City.<br />

With the help of technicians (whom he later cheated) he built<br />

the first “collar box,” an automatic switching device, in 1887.<br />

The round contraption held a pencil that could be revolved to<br />

different pins arrange around it in order to change phone connections.<br />

Two years later he produced a more sophisticated device<br />

that was operated by push-button, <strong>and</strong> despite initial misgivings<br />

brought out a rotary dial device in 1896. That same year<br />

he sold the rights to his patents to business partners for $1,800<br />

<strong>and</strong> his share in Strowger Automatic Dial Telephone Exchange<br />

for $10,000 in 1898. He moved to St. Petersburg, Florida, <strong>and</strong><br />

opened a small hotel, dying there in 1902. It surely would have<br />

done his temper no good to learn that fourteen years later the<br />

Bell system bought his patents for $2.5 million.<br />

of digital switching was not pursued. New versions of the ESS continued<br />

to employ electromechanical technology, although mechanical<br />

switching elements can cause impulse noise in voice signals <strong>and</strong><br />

are larger <strong>and</strong> more difficult to maintain than electronic switching<br />

elements. Ten years later, however, Bell Labs began to develop a digital<br />

toll switch, the ESS-4, in which both switching <strong>and</strong> control functions<br />

were electronic.<br />

Although the ESS-1 was the first electronically controlled switching<br />

system, it did not switch voices electronically. The ESS-1 used<br />

computer control to move mechanical contacts in order to establish<br />

a conversation. In a fully electronic switching system, the voices are


754 / Telephone switching<br />

digitized before switching is performed. This technique, which is<br />

called “digital switching,” is still used.<br />

The advent of electronically controlled switching systems made<br />

possible features such as call forwarding, call waiting, <strong>and</strong> detailed<br />

billing for long-distance calls. Changing these services became a<br />

matter of simply changing tables in computer programs. Telephone<br />

maintenance personnel could communicate with the central processor<br />

of the ESS-1 by using a teletype, <strong>and</strong> they could change numbers<br />

simply by typing comm<strong>and</strong>s on the teletype. In electromechanically<br />

controlled telephone switching systems, however, changing numbers<br />

required rewiring.<br />

Consequences<br />

Electronic switching has greatly decreased the size of switching<br />

offices. Digitization of the voice prior to transmission improves<br />

voice quality. When telephone switches were electromechanical, a<br />

large area was needed to house the many mechanical switches that<br />

were required. In the era of electronic switching, voices are switched<br />

digitally by computer. In this method, voice samples are read into a<br />

computer memory <strong>and</strong> then read out of the memory when it is time<br />

to connect a caller with a desired number. Basically, electronic telephone<br />

systems are specialized computer systems that move digitized<br />

voice samples between customers.<br />

Telephone networks are moving toward complete digitization.<br />

Digitization was first applied to the transmission of voice signals.<br />

This made it possible for a single pair of copper wires to be shared<br />

by a number of telephone users. Currently, voices are digitized<br />

upon their arrival at the switching office. If the final destination of<br />

the telephone call is not connected to the particular switching office,<br />

the voice is sent to the remote office by means of digital circuits.<br />

Currently, voice signals are sent between the switching office <strong>and</strong><br />

homes or businesses. In the future, digitization of the voice signal<br />

will occur in the telephone sets themselves. Digital voice signals<br />

will be sent directly from one telephone to another. This will provide<br />

homes with direct digital communication. A network that provides<br />

such services is called the “integrated services digital network”<br />

(ISDN).


See also Cell phone; Long-distance telephone; Rotary dial telephone;<br />

Touch-tone telephone.<br />

Further Reading<br />

Telephone switching / 755<br />

Briley, Bruce E. Introduction to Telephone Switching. Reading, Mass.:<br />

Addison-Wesley, 1983.<br />

Talley, David. Basic Electronic Switching for Telephone Systems. 2ded.<br />

Rochelle Park, N.J.: Hayden, 1982.<br />

Thompson, Richard A. Telephone Switching Systems. Boston: Artech<br />

House, 2000.


756<br />

Television<br />

Television<br />

The invention: System that converts moving pictures <strong>and</strong> sounds<br />

into electronic signals that can be broadcast at great distances.<br />

The people behind the invention:<br />

Vladimir Zworykin (1889-1982), a Soviet electronic engineer <strong>and</strong><br />

recipient of the National Medal of Science in 1967<br />

Paul Gottlieb Nipkow (1860-1940), a German engineer <strong>and</strong><br />

inventor<br />

Alan A. Campbell Swinton (1863-1930), a Scottish engineer <strong>and</strong><br />

Fellow of the Royal Society<br />

Charles F. Jenkins (1867-1934), an American physicist, engineer,<br />

<strong>and</strong> inventor<br />

The Persistence of Vision<br />

In 1894, an American inventor, Charles F. Jenkins, described a<br />

scheme for electrically transmitting moving pictures. Jenkins’s idea,<br />

however, was only one in an already long tradition of theoretical<br />

television systems. In 1842, for example, the English physicist Alex<strong>and</strong>er<br />

Bain had invented an automatic copying telegraph for sending<br />

still pictures. Bain’s system scanned images line by line. Similarly,<br />

the wide recognition of the persistence of vision—the mind’s<br />

ability to retain a visual image for a short period of time after the image<br />

has been removed—led to experiments with systems in which<br />

the image to be projected was repeatedly scanned line by line. Rapid<br />

scanning of images became the underlying principle of all television<br />

systems, both electromechanical <strong>and</strong> all-electronic.<br />

In 1884, a German inventor, Paul Gottlieb Nipkow, patented a<br />

complete television system that utilized a mechanical sequential<br />

scanning system <strong>and</strong> a photoelectric cell sensitized with selenium<br />

for transmission. The selenium photoelectric cell converted the light<br />

values of the image being scanned into electrical impulses to be<br />

transmitted to a receiver where the process would be reversed. The<br />

electrical impulses led to light of varying brightnesses being produced<br />

<strong>and</strong> projected on to a rotating disk that was scanned to repro-


Electron Gun<br />

Electron Beam<br />

Deflection <strong>and</strong> Focus Coils<br />

Phosphor Screen<br />

Glass Envelope<br />

Schematic of a television picture tube.<br />

Television / 757<br />

duce the original image. If the system—that is, the transmitter <strong>and</strong><br />

the receiver—were in perfect synchronization <strong>and</strong> if the disk rotated<br />

quickly enough, persistence of vision enabled the viewer to<br />

see a complete image rather than a series of moving points of light.<br />

For a television image to be projected onto a screen of reasonable<br />

size <strong>and</strong> retain good quality <strong>and</strong> high resolution, any system employing<br />

only thirty to one hundred lines (as early mechanical systems<br />

did) is inadequate. Afew systems were developed that utilized<br />

two hundred or more lines, but the difficulties these presented<br />

made the possibility of an all-electronic system increasingly attractive.<br />

These difficulties were not generally recognized until the early<br />

1930’s, when television began to move out of the laboratory <strong>and</strong> into<br />

commercial production.<br />

Interest in all-electronic television paralleled interest in mechanical<br />

systems, but solutions to technical problems proved harder to<br />

achieve. In 1908, a Scottish engineer, Alan A. Campbell Swinton,<br />

proposed what was essentially an all-electronic television system.<br />

Swinton theorized that the use of magnetically deflected cathoderay<br />

tubes for both the transmitter <strong>and</strong> receiver in a system was possible.<br />

In 1911, Swinton formally presented his idea to the Röntgen


758 / Television<br />

Society in London, but the technology available did not allow for<br />

practical experiments.<br />

Zworykin’s Picture Tube<br />

Vladimir Zworykin<br />

Born in 1889, Vladimir Kosma Zworykin grew up in Murom,<br />

a small town two hundred miles east of Moscow. His father ran<br />

a riverboat service, <strong>and</strong> Zworykin sometimes helped him, but<br />

his mind was on electricity, which he studied on his own while<br />

aboard his father’s boats. In 1906, he entered the St. Petersburg<br />

Institute of Technology, <strong>and</strong> there he became acquainted with<br />

the idea of television through the work of Professor Boris von<br />

Rosing.<br />

Zworykin assisted Rosing in his attempts to transmit pictures<br />

with a cathode-ray tube. He served with the Russian Signal<br />

Corps during World War I, but then fled to the United States<br />

after the Bolshevist Revolution. In 1920 he got a job at Westinghouse’s<br />

research laboratory in Pittsburgh, helping develop radio<br />

tubes <strong>and</strong> photoelectric cells. He became an American citizen<br />

in 1924 <strong>and</strong> completed a doctorate at the University of<br />

Pittsburgh in 1926. By then he had already demonstrated his<br />

iconoscope <strong>and</strong> applied for a patent. Unable to interest Westinghouse<br />

in his invention, he moved to the Radio Corporation<br />

of America (RCA) in 1929, <strong>and</strong> later became director of its electronics<br />

research laboratory. RCA’s president, David Sarnoff,<br />

also a Russian immigrant, had faith in Zworykin <strong>and</strong> his ideas.<br />

Before Zworykin retired in 1954, RCA had invested $50 million<br />

in television.<br />

Among the many awards Zworykin received for his culturechanging<br />

invention was the National Medal of Science, presented<br />

by President Lyndon Johnson in 1966. Zworykin died on<br />

his birthday in 1982.<br />

In 1923, Vladimir Zworykin, a Soviet electronic engineer working<br />

for the Westinghouse Electric Corporation, filed a patent application<br />

for the “iconoscope,” or television transmission tube. On<br />

March 17, 1924, Zworykin applied for a patent for a two-way system.<br />

The first cathode-ray tube receiver had a cathode, a modulating<br />

grid, an anode, <strong>and</strong> a fluorescent screen.


Early console model television. (PhotoDisc)<br />

Television / 759<br />

Zworykin later admitted that the results were very poor <strong>and</strong> the<br />

system, as shown, was still far removed from a practical television<br />

system. Zworykin’s employers were so unimpressed that they admonished<br />

him to forget television <strong>and</strong> work on something more<br />

useful. Zworykin’s interest in television was thereafter confined to<br />

his nonworking hours, as he spent the next year working on photographic<br />

sound recording.<br />

It was not until the late 1920’s that he was able to devote his full<br />

attention to television. Ironically, Westinghouse had by then resumed<br />

research in television, but Zworykin was not part of the<br />

team. After he returned from a trip to France, where in 1928 he had<br />

witnessed an exciting demonstration of an electrostatic tube, Westinghouse<br />

indicated that it was not interested. This lack of corporate<br />

support in Pittsburgh led Zworykin to approach the Radio Corporation<br />

of America (RCA). According to reports, Zworykin demonstrated<br />

his system to the Institute of Radio Engineers at Rochester,<br />

New York, on November 18, 1929, claiming to have developed a


760 / Television<br />

working picture tube, a tube that would revolutionize television development.<br />

Finally, RCA recognized the potential.<br />

Impact<br />

The picture tube, or “kinescope,” developed by Zworykin changed<br />

the history of television. Within a few years, mechanical systems<br />

disappeared <strong>and</strong> television technology began to utilize systems<br />

similar to Zworykin’s by use of cathode-ray tubes at both ends of<br />

the system. At the transmitter, the image is focused upon a mosaic<br />

screen composed of light-sensitive cells. A stream of electrons sweeps<br />

the image, <strong>and</strong> each cell sends off an electric current pulse as it is hit<br />

by the electrons, the light <strong>and</strong> shade of the focused image regulating<br />

the amount of current.<br />

This string of electrical impulses, after amplification <strong>and</strong> modification<br />

into ultrahigh frequency wavelengths, is broadcast by antenna<br />

to be picked up by any attuned receiver, where it is retransformed<br />

into a moving picture in the cathode-ray tube receiver. The<br />

cathode-ray tubes contain no moving parts, as the electron stream is<br />

guided entirely by electric attraction.<br />

Although both the iconoscope <strong>and</strong> the kinescope were far from<br />

perfect when Zworykin initially demonstrated them, they set the<br />

stage for all future television development.<br />

See also Color television; Community antenna television; Communications<br />

satellite; Fiber-optics; FM radio; Holography; Internet;<br />

Radio; Talking motion pictures.<br />

Further Reading<br />

Abramson, Albert. Zworykin: Pioneer of Television. Urbana: University<br />

of Illinois Press, 1995.<br />

Sconce, Jeffrey. Haunted Media: Electronic Presence from Telegraphy to<br />

Television. Durham, N.C.: Duke University Press, 2000.<br />

Zworykin, Vladimir Kosma, <strong>and</strong> George Ashmun Morton. Television:<br />

The Electronics of Image Transmission in Color <strong>and</strong> Monochrome.<br />

2d ed. New York: J. Wiley, 1954.


Tevatron accelerator<br />

Tevatron accelerator<br />

The invention: A particle accelerator that generated collisions between<br />

beams of protons <strong>and</strong> antiprotons at the highest energies<br />

ever recorded.<br />

The people behind the invention:<br />

Robert Rathbun Wilson (1914- ), an American physicist <strong>and</strong><br />

director of Fermilab from 1967 to 1978<br />

John Peoples (1933- ), an American physicist <strong>and</strong> deputy<br />

director of Fermilab from 1987<br />

Putting Supermagnets to Use<br />

761<br />

The Tevatron is a particle accelerator, a large electromagnetic device<br />

used by high-energy physicists to generate subatomic particles<br />

at sufficiently high energies to explore the basic structure of matter.<br />

The Tevatron is a circular, tubelike track 6.4 kilometers in circumference<br />

that employs a series of superconducting magnets to accelerate<br />

beams of protons, which carry a positive charge in the atom, <strong>and</strong><br />

antiprotons, the proton’s negatively charged equivalent, at energies<br />

up to 1 trillion electronvolts (equal to 1 teraelectronvolt, or 1 TeV;<br />

hence the name Tevatron). An electronvolt is the unit of energy that<br />

an electron gains through an electrical potential of 1 volt.<br />

The Tevatron is located at the Fermi National Accelerator Laboratory,<br />

which is also known as Fermilab. The laboratory was one of<br />

several built in the United States during the 1960’s.<br />

The heart of the original Fermilab was the 6.4-kilometer main accelerator<br />

ring. This main ring was capable of accelerating protons to<br />

energies approaching 500 billion electronvolts, or 0.5 teraelectronvolt.<br />

The idea to build the Tevatron grew out of a concern for the<br />

millions of dollars spent annually on electricity to power the main<br />

ring, the need for higher energies to explore the inner depths of the<br />

atom <strong>and</strong> the consequences of new theories of both matter <strong>and</strong> energy,<br />

<strong>and</strong> the growth of superconductor technology. Planning for a<br />

second accelerator ring, the Tevatron, to be installed beneath the<br />

main ring began in 1972.


762 / Tevatron accelerator<br />

Robert Rathbun Wilson, the director of Fermilab at that time, realized<br />

that the only way the laboratory could achieve the higher energies<br />

needed for future experiments without incurring intolerable<br />

electricity costs was to design a second accelerator ring that employed<br />

magnets made of superconducting material. Extremely powerful<br />

magnets are the heart of any particle accelerator; charged particles<br />

such as protons are given a “push” as they pass through an electromagnetic<br />

field. Each successive push along the path of the circular<br />

accelerator track gives the particle more <strong>and</strong> more energy. The enormous<br />

magnetic fields required to accelerate massive particles such<br />

as protons to energies approaching 1 trillion electronvolts would require<br />

electricity expenditures far beyond Fermilab’s operating budget.<br />

Wilson estimated that using superconducting materials, however,<br />

which have virtually no resistance to electrical current, would<br />

make it possible for the Tevatron to achieve double the main ring’s<br />

magnetic field strength, doubling energy output without significantly<br />

increasing energy costs.<br />

Tevatron to the Rescue<br />

The Tevatron was conceived in three phases. Most important,<br />

however, were Tevatron I <strong>and</strong> Tevatron II, where the highest energies<br />

were to be generated <strong>and</strong> where it was hoped new experimental findings<br />

would emerge. Tevatron II experiments were designed to be<br />

very similar to other proton beam experiments, except that in this<br />

case, the protons would be accelerated to an energy of 1 trillion<br />

electronvolts. More important still are the proton-antiproton colliding<br />

beam experiments of Tevatron I. In this phase, beams of protons<br />

<strong>and</strong> antiprotons rotating in opposite directions are caused to collide<br />

in the Tevatron, producing a combined, or center-of-mass, energy<br />

approaching 2 trillion electronvolts, nearly three times the energy<br />

achievable at the largest accelerator at Centre Européen de Recherche<br />

Nucléaire (the European Center for Nuclear Research, or CERN).<br />

John Peoples was faced with the problem of generating a beam of<br />

antiprotons of sufficient intensity to collide efficiently with a beam<br />

of protons. Knowing that he had the use of a large proton accelerator—the<br />

old main ring—Peoples employed the two-ring mode in<br />

which 120 billion electronvolt protons from the main ring are aimed


at a fixed tungsten target, generating antiprotons, which scatter<br />

from the target. These particles were extracted <strong>and</strong> accumulated in a<br />

smaller storage ring. These particles could be accelerated to relatively<br />

low energies. After sufficient numbers of antiprotons were<br />

collected, they were injected into the Tevatron, along with a beam of<br />

protons for the colliding beam experiments. On October 13, 1985,<br />

Fermilab scientists reported a proton-antiproton collision with a<br />

center-of-mass energy measured at 1.6 trillion electronvolts, the<br />

highest energy ever recorded.<br />

Consequences<br />

Tevatron accelerator / 763<br />

The Tevatron’s success at generating high-energy protonantiproton<br />

collisions affected future plans for accelerator development<br />

in the United States <strong>and</strong> offered the potential for important<br />

discoveries in high-energy physics at energy levels that no other accelerator<br />

could achieve.<br />

Physics recognized four forces in nature: the electromagnetic<br />

force, the gravitational force, the strong nuclear force, <strong>and</strong> the weak<br />

nuclear force. A major goal of the physics community is to formulate<br />

a theory that will explain all these forces: the so-called gr<strong>and</strong><br />

unification theory. In 1967, one of the first of the so-called gauge theories<br />

was developed that unified the weak nuclear force <strong>and</strong> the<br />

electromagnetic force. One consequence of this theory was that the<br />

weak force was carried by massive particles known as “bosons.”<br />

The search for three of these particles—the intermediate vector bosons<br />

W + ,W − , <strong>and</strong> Z 0 —led to the rush to conduct colliding beam experiments<br />

to the early 1970’s. Because the Tevatron was in the planning<br />

phase at this time, these particles were discovered by a team of<br />

international scientists based in Europe. In 1989, Tevatron physicists<br />

reported the most accurate measure to date of the Z 0 mass.<br />

The Tevatron is thought to be the only particle accelerator in the<br />

world with sufficient power to conduct further searches for the elusive<br />

Higgs boson, a particle attributed to weak interactions by University<br />

of Edinburgh physicist Peter Higgs in order to account for<br />

the large masses of the intermediate vector bosons. In addition, the<br />

Tevatron has the ability to search for the so-called top quark. Quarks<br />

are believed to be the constituent particles of protons <strong>and</strong> neutrons.


764 / Tevatron accelerator<br />

Evidence has been gathered of five of the six quarks believed to exist.<br />

Physicists have yet to detect evidence of the most massive quark,<br />

the top quark.<br />

See also Atomic bomb; Cyclotron; Electron microscope; Field ion<br />

microscope; Geiger counter; Hydrogen bomb; Mass spectrograph;<br />

Neutrino detector; Scanning tunneling microscope; Synchrocyclotron.<br />

Further Reading<br />

Hilts, Philip J. Scientific Temperaments: Three Lives in Contemporary<br />

Science. New York: Simon <strong>and</strong> Schuster, 1984.<br />

Ladbury, Ray. “Fermilab Tevatron Collider Group Goes over the<br />

Top—Cautiously.” Physics Today 47, no. 6 (June, 1994).<br />

Lederman, Leon M. “The Tevatron.” Scientific American 264, no. 3<br />

(March, 1991).<br />

Wilson, Robert R., <strong>and</strong> Raphael Littauer. Accelerators: Machines of<br />

Nuclear Physics. London: Heinemann, 1962.


Thermal cracking process<br />

Thermal cracking process<br />

The invention: Process that increased the yield of refined gasoline<br />

extracted from raw petroleum by using heat to convert complex<br />

hydrocarbons into simpler gasoline hydrocarbons, thereby making<br />

possible the development of the modern petroleum industry.<br />

The people behind the invention:<br />

William M. Burton (1865-1954), an American chemist<br />

Robert E. Humphreys (1942- ), an American chemist<br />

Gasoline, Motor Vehicles, <strong>and</strong> Thermal Cracking<br />

765<br />

Gasoline is a liquid mixture of hydrocarbons (chemicals made up<br />

of only hydrogen <strong>and</strong> carbon) that is used primarily as a fuel for internal<br />

combustion engines. It is produced by petroleum refineries<br />

that obtain it by processing petroleum (crude oil), a naturally occurring<br />

mixture of thous<strong>and</strong>s of hydrocarbons, the molecules of which<br />

can contain from one to sixty carbon atoms.<br />

Gasoline production begins with the “fractional distillation” of<br />

crude oil in a fractionation tower, where it is heated to about 400 degrees<br />

Celsius at the tower’s base. This heating vaporizes most of the<br />

hydrocarbons that are present, <strong>and</strong> the vapor rises in the tower,<br />

cooling as it does so. At various levels of the tower, various portions<br />

(fractions) of the vapor containing simple hydrocarbon mixtures become<br />

liquid again, are collected, <strong>and</strong> are piped out as “petroleum<br />

fractions.” Gasoline, the petroleum fraction that boils between 30<br />

<strong>and</strong> 190 degrees Celsius, is mostly a mixture of hydrocarbons that<br />

contain five to twelve carbon atoms.<br />

Only about 25 percent of petroleum will become gasoline via<br />

fractional distillation. This amount of “straight run” gasoline is not<br />

sufficient to meet the world’s needs. Therefore, numerous methods<br />

have been developed to produce the needed amounts of gasoline.<br />

The first such method, “thermal cracking,” was developed in 1913<br />

by William M. Burton of St<strong>and</strong>ard Oil of Indiana. Burton’s cracking<br />

process used heat to convert complex hydrocarbons (whose molecules<br />

contain many carbon atoms) into simpler gasoline hydrocar-


766 / Thermal cracking process<br />

bons (whose molecules contain fewer carbon atoms), thereby increasing<br />

the yield of gasoline from petroleum. Later advances in<br />

petroleum technology, including both an improved Burton method<br />

<strong>and</strong> other methods, increased the gasoline yield still further.<br />

More Gasoline!<br />

Starting in about 1900, gasoline became important as a fuel for<br />

the internal combustion engines of the new vehicles called automobiles.<br />

By 1910, half a million automobiles traveled American roads.<br />

Soon, the great dem<strong>and</strong> for gasoline—which was destined to grow<br />

<strong>and</strong> grow—required both the discovery of new crude oil fields<br />

around the world <strong>and</strong> improved methods for refining the petroleum<br />

mined from these new sources. Efforts were made to increase<br />

the yield of gasoline—at that time, about 15 percent—from petroleum.<br />

The Burton method was the first such method.<br />

At the time that the cracking process was developed, Burton was<br />

the general superintendent of the Whiting refinery, owned by St<strong>and</strong>ard<br />

Oil of Indiana. The Burton process was developed in collaboration<br />

with Robert E. Humphreys <strong>and</strong> F. M. Rogers. This three-person<br />

research group began work knowing that heating petroleum<br />

fractions that contained hydrocarbons more complex than those<br />

present in gasoline—a process called “coking”—produced kerosene,<br />

coke (a form of carbon), <strong>and</strong> a small amount of gasoline. The<br />

process needed to be improved substantially, however, before it<br />

could be used commercially.<br />

Initially, Burton <strong>and</strong> his coworkers used the “heavy fuel” fraction<br />

of petroleum (the 66 percent of petroleum that boils at a temperature<br />

higher than the boiling temperature of kerosene). Soon, they<br />

found that it was better to use only the part of the material that contained<br />

its smaller hydrocarbons (those containing fewer carbon atoms),<br />

all of which were still much larger than those present in gasoline.<br />

The cracking procedure attempted first involved passing the<br />

starting material through a hot tube. This hot-tube treatment vaporized<br />

the material <strong>and</strong> broke down 20 to 30 percent of the larger hydrocarbons<br />

into the hydrocarbons found in gasoline. Various tarry<br />

products were also produced, however, that reduced the quality of<br />

the gasoline that was obtained in this way.


Asphalt Industrial<br />

Fuel Oil<br />

Roofing<br />

Paints<br />

CRUDE OIL IN<br />

PETROLEUM REFINERY<br />

Thermal cracking process / 767<br />

SEPARATING PURIFICATION CONVERSION<br />

C<strong>and</strong>les<br />

Polish<br />

Waxed Paper<br />

Ointments<br />

<strong>and</strong><br />

Creams<br />

De-waxing<br />

Lubricants<br />

<strong>and</strong><br />

Greases<br />

Plastics<br />

Photographic<br />

Film<br />

Synthetic<br />

Rubber<br />

Weed-killers<br />

<strong>and</strong><br />

Fertilizers<br />

Diesel<br />

Oils<br />

Cracking<br />

Medicines<br />

Detergents<br />

Enamel<br />

Synthetic<br />

Fibers<br />

Next, the investigators attempted to work at a higher temperature<br />

by bubbling the starting material through molten lead. More<br />

gasoline was made in this way, but it was so contaminated with<br />

gummy material that it could not be used. Continued investigation<br />

showed, however, that moderate temperatures (between those used<br />

in the hot-tube experiments <strong>and</strong> that of molten lead) produced the<br />

best yield of useful gasoline.<br />

The Burton group then had the idea of using high pressure to<br />

“keep starting materials still.” Although the theoretical basis for the<br />

use of high pressure was later shown to be incorrect, the new<br />

method worked quite well. In 1913, the Burton method was patented<br />

<strong>and</strong> put into use. The first cracked gasoline, called Motor<br />

Spirit, was not very popular, because it was yellowish <strong>and</strong> had a<br />

somewhat unpleasant odor. The addition of some minor refining<br />

procedures, however, soon made cracked gasoline indistinguishable<br />

from straight run gasoline. St<strong>and</strong>ard Oil of Indiana made huge<br />

profits from cracked gasoline over the next ten years. Ultimately,<br />

thermal cracking subjected the petroleum fractions that were uti-<br />

Fuel<br />

Oil<br />

Bottled<br />

Gas<br />

Gasoline<br />

Jet Fuel<br />

Solvents<br />

Insecticides<br />

Burton’s process contributed to the development of petroleum refining, shown in this<br />

diagram.


768 / Thermal cracking process<br />

lized to temperatures between 550 <strong>and</strong> 750 degrees Celsius, under<br />

pressures between 250 <strong>and</strong> 750 pounds per square inch.<br />

Impact<br />

In addition to using thermal cracking to make gasoline for sale,<br />

St<strong>and</strong>ard Oil of Indiana also profited by licensing the process for use<br />

by other gasoline producers. Soon, the method was used throughout<br />

the oil industry. By 1920, it had been perfected as much as it<br />

could be, <strong>and</strong> the gasoline yield from petroleum had been significantly<br />

increased. The disadvantages of thermal cracking include a<br />

relatively low yield of gasoline (compared to those of other methods),<br />

the waste of hydrocarbons in fractions converted to tar <strong>and</strong><br />

coke, <strong>and</strong> the relatively high cost of the process.<br />

A partial solution to these problems was found in “catalytic<br />

cracking”—the next logical step from the Burton method—in which<br />

petroleum fractions to be cracked are mixed with a catalyst (a substance<br />

that causes a chemical reaction to proceed more quickly,<br />

without reacting itself). The most common catalysts used in such<br />

cracking were minerals called “zeolites.” The wide use of catalytic<br />

cracking soon enabled gasoline producers to work at lower temperatures<br />

(450 to 550 degrees Celsius) <strong>and</strong> pressures (10 to 50 pounds<br />

per square inch). This use decreased manufacturing costs because<br />

catalytic cracking required relatively little energy, produced only<br />

small quantities of undesirable side products, <strong>and</strong> produced highquality<br />

gasoline.<br />

Various other methods of producing gasoline have been developed—among<br />

them catalytic reforming, hydrocracking, alkylation,<br />

<strong>and</strong> catalytic isomerization—<strong>and</strong> now about 60 percent of the petroleum<br />

starting material can be turned into gasoline. These methods,<br />

<strong>and</strong> others still to come, are expected to ensure that the world’s<br />

needs for gasoline will continue to be satisfied—as long as petroleum<br />

remains available.<br />

See also Fuel cell; Gas-electric car; Geothermal power; Internal<br />

combustion engine; Oil-well drill bit; Solar thermal engine.


Further Reading<br />

Thermal cracking process / 769<br />

Gorman, Hugh S. Redefining Efficiency: Pollution Concerns, Regulatory<br />

Mechanisms, <strong>and</strong> Technological Change in the U.S. Petroleum Industry.<br />

Akron, Ohio: University of Akron Press, 2001.<br />

Sung, Hsun-chang, Robert Roy White, <strong>and</strong> George Granger Brown.<br />

Thermal Cracking of Petroleum. Ann Arbor: University of Michigan,<br />

1945.<br />

William Meriam Burton: A Pioneer in Modern Petroleum Technology.<br />

Cambridge, Mass.: University Press, 1952.


770<br />

Tidal power plant<br />

Tidal power plant<br />

The invention: Plant that converts the natural ocean tidal forces<br />

into electrical power.<br />

The people behind the invention:<br />

Mariano di Jacopo detto Taccola (Mariano of Siena, 1381-1453),<br />

an Italian notary, artist, <strong>and</strong> engineer<br />

Bernard Forest de Bélidor (1697 or 1698-1761), a French engineer<br />

Franklin D. Roosevelt (1882-1945), president of the United States<br />

Tidal Enersgy<br />

Ocean tides have long been harnessed to perform useful work.<br />

Ancient Greeks, Romans, <strong>and</strong> medieval Europeans all left records<br />

<strong>and</strong> ruins of tidal mills, <strong>and</strong> Mariano di Jacopo included tidal power<br />

in his treatise De Ingeneis (1433; on engines). Some mills consisted of<br />

water wheels suspended in tidal currents, others lifted weights that<br />

powered machinery as they fell, <strong>and</strong> still others trapped the high<br />

tide to run a mill.<br />

Bernard Forest de Bélidor’s Architecture hydraulique (1737; hydraulic<br />

architecture) is often cited as initiating the modern era of<br />

tidal power exploitation. Bélidor was an instructor in the French<br />

École d’Artillerie et du Génie (School of Artillery <strong>and</strong> Engineering).<br />

Industrial expansion between 1700 <strong>and</strong> 1800 led to the construction<br />

of many tidal mills. In these mills, waterwheels or simple turbines<br />

rotated shafts that drove machinery by means of gears or<br />

belts. They powered small enterprises located on the seashore.<br />

Steam engines, however, soon began to replace tidal mills. Steam<br />

could be generated wherever it was needed, <strong>and</strong> steam mills were<br />

not dependent upon the tides or limited in their production capacity<br />

by the amount of tidal flow. Thus, tidal mills gradually were ab<strong>and</strong>oned,<br />

although a few still operate in New Engl<strong>and</strong>, Great Britain,<br />

France, <strong>and</strong> elsewhere.


Electric Power from Tides<br />

Tidal power plant / 771<br />

Modern society requires tremendous amounts of electric energy<br />

generated by large power stations. This need was first met by<br />

using coal <strong>and</strong> by damming rivers. Later, oil <strong>and</strong> nuclear power became<br />

important. Although small mechanical tidal mills are inadequate<br />

for modern needs, tidal power itself remains an attractive<br />

source of energy. Periodic alarms about coal or oil supplies <strong>and</strong><br />

concern about the negative effects on the environment of using<br />

coal, oil, or nuclear energy continue to stimulate efforts to develop<br />

renewable energy sources with fewer negative effects. Every crisis—for<br />

example, the perceived European coal shortages in the<br />

early 1900’s, oil shortages in the 1920’s <strong>and</strong> 1970’s, <strong>and</strong> growing<br />

anxiety about nuclear power—revives interest in tidal power.<br />

In 1912, a tidal power plant was proposed at Busum, Germany.<br />

The English, in 1918 <strong>and</strong> more recently, promoted elaborate schemes<br />

for the Severn Estuary. In 1928, the French planned a plant at Aber-<br />

Wrach in Brittany. In 1935, under the leadership of Franklin Delano<br />

Roosevelt, the United States began construction of a tidal power<br />

plant at Passamaquoddy, Maine. These plants, however, were never<br />

built. All of them had to be located at sites where tides were extremely<br />

high, <strong>and</strong> such sites are often far from power users. So<br />

much electricity was lost in transmission that profitable quantities<br />

of power could not be sent where they were needed. Also, large<br />

tidal power stations were too expensive to compete with existing<br />

steam plants <strong>and</strong> river dams. In addition, turbines <strong>and</strong> generators<br />

capable of using the large volumes of slow-moving tidal water that<br />

reversed flow had not been invented. Finally, large tidal plants inevitably<br />

hampered navigation, fisheries, recreation, <strong>and</strong> other uses<br />

of the sea <strong>and</strong> shore.<br />

French engineers, especially Robert Gibrat, the father of the La<br />

Rance project, have made the most progress in solving the problems<br />

of tidal power plants. France, a highly industrialized country, is<br />

short of coal <strong>and</strong> petroleum, which has brought about an intense<br />

search by the French for alternative energy supplies.<br />

La Rance, which was completed in December, 1967, is the first<br />

full-scale tidal electric power plant in the world. The Chinese, however,<br />

have built more than a hundred small tidal electric stations


772 / Tidal power plant<br />

about the size of the old mechanical tidal mills, <strong>and</strong> the Canadians<br />

<strong>and</strong> the Russians have both operated plants of pilot-plant size.<br />

La Rance, which was selected from more than twenty competing<br />

localities in France, is one of a few places in the world where the<br />

tides are extremely high. It also has a large reservoir that is located<br />

above a narrow constriction in the estuary. Finally, interference with<br />

navigation, fisheries, <strong>and</strong> recreational activities is minimal at La<br />

Rance.<br />

Submersible “bulbs” containing generators <strong>and</strong> mounting propeller<br />

turbines were specially designed for the La Rance project.<br />

These turbines operate using both incoming <strong>and</strong> outgoing tides,<br />

<strong>and</strong> they can pump water either into or out of the reservoir. These<br />

features allow daily <strong>and</strong> seasonal changes in power generation to be<br />

“smoothed out.” These turbines also deliver electricity most economically.<br />

Many engineering problems had to be solved, however,<br />

before the dam could be built in the tidal estuary.<br />

The La Rance plant produces 240 megawatts of electricity. Its<br />

twenty-four highly reliable turbine generator sets operate about 95<br />

percent of the time. Output is coordinated with twenty-four other<br />

hydroelectric plants by means of a computer program. In this system,<br />

pump-storage stations use excess La Rance power during periods<br />

of low dem<strong>and</strong> to pump water into elevated reservoirs. Later,<br />

during peak dem<strong>and</strong>, this water is fed through a power plant, thus<br />

“saving” the excess generated at La Rance when it was not immediately<br />

needed. In this way, tidal energy, which must be used or lost as<br />

the tides continue to flow, can be saved.<br />

Consequences<br />

The operation of La Rance proved the practicality of tide-generated<br />

electricity. The equipment, engineering practices, <strong>and</strong> operating<br />

procedures invented for La Rance have been widely applied. Submersible,<br />

low-head, high-flow reversible generators of the La Rance<br />

type are now used in Austria, Switzerl<strong>and</strong>, Sweden, Russia, Canada,<br />

the United States, <strong>and</strong> elsewhere.<br />

Economic problems have prevented the building of more large<br />

tidal power plants. With technological advances, the inexorable<br />

depletion of oil <strong>and</strong> coal resources, <strong>and</strong> the increasing cost of nu-


clear power, tidal power may be used more widely in the future.<br />

Construction costs may be significantly lowered by using preconstructed<br />

power units <strong>and</strong> dam segments that are floated into place<br />

<strong>and</strong> submerged, thus making unnecessary expensive dams <strong>and</strong> reducing<br />

pumping costs.<br />

See also Compressed-air-accumulating power plant; Geothermal<br />

power; Nuclear power plant; Nuclear reactor; Solar thermal engine;<br />

Thermal cracking process.<br />

Further Reading<br />

Tidal power plant / 773<br />

Bernshtein, L. B. Tidal Power Plants. Seoul, Korea: Korea Ocean Research<br />

<strong>and</strong> Development Institute, 1996.<br />

Boyle, Godfrey. Renewable Energy: Power for a Sustainable Future. Oxford:<br />

Oxford University Press, 1998.<br />

Ross, David. Power from the Waves. New York: Oxford University<br />

Press, 1995.<br />

Seymour, Richard J. Ocean Energy Recovery: The State of the Art. New<br />

York: American Society of Civil Engineers, 1992.


774<br />

Touch-tone telephone<br />

Touch-tone telephone<br />

The invention: A push-button dialing system for telephones that<br />

replaced the earlier rotary-dial phone.<br />

The person behind the invention:<br />

Bell Labs, the research <strong>and</strong> development arm of the American<br />

Telephone <strong>and</strong> Telegraph Company<br />

Dialing Systems<br />

A person who wishes to make a telephone call must inform the<br />

telephone switching office which number he or she wishes to reach.<br />

A telephone call begins with the customer picking up the receiver<br />

<strong>and</strong> listening for a dial tone. The action of picking up the telephone<br />

causes a switch in the telephone to close, allowing electric current to<br />

flow between the telephone <strong>and</strong> the switching office. This signals<br />

the telephone office that the user is preparing to dial a number. To<br />

acknowledge its readiness to receive the digits of the desired number,<br />

the telephone office sends a dial tone to the user. Two methods<br />

have been used to send telephone numbers to the telephone office:<br />

dial pulsing <strong>and</strong> touch-tone dialing.<br />

“Dial pulsing” is the method used by telephones that have rotary<br />

dials. In this method, the dial is turned until it stops, after which it is<br />

released <strong>and</strong> allowed to return to its resting position. When the dial<br />

is returning to its resting position, the telephone breaks the current<br />

between the telephone <strong>and</strong> the switching office. The switching office<br />

counts the number of times that current flow is interrupted,<br />

which indicates the number that had been dialed.<br />

Introduction of Touch-tone Dialing<br />

The dial-pulsing technique was particularly appropriate for use<br />

in the first electromechanical telephone switching offices, because<br />

the dial pulses actually moved mechanical switches in the switching<br />

office to set up the telephone connection. The introduction of<br />

touch-tone dialing into electromechanical systems was made possi-


Touch-tone telephone / 775<br />

ble by a special device that converted the touch-tones into rotary<br />

dial pulses that controlled the switches. At the American Telephone<br />

<strong>and</strong> Telegraph Company’s Bell Labs, experimental studies were<br />

pursued that explored the use of “multifrequency key pulsing” (in<br />

other words, using keys that emitted tones of various frequencies)<br />

by both operators <strong>and</strong> customers. Initially, plucked tuned reeds<br />

were proposed. These were, however, replaced with “electronic<br />

transistor oscillators,” which produced the required signals electronically.<br />

The introduction of “crossbar switching” made dial pulse signaling<br />

of the desired number obsolete. The dial pulses of the telephone<br />

were no longer needed to control the mechanical switching process<br />

at the switching office. When electronic control was introduced into<br />

switching offices, telephone numbers could be assigned by computer<br />

rather than set up mechanically. This meant that a single<br />

touch-tone receiver at the switching office could be shared by a<br />

large number of telephone customers.<br />

Before 1963, telephone switching offices relied upon rotary dial<br />

pulses to move electromechanical switching elements. Touch-tone<br />

dialing was difficult to use in systems that were not computer controlled,<br />

such as the electromechanical step-by-step method. In about<br />

1963, however, it became economically feasible to implement centralized<br />

computer control <strong>and</strong> touch-tone dialing in switching offices.<br />

Computerized switching offices use a central touch-tone receiver<br />

to detect dialed numbers, after which the receiver sends the<br />

number to a call processor so that a voice connection can be established.<br />

Touch-tone dialing transmits two tones simultaneously to represent<br />

a digit. The tones that are transmitted are divided into two<br />

groups: a high-b<strong>and</strong> group <strong>and</strong> a low-b<strong>and</strong> group. For each digit<br />

that is dialed, one tone from the low-frequency (low-b<strong>and</strong>) group<br />

<strong>and</strong> one tone from the high-frequency (high-b<strong>and</strong>) group are transmitted.<br />

The two frequencies of a tone are selected so that they are<br />

not too closely related harmonically. In addition, touch-tone receivers<br />

must be designed so that false digits cannot be generated when<br />

people are speaking into the telephone.<br />

For a call to be completed, the first digit dialed must be detected<br />

in the presence of a dial tone, <strong>and</strong> the receiver must not interpret


776 / Touch-tone telephone<br />

background noise or speech as valid digits. In order to avoid such<br />

misinterpretation, the touch-tone receiver uses both the relative <strong>and</strong><br />

the absolute strength of the two simultaneous tones of the first digit<br />

dialed to determine what that digit is.<br />

A system similar to the touch-tone system is used to send telephone<br />

numbers between telephone switching offices. This system,<br />

which is called “multifrequency signaling,” also uses two tones to<br />

indicate a single digit, but the frequencies used are not the same frequencies<br />

that are used in the touch-tone system. Multifrequency<br />

signaling is currently being phased out; new computer-based systems<br />

are being introduced to replace it.<br />

Impact<br />

Touch-tone dialing has made new caller features available. The<br />

touch-tone system can be used not only to signal the desired number<br />

to the switching office but also to interact with voice-response<br />

systems. This means that touch-tone dialing can be used in conjunction<br />

with such devices as bank teller machines. A customer can also<br />

dial many more digits per second with a touch-tone telephone than<br />

with a rotary dial telephone.<br />

Touch-tone dialing has not been implemented in Europe, <strong>and</strong><br />

one reason may be that the economics of touch-tone dialing change<br />

as a function of technology. In the most modern electronic switching<br />

offices, rotary signaling can be performed at no additional cost,<br />

whereas the addition of touch-tone dialing requires a centralized<br />

touch-tone receiver at the switching office. Touch-tone signaling<br />

was developed in an era of analog telephone switching offices, <strong>and</strong><br />

since that time, switching offices have become overwhelmingly digital.<br />

When the switching network becomes entirely digital, as will<br />

be the case when the integrated services digital network (ISDN) is<br />

implemented, touch-tone dialing will become unnecessary. In the<br />

future, ISDN telephone lines will use digital signaling methods exclusively.<br />

See also Cell phone; Rotary dial telephone; Telephone switching.


Further Reading<br />

Touch-tone telephone / 777<br />

Coe, Lewis. The Telephone <strong>and</strong> Its Several <strong>Inventors</strong>: A History. Jefferson,<br />

N.C.: McFarl<strong>and</strong>, 1995.<br />

Young, Peter. Person to Person: The International Impact of the Telephone.<br />

Cambridge: Granta Editions, 1991.


778<br />

Transistor<br />

Transistor<br />

The invention: A miniature electronic device, comprising a tiny<br />

semiconductor <strong>and</strong> multiple electrical contacts, used in circuits<br />

as an amplifier, detector, or switch, that revolutionized electronics<br />

in the mid-twentieth century.<br />

The people behind the invention:<br />

William B. Shockley (1910-1989), an American physicist who led<br />

the Bell Laboratories team that produced the first transistors<br />

Akio Morita (1921-1999), a Japanese physicist <strong>and</strong> engineer who<br />

was the cofounder of the Sony electronics company<br />

Masaru Ibuka (1908-1997), a Japanese electrical engineer <strong>and</strong><br />

businessman who cofounded Sony with Morita<br />

The Birth of Sony<br />

In 1952, a Japanese engineer visiting the United States learned<br />

that the Western Electric company was granting licenses to use its<br />

transistor technology. He was aware of the development of this device<br />

<strong>and</strong> thought that it might have some commercial applications.<br />

Masaru Ibuka told his business partner in Japan about the opportunity,<br />

<strong>and</strong> they decided to raise the $25,000 required to obtain a license.<br />

The following year, his partner, Akio Morita, traveled to New<br />

York City <strong>and</strong> concluded negotiations with Western Electric. This<br />

was a turning point in the history of the Sony company <strong>and</strong> in the<br />

electronics industry, for transistor technology was to open profitable<br />

new fields in home entertainment.<br />

The origins of the Sony corporation were in the ruins of postwar<br />

Japan. The Tokyo Telecommunications Company was incorporated<br />

in 1946 <strong>and</strong> manufactured a wide range of electrical equipment<br />

based on the existing vacuum tube technology. Morita <strong>and</strong> Ibuka<br />

were involved in research <strong>and</strong> development of this technology during<br />

the war <strong>and</strong> intended to put it to use in the peacetime economy.<br />

In the United States <strong>and</strong> Europe, electrical engineers who had done<br />

the same sort of research founded companies to build advanced<br />

audio products such as high-performance amplifiers, but Morita


<strong>and</strong> Ibuka did not have the resources to make such sophisticated<br />

products <strong>and</strong> concentrated on simple items such as electric water<br />

heaters <strong>and</strong> small electric motors for record players.<br />

In addition to their experience as electrical engineers, both men<br />

were avid music lovers, as a result of their exposure to Americanbuilt<br />

phonographs <strong>and</strong> gramophones exported to Japan in the early<br />

twentieth century. They decided to combine their twin interests by<br />

devising innovative audio products <strong>and</strong> looked to the new field of<br />

magnetic recording as a likely area for exploitation. They had learned<br />

about tape recorders from technical journals <strong>and</strong> had seen them in<br />

use by the American occupation force.<br />

They developed a reel-to-reel tape recorder <strong>and</strong> introduced it in<br />

1950. It was a large machine with vacuum tube amplifiers, so heavy<br />

that they transported it by truck. Although it worked well, they had<br />

a hard job selling it. Ibuka went to the United States in 1952 partly<br />

on a fact-finding mission <strong>and</strong> partly to get some ideas about marketing<br />

the tape recorder to schools <strong>and</strong> businesses. It was not seen as a<br />

consumer product.<br />

Ibuka <strong>and</strong> Morita had read about the invention of the transistor<br />

in Western Electric’s laboratories shortly after the war. John Bardeen<br />

<strong>and</strong> Walter H. Brattain had discovered that a semiconducting material<br />

could be used to amplify or control electric current. Their point<br />

contact transistor of 1948 was a crude laboratory apparatus that<br />

served as the basis for further research. The project was taken over<br />

by William B. Shockley, who had suggested the theory of the transistor<br />

effect. A new generation of transistors was devised; they were<br />

simpler <strong>and</strong> more efficient than the original. The junction transistors<br />

were the first to go into production.<br />

Ongoing Research<br />

Transistor / 779<br />

Bell Laboratories had begun transistor research because Western<br />

Electric, one of its parent companies along with American Telephone<br />

<strong>and</strong> Telegraph, was interested in electronic amplification.<br />

This was seen as a means to increase the strength of telephone signals<br />

traveling over long distances, a job carried out by vacuum<br />

tubes. The junction transistor was developed as an amplifier. Western<br />

Electric thought that the hearing aid was the only consumer


780 / Transistor<br />

product that could be based on it <strong>and</strong> saw the transistor solely as a<br />

telecommunications technology. The Japanese purchased the license<br />

with only the slightest underst<strong>and</strong>ing of the workings of<br />

semiconductors <strong>and</strong> despite the belief that transistors could not be<br />

used at the high frequencies associated with radio.<br />

The first task of Ibuka <strong>and</strong> Morita was to develop a highfrequency<br />

transistor. Once this was accomplished, in 1954, a method<br />

had to be found to manufacture it cheaply. Transistors were made<br />

from crystals, which had to be grown <strong>and</strong> doped with impurities to<br />

form different layers of conductivity. This was not an exact science,<br />

<strong>and</strong> Sony engineers found that the failure rate for high-frequency<br />

transistors was very high. This increased costs <strong>and</strong> put the entire<br />

project into doubt, because the adoption of transistors was based on<br />

simplicity, reliability, <strong>and</strong> low cost.<br />

The introduction of the first Sony transistor radio, the TR-55, in<br />

1955 was the result of basic research combined with extensive industrial<br />

engineering. Morita admitted that its sound was poor, but<br />

because it was the only transistor radio in Japan, it sold well. These<br />

were not cheap products, nor were they particularly compact. The<br />

selling point was that they consumed much less battery power than<br />

the old portable radios.<br />

The TR-55 carried the br<strong>and</strong> name Sony, a relative of the Soni<br />

magnetic tape made by the company <strong>and</strong> a name influenced by the<br />

founders’ interest in sound. Morita <strong>and</strong> Ibuka had already decided<br />

that the future of their company would be in international trade <strong>and</strong><br />

wanted its name to be recognized all over the world. In 1957, they<br />

changed the company’s name from Tokyo Telecomunications Engineering<br />

to Sony.<br />

The first product intended for the export market was a small<br />

transistor radio. Ibuka was disappointed at the large size of the TR-<br />

55 because one of the advantages of the transistor over the vacuum<br />

tube was supposed to be smaller size. He saw a miniature radio as a<br />

promising consumer product <strong>and</strong> gave his engineers the task of designing<br />

one small enough to fit into his shirt pocket.<br />

All elements of the radio had to be reduced in size: amplifier,<br />

transformer, capacitor, <strong>and</strong> loudspeaker. Like many other Japanese<br />

manufacturers, Sony bought many of the component parts of its<br />

products from small manufacturers, all of which had to be cajoled


into decreasing the size of their parts. Morita <strong>and</strong> Ibuka stated that<br />

the hardest task in developing this new product was negotiating<br />

with the subcontractors. Finally, the Type 63 pocket transistor radio—the<br />

“Transistor Six”—was introduced in 1957.<br />

Impact<br />

Transistor / 781<br />

When the transistor radio was introduced, the market for radios<br />

was considered to be saturated. People had rushed to buy them<br />

when they were introduced in the 1920’s, <strong>and</strong> by the time of the<br />

Great Depression, the majority of American households had one.<br />

Improvements had been made to the receiver <strong>and</strong> more attractive<br />

radio/phonograph console sets had been introduced, but these developments<br />

did not add many new customers. The most manufacturers<br />

could hope for was the replacement market with a few sales<br />

as children moved out of their parents’ homes <strong>and</strong> established new<br />

households.<br />

The pocket radio created a new market. It could be taken anywhere<br />

<strong>and</strong> used at any time. Its portability was its major asset, <strong>and</strong> it<br />

became an indispensable part of youth-oriented popular culture of<br />

the 1950’s <strong>and</strong> 1960’s. It provided an outlet for the crowded airwaves<br />

of commercial AM radio <strong>and</strong> was the means to bring the new<br />

music of rock <strong>and</strong> roll to a mass audience.<br />

As soon as Sony introduced the Transistor Six, it began to redesign<br />

it to reduce manufacturing cost. Subsequent transistor radios<br />

were smaller <strong>and</strong> cheaper. Sony sold them by the millions, <strong>and</strong> millions<br />

more were made by other companies under br<strong>and</strong> names such<br />

as “Somy” <strong>and</strong> “Sonny.” By 1960, more than twelve million transistor<br />

radios had been sold.<br />

The transistor radio was the product that established Sony as an<br />

international audio concern. Morita had resisted the temptation to<br />

make radios for other companies to sell under their names. Exports<br />

of Sony radios increased name recognition <strong>and</strong> established a bridgehead<br />

in the United States, the biggest market for electronic consumer<br />

products. Morita planned to follow the radio with other transistorized<br />

products.<br />

The television had challenged radio’s position as the mechanical<br />

entertainer in the home. Like the radio, it stood in nearly every


782 / Transistor<br />

William Shockley<br />

William Shockley’s reputation contains extremes. He helped<br />

invent one of the basic devices supporting modern technological<br />

society, the transistor. He also tried to revive one of the most<br />

infamous social theories, eugenics.<br />

His parents, mining engineer William Hillman Shockley,<br />

<strong>and</strong> surveyor May Bradford Shockley, were on assignment in<br />

Engl<strong>and</strong> in 1910 when he was born. The family returned to<br />

Northern California when the younger William was three, <strong>and</strong><br />

they schooled him at home until he was eight. He acquired an<br />

early interest in physics from a neighbor who taught at Stanford<br />

University. Shockley pursed that interest at the California Institute<br />

of Technology <strong>and</strong> the Massachusetts Institute of Technology,<br />

which awarded him a doctorate in 1936.<br />

Shockley went to work for Bell Telephone Laboratories in<br />

the same year. While trying to design a vacuum tube that could<br />

amplify current, it occurred to him that solid state components<br />

might work better than the fragile tubes. He experimented with<br />

the semiconductors germanium <strong>and</strong> silicon, but the materials<br />

available were too impure for his purpose. World War II interrupted<br />

the experiments, <strong>and</strong> he worked instead to improve radar<br />

<strong>and</strong> anti-submarine devices for the military. Back at Bell<br />

Labs in 1945, Shockley teamed with theorist John Bardeen <strong>and</strong><br />

experimentalist Walter Brattain. Two years later they succeeded<br />

in making the first amplifier out of semiconductor materials<br />

<strong>and</strong> called it a transistor (short for transfer resistor). Its effect on<br />

the electronics industry was revolutionary, <strong>and</strong> the three shared<br />

the 1956 Nobel Prize in Physics for their achievement.<br />

In the mid-1950’s Shockley left Bell Labs to start Shockley<br />

Transistor, then switched to academia in 1963, becoming Stanford<br />

University’s Alex<strong>and</strong>er M. Poniatoff Professor of Engineering<br />

<strong>and</strong> Applied Science. He grew interested in the relation<br />

between race <strong>and</strong> intellectual ability. Teaching himself psychology<br />

<strong>and</strong> genetics, he conceived the theory that Caucasians were<br />

inherently more intelligent than other races because of their genetic<br />

make-up. When he lectured on his br<strong>and</strong> of eugenics, he<br />

was denounced by the public as a racist <strong>and</strong> by scientists for<br />

shoddy thinking. Shockley retired in 1975 <strong>and</strong> died in 1989.


Transistor / 783<br />

American living room <strong>and</strong> used the same vacuum tube amplification<br />

unit. The transistorized portable television set did for images<br />

what the transistor radio did for sound. Sony was the first to develop<br />

an all-transistor television, in 1959. At a time when the trend<br />

in television receivers was toward larger screens, Sony produced<br />

extremely small models with eight-inch screens. Ignoring the marketing<br />

experts who said that Americans would never buy such a<br />

product, Sony introduced these models into the United States in<br />

1960 <strong>and</strong> found that there was a huge dem<strong>and</strong> for them.<br />

As in radio, the number of television stations on the air <strong>and</strong><br />

broadcasts for the viewer to choose from grew. Apersonal television<br />

or radio gave the audience more choices. Instead of one machine in<br />

the family room, there were now several around the house. The<br />

transistorization of mechanical entertainers allowed each family<br />

member to choose his or her own entertainment. Sony learned several<br />

important lessons from the success of the transistor radio <strong>and</strong><br />

television. The first was that small size <strong>and</strong> low price could create<br />

new markets for electronic consumer products. The second was that<br />

constant innovation <strong>and</strong> cost reduction were essential to keep ahead<br />

of the numerous companies that produced cheaper copies of original<br />

Sony products.<br />

In 1962, Sony introduced a tiny television receiver with a fiveinch<br />

screen. In the 1970’s <strong>and</strong> 1980’s, it produced even smaller models,<br />

until it had a TV set that could sit in the palm of the h<strong>and</strong>—the<br />

Video Walkman. Sony’s scientists had developed an entirely new<br />

television screen that worked on a new principle <strong>and</strong> gave better<br />

color resolution; the company was again able to blend the fruits of<br />

basic scientific research with innovative industrial engineering.<br />

The transistorized amplifier unit used in radio <strong>and</strong> television sets<br />

was applied to other products, including amplifiers for record players<br />

<strong>and</strong> tape recorders. Japanese manufacturers were slow to take<br />

part in the boom in high-fidelity audio equipment that began in the<br />

United States in the 1950’s. The leading manufacturers of highquality<br />

audio components were small American companies based<br />

on the talents of one engineer, such as Avery Fisher or Henry Koss.<br />

They sold expensive amplifiers <strong>and</strong> loudspeakers to audiophiles.<br />

The transistor reduced the size, complexity, <strong>and</strong> price of these components.<br />

The Japanese took the lead devising complete audio units


784 / Transistor<br />

based on transistorized integrated circuits, thus developing the basic<br />

home stereo.<br />

In the 1960’s, companies such as Sony <strong>and</strong> Matsushita dominated<br />

the market for inexpensive home stereos. These were the basic<br />

radio/phonograph combination, with two detached speakers.<br />

The finely crafted wooden consoles that had been the st<strong>and</strong>ard for<br />

the home phonograph were replaced by small plastic boxes. The<br />

Japanese were also quick to exploit the opportunities of the tape cassette.<br />

The Philips compact cassette was enthusiastically adopted by<br />

Japanese manufacturers <strong>and</strong> incorporated into portable tape recorders.<br />

This was another product with its ancestry in the transistor<br />

radio. As more of them were sold, the price dropped, encouraging<br />

more consumers to buy. The cassette player became as commonplace<br />

in American society in the 1970’s as the transistor radio had<br />

been in the 1960’s.<br />

The Walkman<br />

The transistor took another step in miniaturization in the Sony<br />

Walkman, a personal stereo sound system consisting of a cassette<br />

player <strong>and</strong> headphones. It was based on the same principles as the<br />

transistor radio <strong>and</strong> television. Sony again confounded marketing<br />

experts by creating a new market for a personal electronic entertainer.<br />

In the ten years following the introduction of the Walkman in<br />

1979, Sony sold fifty million units worldwide, half of those in the<br />

United States. Millions of imitation products were sold by other<br />

companies.<br />

Sony’s acquisition of the Western Electric transistor technology<br />

was a turning point in the fortunes of that company <strong>and</strong> of Japanese<br />

manufacturers in general. Less than ten years after suffering defeat<br />

in a disastrous war, Japanese industry served notice that it had lost<br />

none of its engineering capabilities <strong>and</strong> innovative skills. The production<br />

of the transistor radio was a testament to the excellence of<br />

Japanese research <strong>and</strong> development. Subsequent products proved<br />

that the Japanese had an uncanny sense of the potential market for<br />

consumer products based on transistor technology. The ability to incorporate<br />

solid-state electronics into innovative home entertainment<br />

products allowed Japanese manufacturers to dominate the


world market for electronic consumer products <strong>and</strong> to eliminate<br />

most of their American competitors.<br />

The little transistor radio was the vanguard of an invasion of new<br />

products unparalleled in economic history. Japanese companies<br />

such as Sony <strong>and</strong> Panasonic later established themselves at the leading<br />

edge of digital technology, the basis of a new generation of entertainment<br />

products. Instead of Japanese engineers scraping together<br />

the money to buy a license for an American technology, the<br />

great American companies went to Japan to license compact disc<br />

<strong>and</strong> other digital technologies.<br />

See also Cassette recording; Color television; FM radio; Radio;<br />

Television; Transistor radio; Videocassette recorder; Walkman cassette<br />

player.<br />

Further Reading<br />

Transistor / 785<br />

Lyons, Nick. The Sony Vision. New York: Crown Publishers, 1976.<br />

Marshall, David V. Akio Morita <strong>and</strong> Sony. Watford: Exley, 1995.<br />

Morita, Akio, with Edwin M. Reingold, <strong>and</strong> Mitsuko Shimomura.<br />

Made in Japan: Akio Morita <strong>and</strong> Sony. London: HarperCollins,<br />

1994.<br />

Reid, T. R. The Chip: How Two Americans Invented the Microchip <strong>and</strong><br />

Launched a Revolution. New York: Simon <strong>and</strong> Schuster, 1984.<br />

Riordan, Michael. Crystal Fire: The Invention of the Transistor <strong>and</strong> the<br />

Birth of the Information Age. New York: Norton, 1998.<br />

Scott, Otto. The Creative Ordeal: The Story of Raytheon. New York:<br />

Atheneum, 1974.


786<br />

Transistor radio<br />

Transistor radio<br />

The invention: Miniature portable radio that used transistors <strong>and</strong><br />

created a new mass market for electronic products.<br />

The people behind the invention:<br />

John Bardeen (1908-1991), an American physicist<br />

Walter H. Brattain (1902-1987), an American physicist<br />

William Shockley (1910-1989), an American physicist<br />

Akio Morita (1921-1999), a Japanese physicist <strong>and</strong> engineer<br />

Masaru Ibuka (1907-1997), a Japanese electrical engineer <strong>and</strong><br />

industrialist<br />

A Replacement for Vacuum Tubes<br />

The invention of the first transistor by William Shockley, John<br />

Bardeen, <strong>and</strong> Walter H. Brattain of Bell Labs in 1947 was a scientific<br />

event of great importance. Its commercial importance at the time,<br />

however, was negligible. The commercial potential of the transistor<br />

lay in the possibility of using semiconductor materials to carry out<br />

the functions performed by vacuum tubes, the fragile <strong>and</strong> expensive<br />

tubes that were the electronic hearts of radios, sound amplifiers,<br />

<strong>and</strong> telephone systems. Transistors were smaller, more rugged,<br />

<strong>and</strong> less power-hungry than vacuum tubes. They did not suffer<br />

from overheating. They offered an alternative to the unreliability<br />

<strong>and</strong> short life of vacuum tubes.<br />

Bell Labs had begun the semiconductor research project in an effort<br />

to find a better means of electronic amplification. This was<br />

needed to increase the strength of telephone signals over long distances.<br />

Therefore, the first commercial use of the transistor was<br />

sought in speech amplification, <strong>and</strong> the small size of the device<br />

made it a perfect component for hearing aids. Engineers from the<br />

Raytheon Company, the leading manufacturer of hearing aids, were<br />

invited to Bell Labs to view the new transistor <strong>and</strong> to help assess the<br />

commercial potential of the technology. The first transistorized consumer<br />

product, the hearing aid, was soon on the market. The early<br />

models built by Raytheon used three junction-type transistors <strong>and</strong><br />

cost more than two hundred dollars. They were small enough to go


directly into the ear or to be incorporated into eyeglasses.<br />

The commercial application of semiconductors was aimed largely<br />

at replacing the control <strong>and</strong> amplification functions carried out by<br />

vacuum tubes. The perfect vehicle for this substitution was the radio<br />

set. Vacuum tubes were the most expensive part of a radio set<br />

<strong>and</strong> the most prone to break down. The early junction transistors<br />

operated best at low frequencies, <strong>and</strong> subsequently more research<br />

was needed to produce a commercial high-frequency transistor.<br />

Several of the licensees embarked on this quest, including the Radio<br />

Corporation of America (RCA), Texas Instruments, <strong>and</strong> the Tokyo<br />

Telecommunications Engineering Company of Japan.<br />

Perfecting the Transistor<br />

Transistor radio / 787<br />

The Tokyo Telecommunications Engineering Company of Japan,<br />

formed in 1946, had produced a line of instruments <strong>and</strong> consumer<br />

products based on vacuum-tube technology. Its most successful<br />

product was a magnetic tape recorder. In 1952, one of the founders<br />

of the company, Masaru Ibuka, visited the United States to learn<br />

more about the use of tape recorders in schools <strong>and</strong> found out that<br />

Western Electric was preparing to license the transistor patent. With<br />

only the slightest underst<strong>and</strong>ing of the workings of semiconductors,<br />

Tokyo Telecommunications purchased a license in 1954 with<br />

the intention of using transistors in a radio set.<br />

The first task facing the Japanese was to increase the frequency<br />

response of the transistor to make it suitable for radio use. Then a<br />

method of manufacturing transistors cheaply had to be found. At<br />

the time, junction transistors were made from slices of germanium<br />

crystal. Growing the crystal was not an exact science, nor was the<br />

process of “doping” it with impurities to form the different layers of<br />

conductivity that made semiconductors useful. The Japanese engineers<br />

found that the failure rate for high-frequency transistors was<br />

extremely high. The yield of good transistors from one batch ran as<br />

low as 5 percent, which made them extremely expensive <strong>and</strong> put the<br />

whole project in doubt. The effort to replace vacuum tubes with<br />

components made of semiconductors was motivated by cost rather<br />

than performance; if transistors proved to be more expensive, then<br />

it was not worth using them.


788 / Transistor radio<br />

Engineers from Tokyo Telecommunications again came to the<br />

United States to search for information about the production of<br />

transistors. In 1954, the first high-frequency transistor was produced<br />

in Japan. The success of Texas Instruments in producing the<br />

components for the first transistorized radio (introduced by the Regency<br />

Company in 1954) spurred the Japanese to greater efforts.<br />

Much of their engineering <strong>and</strong> research work was directed at the<br />

manufacture <strong>and</strong> quality control of transistors. In 1955, they introduced<br />

their transistor radio, the TR-55, which carried the br<strong>and</strong><br />

name “Sony.” The name was chosen because the executives of the<br />

company believed that the product would have an international appeal<br />

<strong>and</strong> therefore needed a br<strong>and</strong> name that could be recognized<br />

easily <strong>and</strong> remembered in many languages. In 1957, the name of the<br />

entire company was changed to Sony.<br />

Impact<br />

Although Sony’s transistor radios were successful in the marketplace,<br />

they were still relatively large <strong>and</strong> cumbersome. Ibuka saw a<br />

consumer market for a miniature radio <strong>and</strong> gave his engineers the<br />

task of designing a radio small enough to fit into a shirt pocket. The<br />

realization of this design—“Transistor Six”—was introduced in 1957.<br />

It was an immediate success. Sony sold the radios by the millions,<br />

<strong>and</strong> numerous imitations were also marketed under br<strong>and</strong> names<br />

such as “Somy” <strong>and</strong> “Sonny.” The product became an indispensable<br />

part of popular culture of the late 1950’s <strong>and</strong> 1960’s; its low cost enabled<br />

the masses to enjoy radio wherever there were broadcasts.<br />

The pocket-sized radio was the first of a line of electronic consumer<br />

products that brought technology into personal contact with<br />

the user. Sony was convinced that miniaturization did more than<br />

make products more portable; it established a one-on-one relationship<br />

between people <strong>and</strong> machines. Sony produced the first alltransistor<br />

television in 1960. Two years later, it began to market a<br />

miniature television in the United States. The continual reduction in<br />

the size of Sony’s tape recorders reached a climax with the portable<br />

tape player introduced in the 1980’s. The Sony Walkman was a marketing<br />

triumph <strong>and</strong> a further reminder that Japanese companies led<br />

the way in the design <strong>and</strong> marketing of electronic products.


John Bardeen<br />

Transistor radio / 789<br />

The transistor reduced the size of electronic circuits <strong>and</strong> at<br />

the same time the amount of energy lost from them as heat.<br />

Superconduction gave rise to electronic circuits with practically<br />

no loss of energy at all. John Bardeen helped unlock the secrets<br />

of both.<br />

Bardeen was born in 1908 in Madison, Wisconsin, where his<br />

mother was an artist <strong>and</strong> his father was a professor of anatomy<br />

at the University of Wisconsin. Bardeen attended the university,<br />

earning a bachelor’s degree in electrical engineering in 1928<br />

<strong>and</strong> a master’s degree in geophysics in 1929. After working as a<br />

geophysicist, he entered Princeton University, studying with<br />

Eugene Wigner, the leading authority on solid-state physics,<br />

<strong>and</strong> received a doctorate in mathematics <strong>and</strong> physics in 1936.<br />

Bardeen taught at Harvard University <strong>and</strong> the University of<br />

Minnesota until World War II, when he moved to the Naval<br />

Ordnance Laboratory. Finding academic salaries too low to<br />

support his family after the war, he accepted a position at Bell<br />

Telephone Laboratories. There, with Walter Brattain, he turned<br />

William Shockley’s theory of semiconductors into a practical<br />

device—the transfer resistor, or transistor.<br />

He returned to academia as a professor at the University of<br />

Illinois <strong>and</strong> began to investigate a long-st<strong>and</strong>ing mystery in<br />

physics, superconductivity, with a postdoctoral associate, Leon<br />

Cooper, <strong>and</strong> a graduate student, J. Robert Schrieffer. In 1956<br />

Cooper made a key discovery—superconducting electrons<br />

travel in pairs. And while Bardeen was in Stockholm, Sweden,<br />

collecting a share of the 1956 Nobel Prize in Physics for his work<br />

on transistors, Schrieffer worked out a mathematical analysis of<br />

the phenomenon. The theory that the three men published since<br />

became known as BCS theory from the first letters of their last<br />

names, <strong>and</strong> as well as explain superconductors, it pointed toward<br />

a great deal of technology <strong>and</strong> additional basic research.<br />

The team won the 1972 Nobel Prize in Physics for BCS theory,<br />

making Bardeen the only person to ever win two Nobel Prizes<br />

for physics. He retired in 1975 <strong>and</strong> died sixteen years later.<br />

See also Compact disc; FM radio; Radio; Radio crystal sets; Television;<br />

Transistor; Walkman cassette player.


790 / Transistor radio<br />

Further Reading<br />

H<strong>and</strong>y, Roger, Maureen Erbe, <strong>and</strong> Aileen Antonier. Made in Japan:<br />

Transistor Radios of the 1950s <strong>and</strong> 1960s. San Francisco: Chronicle<br />

Books, 1993.<br />

Marshall, David V. Akio Morita <strong>and</strong> Sony. Watford: Exley, 1995.<br />

Morita, Akio, with Edwin M. Reingold, <strong>and</strong> Mitsuko Shimomura.<br />

Made in Japan: Akio Morita <strong>and</strong> Sony. London: HarperCollins, 1994.<br />

Nathan, John. Sony: The Private Life. London: HarperCollins-<br />

Business, 2001.


Tuberculosis vaccine<br />

Tuberculosis vaccine<br />

The invention: Vaccine that uses an avirulent (nondisease) strain<br />

of bovine tuberculosis bacilli that is safer than earlier vaccines.<br />

The people behind the invention:<br />

Albert Calmette (1863-1933), a French microbiologist<br />

Camille Guérin (1872-1961), a French veterinarian <strong>and</strong><br />

microbiologist<br />

Robert Koch (1843-1910), a German physician <strong>and</strong><br />

microbiologist<br />

Isolating Bacteria<br />

791<br />

Tuberculosis, once called “consumption,” is a deadly, contagious<br />

disease caused by the bacterium Mycobacterium tuberculosis,<br />

first identified by the eminent German physician Robert Koch in<br />

1882. The bacterium can be transmitted from person to person by<br />

physical contact or droplet infection (for example, sneezing). The<br />

condition eventually inflames <strong>and</strong> damages the lungs, causing difficulty<br />

in breathing <strong>and</strong> failure of the body to deliver sufficient oxygen<br />

to various tissues. It can spread to other body tissues, where<br />

further complications develop. Without treatment, the disease progresses,<br />

disabling <strong>and</strong> eventually killing the victim. Tuberculosis<br />

normally is treated with a combination of antibiotics <strong>and</strong> other<br />

drugs.<br />

Koch developed his approach for identifying bacterial pathogens<br />

(disease producers) with simple equipment, primarily microscopy.<br />

Having taken blood samples from diseased animals, he would<br />

identify <strong>and</strong> isolate the bacteria he found in the blood. Each strain of<br />

bacteria would be injected into a healthy animal. The latter would<br />

then develop the disease caused by the particular strain.<br />

In 1890, he discovered that a chemical released from tubercular<br />

bacteria elicits a hypersensitive (allergic) reaction in individuals<br />

previously exposed to or suffering from tuberculosis. This chemical,<br />

called “tuberculin,” was isolated from culture extracts in which tubercular<br />

bacteria were being grown.


792 / Tuberculosis vaccine<br />

When small amounts of tuberculin are injected into a person subcutaneously<br />

(beneath the skin), a reddened, inflamed patch approximately<br />

the size of a quarter develops if the person has been exposed<br />

to or is suffering from tuberculosis. Injection of tuberculin into an<br />

uninfected person yields a negative response (that is, no inflammation).<br />

Tuberculin does not harm those being tested.<br />

Tuberculosis’s Weaker Gr<strong>and</strong>children<br />

The first vaccine to prevent tuberculosis was developed in 1921<br />

by two French microbiologists, Albert Calmette <strong>and</strong> Camille Guérin.<br />

Calmette was a student of the eminent French microbiologist Louis<br />

Pasteur at Pasteur’s Institute in Paris. Guérin was a veterinarian<br />

who joined Calmette’s laboratory in 1897. At Lille, Calmette <strong>and</strong><br />

Guérin focused their research upon the microbiology of infectious<br />

diseases, especially tuberculosis.<br />

In 1906, they discovered that individuals who had been exposed to<br />

tuberculosis or who had mild infections were developing resistance to<br />

the disease. They found that resistance to tuberculosis was initiated by<br />

the body’s immune system. They also discovered that tubercular bacteria<br />

grown in culture over many generations become progressively<br />

weaker <strong>and</strong> avirulent, losing their ability to cause disease.<br />

From 1906 through 1921, Calmette <strong>and</strong> Guérin cultured tubercle<br />

bacilli from cattle. With proper nutrients <strong>and</strong> temperature, bacteria<br />

can reproduce by fission (that is, one bacterium splits into two bacteria)<br />

in as little time as thirty minutes. Calmette <strong>and</strong> Guérin cultivated<br />

these bacteria in a bile-derived food medium for thous<strong>and</strong>s of<br />

generations over fifteen years, periodically testing the bacteria for<br />

virulence by injecting them into cattle. After many generations, the<br />

bacteria lost their virulence, their ability to cause disease. Nevertheless,<br />

these weaker, or “avirulent” bacteria still stimulated the animals’<br />

immune systems to produce antibodies. Calmette <strong>and</strong> Guérin<br />

had successfully bred a strain of avirulent bacteria that could not<br />

cause tuberculosis in cows but could also stimulate immunity against<br />

the disease.<br />

There was considerable concern over whether the avirulent strain<br />

was harmless to humans. Calmette <strong>and</strong> Guérin continued cultivating<br />

weaker versions of the avirulent strain that retained antibody-


stimulating capacity. By 1921, they had isolated an avirulent antibody-stimulating<br />

strain that was harmless to humans, a strain they<br />

called “Bacillus Calmette-Guérin” (BCG).<br />

In 1922, they began BCG-vaccinating newborn children against<br />

tuberculosis at the Charité Hospital in Paris. The immunized children<br />

exhibited no ill effects from the BCG vaccination. Calmette <strong>and</strong><br />

Guérin’s vaccine was so successful in controlling the spread of tuberculosis<br />

in France that it attained widespread use in Europe <strong>and</strong><br />

Asia beginning in the 1930’s.<br />

Impact<br />

Tuberculosis vaccine / 793<br />

Most bacterial vaccines involve the use of antitoxin or heat- or<br />

chemical-treated bacteria. BCG is one of the few vaccines that use<br />

specially bred live bacteria. Its use sparked some controversy in<br />

the United States <strong>and</strong> Engl<strong>and</strong>, where the medical community<br />

questioned its effectiveness <strong>and</strong> postponed BCG immunization<br />

until the late 1950’s. Extensive testing of the vaccine was performed<br />

at the University of Illinois before it was adopted in the<br />

United States. Its effectiveness is questioned by some physicians to<br />

this day.<br />

Some of the controversy stems from the fact that the avirulent,<br />

antibody-stimulating BCG vaccine conflicts with the tuberculin<br />

skin test. The tuberculin skin test is designed to identify people<br />

suffering from tuberculosis so that they can be treated. A BCGvaccinated<br />

person will have a positive tuberculin skin test similar<br />

to that of a tuberculosis sufferer. If a physician does not know that<br />

a patient has had a BCG vaccination, it will be presumed (incorrectly)<br />

that the patient has tuberculosis. Nevertheless, the BCG<br />

vaccine has been invaluable in curbing the worldwide spread of<br />

tuberculosis, although it has not eradicated the disease.<br />

See also Antibacterial drugs; Birth control pill; Penicillin; Polio<br />

vaccine (Sabin); Polio vaccine (Salk); Salvarsan; Typhus vaccine;<br />

Yellow fever vaccine.


794 / Tuberculosis vaccine<br />

Further Reading<br />

Daniel, Thomas M. Pioneers of Medicine <strong>and</strong> their Impact on Tuberculosis.<br />

Rochester, N.Y.: University of Rochester Press, 2000.<br />

DeJauregui, Ruth. 100 Medical Milestones That Shaped World History.<br />

San Mateo, Calif.: Bluewood Books, 1998.<br />

Fry, William F. “Prince Hamlet <strong>and</strong> Professor Koch.” Perspectives in<br />

Biology <strong>and</strong> Medicine 40, no. 3 (Spring, 1997).<br />

Lutwick, Larry I. New Vaccines <strong>and</strong> New Vaccine Technology. Philadelphia:<br />

Saunders, 1999.


Tungsten filament<br />

Tungsten filament<br />

The invention: Metal filament used in the inc<strong>and</strong>escent light bulbs<br />

that have long provided most of the world’s electrical lighting.<br />

The people behind the invention:<br />

William David Coolidge (1873-1975), an American electrical<br />

engineer<br />

Thomas Alva Edison (1847-1931), an American inventor<br />

The Inc<strong>and</strong>escent Light Bulb<br />

The electric lamp developed along with an underst<strong>and</strong>ing of<br />

electricity in the latter half of the nineteenth century. In 1841, the<br />

first patent for an inc<strong>and</strong>escent lamp was granted in Great Britain. A<br />

patent is a legal claim that protects the patent holder for a period of<br />

time from others who might try to copy the invention <strong>and</strong> make a<br />

profit from it. Although others tried to improve upon the inc<strong>and</strong>escent<br />

lamp, it was not until 1877, when Thomas Alva Edison, the famous<br />

inventor, became interested in developing a successful electric<br />

lamp, that real progress was made. The Edison Electric Light<br />

Company was founded in 1878, <strong>and</strong> in 1892, it merged with other<br />

companies to form the General Electric Company.<br />

Early electric lamps used platinum wire as a filament. Because<br />

platinum is expensive, alternative filament materials were sought.<br />

After testing many substances, Edison finally decided to use carbon<br />

as a filament material. Although carbon is fragile, making it difficult<br />

to manufacture filaments, it was the best choice available at the time.<br />

The Manufacture of Ductile Tungsten<br />

795<br />

Edison <strong>and</strong> others had tested tungsten as a possible material for<br />

lamp filaments but discarded it as unsuitable. Tungsten is a hard,<br />

brittle metal that is difficult to shape <strong>and</strong> easy to break, but it possesses<br />

properties that are needed for lamp filaments. It has the highest<br />

melting point (3,410 degrees Celsius) of any known metal; therefore,<br />

it can be heated to a very high temperature, giving off a


796 / Tungsten filament<br />

relatively large amount of radiation without melting (as platinum<br />

does) or decomposing (as carbon does). The radiation it emits when<br />

heated is primarily visible light. Its resistance to the passage of electricity<br />

is relatively high, so it requires little electric current to reach<br />

its operating voltage. It also has a high boiling point (about 5,900 degrees<br />

Celsius) <strong>and</strong> therefore does not tend to boil away, or vaporize,<br />

when heated. In addition, it is mechanically strong, resisting breaking<br />

caused by mechanical shock.<br />

William David Coolidge, an electrical engineer with the General<br />

Electric Company, was assigned in 1906 the task of transforming<br />

tungsten from its natural state into a form suitable for lamp filaments.<br />

The accepted procedure for producing fine metal wires was<br />

(<strong>and</strong> still is) to force a wire rod through successively smaller holes in<br />

a hard metal block until a wire of the proper diameter is achieved.<br />

The property that allows a metal to be drawn into a fine wire by<br />

means of this procedure is called “ductility.” Tungsten is not naturally<br />

ductile, <strong>and</strong> it was Coolidge’s assignment to make it into a ductile<br />

form. Over a period of five years, <strong>and</strong> after many failures, Coolidge<br />

<strong>and</strong> his workers achieved their goal. By 1911, General Electric<br />

was selling lamps that contained tungsten filaments.<br />

Originally, Coolidge attempted to mix powdered tungsten with a<br />

suitable substance, form a paste, <strong>and</strong> squirt that paste through a die<br />

to form the wire. The paste-wire was then sintered (heated at a temperature<br />

slightly below its melting point) in an effort to fuse the<br />

powder into a solid mass. Because of its higher boiling point, the<br />

tungsten would remain after all the other components in the paste<br />

boiled away. At about 300 degrees Celsius, tungsten softens sufficiently<br />

to be hammered into an elongated form. Upon cooling, however,<br />

tungsten again becomes brittle, which prevents it from being<br />

shaped further into filaments. It was suggested that impurities in<br />

the tungsten caused the brittleness, but specially purified tungsten<br />

worked no better than the unpurified form.<br />

Many metals can be reduced from rods to wires if the rods are<br />

passed through a series of rollers that are successively closer together.<br />

Some success was achieved with this method when the rollers<br />

were heated along with the metal, but it was still not possible to<br />

produce sufficiently fine wire. Next, Coolidge tried a procedure<br />

called “swaging,” in which a thick wire is repeatedly <strong>and</strong> rapidly


struck by a series of rotating hammers as the wire is drawn past<br />

them. After numerous failures, a fine wire was successfully produced<br />

using this procedure. It was still too thick for lamp filaments,<br />

but it was ductile at room temperature.<br />

Microscopic examination of the wire revealed a change in the<br />

crystalline structure of tungsten as a result of the various treatments.<br />

The individual crystals had elongated, taking on a fiberlike<br />

appearance. Now the wire could be drawn through a die to achieve<br />

the appropriate thickness. Again, the wire had to be heated, <strong>and</strong> if<br />

the temperature was too high, the tungsten reverted to a brittle<br />

state. The dies themselves were heated, <strong>and</strong> the reduction progressed<br />

in stages, each of which reduced the wire’s diameter by a<br />

thous<strong>and</strong>th of an inch.<br />

Finally, Coolidge had been successful. Pressed tungsten bars<br />

measuring 1 4 × 3 8 × 6 inches were hammered <strong>and</strong> rolled into rods 1 8<br />

inch, or 125 1000 inch, in diameter. The unit 1 1000 inch is often called a<br />

“mil.” These rods were then swaged to approximately 30 mil <strong>and</strong><br />

then passed through dies to achieve the filament size of 25 mil or<br />

smaller, depending on the power output of the lamp in which the<br />

filament was to be used. Tungsten wires of 1 mil or smaller are now<br />

readily available.<br />

Impact<br />

Tungsten filament / 797<br />

Ductile tungsten wire filaments are superior in several respects<br />

to platinum, carbon, or sintered tungsten filaments. Ductile filament<br />

lamps can withst<strong>and</strong> more mechanical shock without breaking.<br />

This means that they can be used in, for example, automobile<br />

headlights, in which jarring frequently occurs. Ductile wire can also<br />

be coiled into compact cylinders within the lamp bulb, which makes<br />

for a more concentrated source of light <strong>and</strong> easier focusing. Ductile<br />

tungsten filament lamps require less electricity than do carbon filament<br />

lamps, <strong>and</strong> they also last longer. Because the size of the filament<br />

wire can be carefully controlled, the light output from lamps<br />

of the same power rating is more reproducible. One 60-watt bulb is<br />

therefore exactly like another in terms of light production.<br />

Improved production techniques have greatly reduced the cost<br />

of manufacturing ductile tungsten filaments <strong>and</strong> of light-bulb man-


798 / Tungsten filament<br />

ufacturing in general. The modern world is heavily dependent<br />

upon this reliable, inexpensive light source, which turns darkness<br />

into daylight.<br />

See also Fluorescent lighting; Memory metal; Steelmaking<br />

process.<br />

Further Reading<br />

Baldwin, Neil. Edison: Inventing the Century. Chicago: University of<br />

Chicago Press, 2001.<br />

Cramer, Carol. Thomas Edison. San Diego, Calif.: Greenhaven Press,<br />

2001.<br />

Israel, Paul. Edison: A Life of Invention. New York: John Wiley, 1998.<br />

Liebhafsky, H. A. William David Coolidge: A Centenarian <strong>and</strong> His Work.<br />

New York: Wiley, 1974.<br />

Miller, John A. Yankee Scientist: William David Coolidge. Schenectady,<br />

N.Y.: Mohawk Development Service, 1963.


Tupperware<br />

Tupperware<br />

The invention: Trademarked food-storage products that changed<br />

the way Americans viewed plastic products <strong>and</strong> created a model<br />

for selling products in consumers’ homes.<br />

The people behind the invention:<br />

Earl S. Tupper (1907-1983), founder of Tupperware<br />

Brownie Wise, the creator of the vast home sales network for<br />

Tupperware<br />

Morison Cousins (1934-2001), a designer hired by Tupperware<br />

to modernize its products in the early 1990’s<br />

“The Wave of the Future”?<br />

799<br />

Relying on a belief that plastic was the wave of the future <strong>and</strong><br />

wanting to improve on the newest refrigeration technology, Earl S.<br />

Tupper, who called himself “a ham inventor <strong>and</strong> Yankee trader,”<br />

created an empire of products that changed America’s kitchens.<br />

Tupper, a self-taught chemical engineer, began working at Du Pont<br />

in the 1930’s. This was a time of important developments in the<br />

field of polymers <strong>and</strong> the technology behind plastics. Wanting to<br />

experiment with this new material yet unable to purchase the<br />

needed supplies, Tupper went to his employer for help. Because of<br />

the limited availability of materials, major chemical companies<br />

had been receiving all the raw goods for plastic production. Although<br />

Du Pont would not part with raw materials, the company<br />

was willing to let Tupper have the slag.<br />

Polyethylene slag was a black, rock-hard, malodorous waste<br />

product of oil refining. It was virtually unusable. Undaunted,<br />

Tupper developed methods to purify the slag. He then designed<br />

an injection molding machine to form bowls <strong>and</strong> other containers<br />

out of his “Poly-T.” Tupper did not want to call the substance plastic<br />

because of a public distrust of that substance. In 1938, he<br />

founded the Tupper Plastics Company to pursue his dream. It was<br />

during those first years that he formulated the design for the famous<br />

Tupperware seal.


800 / Tupperware<br />

Refrigeration techniques had improved tremendously during<br />

the first part of the twentieth century. The iceboxes in use prior to<br />

the 1940’s were inconsistent in their interior conditions <strong>and</strong> were<br />

usually damp inside because of melting of the ice. In addition, the<br />

metal, glass, or earthenware food storage containers used during<br />

the first half of the century did not seal tightly <strong>and</strong> allowed food to<br />

stay moist. Iceboxes allowed mixing of food odors, particularly evident<br />

with strong-smelling items such as onions <strong>and</strong> fish.<br />

Electric Refrigerators<br />

In contrast to iceboxes, the electric refrigerators available starting<br />

in the 1940’s maintained dry interiors <strong>and</strong> low temperatures. This<br />

change in environment resulted in food drying out <strong>and</strong> wilting.<br />

Tupper set out to alleviate this problem through his plastic containers.<br />

The key to Tupper’s solution was his containers’ seal. He took<br />

his design from paint can lids <strong>and</strong> inverted it. This tight seal created<br />

a partial vacuum that protected food from the dry refrigeration process<br />

<strong>and</strong> kept food odors sealed within containers.<br />

In 1942, Tupper bought his first manufacturing plant, in Farnumsville,<br />

Massachusetts. There he continued to improve on his designs.<br />

In 1945, Tupper introduced Tupperware, selling it through<br />

hardware <strong>and</strong> department stores as well as through catalog sales.<br />

Tupperware products were made of flexible, translucent plastic.<br />

Available in frosted crystal <strong>and</strong> five pastel colors, the new containers<br />

were airtight <strong>and</strong> waterproof. In addition, they carried a lifetime<br />

warranty against chipping, cracking, peeling, <strong>and</strong> breaking in normal<br />

noncommercial use. Early supporters of Tupperware included<br />

the American Thermos Bottle Company, which purchased seven<br />

million nesting cups, <strong>and</strong> the Tek Corporation, which ordered fifty<br />

thous<strong>and</strong> tumblers to sell with toothbrushes.<br />

Even though he benefited from this type of corporate support,<br />

Tupper wanted his products to be for home use. Marketing the new<br />

products proved to be difficult in the early years. Tupperware sat on<br />

hardware <strong>and</strong> department store shelves, <strong>and</strong> catalog sales were<br />

nearly nonexistent. The problem appeared to involve a basic distrust<br />

of plastic by consumers <strong>and</strong> an unfamiliarity with how to use<br />

the new products. The product did not come with instructions on


how to seal the containers or descriptions of how the closed container<br />

protected the food within. Brownie Wise, an early direct seller<br />

<strong>and</strong> veteran distributor of Stanley Home Products, stated that it<br />

took her several days to underst<strong>and</strong> the technology behind the seal<br />

<strong>and</strong> the now-famous Tupperware “burp,” the sound made when air<br />

leaves the container as it seals.<br />

Wise <strong>and</strong> two other direct sellers, Tom Damigella <strong>and</strong> Harvey<br />

Hollenbush, found the niche for selling Tupperware for daily use—<br />

home sales. Wise approached Tupper with a home party sales strategy<br />

<strong>and</strong> detailed how it provided a relaxed atmosphere in which to<br />

learn about the products <strong>and</strong> thus lowered sales resistance. In April,<br />

1951, Tupper took his product off store shelves <strong>and</strong> hired Wise to<br />

create a new direct selling system under the name of Tupperware<br />

Home Parties, Inc.<br />

Impact<br />

Tupperware / 801<br />

Home sales had already proved to be successful for the Fuller<br />

Brush Company <strong>and</strong> numerous encyclopedia publishers, yet Brownie<br />

Wise wanted to exp<strong>and</strong> the possibilities. Her first step was to found<br />

a campus-like headquarters in Kissimmee, Florida. There, Tupper <strong>and</strong><br />

a design department worked to develop new products, <strong>and</strong> Tupperware<br />

Home Parties, Inc., under Wise’s direction, worked to develop<br />

new incentives for Tupperware’s direct sellers, called hostesses.<br />

Wise added spark to the notion of home demonstrations. “Parties,”<br />

as they were called, included games, recipes, giveaways, <strong>and</strong> other<br />

ideas designed to help housewives learn how to use Tupperware<br />

products. The marketing philosophy was to make parties appealing<br />

events at which women could get together while their children were<br />

in school. This fit into the suburban lifestyle of the 1950’s. These parties<br />

offered a nonthreatening means for home sales representatives<br />

to attract audiences for their demonstrations <strong>and</strong> gave guests a chance<br />

to meet <strong>and</strong> socialize with their neighbors. Often compared to<br />

the barbecue parties of the 1950’s, Tupperware parties were social,<br />

yet educational, affairs. While guests ate lunch or snacked on desserts,<br />

the Tupperware hostess educated them about the technology<br />

behind the bowls <strong>and</strong> their seals as well as suggesting a wide variety<br />

of uses for the products. For example, a party might include


802 / Tupperware<br />

recipes for dinner parties, with information provided on how<br />

party leftovers could be stored efficiently <strong>and</strong> economically with<br />

Tupperware products.<br />

While Tupperware products were changing the kitchens of America,<br />

they were also changing the women who sold them (almost all the<br />

hosts were women). Tupperware sales offered employment for women<br />

at a time when society disapproved of women working outside the<br />

home. Being a hostess, however, was not a nine-to-five position. The<br />

job allowed women freedom to tailor their schedules to meet family<br />

needs. Employment offered more than the economic incentive of 35<br />

percent of gross sales. Hostesses also learned new skills <strong>and</strong> developed<br />

self-esteem. An acclaimed mentoring program for new <strong>and</strong> advancing<br />

employees provided motivational training. Managers came only from<br />

the ranks of hostesses; moving up the corporate ladder meant spending<br />

time selling Tupperware at home parties.<br />

The opportunity to advance offered incentive. In addition, annual<br />

sales conventions were renowned for teaching new marketing<br />

strategies in fun-filled classes. These conventions also gave women<br />

an opportunity to network <strong>and</strong> establish contacts. These experiences<br />

proved to be invaluable as women entered the workforce in<br />

increasing numbers in later decades.<br />

Exp<strong>and</strong>ing Home-Sales Business<br />

The tremendous success of Tupperware’s marketing philosophy<br />

helped to set the stage for other companies to enter home sales.<br />

These companies used home-based parties to educate potential customers<br />

in familiar surroundings, in their own homes or in the<br />

homes of friends. The Mary Kay Cosmetics Company, founded in<br />

1963, used beauty makeovers in the home party setting as its chief<br />

marketing tool. Discovery Toys, founded in 1978, encouraged guests<br />

to get on the floor <strong>and</strong> play with the toys demonstrated at its home<br />

parties. Both companies extended the socialization aspects found in<br />

Tupperware parties.<br />

In addition to setting the st<strong>and</strong>ard for home sales, Tupperware<br />

is also credited with starting the plastic revolution. Early plastics<br />

were of poor quality <strong>and</strong> cracked or broke easily. This created distrust<br />

of plastic products among consumers. Earl Tupper’s dem<strong>and</strong>


Tupperware / 803<br />

for quality set the stage for the future of plastics. He started with<br />

high-quality resin <strong>and</strong> developed a process that kept the “Poly-T”<br />

from splitting. He then invented an injection molding machine that<br />

mass-produced his bowl <strong>and</strong> cup designs. His st<strong>and</strong>ards of quality<br />

from start to finish helped other companies exp<strong>and</strong> into plastics.<br />

The 1950’s saw a wide variety of products appear in the improved<br />

material, including furniture <strong>and</strong> toys. This shift from wood, glass,<br />

<strong>and</strong> metal to plastic continued for decades.<br />

Maintaining the position of Tupperware within the housewares<br />

Earl S. Tupper<br />

Born in 1907, Earl Silas Tupper came from a family of go-getters.<br />

His mother, Lulu Clark Tupper, kept a boardinghouse <strong>and</strong><br />

took in laundry, while his father, Earnest, ran a small farm <strong>and</strong><br />

greenhouse in New Hampshire. The elder Tupper was also a<br />

small-time inventor, patenting a device for stretching out chickens<br />

to make cleaning them easier. Earl absorbed the family’s<br />

taste for invention <strong>and</strong> enterprise.<br />

Fresh out of high school in 1925, Tupper vowed to turn himself<br />

into a millionaire by the time he was thirty. He started a<br />

l<strong>and</strong>scaping <strong>and</strong> nursery business in 1928, but the Depression<br />

led his company, Tupper Tree, into bankruptcy in 1936. Tupper<br />

was undeterred. He hired on with Du Pont the next year. Du<br />

Pont taught him a great deal about the chemistry <strong>and</strong> manufacturing<br />

of plastics, but it did not give him scope to apply his<br />

ideas, so in 1938 he founded the Earl S. Tupper Company. He<br />

continued to work as a contractor for Du Pont to make the<br />

fledgling company profitable, <strong>and</strong> during World War II the<br />

company made plastic moldings for gas masks <strong>and</strong> Navy signal<br />

lamps. Finally, in the 1940’s Tupper could devote himself to<br />

his dream—designing plastic food containers, cups, <strong>and</strong> such<br />

small household conveniences as cases for cigarette packs.<br />

Thanks to aggressive, innovative direct marketing, Tupper’s<br />

kitchenware, Tupperware, became synonymous with plastic<br />

containers during the 1950’s. In 1958 Tupper sold his company<br />

to Rexall for $16 million, having finally realized his youthful<br />

ambition to make himself wealthy through Yankee wit <strong>and</strong><br />

hard work. He died in 1983.


804 / Tupperware<br />

market meant keeping current. As more Americans were able to purchase<br />

the newest refrigerators, Tupperware exp<strong>and</strong>ed to meet their<br />

needs. The company added new products, improved marketing<br />

strategies, <strong>and</strong> changed or updated designs. Over the years, Tupperware<br />

added baking items, toys, <strong>and</strong> home storage containers for such<br />

items as photographs, sewing materials, <strong>and</strong> holiday ornaments. The<br />

1980’s <strong>and</strong> 1990’s brought microwaveable products.<br />

As women moved into the work force in great numbers, Tupperware<br />

moved with them. The company introduced lunchtime parties<br />

at the workplace <strong>and</strong> parties at daycare centers for busy working<br />

parents. Tupperware also started a fund-raising line, in special colors,<br />

that provided organizations with a means to bring in money<br />

while not necessitating full-fledged parties. New party themes developed<br />

around time-saving techniques <strong>and</strong> health concerns such<br />

as diet planning. Beginning in 1992, customers too busy to attend a<br />

party could call a toll-free number, request a catalog, <strong>and</strong> be put in<br />

contact with a “consultant,” as “hostesses” now were called.<br />

Another marketing strategy developed out of a public push for<br />

environmentally conscious products. Tupperware consultants stressed<br />

the value of buying food in bulk to create less trash as well as saving<br />

money. To store these increased purchases, the company developed<br />

a new line for kitchen staples called Modular Mates. These stackable<br />

containers came in a wide variety of shapes <strong>and</strong> sizes to hold everything<br />

from cereal to flour to pasta. They were made of see-through<br />

plastic, allowing the user to see if the contents needed replenishing.<br />

Some consultants tailored parties around ideas to better organize<br />

kitchen cabinets using the new line. Another environmentally conscious<br />

product idea was the Tupperware lunch kit. These kits did<br />

away with the need for throwaway products such as paper plates,<br />

plastic storage bags, <strong>and</strong> aluminum foil. Lunch kits marketed in<br />

other countries were developed to accommodate the countries’ particular<br />

needs. For example, Japanese designs included chopsticks,<br />

while Latin American styles were designed to hold tortillas.<br />

Design Changes<br />

Tupperware designs have been well received over the years.<br />

Early designs prompted a 1947 edition of House Beautiful to call the


Tupperware / 805<br />

product “Fine Art for 39 cents.” Fifteen of Tupper’s earliest designs<br />

are housed in a permanent collection at the Museum of Modern Art<br />

in New York City. Other museums, such as the Metropolitan Museum<br />

of Art <strong>and</strong> the Brooklyn Museum, also house Tupperware designs.<br />

Tupperware established its own Museum of Historic Food<br />

Containers at its international headquarters in Florida. Despite this<br />

critical acclaim, the company faced a constant struggle to keep<br />

product lines competitive with more accessible products, such as<br />

those made by Rubbermaid, that could be found on the shelves of<br />

local grocery or department stores.<br />

Some of the biggest design changes came with the hiring of<br />

Morison Cousins in the early 1990’s. Cousins, an accomplished designer,<br />

set out to modernize the Tupperware line. He sought to return<br />

to simple, traditional styles while bringing in time-saving aspects.<br />

He changed lid designs to make them easier to clean <strong>and</strong><br />

rounded the bottoms of bowls so that every portion could be scooped<br />

out. Cousins also added thumb h<strong>and</strong>les to bowls.<br />

Backed by a knowledgeable sales force <strong>and</strong> quality product, the<br />

company experienced tremendous growth. Tupperware sales reached<br />

$25 million in 1954. By 1958, the company had grown from seven<br />

distributorships to a vast system covering the United States <strong>and</strong><br />

Canada. That same year, Brownie Wise left the company, <strong>and</strong> Tupper<br />

Plastics was sold to Rexall Drug Company for $9 million. Rexall<br />

Drug changed its name to Dart Industries, Inc., in 1969, then merged<br />

with Kraft, Inc., eleven years later to become Dart <strong>and</strong> Kraft, Inc.<br />

During this time of parent-company name changing, Tupperware<br />

continued to be an important subsidiary. Through the 1960’s <strong>and</strong><br />

1970’s, the company spread around the world, with sales in Western<br />

Europe, the Far East, <strong>and</strong> Latin America. In 1986, Dart <strong>and</strong> Kraft,<br />

Inc., split into Kraft, Inc., <strong>and</strong> Premark International, Inc., of which<br />

Dart (<strong>and</strong> therefore Tupperware) was a subsidiary. Premark International<br />

included other home product companies such as West<br />

Bend, Precor, <strong>and</strong> Florida Tile.<br />

By the early 1990’s, annual sales of Tupperware products reached<br />

$1.1 billion. Manufacturing plants in Halls, Tennessee, <strong>and</strong> Hemingway,<br />

South Carolina, worked to meet the high dem<strong>and</strong> for Tupperware<br />

products in more than fifty countries. Foreign sales accounted<br />

for almost 75 percent of the company’s business. By meeting the


806 / Tupperware<br />

needs of consumers <strong>and</strong> keeping current with design changes, new<br />

sales techniques, <strong>and</strong> new products, Tupperware was able to reach<br />

90 percent of America’s homes.<br />

See also Electric refrigerator; Food freezing; Freeze-drying; Microwave<br />

cooking; Plastic; Polystyrene; Pyrex glass; Teflon.<br />

Further Reading<br />

Brown, Patricia Leigh. “New Designs to Keep Tupperware Fresh.”<br />

New York Times (June 10, 1993).<br />

Clarke, Alison J. Tupperware: The Promise of Plastic in 1950s America.<br />

Washington, D.C.: Smithsonian Institution Press, 1999.<br />

Gershman, Michael. Getting It Right the Second Time. Reading, Mass.:<br />

Addison-Wesley, 1990.<br />

Martin, Douglas. “Morison S. Cousins, Sixty-six, Designer, Dies; Revamped<br />

Tupperware’s Look with Flair.” New York Times (February<br />

18, 2001).<br />

Sussman, Vic. “I Was the Only Virgin at the Party.” Sales <strong>and</strong> Marketing<br />

Management 141 (September 1, 1989).


Turbojet<br />

Turbojet<br />

The invention: A jet engine with a turbine-driven compressor that<br />

uses its hot-gas exhaust to develop thrust.<br />

The people behind the invention:<br />

Henry Harley Arnold (1886-1950), a chief of staff of the U.S.<br />

Army Air Corps<br />

Gerry Sayer, a chief test pilot for Gloster Aircraft Limited<br />

Hans Pabst von Ohain (1911- ), a German engineer<br />

Sir Frank Whittle (1907-1996), an English Royal Air Force<br />

officer <strong>and</strong> engineer<br />

Developments in Aircraft Design<br />

807<br />

On the morning of May 15, 1941, some eleven months after<br />

France had fallen to Adolf Hitler’s advancing German army, an experimental<br />

jet-propelled aircraft was successfully tested by pilot<br />

Gerry Sayer. The airplane had been developed in a little more than<br />

two years by the English company Gloster Aircraft under the supervision<br />

of Sir Frank Whittle, the inventor of Engl<strong>and</strong>’s first jet engine.<br />

Like the jet engine that powered it, the plane had a number of<br />

predecessors. In fact, the May, 1941, flight was not the first jetpowered<br />

test flight: That flight occurred on August 27, 1939, when a<br />

Heinkel aircraft powered by a jet engine developed by Hans Pabst<br />

von Ohain completed a successful test flight in Germany. During<br />

this period, Italian airplane builders were also engaged in jet aircraft<br />

testing, with lesser degrees of success.<br />

Without the knowledge that had been gained from Whittle’s experience<br />

in experimental aviation, the test flight at the Royal Air<br />

Force’s Cranwell airfield might never have been possible. Whittle’s<br />

repeated efforts to develop turbojet propulsion engines had begun<br />

in 1928, when, as a twenty-one-year-old Royal Air Force (RAF)<br />

flight cadet at Cranwell Academy, he wrote a thesis entitled “Future<br />

Developments in Aircraft Design.” One of the principles of Whittle’s<br />

earliest research was that if aircraft were eventually to achieve<br />

very high speeds over long distances, they would have to fly at very


808 / Turbojet<br />

high altitudes, benefiting from the reduced wind resistance encountered<br />

at such heights.<br />

Whittle later stated that the speed he had in mind at that time<br />

was about 805 kilometers per hour—close to that of the first jetpowered<br />

aircraft. His earliest idea of the engines that would be necessary<br />

for such planes focused on rocket propulsion (that is, “jets” in<br />

which the fuel <strong>and</strong> oxygen required to produce the explosion needed<br />

to propel an air vehicle are entirely contained in the engine, or, alternatively,<br />

in gas turbines driving propellers at very high speeds).<br />

Later, it occurred to him that gas turbines could be used to provide<br />

forward thrust by what would become “ordinary” jet propulsion<br />

(that is, “thermal air” engines that take from the surrounding atmosphere<br />

the oxygen they need to ignite their fuel). Eventually, such<br />

ordinary jet engines would function according to one of four possible<br />

systems: the so-called athodyd, or continuous-firing duct; the<br />

pulsejet, or intermittent-firing duct; the turbojet, or gas-turbine jet;<br />

or the propjet, which uses a gas turbine jet to rotate a conventional<br />

propeller at very high speeds.<br />

Passing the Test<br />

The aircraft that was to be used to test the flight performance<br />

was completed by April, 1941. On April 7, tests were conducted<br />

on the ground at Gloster Aircraft’s l<strong>and</strong>ing strip at Brockworth<br />

by chief test pilot Sayer. At this point, all parties concerned tried<br />

to determine whether the jet engine’s capacity would be sufficient<br />

to push the aircraft forward with enough speed to make it<br />

airborne. Sayer dared to take the plane off the ground for a limited<br />

distance of between 183 meters <strong>and</strong> 273 meters, despite the<br />

technical staff’s warnings against trying to fly in the first test<br />

flights.<br />

On May 15, the first real test was conducted at Cranwell. During<br />

that test, Sayer flew the plane, now called the Pioneer, for seventeen<br />

minutes at altitudes exceeding 300 meters <strong>and</strong> at a conservative test<br />

speed exceeding 595 kilometers per hour, which was equivalent to<br />

the top speed then possible in the RAF’s most versatile fighter<br />

plane, the Spitfire.


Once it was clear that the tests undertaken at Cranwell were not<br />

only successful but also highly promising in terms of even better<br />

performance, a second, more extensive test was set for May 21, 1941.<br />

It was this later demonstration that caused the Ministry of Air Production<br />

(MAP) to initiate the first steps to produce the Meteor jet<br />

fighter aircraft on a full industrial scale barely more than a year after<br />

the Cranwell test flight.<br />

Impact<br />

Turbojet / 809<br />

Since July, 1936, the Junkers engine <strong>and</strong> aircraft companies in<br />

Hitler’s Germany had been a part of a new secret branch dedicated<br />

to the development of a turbojet-driven aircraft. In the same period,<br />

Junkers’ rival in the German aircraft industry, Heinkel, Inc., approached<br />

von Ohain, who was far enough along in his work on the<br />

turbojet principle to have patented a device very similar to Whittle’s<br />

in 1935. Alater model of this jet engine would power a test aircraft in<br />

August, 1939.<br />

In the meantime, the wider impact of the flight was the result of<br />

decisions made by General Henry Harley Arnold, chief of staff of<br />

the U.S. Army Air Corps. Even before learning of the successful<br />

flight in May, he made arrangements to have one of Whittle’s engines<br />

shipped to the United States to be used by General Electric<br />

Company as a model for U.S. production. The engine arrived in<br />

October, 1941, <strong>and</strong> within one year, a General Electric-built engine<br />

powered a Bell Aircraft plane, the XP-59 A Airacomet, in its<br />

maiden flight.<br />

The jet airplane was not perfected in time to have any significant<br />

impact on the outcome of World War II, but all of the wartime experimental<br />

jet aircraft developments that were either sparked by the<br />

flight in 1941 or preceded it prepared the way for the research <strong>and</strong><br />

development projects that would leave a permanent revolutionary<br />

mark on aviation history in the early 1950’s.<br />

See also Airplane; Dirigible; Rocket; Rocket; Stealth aircraft; Supersonic<br />

passenger plane; V-2 rocket.


810 / Turbojet<br />

Further Reading<br />

Adams, Robert. “Smithsonian Horizons.” Smithsonian 18 (July, 1987).<br />

Boyne, Walter J., Donald S. Lopez, <strong>and</strong> Anselm Franz. The Jet Age:<br />

Forty Years of Jet Aviation. Washington: National Air <strong>and</strong> Space<br />

Museum, 1979.<br />

Constant, Edward W. The Origins of the Turbojet Revolution. Baltimore:<br />

Johns Hopkins University Press, 1980.<br />

Launius, Roger D. Innovation <strong>and</strong> the Development of Flight. College<br />

Station: Texas A&M University Press, 1999.


Typhus vaccine<br />

Typhus vaccine<br />

The invention: the first effective vaccine against the virulent typhus<br />

disease.<br />

The person behind the invention:<br />

Hans Zinsser (1878-1940), an American bacteriologist <strong>and</strong><br />

immunologist<br />

Studying Diseases<br />

811<br />

As a bacteriologist <strong>and</strong> immunologist, Hans Zinsser was interested<br />

in how infectious diseases spread. During an outbreak of typhus<br />

in Serbia in 1915, he traveled with a Red Cross team so that he<br />

could study the disease. He made similar trips to the Soviet Union<br />

in 1923, Mexico in 1931, <strong>and</strong> China in 1938. His research showed<br />

that, as had been suspected, typhus was caused by the rickettsia, an<br />

organism that had been identified in 1916 by Henrique da Rocha-<br />

Lima. The organism was known to be carried by a louse or a rat flea<br />

<strong>and</strong> transmitted to humans through a bite. Poverty, dirt, <strong>and</strong> overcrowding<br />

led to environments that helped the typhus disease to<br />

spread.<br />

The rickettsia is a microorganism that is rod-shaped or spherical.<br />

Within the insect’s body, it works its way into the cells that line the<br />

gut. Multiplying within this tissue, the rickettsia passes from the insect<br />

body with the feces. Since its internal cells are being destroyed,<br />

the insect dies within three weeks after it has been infected with the<br />

microorganism. As the infected flea or louse feeds on a human, it<br />

causes itching. When the bite is scratched, the skin may be opened,<br />

<strong>and</strong> the insect feces, carrying rickettsia, can then enter the body.<br />

Also, dried airborne feces can be inhaled.<br />

Once inside the human, the rickettsia invades endothelial cells<br />

<strong>and</strong> causes an inflammation of the blood vessels. Cell death results,<br />

<strong>and</strong> this leads to tissue death. In a few days, the infected person may<br />

have a rash, a severe headache, a fever, dizziness, ringing in the ears,<br />

or deafness. Also, light may hurt the person’s eyes, <strong>and</strong> the thinking<br />

processes become foggy <strong>and</strong> mixed up. (The word “typhus” comes


812 / Typhus vaccine<br />

from a Greek word meaning “cloudy” or “misty.”) Without treatment,<br />

the victim dies within nine to eighteen days.<br />

Medical science now recognizes three forms of typhus: the epidemic<br />

louse-borne, the Brill-Zinsser, <strong>and</strong> the murine (or rodentrelated)<br />

form. The epidemic louse-borne (or “classical”) form is the<br />

most severe. The Brill-Zinsser (or “endemic”) form is similar but<br />

less severe. The murine form of typhus is also milder then the epidemic<br />

type.<br />

In 1898, a researcher named Brill studied typhus among immigrants<br />

in New York City; the form of typhus he found was called<br />

“Brill’s disease.” In the late 1920’s, Hermann Mooser proved that<br />

Brill’s disease was carried by the rat flea.<br />

When Zinsser began his work on typhus, he realized that what<br />

was known about the disease had never been properly organized.<br />

Zinsser <strong>and</strong> his coworkers, including Mooser <strong>and</strong> others, worked to<br />

identify the various types of typhus. In the 1930’s, Zinsser suggested<br />

that the typhus studied by Brill in New York City had actually<br />

included two types: the rodent-associated form <strong>and</strong> Brill’s disease.<br />

As a result of Zinsser’s effort to identify the types of typhus<br />

disease, it was renamed Brill-Zinsser disease.<br />

Making a Vaccine<br />

Zinsser’s studies had shown him that the disease-causing organism<br />

in typhus contained some kind of antigen, most likely a polysaccharide.<br />

In 1932, Zinsser would identify agglutinins, or antibodies,<br />

in the blood serum of patients who had the murine <strong>and</strong> classical<br />

forms of typhus. Zinsser believed that a vaccine could be developed<br />

to prevent the spread of typhus. He realized, however, that a large<br />

number of dead microorganisms was needed to help people develop<br />

an immunity.<br />

Zinsser <strong>and</strong> his colleagues set out to develop a method of growing<br />

organisms in large quantities in tissue culture. The infected tissue<br />

was used to inoculate large quantities of normal chick tissue,<br />

<strong>and</strong> this tissue was then grown in flasks. In this way, Zinsser’s team<br />

was able to produce the quantities of microorganisms they needed.<br />

The type of immunization that Zinsser developed (in 1930) is<br />

known as “passive immunity.” The infecting organisms carry anti-


gens, which stimulate the production of antibodies. The antigens<br />

can elicit an immune reaction even if the cell is weak or dead.<br />

“B” cells <strong>and</strong> macrophages, both of which are used in fighting<br />

disease organisms, recognize <strong>and</strong> respond to the antigen. The B cells<br />

produce antibodies that can destroy the invading organism directly<br />

or attract more macrophages to the area so that they can attack the<br />

organism. B cells also produce “memory cells,” which remain in the<br />

blood <strong>and</strong> trigger a quick second response if there is a later infection.<br />

Since the vaccine contains weakened or dead organisms, the<br />

person who is vaccinated may have a mild reaction but does not actually<br />

come down with the disease.<br />

Impact<br />

Typhus vaccine / 813<br />

Typhus is still common in many parts of the world, especially<br />

where there is poverty <strong>and</strong> overcrowding. Classical typhus is quite<br />

rare; the last report of this type of typhus in the United States was in<br />

1921. Endemic <strong>and</strong> murine typhus are more common. In the United<br />

States, where children are vaccinated against the disease, only about<br />

fifty cases are now reported each year. Antibiotics such as tetracycline<br />

<strong>and</strong> chloramphenicol are effective in treating the disease, so<br />

few infected people now die of the disease in areas where medical<br />

care is available.<br />

The work of Zinsser <strong>and</strong> his colleagues was very important in<br />

stopping the spread of typhus. Zinsser’s classification of different<br />

types of the disease meant that it was better understood, <strong>and</strong> this<br />

led to the development of cures. The control of lice <strong>and</strong> rodents <strong>and</strong><br />

improved cleanliness in living conditions helped bring typhus under<br />

control. Once Zinsser’s vaccine was available, even people who<br />

lived in crowded inner cities could be protected against the disease.<br />

Zinsser’s research in growing the rickettsia in tissue culture also<br />

inspired further work. Other researchers modified <strong>and</strong> improved<br />

his technique so that the use of tissue culture is now st<strong>and</strong>ard in laboratories.<br />

See also Antibacterial drugs; Birth control pill; Penicillin; Polio<br />

vaccine (Sabin); Polio vaccine (Salk); Salvarsan; Tuberculosis vaccine;<br />

Yellow fever vaccine.


814 / Typhus vaccine<br />

Further Reading<br />

DeJauregui, Ruth. 100 Medical Milestones That Shaped World History.<br />

San Mateo, Calif.: Bluewood Books, 1998.<br />

Gray, Michael W. “Rickettsia in Medicine <strong>and</strong> History.” Nature 396,<br />

no. 6707 (November, 1998).<br />

Hoff, Brent H., Carter Smith, <strong>and</strong> Charles H. Calisher. Mapping Epidemics:<br />

A Historical Atlas of Disease. New York: Franklin Watts,<br />

2000.


Ultracentrifuge<br />

Ultracentrifuge<br />

The invention: A super-high-velocity centrifuge designed to separate<br />

colloidal or submicroscopic substances, the ultracentrifuge<br />

was used to measure the molecular weight of proteins <strong>and</strong><br />

proved that proteins are large molecules.<br />

The people behind the invention:<br />

Theodor Svedberg (1884-1971), a Swedish physical chemist <strong>and</strong><br />

1926 Nobel laureate in chemistry<br />

Jesse W. Beams (1898-1977), an American physicist<br />

Arne Tiselius (1902-1971), a Swedish physical biochemist <strong>and</strong><br />

1948 Nobel laureate in chemistry<br />

Svedberg Studies Colloids<br />

815<br />

Theodor “The” Svedberg became the principal founder of molecular<br />

biology when he invented the ultracentrifuge <strong>and</strong> used it to<br />

examine proteins in the mid-1920’s. He began to study materials<br />

called “colloids” as a Swedish chemistry student at the University<br />

of Uppsala <strong>and</strong> continued to conduct experiments with colloidal<br />

systems when he joined the faculty in 1907. A colloid is a kind of<br />

mixture in which very tiny particles of one substance are mixed<br />

uniformly with a dispersing medium (often water) <strong>and</strong> remain<br />

suspended indefinitely. These colloidal dispersions play an important<br />

role in many chemical <strong>and</strong> biological systems.<br />

The size of the colloid particles must fall within a certain<br />

range. The force of gravity will cause them to settle if they are too<br />

large. If they are too small, the properties of the mixture change,<br />

<strong>and</strong> a solution is formed. Some examples of colloidal systems include<br />

mayonnaise, soap foam, marshmallows, the mineral opal,<br />

fog, India ink, jelly, whipped cream, butter, paint, <strong>and</strong> milk.<br />

Svedberg wondered what such different materials could have in<br />

common. His early work helped to explain why colloids remain<br />

in suspension. Later, he developed the ultracentrifuge to measure<br />

the weight of colloid particles by causing them to settle in a controlled<br />

way.


816 / Ultracentrifuge<br />

Svedberg Builds an Ultracentrifuge<br />

Svedberg was a successful chemistry professor at the University<br />

of Uppsala in Sweden when he had the idea that colloids could be<br />

made to separate from suspension by means of centrifugal force.<br />

Centrifugal force is caused by circular motion <strong>and</strong> acts on matter<br />

much as gravity does. A person can feel this force by tying a ball to a<br />

rope <strong>and</strong> whirling it rapidly in a circle. The pull on the rope becomes<br />

stronger as the ball moves faster in its circular orbit. A centrifuge<br />

works the same way: It is a device that spins balanced containers of<br />

substances very rapidly.<br />

Svedberg figured that it would take a centrifugal force thous<strong>and</strong>s<br />

of times the force of gravity to cause colloid particles to settle. How<br />

fast they settle depends on their size <strong>and</strong> weight, so the ultracentrifuge<br />

can also provide a measure of these properties. Centrifuges were<br />

already used to separate cream from whole milk <strong>and</strong> blood corpuscles<br />

from plasma, but these centrifuges were too slow to cause the<br />

separation of colloids. An ultracentrifuge—one that could spin samples<br />

much faster—was needed, <strong>and</strong> Svedberg made plans to build one.<br />

The opportunity came in 1923, when Svedberg spent eight months<br />

as visiting professor in the chemistry department of the University<br />

of Wisconsin at Madison <strong>and</strong> worked with J. Burton Nichols, one of<br />

the six graduate students assigned to assist him. Here, Svedberg announced<br />

encouraging results with an electrically driven centrifuge—not<br />

yet an ultracentrifuge—which attained a rotation equal<br />

to about 150 times the force of gravity. Svedberg returned to Sweden<br />

<strong>and</strong>, within a year, built a centrifuge capable of generating 7,000<br />

times the force of gravity. He used it with Herman Rinde, a colleague<br />

at the University of Uppsala, to separate the suspended particles<br />

of colloidal gold. This was in 1924, which is generally accepted<br />

as the date of the first use of a true ultracentrifuge. From 1925 to<br />

1926, Svedberg raised the funds to build an even more powerful ultracentrifuge.<br />

It would be driven by an oil turbine, a machine capable<br />

of producing more than 40,000 revolutions per minute to generate<br />

a force 100,000 times that of gravity.<br />

Svedberg <strong>and</strong> Robin Fahraeus used the new ultracentrifuge to<br />

separate the protein hemoglobin from its colloidal suspension. Together<br />

with fats <strong>and</strong> carbohydrates, proteins are one of the most


abundant organic constituents of living organisms. No protein had<br />

been isolated in pure form before Svedberg began this study, <strong>and</strong> it<br />

was uncertain whether proteins consisted of molecules of a single<br />

compound or mixtures of different substances working together in<br />

biological systems. The colloid particles of Svedberg’s previous<br />

studies separated at different rates, some settling faster than others,<br />

showing that they had different sizes <strong>and</strong> weights. Colloid particles<br />

of the protein, however, separated together. The uniform separation<br />

observed for proteins, such as hemoglobin, demonstrated for the<br />

first time that each protein consists of identical well-defined molecules.<br />

More than one hundred proteins were studied by Svedberg<br />

<strong>and</strong> his coworkers, who extended their technique to carbohydrate<br />

polymers such as cellulose <strong>and</strong> starch.<br />

Impact<br />

Svedberg built more <strong>and</strong> more powerful centrifuges so that smaller<br />

<strong>and</strong> smaller molecules could be studied. In 1936, he built an ultracentrifuge<br />

that produced a centrifugal force of more than a halfmillion<br />

times the force of gravity. Jesse W. Beams was an American<br />

pioneer in ultracentrifuge design. He reduced the friction of an airdriven<br />

rotor by first housing it in a vacuum, in 1934, <strong>and</strong> later by<br />

supporting it with a magnetic field.<br />

The ultracentrifuge was a central tool for providing a modern underst<strong>and</strong>ing<br />

of the molecular basis of living systems, <strong>and</strong> it is employed<br />

in thous<strong>and</strong>s of laboratories for a variety of purposes. It is<br />

used to analyze the purity <strong>and</strong> the molecular properties of substances<br />

containing large molecules, from the natural products of the biosciences<br />

to the synthetic polymers of chemistry. The ultracentrifuge is<br />

also employed in medicine to analyze body fluids, <strong>and</strong> it is used in biology<br />

to isolate viruses <strong>and</strong> the components of fractured cells.<br />

Svedberg, while at Wisconsin in 1923, invented a second, very<br />

different method to separate proteins in suspension using electric<br />

currents. It is called “electrophoresis,” <strong>and</strong> it was later improved by<br />

his student, Arne Tiselius, for use in his famous study of the proteins<br />

in blood serum. The technique of electrophoresis is as widespread<br />

<strong>and</strong> important as is the ultracentrifuge.<br />

See also Ultramicroscope; X-ray crystallography.<br />

Ultracentrifuge / 817


818 / Ultracentrifuge<br />

Further Reading<br />

Lechner, M. D. Ultracentrifugation. New York: Springer, 1994.<br />

Rickwood, David. Preparative Centrifugation: A Practical Approach.<br />

New York: IRL Press at Oxford University Press, 1992.<br />

Schuster, Todd M. Modern Analytical Ultracentrifugation: Acquisition<br />

<strong>and</strong> Interpretation of Data for Biological <strong>and</strong> Synthetic Polymer Systems.<br />

Boston: Birkhäuser, 1994.<br />

Svedberg, Theodor B., Kai Oluf Pedersen, <strong>and</strong> Johannes Henrik<br />

Bauer. The Ultracentrifuge. Oxford: Clarendon Press, 1940.


Ultramicroscope<br />

Ultramicroscope<br />

The invention: A microscope characterized by high-intensity illumination<br />

for the study of exceptionally small objects, such as colloidal<br />

substances.<br />

The people behind the invention:<br />

Richard Zsigmondy (1865-1929), an Austrian-born German<br />

organic chemist who won the 1925 Nobel Prize in Chemistry<br />

H. F. W. Siedentopf (1872-1940), a German physicist-optician<br />

Max von Smouluchowski (1879-1961), a German organic<br />

chemist<br />

Accidents of Alchemy<br />

819<br />

Richard Zsigmondy’s invention of the ultramicroscope grew out<br />

of his interest in colloidal substances. Colloids consist of tiny particles<br />

of a substance that are dispersed throughout a solution of another<br />

material or substance (for example, salt in water). Zsigmondy<br />

first became interested in colloids while working as an assistant to<br />

the physicist Adolf Kundt at the University of Berlin in 1892. Although<br />

originally trained as an organic chemist, in which discipline<br />

he took his Ph.D. at the University of Munich in 1890, Zsigmondy<br />

became particularly interested in colloidal substances containing<br />

fine particles of gold that produce lustrous colors when painted on<br />

porcelain. For this reason, he ab<strong>and</strong>oned organic chemistry <strong>and</strong> devoted<br />

his career to the study of colloids.<br />

Zsigmondy began intensive research into his new field of interest<br />

in 1893, when he returned to Austria to accept a post as lecturer at a<br />

technical school at Graz. Zsigmondy became especially interested<br />

in gold-ruby glass, the accidental invention of the seventeenth century<br />

alchemist Johann Kunckle. Kunckle, while pursuing the alchemist’s<br />

pipe dream of transmuting base substances (such as lead)<br />

into gold, discovered instead a method of producing glass with a<br />

beautiful, deep red luster by suspending very fine particles of gold<br />

throughout the liquid glass before it was cooled. Zsigmondy also<br />

began studying a colloidal pigment called “purple of Cassius,” the<br />

discovery of another seventeenth century alchemist, Andreas Cassius.


820 / Ultramicroscope<br />

Zsigmondy soon discovered that purple of Cassius was a colloidal<br />

solution <strong>and</strong> not, as most chemists believed at the time, a chemical<br />

compound. This fact allowed him to develop techniques for<br />

glass <strong>and</strong> porcelain coloring with great commercial value, which led<br />

directly to his 1897 appointment to a research post with the Schott<br />

Glass Manufacturing Company in Jena, Germany. With the Schott<br />

Company, Zsigmondy concentrated on the commercial production<br />

of colored glass objects. His most notable achievement during this<br />

period was the invention of Jena milk glass, which is still prized by<br />

collectors throughout the world.<br />

Brilliant Proof<br />

While studying colloids, Zsigmondy devised experiments that<br />

proved that purple of Cassius was colloidal. When he published the<br />

results of his research in professional journals, however, they were<br />

not widely accepted by the scientific community. Other scientists<br />

were not able to replicate Zsigmondy’s experiments <strong>and</strong> consequently<br />

denounced them as flawed. The criticism of his work in<br />

technical literature stimulated Zsigmondy to make his greatest discovery,<br />

the ultramicroscope, which he developed to prove his theories<br />

regarding purple of Cassius.<br />

The problem with proving the exact nature of purple of Cassius<br />

was that the scientific instruments available at the time were not<br />

sensitive enough for direct observation of the particles suspended<br />

in a colloidal substance. Using the facilities <strong>and</strong> assisted by the staff<br />

(especially H. F. W. Siedentopf, an expert in optical lens grinding) of<br />

the Zeiss Glass Manufacturing Company of Jena, Zsigmondy developed<br />

an ingenious device that permitted direct observation of individual<br />

colloidal particles.<br />

This device, which its developers named the “ultramicroscope,”<br />

made use of a principle that already existed. Sometimes called “darkfield<br />

illumination,” this method consisted of shining a light (usually<br />

sunlight focused by mirrors) through the solution under the microscope<br />

at right angles to the observer, rather than shining the light directly<br />

from the observer into the solution. The resulting effect is similar<br />

to that obtained when a beam of sunlight is admitted to a closed<br />

room through a small window. If an observer st<strong>and</strong>s back from <strong>and</strong> at


ight angles to such a beam, many dust particles suspended in the air<br />

will be observed that otherwise would not be visible.<br />

Zsigmondy’s device shines a very bright light through the substance<br />

or solution being studied. From the side, the microscope then<br />

focuses on the light shaft. This process enables the observer using<br />

the ultramicroscope to view colloidal particles that are ordinarily<br />

invisible even to the strongest conventional microscope. To a scientist<br />

viewing purple of Cassius, for example, colloidal gold particles<br />

as small as one ten-millionth of a millimeter in size become visible.<br />

Impact<br />

Richard Zsigmondy<br />

Ultramicroscope / 821<br />

Born in Vienna, Austria, in 1865, Richard Adolf Zsigmondy<br />

came from a talented, energetic family. His father, a celebrated<br />

dentist <strong>and</strong> inventor of medical equipment, inspired his children<br />

to study the sciences, while his mother urged them to<br />

spend time outdoors in strenuous exercise. Although his father<br />

died when Zsigmondy was fifteen, the teenager’s interest in<br />

chemistry was already firmly established. He read advanced<br />

chemistry textbooks <strong>and</strong> worked on experiments in his own<br />

home laboratory.<br />

After taking his doctorate at the University of Munich <strong>and</strong><br />

teaching in Berlin <strong>and</strong> Graz, Austria, he became an industrial<br />

chemist at the glassworks in Jena, Germany. However, pure research<br />

was his love, <strong>and</strong> he returned to it, working entirely on<br />

his own after 1900. In 1907 he received an appointment as professor<br />

<strong>and</strong> director of the Institute of Inorganic Chemistry at the<br />

University of Göttingen, one of the scientific centers of the<br />

world. There he accomplished much of his ground-breaking<br />

work on colloids <strong>and</strong> Brownian motion, despite the severe<br />

shortages that hampered him during the economic depression<br />

in Germany following World War I. His 1925 Nobel Prize in<br />

Chemistry, especially the substantial money award, helped him<br />

overcome his supply problems. He retired in early 1929 <strong>and</strong><br />

died seven months later.<br />

After Zsigmondy’s invention of the ultramicroscope in 1902,<br />

the University of Göttingen appointed him professor of inorganic


822 / Ultramicroscope<br />

chemistry <strong>and</strong> director of its Institute for Inorganic Chemistry.<br />

Using the ultramicroscope, Zsigmondy <strong>and</strong> his associates quickly<br />

proved that purple of Cassius is indeed a colloidal substance.<br />

That finding, however, was the least of the spectacular discoveries<br />

that resulted from Zsigmondy’s invention. In the next decade,<br />

Zsigmondy <strong>and</strong> his associates found that color changes in colloidal<br />

gold solutions result from coagulation—that is, from changes in the<br />

size <strong>and</strong> number of gold particles in the solution caused by particles<br />

bonding together. Zsigmondy found that coagulation occurs when<br />

the negative electrical charge of the individual particles is removed<br />

by the addition of salts. Coagulation can be prevented or slowed by<br />

the addition of protective colloids.<br />

These observations also made possible the determination of the<br />

speed at which coagulation takes place, as well as the number of particles<br />

in the colloidal substance being studied. With the assistance of<br />

the organic chemist Max von Smouluchowski, Zsigmondy worked<br />

out a complete mathematical formula of colloidal coagulation that is<br />

valid not only for gold colloidal solutions but also for all other<br />

colloids. Colloidal substances include blood <strong>and</strong> milk, which both coagulate,<br />

thus giving Zsigmondy’s work relevance to the fields of<br />

medicine <strong>and</strong> agriculture. These observations <strong>and</strong> discoveries concerning<br />

colloids—in addition to the invention of the ultramicroscope—earned<br />

for Zsigmondy the 1925 Nobel Prize in Chemistry.<br />

See also Scanning tunneling microscope; Ultracentrifuge; X-ray<br />

crystallography.<br />

Further Reading<br />

Zsigmondy, Richard, <strong>and</strong> Jerome Alex<strong>and</strong>er. Colloids <strong>and</strong> the Ultramicroscope.<br />

New York: J. Wiley & Sons, 1909.<br />

Zsigmondy, Richard, Ellwood Barker Spear, <strong>and</strong> John Foote Norton.<br />

The Chemistry of Colloids. New York: John Wiley & Sons, 1917.


Ultrasound<br />

Ultrasound<br />

The invention: A medically safe alternative to X-ray examination,<br />

ultrasound uses sound waves to detect fetal problems in pregnant<br />

women.<br />

The people behind the invention:<br />

Ian T. Donald (1910-1987), a British obstetrician<br />

Paul Langévin (1872-1946), a French physicist<br />

Marie Curie (1867-1946) <strong>and</strong> Pierre Curie (1859-1906), the<br />

French husb<strong>and</strong>-<strong>and</strong>-wife team that researched <strong>and</strong><br />

developed the field of radioactivity<br />

Alice Stewart, a British researcher<br />

An Underwater Beginning<br />

823<br />

In the early 1900’s, two major events made it essential to develop<br />

an appropriate means for detecting unseen underwater objects. The<br />

first event was the Titanic disaster in 1912, which involved a largely<br />

submerged, unseen, <strong>and</strong> silent iceberg. This iceberg caused the sinking<br />

of the Titanic <strong>and</strong> resulted in the loss of many lives as well as<br />

valuable treasure. The second event was the threat to the Allied<br />

Powers from German U-boats during World War I (1914-1918). This<br />

threat persuaded the French <strong>and</strong> English Admiralties to form a joint<br />

committee in 1917. The Anti-Submarine Detection <strong>and</strong> Investigation<br />

Committee (ASDIC) found ways to counter the German naval<br />

developments. Paul Langévin, a former colleague of Pierre Curie<br />

<strong>and</strong> Marie Curie, applied techniques developed in the Curies’ laboratories<br />

in 1880 to formulate a crude ultrasonic system to detect submarines.<br />

These techniques used beams of sound waves of very high<br />

frequency that were highly focused <strong>and</strong> directional.<br />

The advent of World War II (1939-1945) made necessary the development<br />

of faster electronic detection technology to improve the efforts<br />

of ultrasound researchers. Langévin’s crude invention evolved<br />

into the sophisticated system called “sonar” (sound navigation ranging),<br />

which was important in the success of the Allied forces. Sonar<br />

was based on pulse echo principles <strong>and</strong>, like the system called “ra-


824 / Ultrasound<br />

dar” (radio detecting <strong>and</strong> ranging), had military implications. This vital<br />

technology was classified as a military secret <strong>and</strong> was kept hidden<br />

until after the war.<br />

An Alternative to X Rays<br />

Ian Donald<br />

Ian Donald was born in Paisley, Scotl<strong>and</strong>, in 1910 <strong>and</strong> educated<br />

in Edinburgh until he was twenty, when he moved to<br />

South Africa with his parents. He graduated with a bachelor of<br />

arts degree from Diocesan College, Cape Town, <strong>and</strong> then moved<br />

to London to study medicine, graduating from the University of<br />

London in 1937. During World War II he served as a medical officer<br />

in the Royal Air Force <strong>and</strong> received a medal for rescuing flyers<br />

from a burning airplane. After the war he began his long<br />

teaching career in medicine, first at St. Thomas Hospital Medical<br />

School <strong>and</strong> then as the Regius Professor of Midwifery at Glasgow<br />

University. His specialties were obstetrics <strong>and</strong> gynecology.<br />

While at Glasgow he accomplished his pioneering work<br />

with diagnostic ultrasound technology, but he also championed<br />

laparoscopy, breast feeding, <strong>and</strong> the preservation of membranes<br />

during the delivery of babies. In addition to his teaching<br />

duties <strong>and</strong> medical practice he wrote a widely used textbook,<br />

oversaw the building of the Queen Mother’s Hospital in Glasgow,<br />

<strong>and</strong> campaigned against Engl<strong>and</strong>’s 1967 Abortion Act.<br />

His expertise with ultrasound came to his own rescue after<br />

he had cardiac surgery in the 1960’s. He diagnosed himself as<br />

having internal bleeding from a broken blood vessel. The cardiologists<br />

taking care of him were skeptical until an ultrasound<br />

proved him right. Widely honored among physicians, he died<br />

in Engl<strong>and</strong> in 1987.<br />

Ian Donald’s interest in engineering <strong>and</strong> the principles of<br />

sound waves began when he was a schoolboy. Later, while he was<br />

in the British Royal Air Force, he continued <strong>and</strong> maintained his<br />

enthusiasm by observing the development of the anti-U-boat<br />

warfare efforts. He went to medical school after World War II <strong>and</strong><br />

began a career in obstetrics. By the early 1950’s, Donald had em-


Ultrasound / 825<br />

Safe <strong>and</strong> not requiring surgery, ultrasonography has become the principal means for obtaining<br />

information about fetal structures. (Digital Stock)<br />

barked on a study of how to apply sonar technology in medicine.<br />

He moved to Glasgow, Scotl<strong>and</strong>, a major engineering center in<br />

Europe that presented a fertile environment for interdisciplinary<br />

research. There Donald collaborated with engineers <strong>and</strong> technicians<br />

in his medical ultrasound research. They used inanimate<br />

<strong>and</strong> tissue materials in many trials. Donald hoped to apply ultrasound<br />

technology to medicine, especially to gynecology <strong>and</strong> obstetrics,<br />

his specialty.<br />

His efforts led to new pathways <strong>and</strong> new discoveries. He was interested<br />

in adapting a certain type of ultrasound technology method<br />

(used to probe metal structures <strong>and</strong> welds for cracks <strong>and</strong> flaws) to<br />

medicine. Kelvin Hughes, the engineering manufacturing company<br />

that produced the flaw detector apparatus, gave advice, expertise,<br />

<strong>and</strong> equipment to Donald <strong>and</strong> his associates, who were then able to<br />

devise water tanks with flexible latex bottoms. These were coated<br />

with a film of grease <strong>and</strong> placed into contact with the abdomens of<br />

pregnant women.<br />

The use of diagnostic radiography (such as X rays) became controversial<br />

when it was evident that it caused potential leukemias


826 / Ultrasound<br />

<strong>and</strong> other injuries to the fetus. It was realized from the earliest days<br />

of radiology that radiation could cause tumors, particularly of the<br />

skin. The aftereffects of radiological studies were recognized much<br />

later <strong>and</strong> confirmed by studies of atomic bomb survivors <strong>and</strong> of patients<br />

receiving therapeutic irradiation. The use of radiation in obstetrics<br />

posed several major threats to the developing fetus, most<br />

notably the production of tumors later in life, genetic damage, <strong>and</strong><br />

developmental anomalies in the unborn fetus.<br />

In 1958, bolstered by earlier clinical reports <strong>and</strong> animal research<br />

findings, Alice Stewart <strong>and</strong> her colleagues presented a major case<br />

study of more than thirteen hundred children in Engl<strong>and</strong> <strong>and</strong> Wales<br />

who had died of cancer before the age of ten between 1953 <strong>and</strong> 1958.<br />

There was a 91 percent increase in leukemias in children who were<br />

exposed to intrauterine radiation, as well as a higher percentage of<br />

fetal death. Although controversial, this report led to a reduction in<br />

the exposure of pregnant women to X rays, with subsequent reductions<br />

in fetal abnormalities <strong>and</strong> death.<br />

These reports came at a very opportune time for Donald: The development<br />

of ultrasonography would provide useful information<br />

about the unborn fetus without the adverse effects of radiation.<br />

Stewart’s findings <strong>and</strong> Donald’s experiments convinced others of<br />

the need for ultrasonography in obstetrics.<br />

Consequences<br />

Diagnostic ultrasound first gained clinical acceptance in obstetrics,<br />

<strong>and</strong> its major contributions have been in the assessment of fetal<br />

size <strong>and</strong> growth. In combination with amniocentesis (the study of<br />

fluid taken from the womb), ultrasound is an invaluable tool in operative<br />

procedures necessary to improve the outcomes of pregnancies.<br />

As can be expected, safety has been a concern, especially for a developing,<br />

vulnerable fetus that is exposed to high-frequency sound.<br />

Research has not been able to document any harmful effect of ultrasonography<br />

on the developing fetus. The procedure produces neither<br />

heat nor cold. It has not been shown to produce any toxic or destructive<br />

effect on the auditory or balancing organs of the<br />

developing fetus. Chromosomal abnormalities have not been reported<br />

in any of the studies conducted.


Ultrasonography, because it is safe <strong>and</strong> does not require surgery,<br />

has become the principal means for obtaining information about fetal<br />

structures. With this procedure, the contents of the uterus—as<br />

well as the internal structure of the placenta, fetus, <strong>and</strong> fetal organs—can<br />

be evaluated at any time during pregnancy. The use of<br />

ultrasonography remains a most valued tool in medicine, especially<br />

obstetrics, because of Donald’s work.<br />

See also Amniocentesis; Birth control pill; CAT scanner; Electrocardiogram;<br />

Electroencephalogram; Mammography; Nuclear magnetic<br />

resonance; Pap test; Sonar; Syphilis test; X-ray image intensifier.<br />

Further Reading<br />

Ultrasound / 827<br />

Danforth, David N., <strong>and</strong> James R. Scott. Danforth’s Obstetrics <strong>and</strong> Gynecology.<br />

7th ed. Philadelphia: Lippincott, 1994.<br />

DeJauregui, Ruth. 100 Medical Milestones That Shaped World History.<br />

San Mateo, Calif.: Bluewood Books, 1998.<br />

Rozycki, Grace S. Surgeon-Performed Ultrasound: Its Use in Clinical<br />

Practice. Philadelphia: W. B. Saunders, 1998.<br />

Wolbarst, Anthony B. Looking Within: How X-ray, CT, MRI, Ultrasound,<br />

<strong>and</strong> Other Medical Images Are Created, <strong>and</strong> How They Help<br />

Physicians Save Lives. Berkeley: University of California Press,<br />

1999.


828<br />

UNIVAC computer<br />

UNIVAC computer<br />

The invention: The first commercially successful computer system.<br />

The people behind the invention:<br />

John Presper Eckert (1919-1995), an American electrical engineer<br />

John W. Mauchly (1907-1980), an American physicist<br />

John von Neumann (1903-1957), a Hungarian American<br />

mathematician<br />

Howard Aiken (1900-1973), an American physicist<br />

George Stibitz (1904-1995), a scientist at Bell Labs<br />

The Origins of Computing<br />

On March 31, 1951, the U.S. Census Bureau accepted delivery of<br />

the first Universal Automatic Computer (UNIVAC). This powerful<br />

electronic computer, far surpassing anything then available in technological<br />

features <strong>and</strong> capability, ushered in the first computer generation<br />

<strong>and</strong> pioneered the commercialization of what had previously<br />

been the domain of academia <strong>and</strong> the interest of the military. The fanfare<br />

that surrounded this historic occasion, however, masked the turbulence<br />

of the previous five years for the young upstart Eckert-<br />

Mauchly Computer Corporation (EMCC), which by this time was a<br />

wholly owned subsidiary of Remington R<strong>and</strong> Corporation.<br />

John Presper Eckert <strong>and</strong> John W. Mauchly met in the summer of<br />

1941 at the University of Pennsylvania. A short time later, Mauchly,<br />

then a physics professor at Ursinus College, joined the Moore School<br />

of Engineering at the University of Pennsylvania <strong>and</strong> embarked on a<br />

crusade to convince others of the feasibility of creating electronic digital<br />

computers. Up to this time, the only computers available were<br />

called “differential analyzers,” which were used to solve complex<br />

mathematical equations known as “differential equations.” These<br />

slow machines were good only for solving a relatively narrow range<br />

of mathematical problems.<br />

Eckert <strong>and</strong> Mauchly l<strong>and</strong>ed a contract that eventually resulted in<br />

the development <strong>and</strong> construction of the world’s first operational


general-purpose electronic computer, the Electronic Numerical Integrator<br />

<strong>and</strong> Calculator (ENIAC). This computer, used eventually<br />

by the Army for the calculation of ballistics tables, was deficient in<br />

many obvious areas, but this was caused by economic rather than<br />

engineering constraints. One major deficiency was the lack of automatic<br />

program control; the ENIAC did not have stored program<br />

memory. This was addressed in the development of the Electronic<br />

Discrete Variable Automatic Computer (EDVAC), the successor to<br />

the ENIAC.<br />

Fighting the Establishment<br />

UNIVAC computer / 829<br />

A symbiotic relationship had developed between Eckert <strong>and</strong><br />

Mauchly that worked to their advantage on technical matters.<br />

They worked well with each other, <strong>and</strong> this contributed to their<br />

success in spite of external obstacles. They both were interested in<br />

the commercial applications of computers <strong>and</strong> envisioned uses for<br />

these machines far beyond the narrow applications required by<br />

the military.<br />

This interest brought them into conflict with the administration<br />

at the Moore School of Engineering as well as with the noted mathematician<br />

John von Neumann, who “joined” the ENIAC/EDVAC<br />

development team in 1945. Von Neumann made significant contributions<br />

<strong>and</strong> added credibility to the Moore School group, which often<br />

had to fight against the conservative scientific establishment<br />

characterized by Howard Aiken at Harvard University <strong>and</strong> George<br />

Stibitz at Bell Labs. Philosophical differences between von Neumann<br />

<strong>and</strong> Eckert <strong>and</strong> Mauchly, as well as patent issue disputes with<br />

the Moore School administration, eventually caused the resignation<br />

of Eckert <strong>and</strong> Mauchly on March 31, 1946.<br />

Eckert <strong>and</strong> Mauchly, along with some of their engineering colleagues<br />

at the University of Pennsylvania, formed the Electronic<br />

Control Company <strong>and</strong> proceeded to interest potential customers<br />

(including the Census Bureau) in an “EDVAC-type” machine. On<br />

May 24, 1947, the EDVAC-type machine became the UNIVAC. This<br />

new computer would overcome the shortcomings of the ENIAC<br />

<strong>and</strong> the EDVAC (which was eventually completed by the Moore<br />

School in 1951). It would be a stored-program computer <strong>and</strong> would


830 / UNIVAC computer<br />

allow input to <strong>and</strong> output from the computer via magnetic tape. The<br />

prior method of input/output used punched paper cards that were<br />

extremely slow compared to the speed at which data in the computer<br />

could be processed.<br />

A series of poor business decisions <strong>and</strong> other unfortunate circumstances<br />

forced the newly renamed Eckert-Mauchly Computer<br />

Corporation to look for a buyer. They found one in Remington R<strong>and</strong><br />

in 1950. Remington R<strong>and</strong> built tabulating equipment <strong>and</strong> was a<br />

competitor of International Business Machines Corporation (IBM).<br />

IBM was approached about buying EMCC, but the negotiations fell<br />

apart. EMCC became a division of Remington R<strong>and</strong> <strong>and</strong> had access<br />

to the resources necessary to finish the UNIVAC.<br />

Consequences<br />

Eckert <strong>and</strong> Mauchly made a significant contribution to the advent<br />

of the computer age with the introduction of the UNIVAC I.<br />

The words “computer” <strong>and</strong> “UNIVAC” entered the popular vocabulary<br />

as synonyms. The efforts of these two visionaries were rewarded<br />

quickly as contracts started to pour in, taking IBM by surprise<br />

<strong>and</strong> propelling the inventors into the national spotlight.<br />

This spotlight shone brightest, perhaps, on the eve of the national<br />

presidential election of 1952, which pitted war hero General Dwight<br />

D. Eisenhower against statesman Adlai Stevenson. At the suggestion<br />

of Remington R<strong>and</strong>, CBS was invited to use UNIVAC to predict<br />

the outcome of the election. Millions of television viewers watched<br />

as CBS anchorman Walter Cronkite “asked” UNIVAC for its predictions.<br />

A program had been written to analyze the results of thous<strong>and</strong>s<br />

of voting districts in the elections of 1944 <strong>and</strong> 1948. Based on<br />

only 7 percent of the votes coming in, UNIVAC had Eisenhower<br />

winning by a l<strong>and</strong>slide, in contrast with all the prior human forecasts<br />

of a close election. Surprised by this answer <strong>and</strong> not willing to<br />

suffer the embarrassment of being wrong, the programmers quickly<br />

directed the program to provide an answer that was closer to the<br />

perceived situation. The outcome of the election, however, matched<br />

UNIVAC’s original answer. This prompted CBS commentator Edward<br />

R. Murrow’s famous quote, “The trouble with machines is<br />

people.”


The development of the UNIVAC I produced many technical innovations.<br />

Primary among these is the use of magnetic tape for input<br />

<strong>and</strong> output. All machines that preceded the UNIVAC (with one<br />

exception) used either paper tape or cards for input <strong>and</strong> cards for<br />

output. These methods were very slow <strong>and</strong> created a bottleneck of<br />

information. The great advantage of magnetic tape was the ability<br />

to store the equivalent of thous<strong>and</strong>s of cards of data on one 30-centimeter<br />

reel of tape. Another advantage was its speed.<br />

See also Apple II computer; BINAC computer; Colossus computer;<br />

ENIAC computer; IBM Model 1401 computer; Personal computer;<br />

Supercomputer.<br />

Further Reading<br />

UNIVAC computer / 831<br />

Metropolis, Nicholas, Jack Howlett, <strong>and</strong> Gian Carlo Rota. A History<br />

of Computing in the Twentieth Century: A Collection of Essays. New<br />

York: Academic Press, 1980.<br />

Slater, Robert. Portraits in Silicon. Cambridge, Mass.: MIT Press,<br />

1987.<br />

Stern, Nancy B. From ENIAC to UNIVAC: An Appraisal of the Eckert-<br />

Mauchly Computers. Bedford, Mass.: Digital Press, 1981.


832<br />

Vacuum cleaner<br />

Vacuum cleaner<br />

The invention: The first portable domestic vacuum cleaner successfully<br />

adapted to electricity, the original machine helped begin<br />

the electrification of domestic appliances in the early twentieth<br />

century.<br />

The people behind the invention:<br />

H. Cecil Booth (1871-1955), a British civil engineer<br />

Melville R. Bissell (1843-1889), the inventor <strong>and</strong> marketer of the<br />

Bissell carpet sweeper in 1876<br />

William Henry Hoover (1849-1932), an American industrialist<br />

James Murray Spangler (1848-1915), an American inventor<br />

From Brooms to Bissells<br />

During most of the nineteenth century, the floors of homes<br />

were cleaned primarily with brooms. Carpets were periodically<br />

dragged out of the home by the boys <strong>and</strong> men of the family,<br />

stretched over rope lines or fences, <strong>and</strong> given a thorough beating<br />

to remove dust <strong>and</strong> dirt. In the second half of the century, carpet<br />

sweepers, perhaps inspired by the success of street-sweeping machines,<br />

began to appear. Although there were many models, nearly<br />

all were based upon the idea of a revolving brush within an outer<br />

casing that moved on rollers or wheels when pushed by a long<br />

h<strong>and</strong>le.<br />

Melville Bissell’s sweeper, patented in 1876, featured a knob for<br />

adjusting the brushes to the surface. The Bissell Carpet Company,<br />

also formed in 1876, became the most successful maker of carpet<br />

sweepers <strong>and</strong> dominated the market well into the twentieth century.<br />

Electric vacuum cleaners were not feasible until homes were<br />

wired for electricity <strong>and</strong> the small electric motor was invented.<br />

Thomas Edison’s success with an inc<strong>and</strong>escent lighting system in<br />

the 1880’s <strong>and</strong> Nikola Tesla’s invention of a small electric motor<br />

that was used in 1889 to drive a Westinghouse Electric Corporation<br />

fan opened the way for the application of electricity to household<br />

technologies.


Cleaning with Electricity<br />

Vacuum cleaner / 833<br />

In 1901, H. Cecil Booth, a British civil engineer, observed a London<br />

demonstration of an American carpet cleaner that blew compressed<br />

air at the fabric. Booth was convinced that the process<br />

should be reversed so that dirt would be sucked out of the carpet. In<br />

developing this idea, Booth invented the first successful suction<br />

vacuum sweeper.<br />

Booth’s machines, which were powered by gasoline or electricity,<br />

worked without brushes. Dust was extracted by means of a<br />

suction action through flexible tubes with slot-shaped nozzles.<br />

Some machines were permanently installed in buildings that had<br />

wall sockets for the tubes in every room. Booth’s British Vacuum<br />

Cleaner Company also employed horse-drawn mobile units from<br />

which white-uniformed men unrolled long tubes that they passed<br />

into buildings through windows <strong>and</strong> doors. His company’s commercial<br />

triumph came when it cleaned Westminster Abbey for the<br />

coronation of Edward VII in 1902. Booth’s company also manufactured<br />

a 1904 domestic model that had a direct-current electric motor<br />

<strong>and</strong> a vacuum pump mounted on a wheeled carriage. Dust was<br />

sucked into the nozzle of a long tube <strong>and</strong> deposited into a metal<br />

container. Booth’s vacuum cleaner used electricity from overhead<br />

light sockets.<br />

The portable electric vacuum cleaner was invented in 1907 in the<br />

United States by James Murray Spangler. When Spangler was a janitor<br />

in a department store in Canton, Ohio, his asthmatic condition<br />

was worsened by the dust he raised with a large Bissell carpet<br />

sweeper. Spangler’s modifications of the Bissell sweeper led to his<br />

own invention. On June 2, 1908, he received a patent for his Electric<br />

Suction Sweeper. The device consisted of a cylindrical brush in the<br />

front of the machine, a vertical-shaft electric motor above a fan in<br />

the main body, <strong>and</strong> a pillowcase attached to a broom h<strong>and</strong>le behind<br />

the main body. The brush dislodged the dirt, which was sucked into<br />

the pillowcase by the movement of air caused by a fan powered by<br />

the electric motor. Although Spangler’s initial attempt to manufacture<br />

<strong>and</strong> sell his machines failed, Spangler had, luckily for him, sold<br />

one of his machines to a cousin, Susan Troxel Hoover, the wife of<br />

William Henry Hoover.


834 / Vacuum cleaner<br />

The Hoover family was involved in the production of leather<br />

goods, with an emphasis on horse saddles <strong>and</strong> harnesses. William<br />

Henry Hoover, president of the Hoover Company, recognizing that<br />

the adoption of the automobile was having a serious impact on the<br />

family business, was open to investigating another area of production.<br />

In addition, Mrs. Hoover liked the Spangler machine that she<br />

had been using for a couple of months, <strong>and</strong> she encouraged her husb<strong>and</strong><br />

to enter into an agreement with Spangler. An agreement made<br />

on August 5, 1908, allowed Spangler, as production manager, to<br />

manufacture his machine with a small work force in a section of<br />

Hoover’s plant. As sales of vacuum cleaners increased, what began<br />

as a sideline for the Hoover Company became the company’s main<br />

line of production.<br />

Few American homes were wired for electricity when Spangler<br />

<strong>and</strong> Hoover joined forces; not until 1920 did 35 percent of American<br />

homes have electric power. In addition to this inauspicious fact, the<br />

first Spangler-Hoover machine, the Model O, carried the relatively<br />

high price of seventy-five dollars. Yet a full-page ad for the Model O<br />

in the December, 1908, issue of the Saturday Evening Post brought a<br />

deluge of requests. American women had heard of the excellent performance<br />

of commercial vacuum cleaners, <strong>and</strong> they hoped that the<br />

Hoover domestic model would do as well in the home.<br />

Impact<br />

As more <strong>and</strong> more homes in the United States <strong>and</strong> abroad became<br />

wired for electric lighting, a clean <strong>and</strong> accessible power<br />

source became available for household technologies. Whereas electric<br />

lighting was needed only in the evening, the electrification of<br />

household technologies made it necessary to use electricity during<br />

the day. The electrification of domestic technologies therefore<br />

matched the needs of the utility companies, which sought to maximize<br />

the use of their facilities. They became key promoters of electric<br />

appliances. In the first decades of the twentieth century, many<br />

household technologies became electrified. In addition to fans <strong>and</strong><br />

vacuum cleaners, clothes-washing machines, irons, toasters, dishwashing<br />

machines, refrigerators, <strong>and</strong> kitchen ranges were being<br />

powered by electricity.


Vacuum cleaner / 835<br />

The application of electricity to household technologies came as<br />

large numbers of women entered the work force. During <strong>and</strong> after<br />

World War I, women found new employment opportunities in industrial<br />

manufacturing, department stores, <strong>and</strong> offices. The employment<br />

of women outside the home continued to increase throughout the<br />

twentieth century. Electrical appliances provided the means by which<br />

families could maintain the same st<strong>and</strong>ards of living in the home while<br />

both parents worked outside the home.<br />

It is significant that Bissell was motivated by an allergy to dust<br />

<strong>and</strong> Spangler by an asthmatic condition. The employment of the<br />

carpet sweeper, <strong>and</strong> especially the electric vacuum cleaner, not only<br />

H. Cecil Booth<br />

Although Hubert Cecil Booth (1871-1955), an English civil<br />

engineer, designed battleship engines, factories, <strong>and</strong> bridges, he<br />

was not above working on homier problems when they intrigued<br />

him. That happened in 1900 when he watched the demonstration<br />

of a device that used forced air to blow the dirt out of<br />

railway cars. It worked poorly, <strong>and</strong> the reason, it seemed to<br />

Booth, was that blowing just stirred up the dirt. Sucking it into a<br />

receptacle, he thought, would work better. He tested his idea by<br />

placing a wet cloth over furniture upholstery <strong>and</strong> sucking through<br />

it. The grime that collected on the side of the cloth facing the upholstery<br />

proved him right.<br />

He built his first vacuum cleaner—a term that he coined—in<br />

1901. It cleaned houses, but only with considerable effort. Measuring<br />

54 inches by 42 inches by 10 inches, it had to be carried in<br />

a horse-driven van to the cleaning site. A team of workmen<br />

from Booth’s Vacuum Cleaner Company then did the cleaning<br />

with hoses that reached inside the house through windows <strong>and</strong><br />

doors. Moreover, the machine cost the equivalent of more than<br />

fifteen hundred dollars. It was beyond the finances <strong>and</strong> physical<br />

powers of home owners.<br />

Booth marketed the first successful British one-person vacuum<br />

cleaner, the Trolley-Vac, in 1906. Weighing one hundred<br />

pounds, it was still difficult to wrestle into position, but it came<br />

with hoses <strong>and</strong> attachments that made possible the cleaning of<br />

different types of surfaces <strong>and</strong> hard-to-reach areas.


836 / Vacuum cleaner<br />

made house cleaning more efficient <strong>and</strong> less physical but also led to<br />

a healthier home environment. Whereas sweeping with a broom<br />

tended only to move dust to a different location, the carpet sweeper<br />

<strong>and</strong> the electric vacuum cleaner removed the dirt from the house.<br />

See also Disposable razor; Electric refrigerator; Microwave cooking;<br />

Robot (household); Washing machine.<br />

Further Reading<br />

Jailer-Chamberlain, Mildred. “This Is the Way We Cleaned Our<br />

Floors.” Antiques & Collecting Magazine 101, no. 4 (June, 1996).<br />

Kirkpatrick, David D. “The Ultimate Victory of Vacuum Cleaners.”<br />

New York Times (April 14, 2001).<br />

Shapiro, Laura. “Household Appliances.” Newsweek 130, no. 24A<br />

(Winter, 1997/1998).


Vacuum tube<br />

Vacuum tube<br />

The invention: A sealed glass tube from which air <strong>and</strong> gas have<br />

been removed to permit electrons to move more freely, the vacuum<br />

tube was the heart of electronic systems until it was displaced<br />

by transistors.<br />

The people behind the invention:<br />

Sir John Ambrose Fleming (1849-1945), an English physicist<br />

<strong>and</strong> professor of electrical engineering<br />

Thomas Alva Edison (1847-1931), an American inventor<br />

Lee de Forest (1873-1961), an American scientist <strong>and</strong> inventor<br />

Arthur Wehnelt (1871-1944), a German inventor<br />

A Solution in Search of a Problem<br />

837<br />

The vacuum tube is a sealed tube or container from which almost<br />

all the air has been pumped out, thus creating a near vacuum. When<br />

the tube is in operation, currents of electricity are made to travel<br />

through it. The most widely used vacuum tubes are cathode-ray<br />

tubes (television picture tubes).<br />

The most important discovery leading to the invention of the<br />

vacuum tube was the Edison effect by Thomas Alva Edison in<br />

1884. While studying why the inner glass surface of light bulbs<br />

blackened, Edison inserted a metal plate near the filament of one<br />

of his light bulbs. He discovered that electricity would flow from<br />

the positive side of the filament to the plate, but not from the neg<br />

ative side to the plate. Edison offered no explanation for the effect.<br />

Edison had, in fact, invented the first vacuum tube, which was<br />

later termed the diode; at that time there was no use for this device.<br />

Therefore, the discovery was not recognized for its true significance.<br />

A diode converts electricity that alternates in direction (alternating<br />

current) to electricity that flows in the same direction (direct<br />

current). Since Edison was more concerned with producing<br />

direct current in generators, <strong>and</strong> not household electric lamps, he<br />

essentially ignored this aspect of his discovery. Like many other in-


838 / Vacuum tube<br />

ventions or discoveries that were ahead of their time—such as the<br />

laser—for a number of years, the Edison effect was “a solution in<br />

search of a problem.”<br />

The explanation for why this phenomenon occurred would not<br />

come until after the discovery of the electron in 1897 by Sir Joseph<br />

John Thomson, an English physicist. In retrospect, the Edison effect<br />

can be identified as one of the first observations of thermionic emission,<br />

the freeing up of electrons by the application of heat. Electrons<br />

were attracted to the positive charges <strong>and</strong> would collect on the positively<br />

charged plate, thus providing current; but they were repelled<br />

from the plate when it was made negative, meaning that no current<br />

was produced. Since the diode permitted the electrical current to<br />

flow in only one direction, it was compared to a valve that allowed a<br />

liquid to flow in only one direction. This analogy is popular since<br />

the behavior of water has often been used as an analogy for electricity,<br />

<strong>and</strong> this is the reason that the term valves became popular for<br />

vacuum tubes.<br />

Same Device, Different Application<br />

Sir John Ambrose Fleming, acting as adviser to the Edison Electric<br />

Light Company, had studied the light bulb <strong>and</strong> the Edison effect<br />

starting in the early 1880’s, before the days of radio. Many years<br />

later, he came up with an application for the Edison effect as a radio<br />

detector when he was a consultant for the Marconi Wireless Telegraph<br />

Company. Detectors (devices that conduct electricity in one<br />

direction only, just as the diode does, but at higher frequencies)<br />

were required to make the high-frequency radio waves audible by<br />

converting them from alternating current to direct current. Fleming<br />

was able to detect radio waves quite effectively by using the Edison<br />

effect. Fleming used essentially the same device that Edison had created,<br />

but for a different purpose. Fleming applied for a patent on his<br />

detector on November 16, 1904.<br />

In 1906, Lee de Forest refined Fleming’s invention by adding a<br />

zigzag piece of wire between the metal plate <strong>and</strong> the filament of the<br />

vacuum tube. The zigzag piece of wire was later replaced by a<br />

screen called a “grid.” The grid allowed a small voltage to control a<br />

larger voltage between the filament <strong>and</strong> plate. It was the first com-


John Ambrose Fleming<br />

Vacuum tube / 839<br />

John Ambrose Fleming had a remarkably long <strong>and</strong> fruitful<br />

scientific career. He was born in Lancaster, Engl<strong>and</strong>, in 1849, the<br />

eldest son of a minister. When he was a boy, the family moved<br />

to London, which remained his home for the rest of his life. An<br />

outst<strong>and</strong>ing student, Fleming matriculated at University College,<br />

London, graduating in 1870 with honors. Scholarships<br />

took him to other colleges until his skill with electrical experiments<br />

earned him a job as a lab instructor at Cambridge University<br />

in 1880. In 1885, he returned to University College, London,<br />

as professor of electrical technology. He taught there for the following<br />

forty-one years, occasionally taking time off to serve as a<br />

consultant for such electronics industry leaders as Thomas Edison<br />

<strong>and</strong> Guglielmo Marconi.<br />

Fleming’s passion was electricity <strong>and</strong> electronics, <strong>and</strong> he<br />

was sought after as a teacher with a knack for memorable explanations.<br />

For instance, he thought up the “right-h<strong>and</strong>” rule (also<br />

called Fleming’s rule) to illustrate the relation of electromagnetic<br />

forces during induction: When the thumb, index finger,<br />

<strong>and</strong> middle finger of a human h<strong>and</strong> are held at right angles to<br />

one another so that the thumb points in the direction of motion<br />

through a magnetic field—which is indicated by the index finger—then<br />

the middle finger shows the direction of induced<br />

current. During his extensive research, Fleming investigated<br />

transformers, high-voltage transmitters, electrical conduction,<br />

cryogenic electrical effects, radio, <strong>and</strong> television, <strong>and</strong> also invented<br />

the vacuum tube.<br />

Advanced age hardly slowed him down. He wrote three<br />

books <strong>and</strong> more than one hundred articles <strong>and</strong> remarried at<br />

eighty-four. He also delivered public lectures—to audiences at<br />

the Royal Institution <strong>and</strong> the Royal Society among other venues—<br />

until he was ninety. He died in 1945, ninety-five years old,<br />

having helped give birth to telecommunications.<br />

plete vacuum tube <strong>and</strong> the first device ever constructed capable of<br />

amplifying a signal—that is, taking a small-voltage signal <strong>and</strong> making<br />

it much larger. He named it the “audion” <strong>and</strong> was granted a U.S.<br />

patent in 1907.


840 / Vacuum tube<br />

In 1907-1908, the American Navy carried radios equipped with<br />

de Forest’s audion in its goodwill tour around the world. While useful<br />

as an amplifier of the weak radio signals, it was not useful at this<br />

point for the more powerful signals of the telephone. Other developments<br />

were made quickly as the importance of the emerging<br />

fields of radio <strong>and</strong> telephony were realized.<br />

Impact<br />

With many industrial laboratories working on vacuum tubes,<br />

improvements came quickly. For example, tantalum <strong>and</strong> tungsten<br />

filaments quickly replaced the early carbon filaments. In 1904, Arthur<br />

Wehnelt, a German inventor, discovered that if metals were<br />

coated with certain materials such as metal oxides, they emitted far<br />

more electrons at a given temperature. These materials enabled<br />

electrons to escape the surface of the metal oxides more easily. Thermionic<br />

emission <strong>and</strong>, therefore, tube efficiencies were greatly improved<br />

by this method.<br />

Another important improvement in the vacuum tube came with<br />

the work of the American chemist Irving Langmuir of the General<br />

Electric Research Laboratory, starting in 1909, <strong>and</strong> Harold D. Arnold<br />

of Bell Telephone Laboratories. They used new devices such as<br />

the mercury diffusion pump to achieve higher vacuums. Working<br />

independently, Langmuir <strong>and</strong> Arnold discovered that very high<br />

vacuum used with higher voltages increased the power these tubes<br />

could h<strong>and</strong>le from small fractions of a watt to hundreds of watts.<br />

The de Forest tube was now useful for the higher-power audio signals<br />

of the telephone. This resulted in the introduction of the first<br />

transamerican speech transmission in 1914, followed by the first<br />

transatlantic communication in 1915.<br />

The invention of the transistor in 1948 by the American physicists<br />

William Shockley, Walter H. Brattain, <strong>and</strong> John Bardeen ultimately<br />

led to the downfall of the tube. With the exception of the cathode-ray<br />

tube, transistors could accomplish the jobs of nearly all vacuum tubes<br />

much more efficiently. Also, the development of the integrated circuit<br />

allowed the creation of small, efficient, highly complex devices that<br />

would be impossible with radio tubes. By 1977, the major producers<br />

of the vacuum tube had stopped making it.


See also Color television; FM radio; Radar; Radio; Radio crystal<br />

sets; Television; Transistor; Transistor radio.<br />

Further Reading<br />

Vacuum tube / 841<br />

Baldwin, Neil. Edison: Inventing the Century. Chicago: University of<br />

Chicago Press, 2001.<br />

Fleming, John Ambrose. Memories of a Scientific Life. London: Marshall,<br />

Morgan & Scott, 1934.<br />

Hijiya, James A. Lee de Forest <strong>and</strong> the Fatherhood of Radio. Bethlehem,<br />

Pa.: Lehigh University Press, 1992.<br />

Read, Oliver, <strong>and</strong> Walter L. Welch. From Tin Foil to Stereo: Evolution of<br />

the Phonograph. 2d ed. Indianapolis: H. W. Sams, 1976.


842<br />

Vat dye<br />

Vat dye<br />

The invention: The culmination of centuries of efforts to mimic the<br />

brilliant colors displayed in nature in dyes that can be used in<br />

many products.<br />

The people behind the invention:<br />

Sir William Henry Perkin (1838-1907), an English student in<br />

Hofmann’s laboratory<br />

René Bohn (1862-1922), a synthetic organic chemist<br />

Karl Heumann (1850-1894), a German chemist who taught Bohn<br />

Rol<strong>and</strong> Scholl (1865-1945), a Swiss chemist who established the<br />

correct structure of Bohn’s dye<br />

August Wilhelm von Hofmann (1818-1892), an organic chemist<br />

Synthesizing the Compounds of Life<br />

From prehistoric times until the mid-nineteenth century, all dyes<br />

were derived from natural sources, primarily plants. Among the<br />

most lasting of these dyes were the red <strong>and</strong> blue dyes derived from<br />

alizarin <strong>and</strong> indigo.<br />

The process of making dyes took a great leap forward with the<br />

advent of modern organic chemistry in the early years of the nineteenth<br />

century. At the outset, this branch of chemistry, dealing with<br />

the compounds of the element carbon <strong>and</strong> associated with living<br />

matter, hardly existed, <strong>and</strong> synthesis of carbon compounds was not<br />

attempted. Considerable data had accumulated showing that organic,<br />

or living, matter was basically different from the compounds<br />

of the nonliving mineral world. It was widely believed that although<br />

one could work with various types of organic matter in<br />

physical ways <strong>and</strong> even analyze their composition, they could be<br />

produced only in a living organism.<br />

Yet, in 1828, the German chemist Friedrich Wöhler found that it<br />

was possible to synthesize the organic compound urea from mineral<br />

compounds. As more chemists reported the successful preparation<br />

of compounds previously isolated only from plants or animals,<br />

the theory that organic compounds could be produced only in a living<br />

organism faded.


One field ripe for exploration was that committed to exploiting the<br />

uses of coal tar. Here, August Wilhelm von Hofmann was an active<br />

worker. He <strong>and</strong> his students made careful studies of this complex<br />

mixture. The high-quality stills they designed allowed for the isolation<br />

of pure samples of important compounds for further study.<br />

Of greater importance was the collection of able students Hofmann<br />

attracted. Among them was Sir William Henry Perkin, who is regarded<br />

as the founder of the dyestuffs industry. In 1856, Perkin undertook<br />

the task of synthesizing quinine (a bitter crystalline alkaloid<br />

used in medicine) from a nitrogen-containing coal tar material<br />

called toluidine. Luck played a decisive role in the outcome of his<br />

experiment. The sticky compound Perkin obtained contained no<br />

quinine, so he decided to investigate the simpler related compound<br />

aniline. A small amount of the impurity toluidine in his aniline gave<br />

Perkin the first synthetic dye, Mauveine.<br />

Searching for Structure<br />

Vat dye / 843<br />

From this beginning, the great dye industries of Europe, particularly<br />

Germany, grew. The trial-<strong>and</strong>-error methods gave way to more<br />

systematic searches as the structural theory of organic chemistry<br />

was formulated.<br />

As the twentieth century began, great progress had been made,<br />

<strong>and</strong> German firms dominated the industry. Badische Anilin- und<br />

Soda-Fabrik (BASF) was incorporated at Ludwigshafen in 1865 <strong>and</strong><br />

undertook extensive explorations of both alizarin <strong>and</strong> indigo. A<br />

chemist, René Bohn, had made important discoveries in 1888, which<br />

helped the company recover lost ground in the alizarin field. In<br />

1901, he undertook the synthesis of a dye he hoped would combine<br />

the desirable attributes of both alizarin <strong>and</strong> indigo.<br />

As so often happens in science, nothing like the expected occurred.<br />

Bohn realized that the beautiful blue crystals that resulted<br />

from his synthesis represented a far more important product. Not<br />

only was this the first synthetic vat dye, Indanthrene, ever prepared,<br />

but also, by studying the reaction at higher temperature, a useful<br />

yellow dye, Flavanthrone, could be produced.<br />

The term vat dye is used to describe a method of applying the dye,<br />

but it also serves to characterize the structure of the dye, because all


844 / Vat dye<br />

William Henry Perkin<br />

Born in Engl<strong>and</strong> in 1838, William Henry Perkin saw a chemical<br />

experiment for the first time when he was a small boy. He<br />

found his calling there <strong>and</strong> then, much to the dismay of his father,<br />

who wanted him to be a builder <strong>and</strong> architect like himself.<br />

Perkin studied chemistry every chance he found as a teenager<br />

<strong>and</strong> was only seventeen when he won an appointment as<br />

the assistant to the German chemist August Wilhelm von Hofmann.<br />

A year later, while trying to synthesize quinine at Hofmann’s<br />

suggestion, Perkin discovered a deep purple dye—now<br />

known as aniline purple or Mauveine, but popularly called<br />

mauve. In 1857 he opened a small dyeworks by the Gr<strong>and</strong><br />

Union Canal in West London, hoping to make his fortune by<br />

manufacturing the dye.<br />

He succeeded brilliantly. His ambitions were helped along<br />

royally when Queen Victoria wore a silk gown dyed with Mauveine<br />

to the Royal Exhibition of 1862. In 1869, he perfected a<br />

method for producing another new dye, alizarin, which is red.<br />

A wealthy man, he sold his business in 1874 when he was just<br />

thirty-six years old <strong>and</strong> devoted himself to research, which included<br />

isolation of the first synthetic perfume, coumarin, from<br />

coal tar.<br />

Perkin died in 1907, a year after receiving a knighthood, one<br />

of his many awards <strong>and</strong> honors for starting the artificial dye industry.<br />

His son William Henry Perkin, Jr. (1860-1927) also became<br />

a well-known researcher in organic chemistry.<br />

currently useful vat dyes share a common unit. One fundamental<br />

problem in dyeing relates to the extent to which the dye is watersoluble.<br />

A beautifully colored molecule that is easily soluble in water<br />

might seem attractive given the ease with which it binds with the fiber;<br />

however, this same solubility will lead to the dye’s rapid loss in<br />

daily use.<br />

Vat dyes are designed to solve this problem by producing molecules<br />

that can be made water-soluble, but only during the dyeing or<br />

vatting process. This involves altering the chemical structure of the<br />

dye so that it retains its color throughout the life of the cloth.<br />

By 1907, Rol<strong>and</strong> Scholl had showed unambiguously that the


chemical structure proposed by Bohn for Indanthrene was correct,<br />

<strong>and</strong> a major new area of theoretical <strong>and</strong> practical importance was<br />

opened for organic chemists.<br />

Impact<br />

Bohn’s discovery led to the development of many new <strong>and</strong> useful<br />

dyes. The list of patents issued in his name fills several pages in<br />

Chemical Abstracts indexes.<br />

The true importance of this work is to be found in a consideration<br />

of all synthetic chemistry, which may perhaps be represented by<br />

this particular event. More than two hundred dyes related to Indanthrene<br />

are in commercial use. The colors represented by these substances<br />

are a rainbow making nature’s finest hues available to all.<br />

The dozen or so natural dyes have been synthesized into more than<br />

seven thous<strong>and</strong> superior products through the creativity of the<br />

chemist.<br />

Despite these desirable outcomes, there is doubt whether there is<br />

any real benefit to society from the development of new dyes. This<br />

doubt is the result of having to deal with limited natural resources.<br />

With so many urgent problems to be solved, scientists are not sure<br />

whether to search for greater luxury. If the field of dye synthesis reveals<br />

a single theme, however, it must be to expect the unexpected.<br />

Time after time, the search for one goal has led to something quite<br />

different—<strong>and</strong> useful.<br />

See also Buna rubber; Color film; Neoprene.<br />

Further Reading<br />

Vat dye / 845<br />

Clark, Robin J. H., et al. “Indigo, Woad, <strong>and</strong> Tyrian Purple: Important<br />

Vat Dyes from Antiquity to the Present.” Endeavour 17, no. 4<br />

(December, 1993).<br />

Farber, Eduard. The Evolution of Chemistry: A History of Its Ideas,<br />

Methods, <strong>and</strong> Materials. 2d ed. New York: Ronald Press, 1969.<br />

Partington, J. R. A History of Chemistry. Staten Isl<strong>and</strong>, N.Y.: Martino,<br />

1996.<br />

Schatz, Paul F. “Anniversaries: 2001.” Journal of Chemical Education<br />

78, no. 1 (January, 2001).


846<br />

Velcro<br />

Velcro<br />

The invention: A material comprising millions of tiny hooks <strong>and</strong><br />

loops that work together to create powerful <strong>and</strong> easy-to-use fasteners<br />

for a wide range of applications.<br />

The person behind the invention:<br />

Georges de Mestral (1904-1990), a Swiss engineer <strong>and</strong> inventor<br />

From Cockleburs to Fasteners<br />

Since prehistoric times, people have walked through weedy fields<br />

<strong>and</strong> arrived at home with cockleburs all over their clothing. In 1948, a<br />

Swiss engineer <strong>and</strong> inventor, Georges de Mestral, found his clothing<br />

full of cockleburs after walking in the Swiss Alps near Geneva. Wondering<br />

why cockleburs stuck to clothing, he began to examine them<br />

under a microscope. De Mestral’s initial examination showed that<br />

each of the thous<strong>and</strong>s of fibrous ends of the cockleburs was tipped<br />

with a tiny hook; it was the hooks that made the cockleburs stick to<br />

fabric. This observation, combined with much subsequent work, led<br />

de Mestral to invent velcro, which was patented in 1957 in the form of<br />

two strips of nylon material. One of the strips contained millions of<br />

tiny hooks, while the other contained a similar number of tiny loops.<br />

When the two strips were pushed together, the hooks were inserted<br />

into the loops, joining the two strips of nylon very firmly. This design<br />

makes velcro extremely useful as a material for fasteners that is used<br />

in applications ranging from sneaker fasteners to fasteners used to<br />

join heart valves during surgery.<br />

Making Velcro Practical<br />

Velcro is not the only invention credited to de Mestral, who also<br />

invented such items as a toy airplane <strong>and</strong> an asparagus peeler, but it<br />

was his greatest achievement. It is said that his idea for the material<br />

was partly the result of a problem his wife had with a jammed dress<br />

zipper just before an important social engagement. De Mestral’s<br />

idea was to design a sort of locking tape that used the hook-<strong>and</strong>loop<br />

principle that he had observed under the microscope. Such a


tape, he believed, would never jam. He also believed that the tape<br />

would do away with such annoyances as buttons that popped open<br />

unexpectedly <strong>and</strong> knots in shoelaces that refused to be untied.<br />

The design of the material envisioned by de Mestral took seven<br />

years of painstaking effort. When it was finished, de Mestral named<br />

it “velcro” (a contraction of the French phrase velvet crochet, meaning<br />

velvet hook), patented it, <strong>and</strong> opened a factory to manufacture<br />

it. Velcro’s design required that de Mestral identify the optimal<br />

number of hooks <strong>and</strong> loops to be used. He eventually found that using<br />

approximately three hundred per square inch worked best. In<br />

addition, his studies showed that nylon was an excellent material<br />

for his purposes, although it had to be stiffened somewhat to work<br />

well. Much additional experimentation showed that the most effective<br />

way of producing the necessary stiffening was to subject the<br />

velcro to infrared light after manufacturing it.<br />

Other researchers have demonstrated that velcrolike materials<br />

need not be made of nylon. For example, a new micromechanical<br />

velcrolike material (microvelcro) that medical researchers believe<br />

will soon be used to hold together blood vessels after surgery is<br />

made of minute silicon loops <strong>and</strong> hooks. This material is thought to<br />

be superior to other materials for such applications because it will<br />

not be redissolved prematurely by the body. Other uses for microvelcro<br />

may be to hold together tiny electronic components in miniaturized<br />

computers without the use of glue or other adhesives. A major<br />

advantage of the use of microvelcro in such situations is that it is<br />

resistant to changes of temperature as well as to most chemicals that<br />

destroy glue <strong>and</strong> other adhesives.<br />

Impact<br />

Velcro / 847<br />

In 1957, when velcro was patented, there were four main ways to<br />

hold things together. These involved the use of buttons, laces, snaps,<br />

<strong>and</strong> zippers (which had been invented by Chicagoan Whitcomb L.<br />

Judson in 1892). All these devices had drawbacks; zippers can jam,<br />

buttons can come open at embarrassing times, <strong>and</strong> shoelaces can<br />

form knots that are difficult to unfasten. Almost immediately after<br />

velcro was introduced, its use became widespread; velcro fasteners<br />

can be found on or in clothing, shoes, watchb<strong>and</strong>s, wallets, back-


848 / Velcro<br />

packs, bookbags, motor vehicles, space suits, blood-pressure cuffs,<br />

<strong>and</strong> in many other places. There is even a “wall jumping” game incorporating<br />

velcro in which a wall is covered with a well-supported<br />

piece of velcro. People who want to play put on jackets made of<br />

velcro <strong>and</strong> jump as high as they can. Wherever they l<strong>and</strong> on the wall,<br />

the velcro will join together, making them stick.<br />

Wall jumping, silly though it may be, demonstrates the tremendous<br />

holding power of velcro; a velcro jacket can keep a twohundred-pound<br />

person suspended from a wall. This great strength is<br />

used in a more serious way in the design of the items used to anchor<br />

astronauts to space shuttles <strong>and</strong> to buckle on parachutes. In addition,<br />

velcro is washable, comes in many colors, <strong>and</strong> will not jam. No<br />

doubt many more uses for this innovative product will be found.<br />

See also Artificial heart.<br />

Georges de Mestral<br />

Georges de Mestral got his idea for Velcro in part during a<br />

hunting trip on his estates <strong>and</strong> in part before an important formal<br />

social function. These contexts are evidence of the high<br />

st<strong>and</strong>ing in Swiss society held by de Mestral, an engineer <strong>and</strong><br />

manufacturer. In fact, de Mestral, who was born in 1904, came<br />

from a illustrious line of noble l<strong>and</strong>owners. Their prize possession<br />

was one of Switzerl<strong>and</strong>’s famous residences, the castle of<br />

Saint Saphorin on Morges.<br />

Built on the site of yet older fortifications, the castle was<br />

completed by François-Louis de Pesme in 1710. An enemy of<br />

King Louis XIV, de Pesme served in the military forces of Austria,<br />

Holl<strong>and</strong>, <strong>and</strong> Engl<strong>and</strong>, rising to the rank of lieutenant general,<br />

but he is best known for driving off a Turkish invasion fleet<br />

on the Danube in 1695. Other forebears include the diplomat<br />

Arm<strong>and</strong>- François Louis de Mestral (1738-1805) <strong>and</strong> his father,<br />

Albert-Georges-Constantin de Mestral (1878-1966), an agricultural<br />

engineer.<br />

The castle passed to the father’s four sons <strong>and</strong> eventually<br />

into the care of the inventor. It in turn was inherited by Georges<br />

de Mestral’s sons Henri <strong>and</strong> François when he died in 1990 in<br />

Genolier, Switzerl<strong>and</strong>.


Further Reading<br />

Velcro / 849<br />

“George De Mestral: Inventor of Velcro Fastener.” Los Angeles Times<br />

(February 13, 1990).<br />

LaFavre Yorks, Cindy. “Hidden Helpers Velcro Fasteners, Pull-On<br />

Loops <strong>and</strong> Other Extras Make Dressing Easier for People with<br />

Disabilities.” Los Angeles Times (November 1, 1991).<br />

Roberts, Royston M., <strong>and</strong> Jeanie Roberts. Lucky Science: Accidental<br />

Discoveries from Gravity to Velcro, with Experiments. New York:<br />

John Wiley, 1994.<br />

Stone, Judith. “Stuck on Velcro!” Reader’s Digest (September, 1988).<br />

“Velcro-wrapped Armor Saves Lives in Bosnia.” Design News 52, no.<br />

7 (April 7, 1997).


850<br />

Vending machine slug rejector<br />

Vending machine slug rejector<br />

The invention: A device that separates real coins from counterfeits,<br />

the slug rejector made it possible for coin-operated vending<br />

machines to become an important marketing tool for many<br />

products<br />

The people behind the invention:<br />

Thomas Adams, the founder of Adams Gum Company<br />

Frederick C. Lynde, an Englishman awarded the first American<br />

patent on a vending machine<br />

Nathaniel Leverone (1884-1969), a founder of the Automatic<br />

Canteen Company of America<br />

Louis E. Leverone (1880-1957), a founder, with his brother, of the<br />

Automatic Canteen Company of America<br />

The Growth of Vending Machines<br />

One of the most imposing phenomena to occur in the United<br />

States economy following World War II was the growth of vending<br />

machines. Following the 1930’s invention <strong>and</strong> perfection of the slug<br />

rejector, vending machines became commonplace as a means of<br />

marketing gum <strong>and</strong> c<strong>and</strong>y. By the 1960’s, almost every building had<br />

soft drink <strong>and</strong> coffee machines. Street corners featured machines<br />

that dispensed newspapers, <strong>and</strong> post offices even used vending machines<br />

to sell stamps. Occasionally someone fishing in the backwoods<br />

could find a vending machine next to a favorite fishing hole<br />

that would dispense a can of fishing worms upon deposit of the correct<br />

amount of money. The primary advantage offered by vending<br />

machines is their convenience. Unlike people, machines can provide<br />

goods <strong>and</strong> services around the clock, with no charge for the “labor”<br />

of st<strong>and</strong>ing duty.<br />

The decade of the 1950’s brought not only an increase in the number<br />

of vending machines but also an increase in the types of goods<br />

that were marketed through them. Before World War II, the major<br />

products had been cigarettes, c<strong>and</strong>y, gum, <strong>and</strong> soft drinks. The<br />

1950’s brought far more products into the vending machine market.


Vending machine slug rejector / 851<br />

The first recognized vending machine in history was invented in<br />

the third century b.c.e. by the mathematician Hero. This first machine<br />

was a coin-activated device that dispensed sacrificial water in<br />

an Egyptian temple. It was not until the year 1615 that another<br />

vending machine was recorded. In that year, snuff <strong>and</strong> tobacco<br />

vending boxes began appearing in English pubs <strong>and</strong> taverns. These<br />

tobacco boxes were less sophisticated machines than was Hero’s,<br />

since they left much to the honesty of the customer. Insertion of a<br />

coin opened the box; once it was open, the customer could take out<br />

as much tobacco as desired. One of the first United States patents on<br />

a machine was issued in 1886 to Frederick C. Lynde. That machine<br />

was used to vend postcards.<br />

If any one person can be considered the father of vending machines<br />

in the United States, it would probably be Thomas Adams,<br />

the founder of Adams Gum Company. Adams began the first successful<br />

vending operation in America in 1888 when he placed gum<br />

machines on train platforms in New York City.<br />

Other early vending machines included scales (which vended a<br />

service rather than a product), photograph machines, strength testers,<br />

beer machines, <strong>and</strong> hot water vendors (to supply poor people<br />

who had no other source of hot water). These were followed, around<br />

1900, by complete automatic restaurants in Germany, cigar vending<br />

machines in Chicago, perfume machines in Paris, <strong>and</strong> an automatic<br />

divorce machine in Utah.<br />

Also around 1900 came the introduction of coin-operated gambling<br />

machines. These “slot machines” are differentiated from normal<br />

vending machines. The vending machine industry does not<br />

consider gambling machines to be a part of the vending industry<br />

since they do not vend merch<strong>and</strong>ise. The primary importance of the<br />

gambling machines was that they induced the industry to do research<br />

into slug rejection. Early machines allowed coins to be retrieved<br />

by the use of strings tied to them <strong>and</strong> accepted counterfeit<br />

lead coins, called slugs. It was not until the 1930’s that the slug<br />

rejector was perfected. Invention of the slug rejection device gave<br />

rise to the tremendous growth in the vending machine industry in<br />

the 1930’s by giving vendors more confidence that they would be<br />

paid for their products or services.<br />

Soft drink machines got their start just prior to the beginning of


852 / Vending machine slug rejector<br />

the twentieth century. By 1906, improved models of these machines<br />

could dispense up to ten different flavors of soda pop. The<br />

drinks were dispensed into a drinking glass or tin cup that was<br />

placed near the machine (there was usually only one glass or cup<br />

to a machine, since paper cups had not been invented). <strong>Public</strong><br />

health officials became concerned that everyone was drinking<br />

from the same cup. At that point, someone came up with the idea<br />

of setting a bucket of water next to the machine so that each customer<br />

could rinse off the cup before drinking from it. The year 1909<br />

witnessed one of the monumental inventions in the history of<br />

vending machines, the pay toilet.<br />

Impact<br />

The 1930’s witnessed improved vending machines. Slug rejectors<br />

were the most important introduction. In addition, change-making<br />

machines were instituted, <strong>and</strong> a few machines would even say<br />

“thank you” after a coin was deposited. These improved machines<br />

led many marketers to experiment with automatic vending. Coinoperated<br />

washing machines were one of the new applications of the<br />

1930’s. During the Depression, many appliance dealers attached<br />

coin metering devices to washing machines, allowing the user to accumulate<br />

money to make the monthly payments by using the appliance.<br />

This was a form of forced saving. It was not long before some<br />

enterprising appliance dealer got the idea of placing washing machines<br />

in apartment house basements. This idea was soon followed<br />

by stores full of coin-operated laundry machines, giving rise to a<br />

new kind of automatic vending business.<br />

Following World War II, there was a surge of innovation in the<br />

vending machine industry. Much of that surge resulted from the<br />

discovery of vending machines by industrial management. Prior to<br />

the war, the managements of most factories had been tolerant of<br />

vending machines. Following the war, managers discovered that<br />

the machines could be an inexpensive means of keeping workers<br />

happy. They became aware that worker productivity could be increased<br />

by access to c<strong>and</strong>y bars or soft drinks. As a result, the dem<strong>and</strong><br />

for machines exceeded the supply offered by the industry<br />

during the late 1940’s.


Vending machines have had a surprising effect on the total retail<br />

sales of the U.S. economy. In 1946, sales through vending machines<br />

totaled $600 million. By 1960, that figure had increased to $2.5 billion;<br />

by 1970, it exceeded $6 billion. The decade of the 1950’s began<br />

with individual machines that would dispense cigarettes, c<strong>and</strong>y,<br />

gum, coffee, <strong>and</strong> soft drinks. By the end of that decade, it was much<br />

more common to see vending machines in groups. The combination<br />

of machines in a group could, in many cases, meet the requirements<br />

to assemble a complete meal.<br />

Convenience is the key to the popularity of vending machines.<br />

Their ability to sell around the clock has probably been the major<br />

impetus to vending machine sales as opposed to more conventional<br />

marketing. Lower labor costs have also played a role in their popularity,<br />

<strong>and</strong> their location in areas of dense pedestrian traffic prompts<br />

impulse purchases.<br />

Despite the advances made by the vending machine industry<br />

during the 1950’s, there was still one major limitation to growth, to<br />

be solved during the early 1960’s. That problem was that vending<br />

machines were effectively limited to low-priced items, since the machines<br />

would accept nothing but coins. The inconvenience of inserting<br />

many coins kept machine operators from trying to market expensive<br />

items; as they expected consumer reluctance. The early<br />

1960’s witnessed the invention of vending machines that would accept<br />

<strong>and</strong> make change for $1, $5, <strong>and</strong> $10 bills. This invention paved<br />

the way for expansion into lines of grocery items <strong>and</strong> tickets.<br />

The first use of vending machines to issue tickets was at an Illinois<br />

race track, where pari-mutuel tickets were dispensed upon deposit of<br />

$2. Penn Central Railroad was one of the first transportation companies<br />

to sell tickets by means of vending machines. These machines,<br />

used in high-traffic areas on the East Coast, permitted passengers to<br />

deal directly with a computer when buying reserved-seat train tickets.<br />

The machines would accept $1 bills <strong>and</strong> $5 bills as well as coins.<br />

Limitations to Vending Machines<br />

Vending machine slug rejector / 853<br />

There are limitations to the use of vending machines. Primary<br />

among these are mechanical failure <strong>and</strong> v<strong>and</strong>alism of machines.<br />

Another limitation often mentioned is that not every product can be


854 / Vending machine slug rejector<br />

sold by machine. There are several factors that make some goods<br />

more vendable than others. National advertising <strong>and</strong> wide consumer<br />

acceptance help. Product must have a high turnover in order<br />

to justify the cost of a machine <strong>and</strong> the cost of servicing it. A third<br />

factor in measuring the potential success of an item is where it will<br />

be consumed or used. The most successful products are used within<br />

a short distance of the machine; consumers must be made willing to<br />

pay the usually higher prices of machine-bought products by the<br />

convenience of machine location.<br />

The automatic vending of merch<strong>and</strong>ise plays the largest role in<br />

the vending machine industry, but the vending of services also<br />

plays a role. The largest percentage of service vending comes from<br />

coin laundries. Other types of services are vended by weighing machines,<br />

parcel lockers, <strong>and</strong> pay toilets. By depositing a coin, a person<br />

can even get shoes shined. Some motel beds offer a “massage.” Even<br />

the lowly parking meter is an example of a vending machine that<br />

dispenses services. Coin-operated photocopy machines account for<br />

a large portion of service vending.<br />

A later advance in the vending machine industry is the use of<br />

credit. The cashless society began to make strides with vending machines<br />

as well as conventional vendors. As of the early 1990’s, credit<br />

cards could be used to operate only a few types of vending machines,<br />

primarily those that dispense transportation tickets. Vending machines<br />

operated by banks dispense money upon deposit of a credit<br />

card. Credit-card gasoline pumps reduced labor requirements at gasoline<br />

stations, pushing the concept of self-service a step further. As<br />

credit card transactions become more common in general <strong>and</strong> as the<br />

cost of making them falls, use of credit cards for vending machines<br />

will increase.<br />

Thous<strong>and</strong>s of items have been marketed through vending machines,<br />

<strong>and</strong> firms must continue to evaluate the use of automatic retailing<br />

as a marketing channel. Many products are not conducive to<br />

automatic vending, but before dismissing that option for a particular<br />

product, a marketer should consider the range of products sold<br />

through vending machines. The producers of B<strong>and</strong>-Aid flexible plastic<br />

b<strong>and</strong>ages saw the possibilities in the vending field. The only product<br />

modification necessary was to put B<strong>and</strong>-Aids in a package the<br />

size of a c<strong>and</strong>y bar, able to be sold from renovated c<strong>and</strong>y machines.


The next problem was to determine areas where there would be a<br />

high turnover of B<strong>and</strong>-Aids. Bowling alleys were an obvious answer,<br />

since many bowlers suffered from abrasions on their fingers.<br />

The United States is not alone in the development of vending machines;<br />

in fact, it is not as advanced as some nations of the world. In<br />

Japan, machines operated by credit cards have been used widely<br />

since the mid-1960’s, <strong>and</strong> the range of products offered has been<br />

larger than in the United States. Western Europe is probably the<br />

most advanced area of the world in terms of vending machine technology.<br />

Germany of the early 1990’s probably had the largest selection<br />

of vending machines of any European country. Many gasoline<br />

stations in Germany featured beer dispensing machines. In rural areas<br />

of the country, vending machines hung from utility poles. These<br />

rural machines provided c<strong>and</strong>y <strong>and</strong> gum, among other products, to<br />

farmers who did not often travel into town.<br />

Most vending machine business in Europe was done not in individual<br />

machines but in automated vending shops. The machines offered<br />

a creative solution to obstacles created by regulations <strong>and</strong><br />

laws. Some countries had laws stating that conventional retail stores<br />

could not be open at night or on Sundays. To increase sales <strong>and</strong> satisfy<br />

consumer needs, stores built vending operations that could be<br />

used by customers during off hours. The machines, or combinations<br />

of them, often stocked a tremendous variety of items. At one German<br />

location, consumers could choose among nearly a thous<strong>and</strong><br />

grocery items.<br />

The Future<br />

Vending machine slug rejector / 855<br />

The future will see a broadening of product lines offered in vending<br />

machines as marketers come to recognize the opportunities that<br />

exist in automatic retailing. In the United States, vending machines<br />

of the early 1990’s primarily dispensed products for immediate consumption.<br />

If labor costs increase, it will become economically feasible<br />

to sell more items from vending machines. Grocery items <strong>and</strong><br />

tickets offered the most potential for expansion.<br />

Vending machines offer convenience to the consumer. Virtually<br />

any company that produces for the retail market must consider<br />

vending machines as a marketing channel. Machines offer an alter-


856 / Vending machine slug rejector<br />

native to conventional stores that cannot be ignored as the range of<br />

products offered through machines increases.<br />

Vending machines appear to be a permanent fixture <strong>and</strong> have<br />

only scratched the surface of the market. Although machines have a<br />

long history, their popularization came from innovations of the<br />

1930’s, particularly the slug rejector. Marketing managers came to<br />

recognize that vending machine sales are more than a sideline. Increasingly,<br />

firms established separate departments to h<strong>and</strong>le sales<br />

through vending machines. Successful companies make the best<br />

use of all channels of distribution, <strong>and</strong> vending machines had become<br />

an important marketing channel.<br />

See also Geiger counter; Sonar; Radio interferometer.<br />

Further Reading<br />

Ho, Rodney. “Vending Machines Make Change—-Now They Sell<br />

Movie Soundtracks, Underwear—Even Art.” Wall Street Journal<br />

(July 7, 1999).<br />

Rosen, Cheryl. “Vending Machines Get a High-Tech Makeover.<br />

Informationweek 822 (January 29, 2001).<br />

Ryan, James. “In Vending Machine, Brains That Tell Good Money<br />

from Bad.” New York Times (April 8, 1999).<br />

Tagliabue, John. “Vending Machines Face an Upheaval of Change.”<br />

New York Times (February 16, 1999).


Videocassette recorder<br />

Videocassette recorder<br />

The invention: A device for recording <strong>and</strong> playing back movies<br />

<strong>and</strong> television programs, the videocassette recorder (VCR) revolutionized<br />

the home entertainment industry in the late 1970’s.<br />

The company behind the invention:<br />

Philips Corporation, a Dutch Company<br />

Videotape Recording<br />

857<br />

Although television sets first came on the market before World<br />

War II, video recording on magnetic tape was not developed until<br />

the 1950’s. Ampex marketed the first practical videotape recorder<br />

in 1956. Unlike television, which manufacturers aimed at retail<br />

consumers from its inception, videotape recording was never expected<br />

to be attractive to the individual consumer. The first videotape<br />

recorders were meant for use within the television industry.<br />

Developed not long after the invention of magnetic tape recording<br />

of audio signals, the early videotape recorders were large machines<br />

that employed an open reel-to-reel tape drive similar to that<br />

of a conventional audiotape recorder. Recording <strong>and</strong> playback heads<br />

scanned the tape longitudinally (lengthwise). Because video signals<br />

have a much wider frequency (“frequency” is the distance between<br />

the tops <strong>and</strong> the bottoms of the signal waves) than audio signals do,<br />

this scanning technique meant that the amount of recording time<br />

available on one reel of tape was extremely limited. In addition,<br />

open reels were large <strong>and</strong> awkward, <strong>and</strong> the magnetic tape itself<br />

was quite expensive.<br />

Still, within the limited application area of commercial television,<br />

videotape recording had its uses. It made it possible to play<br />

back recorded material immediately rather than having to wait for<br />

film to be processed in a laboratory. As television became more popular<br />

<strong>and</strong> production schedules became more hectic, with more material<br />

being produced in shorter <strong>and</strong> shorter periods of time, videotape<br />

solved some significant problems.


858 / Videocassette recorder<br />

Helical Scanning Breakthrough<br />

Engineers in the television industry continued to search for innovations<br />

<strong>and</strong> improvements in videotape recording following<br />

Ampex’s marketing of the first practical videotape recorder in the<br />

1950’s. It took more than ten years, however, for the next major<br />

breakthrough to occur. The innovation that proved to be the key to<br />

reducing the size <strong>and</strong> awkwardness of video recording equipment<br />

came in 1967 with the invention by the Philips Corporation of helical<br />

scanning.<br />

All videocassette recorders eventually employed multiple-head<br />

helical scanning systems. In a helical scanning system, the record<br />

<strong>and</strong> playback heads are attached to a spinning drum or head that rotates<br />

at exactly 1,800 revolutions per minute, or 30 revolutions per<br />

second. This is the number of video frames per second used in the<br />

NTSC-TV broadcasts in the United States <strong>and</strong> Canada. The heads<br />

are mounted in pairs 180 degrees apart on the drum. Two fields on<br />

the tape are scanned for each revolution of the drum. Perhaps the<br />

easiest way to underst<strong>and</strong> the helical scanning system is to visualize<br />

the spiral path followed by the stripes on a barber’s pole.<br />

Helical scanning deviated sharply from designs based on audio<br />

recording systems. In an audiotape recorder, the tape passes over<br />

stationary playback <strong>and</strong> record heads; in a videocassette recorder,<br />

both the heads <strong>and</strong> the tape move. Helical scanning is, however, one<br />

of the few things that competing models <strong>and</strong> formats of videocassette<br />

recorders have in common. Different models employ different<br />

tape delivery systems <strong>and</strong>, in the case of competing formats such as<br />

Beta <strong>and</strong> VHS, there may be differences in the composition of the<br />

video signal to be recorded. Beta uses a 688-kilohertz (kHz) frequency,<br />

while VHS employs a frequency of 629 kHz. This difference<br />

in frequency is what allows Beta videocassette recorders (VCRs) to<br />

provide more lines of resolution <strong>and</strong> thus a superior picture quality;<br />

VHS provides 240 lines of resolution, while Beta has 400. (For this<br />

reason, it is perhaps unfortunate that the VHS format eventually<br />

dominated the market.)<br />

In addition to helical scanning, Philips introduced another innovation:<br />

the videocassette. Existing videotape recorders employed a<br />

reel-to-reel tape drive, as do videocassettes, but videocassettes en-


close the tape reels in a protective case. The case prevents the tape<br />

from being damaged in h<strong>and</strong>ling.<br />

The first VCRs were large <strong>and</strong> awkward compared to later models.<br />

Industry analysts still thought that the commercial television<br />

<strong>and</strong> film industries would be the primary markets for VCRs. The<br />

first videocassettes employed wide— 3 4-inch or 1-inch—videotapes,<br />

<strong>and</strong> the machines themselves were cumbersome. Although Philips<br />

introduced a VCR in 1970, it took until 1972 before the machines actually<br />

became available for purchase, <strong>and</strong> it would be another ten<br />

years before VCRs became common appliances in homes.<br />

Consequences<br />

Videocassette recorder / 859<br />

Following the introduction of the VCR in 1970, the home entertainment<br />

industry changed radically. Although the industry did not<br />

originally anticipate that the VCR would have great commercial potential<br />

as a home entertainment device, it quickly became obvious<br />

that it did. By the late 1970’s, the size of the cassette had been reduced<br />

<strong>and</strong> the length of recording time available per cassette had<br />

been increased from one hour to six. VCRs became so widespread<br />

that advertisers on television became concerned with a phenomenon<br />

known as “timeshifting,” which refers to viewers setting the<br />

VCR to record programs for later viewing. Jokes about the complexity<br />

of programming VCRs appeared in the popular culture, <strong>and</strong> an<br />

inability to cope with the VCR came to be seen as evidence of technological<br />

illiteracy.<br />

Consumer dem<strong>and</strong> for VCRs was so great that, by the late 1980’s,<br />

compact portable video cameras became widely available. The same<br />

technology—helical scanning with multiple heads—was successfully<br />

miniaturized, <strong>and</strong> “camcorders” were developed that were not much<br />

larger than a paperback book. By the early 1990’s, “reality television”—that<br />

is, television shows based on actual events—began relying<br />

on video recordings supplied by viewers rather than material<br />

produced by professionals. The video recorder had completed a circle:<br />

It began as a tool intended for use in the television studio, <strong>and</strong> it<br />

returned there four decades later. Along the way, it had an effect no<br />

one could have predicted; passive viewers in the audience had evolved<br />

into active participants in the production process.


860 / Videocassette recorder<br />

See also Cassette recording; Color television; Compact disc;<br />

Dolby noise reduction; Television; Walkman cassette player.<br />

Further Reading<br />

Gilder, George. Life After Television. New York: W. W. Norton, 1992.<br />

Lardner, James. Fast Forward: Hollywood, the Japanese, <strong>and</strong> the Onslaught<br />

of the VCR. New York: Norton, 1987.<br />

Luther, Arch C. Digital Video in the PC Environment. New York:<br />

McGraw-Hill, 1989.<br />

Wassser, Frederick. Veni, Vidi, Video: The Hollywood Empire <strong>and</strong> the<br />

VCR. Austin: University of Texas Press, 2001.


Virtual machine<br />

Virtual machine<br />

The invention: The first computer to swap storage space between<br />

its r<strong>and</strong>om access memory (RAM) <strong>and</strong> hard disk to create a<br />

larger “virtual” memory that enabled it to increase its power.<br />

The people behind the invention:<br />

International Business Machines (IBM) Corporation, an<br />

American data processing firm<br />

Massachusetts Institute of Technology (MIT), an American<br />

university<br />

Bell Labs, the research <strong>and</strong> development arm of the American<br />

Telephone <strong>and</strong> Telegraph Company<br />

A Shortage of Memory<br />

861<br />

During the late 1950’s <strong>and</strong> the 1960’s, computers generally used<br />

two types of data storage areas. The first type, called “magnetic<br />

disk storage,” was slow <strong>and</strong> large, but its storage space was relatively<br />

cheap <strong>and</strong> abundant. The second type, called “main memory”<br />

(also often called “r<strong>and</strong>om access memory,” or RAM), was<br />

much faster. Computation <strong>and</strong> program execution occurred primarily<br />

in the “central processing unit” (CPU), which is the “brain”<br />

of the computer. The CPU accessed RAM as an area in which to<br />

perform intermediate computations, store data, <strong>and</strong> store program<br />

instructions.<br />

To run programs, users went through a lengthy process. At that<br />

time, keyboards with monitors that allowed on-line editing <strong>and</strong><br />

program storage were very rare. Instead, most users used typewriter-like<br />

devices to type their programs or text on paper cards.<br />

Holding decks of such cards, users waited in lines to use card readers.<br />

The cards were read <strong>and</strong> returned to the user, <strong>and</strong> the programs<br />

were scheduled to run later. Hours later or even overnight,<br />

the output of each program was printed in some predetermined<br />

order, after which all the outputs were placed in user bins. It might<br />

take as long as several days to make any program corrections that<br />

were necessary.


862 / Virtual machine<br />

Because CPUs were expensive, many users had to share a single<br />

CPU. If a computer had a monitor that could be used for editing or<br />

could run more than one program at a time, more memory was required.<br />

RAM was extremely expensive, <strong>and</strong> even multimilliondollar<br />

computers had small memories. In addition, this primitive<br />

RAM was extremely bulky.<br />

Virtually Unlimited Memory<br />

The solution to the problem of creating affordable, convenient<br />

memory came in a revolutionary reformulation of the relationship<br />

between main memory <strong>and</strong> disk space. Since disk space was large<br />

<strong>and</strong> cheap, it could be treated as an extended “scratch pad,” or temporary-use<br />

area, for main memory. While a program ran, only small<br />

parts of it (called pages or segments), normally the parts in use at<br />

that moment, would be kept in the main memory. If only a few<br />

pages of each program were kept in memory at any time, more programs<br />

could coexist in memory. When pages lay idle, they would be<br />

sent from RAM to the disk, as newly requested pages were loaded<br />

from the disk to the RAM. Each user <strong>and</strong> program “thought” it had<br />

essentially unlimited memory (limited only by disk space), hence<br />

the term “virtual memory.”<br />

The system did, however, have its drawbacks. The swapping <strong>and</strong><br />

paging processes reduced the speed at which the computer could<br />

process information. Coordinating these activities also required<br />

more circuitry. Integrating each program <strong>and</strong> the amount of virtual<br />

memory space it required was critical. To keep the system operating<br />

accurately, stably, <strong>and</strong> fairly among users, all computers have an<br />

“operating system.” Operating systems that support virtual memory<br />

are more complex than the older varieties are.<br />

Many years of research, design, simulations, <strong>and</strong> prototype testing<br />

were required to develop virtual memory. CPUs <strong>and</strong> operating<br />

systems were designed by large teams, not individuals. Therefore,<br />

the exact original discovery of virtual memory is difficult to trace.<br />

Many people contributed at each stage.<br />

The first rudimentary implementation of virtual memory concepts<br />

was on the Atlas computer, which was constructed in the early<br />

1960’s in Engl<strong>and</strong>, at the University of Manchester. It coupled RAM


Virtual machine / 863<br />

with a device that read a magnetizable cylinder, or drum, which<br />

meant that it was a two-part storage system.<br />

In the late 1960’s, the Massachusetts Institute of Technology<br />

(MIT), Bell Telephone Labs, <strong>and</strong> the General Electric Company<br />

(later Honeywell) jointly designed a high-level operating system<br />

called MULTICS, which had virtual memory.<br />

During the 1960’s, IBM worked on virtual memory, <strong>and</strong> the IBM<br />

360 series supported the new memory system. With the evolution<br />

of engineering concepts such as circuit integration, IBM produced<br />

a new line of computers called the IBM 370 series. The IBM 370<br />

supported several advances in hardware (equipment) <strong>and</strong> software<br />

(program instructions), including full virtual memory capabilities.<br />

It was a platform for a new <strong>and</strong> powerful “environment,”<br />

or set of conditions, in which software could be run; IBM called<br />

this environment the VM/370. The VM/370 went far beyond virtual<br />

memory, using virtual memory to create virtual machines. In a<br />

virtual machine environment, each user can select a separate <strong>and</strong><br />

complete operating system. This means that separate copies of operating<br />

systems such as OS/360, CMS, DOS/360, <strong>and</strong> UNIX can all<br />

run in separate “compartments” on a single computer. In effect,<br />

each operating system has its own machine. Reliability <strong>and</strong> security<br />

were also increased. This was a major breakthrough, a second<br />

computer revolution.<br />

Another measure of the significance of the IBM 370 was the commercial<br />

success <strong>and</strong> rapid, widespread distribution of the system.<br />

The large customer base for the older IBM 360 also appreciated the<br />

IBM 370’s compatibility with that machine. The essentials of the<br />

IBM 370 virtual memory model were retained even in the 1990’s<br />

generation of large, powerful mainframe computers. Furthermore,<br />

its success carried over to the design decisions of other computers in<br />

the 1970’s.<br />

The second-largest computer manufacturer, Digital Equipment<br />

Corporation (DEC), followed suit; its popular VAX minicomputers<br />

had virtual memory in the late 1970’s. The celebrated UNIX operating<br />

system also added virtual memory. IBM’s success had led to industry-wide<br />

acceptance.


864 / Virtual machine<br />

Consequences<br />

The impact of virtual memory extends beyond large computers<br />

<strong>and</strong> the 1970’s. During the late 1970’s <strong>and</strong> early 1980’s, the computer<br />

world took a giant step backward. Small, single-user computers<br />

called personal computers (PCs) became very popular. Because<br />

they were single-user models <strong>and</strong> were relatively cheap,<br />

they were sold with weak CPUs <strong>and</strong> deplorable operating systems<br />

that did not support virtual memory. Only one program could run<br />

at a time. Larger <strong>and</strong> more powerful programs required more<br />

memory than was physically installed. These computers crashed<br />

often.<br />

Virtual memory raises PC user productivity. With virtual memory<br />

space, during data transmissions or long calculations, users can<br />

simultaneously edit files if physical memory runs out. Most major<br />

PCs now have improved CPUs <strong>and</strong> operating systems, <strong>and</strong> these<br />

advances support virtual memory. Popular virtual memory systems<br />

such as OS/2, Windows/DOS, <strong>and</strong> MAC-OS are available. Even old<br />

virtual memory UNIX has been used in PCs.<br />

The concept of a virtual machine has been revived, in a weak<br />

form, on PCs that have dual operating systems (such as UNIX <strong>and</strong><br />

DOS, OS/2 <strong>and</strong> DOS, MAC <strong>and</strong> DOS, <strong>and</strong> UNIX <strong>and</strong> DOS combinations).<br />

Most powerful programs benefit from virtual memory. Many<br />

dazzling graphics programs require massive RAM but run safely in<br />

virtual memory. Scientific visualization, high-speed animation, <strong>and</strong><br />

virtual reality all benefit from it. Artificial intelligence <strong>and</strong> computer<br />

reasoning are also part of a “virtual” future.<br />

See also Colossus computer; Differential analyzer; ENIAC computer;<br />

IBM Model 1401 computer; Personal computer; Robot (industrial);<br />

SAINT; Virtual reality.<br />

Further Reading<br />

Bashe, Charles J. IBM’s Early Computers. Cambridge, Mass.: MIT<br />

Press, 1986.<br />

Ceruzzi, Paul E. A History of Modern Computing. Cambridge, Mass.:<br />

MIT Press, 2000.


Virtual machine / 865<br />

Chposky, James, <strong>and</strong> Ted Leonsis. Blue Magic: The People, Power, <strong>and</strong><br />

Politics Behind the IBM Personal Computer. New York: Facts on File,<br />

1988.<br />

Seitz, Frederick, <strong>and</strong> Norman G. Einspruch. Electronic Genie: The<br />

Tangled History of Silicon. Urbana: University of Illinois Press,<br />

1998.


866<br />

Virtual reality<br />

Virtual reality<br />

The invention: The creation of highly interactive, computer-based<br />

multimedia environments in which the user becomes a participant<br />

with the computer in a “virtually real” world.<br />

The people behind the invention:<br />

Ivan Sutherl<strong>and</strong> (1938- ), an American computer scientist<br />

Myron W. Krueger (1942- ), an American computer scientist<br />

Fred P. Brooks (1931- ), an American computer scientist<br />

Human/Computer Interface<br />

In the early 1960’s, the encounter between humans <strong>and</strong> computers<br />

was considered to be the central event of the time. The computer<br />

was evolving more rapidly than any technology in history; humans<br />

seemed not to be evolving at all. The “user interface” (the devices<br />

<strong>and</strong> language with which a person communicates with a computer)<br />

was a veneer that had been applied to the computer to make it<br />

slightly easier to use, but it seemed obvious that the ultimate interface<br />

would be connecting the human body <strong>and</strong> senses directly to the<br />

computer.<br />

Against this background, Ivan Sutherl<strong>and</strong> of the University of<br />

Utah identified the next logical step in the development of computer<br />

graphics. He implemented a head-mounted display that allowed<br />

a person to look around in a graphically created “room” simply<br />

by turning his or her head. Two small cathode-ray tubes, or<br />

CRTs (which are the basis of television screens <strong>and</strong> computer monitors),<br />

driven by vector graphics generators (mathematical imagecreating<br />

devices) provided the appropriate view for each eye, <strong>and</strong><br />

thus, stereo vision.<br />

In the early 1970’s, Fred P. Brooks of the University of North Carolina<br />

created a system that allowed a person to h<strong>and</strong>le graphic objects<br />

by using a mechanical manipulator. When the user moved the<br />

physical manipulator, a graphic manipulator moved accordingly. If<br />

a graphic block was picked up, the user felt its weight <strong>and</strong> its resistance<br />

to his or her fingers closing around it.


A New Reality<br />

Virtual reality / 867<br />

Beginning in 1969, Myron W. Krueger of the University of Wisconsin<br />

created a series of interactive environments that emphasized<br />

unencumbered, full-body, multisensory participation in computer<br />

events. In one demonstration, a sensory floor detected participants’<br />

movements around a room. A symbol representing each participant<br />

moved through a projected graphic maze that changed in playful<br />

ways if participants tried to cheat. In another demonstration, participants<br />

could use the image of a finger to draw on the projection<br />

screen. In yet another, participants’ views of a projected threedimensional<br />

room changed appropriately as they moved around<br />

the physical space.<br />

It was interesting that people naturally accepted these projected<br />

experiences as reality. They expected their bodies to influence graphic<br />

objects <strong>and</strong> were delighted when they did. They regarded their electronic<br />

images as extensions of themselves. What happened to their<br />

images also happened to them; they felt what touched their images.<br />

These observations led to the creation of the Videoplace, a graphic<br />

world that people could enter from different places to interact with<br />

each other <strong>and</strong> with graphic creatures. Videoplace is an installation<br />

at the Connecticut Museum of Natural History in Storrs, Connecticut.<br />

Videoplace visitors in separate rooms can fingerpaint together,<br />

perform free-fall gymnastics, tickle each other, <strong>and</strong> experience additional<br />

interactive events.<br />

The computer combines <strong>and</strong> alters inputs from separate cameras<br />

trained on each person, each of whom responds in turn to the computer’s<br />

output, playing games in the world created by Videoplace<br />

software. Since participants’ live video images can be manipulated<br />

(moved, scaled, or rotated) in real time, the world that is created is<br />

not bound by the laws of physics. In fact, the result is a virtual reality<br />

in which new laws of cause <strong>and</strong> effect are created, <strong>and</strong> can be<br />

changed, from moment to moment. Indeed, the term “virtual reality”<br />

describes the type of experience that can be created with Videoplace<br />

or with the technology invented by Ivan Sutherl<strong>and</strong>.<br />

Virtual realities are part of certain ongoing trends. Most obvious<br />

are the trend from interaction to participation in computer events<br />

<strong>and</strong> the trend from passive to active art forms. In addition, artificial


868 / Virtual reality<br />

Ivan Sutherl<strong>and</strong><br />

Ivan Sutherl<strong>and</strong> was born in Hastings, Nebraska, in 1938.<br />

His father was an engineer, <strong>and</strong> from an early age Sutherl<strong>and</strong><br />

considered engineering his own destiny, too. He earned a<br />

bachelor’s degree from the Carnegie Institute of Technology in<br />

1959, a master’s degree from the California Institute of Technology<br />

in 1960, <strong>and</strong> a doctorate from the Massachusetts Institute<br />

of Technology (MIT) in 1963.<br />

His adviser at MIT was Claude Shannon, creator of information<br />

theory, who directed Sutherl<strong>and</strong> to find ways to simplify<br />

the interface between people <strong>and</strong> computers. Out of this<br />

research came Sketchpad. It was software that allowed people<br />

to draw designs on a computer terminal with a light pen,<br />

an early form of computer-assisted design (CAD). The U.S.<br />

Defense Department’s Advanced Research Projects Center<br />

became interested in Sutherl<strong>and</strong>’s work <strong>and</strong> hired him to direct<br />

its Information Processing Techniques Office in 1964. In<br />

1966 he left to become an associate professor of electrical engineering<br />

at Harvard University, moving to the University of<br />

Utah in 1968, <strong>and</strong> then to Caltech in 1975. During his academic<br />

career he developed the graphic interface for virtual<br />

reality, first announced in his ground-breaking 1968 article<br />

“A Head-Mounted Three-Dimensional Display.”<br />

In 1980 Sutherl<strong>and</strong> left academia for industry. He already<br />

had business experience as cofounder of Evans & Sutherl<strong>and</strong><br />

in Salt Lake City. The new firm, Sutherl<strong>and</strong>, Sproull, <strong>and</strong> Associates,<br />

which provided consulting services <strong>and</strong> venture capital,<br />

later became part of Sun Microsystems, Inc. Sutherl<strong>and</strong> remained<br />

as a research fellow <strong>and</strong> vice president. A member of<br />

the National Academy of Engineering <strong>and</strong> National Academy<br />

of Sciences, in 1988 Sutherl<strong>and</strong> was awarded the AM Turing<br />

Award, the highest honor in information technology.<br />

experiences are taking on increasing significance. Businesspersons<br />

like to talk about “doing it right the first time.” This can now be<br />

done in many cases, not because fewer mistakes are being made by<br />

people but because those mistakes are being made in simulated environments.<br />

Most important is that virtual realities provide means of express-


ing <strong>and</strong> experiencing, as well as new ways for people to interact. Entertainment<br />

uses of virtual reality will be as economically significant<br />

as more practical uses, since entertainment is the United States’<br />

number-two export. Vicarious experience through theater, novels,<br />

movies, <strong>and</strong> television represents a significant percentage of people’s<br />

lives in developed countries. The addition of a radically new<br />

form of physically involving, interactive experience is a major cultural<br />

event that may shape human consciousness as much as earlier<br />

forms of experience have.<br />

Consequences<br />

Most religions offer their believers an escape from this world,<br />

but few technologies have been able to do likewise. Not so with<br />

virtual reality, the fledgling technology in which people explore a<br />

simulated three-dimensional environment generated by a computer.<br />

Using this technology, people can not only escape from this<br />

world but also design the world in which they want to live.<br />

In most virtual reality systems, many of which are still experimental,<br />

one watches the scene, or alternative reality, through threedimensional<br />

goggles in a headset. Sound <strong>and</strong> tactile sensations enhance<br />

the illusion of reality. Because of the wide variety of actual<br />

<strong>and</strong> potential applications of virtual reality, from three-dimensional<br />

video games <strong>and</strong> simulators to remotely operated “telepresence”<br />

systems for the nuclear <strong>and</strong> undersea industries, interest in the field<br />

is intense.<br />

The term “virtual reality” describes the computer-generated<br />

simulation of reality with physical, tactile, <strong>and</strong> visual dimensions.<br />

The interactive technology is used by science <strong>and</strong> engineering researchers<br />

as well as by the entertainment industry, especially in<br />

the form of video games. Virtual reality systems can, for example,<br />

simulate a walk-through of a building in an architectural graphics<br />

program. Virtual reality technology in which the artificial world<br />

overlaps with reality will have major social <strong>and</strong> psychological implications.<br />

See also Personal computer; Virtual machine.<br />

Virtual reality / 869


870 / Virtual reality<br />

Further Reading<br />

Earnshaw, Rae A., M. A. Gigante, <strong>and</strong> H. Jones. Virtual Reality Systems.<br />

San Diego: Academic Press, 1993.<br />

Moody, Fred. The Visionary Position: The Inside Story of the Digital<br />

Dreamers Who Are Making Virtual Reality a Reality. New York:<br />

Times Business, 1999.<br />

Sutherl<strong>and</strong>, Ivan Edward. Sketchpad: A Man-Machine Graphical Communication<br />

System. New York: Garl<strong>and</strong>, 1980.


V-2 rocket<br />

V-2 rocket<br />

The invention: The first first long-range, liquid-fueled rocket, the<br />

V-2 was developed by Germany to carry bombs during World<br />

War II.<br />

The people behind the invention:<br />

Wernher von Braun (1912-1977), the chief engineer <strong>and</strong> prime<br />

motivator of rocket research in Germany during the 1930’s<br />

<strong>and</strong> 1940’s<br />

Walter Robert Dornberger (1895-1980), the former comm<strong>and</strong>er<br />

of the Peenemünde Rocket Research Institute<br />

Ing Fritz Gosslau, the head of the V-1 development team<br />

Paul Schmidt, the designer of the impulse jet motor<br />

The “Buzz Bomb”<br />

871<br />

On May 26, 1943, in the middle of World War II, key German military<br />

officials were briefed by two teams of scientists, one representing<br />

the air force <strong>and</strong> the other representing the army. Each team had<br />

launched its own experimental aerial war craft. The military chiefs<br />

were to decide which project merited further funding <strong>and</strong> development.<br />

Each experimental craft had both advantages <strong>and</strong> disadvantages,<br />

<strong>and</strong> each counterbalanced the other. Therefore, it was decided<br />

that both craft were to be developed. They were to become the V-1<br />

<strong>and</strong> the V-2 aircraft.<br />

The impulse jet motor used in the V-1 craft was designed by Munich<br />

engineer Paul Schmidt. On April 30, 1941, the motor had been<br />

used to assist power on a biplane trainer. The development team for<br />

the V-1 was headed by Ing Fritz Gosslau; the aircraft was designed<br />

by Robert Lusser.<br />

The V-1, or “buzz bomb,” was capable of delivering a one-ton warhead<br />

payload. While still in a late developmental stage, it was<br />

launched, under Adolf Hitler’s orders, to terrorize inhabited areas of<br />

London in retaliation for the damage that had been wreaked on Germany<br />

during the war. More than one hundred V-1’s were launched<br />

daily between June 13 <strong>and</strong> early September, 1944. Because the V-1


872 / V-2 rocket<br />

flew in a straight line <strong>and</strong> at a constant speed, Allied aircraft were<br />

able to intercept it more easily than they could the V-2.<br />

Two innovative systems made the V-1 unique: the drive operation<br />

<strong>and</strong> the guidance system. In the motor, oxygen entered the<br />

grid valves through many small flaps. Fuel oil was introduced <strong>and</strong><br />

the mixture of fuel <strong>and</strong> oxygen was ignited. After ignition, the exp<strong>and</strong>ed<br />

gases produced the reaction propulsion. When the exp<strong>and</strong>ed<br />

gases had vacated, the reduced internal pressure allowed the valve<br />

flaps to reopen, admitting more air for the next cycle.<br />

The guidance system included a small propeller connected to a<br />

revolution counter that was preset based on the distance to the target.<br />

The number of propeller revolutions that it would take to reach<br />

the target was calculated before launch <strong>and</strong> punched into the counter.<br />

During flight, after the counter had measured off the selected<br />

number of revolutions, the aircraft’s elevator flaps became activated,<br />

causing the craft to dive at the target. Underst<strong>and</strong>ably, the accuracy<br />

was not what the engineers had hoped.<br />

Vengeance Weapon 2<br />

According to the Treaty of Versailles (1919), world military forces<br />

were restricted to 100,000 men <strong>and</strong> a certain level of weaponry. The<br />

German military powers realized very early, however, that the<br />

treaty had neglected to restrict rocket-powered weaponry, which<br />

did not exist at the end of World War I (1914-1918). Wernher von<br />

Braun was hired as chief engineer for developing the V-2 rocket.<br />

The V-2 had a lift-off thrust of 11,550.5 newtons <strong>and</strong> was propelled<br />

by the combustion of liquid oxygen <strong>and</strong> alcohol. The propellants<br />

were pumped into the combustion chamber by a steampowered<br />

turboprop. The steam was generated by the decomposition<br />

of hydrogen peroxide, using sodium permanganate as a catalyst.<br />

One innovative feature of the V-2 that is still used was regenerative<br />

cooling, which used alcohol to cool the double-walled<br />

combustion chamber.<br />

The guidance system included two phases: powered <strong>and</strong> ballistic.<br />

Four seconds after launch, a preprogrammed tilt to 17 degrees<br />

was begun, then acceleration was continued to achieve the desired<br />

trajectory. At the desired velocity, the engine power was cut off via


one of two systems. In the automatic system, a device shut off the<br />

engine at the velocity desired; this method, however, was inaccurate.<br />

The second system sent a radio signal to the rocket’s receiver,<br />

which cut off the power. This was a far more accurate method, but<br />

the extra equipment required at the launch site was an attractive target<br />

for Allied bombers. This system was more often employed toward<br />

the end of the war.<br />

Even the 907-kilogram warhead of the V-2 was a carefully tested<br />

device. The detonators had to be able to withst<strong>and</strong> 6 g’s of force during<br />

lift-off <strong>and</strong> reentry, as well as the vibrations inherent in a rocket<br />

flight. Yet they also had to be sensitive enough to ignite the bomb<br />

upon impact <strong>and</strong> before the explosive became buried in the target<br />

<strong>and</strong> lost power through diffusion of force.<br />

The V-2’s first successful test was in October of 1942, but it continued<br />

to be developed until August of 1944. During the next eight<br />

months, more than three thous<strong>and</strong> V-2’s were launched against Engl<strong>and</strong><br />

<strong>and</strong> the Continent, causing immense devastation <strong>and</strong> living<br />

up to its name: Vergeltungswaffe zwei (vengeance weapon 2). Unfortunately<br />

for Hitler’s regime, the weapon that took fourteen years of<br />

research <strong>and</strong> testing to perfect entered the war too late to make an<br />

impact upon the outcome.<br />

Impact<br />

V-2 rocket / 873<br />

The V-1 <strong>and</strong> V-2 had a tremendous impact on the history <strong>and</strong> development<br />

of space technology. Even during the war, captured V-2’s<br />

were studied by Allied scientists. American rocket scientists were<br />

especially interested in the technology, since they too were working<br />

to develop liquid-fueled rockets.<br />

After the war, German military personnel were sent to the United<br />

States, where they signed contracts to work with the U.S. Army in a<br />

program known as “Operation Paperclip.” Testing of the captured<br />

V-2’s was undertaken at White S<strong>and</strong>s Missile Range near Alamogordo,<br />

New Mexico. The JB-2 Loon Navy jet-propelled bomb was<br />

developed following the study of the captured German craft.<br />

The Soviet Union also benefited from captured V-2’s <strong>and</strong> from the<br />

German V-2 factories that were dismantled following the war. With<br />

these resources, the Soviet Union developed its own rocket technol-


874 / V-2 rocket<br />

ogy, which culminated in the launch of Sputnik 1, the world’s first artificial<br />

satellite, on October 4, 1957. The United States was not far behind.<br />

It launched its first satellite, Explorer 1, on January 31, 1958. On<br />

April 12, 1961, the world’s first human space traveler, Soviet cosmonaut<br />

Yuri A. Gagarin, was launched into Earth orbit.<br />

See also Airplane; Cruise missile; Hydrogen bomb; Radar; Rocket;<br />

Stealth aircraft.<br />

Further Reading<br />

Bergaust, Erik. Wernher von Braun: The Authoritative <strong>and</strong> Definitive<br />

Biographical Profile of the Father of Modern Space Flight. Washington:<br />

National Space Institute, 1976.<br />

De Maeseneer, Guido. Peenemünde: The Extraordinary Story of Hitler’s<br />

Secret Weapons V-1 <strong>and</strong> V-2. Vancouver: AJ Publishing, 2001.<br />

Piszkiewicz, Dennis. Wernher von Braun: The Man Who Sold the Moon.<br />

Westport, Conn.: Praeger, 1998.


Walkman cassette player<br />

Walkman cassette player<br />

The invention: Inexpensive portable device for listening to stereo<br />

cassettes that was the most successful audio product of the 1980’s<br />

<strong>and</strong> the forerunner of other portable electronic devices.<br />

The people behind the invention:<br />

Masaru Ibuka (1908-1997), a Japanese engineer who cofounded<br />

Sony<br />

Akio Morita (1921-1999), a Japanese physicist <strong>and</strong> engineer,<br />

cofounder of Sony<br />

Norio Ohga (1930- ), a Japanese opera singer <strong>and</strong><br />

businessman who ran Sony’s tape recorder division before<br />

becoming president of the company in 1982<br />

Convergence of Two Technologies<br />

875<br />

The Sony Walkman was the result of the convergence of two<br />

technologies: the transistor, which enabled miniaturization of electronic<br />

components, <strong>and</strong> the compact cassette, a worldwide st<strong>and</strong>ard<br />

for magnetic recording tape. As the smallest tape player devised,<br />

the Walkman was based on a systems approach that made use of advances<br />

in several unrelated areas, including improved loudspeaker<br />

design <strong>and</strong> reduced battery size. The Sony company brought them<br />

together in an innovative product that found a mass market in a remarkably<br />

short time.<br />

Tokyo Telecommunications Engineering, which became Sony,<br />

was one of many small entrepreneurial companies that made audio<br />

products in the years following World War II. It was formed in the<br />

ruins of Tokyo, Japan, in 1946, <strong>and</strong> got its start manufacturing components<br />

for inexpensive radios <strong>and</strong> record players. They were the<br />

ideal products for a company with some expertise in electrical engineering<br />

<strong>and</strong> a limited manufacturing capability.<br />

Akio Morita <strong>and</strong> Masaru Ibuka formed Tokyo Telecommunications<br />

Engineering to make a variety of electrical testing devices <strong>and</strong><br />

instruments, but their real interests were in sound, <strong>and</strong> they decided<br />

to concentrate on audio products. They introduced a reel-to-reel


876 / Walkman cassette player<br />

tape recorder in 1946. Its success ensured that the company would<br />

remain in the audio field. The trade name of the magnetic tape they<br />

manufactured was “Soni,” this was the origin of the company’s new<br />

name, adopted in 1957. The 1953 acquisition of a license to use Bell<br />

Laboratories’ transistor technology was a turning point in the fortunes<br />

of Sony, for it led the company to the highly popular transistor<br />

radio <strong>and</strong> started it along the path to reducing the size of consumer<br />

products. In the 1960’s, Sony led the way to smaller <strong>and</strong> cheaper radios,<br />

tape recorders, <strong>and</strong> television sets, all using transistors instead<br />

of vacuum tubes.<br />

The Consumer Market<br />

The original marketing strategy for manufacturers of mechanical<br />

entertainment devices had been to put one into every home. This<br />

was the goal for Edison’s phonograph, the player piano, the Victrola,<br />

<strong>and</strong> the radio receiver. Sony <strong>and</strong> other Japanese manufacturers<br />

found out that if a product were small enough <strong>and</strong> cheap enough,<br />

two or three might be purchased for home use, or even for outdoor<br />

use. This was the marketing lesson of the transistor radio.<br />

The unparalleled sales of transistor radios indicated that consumer<br />

durables intended for entertainment were not exclusively<br />

used in the home. The appeal of the transistor radio was that it made<br />

entertainment portable. Sony applied this concept to televisions<br />

<strong>and</strong> tape recorders, developing small portable units powered by<br />

batteries. Sony was first to produce a “personal” television set, with<br />

a five-inch screen. To the surprise of many manufacturers who said<br />

there would never be a market for such a novelty item, it sold well.<br />

It was impossible to reduce tape recorders to the size of transistor<br />

radios because of the problems of h<strong>and</strong>ling very small reels of tape<br />

<strong>and</strong> the high power required to turn them. Portable tape recorders required<br />

several large flashlight batteries. Although tape had the advantage<br />

of recording capability, it could not challenge the popularity<br />

of the microgroove 45 revolution-per-minute (rpm) disc because the<br />

tape player was much more difficult to operate. In the 1960’s, several<br />

types of tape cartridge were introduced to overcome this problem, including<br />

the eight-track tape cartridge <strong>and</strong> the Philips compact cassette.<br />

Sony <strong>and</strong> Matsushita were two of the leading Japanese manu-


facturers that quickly incorporated the compact cassette into their<br />

audio products, producing the first cassette players available in the<br />

United States.<br />

The portable cassette players of the 1960’s <strong>and</strong> 1970’s were based<br />

on the transistor radio concept: small loudspeaker, transistorized<br />

amplifier, <strong>and</strong> flashlight batteries all enclosed in a plastic case. The<br />

size of transistorized components was being reduced constantly,<br />

<strong>and</strong> new types of batteries, notably the nickel cadmium combination,<br />

offered higher power output in smaller sizes. The problem of<br />

reducing the size of the loudspeaker without serious deterioration<br />

of sound quality blocked the path to very small cassette players.<br />

Sony’s engineers solved the problem with a very small loudspeaker<br />

device using plastic diaphragms <strong>and</strong> new, lighter materials for the<br />

magnets. These devices were incorporated into tiny stereo headphones<br />

that set new st<strong>and</strong>ards of fidelity.<br />

The first “walkman” was made by Sony engineers for the personal<br />

use of Masaru Ibuka. He wanted to be able to listen to high-fidelity<br />

recorded sound wherever he went, <strong>and</strong> the tiny player was small<br />

enough to fit inside a pocket. Sony was experienced in reducing the<br />

size of machines. At the same time the walkman was being made up,<br />

Sony engineers were struggling to produce a video recording cassette<br />

that was also small enough to fit into Ibuka’s pocket.<br />

Although the portable stereo was part of a long line of successful<br />

miniaturized consumer products, it was not immediately recognized<br />

as a commercial technology. There were already plenty of cassette<br />

players in home units, in automobiles, <strong>and</strong> in portable players.<br />

Marketing experts questioned the need for a tiny version. The board<br />

of directors of Sony had to be convinced by Morita that the new<br />

product had commercial potential. The Sony Soundabout portable<br />

cassette player was introduced to the market in 1979.<br />

Impact<br />

Walkman cassette player / 877<br />

The Soundabout was initially treated as a novelty in the audio<br />

equipment industry. At a price of $200, it could not be considered as<br />

a product for the mass market. Although it sold well in Japan, where<br />

people were used to listening to music on headphones, sales in the<br />

United States were not encouraging. Sony’s engineers, working un-


878 / Walkman cassette player<br />

der the direction of Kozo Ohsone, reduced the size <strong>and</strong> cost of the<br />

machine. In 1981, the Walkman II was introduced. It was 25 percent<br />

smaller than the original version <strong>and</strong> had 50 percent fewer moving<br />

parts. Its price was considerably lower <strong>and</strong> continued to fall.<br />

The Walkman opened a huge market for audio equipment that<br />

nobody knew existed. Sony had again confounded the marketing<br />

experts who doubted the appeal of a new consumer electronics<br />

product. It took about two years for Sony’s Japanese competitors,<br />

including Matsushita, Toshiba, <strong>and</strong> Aiwa, to bring out portable personal<br />

stereos. Such was the popularity of the device that any miniature<br />

cassette player was called a “walkman,” irrespective of the<br />

manufacturer. Sony kept ahead of the competition by constant innovation:<br />

Dolby noise reduction circuits were added in 1982, <strong>and</strong> a rechargeable<br />

battery feature was introduced in 1985. The machine became<br />

smaller, until it was barely larger than the audio cassette it<br />

played.<br />

Sony developed a whole line of personal stereos. Waterproofed<br />

Walkmans were marketed to customers who wanted musical accompaniment<br />

to water sports. There were special models for tennis<br />

players <strong>and</strong> joggers. The line grew to encompass about forty different<br />

types of portable cassette players, priced from about $30 to $500<br />

for a high-fidelity model.<br />

In the ten years following the introduction of the Walkman,<br />

Sony sold fifty million units, including twenty-five million in the<br />

United States. Its competitors sold millions more. They were manufactured<br />

all over the Far East <strong>and</strong> came in a broad range of sizes<br />

<strong>and</strong> prices, with the cheapest models about $20. Increased competition<br />

in the portable tape player market continually forced down<br />

prices. Sony had to respond to the huge numbers of cheap copies<br />

by redesigning the Walkman to bring down its cost <strong>and</strong> by automating<br />

its production. The playing mechanism became part of the<br />

integrated circuit that provided amplification, allowing manufacturing<br />

as one unit.<br />

The Walkman did more than revive sales of audio equipment in<br />

the sagging market of the late 1970’s. It stimulated dem<strong>and</strong> for cassette<br />

tapes <strong>and</strong> helped make the compact cassette the worldwide<br />

st<strong>and</strong>ard for magnetic tape. At the time the Walkman was introduced,<br />

the major form of prerecorded sound was the vinyl micro-


Masaru Ibuka<br />

Walkman cassette player / 879<br />

Nicknamed “genius inventor” in college, Masaru Ibuka developed<br />

into a visionary corporate leader <strong>and</strong> business philosopher.<br />

Born in Nikko City, Japan, in 1908, he took a degree in engineering<br />

from Waseda University in 1933 <strong>and</strong> went to work<br />

at Photo-Chemical Laboratory, which developed movie film.<br />

Changing to naval research during World War II, he met Akio<br />

Morita, another engineer. After the war they opened an electronics<br />

shop together, calling it the Tokyo Telecommunications<br />

Engineering Corporation, <strong>and</strong> began experimenting with tape<br />

recorders.<br />

Their first model was a modest success, <strong>and</strong> the business<br />

grew under Ibuka, who was president <strong>and</strong> later chairman He<br />

thought up a new, less daunting name for his company, Sony, in<br />

the 1950’s, when it rapidly became a leader in consumer electronics.<br />

His goal was to make existing technology useful to people<br />

in everyday life. “He sowed the seeds of a deep conviction<br />

that our products must bring joy <strong>and</strong> fun to users,” one of his<br />

successors as president, Nobuyuki Idei, said in 1997.<br />

While American companies were studying military applications<br />

for the newly developed transistor in the 1950’s, Ibuka<br />

<strong>and</strong> Morita put it to use in an affordable transistor radio <strong>and</strong><br />

then found ways to shrink its size <strong>and</strong> power it with batteries so<br />

that it could be taken anywhere. In a similar fashion, they made<br />

tape recorders <strong>and</strong> players (such as the Walkman), video players,<br />

compact disc players, <strong>and</strong> televisions ever cheaper, more reliable,<br />

<strong>and</strong> more efficiently designed.<br />

A hero in the Japanese business world, Ibuka retired as Sony<br />

chairman in 1976 but continued to help out as a consultant until<br />

his death in 1997.<br />

groove record. In 1983, the ratio of vinyl to cassette sales was 3:2. By<br />

the end of the decade, the audio cassette was the bestselling format<br />

for recorded sound, outselling vinyl records <strong>and</strong> compact discs<br />

combined by a ratio of 2:1. The compatibility of the audio cassette<br />

used in personal players with the home stereo ensured that it would<br />

be the most popular tape recording medium.<br />

The market for portable personal players in the United States<br />

during the decade of the 1990’s was estimated to be more than


880 / Walkman cassette player<br />

twenty million units each year. Sony accounted for half of the 1991<br />

American market of fifteen million units selling at an average price<br />

of $50. It appeared that there would be more than one in every<br />

home. In some parts of Western Europe, there were more cassette<br />

players than people, reflecting the level of market penetration<br />

achieved by the Walkman.<br />

The ubiquitous Walkman had a noticeable effect on the way<br />

that people listen to music. The sound from the headphones of a<br />

portable player is more intimate <strong>and</strong> immediate than the sound<br />

coming from the loudspeakers of a home stereo. The listener can<br />

hear a wider range of frequencies <strong>and</strong> more of the lower amplitudes<br />

of music, while the reverberation caused by sound bouncing<br />

off walls is reduced. The listening public has become accustomed<br />

to the Walkman sound <strong>and</strong> expects it to be duplicated on<br />

commercial recordings. Recording studios that once mixed their<br />

master recordings to suit the reproduction characteristics of car<br />

or transistor radios began to mix them for Walkman headphones.<br />

Personal stereos also enable the listener to experience more of the<br />

volume of recorded sound because it is injected directly into the<br />

ear.<br />

The Walkman established a market for portable tape players that<br />

exerted an influence on all subsequent audio products. The introduction<br />

of the compact disc (CD) in 1983 marked a completely new<br />

technology of recording based on digital transformation of sound. It<br />

was jointly developed by the Sony <strong>and</strong> Philips companies. Despite<br />

the enormous technical difficulties of reducing the size of the laser<br />

reader <strong>and</strong> making it portable, Sony’s engineers devised the Discman<br />

portable compact disc player, which was unveiled in 1984. It<br />

followed the Walkman concept exactly <strong>and</strong> offered higher fidelity<br />

than the cassette tape version. The Discman sold for about $300<br />

when it was introduced, but its price soon dropped to less than<br />

$100. It did not achieve the volume of sales of the audio cassette version<br />

because fewer CDs than audio cassettes were in use. The slow<br />

acceptance of the compact disc hindered sales growth. The Discman<br />

could not match the portability of the Walkman because vibrations<br />

caused the laser reader to skip tracks.<br />

In the competitive market for consumer electronics products, a<br />

company must innovate to survive. Sony had watched cheap compe-


Walkman cassette player / 881<br />

tition erode the sales of many of its most successful products, particularly<br />

the transistor radio <strong>and</strong> personal television, <strong>and</strong> was committed<br />

to both product improvement <strong>and</strong> new entertainment technologies.<br />

It knew that the personal cassette player had a limited sales potential<br />

in the advanced industrial countries, especially after the introduction<br />

of digital recording in the 1980’s. It therefore sought new technology<br />

to apply to the Walkman concept. Throughout the 1980’s, Sony <strong>and</strong><br />

its many competitors searched for a new version of the Walkman.<br />

The next generation of personal players was likely to be based on<br />

digital recording. Sony introduced its digital audio tape (DAT) system<br />

in 1990. This used the same digital technology as the compact<br />

disc but came in tape form. It was incorporated into expensive<br />

home players; naturally, Sony engineered a portable version. The<br />

tiny DAT Walkman offered unsurpassed fidelity of reproduction,<br />

but its incompatibility with any other tape format <strong>and</strong> its high price<br />

limited its sales to professional musicians <strong>and</strong> recording engineers.<br />

After the failure of DAT, Sony refocused its digital technology<br />

into a format more similar to the Walkman. Its Mini Disc (MD) used<br />

the same technology as the compact disc but had the advantage of a<br />

recording capability. The 2.5-inch disc was smaller than the CD, <strong>and</strong><br />

the player was smaller than the Walkman. The play-only version fit<br />

in the palm of a h<strong>and</strong>. A special feature prevented the skipping of<br />

tracks that caused problems with the Discman. The Mini Disc followed<br />

the path blazed by the Walkman <strong>and</strong> represented the most<br />

advanced technology applied to personal stereo players. At a price<br />

of about $500 in 1993, it was still too expensive to compete in the<br />

audio cassette Walkman market, but the history of similar products<br />

illustrates that rapid reduction of price could be achieved even with<br />

a complex technology.<br />

The Walkman had a powerful influence on the development of<br />

other digital <strong>and</strong> optical technologies. The laser readers of compact<br />

disc players can access visual <strong>and</strong> textual information in addition to<br />

sound. Sony introduced the Data Discman, a h<strong>and</strong>held device that<br />

displayed text <strong>and</strong> pictures on a tiny screen. Several other manufacturers<br />

marketed electronic books. Whatever the shape of future entertainment<br />

<strong>and</strong> information technologies, the legacy of the Walkman<br />

will put a high premium on portability, small size, <strong>and</strong> the<br />

interaction of machine <strong>and</strong> user.


882 / Walkman cassette player<br />

See also Cassette recording; Compact disc; Dolby noise reduction;<br />

Electronic synthesizer; Laser; Transistor; Videocassette recorder.<br />

Further Reading<br />

Bull, Michael. Sounding Out the City: Personal Stereos <strong>and</strong> the Management<br />

of Everyday Life. New York: Berg, 2000.<br />

Lyons, Nick. The Sony Vision. New York: Crown Publishers, 1976.<br />

Morita, Akio, with Edwin M. Reingold, <strong>and</strong> Mitsuko Shimomura.<br />

Made in Japan: Akio Morita <strong>and</strong> Sony. London: HarperCollins, 1994.<br />

Nathan, John. Sony: The Private Life. London: HarperCollins Business,<br />

2001.<br />

Schlender, Brenton R. “How Sony Keeps the Magic Going.” Fortune<br />

125 (February 24, 1992).


Washing machine<br />

Washing machine<br />

The invention: Electrical-powered machines that replaced h<strong>and</strong>operated<br />

washing tubs <strong>and</strong> wringers, making the job of washing<br />

clothes much easier.<br />

The people behind the invention:<br />

O. B. Woodrow, a bank clerk who claimed to be the first to<br />

adapt electricity to a remodeled h<strong>and</strong>-operated washing<br />

machine<br />

Alva J. Fisher (1862-1947), the founder of the Hurley Machine<br />

Company, who designed the Thor electric washing machine,<br />

claiming that it was the first successful electric washer<br />

Howard Snyder, the mechanical genius of the Maytag<br />

Company<br />

H<strong>and</strong> Washing<br />

883<br />

Until the development of the electric washing machine in the<br />

twentieth century, washing clothes was a tiring <strong>and</strong> time-consuming<br />

process. With the development of the washboard, dirt was loosened<br />

by rubbing. Clothes <strong>and</strong> tubs had to be carried to the water, or the<br />

water had to be carried to the tubs <strong>and</strong> clothes. After washing <strong>and</strong><br />

rinsing, clothes were h<strong>and</strong>-wrung, hang-dried, <strong>and</strong> ironed with<br />

heavy, heated irons. In nineteenth century America, the laundering<br />

process became more arduous with the greater use of cotton fabrics.<br />

In addition, the invention of the sewing machine resulted in the mass<br />

production of inexpensive ready-to-wear cotton clothing. With more<br />

clothing, there was more washing.<br />

One solution was h<strong>and</strong>-operated washing machines. The first<br />

American patent for a h<strong>and</strong>-operated washing machine was issued<br />

in 1805. By 1857, more than 140 patents had been issued; by 1880, between<br />

4,000 <strong>and</strong> 5,000 patents had been granted. While most of<br />

these machines were never produced, they show how much the<br />

public wanted to find a mechanical means of washing clothes.<br />

Nearly all the early types prior to the Civil War (1861-1865) were<br />

modeled after the washboard.


884 / Washing machine<br />

Washing machines based upon the rubbing principle had two<br />

limitations: They washed only one item at a time, <strong>and</strong> the rubbing<br />

was hard on clothes. The major conceptual breakthrough was to<br />

move away from rubbing <strong>and</strong> to design machines that would clean<br />

by forcing water through a number of clothes at the same time.<br />

An early suction machine used a plunger to force water through<br />

clothes. Later electric machines would have between two <strong>and</strong> four<br />

suction cups, similar to plungers, attached to arms that went up <strong>and</strong><br />

down <strong>and</strong> rotated on a vertical shaft. Another h<strong>and</strong>-operated washing<br />

machine was used to rock a tub on a frame back <strong>and</strong> forth. An<br />

electric motor was later substituted for the h<strong>and</strong> lever that rocked<br />

the tub. A third h<strong>and</strong>-operated washing machine was the dolly<br />

type. The dolly, which looked like an upside-down three-legged<br />

milking stool, was attached to the inside of the tub cover <strong>and</strong> was<br />

turned by a two-h<strong>and</strong>led lever on top of the enclosed tub.<br />

Machine Washing<br />

The h<strong>and</strong>-operated machines that would later dominate the<br />

market as electric machines were the horizontal rotary cylinder<br />

<strong>and</strong> the underwater agitator types. In 1851, James King patented<br />

a machine of the first type that utilized two concentric half-full<br />

cylinders. Water in the outer cylinder was heated by a fire beneath<br />

it; a h<strong>and</strong> crank turned the perforated inner cylinder that<br />

contained clothing <strong>and</strong> soap. The inner-ribbed design of the rotating<br />

cylinder raised the clothes as the cylinder turned. Once the<br />

clothes reached the top of the cylinder, they dropped back down<br />

into the soapy water.<br />

The first underwater agitator-type machine, the second type,<br />

was patented in 1869. In this machine, four blades at the bottom of<br />

the tub were attached to a central vertical shaft that was turned by<br />

a h<strong>and</strong> crank on the outside. The agitation created by the blades<br />

washed the clothes by driving water through the fabric. It was not<br />

until 1922, when Howard Snyder of the Maytag Company developed<br />

an underwater agitator with reversible motion, that this type<br />

of machine was able to compete with the other machines. Without<br />

reversible action, clothes would soon wrap around the blades <strong>and</strong><br />

not be washed.


Claims for inventing the first electric washing machine came<br />

from O. B. Woodrow, who founded the Automatic Electric Washer<br />

Company, <strong>and</strong> Alva J. Fisher, who developed the Thor electric<br />

washing machine for the Hurley Machine Corporation. Both Woodrow<br />

<strong>and</strong> Fisher made their innovations in 1907 by adapting electric<br />

power to modified h<strong>and</strong>-operated, dolly-type machines. Since only<br />

8 percent of American homes were wired for electricity in 1907, the<br />

early machines were advertised as adaptable to electric or gasoline<br />

power but could be h<strong>and</strong>-operated if the power source failed. Soon,<br />

electric power was being applied to the rotary cylinder, oscillating,<br />

<strong>and</strong> suction-type machines. In 1910, a number of companies introduced<br />

washing machines with attached wringers that could be operated<br />

by electricity. The introduction of automatic washers in 1937<br />

meant that washing machines could change phases without the action<br />

of the operator.<br />

Impact<br />

Washing machine / 885<br />

By 1907 (the year electricity was adapted to washing machines),<br />

electric power was already being used to operate fans, ranges, coffee<br />

percolators, flatirons, <strong>and</strong> sewing machines. By 1920, nearly 35<br />

percent of American residences had been wired for electricity; by<br />

1941, nearly 80 percent had been wired. The majority of American<br />

homes had washing machines by 1941; by 1958, this had risen to an<br />

estimated 90 percent.<br />

The growth of electric appliances, especially washing machines,<br />

is directly related to the decline in the number of domestic servants<br />

in the United States. The development of the electric washing machine<br />

was, in part, a response to a decline in servants, especially<br />

laundresses. Also, rather than easing the work of laundresses with<br />

technology, American families replaced their laundresses with washing<br />

machines.<br />

Commercial laundries were also affected by the growth of electric<br />

washing machines. At the end of the nineteenth century, they<br />

were in every major city <strong>and</strong> were used widely. Observers noted<br />

that just as spinning, weaving, <strong>and</strong> baking had once been done in<br />

the home but now were done in commercial establishments, laundry<br />

work had now begun its move out of the home. After World


886 / Washing machine<br />

War II (1939-1945), however, although commercial laundries continued<br />

to grow, their business was centered more <strong>and</strong> more on institutional<br />

laundry, rather than residential laundry, which they had lost<br />

to the home washing machine.<br />

Some scholars have argued that, on one h<strong>and</strong>, the return of laundry<br />

to the home resulted from marketing strategies that developed<br />

the image of the American woman as one who is home operating<br />

her appliances. On the other h<strong>and</strong>, it was probably because the electric<br />

washing machine made the task much easier that American<br />

women, still primarily responsible for the family laundry, were able<br />

to pursue careers outside the home.<br />

See also Electric refrigerator; Microwave cooking; Robot (household);<br />

Vacuum cleaner; Vending machine slug rejector.<br />

Further Reading<br />

Ierley, Merritt. Comforts of Home: The American House <strong>and</strong> the Evolution<br />

of Modern Convenience. New York: C. Potter, 1999.<br />

“Maytag Heritage Embraces Innovation, Dependable Products.”<br />

Machine Design 71, no. 18 (September, 1999).<br />

Shapiro, Laura. “Household Appliances.” Newsweek 130, no. 24A<br />

(Winter, 1997/1998).


Weather satellite<br />

Weather satellite<br />

The invention: A series of cloud-cover meteorological satellites<br />

that pioneered the reconnaissance of large-scale weather systems<br />

<strong>and</strong> led to vast improvements in weather forecasting.<br />

The person behind the invention:<br />

Harry Wexler (1911-1962), director of National Weather Bureau<br />

meteorological research<br />

Cameras in Space<br />

887<br />

The first experimental weather satellite, Tiros 1, was launched<br />

from Cape Canaveral on April 1, 1960. Tiros’s orbit was angled to<br />

cover the area from Montreal, Canada, to Santa Cruz, Argentina, in<br />

the Western Hemisphere. Tiros completed an orbit every ninetynine<br />

minutes <strong>and</strong>, when launched, was expected to survive at least<br />

three months in space, returning thous<strong>and</strong>s of images of large-scale<br />

weather systems.<br />

Tiros 1 was equipped with a pair of vidicon scanner television<br />

cameras, one equipped with a wide-angle lens <strong>and</strong> the other with a<br />

narrow-angle lens. Both cameras created pictures with five hundred<br />

lines per frame at a shutter speed of 1.5 milliseconds. Each television<br />

camera’s imaging data were stored on magnetic tape for downloading<br />

to ground stations when Tiros 1 was in range. The wideangle<br />

lens provided a low-resolution view of an area covering 2,048<br />

square kilometers. The narrow-angle lens had a resolution of half a<br />

kilometer within a viewing area of 205 square kilometers.<br />

Tiros transmitted its data to ground stations, which displayed<br />

the data on television screens. Photographs of these displays were<br />

then made for permanent records. Tiros weather data were sent to<br />

the Naval Photographic Interpretation Center for detailed meteorological<br />

analysis. Next, the photographs were passed along to the<br />

National Weather Bureau for further study.<br />

Tiros caused some controversy because it was able to image large<br />

areas of the communist world: the Soviet Union, Cuba, <strong>and</strong> Mongolia.<br />

The weather satellite’s imaging system was not, however, partic-


888 / Weather satellite<br />

Hurricane off the coast of Florida photographed from space. (PhotoDisc)<br />

ularly useful as a spy satellite, <strong>and</strong> only large-scale surface features<br />

were visible in the images. Nevertheless, the National Aeronautics<br />

<strong>and</strong> Space Administration (NASA) skirted adverse international reactions<br />

by carefully scrutinizing Tiros’s images for evidence of sensitive<br />

surface features before releasing them publicly.<br />

A Startling Discovery<br />

Tiros 1 was not in orbit very long before it made a significant <strong>and</strong><br />

startling discovery. It was the first satellite to document that large<br />

storms have vortex patterns that resemble whirling pinwheels. Within<br />

its lifetime, Tiros photographed more than forty northern mid-latitude<br />

storm systems, <strong>and</strong> each one had a vortex at its center. These storms<br />

were in various stages of development <strong>and</strong> were between 800 <strong>and</strong><br />

1,600 kilometers in diameter. The storm vortex in most of these was located<br />

inside a 560-kilometer-diameter circle around the center of the<br />

storm’s low-pressure zone. Nevertheless, Tiros’s images did not reveal<br />

at what stage in a storm’s development the vortex pattern formed.


This was typical of Tiros’s data. The satellite was truly an experiment,<br />

<strong>and</strong>, as is the case with most initial experiments, various new<br />

phenomena were uncovered but were not fully understood. The data<br />

showed clearly that weather systems could be investigated from orbit<br />

<strong>and</strong> that future weather satellites could be outfitted with sensors that<br />

would lead to better underst<strong>and</strong>ing of meteorology on a global scale.<br />

Tiros 1 did suffer from a few difficulties during its lifetime in orbit.<br />

Low contrast in the television imaging system often made it difficult<br />

to distinguish between cloud cover <strong>and</strong> snow cover. The magnetic<br />

tape system for the high-resolution camera failed at an early<br />

stage. Also, Earth’s magnetic field tended to move Tiros 1 away<br />

from an advantageous Earth observation attitude. Experience with<br />

Tiros 1 led to improvements in later Tiros satellites <strong>and</strong> many other<br />

weather-related satellites.<br />

Consequences<br />

Weather satellite / 889<br />

Prior to Tiros 1, weather monitoring required networks of groundbased<br />

instrumentation centers, airborne balloons, <strong>and</strong> instrumented<br />

aircraft. Brief high-altitude rocket flights provided limited coverage<br />

of cloud systems from above. Tiros 1 was the first step in the development<br />

of the permanent monitoring of weather systems. The resulting<br />

early detection <strong>and</strong> accurate tracking of hurricanes alone have resulted<br />

in savings in both property <strong>and</strong> human life.<br />

As a result of the Tiros 1 experiment, meteorologists were not<br />

ready to discard ground-based <strong>and</strong> airborne weather systems in<br />

favor of satellites alone. Such systems could not provide data<br />

about pressure, humidity, <strong>and</strong> temperature, for example. Tiros 1<br />

did, however, introduce weather satellites as a necessary supplement<br />

to ground-based <strong>and</strong> airborne systems for large-scale monitoring<br />

of weather systems <strong>and</strong> storms. Satellites could provide<br />

more reliable <strong>and</strong> expansive coverage at a far lower cost than a<br />

large contingent of aircraft. Tiros 1, which was followed by nine<br />

similar spacecraft, paved the way for modern weather satellite<br />

systems.<br />

See also Artificial satellite; Communications satellite; Cruise<br />

missile; Radio interferometer; Rocket.


890 / Weather satellite<br />

Further Reading<br />

Fishman, Jack, <strong>and</strong> Robert Kalish. The Weather Revolution: Innovations<br />

<strong>and</strong> Imminent Breakthroughs in Accurate Forecasting. New<br />

York: Plenum Press, 1994.<br />

Kahl, Jonathan D. Weather Watch: Forecasting the Weather. Minneapolis,<br />

Minn.: Lerner, 1996.<br />

Rao, Krishna P. Weather Satellites: Systems, Data, <strong>and</strong> Environmental<br />

Applications. Boston: American Meteorological Society, 1990.<br />

Artist’s depiction of a weather satellite. (PhotoDisc)


Xerography<br />

Xerography<br />

The invention: Process that makes identical copies of documents<br />

with a system of lenses, mirrors, electricity, chemicals that conduct<br />

electricity in bright light, <strong>and</strong> dry inks (toners) that fuse to<br />

paper by means of heat.<br />

The people behind the invention:<br />

Chester F. Carlson (1906-1968), an American inventor<br />

Otto Kornei (1903- ), a German physicist <strong>and</strong> engineer<br />

Xerography, Xerography, Everywhere<br />

The term xerography is derived from the Greek for “dry writing.”<br />

The process of xerography was invented by an American, Chester F.<br />

Carlson, who made the first xerographic copy of a document in<br />

1938. Before the development of xerography, the preparation of copies<br />

of documents was often difficult <strong>and</strong> tedious. Most often, unclear<br />

carbon copies of typed documents were the only available medium<br />

of information transfer.<br />

The development of xerography led to the birth of the giant<br />

Xerox Corporation, <strong>and</strong> the term xerographic was soon shortened to<br />

Xerox. The process of xerography makes identical copies of a document<br />

by using lens systems, mirrors, electricity, chemicals that conduct<br />

electricity in bright light (“semiconductors”), <strong>and</strong> dry inks<br />

called “toners” that are fused to copy paper by means of heat. The<br />

process makes it easy to produce identical copies of a document<br />

quickly <strong>and</strong> cheaply. In addition, xerography has led to huge advances<br />

in information transfer, the increased use of written documents,<br />

<strong>and</strong> rapid decision-making in all areas of society. Xeroxing<br />

can produce both color <strong>and</strong> black-<strong>and</strong>-white copies.<br />

From the First Xerox Copy to Modern Photocopies<br />

891<br />

On October 22, 1938, after years of effort, Chester F. Carlson produced<br />

the first Xerox copy. Reportedly, his efforts grew out of his<br />

1930’s job in the patent department of the New York firm P. R.


892 / Xerography<br />

Mallory <strong>and</strong> Company. He was looking for a quick, inexpensive<br />

method for making copies of patent diagrams <strong>and</strong> other patent<br />

specifications. Much of Carlson’s original work was conducted in<br />

the kitchen of his New York City apartment or in a room behind a<br />

beauty parlor in Astoria, Long Isl<strong>and</strong>. It was in Astoria that Carlson,<br />

with the help of Otto Kornei, produced the first Xerox copy (of the<br />

inscription “10-22-38 Astoria”) on waxed paper.<br />

The first practical method of xerography used the element selenium,<br />

a substance that conducts electricity only when it is exposed<br />

to light. The prototype Xerox copying machines were developed as<br />

a result of the often frustrating, nerve-wracking, fifteen-year collaboration<br />

of Carlson, scientists <strong>and</strong> engineers at the Battelle Memorial<br />

Institute in Columbus, Ohio, <strong>and</strong> the Haloid Company of Rochester,<br />

New York. The Haloid Company financed the effort after 1947,<br />

based on an evaluation made by an executive, John H. Dessauer. In<br />

return, the company obtained the right to manufacture <strong>and</strong> market<br />

Xerox machines. The company, which was originally a manufacturer<br />

of photographic paper, evolved into the giant Xerox Corporation.<br />

Carlson became very wealthy as a result of the royalties <strong>and</strong><br />

dividends paid to him by the company.<br />

Early xerographic machines operated in several stages. First, the<br />

document to be copied was positioned above a mirror so that its image,<br />

lit by a flash lamp <strong>and</strong> projected by a lens, was reflected onto a<br />

drum coated with electrically charged selenium. Wherever dark<br />

sections of the document’s image were reflected, the selenium coating<br />

retained its positive charge. Where the image was light, the<br />

charge of the selenium was lost, because of the photoactive properties<br />

of the selenium.<br />

Next, the drum was dusted with a thin layer of a negatively<br />

charged black powder called a “toner.” Toner particles stuck to positively<br />

charged dark areas of the drum <strong>and</strong> produced a visible image<br />

on the drum. Then, Xerox copy paper, itself positively charged, was<br />

put in contact with the drum, where it picked up negatively charged<br />

toner. Finally, an infrared lamp heated the paper <strong>and</strong> the toner, fusing<br />

the toner to the paper <strong>and</strong> completing the copying process.<br />

In ensuing years, the Xerox Corporation engineered many changes<br />

in the materials <strong>and</strong> mechanics of Xerox copiers. For example, the<br />

semiconductors <strong>and</strong> toners were changed, which increased both the


Chester F. Carlson<br />

Xerography / 893<br />

The copying machine changed Chester Floyd Carlson’s life<br />

even before he invented it. While he was experimenting with<br />

photochemicals in his apartment, the building owner’s daughter<br />

came by to complain about the stench Carlson was creating.<br />

However, she found Carlson himself more compelling than her<br />

complaints <strong>and</strong> married him not long afterward. Soon Carlson<br />

transferred his laboratory to a room behind his mother-in-law’s<br />

beauty parlor, where he devoted ten dollars a month from his<br />

meager wages to spend on research.<br />

Born in Seattle, Washington, in 1906, Carlson learned early<br />

to husb<strong>and</strong> his resources, set his goals high, <strong>and</strong> never give up.<br />

Both his father <strong>and</strong> mother were sickly, <strong>and</strong> so after he was fourteen,<br />

Carlson was the family’s main breadwinner. His relentless<br />

drive <strong>and</strong> native intelligence got him through high school <strong>and</strong><br />

into a community college, where an impressed teacher inspired<br />

him to go even further—into the California Institute of Technology.<br />

After he graduated, he worked for General Electric but lost<br />

his job during the layoffs caused by the Great Depression. In<br />

1933 he hired on with P. R. Mallory Company, an electrical component<br />

manufacturer, which, although not interested in his invention,<br />

at least paid him enough in wages to keep going.<br />

His thirteen-year crusade to invent a copier <strong>and</strong> then find a<br />

manufacturer to build it ended just as Carlson was nearly<br />

broke. In 1946 Haloid Corporation licensed the rights to Carlson’s<br />

copying machine, but even then the invention did not become<br />

an important part of American communications culture<br />

until the company marketed the Xerox 914 in 1960. The earnings<br />

for Xerox Corporation (as it was called after 1961) leapt<br />

from $33 million to more than $500 million in the next six years,<br />

<strong>and</strong> Carlson became enormously wealthy. He won the Inventor<br />

of the Year Award in 1964 <strong>and</strong> the Horatio Alger Award in 1966.<br />

Before he died in 1968, he remembered the hardships of his<br />

youth by donating $100 million to research organizations <strong>and</strong><br />

charitable foundations.<br />

quality of copies <strong>and</strong> the safety of the copying process. In addition,<br />

auxiliary lenses of varying focal length were added, along with other<br />

features, which made it possible to produce enlarged or reduced copies.<br />

Furthermore, modification of the mechanical <strong>and</strong> chemical prop-


894 / Xerography<br />

erties of the components of the system made it possible to produce<br />

thous<strong>and</strong>s of copies per hour, sort them, <strong>and</strong> staple them.<br />

The next development was color Xerox copying. Color systems<br />

use the same process steps that the black-<strong>and</strong>-white systems use,<br />

but the document exposure <strong>and</strong> toning operations are repeated<br />

three times to yield the three overlaid colored layers (yellow, magenta,<br />

<strong>and</strong> cyan) that are used to produce multicolored images in<br />

any color printing process. To accomplish this, blue, green, <strong>and</strong> red<br />

filters are rotated in front of the copier’s lens system. This action<br />

produces three different semiconductor images on three separate<br />

rollers. Next, yellow, magenta, <strong>and</strong> cyan toners are used—each on<br />

its own roller—to yield three images. Finally, all three images are<br />

transferred to one sheet of paper, which is heated to produce the<br />

multicolored copy. The complex color procedure is slower <strong>and</strong><br />

much more expensive than the black-<strong>and</strong>-white process.<br />

Impact<br />

The quick, inexpensive copying of documents is commonly performed<br />

worldwide. Memor<strong>and</strong>a that must be distributed to hundreds<br />

of business employees can now be copied in moments, whereas in the<br />

past such a process might have occupied typists for days <strong>and</strong> cost<br />

hundreds of dollars. Xerox copying also has the advantage that each<br />

copy is an exact replica of the original; no new errors can be introduced,<br />

as was the case when documents had to be retyped. Xerographic<br />

techniques are also used to reproduce X rays <strong>and</strong> many other<br />

types of medical <strong>and</strong> scientific data, <strong>and</strong> the facsimile (fax) machines<br />

that are now used to send documents from one place to another over<br />

telephone lines are a variation of the Xerox process.<br />

All this convenience is not without some problems: The ease of<br />

photocopying has made it possible to reproduce copyrighted publications.<br />

Few students at libraries, for example, think twice about<br />

copying portions of books, since it is easy <strong>and</strong> inexpensive to do so.<br />

However, doing so can be similar to stealing, according to the law.<br />

With the advent of color photocopying, an even more alarming<br />

problem has arisen: Thieves are now able to use this technology to<br />

create counterfeit money <strong>and</strong> checks. Researchers will soon find a<br />

way to make such important documents impossible to copy.


See also Fax machine; Instant photography; Laser-diode recording<br />

process.<br />

Further Reading<br />

Xerography / 895<br />

Kelley, Neil D. “Xerography: The Greeks Had a Word for It.”<br />

Infosystems 24, no. 1 (January, 1977).<br />

McClain, Dylan L. “Duplicate Efforts.” New York Times (November<br />

30, 1998).<br />

Mort, J. The Anatomy of Xerography: Its Invention <strong>and</strong> Evolution. Jefferson,<br />

N.C.: McFarl<strong>and</strong>, 1989.


896<br />

X-ray crystallography<br />

X-ray crystallography<br />

The invention: Technique for using X rays to determine the crystal<br />

structures of many substances.<br />

The people behind the invention:<br />

Sir William Lawrence Bragg (1890-1971), the son of Sir William<br />

Henry Bragg <strong>and</strong> cowinner of the 1915 Nobel Prize in Physics<br />

Sir William Henry Bragg (1862-1942), an English mathematician<br />

<strong>and</strong> physicist <strong>and</strong> cowinner of the 1915 Nobel Prize in<br />

Physics<br />

Max von Laue (1879-1960), a German physicist who won the<br />

1914 Nobel Prize in Physics<br />

Wilhelm Conrad Röntgen (1845-1923), a German physicist who<br />

won the 1901 Nobel Prize in Physics<br />

René-Just Haüy (1743-1822), a French mathematician <strong>and</strong><br />

mineralogist<br />

Auguste Bravais (1811-1863), a French physicist<br />

The Elusive Crystal<br />

A crystal is a body that is formed once a chemical substance has<br />

solidified. It is uniformly shaped, with angles <strong>and</strong> flat surfaces that<br />

form a network based on the internal structure of the crystal’s atoms.<br />

Determining what these internal crystal structures look like is<br />

the goal of the science of X-ray crystallography. To do this, it studies<br />

the precise arrangements into which the atoms are assembled.<br />

Central to this study is the principle of X-ray diffraction. This<br />

technique involves the deliberate scattering of X rays as they are<br />

shot through a crystal, an act that interferes with their normal path<br />

of movement. The way in which the atoms are spaced <strong>and</strong> arranged<br />

in the crystal determines how these X rays are reflected off them<br />

while passing through the material. The light waves thus reflected<br />

form a telltale interference pattern. By studying this pattern, scientists<br />

can discover variations in the crystal structure.<br />

The development of X-ray crystallography in the early twentieth<br />

century helped to answer two major scientific questions: What are X


ays? <strong>and</strong> What are crystals? It gave birth to a new technology for<br />

the identification <strong>and</strong> classification of crystalline substances.<br />

From studies of large, natural crystals, chemists <strong>and</strong> geologists<br />

had established the elements of symmetry through which one<br />

could classify, describe, <strong>and</strong> distinguish various crystal shapes.<br />

René-Just Haüy, about a century before, had demonstrated that diverse<br />

shapes of crystals could be produced by the repetitive stacking<br />

of tiny solid cubes.<br />

Auguste Bravais later showed, through mathematics, that all<br />

crystal forms could be built from a repetitive stacking of three-dimensional<br />

arrangements of points (lattice points) into “space lattices,”<br />

but no one had ever been able to prove that matter really was<br />

arranged in space lattices. Scientists did not know if the tiny building<br />

blocks modeled by space lattices actually were solid matter<br />

throughout, like Haüy’s cubes, or if they were mostly empty space,<br />

with solid matter located only at the lattice points described by<br />

Bravais.<br />

With the disclosure of the atomic model of Danish physicist Niels<br />

Bohr in 1913, determining the nature of the building blocks of crystals<br />

took on a special importance. If crystal structure could be<br />

shown to consist of atoms at lattice points, then the Bohr model<br />

would be supported, <strong>and</strong> science then could ab<strong>and</strong>on the theory<br />

that matter was totally solid.<br />

X Rays Explain Crystal Structure<br />

X-ray crystallography / 897<br />

In 1912, Max von Laue first used X rays to study crystalline matter.<br />

Laue had the idea that irradiating a crystal with X rays might<br />

cause diffraction. He tested this idea <strong>and</strong> found that X rays were<br />

scattered by the crystals in various directions, revealing on a photographic<br />

plate a pattern of spots that depended on the orientation<br />

<strong>and</strong> the symmetry of the crystal.<br />

The experiment confirmed in one stroke that crystals were not<br />

solid <strong>and</strong> that their matter consisted of atoms occupying lattice sites<br />

with substantial space in between. Further, the atomic arrangements<br />

of crystals could serve to diffract light rays. Laue received the 1914<br />

Nobel Prize in Physics for his discovery of the diffraction of X rays in<br />

crystals.


(Library of Congess)<br />

898 / X-ray crystallography<br />

Sir William Henry Bragg<br />

<strong>and</strong> Sir William Lawrence Bragg<br />

William Henry Bragg, senior member of one of the most illustrious<br />

father-son scientific teams in history, was born in Cumberl<strong>and</strong>,<br />

Engl<strong>and</strong>, in 1862. Talented at mathematics, he<br />

studied that field at Trinity College, Cambridge, <strong>and</strong><br />

physics at the Cavendish Laboratory, then moved into<br />

a professorship at the University of Adelaide in<br />

Australia. Despite an underequipped laboratory, he<br />

proved that the atom is not a solid body, <strong>and</strong> his<br />

work with X rays attracted the attention of Ernest<br />

Rutherford in Engl<strong>and</strong>, who helped him win a professorship<br />

at the University of Leeds in 1908.<br />

By then his eldest son, William Lawrence Bragg,<br />

William Henry Bragg<br />

was showing considerable scientific abilities of his<br />

own. Born in Adelaide in 1890, he also attended Trinity<br />

College, Cambridge, <strong>and</strong> performed research at the Cavendish.<br />

It was while there that father <strong>and</strong> son worked together to<br />

establish the specialty of X-ray crystallography. When they<br />

shared the 1915 Nobel Prize in Physics for their work, the son<br />

was only twenty-five years old—the youngest person ever to<br />

receive a Nobel Prize in any field.<br />

The younger Bragg was also an artillery officer in France<br />

during World War I. Meanwhile, his father worked for the<br />

Royal Admiralty. The hydrophone he invented to detect submarines<br />

underwater earned him a knighthood in 1920. The father<br />

moved to University College, London, <strong>and</strong> became director<br />

of the Royal Institution. His popular lectures about the latest<br />

scientific developments made him famous among the public,<br />

while his elevation to president of the Royal Society in 1935<br />

placed him among the most influential scientists in the world.<br />

He died in 1942.<br />

The son taught at the University of Manchester in 1919 <strong>and</strong><br />

then in 1938 became director of the National Physics Laboratory<br />

<strong>and</strong> professor of physics at the Cavendish. Following the<br />

father’s example, he became an administrator <strong>and</strong> professor at<br />

the Royal Institution, where he also distinguished himself with<br />

his popular lectures. He encouraged research using X-ray crystallography,<br />

including the work that unlocked the structure of<br />

deoxyribonucleic acid (DNA). Knighted in 1941, he became a<br />

royal Companion of Honor in 1967. He died in 1971.


Still, the diffraction of X rays was not yet a proved scientific fact.<br />

Sir William Henry Bragg contributed the final proof by passing one of<br />

the diffracted beams through a gas <strong>and</strong> achieving ionization of the<br />

gas, the same effect that true X rays would have caused. He also used<br />

the spectrometer he built for this purpose to detect <strong>and</strong> measure specific<br />

wavelengths of X rays <strong>and</strong> to note which orientations of crystals<br />

produced the strongest reflections. He noted that X rays, like visible<br />

light, occupy a definite part of the electromagnetic spectrum. Yet<br />

most of Bragg’s work focused on actually using X rays to deduce<br />

crystal structures.<br />

Sir Lawrence Bragg was also deeply interested in this new phenomenon.<br />

In 1912, he had the idea that the pattern of spots was an indication<br />

that the X rays were being reflected from the planes of atoms in the crystal.<br />

If that were true, Laue pictures could be used to obtain information<br />

about the structures of crystals. Bragg developed an equation that described<br />

the angles at which X rays would be most effectively diffracted<br />

by a crystal. This was the start of the X-ray analysis of crystals.<br />

Henry Bragg had at first used his spectrometer to try to determine<br />

whether X rays had a particulate nature. It soon became evident,<br />

however, that the device was a far more powerful way of analyzing<br />

crystals than the Laue photograph method had been. Not<br />

long afterward, father <strong>and</strong> son joined forces <strong>and</strong> founded the new<br />

science of X-ray crystallography. By experimenting with this technique,<br />

Lawrence Bragg came to believe that if the lattice models of<br />

Bravais applied to actual crystals, a crystal structure could be<br />

viewed as being composed of atoms arranged in a pattern consisting<br />

of a few sets of flat, regularly spaced, parallel planes.<br />

Diffraction became the means by which the Braggs deduced the<br />

detailed structures of many crystals. Based on these findings, they<br />

built three-dimensional scale models out of wire <strong>and</strong> spheres that<br />

made it possible for the nature of crystal structures to be visualized<br />

clearly even by nonscientists. Their results were published in the<br />

book X-Rays <strong>and</strong> Crystal Structure (1915).<br />

Impact<br />

X-ray crystallography / 899<br />

The Braggs founded an entirely new discipline, X-ray crystallography,<br />

which continues to grow in scope <strong>and</strong> application. Of partic-


900 / X-ray crystallography<br />

ular importance was the early discovery that atoms, rather than<br />

molecules, determine the nature of crystals. X-ray spectrometers of<br />

the type developed by the Braggs were used by other scientists to<br />

gain insights into the nature of the atom, particularly the innermost<br />

electron shells. The tool made possible the timely validation of some<br />

of Bohr’s major concepts about the atom.<br />

X-ray diffraction became a cornerstone of the science of mineralogy.<br />

The Braggs, chemists such as Linus Pauling, <strong>and</strong> a number of<br />

mineralogists used the tool to do pioneering work in deducing the<br />

structures of all major mineral groups. X-ray diffraction became the<br />

definitive method of identifying crystalline materials.<br />

Metallurgy progressed from a technology to a science as metallurgists<br />

became able, for the first time, to deduce the structural order of<br />

various alloys at the atomic level. Diffracted X rays were applied in<br />

the field of biology, particularly at the Cavendish Laboratory under<br />

the direction of Lawrence Bragg. The tool proved to be essential for<br />

deducing the structures of hemoglobin, proteins, viruses, <strong>and</strong> eventually<br />

the double-helix structure of deoxyribonucleic acid (DNA).<br />

See also Field ion microscope; Geiger counter; Holography;<br />

Mass spectrograph; Neutrino detector; Scanning tunneling microscope;<br />

Thermal cracking process; Ultramicroscope.<br />

Further Reading<br />

Achilladelis, Basil, <strong>and</strong> Mary Ellen Bowden. Structures of Life. Philadelphia:<br />

The Center, 1989.<br />

Bragg, William Lawrence. The Development of X-Ray Analysis. New<br />

York: Hafner Press, 1975.<br />

Thomas, John Meurig. “Architecture of the Invisible.” Nature 364<br />

(August 5, 1993).


X-ray image intensifier<br />

X-ray image intensifier<br />

The invention: A complex electronic device that increases the intensity<br />

of the light in X-ray beams exiting patients, thereby making<br />

it possible to read finer details.<br />

The people behind the invention:<br />

Wilhelm Conrad Röntgen (1845-1923), a German physicist<br />

Thomas Alva Edison (1847-1931), an American inventor<br />

W. Edward Chamberlain, an American physician<br />

Thomson Electron Tubes, a French company<br />

Radiologists Need Dark Adaptation<br />

901<br />

Thomas Alva Edison invented the fluoroscope in 1896, only one<br />

year after Wilhelm Conrad Röntgen’s discovery of X rays. The primary<br />

function of the fluoroscope is to create images of the internal<br />

structures <strong>and</strong> fluids in the human body. During fluoroscopy, the radiologist<br />

who performs the procedure views a continuous image of<br />

the motion of the internal structures.<br />

Although much progress was made during the first half of the<br />

twentieth century in recording X-ray images on plates <strong>and</strong> film,<br />

fluoroscopy lagged behind. In conventional fluoroscopy, a radiologist<br />

observed an image on a dim fluoroscopic screen. In the same<br />

way that it is more difficult to read a telephone book in dim illumination<br />

than in bright light, it is much harder to interpret a dim<br />

fluoroscopic image than a bright one. In the early years of fluoroscopy,<br />

the radiologist’s eyes had to be accustomed to dim illumination<br />

for at least fifteen minutes before performing fluoroscopy.<br />

“Dark adaptation” was the process of wearing red goggles under<br />

normal illumination so that the amount of light entering the eye<br />

was reduced.<br />

The human retina contains two kinds of light-sensitive elements:<br />

rods <strong>and</strong> cones. The dim light emitted by the screen of the<br />

fluoroscope, even under the best conditions, required the radiologist<br />

to see only with the rods, <strong>and</strong> vision is much less accurate in<br />

such circumstances. For normal rod-<strong>and</strong>-cone vision, the bright-


902 / X-ray image intensifier<br />

ness of the screen might have to be increased a thous<strong>and</strong>fold.<br />

Such an increase was impossible; even if an X-ray tube could have<br />

been built that was capable of emitting a beam of sufficient intensity,<br />

its rays would have been fatal to the patient in less than a<br />

minute.<br />

Fluoroscopy in an Undarkened Room<br />

In a classic paper delivered at the December, 1941, meeting of<br />

the Radiological Society of North America, Dr. W. Edward Chamberlain<br />

of Temple University Medical School proposed applying to<br />

fluoroscopy the techniques of image amplification (also known as<br />

image intensification) that had already been adapted for use in the<br />

electron microscope <strong>and</strong> in television. The idea was not original<br />

with him. Four or five years earlier, Irving Langmuir of General<br />

Electric Company had applied for a patent for a device that would<br />

intensify a fluoroscopic image. “It is a little hard to underst<strong>and</strong> the<br />

delay in the creation of a practical device,” Chamberlain noted.<br />

“Perhaps what is needed is a realization by the physicists <strong>and</strong> the<br />

engineers of the great need for brighter fluoroscopic images <strong>and</strong><br />

the great advantage to humanity which their arrival would entail.”<br />

Chamberlain’s brilliant analysis provided precisely that awareness.<br />

World War II delayed the introduction of fluoroscopic image<br />

intensification, but during the 1950’s, a number of image intensifiers<br />

based on the principles Chamberlain had outlined came on the<br />

market.<br />

The image-intensifier tube is a complex electronic device that receives<br />

the X-ray beam exiting the patient, converts it into light, <strong>and</strong><br />

increases the intensity of that light. The tube is usually contained in<br />

a glass envelope that provides some structural support <strong>and</strong> maintains<br />

a vacuum. The X rays, after passing through the patient, impinge<br />

on the face of a screen <strong>and</strong> trigger the ejection of electrons,<br />

which are then speeded up <strong>and</strong> focused within the tube by means of<br />

electrical fields. When the speeded-up electrons strike the phosphor<br />

at the output end of the tube, they trigger the emission of light photons<br />

that re-create the desired image, which is several thous<strong>and</strong><br />

times brighter than is the case with the conventional fluoroscopic<br />

screen. The output of the image intensifier can be viewed in an


undarkened room without prior dark adaptation, thus saving the<br />

radiologist much valuable time.<br />

Moving pictures can be taken of the output phosphor of the intensifying<br />

tube or of the television receiver image, <strong>and</strong> they can be<br />

stored on motion picture film or on magnetic tape. This permanently<br />

records the changing image <strong>and</strong> makes it possible to reduce<br />

further the dose of radiation that a patient must receive. Instead of<br />

prolonging the radiation exposure while examining various parts of<br />

the image or checking for various factors, the radiologist can record<br />

a relatively short exposure <strong>and</strong> then rerun the motion picture film or<br />

tape as often as necessary to analyze the information that it contains.<br />

The radiation dosage that is administered to the patient can be<br />

reduced to a tenth or even a hundredth of what it had been previously,<br />

<strong>and</strong> the same amount of diagnostic information or more can<br />

be obtained. The radiation dose that the radiologist receives is reduced<br />

to zero or almost zero. In addition, the combination of the<br />

brighter image <strong>and</strong> the lower radiation dosage administered to the<br />

patient has made it possible for radiologists to develop a number of<br />

important new diagnostic procedures that could not have been accomplished<br />

at all without image intensification.<br />

Impact<br />

X-ray image intensifier / 903<br />

The image intensifier that was developed by the French company<br />

Thomson Electron Tubes in 1959 had an input-phosphor diameter,<br />

or field, of four inches. Later on, image intensifiers with<br />

field sizes of up to twenty-two inches became available, making it<br />

possible to create images of much larger portions of the human<br />

anatomy.<br />

The most important contribution made by image intensifiers was<br />

to increase fluoroscopic screen illumination to the level required for<br />

cone vision. These devices have made dark adaptation a thing of the<br />

past. They have also brought the television camera into the fluoroscopic<br />

room <strong>and</strong> opened up a whole new world of fluoroscopy.<br />

See also Amniocentesis; CAT scanner; Electrocardiogram; Electroencephalogram;<br />

Mammography; Nuclear magnetic resonance;<br />

Ultrasound.


904 / X-ray image intensifier<br />

Further Reading<br />

Glasser, Otto. Dr. W. C. Röntgen. 2d ed. Springfield, Ill.: Charles C.<br />

Thomas, 1972.<br />

Isherwood, Ian, Adrian Thomas, <strong>and</strong> Peter Neil Temple Wells. The<br />

Invisible Light: One Hundred Years of Medical Radiology. Cambridge,<br />

Mass.: Blackwell Science, 1995.<br />

Lewis, Ricki. “Radiation Continuing Concern with Fluoroscopy.”<br />

FDA Consumer 27 (November, 1993).


Yellow fever vaccine<br />

Yellow fever vaccine<br />

The invention: The first safe vaccine agaisnt the virulent yellow fever<br />

virus, which caused some of the deadliest epidemics of the<br />

nineteenth <strong>and</strong> early twentieth centuries.<br />

The people behind the invention:<br />

Max Theiler (1899-1972), a South African microbiologist<br />

Wilbur Augustus Sawyer (1879-1951), an American physician<br />

Hugh Smith (1902-1995), an American physician<br />

A Yellow Flag<br />

905<br />

Yellow fever, caused by a virus <strong>and</strong> transmitted by mosquitoes,<br />

infects humans <strong>and</strong> monkeys. After the bite of the infecting mosquito,<br />

it takes several days before symptoms appear. The onset of<br />

symptoms is abrupt, with headache, nausea, <strong>and</strong> vomiting. Because<br />

the virus destroys liver cells, yellowing of the skin <strong>and</strong> eyes is common.<br />

Approximately 10 to 15 percent of patients die after exhibiting<br />

terrifying signs <strong>and</strong> symptoms. Death occurs usually from liver necrosis<br />

(decay) <strong>and</strong> liver shutdown. Those that survive recover completely<br />

<strong>and</strong> are immunized.<br />

At the beginning of the twentieth century, there was no cure for<br />

yellow fever. The best that medical authorities could do was to quarantine<br />

the afflicted. Those quarantines usually waved the warning yellow<br />

flag, which gave the disease its colloquial name, “yellow jack.”<br />

After the Aëdes aegypti mosquito was clearly identified as the carrier<br />

of the disease in 1900, efforts were made to combat the disease<br />

by wiping out the mosquito. Most famous in these efforts were the<br />

American army surgeon Walter Reed <strong>and</strong> the Cuban physician<br />

Carlos J. Finlay. This strategy was successful in Panama <strong>and</strong> Cuba<br />

<strong>and</strong> made possible the construction of the Panama Canal. Still, the<br />

yellow fever virus persisted in the tropics, <strong>and</strong> the opening of the<br />

Panama Canal increased the danger of its spreading aboard the<br />

ships using this new route.<br />

Moreover, the disease, which was thought to be limited to the<br />

jungles of South <strong>and</strong> Central America, had begun to spread arounds


906 / Yellow fever vaccine<br />

the world to wherever the mosquito Aëdes aegypti could carry the<br />

virus. Mosquito larvae traveled well in casks of water aboard<br />

trading vessels <strong>and</strong> spread the disease to North America <strong>and</strong> Europe.<br />

Immunization by Mutation<br />

Max Theiler received his medical education in London. Following<br />

that, he completed a four-month course at the London School of<br />

Hygiene <strong>and</strong> Tropical Medicine, after which he was invited to come<br />

to the United States to work in the department of tropical medicine<br />

at Harvard University.<br />

While there, Theiler started working to identify the yellow fever<br />

organism. The first problem he faced was finding a suitable<br />

laboratory animal that could be infected with yellow fever. Until<br />

that time, the only animal successfully infected with yellow fever<br />

was the rhesus monkey, which was expensive <strong>and</strong> difficult to care<br />

for under laboratory conditions. Theiler succeeded in infecting<br />

laboratory mice with the disease by injecting the virus directly into<br />

their brains.<br />

Laboratory work for investigators <strong>and</strong> assistants coming in contact<br />

with the yellow fever virus was extremely dangerous. At least<br />

six of the scientists at the Yellow Fever Laboratory at the Rockefeller<br />

Institute died of the disease, <strong>and</strong> many other workers were<br />

infected. In 1929, Theiler was infected with yellow fever; fortunately,<br />

the attack was so mild that he recovered quickly <strong>and</strong> resumed<br />

his work.<br />

During one set of experiments, Theiler produced successive generations<br />

of the virus. First, he took virus from a monkey that had died<br />

of yellow fever <strong>and</strong> used it to infect a mouse. Next, he extracted the<br />

virus from that mouse <strong>and</strong> injected it into a second mouse, repeating<br />

the same procedure using a third mouse. All of them died of encephalitis<br />

(inflammation of the brain). The virus from the third mouse was<br />

then used to infect a monkey. Although the monkey showed signs of<br />

yellow fever, it recovered completely. When Theiler passed the virus<br />

through more mice <strong>and</strong> then into the abdomen of another monkey,<br />

the monkey showed no symptoms of the disease. The results of these<br />

experiments were published by Theiler in the journal Science.


Yellow fever vaccine / 907<br />

This article caught the attention of Wilbur Augustus Sawyer, director<br />

of the Yellow Fever Laboratory at the Rockefeller Foundation<br />

International Health Division in New York. Sawyer, who was working<br />

on a yellow fever vaccine, offered Theiler a job at the Rockefeller<br />

Foundation, which Theiler accepted. Theiler’s mouse-adapted, “attenuated”<br />

virus was given to the laboratory workers, along with human<br />

immune serum, to protect them against the yellow fever virus.<br />

This type of vaccination, however, carried the risk of transferring<br />

other diseases, such as hepatitis, in the human serum.<br />

In 1930, Theiler worked with Eugen Haagen, a German bacteriologist,<br />

at the Rockefeller Foundation. The strategy of the Rockefeller<br />

laboratory was a cautious, slow, <strong>and</strong> steady effort to culture a strain<br />

of the virus so mild as to be harmless to a human but strong enough<br />

to confer a long-lasting immunity. (To “culture” something—tissue<br />

cells, microorganisms, or other living matter—is to grow it in a specially<br />

prepared medium under laboratory conditions.) They started<br />

with a new strain of yellow fever harvested from a twenty-eightyear-old<br />

West African named Asibi; it was later known as the “Asibi<br />

strain.” It was a highly virulent strain that in four to seven days<br />

killed almost all the monkeys that were infected with it. From time<br />

to time, Theiler or his assistant would test the culture on a monkey<br />

<strong>and</strong> note the speed with which it died.<br />

It was not until April, 1936, that Hugh Smith, Theiler’s assistant,<br />

called to his attention an odd development as noted in the laboratory<br />

records of strain 17D. In its 176th culture, 17D had failed to kill<br />

the test mice. Some had been paralyzed, but even these eventually<br />

recovered. Two monkeys who had received a dose of 17D in their<br />

brains survived a mild attack of encephalitis, but those who had<br />

taken the infection in the abdomen showed no ill effects whatever.<br />

Oddly, subsequent subcultures of the strain killed monkeys <strong>and</strong><br />

mice at the usual rate. The only explanation possible was that a mutation<br />

had occurred unnoticed.<br />

The batch of strain 17D was tried over <strong>and</strong> over again on monkeys<br />

with no harmful effects. Instead, the animals were immunized<br />

effectively. Then it was tried on the laboratory staff, including<br />

Theiler <strong>and</strong> his wife, Lillian. The batch injected into humans had the<br />

same immunizing effect. Neither Theiler nor anyone else could explain<br />

how the mutation of the virus had resulted. Attempts to dupli-


908 / Yellow fever vaccine<br />

cate the experiment, using the same Asibi virus, failed. Still, this was<br />

the first safe vaccine for yellow fever. In June, 1937, Theiler reported<br />

this crucial finding in the Journal of Experimental Medicine.<br />

Impact<br />

Following the discovery of the vaccine, Theiler’s laboratory became<br />

a production plant for the 17D virus. Before World War II<br />

(1939-1945), more than one million vaccination doses were sent to<br />

Brazil <strong>and</strong> other South American countries. After the United States<br />

entered the war, eight million soldiers were given the vaccine before<br />

being shipped to tropical war zones. In all, approximately fifty million<br />

people were vaccinated in the war years.<br />

Yet although the vaccine, combined with effective mosquito control,<br />

eradicated the disease from urban centers, yellow fever is still<br />

present in large regions of South <strong>and</strong> Central America <strong>and</strong> of Africa.<br />

The most severe outbreak of yellow fever ever known occurred<br />

from 1960 to 1962 in Ethiopia; out of one hundred thous<strong>and</strong> people<br />

infected, thirty thous<strong>and</strong> died.<br />

The 17D yellow fever vaccine prepared by Theiler in 1937 continues<br />

to be the only vaccine used by the World Health Organization,<br />

more than fifty years after its discovery. There is a continuous effort<br />

by that organization to prevent infection by immunizing the people<br />

living in tropical zones.<br />

See also Antibacterial drugs; Penicillin; Polio vaccine (Sabin); Polio<br />

vaccine (Salk); Salvarsan; Tuberculosis vaccine; Typhus vaccine.<br />

Further Reading<br />

DeJauregui, Ruth. One Hundred Medical Milestones That Shaped World<br />

History. San Mateo, Calif.: Bluewood Books, 1998.<br />

Delaporte, François. The History of Yellow Fever: An Essay on the Birth<br />

of Tropical Medicine. Cambridge, Mass.: MIT Press, 1991.<br />

Theiler, Max, <strong>and</strong> Wilbur G. Downs. The Arthropod-borne Viruses of<br />

Vertebrates: An Account of the Rockefeller Foundation Virus Program,<br />

1951-1970. New Haven, Conn.: Yale University Press, 1973.<br />

Williams, Greer. Virus Hunters. London: Hutchinson, 1960.


Time Line<br />

Time Line<br />

Date Invention<br />

c. 1900 Electrocardiogram<br />

1900 Brownie camera<br />

1900 Dirigible<br />

1901 Artificial insemination<br />

1901 Vat dye<br />

1901-1904 Silicones<br />

1902 Ultramicroscope<br />

1903 Airplane<br />

1903 Disposable razor<br />

1903-1909 Laminated glass<br />

1904 Alkaline storage battery<br />

1904 Photoelectric cell<br />

1904 Vacuum tube<br />

1905 Blood transfusion<br />

1905-1907 Plastic<br />

1906 Gyrocompass<br />

1906 Radio<br />

1906-1911 Tungsten filament<br />

1907 Autochrome plate<br />

1908 Ammonia<br />

1908 Geiger counter<br />

1908 Interchangeable parts<br />

1908 Oil-well drill bit<br />

1908 Vacuum cleaner<br />

1910 Radio crystal sets<br />

1910 Salvarsan<br />

1910 Washing machine<br />

1910-1939 Electric refrigerator<br />

1912 Color film<br />

1912 Diesel locomotive<br />

1912-1913 Solar thermal engine<br />

1912-1914 Artificial kidney<br />

1912-1915 X-ray crystallography<br />

909


910 / Time Line<br />

Date Invention<br />

1913 Assembly line<br />

1913 Geothermal power<br />

1913 Mammography<br />

1913 Thermal cracking process<br />

1915 Long-distance telephone<br />

1915 Propeller-coordinated machine gun<br />

1915 Pyrex glass<br />

1915 Long-distance radiotelephony<br />

1916-1922 Internal combustion engine<br />

1917 Food freezing<br />

1917 Sonar<br />

1919 Mass spectrograph<br />

1921 Tuberculosis vaccine<br />

1923 Rotary dial telephone<br />

1923 Television<br />

1923 <strong>and</strong> 1951 Syphilis test<br />

1924 Ultracentrifuge<br />

1925-1930 Differential analyzer<br />

1926 Buna rubber<br />

1926 Rocket<br />

1926 Talking motion pictures<br />

1927 Heat pump<br />

1928 Pap test<br />

1929 Electric clock<br />

1929 Electroencephalogram<br />

1929 Iron lung<br />

1930’s Contact lenses<br />

1930’s Vending machine slug rejector<br />

1930 Refrigerant gas<br />

1930 Typhus vaccine<br />

1930-1935 FM Radio<br />

1931 Cyclotron<br />

1931 Electron microscope<br />

1931 Neoprene<br />

1932 Fuel cell<br />

1932-1935 Antibacterial drugs


Date Invention<br />

1933-1954 Freeze-drying<br />

1934 Bathysphere<br />

1935 Nylon<br />

1935 Radar<br />

1935 Richter scale<br />

1936 Fluorescent lighting<br />

1937 Yellow fever vaccine<br />

1938 Polystyrene<br />

1938 Teflon<br />

1938 Xerography<br />

1940’s Carbon dating<br />

1940 Color television<br />

1940 Penicillin<br />

1940-1955 Microwave cooking<br />

1941 Polyester<br />

1941 Touch-tone telephone<br />

1941 Turbojet<br />

1942 Infrared photography<br />

1942-1950 Orlon<br />

1943 Aqualung<br />

1943 Colossus computer<br />

1943 Nuclear reactor<br />

1943-1946 ENIAC computer<br />

1944 Mark I calculator<br />

1944 V-2 rocket<br />

1945 Atomic bomb<br />

1945 Tupperware<br />

1946 Cloud seeding<br />

1946 Synchrocyclotron<br />

1947 Holography<br />

1948 Atomic clock<br />

1948 Broadcaster guitar<br />

1948 Instant photography<br />

1948-1960 Bathyscaphe<br />

1949 BINAC computer<br />

1949 Community antenna television<br />

Time Line / 911


912 / Time Line<br />

Date Invention<br />

1950 Cyclamate<br />

1950-1964 In vitro plant culture<br />

1951 Breeder reactor<br />

1951 UNIVAC computer<br />

1951-1952 Hydrogen bomb<br />

1952 Amniocentesis<br />

1952 Hearing aid<br />

1952 Polio vaccine (Salk)<br />

1952 Reserpine<br />

1952 Steelmaking process<br />

1952-1956 Field ion microscope<br />

1953 Artificial hormone<br />

1953 Heart-lung machine<br />

1953 Polyethylene<br />

1953 Synthetic amino acid<br />

1953 Transistor<br />

1953-1959 Hovercraft<br />

mid-1950’s Synthetic RNA<br />

1954 Photovoltaic cell<br />

1955 Radio interferometer<br />

1955-1957 FORTRAN programming language<br />

1956 Birth control pill<br />

1957 Artificial satellite<br />

1957 Nuclear power plant<br />

1957 Polio vaccine (Sabin)<br />

1957 Transistor radio<br />

1957 Velcro<br />

1957-1972 Pacemaker<br />

1958 Ultrasound<br />

1959 Atomic-powered ship<br />

1959 COBOL computer language<br />

1959 IBM Model 1401 computer<br />

1959 X-ray image intensifier<br />

1960’s Rice <strong>and</strong> wheat strains<br />

1960’s Virtual machine<br />

1960 Laser


Date Invention<br />

Time Line / 913<br />

1960 Memory metal<br />

1960 Telephone switching<br />

1960 Weather satellite<br />

1961 SAINT<br />

1962 Communications satellite<br />

1962 Laser eye surgery<br />

1962 Robot (industrial)<br />

1963 Cassette recording<br />

1964 Bullet train<br />

1964 Electronic synthesizer<br />

1964-1965 BASIC programming language<br />

1966 Tidal power plant<br />

1967 Coronary artery bypass surgery<br />

1967 Dolby noise reduction<br />

1967 Neutrino detector<br />

1967 Synthetic DNA<br />

1969 Bubble memory<br />

1969 The Internet<br />

1969-1983 Optical disk<br />

1970 Floppy disk<br />

1970 Videocassette recorder<br />

1970-1980 Virtual reality<br />

1972 CAT scanner<br />

1972 Pocket calculator<br />

1975-1979 Laser-diode recording process<br />

1975-1990 Fax machine<br />

1976 Supercomputer<br />

1976 Supersonic passenger plane<br />

1976-1988 Stealth aircraft<br />

1977 Apple II computer<br />

1977 Fiber-optics<br />

1977-1985 Cruise missile<br />

1978 Cell phone<br />

1978 Compressed-air-accumulating power plant<br />

1978 Nuclear magnetic resonance<br />

1978-1981 Scanning tunneling microscope


914 / Time Line<br />

Date Invention<br />

1979 Artificial blood<br />

1979 Walkman cassette player<br />

1980’s CAD/CAM<br />

1981 Personal computer<br />

1982 Abortion pill<br />

1982 Artificial heart<br />

1982 Genetically engineered insulin<br />

1982 Robot (household)<br />

1983 Artificial chromosome<br />

1983 Aspartame<br />

1983 Compact disc<br />

1983 Hard disk<br />

1983 Laser vaporization<br />

1985 Genetic “fingerprinting”<br />

1985 Tevatron accelerator<br />

1997 Cloning<br />

2000 Gas-electric car


Topics by Category<br />

Topics by Category<br />

Agriculture<br />

Artificial insemination<br />

Cloning<br />

Cloud seeding<br />

In vitro plant culture<br />

Rice <strong>and</strong> wheat strains<br />

Astronomy<br />

Artificial satellite<br />

Communications satellite<br />

Neutrino detector<br />

Radio interferometer<br />

Weather satellite<br />

Aviation <strong>and</strong> space<br />

Airplane<br />

Artificial satellite<br />

Communications satellite<br />

Dirigible<br />

Radio interferometer<br />

Rocket<br />

Stealth aircraft<br />

Turbojet<br />

V-2 rocket<br />

Weather satellite<br />

Biology<br />

Artificial chromosome<br />

Artificial insemination<br />

Cloning<br />

Genetic “fingerprinting”<br />

In vitro plant culture<br />

Synthetic amino acid<br />

Synthetic DNA<br />

Synthetic RNA<br />

Ultracentrifuge<br />

Chemistry<br />

Ammonia<br />

Fuel cell<br />

Refrigerant gas<br />

Silicones<br />

Thermal cracking process<br />

Ultracentrifuge<br />

Ultramicroscope<br />

Vat dye<br />

X-ray crystallography<br />

Communications<br />

915<br />

Cassette recording<br />

Cell phone<br />

Color television<br />

Communications satellite<br />

Community antenna television<br />

Dolby noise reduction<br />

Electronic synthesizer<br />

Fax machine<br />

Fiber-optics<br />

FM radio<br />

Hearing aid<br />

Laser-diode recording process<br />

Long-distance radiotelephony<br />

Long-distance telephone


916 / Topics by Category<br />

Radar<br />

Radio<br />

Radio crystal sets<br />

Rotary dial telephone<br />

Sonar<br />

Talking motion pictures<br />

Telephone switching<br />

Television<br />

Touch-tone telephone<br />

Transistor radio<br />

Vacuum tube<br />

Videocassette recorder<br />

Xerography<br />

Computer science<br />

Apple II computer<br />

BASIC programming language<br />

BINAC computer<br />

Bubble memory<br />

COBOL computer language<br />

Colossus computer<br />

Computer chips<br />

Differential analyzer<br />

ENIAC computer<br />

Floppy disk<br />

FORTRAN programming<br />

language<br />

Hard disk<br />

IBM Model 1401 computer<br />

Internet<br />

Mark I calculator<br />

Optical disk<br />

Personal computer<br />

Pocket calculator<br />

SAINT<br />

Supercomputer<br />

UNIVAC computer<br />

Virtual machine<br />

Virtual reality<br />

Consumer products<br />

Apple II computer<br />

Aspartame<br />

Birth control pill<br />

Broadcaster guitar<br />

Brownie camera<br />

Cassette recording<br />

Cell phone<br />

Color film<br />

Color television<br />

Compact disc<br />

Cyclamate<br />

Disposable razor<br />

Electric refrigerator<br />

FM radio<br />

Gas-electric car<br />

Hearing aid<br />

Instant photography<br />

Internet<br />

Nylon<br />

Orlon<br />

Personal computer<br />

Pocket calculator<br />

Polyester<br />

Pyrex glass<br />

Radio<br />

Rotary dial telephone<br />

Teflon<br />

Television<br />

Touch-tone telephone<br />

Transistor radio<br />

Tupperware<br />

Vacuum cleaner<br />

Velcro


Videocassette recorder<br />

Walkman cassette player<br />

Washing machine<br />

Drugs <strong>and</strong> vaccines<br />

Abortion pill<br />

Antibacterial drugs<br />

Artificial hormone<br />

Birth control pill<br />

Genetically engineered insulin<br />

Penicillin<br />

Polio vaccine (Sabin)<br />

Polio vaccine (Salk)<br />

Reserpine<br />

Salvarsan<br />

Tuberculosis vaccine<br />

Typhus vaccine<br />

Yellow fever vaccine<br />

Earth science<br />

Aqualung<br />

Bathyscaphe<br />

Bathysphere<br />

Cloud seeding<br />

Richter scale<br />

X-ray crystallography<br />

Electronics<br />

Cassette recording<br />

Cell phone<br />

Color television<br />

Communications satellite<br />

Compact disc<br />

Dolby noise reduction<br />

Electronic synthesizer<br />

Fax machine<br />

Fiber-optics<br />

FM radio<br />

Hearing aid<br />

Laser-diode recording process<br />

Long-distance radiotelephony<br />

Long-distance telephone<br />

Radar<br />

Radio<br />

Radio crystal sets<br />

Rotary dial telephone<br />

Sonar<br />

Telephone switching<br />

Television<br />

Touch-tone telephone<br />

Transistor<br />

Transistor radio<br />

Vacuum tube<br />

Videocassette recorder<br />

Walkman cassette player<br />

Xerography<br />

Energy<br />

Topics by Category / 917<br />

Alkaline storage battery<br />

Breeder reactor<br />

Compressed-air-accumulating<br />

power plant<br />

Fluorescent lighting<br />

Fuel cell<br />

Gas-electric car<br />

Geothermal power<br />

Heat pump<br />

Nuclear power plant<br />

Nuclear reactor<br />

Oil-well drill bit<br />

Photoelectric cell<br />

Photovoltaic cell


918 / Topics by Category<br />

Solar thermal engine<br />

Tidal power plant<br />

Vacuum tube<br />

Engineering<br />

Airplane<br />

Assembly line<br />

Bullet train<br />

CAD/CAM<br />

Differential analyzer<br />

Dirigible<br />

ENIAC computer<br />

Gas-electric car<br />

Internal combustion engine<br />

Oil-well drill bit<br />

Robot (household)<br />

Robot (industrial)<br />

Steelmaking process<br />

Tidal power plant<br />

Vacuum cleaner<br />

Washing machine<br />

Exploration<br />

Carbon dating<br />

Aqualung<br />

Bathyscaphe<br />

Bathysphere<br />

Neutrino detector<br />

Radar<br />

Radio interferometer<br />

Sonar<br />

Food science<br />

Aspartame<br />

Cyclamate<br />

Electric refrigerator<br />

Food freezing<br />

Freeze-drying<br />

Genetically engineered insulin<br />

In vitro plant culture<br />

Microwave cooking<br />

Polystyrene<br />

Refrigerant gas<br />

Rice <strong>and</strong> wheat strains<br />

Teflon<br />

Tupperware<br />

Genetic engineering<br />

Amniocentesis<br />

Artificial chromosome<br />

Artificial insemination<br />

Cloning<br />

Genetic “fingerprinting”<br />

Genetically engineered insulin<br />

In vitro plant culture<br />

Rice <strong>and</strong> wheat strains<br />

Synthetic amino acid<br />

Synthetic DNA<br />

Synthetic RNA<br />

Home products<br />

Cell phone<br />

Color television<br />

Community antenna television<br />

Disposable razor<br />

Electric refrigerator<br />

Fluorescent lighting<br />

FM radio<br />

Microwave cooking<br />

Radio<br />

Refrigerant gas


Robot (household)<br />

Rotary dial telephone<br />

Television<br />

Touch-tone Telephone<br />

Transistor radio<br />

Tungsten filament<br />

Tupperware<br />

Vacuum cleaner<br />

Videocassette recorder<br />

Washing machine<br />

Manufacturing<br />

Assembly line<br />

CAD/CAM<br />

Interchangeable parts<br />

Memory metal<br />

Polystyrene<br />

Steelmaking process<br />

Materials<br />

Buna rubber<br />

Contact lenses<br />

Disposable razor<br />

Laminated glass<br />

Memory metal<br />

Neoprene<br />

Nylon<br />

Orlon<br />

Plastic<br />

Polyester<br />

Polyethylene<br />

Polystyrene<br />

Pyrex glass<br />

Silicones<br />

Steelmaking process<br />

Teflon<br />

Topics by Category / 919<br />

Tungsten filament<br />

Velcro<br />

Measurement <strong>and</strong> detection<br />

Amniocentesis<br />

Atomic clock<br />

Carbon dating<br />

CAT scanner<br />

Cyclotron<br />

Electric clock<br />

Electrocardiogram<br />

Electroencephalogram<br />

Electron microscope<br />

Geiger counter<br />

Gyrocompass<br />

Mass spectrograph<br />

Neutrino detector<br />

Radar<br />

Sonar<br />

Radio interferometer<br />

Richter scale<br />

Scanning tunneling microscope<br />

Synchrocyclotron<br />

Tevatron accelerator<br />

Ultracentrifuge<br />

Ultramicroscope<br />

Vending machine slug rejector<br />

X-ray crystallography<br />

Medical procedures<br />

Amniocentesis<br />

Blood transfusion<br />

CAT scanner<br />

Cloning<br />

Coronary artery bypass surgery<br />

Electrocardiogram


920 / Topics by Category<br />

Electroencephalogram<br />

Heart-lung machine<br />

Iron lung<br />

Laser eye surgery<br />

Laser vaporization<br />

Mammography<br />

Nuclear magnetic resonance<br />

Pap test<br />

Syphilis test<br />

Ultrasound<br />

X-ray image intensifier<br />

Medicine<br />

Abortion pill<br />

Amniocentesis<br />

Antibacterial drugs<br />

Artificial blood<br />

Artificial heart<br />

Artificial hormone<br />

Artificial kidney<br />

Birth control pill<br />

Blood transfusion<br />

CAT scanner<br />

Contact lenses<br />

Coronary artery bypass surgery<br />

Electrocardiogram<br />

Electroencephalogram<br />

Genetically engineered insulin<br />

Hearing aid<br />

Heart-lung machine<br />

Iron lung<br />

Laser eye surgery<br />

Laser vaporization<br />

Mammography<br />

Nuclear magnetic resonance<br />

Pacemaker<br />

Pap test<br />

Penicillin<br />

Polio vaccine (Sabin)<br />

Polio vaccine (Salk)<br />

Reserpine<br />

Salvarsan<br />

Syphilis test<br />

Tuberculosis vaccine<br />

Typhus vaccine<br />

Ultrasound<br />

X-ray image intensifier<br />

Yellow fever vaccine<br />

Music<br />

Broadcaster guitar<br />

Cassette recording<br />

Dolby noise reduction<br />

Electronic synthesizer<br />

FM Radio<br />

Radio<br />

Transistor radio<br />

Photography<br />

Autochrome plate<br />

Brownie camera<br />

Color film<br />

Electrocardiogram<br />

Electron microscope<br />

Fax machine<br />

Holography<br />

Infrared photography<br />

Instant photography<br />

Mammography<br />

Mass spectrograph<br />

Optical disk<br />

Talking motion pictures<br />

Weather satellite


Xerography<br />

X-ray crystallography<br />

Physics<br />

Atomic bomb<br />

Cyclotron<br />

Electron microscope<br />

Field ion microscope<br />

Geiger counter<br />

Hydrogen bomb<br />

Holography<br />

Laser<br />

Mass spectrograph<br />

Scanning tunneling microscope<br />

Synchrocyclotron<br />

Tevatron accelerator<br />

X-ray crystallography<br />

Synthetics<br />

Artificial blood<br />

Artificial chromosome<br />

Artificial heart<br />

Artificial hormone<br />

Artificial insemination<br />

Artificial kidney<br />

Artificial satellite<br />

Aspartame<br />

Buna rubber<br />

Cyclamate<br />

Electronic synthesizer<br />

Genetically engineered insulin<br />

Neoprene<br />

Topics by Category / 921<br />

Synthetic amino acid<br />

Synthetic DNA<br />

Synthetic RNA<br />

Vat dye<br />

Transportation<br />

Airplane<br />

Atomic-powered ship<br />

Bullet train<br />

Diesel locomotive<br />

Dirigible<br />

Gas-electric car<br />

Gyrocompass<br />

Hovercraft<br />

Internal combustion engine<br />

Supersonic passenger plane<br />

Turbojet<br />

Weapons technology<br />

Airplane<br />

Atomic bomb<br />

Cruise missile<br />

Dirigible<br />

Hydrogen bomb<br />

Propeller-coordinated machine<br />

gun<br />

Radar<br />

Rocket<br />

Sonar<br />

Stealth aircraft<br />

V-2 rocket


This Page Intentionally Left Blank


Index<br />

Index<br />

Abbe, Ernst, 678<br />

ABC. See American Broadcasting<br />

Company<br />

Abel, John Jacob, 50, 58, 60<br />

Abortion pill, 1-5<br />

Adams, Ansel, 430<br />

Adams, Thomas, 850<br />

Advanced Research Projects Agency,<br />

446-447<br />

AHD. See Audio high density disc<br />

Aiken, Howard H., 187, 417, 490, 828<br />

Airplane, 6-10<br />

Aldrin, Edwin, 8<br />

Alferov, Zhores I., 320-321<br />

Alkaline storage battery, 11-15<br />

Ambrose, James, 167<br />

American Broadcasting Company, 215<br />

American Telephone <strong>and</strong> Telegraph<br />

Company, 741<br />

Amery, Julian, 714<br />

Amino acid, synthetic, 724-728<br />

Ammonia, 16-19; <strong>and</strong> atomic clock, 81-<br />

82; as a refrigerant, 290-291, 345,<br />

631, 746<br />

Amniocentesis, 20-23<br />

Anable, Gloria Hollister, 100<br />

Anschütz-Kaempfe, Hermann, 382<br />

Antibacterial drugs, 24-27<br />

Antibiotics, 24-27, 47, 813; penicillin,<br />

553-557, 676, 738<br />

Apple II computer, 28-32<br />

Appliances. See Electric clock; Electric<br />

refrigerator; Microwave cooking;<br />

Refrigerant gas; Robot (household);<br />

Vacuum cleaner; Washing machine<br />

Aqualung, 33-37<br />

Archaeology, 158-162<br />

Archimedes, 687<br />

Armstrong, Edwin H., 339<br />

Armstrong, Neil, 8<br />

Arnold, Harold D., 477<br />

Arnold, Henry Harley, 807<br />

ARPAnet, 447-448<br />

Arsonval, Jacques Arsène d’, 351<br />

Arteries <strong>and</strong> laser vaporization, 472-476<br />

Artificial blood, 38-40<br />

Artificial chromosome, 41-44<br />

Artificial heart, 45-49<br />

Artificial hormone, 50-53<br />

Artificial insemination, 54-57<br />

Artificial intelligence, 668, 671, 864<br />

Artificial kidney, 58-62<br />

Artificial satellite, 63-66<br />

Artificial sweeteners, 67-70;<br />

Aspartame, 67-70; cyclamates, 248-<br />

251<br />

ASCC. See Automatic Sequence<br />

Controlled Calculator<br />

Aspartame, 67-70<br />

Assembly line, 71-75, 197, 434, 436, 439<br />

Aston, Francis William, 494, 496<br />

Astronauts, 749, 848<br />

AT&T. See American Telephone <strong>and</strong><br />

Telegraph Company<br />

Atanasoff, John Vincent, 312<br />

Atomic bomb, 76-79, 84, 118-119, 255,<br />

412, 414, 521, 525, 697, 721<br />

Atomic clock, 80-83<br />

Atomic Energy Commission, 119, 521,<br />

523<br />

Atomic force microscope, 681<br />

Atomic mass, 494-497<br />

Atomic-powered ship, 84, 86-87<br />

Audiffren, Marcel, 289<br />

Audio high density disc, 220<br />

Audrieth, Ludwig Frederick, 67<br />

Autochrome plate, 88-91<br />

Automatic Sequence Controlled<br />

Calculator, 187<br />

Automobiles; <strong>and</strong> assembly lines, 71,<br />

75; <strong>and</strong> interchangeable parts, 434-<br />

441; <strong>and</strong> internal combustion<br />

engine, 442-445<br />

Avery, Oswald T., 733<br />

Aviation. See Airplane; Dirigible;<br />

Rockets; Stealth aircraft; Supersonic<br />

passenger plane; Turbojet<br />

Babbage, Charles, 417<br />

Backus, John, 347<br />

923


924 / Index<br />

Bacon, Francis Thomas, 355, 358<br />

Baekel<strong>and</strong>, Leo Hendrik, 571<br />

Baeyer, Adolf von, 571<br />

Bahcall, John Norris, 511<br />

Bain, Alex<strong>and</strong>er, 316<br />

Baker, William Oliver, 172, 174<br />

Banting, Frederick G., 375<br />

Baran, Paul, 446<br />

Bardeen, John, 782, 786, 789<br />

Barnay, Antoine, 663<br />

Barton, Otis, 95, 100<br />

BASIC computer language, 29-30, 92-<br />

94, 559<br />

Bathyscaphe, 95-99<br />

Bathysphere, 100-103<br />

Batteries, 11, 227; alkaline storage, 11-<br />

15; <strong>and</strong> electric cars, 360, 363; <strong>and</strong><br />

fuel cells, 356; <strong>and</strong> hearing aids, 390,<br />

392; <strong>and</strong> pacemakers, 547; silicon<br />

solar, 569; <strong>and</strong> transistor radios, 780,<br />

875, 878<br />

Battery jars, 454, 607<br />

Baulieu, Étienne-Émile, 1-2<br />

Bavolek, Cecelia, 394<br />

Bazooka, 659<br />

BCS theory, 789<br />

Beams, Jesse W., 815<br />

Becquerel, Alex<strong>and</strong>re-Edmond, 562<br />

Becquerel, Antoine-Henri, 365<br />

Beebe, William, 95, 100<br />

Bélanger, Alain, 1<br />

Bell, Alex<strong>and</strong>er Graham, 320-322, 390,<br />

482-483, 663-665<br />

Bell Telephone Laboratories, 101, 138-<br />

140, 172-173, 204-205, 217, 229-230,<br />

323, 390-391, 482, 567, 614, 625, 678,<br />

744, 752, 774-775, 778-779, 786, 829,<br />

840, 861, 863, 876<br />

Belzel, George, 558<br />

Benedictus, Edouard, 454<br />

Bennett, Frederick, 434<br />

Bennett, W. R., 217<br />

Berger, Hans, 298<br />

Bergeron, Tor, 183<br />

Berliner, Emil, 279<br />

Berthelot, Marcellin Pierre, 597<br />

Bessemer, Henry, 701, 704<br />

Bessemer converter, 701-702, 704<br />

Best, Charles H., 375<br />

Bethe, Hans, 412, 720<br />

Bevis, Douglas, 20<br />

Billiard balls, 572-573<br />

BINAC. See Binary Automatic<br />

Computer<br />

Binary Automatic Computer, 104-107,<br />

315, 330, 348<br />

Binnig, Gerd, 678, 680<br />

Birdseye, Clarence, 343<br />

Birth control pill, 108-112<br />

Bissell, Melville R., 832<br />

Blodgett, Katherine Ann, 454<br />

Blood plasma, 38<br />

Blood transfusion, 113-117<br />

Bobeck, Andrew H., 138-139<br />

Bohn, René, 842<br />

Bohr, Niels, 76, 520, 695<br />

Bolton, Elmer Keiser, 507, 529<br />

Booth, Andrew D., 330<br />

Booth, H. Cecil, 832, 835<br />

Borlaug, Norman E., 638, 643<br />

Borsini, Fred, 151<br />

Bothe, Walter, 367<br />

Bragg, Lawrence, 896, 898<br />

Bragg, William Henry, 896, 898<br />

Brain, <strong>and</strong> nuclear magnetic<br />

resonance, 516, 519<br />

Brattain, Walter H., 782, 786, 789<br />

Braun, Wernher von, 63, 871<br />

Bravais, Auguste, 896<br />

Breast cancer, 486, 489<br />

Breeder reactor, 118-121<br />

Broadcaster guitar, 122-129<br />

Broadcasting. See FM radio; Radio;<br />

Radio crystal sets; Television;<br />

Transistor radio<br />

Broglie, Louis de, 302, 678<br />

Brooks, Fred P., 866<br />

Brownell, Frank A., 130<br />

Brownie camera, 130-137<br />

Bubble memory, 138-141<br />

Buehler, William, 498<br />

Bullet train, 142-145<br />

Buna rubber, 146-150<br />

Burks, Arthur Walter, 312<br />

Burton, William M., 765<br />

Busch, Adolphus, 259<br />

Busch, Hans, 302, 679<br />

Bush, Vannevar, 262, 264


CAD. See Computer-Aided Design<br />

CAD/CAM, 151-157<br />

Calculators; desktop, 232; digital, 490-<br />

493; electromechanical, 313;<br />

mechanical, 104; pocket, 576-580;<br />

punched-card, 104<br />

California Institute of Technology, 646,<br />

731, 782<br />

Callus tissue, 421<br />

Calmette, Albert, 791<br />

Cameras; Brownie, 130-137; <strong>and</strong> film,<br />

88-91, 192-195; <strong>and</strong> infrared film,<br />

426; instant, 430-433; in space, 887-<br />

889; video, 165, 859; <strong>and</strong> virtual<br />

reality, 867; <strong>and</strong> X rays, 901-904. See<br />

also Photography<br />

Campbell, Charles J., 468<br />

Campbell, Keith H. S., 177<br />

Cancer, 4, 324, 376; <strong>and</strong> cyclamates, 69,<br />

249-250; <strong>and</strong> infrared photography,<br />

428; <strong>and</strong> mammography, 486-489;<br />

therapy, 40; uterine, 549-552<br />

Capek, Karel, 650, 654<br />

Carbohydrates, 374<br />

Carbon dating, 158-162<br />

Carlson, Chester F., 891, 893<br />

Carnot, Sadi, 398<br />

Carothers, Wallace H., 507, 510, 529,<br />

574, 589<br />

Carrel, Alexis, 113<br />

Carty, John J., 477, 484<br />

Cary, Frank, 558<br />

Cascariolo, Vincenzo, 335<br />

Cassette recording, 163-166, 221, 223,<br />

279, 538; <strong>and</strong> Dolby noise reduction,<br />

282; <strong>and</strong> microcomputers, 386; <strong>and</strong><br />

Sony Walkman, 788; <strong>and</strong> transistors,<br />

784<br />

CAT scanner, 167-171<br />

Cathode-ray tubes, 170, 303, 315, 326,<br />

564, 611; <strong>and</strong> television, 757-758,<br />

760, 837<br />

Caton, Richard, 298<br />

CBS. See Columbia Broadcasting<br />

System<br />

CD. See Compact disc<br />

CDC. See Control Data Corporation<br />

Cell phone, 172-176<br />

Celluloid, 454, 571-573<br />

Centrifuge, 815-818<br />

Index / 925<br />

Cerf, Vinton G., 446, 448<br />

Chadwick, James, 367<br />

Chain, Ernst Boris, 553<br />

Chamberlain, W. Edward, 901<br />

Chance, Ronald E., 374<br />

Ch<strong>and</strong>ler, Robert F., Jr., 638<br />

Chang, Min-Chueh, 108<br />

Chanute, Octave, 6<br />

Chapin, Daryl M., 567<br />

Chardonnet, Hilaire de, 589<br />

Chemotherapy, 24, 40, 676<br />

Cho, Fujio, 360<br />

Christian, Charlie, 122, 126<br />

Chromosomes. See Artificial<br />

chromosome<br />

Clark, Barney, 45<br />

Clark, Dugold, 257<br />

Clarke, Arthur C., 63, 204<br />

Cloning, 177-182<br />

Cloud seeding, 183-186<br />

Coal tars, 593, 843<br />

COBOL computer language, 92, 187-<br />

191, 350<br />

Cockerell, Christopher Sydney, 407<br />

Cohen, Robert Waley, 442<br />

Coleman, William T., Jr., 714<br />

Collins, Arnold Miller, 507<br />

Color film, 192-195<br />

Color photography, 88-91<br />

Color television, 196-199<br />

Colossus computer, 200-203<br />

Columbia Broadcasting System, 196,<br />

215, 830<br />

Communications satellite, 204-207<br />

Community antenna television, 208-216<br />

Compaan, Klaas, 537<br />

Compact disc, 217-224<br />

Compressed-air-accumulating power<br />

plant, 225-228<br />

Computer-Aided Design (CAD), 151-<br />

157<br />

Computer chips, 140, 229-234<br />

Computer languages, 154; ALGOL, 92-<br />

93; BASIC, 29-30, 92-94, 559;<br />

COBOL, 92, 187-191, 350;<br />

FORTRAN, 92-93, 189, 347-350<br />

Computerized axial tomography, 167-<br />

171<br />

Computers; <strong>and</strong> information storage,


926 / Index<br />

104-107, 138-141, 165, 330-334, 386-<br />

389, 537-540; <strong>and</strong> Internet, 446-450.<br />

See also Apple II computer; Personal<br />

computers<br />

Concorde, 714-719<br />

Condamine, Charles de la, 146<br />

Contact lenses, 235-239<br />

Conti, Piero Ginori, 378<br />

Contraception, 1-5, 108-112<br />

Control Data Corporation, 709, 711<br />

Cooking; microwave, 502-506; <strong>and</strong><br />

Pyrex glass, 607, 609; <strong>and</strong> Tefloncoating,<br />

748-749<br />

Coolidge, William David, 795<br />

Cormack, Allan M., 167<br />

Corning Glass Works, 323, 606-610<br />

Coronary artery bypass surgery, 240-<br />

243<br />

Cource, Geoffroy de, 714, 716<br />

Cousins, Morison, 799<br />

Cousteau, Jacques-Yves, 33, 35, 102<br />

Cray, Seymour R., 709, 711-712<br />

Crick, Francis, 41, 177, 729, 733<br />

Crile, George Washington, 113<br />

Critical mass, 77, 119, 521<br />

Crookes, William, 365<br />

CRT. See Cathode-ray tubes<br />

Cruise missile, 244-247<br />

Curie, Jacques, 692<br />

Curie, Marie, 823<br />

Curie, Pierre, 692, 695<br />

Curtis, William C., 611-612<br />

Cyclamate, 248-251<br />

Cyclotron, 252-256<br />

DAD. See Digital audio disc<br />

Daimler, Gottlieb, 257<br />

Dale, Henry Hallett, 50<br />

Damadian, Raymond, 516<br />

Datamath, 579<br />

Davis, Raymond, Jr., 511<br />

Deep-sea diving, 95-103<br />

De Forest, Lee, 477-478, 480, 483, 837-<br />

838<br />

Dekker, Wisse, 217<br />

Deoxyribonucleic acid; characteristics,<br />

733. See also DNA<br />

Depp, Wallace Andrew, 751<br />

Desert Storm, 699<br />

Devol, George C., Jr., 654<br />

DeVries, William Castle, 45-46<br />

Diabetes, 51-52, 374-377<br />

Dichlorodifluoromethane, 630-633<br />

Diesel, Rudolf, 257-258<br />

Diesel locomotive, 257-261<br />

Differential analyzer, 262-266<br />

Digital audio disc, 219<br />

Dirigible, 267-271<br />

Disposable razor, 272-278<br />

Diving. See Aqualung<br />

DNA, 41, 177; <strong>and</strong> artificial<br />

chromosomes, 41-44; <strong>and</strong> cloning,<br />

177; <strong>and</strong> genetic “fingerprinting,”<br />

370-373; recombinant, 41; synthetic,<br />

729-732; <strong>and</strong> X-ray crystallography,<br />

900. See also Deoxyribonucleic acid;<br />

Synthetic DNA<br />

Dolby, Ray Milton, 279, 281<br />

Dolby noise reduction, 279-283<br />

Domagk, Gerhard, 24<br />

Donald, Ian T., 823-824<br />

Dornberger, Walter Robert, 871<br />

Drew, Charles, 113, 115<br />

Drinker, Philip, 451<br />

Dulbecco, Renato, 581<br />

Dunwoody, H. H., 621<br />

Du Pont. See Du Pont de Nemours <strong>and</strong><br />

Company<br />

Du Pont de Nemours <strong>and</strong> Company,<br />

77, 149, 248, 508-509, 529, 531, 542,<br />

589-590, 746-748, 799, 803<br />

Durfee, Benjamin M., 490<br />

Durham, Eddie, 126<br />

Durrer, Robert, 701<br />

Dyes, 593; <strong>and</strong> acrylics, 543; <strong>and</strong><br />

infrared radiation, 425, 428; <strong>and</strong><br />

microorganism staining, 24-25; <strong>and</strong><br />

photographic film, 192-194; poison,<br />

674; <strong>and</strong> polyesters, 591; vat, 842-<br />

845<br />

Earthquakes, measuring of, 645-649<br />

Eastman, George, 130, 135<br />

Eckert, John Presper, 104, 312, 828<br />

Edison, Thomas Alva, 11, 335, 479, 616,<br />

744, 839; <strong>and</strong> batteries, 12-14; <strong>and</strong><br />

Edison effect, 837; <strong>and</strong> electric light,<br />

795, 832; <strong>and</strong> fluoroscope, 901; <strong>and</strong><br />

phonograph, 217, 279


Edison effect, 837-838<br />

Edlefsen, Niels, 252<br />

EDVAC. See Electronic Discrete<br />

Variable Automatic Computer<br />

Effler, Donald B., 240<br />

Ehrlich, Paul, 24, 673<br />

Einstein, Albert, 82, 472, 497, 563, 695,<br />

721<br />

Einthoven, Willem, 293, 295<br />

Eisenhower, Dwight D., 84, 415<br />

Elastomers, 148, 507-510, 598<br />

Electric clock, 284-288<br />

Electric refrigerator, 289-292<br />

Electricity, generation of, 79, 378, 569<br />

Electrocardiogram, 293-297<br />

Electroencephalogram, 298-301<br />

Electrolyte detector, 479<br />

Electron microscope, 302-306, 403, 902<br />

Electron theory, 562-565<br />

Electronic Discrete Variable Automatic<br />

Computer, 105-107, 314, 829<br />

Electronic Numerical Integrator <strong>and</strong><br />

Calculator, 105-106, 312-315, 347,<br />

668, 829<br />

Electronic synthesizer, 307-311<br />

Eli Lilly Research Laboratories, 374-377<br />

Elliott, Tom, 360<br />

Elmquist, Rune, 545<br />

Elster, Julius, 562, 564<br />

Engelberger, Joseph F., 654<br />

ENIAC. See Electronic Numerical<br />

Integrator <strong>and</strong> Calculator<br />

Ericsson, John, 687<br />

Espinosa, Chris, 28<br />

Estridge, Philip D., 386<br />

Evans, Oliver, 71<br />

Ewan, Harold Irving, 625<br />

“Excalibur,” 416<br />

Eyeglasses; <strong>and</strong> contact lenses, 235-239<br />

; frames, 498, 500; <strong>and</strong> hearing aids,<br />

391, 787<br />

Fabrics; <strong>and</strong> dyes, 842-845; orlon, 541-<br />

544; polyester, 589-592; <strong>and</strong> washing<br />

machines, 883-886<br />

Fahlberg, Constantin, 67<br />

Fasteners, velcro, 846-849<br />

Favaloro, Rene, 240<br />

Fax machine, 316-319<br />

Index / 927<br />

FCC. See Federal Communications<br />

Commission<br />

Federal Communications Commission;<br />

<strong>and</strong> cell phones, 173, 175; <strong>and</strong><br />

communication satellites, 204; <strong>and</strong><br />

FM radio, 341; <strong>and</strong> microwave<br />

cooking, 505; <strong>and</strong> television, 196-<br />

197, 208-210<br />

Fefrenkiel, Richard H., 172<br />

Feinbloom, William, 235, 237<br />

Fender, Leo, 122<br />

Ferguson, Charles Wesley, 158<br />

Fermi, Enrico, 76, 84, 412, 520, 525<br />

Fessenden, Reginald, 13, 477-480, 616-<br />

618<br />

Fiber-optics, 320-324<br />

Fick, Adolf Eugen, 235<br />

Field ion microscope, 325-329, 679<br />

FIM. See Field ion microscope<br />

Finlay, Carlos J., 905<br />

Fischer, Rudolf, 192<br />

Fisher, Alva J., 883<br />

Fleming, Alex<strong>and</strong>er, 553, 555<br />

Fleming, John Ambrose, 478, 621, 837,<br />

839<br />

Flick, J. B., 394<br />

Floppy disk, 330-334<br />

Florey, Baron, 553<br />

Flosdorf, Earl W., 351<br />

Flowers, Thomas H., 200<br />

FLOW-MATIC, 187<br />

Fluorescent lighting, 335-338<br />

FM radio, 339-342<br />

Fokker, Anthony Herman Gerard, 601,<br />

603<br />

Food; artificial sweeteners, 67-70, 248-<br />

251; freeze-drying, 351-354;<br />

freezing, 343-346; microwave<br />

cooking, 502-506; packaging, 598-<br />

599; <strong>and</strong> refrigeration, 289-292, 343-<br />

346, 630, 632; rice <strong>and</strong> wheat, 638-<br />

644; storage, 799-806<br />

Food <strong>and</strong> Drug Administration, 45,<br />

111, 375<br />

Ford, Henry, 11, 71, 74, 257, 434<br />

Forel, François-Alphonse, 645<br />

Forest de Bélidor, Bernard, 770<br />

FORTRAN programming language,<br />

92-93, 189, 347-350<br />

Foucault, Jean-Bernard-Léon, 382


928 / Index<br />

Fox Network, 215<br />

Francis, Thomas, Jr., 585<br />

Freeze-drying, 351-354<br />

Frerichs, Friedrick von, 673<br />

Frisch, Otto Robert, 76, 520<br />

Fuchs, Klaus Emil Julius, 412<br />

Fuel cell, 355-359<br />

Fuller, Calvin S., 567<br />

Fulton, Robert, 335<br />

Gabor, Dennis, 402, 404<br />

Gagarin, Yuri A., 874<br />

Gagnan, Émile, 33, 102<br />

Gamow, George, 325, 412, 414, 720-721<br />

Garcia, Celso-Ramon, 108<br />

Garros, Rol<strong>and</strong>, 601<br />

Garwin, Richard L., 414<br />

Gas-electric car, 360-364<br />

Gates, Bill, 92, 94<br />

Gaud, William S., 638<br />

Gautheret, Roger, 421<br />

GE. See General Electric Company<br />

Geiger, Hans, 365, 367<br />

Geiger counter, 365-369<br />

Geissler, Heinrich, 335<br />

Geitel, Hans Friedrich, 562, 564<br />

General Electric Company, 101, 183-<br />

185, 219, 264, 290, 341, 356, 384, 440,<br />

455, 477, 617, 683, 685, 795-796, 809,<br />

840, 863, 893, 902<br />

Genetic “fingerprinting,” 370-373<br />

Genetically engineered insulin, 374-377<br />

Geothermal power, 378-381<br />

Gerhardt, Charles, 597<br />

Gershon-Cohen, Jacob, 486<br />

Gibbon, John H., Jr., 394<br />

Gibbon, Mary Hopkinson, 394<br />

Gillette, George, 272<br />

Gillette, King Camp, 272, 276<br />

Glass; coloring of, 819-820; fibers, 322-<br />

323, 591; food containers, 800; goldruby,<br />

819; high-purity, 322;<br />

laminated, 454-458; Pyrex, 606-610<br />

Glass fiber. See Fiber-optics<br />

Goddard, Robert H., 63, 65, 658, 660,<br />

662<br />

Goldmark, Peter Carl, 196<br />

Goldstine, Herman Heine, 312, 347<br />

Goodman, Benny, 126<br />

Goodyear, Charles, 146-147, 335<br />

Gosslau, Ing Fritz, 871<br />

Gould, R. Gordon, 472<br />

Goulian, Mehran, 729<br />

Graf Zeppelin, 271<br />

Gray, Elisha, 663<br />

Greaves, Ronald I. N., 351<br />

Green Revolution, 638-639, 641-644<br />

Grove, William Robert, 355<br />

Groves, Leslie R., 76, 747<br />

Grunberg-Manago, Marianne, 733<br />

Guérin, Camille, 791<br />

Guitar, electric, 122-129<br />

Gutenberg, Beno, 645<br />

Gyrocompass, 382-385<br />

Haas, Georg, 59<br />

Haber, Fritz, 16-19<br />

Haberl<strong>and</strong>t, Gottlieb, 421<br />

Hahn, Otto, 84, 520<br />

Haldane, John Burdon S<strong>and</strong>erson, 724<br />

Haldane, T. G. N., 398<br />

Hall, Charles, 335<br />

Halliday, Don, 151<br />

Hallwachs, Wilhelm, 562<br />

Hamilton, Francis E., 490<br />

Hammond, John, 126<br />

Hanratty, Patrick, 151<br />

Hard disk, 386-389<br />

Hata, Sahachiro, 673<br />

Haüy, René-Just, 896<br />

Hayato, Ikeda, 142<br />

Hayes, Arthur H., Jr., 67<br />

Hazen, Harold L., 262<br />

Health Company, 650<br />

Hearing aid, 390-393<br />

Heart; <strong>and</strong> pacemakers, 545-548. See<br />

also Artificial heart<br />

Heart-lung machine, 394-397<br />

Heat pump, 398-401<br />

Heilborn, Jacob, 272<br />

Henne, Albert, 630, 746<br />

Hero (Greek mathematician), 851<br />

Hero 1 robot, 650-653<br />

Herschel, William, 425, 427<br />

Hertz, Heinrich, 502, 621<br />

Heumann, Karl, 842<br />

Hewitt, Peter Cooper, 335


Hindenburg, 271<br />

Hitler, Adolf, 414, 509, 807, 871<br />

Hoff, Marcian Edward, Jr., 229<br />

Hoffman, Frederick de, 413<br />

Hoffmann, Erich, 676<br />

Hofmann, August Wilhelm von, 593,<br />

842, 844<br />

Hollerith, Herman, 417<br />

Holography, 402-406, 537<br />

Homolka, Benno, 192<br />

Honda Insight, 360<br />

Hoover, Charles Wilson, Jr., 751<br />

Hoover, William Henry, 832<br />

Hopper, Grace Murray, 187-188<br />

Hormones. See Artificial hormone<br />

Hounsfield, Godfrey Newbold, 167,<br />

169<br />

House appliances. See Appliances<br />

Houtz, Ray C., 541<br />

Hovercraft, 407-411<br />

Howe, Elias, 335<br />

Hughes, Howard R., 533, 535<br />

Hulst, Hendrik Christoffel van de, 625<br />

Humphreys, Robert E., 765<br />

Humulin, 374, 377<br />

Hyatt, John Wesley, 571, 573<br />

Hyde, James Franklin, 683<br />

Hydrofoil, 665<br />

Hydrogen bomb, 412-416<br />

IBM. See International Business Machines<br />

IBM Model 1401 computer, 417-420<br />

Ibuka, Masaru, 778, 786, 875, 879<br />

ICBM. See Intercontinental ballistic<br />

missiles<br />

Idaho National Engineering<br />

Laboratory, 119, 521<br />

Immelmann, Max, 601<br />

Immunology. See Polio vaccine;<br />

Tuberculosis vaccine; Typhus<br />

vaccine; Yellow fever vaccine<br />

In vitro plant culture, 108, 421-424<br />

INEL. See Idaho National Engineering<br />

Laboratory<br />

Infantile paralysis. See Polio<br />

Infrared photography, 425-429<br />

Instant photography, 430-433<br />

Insulin, genetically engineered, 374-377<br />

Index / 929<br />

Intel Corporation, 153, 232, 234, 559<br />

Interchangeable parts, 434-441<br />

Intercontinental ballistic missiles, 63-64<br />

Internal combustion engine, 442-445<br />

International Business Machines, 31,<br />

140, 187, 189, 313, 330-331, 333, 347-<br />

350, 386, 388, 395, 420, 490-493, 680-<br />

681, 830, 861-865; Model 1401<br />

computer, 417-420; personal<br />

computers, 558-561<br />

Internet, 446-450<br />

Iron lung, 451-453<br />

Isotopes, <strong>and</strong> atomic mass, 494<br />

Ivanov, Ilya Ivanovich, 54<br />

Ives, Frederick E., 90<br />

Jansky, Karl, 614, 625<br />

Jarvik, Robert, 45<br />

Jarvik-7, 45, 49<br />

The Jazz Singer, 742<br />

Jeffreys, Alec, 370<br />

Jenkins, Charles F., 756<br />

Jet engines; <strong>and</strong> hovercraft, 408, 410;<br />

impulse, 871; <strong>and</strong> missiles, 244;<br />

supersonic, 714-719; turbo, 807-810<br />

Jobs, Steven, 28, 30<br />

Johnson, Irving S., 374<br />

Johnson, Lyndon B., 206<br />

Johnson, Reynold B., 330<br />

Joliot, Frédéric, 76<br />

Jolson, Al, 742<br />

Jones, Am<strong>and</strong>a Theodosia, 343, 345<br />

Joyce, John, 272<br />

Judson, Walter E., 634<br />

Judson, Whitcomb L., 847<br />

Kahn, Reuben Leon, 737<br />

Kamm, Oliver, 50<br />

Kao, Charles K., 320<br />

Kelvin, Lord, 398<br />

Kemeny, John G., 92<br />

Kettering, Charles F., 11, 630<br />

Kidneys, 58, 62, 374; <strong>and</strong> blood, 39;<br />

<strong>and</strong> cyclamate, 248; problems, 634<br />

Kilby, Jack St. Clair, 151, 229, 231, 576,<br />

578<br />

Kipping, Frederic Stanley, 683<br />

Kitchenware. See Polystyrene; Pyrex<br />

glass; Teflon; Tupperware


930 / Index<br />

Knoll, Max, 302<br />

Kober, Theodor, 267<br />

Koch, Robert, 791<br />

Kolff, Willem Johan, 58<br />

Kornberg, Arthur, 729<br />

Kornei, Otto, 891<br />

Korolev, Sergei P., 63-64<br />

Kramer, Piet, 537<br />

Krueger, Myron W., 866<br />

Kruiff, George T. de, 537<br />

Kunitsky, R. W., 54<br />

Kurtz, Thomas E., 92<br />

Lake, Clair D., 490<br />

Laminated glass, 454-458<br />

L<strong>and</strong>, Edwin Herbert, 430, 432<br />

Langévin, Paul, 692, 695, 823<br />

Langmuir, Irving, 183<br />

Laser, 459-463<br />

Laser-diode recording process, 464-467<br />

Laser eye surgery, 468-472<br />

Laser vaporization, 472-476<br />

Laservision, 219, 465<br />

Laue, Max von, 896<br />

Lauterbur, Paul C., 516<br />

Lawrence, Ernest Orl<strong>and</strong>o, 252, 254,<br />

720<br />

Lawrence-Livermore National<br />

Laboratory, 416, 671<br />

Leclanché, Georges, 355<br />

Leeuwenhoek, Antoni van, 678<br />

Leith, Emmett, 402<br />

Lel<strong>and</strong>, Henry M., 434, 437<br />

Lengyel, Peter, 733<br />

Lenses; camera, 130, 132-134;<br />

electromagnetic, 303; electron, 302-<br />

303; <strong>and</strong> fax machines, 317; <strong>and</strong><br />

laser diodes, 465; microscope, 678;<br />

<strong>and</strong> optical disks, 539; Pyrex, 609;<br />

railroad lantern, 606; scleral, 235;<br />

television camera, 887; <strong>and</strong><br />

xerography, 891-894. See also<br />

Contact lenses<br />

Leonardo da Vinci, 235<br />

Leverone, Louis E., 850<br />

Leverone, Nathaniel, 850<br />

Lewis, Thomas, 293<br />

LGOL computer language, 92-93<br />

Libby, Willard Frank, 158, 160<br />

Lidwell, Mark, 545<br />

Lincoln, Abraham, 320, 439<br />

Lindbergh, Charles A., 661<br />

Littleton, Jesse T., 606, 608<br />

Livestock, artificial insemination of,<br />

54-57<br />

Livingston, M. Stanley, 252<br />

Locke, Walter M., 244<br />

Lockhead Corporation, 697<br />

Long-distance radiotelephony, 477-<br />

481<br />

Long-distance telephone, 482-485<br />

Loosley, F. A., 701<br />

Lumière, Auguste, 88-89<br />

Lumière, Louis, 88-89<br />

Lynde, Frederick C., 850-851<br />

Lyons, Harold, 80<br />

McCabe, B. C., 378<br />

McCormick, Cyrus Hall, 335<br />

McCormick, Katherine Dexter, 108<br />

McCune, William J., 430<br />

Machine guns, 601-605<br />

McKay, Dean, 558<br />

McKenna, Regis, 28<br />

McKhann, Charles F., III, 451<br />

McMillan, Edwin Mattison, 720<br />

McWhir, J., 177<br />

Magnetron, 504<br />

Maiman, Theodore Harold, 320, 459,<br />

468, 472<br />

Mallory, Joseph, 432<br />

Mammography, 486-489<br />

Manhattan Project, 77-78, 412, 414, 525,<br />

747-748<br />

Mansfield, Peter, 516<br />

Marconi, Guglielmo, 477, 616, 619, 621,<br />

839<br />

Mariano di Jacopo detto Taccola, 770<br />

Mark I calculator, 490-493<br />

Marrison, Warren Alvin, 284, 286<br />

Marsden, Ernest, 367<br />

Mass spectrograph, 494-497<br />

Massachusetts Institute of Technology,<br />

861<br />

Mauchly, John W., 104, 312, 347, 828<br />

Maxwell, James Clerk, 88, 502, 621<br />

Meitner, Lise, 76, 520<br />

Memory metal, 498-501


Mercalli, Giuseppe, 645<br />

Merrill, John P., 61<br />

Merryman, Jerry D., 576, 578<br />

Mestral, Georges de, 846, 848<br />

Metchnikoff, Élie, 673-674<br />

Microprocessors, 94, 229-234, 287, 419,<br />

538<br />

Microscopes; atomic force, 681;<br />

electron, 302-306, 403, 902; field ion,<br />

325-329, 679; scanning tunneling,<br />

678-682; ultra-, 819-822<br />

Microvelcro, 847<br />

Microwave cooking, 502-506<br />

Midgley, Thomas, Jr., 444, 630, 746<br />

Miller, Bernard J., 394<br />

Miller, Stanley Lloyd, 724<br />

Millikan, Robert A., 646<br />

Millikan, Robert Andrews, 722<br />

Milunsky, Aubrey, 20<br />

Missiles; cruise, 244-247; guided, 385;<br />

intercontinental, 63-64; Sidewinder,<br />

698; Snark, 106. See also Rockets; V-2<br />

rocket<br />

Mixter, Samuel Jason, 113<br />

Mobile Telephone Service, 172<br />

Model T, 14, 71, 75, 439-440<br />

Monitor, 687<br />

Monocot plants, 422<br />

Monomers, 148, 541, 590-591<br />

Moog, Robert A., 307, 309<br />

Moon; distance to, 462; <strong>and</strong> lasers, 459;<br />

<strong>and</strong> radar, 614; <strong>and</strong> radio signals,<br />

614<br />

Morel, Georges Michel, 421-423<br />

Morganthaler, Ottmar, 335<br />

Morita, Akio, 217, 222, 778, 786, 875<br />

Morse, Samuel F. B., 320, 335<br />

Morse code, 477, 616, 621<br />

Motion picture sound, 741-745<br />

Mouchout, Augustin, 687<br />

Movies. See Talking motion pictures<br />

Müller, Erwin Wilhelm, 325, 327, 679<br />

Murray, Andrew W., 41<br />

Murrow, Edward R., 830<br />

Naito, Ryoichi, 38<br />

National Broadcasting Company, 198,<br />

215<br />

National Geographic, 665<br />

Index / 931<br />

National Geographic Society, 665<br />

National Radio Astronomy<br />

Observatory, 628<br />

Natta, Giulio, 593<br />

Nautilus, 84, 521<br />

NBC. See National Broadcasting<br />

Company<br />

Neoprene, 507-510<br />

Neumann, John von, 92, 104, 312, 347,<br />

710, 828<br />

Neurophysiology, 298, 300<br />

Neutrino detector, 511-515<br />

Newman, Max H. A., 200<br />

Newton, Isaac, 659<br />

Nickerson, William Emery, 272<br />

Nieuwl<strong>and</strong>, Julius Arthur, 507<br />

Nipkow, Paul Gottlieb, 756<br />

Nirenberg, Marshall W., 733<br />

Nitinol, 498-501<br />

Nitrogen, 16<br />

Nobécourt, P., 421<br />

Nobel Prize winners, 174; Chemistry,<br />

16, 18-19, 50, 52, 158, 160, 183, 455,<br />

494, 496, 595, 720, 724, 819, 821-822;<br />

Physics, 229, 231, 252, 254-255, 302,<br />

304, 321, 402, 404, 459, 520, 619, 678,<br />

680, 782, 789, 896-898; Physiology<br />

or Medicine, 24, 41, 167, 169, 293,<br />

295, 375, 553, 555, 581, 674, 676, 730,<br />

733<br />

Nordwestdeutsche Kraftwerke, 225<br />

Northrop Corporation, 106, 697<br />

Noyce, Robert, 151, 229<br />

NSFnet, 447<br />

Nuclear fission, 76, 84, 118-121, 185,<br />

412, 520-528<br />

Nuclear fusion, 78<br />

Nuclear magnetic resonance, 516-519<br />

Nuclear power plant, 520-524<br />

Nuclear reactor, 118-121, 520-528<br />

Nylon, 510, 529-532, 541, 574, 590;<br />

Helance, 591; <strong>and</strong> velcro, 846-847<br />

Oak Ridge National Laboratory, 77,<br />

525-528<br />

Ochoa, Severo, 733<br />

Ohain, Hans Pabst von, 807<br />

Ohga, Norio, 875<br />

Oil-well drilling, 345, 533-536<br />

Oparin, Aleks<strong>and</strong>r Ivanovich, 724


932 / Index<br />

Opel, John, 558<br />

Ophthalmology, 468<br />

Oppenheimer, J. Robert, 76, 325<br />

Optical disk, 537-540<br />

Orlon, 541-544<br />

Ottens, Lou F., 537<br />

Otto, Nikolaus, 257<br />

Oxytocin, 50<br />

Pacemaker, 545-548<br />

Painter, William, 272<br />

Paley, William S., 196<br />

Pap test, 549-552<br />

Papanicolaou, George N., 549<br />

Parsons, Charles, 378<br />

Parsons, Ed, 208<br />

Particle accelerators, 252, 256, 720-723,<br />

761-764<br />

Paul, Les, 122<br />

Pauli, Wolfgang, 511<br />

PC. See Personal computers<br />

PCM. See Pulse code modulation<br />

Pearson, Gerald L., 567<br />

Penicillin, 553-557<br />

Peoples, John, 761<br />

Perkin, William Henry, 842, 844<br />

Perrin, Jean, 695<br />

Persian Gulf War, 246, 698-699<br />

Personal computers, 153, 558-561, 864;<br />

Apple, 28-32; <strong>and</strong> floppy disks, 332-<br />

333; <strong>and</strong> hard disks, 389; <strong>and</strong><br />

Internet, 447, 449<br />

Pfleumer, Fritz, 163<br />

Philibert, Daniel, 1<br />

Philips Corporation, 464, 857<br />

Photocopying. See Xerography<br />

Photoelectric cell, 562-566<br />

Photography; film, 88-91, 192-195, 430-<br />

433. See also Cameras<br />

Photovoltaic cell, 567-570<br />

Piccard, Auguste, 36, 95, 97, 103<br />

Piccard, Jacques, 95<br />

Piccard, Jean-Félix, 97<br />

Pickard, Greenleaf W., 621<br />

Pierce, John R., 204<br />

Pincus, Gregory, 108<br />

Planck, Max, 563<br />

Plastic, 571-575; Tupperware, 799-806<br />

Plunkett, Roy J., 746, 748<br />

Pocket calculator, 576-580<br />

Polaroid camera, 170, 430-433<br />

Polio, 451-453<br />

Polio vaccine, 581-588<br />

Polyacrylonitrile, 541-543<br />

Polyester, 589-592<br />

Polyethylene, 593-596<br />

Polystyrene, 597-600<br />

Porter, Steven, 272<br />

Powers, Gary, 245<br />

Pregnancy. See Abortion pill;<br />

Amniocentesis; Birth control pill;<br />

Ultrasound<br />

Priestley, Joseph, 146<br />

Propeller-coordinated machine gun,<br />

601-605<br />

Protein synthesis, 735<br />

Prout, William, 494<br />

Pulse code modulation, 217-220<br />

Purcell, Edward Mills, 625<br />

Purvis, Merton Brown, 751<br />

Pye, David R<strong>and</strong>all, 442<br />

Pyrex glass, 606-610<br />

Quadrophonic sound, 221<br />

Quantum theory, 563<br />

Quartz crystals, 81, 284-288<br />

Radar, 229, 265, 314, 391, 504, 611-612,<br />

614-615; <strong>and</strong> sonar, 693, 824; <strong>and</strong><br />

bathyscaphe, 96; <strong>and</strong> laser<br />

holography, 405; <strong>and</strong> stealth aircraft,<br />

697-699<br />

Radio, 616-620; FM, 339-342<br />

Radio Corporation of America, 196-<br />

199, 210, 213, 219, 340-341, 464, 537,<br />

618-619, 741, 758-759, 787<br />

Radio crystal sets, 621-624<br />

Radio frequency, 616; <strong>and</strong> cell phones,<br />

172, 175; <strong>and</strong> crystral radio, 622; <strong>and</strong><br />

microwave heating, 505<br />

Radio interferometer, 625-629<br />

Radioactivity, 720, 734; <strong>and</strong> barium, 76,<br />

520; carbon dating, 158-162; <strong>and</strong><br />

DNA, 371; <strong>and</strong> isotopes, 494, 497;<br />

measuring, 365-369; <strong>and</strong> neutrinos,<br />

511-512<br />

Radiotelephony, 477-481<br />

Rainfall, induced, 183-186<br />

RAM. See R<strong>and</strong>om access memory


R<strong>and</strong>om access memory, 140, 387, 559,<br />

861-862, 864<br />

Raytheon Company, 503, 505, 786<br />

Razors, 272-278<br />

RCA. See Radio Corporation of<br />

America<br />

Reagan, Ronald, 415<br />

Reber, Grote, 625<br />

Recombinant DNA, 41<br />

Recording; cassettes, 163-166, 538, 784,<br />

788, 875-882; compact discs, 217-224;<br />

Dolby noise reduction, 279-283;<br />

laser-diodes, 464-467; sound, 741-<br />

742; video, 857-860<br />

Reed, Walter, 905<br />

Refrigerant gas, 630-633<br />

Reichenbach, Henry M., 130<br />

Rein, Herbert, 541<br />

Remsen, Ira, 67<br />

Reserpine, 634-637<br />

Ribonucleic acid, 734. See also Synthetic<br />

RNA<br />

Ricardo, Harry Ralph, 442<br />

Rice <strong>and</strong> wheat strains, 638-644<br />

Rice-Wray, Edris, 108<br />

Richter, Charles F., 645-646<br />

Richter scale, 645-649<br />

Rickover, Hyman G., 520<br />

Riffolt, Nils, 661<br />

Ritchie, W. A., 177<br />

Rizzo, Paul, 558<br />

RNA, synthetic, 733-736<br />

Robot, household, 650-653<br />

Robot, industrial, 654-657<br />

Rochow, Eugene G., 683, 685<br />

Rock, John, 108<br />

Rockets; <strong>and</strong> satellites, 63-66; design,<br />

712; liquid-fuel-propelled, 658-662,<br />

871-874. See also Missiles<br />

Rogers, Howard G., 430<br />

Rohrer, Heinrich, 678, 680<br />

Röntgen, Wilhelm Conrad, 167, 365,<br />

896, 901<br />

Roosevelt, Franklin D., 264, 588, 770-771<br />

Root, Elisha King, 71<br />

Rosen, Charles, 362<br />

Rosing, Boris von, 758<br />

Rossi, Michele Stefano de, 645<br />

Rotary cone drill bit, 533, 536<br />

Index / 933<br />

Rotary dial telephone, 663-667, 751,<br />

774-776<br />

Rotary engine, 362<br />

Roux, Pierre-Paul-Émile, 673<br />

Rubber, synthetic, 146-150, 507-510,<br />

530, 593, 595<br />

Ruska, Ernst, 302, 304, 678, 680<br />

Russell, Archibald, 714<br />

Rutherford, Ernest, 252, 365-368, 455,<br />

494, 564, 720-721, 898<br />

Ryle, Martin, 625<br />

Sabin, Albert Bruce, 581, 583<br />

Saccharin, 248<br />

Sachs, Henry, 272<br />

SAINT, 668-672<br />

Salk, Jonas Edward, 581, 585-586<br />

Salomon, Albert, 486<br />

Salvarsan, 673-674, 676-677<br />

Sanger, Margaret, 108, 110, 112<br />

Sarnoff, David, 196-197, 210, 339-340,<br />

758<br />

Satellite, artificial, 63-66<br />

Satre, Pierre, 714<br />

Saulnier, Raymond, 601<br />

Savannah, 85<br />

Sawyer, Wilbur Augustus, 905<br />

Sayer, Gerry, 807<br />

Scanning tunneling microscope, 678-682<br />

Schaefer, Vincent Joseph, 183<br />

Schaudinn, Fritz, 673<br />

Schawlow, Arthur L., 459<br />

Schlatter, James M., 67<br />

Schmidt, Paul, 871<br />

Scholl, Rol<strong>and</strong>, 842<br />

Schönbein, Christian Friedrich, 571<br />

Schrieffer, J. Robert, 789<br />

SDI. See Strategic Defense Initiative<br />

Selectavision, 219<br />

Semiconductors, 139-140, 218, 229-234,<br />

317, 464-466, 568, 786-787, 892; <strong>and</strong><br />

calculators, 577, 579; defined, 229,<br />

232, 891<br />

Senning, Ake, 545<br />

Serviss, Garrett P., 659<br />

Seyewetz, Alphonse, 88<br />

Shannon, Claude, 868<br />

Sharp, Walter B., 533, 535<br />

Sharpey-Schafer, Edward Albert, 50


934 / Index<br />

Shaw, Louis, 451<br />

Shaw, Ronald A., 407<br />

Sheep, cloning of, 177-182<br />

Shellac, 572<br />

Shockley, William B., 229, 778, 782, 786,<br />

789<br />

Shroud of Turin, 161<br />

Shugart, Alan, 330, 386<br />

Shuman, Frank, 687<br />

Sidewinder missile, 698<br />

Siedentopf, H. F. W., 819<br />

Siegrist, H., 192<br />

Silicones, 683-686<br />

Simon, Edward, 597<br />

The Singing Fool, 742<br />

Sinjou, Joop, 537<br />

Sinsheimer, Robert L., 729<br />

Sketchpad, 868<br />

Slagle, James R., 668, 671<br />

Sloan, David, 252<br />

Smith, Hugh, 905<br />

Smith, Robert, 427<br />

Smouluchowski, Max von, 819<br />

Snark missile, 106<br />

Snyder, Howard, 883<br />

Sogo, Shinji, 142<br />

Solar energy, 567-568, 687-688, 690<br />

Solar thermal engine, 687-691<br />

Sonar, 692-696; <strong>and</strong> radar, 823<br />

Sones, F. Mason, 240<br />

Sony Corporation, 165, 218-224, 539,<br />

778, 781, 783-785, 788-789, 875-881<br />

Spaeth, Mary, 459, 461<br />

Spallanzani, Lazzaro, 54<br />

Spangler, James Murray, 832<br />

Spencer, Percy L., 502, 504<br />

Sperry, Elmer Ambrose, 382, 384<br />

Sputnik, 63-66, 446, 874<br />

“Star Wars” (Strategic Defense<br />

Initiative), 699<br />

Staudinger, Hermann, 530<br />

Stealth aircraft, 697-700<br />

Steelmaking process, 701-708<br />

Steenstrup, Christian, 289<br />

Stewart, Alice, 823<br />

Stewart, Edward J., 272<br />

Stibitz, George, 828<br />

Stine, Charles M. A., 529<br />

STM. See Scanning tunneling microscope<br />

Stockard, Charles, 549<br />

Stokes, T. L., 394<br />

Storax, 597<br />

Strassmann, Fritz, 76<br />

Strategic Defense Initiative, 416, 699<br />

Strowger, Almon B., 751, 753<br />

Styrene, 148-149, 597-598<br />

Submarines; detection of, 692, 695, 823;<br />

navigation, 382-385; nuclear, 98, 521;<br />

weapons, 245-246<br />

Sucaryl, 248<br />

Suess, Theodor, 701<br />

Sulfonamides, 24<br />

Sullivan, Eugene G., 606<br />

Sun, 514, 725; energy, 725; <strong>and</strong> nuclear<br />

fusion, 511, 515, 567, 687; <strong>and</strong><br />

timekeeping, 80<br />

Sun Power Company, 688<br />

Supercomputer, 709-713<br />

Supersonic passenger plane, 714-719<br />

Surgery; <strong>and</strong> artificial blood, 39; <strong>and</strong><br />

artificial heart, 46, 48; <strong>and</strong> blood<br />

transfusion, 113-117; <strong>and</strong> breast<br />

cancer, 486-487, 489; cardiac, 499;<br />

coronary artery bypass, 240-243;<br />

<strong>and</strong> heart-lung machine, 394-397;<br />

kidney-transplant, 61; laser eye,<br />

468-472; laser vaporization, 472-476;<br />

transplantation, 61<br />

Sutherl<strong>and</strong>, Ivan, 866, 868<br />

Sveda, Michael, 67, 248<br />

Svedberg, Theodor, 815<br />

Swarts, Frédéric, 630<br />

Swinton, Alan A. Campbell, 756<br />

Sydnes, William L., 558<br />

Synchrocyclotron, 720-723<br />

Synthetic amino acid, 724-728<br />

Synthetic DNA, 729-732<br />

Synthetic RNA, 733-736<br />

Syphilis, 24, 554-556, 673, 676, 744; test,<br />

737-740; treatment of, 673-674, 676-<br />

677<br />

Szostak, Jack W., 41<br />

Talking motion pictures, 741-745<br />

Tarlton, Robert J., 208-209<br />

Tassel, James Van, 576<br />

Taylor, Frederick Winslow, 71<br />

Taylor, William C., 606


Tee-Van, John, 100<br />

Teflon, 746-750<br />

Telecommunications Research<br />

Establishment, 201<br />

Telegraphy, radio, 616<br />

Telephone; cellular, 172-176; longdistance,<br />

482-485; rotary dial, 663-<br />

667, 751, 774-776; touch-tone, 667,<br />

774-777<br />

Telephone switching, 751-755<br />

Television, 756-760<br />

Teller, Edward, 78, 412, 414, 416<br />

Tesla, Nikola, 13, 832<br />

Teutsch, Georges, 1<br />

Tevatron accelerator, 761-764<br />

Texas Instruments, 140, 153, 232-233,<br />

419, 577-579, 787-788<br />

Theiler, Max, 905<br />

Thein, Swee Lay, 370<br />

Thermal cracking process, 765-769<br />

Thermionic valve, 564<br />

Thomson Electron Tubes, 901<br />

Thomson, Joseph John, 494, 496, 563-<br />

564, 838<br />

Thornycroft, John Isaac, 407, 409<br />

Thornycroft, Oliver, 442<br />

Tidal power plant, 770-773<br />

Tiros 1, 887-890<br />

Tiselius, Arne, 815<br />

Tokyo Telecommunications<br />

Engineering Company, 778, 780,<br />

787, 875. See also Sony Corporation<br />

Tomography, 168, 170<br />

Topografiner, 679<br />

Torpedo boat, 409<br />

Touch-tone telephone, 667, 774-777<br />

Townes, Charles Hard, 459<br />

Townsend, John Sealy Edward, 365<br />

Toyota Prius, 363<br />

Transistor radio, 786-790<br />

Transistors, 172, 229, 232, 390-391, 418-<br />

419, 778-788, 875-876; invention of,<br />

840<br />

Traut, Herbert, 549<br />

Tressler, Donald K., 343<br />

Truman, Harry S., 78<br />

Tsiolkovsky, Konstantin, 63, 65<br />

Tuberculosis vaccine, 791-794<br />

Tungsten filament, 795-798<br />

Index / 935<br />

Tuohy, Kevin, 235<br />

Tupper, Earl S., 799, 803<br />

Tupperware, 799-806<br />

Turbojet, 807-810<br />

Turing, Alan Mathison, 104, 200, 668<br />

Turner, Ted, 208, 211<br />

Tuskegee Airmen, 612<br />

Typhus vaccine, 811-814<br />

U2 spyplane, 245, 432<br />

U-boats. See Submarines<br />

Ulam, Stanislaw, 412, 414<br />

Ultracentrifuge, 815-818<br />

Ultramicroscope, 819-822<br />

Ultrasound, 823-827<br />

Unimate robots, 654-656<br />

UNIVAC. See Universal Automatic<br />

Computer<br />

Universal Automatic Computer, 106,<br />

315, 331, 348, 711, 828-831<br />

Upatnieks, Juris, 402<br />

Urey, Harold Clayton, 724<br />

Uterine cancer, 549, 552<br />

V-2 rocket, 65, 244, 659, 662, 871-874<br />

Vaccines. See Polio vaccine;<br />

Tuberculosis vaccine; Typhus<br />

vaccine; Yellow fever vaccine<br />

Vacuum cleaner, 832-836<br />

Vacuum tubes, 339, 837-841; <strong>and</strong><br />

computers, 106, 201-202, 313-314;<br />

<strong>and</strong> radar, 391; <strong>and</strong> radio, 478, 623;<br />

<strong>and</strong> television, 783; thermionic<br />

valve, 564; <strong>and</strong> transistors, 229, 391,<br />

778-780, 786-787, 876. See also<br />

Cathode-ray tubes<br />

Vat dye, 842-845<br />

VCR. See Videocassette recorder<br />

Vectograph, 432<br />

Veksler, Vladimir Iosifovich, 720<br />

Velcro, 846-849<br />

Vending machine slug rejector, 850-856<br />

Videocassette recorder, 214, 218, 857-<br />

860; <strong>and</strong> laservision, 465<br />

Videodisc, 219<br />

Vigneaud, Vincent du, 50<br />

Virtual machine, 861-865<br />

Virtual reality, 866-870<br />

Vitaphone, 742<br />

Vogel, Orville A., 638, 643


936 / Index<br />

Volta, Aless<strong>and</strong>ro, 355<br />

Vonnegut, Bernard, 183<br />

Vulcanization of rubber, 146, 149<br />

Wadati, Kiyoo, 645<br />

Waldeyer, Wilhelm von, 673<br />

Walker, William H., 130<br />

Walkman cassette player, 165, 784, 788,<br />

875-882<br />

Waller, Augustus D., 293<br />

Warner Bros., 741-745<br />

Warner, Albert, 741, 744<br />

Warner, Harry, 741, 744<br />

Warner, Jack, 741, 744<br />

Warner, Samuel, 741, 744<br />

Warren, Stafford L., 487<br />

Washing machine, electric, 883-886<br />

Washington, George, 289<br />

Wassermann, August von, 676, 737<br />

Watson, James D., 41, 177, 729, 733<br />

Watson, Thomas A., 482<br />

Watson, Thomas J., 394, 558<br />

Watson, Thomas J., Jr., 386<br />

Watson-Watt, Robert, 611<br />

Weather; <strong>and</strong> astronomy, 609; cloud<br />

seeding, 183-186; <strong>and</strong> rockets, 712<br />

Weather satellite, 887-890<br />

Wehnelt, Arthur, 837<br />

Wells, H. G., 659<br />

Westinghouse, George, 335<br />

Westinghouse Company, 101, 440, 758-<br />

759, 832<br />

Wexler, Harry, 887<br />

Whinfield, John R., 589<br />

Whitaker, Martin D., 525<br />

White, Philip Cleaver, 421<br />

White S<strong>and</strong>s Missile Range, 873<br />

Whitney, Eli, 71, 335<br />

Whittle, Frank, 807<br />

Wichterle, Otto, 235<br />

Wigginton, R<strong>and</strong>y, 28<br />

Wigner, Eugene, 525, 789<br />

Wilkins, Arnold F., 611<br />

Wilkins, Maurice H. F., 733<br />

Wilkins, Robert Wallace, 634<br />

Williams, Charles Greville, 146<br />

Wilmut, Ian, 177-178<br />

Wilson, Robert Rathbun, 761<br />

Wilson, Victoria, 370<br />

Wise, Brownie, 799<br />

Wolf, Fred, 289-290<br />

Woodrow, O. B., 883<br />

World War I, <strong>and</strong> nitrates, 18<br />

World War II; <strong>and</strong> Aqualung, 36;<br />

atomic bomb, 84, 118, 521, 525-527,<br />

697, 721; <strong>and</strong> computers, 92; spying,<br />

34, 200, 202, 668; V-2 rocket, 65, 244,<br />

659, 662, 871-874<br />

Wouk, Victor, 360, 362<br />

Wozniak, Stephen, 28, 30<br />

Wright, Almroth, 555<br />

Wright, Orville, 6-10, 658<br />

Wright, Wilbur, 6-10, 335, 658<br />

Wynn-Williams, C. E., 200<br />

Xerography, 891-895<br />

Xerox Corporation, 891-894<br />

X-ray crystallography, 896-900<br />

X-ray image intensifier, 901-904<br />

X-ray mammography, 486-489<br />

Yellow fever vaccine, 905-908<br />

Yoshino, Hiroyuki, 360<br />

Zaret, Milton M., 468<br />

Zenith Radio Corporation, 209, 341<br />

Zeppelin, Ferdin<strong>and</strong> von, 267-270<br />

Ziegler, Karl, 593<br />

Zinn, Walter Henry, 118<br />

Zinsser, Hans, 811<br />

Zippers, 846-847<br />

Zoll, Paul Maurice, 545<br />

Zsigmondy, Richard, 819, 821<br />

Zweng, H. Christian, 468<br />

Zworykin, Vladimir, 756, 758

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!