Tuesday, July 28, 2009

In vitro plant culture




The invention: Method for propagating plants in artificial media
that has revolutionized agriculture.
The people behind the invention:
Georges Michel Morel (1916-1973), a French physiologist
Philip Cleaver White (1913- ), an American chemist
Plant Tissue Grows “In Glass”
In the mid-1800’s, biologists began pondering whether a cell isolated
from a multicellular organism could live separately if it were
provided with the proper environment. In 1902, with this question in
mind, the German plant physiologist Gottlieb Haberlandt attempted
to culture (grow) isolated plant cells under sterile conditions on an artificial
growth medium. Although his cultured cells never underwent
cell division under these “in vitro” (in glass) conditions, Haberlandt
is credited with originating the concept of cell culture.
Subsequently, scientists attempted to culture plant tissues and
organs rather than individual cells and tried to determine the medium
components necessary for the growth of plant tissue in vitro.
In 1934, Philip White grew the first organ culture, using tomato
roots. The discovery of plant hormones, which are compounds that
regulate growth and development, was crucial to the successful culture
of plant tissues; in 1939, Roger Gautheret, P. Nobécourt, and
White independently reported the successful culture of plant callus
tissue. “Callus” is an irregular mass of dividing cells that often results
from the wounding of plant tissue. Plant scientists were fascinated
by the perpetual growth of such tissue in culture and spent
years establishing optimal growth conditions and exploring the nutritional
and hormonal requirements of plant tissue.
Plants by the Millions
A lull in botanical research occurred during World War II, but
immediately afterward there was a resurgence of interest in applying
tissue culture techniques to plant research. Georges Morel, a plant physiologist at the National Institute for Agronomic Research
in France, was one of many scientists during this time who
had become interested in the formation of tumors in plants as well
as in studying various pathogens such as fungi and viruses that
cause plant disease.
To further these studies, Morel adapted existing techniques in order
to grow tissue from a wider variety of plant types in culture, and
he continued to try to identify factors that affected the normal
growth and development of plants. Morel was successful in culturing
tissue from ferns and was the first to culture monocot plants.
Monocots have certain features that distinguish them fromthe other
classes of seed-bearing plants, especially with respect to seed structure.
More important, the monocots include the economically important
species of grasses (the major plants of range and pasture)
and cereals.
For these cultures, Morel utilized a small piece of the growing tip
of a plant shoot (the shoot apex) as the starting tissue material. This
tissue was placed in a glass tube, supplied with a medium containing
specific nutrients, vitamins, and plant hormones, and allowed
to grow in the light. Under these conditions, the apex tissue grew
roots and buds and eventually developed into a complete plant.
Morel was able to generate whole plants from pieces of the shoot
apex that were only 100 to 250 micrometers in length.
Morel also investigated the growth of parasites such as fungi and
viruses in dual culture with host-plant tissue. Using results from
these studies and culture techniques that he had mastered, Morel
and his colleague Claude Martin regenerated virus-free plants from
tissue that had been taken from virally infected plants. Tissues from
certain tropical species, dahlias, and potato plants were used for the
original experiments, but after Morel adapted the methods for the
generation of virus-free orchids, plants that had previously been
difficult to propagate by any means, the true significance of his
work was recognized.
Morel was the first to recognize the potential of the in vitro culture
methods for the mass propagation of plants. He estimated that several
million plants could be obtained in one year from a single small
piece of shoot-apex tissue. Plants generated in this manner were
clonal (genetically identical organisms prepared from a single plant).With other methods of plant propagation, there is often a great variation
in the traits of the plants produced, but as a result of Morel’s
ideas, breeders could select for some desirable trait in a particular
plant and then produce multiple clonal plants, all of which expressed
the desired trait. The methodology also allowed for the production of
virus-free plant material, which minimized both the spread of potential
pathogens during shipping and losses caused by disease.
Consequences
Variations on Morel’s methods are used to propagate plants used
for human food consumption; plants that are sources of fiber, oil,
and livestock feed; forest trees; and plants used in landscaping and
in the floral industry. In vitro stocks are preserved under deepfreeze
conditions, and disease-free plants can be proliferated quickly
at any time of the year after shipping or storage.
The in vitro multiplication of plants has been especially useful
for species such as coconut and certain palms that cannot be propagated
by other methods, such as by sowing seeds or grafting, and
has also become important in the preservation and propagation of rare plant species that might otherwise have become extinct. Many
of these plants are sources of pharmaceuticals, oils, fragrances, and
other valuable products.
The capability of regenerating plants from tissue culture has also
been crucial in basic scientific research. Plant cells grown in culture
can be studied more easily than can intact plants, and scientists have
gained an in-depth understanding of plant physiology and biochemistry
by using this method. This information and the methods
of Morel and others have made possible the genetic engineering and
propagation of crop plants that are resistant to disease or disastrous
environmental conditions such as drought and freezing. In vitro
techniques have truly revolutionized agriculture.

IBM Model 1401 Computer




The invention: A relatively small, simple, and inexpensive computer
that is often credited with having launched the personal
computer age.
The people behind the invention:
Howard H. Aiken (1900-1973), an American mathematician
Charles Babbage (1792-1871), an English mathematician and
inventor
Herman Hollerith (1860-1929), an American inventor
Computers: From the Beginning
Computers evolved into their modern form over a period of
thousands of years as a result of humanity’s efforts to simplify the
process of counting. Two counting devices that are considered to be
very simple, early computers are the abacus and the slide rule.
These calculating devices are representative of digital and analog
computers, respectively, because an abacus counts numbers of things,
while the slide rule calculates length measurements.
The first modern computer, which was planned by Charles Babbage
in 1833, was never built. It was intended to perform complex
calculations with a data processing/memory unit that was controlled
by punched cards. In 1944, Harvard University’s Howard H.
Aiken and the International Business Machines (IBM) Corporation
built such a computer—the huge, punched-tape-controlled Automatic
Sequence Controlled Calculator, or Mark I ASCC, which
could perform complex mathematical operations in seconds. During
the next fifteen years, computer advances produced digital computers
that used binary arithmetic for calculation, incorporated
simplified components that decreased the sizes of computers, had
much faster calculating speeds, and were transistorized.
Although practical computers had become much faster than
they had been only a few years earlier, they were still huge and extremely
expensive. In 1959, however, IBM introduced the Model
1401 computer. Smaller, simpler, and much cheaper than the multimillion-dollar computers that were available, the IBM Model 1401
computer was also relatively easy to program and use. Its low cost,
simplicity of operation, and very wide use have led many experts
to view the IBM Model 1401 computer as beginning the age of the
personal computer.
Computer Operation and IBM’s Model 1401
Modern computers are essentially very fast calculating machines
that are capable of sorting, comparing, analyzing, and outputting information,
as well as storing it for future use. Many sources credit
Aiken’s Mark I ASCC as being the first modern computer to be built.
This huge, five-ton machine used thousands of relays to perform complex
mathematical calculations in seconds. Soon after its introduction,
other companies produced computers that were faster and more versatile
than the Mark I. The computer development race was on.
All these early computers utilized the decimal system for calculations
until it was found that binary arithmetic, whose numbers are
combinations of the binary digits 1 and 0, was much more suitable
for the purpose. The advantage of the binary system is that the electronic
switches that make up a computer (tubes, transistors, or
chips) can be either on or off; in the binary system, the on state can
be represented by the digit 1, the off state by the digit 0. Strung together
correctly, binary numbers, or digits, can be inputted rapidly
and used for high-speed computations. In fact, the computer term
bit is a contraction of the phrase “binary digit.”
A computer consists of input and output devices, a storage device
(memory), arithmetic and logic units, and a control unit. In
most cases, a central processing unit (CPU) combines the logic,
arithmetic, memory, and control aspects. Instructions are loaded
into the memory via an input device, processed, and stored. Then,
the CPU issues commands to the other parts of the system to carry
out computations or other functions and output the data as needed.
Most output is printed as hard copy or displayed on cathode-ray
tube monitors, or screens.
The early modern computers—such as the Mark I ASCC—were
huge because their information circuits were large relays or tubes.
Computers became smaller and smaller as the tubes were replaced first with transistors, then with simple integrated circuits, and then
with silicon chips. Each technological changeover also produced
more powerful, more cost-effective computers.
In the 1950’s, with reliable transistors available, IBM began the
development of two types of computers that were completed by
about 1959. The larger version was the Stretch computer, which was
advertised as the most powerful computer of its day. Customized
for each individual purchaser (for example, the Atomic Energy
Commission), a Stretch computer cost $10 million or more. Some innovations
in Stretch computers included semiconductor circuits,
new switching systems that quickly converted various kinds of data
into one language that was understood by the CPU, rapid data readers,
and devices that seemed to anticipate future operations.
Consequences
The IBM Model 1401 was the first computer sold in very large
numbers. It led IBM and other companies to seek to develop less expensive,
more versatile, smaller computers that would be sold to
small businesses and to individuals. Six years after the development
of the Model 1401, other IBM models—and those made by
other companies—became available that were more compact and
had larger memories. The search for compactness and versatility
continued. A major development was the invention of integrated
circuits by Jack S. Kilby of Texas Instruments; these integrated circuits
became available by the mid-1960’s. They were followed by
even smaller “microprocessors” (computer chips) that became available
in the 1970’s. Computers continued to become smaller and more
powerful.
Input and storage devices also decreased rapidly in size. At first,
the punched cards invented by Herman Hollerith, founder of the
Tabulation Machine Company (which later became IBM), were read
by bulky readers. In time, less bulky magnetic tapes and more compact
readers were developed, after which magnetic disks and compact
disc drives were introduced.
Many other advances have been made. Modern computers can
talk, create art and graphics, compose music, play games, and operate
robots. Further advancement is expected as societal needs change. Many experts believe that it was the sale of large numbers
of IBM Model 1401 computers that began the trend.

Monday, July 20, 2009

Hydrogen bomb



The invention: Popularly known as the “H-Bomb,” the hydrogen
bomb differs from the original atomic bomb in using fusion,
rather than fission, to create a thermonuclear explosion almost a
thousand times more powerful.
The people behind the invention:
Edward Teller (1908- ), a Hungarian-born theoretical
physicist
Stanislaw Ulam (1909-1984), a Polish-born mathematician
Crash Development
Afew months before the 1942 creation of the Manhattan Project,
the United States-led effort to build the atomic (fission) bomb, physicist
Enrico Fermi suggested to Edward Teller that such a bomb
could release more energy by the process of heating a mass of the
hydrogen isotope deuterium and igniting the fusion of hydrogen
into helium. Fusion is the process whereby two atoms come together
to form a larger atom, and this process usually occurs only in stars,
such as the Sun. Physicists Hans Bethe, George Gamow, and Teller
had been studying fusion since 1934 and knew of the tremendous
energy than could be released by this process—even more energy
than the fission (atom-splitting) process that would create the atomic
bomb. Initially, Teller dismissed Fermi’s idea, but later in 1942, in
collaboration with Emil Konopinski, he concluded that a hydrogen
bomb, or superbomb, could be made.
For practical considerations, it was decided that the design of the
superbomb would have to wait until after the war. In 1946, a secret
conference on the superbomb was held in Los Alamos, New Mexico,
that was attended by, among other Manhattan Project veterans,
Stanislaw Ulam and Klaus Emil Julius Fuchs. Supporting the investigation
of Teller’s concept, the conferees requested a more complete
mathematical analysis of his own admittedly crude calculations
on the dynamics of the fusion reaction. In 1947, Teller believed
that these calculations might take years. Two years later, however,the Soviet explosion of an atomic bomb convinced Teller that America’s
ColdWar adversary was hard at work on its own superbomb.
Even when new calculations cast further doubt on his designs,
Teller began a vigorous campaign for crash development of the hydrogen
bomb, or H-bomb.
The Superbomb
Scientists knew that fusion reactions could be induced by the explosion
of an atomic bomb. The basic problem was simple and formidable:
How could fusion fuel be heated and compressed long
enough to achieve significant thermonuclear burning before the
atomic fission explosion blew the assembly apart? A major part of
the solution came from Ulam in 1951. He proposed using the energy
from an exploding atomic bomb to induce significant thermonuclear
reactions in adjacent fusion fuel components.
This arrangement, in which the A-bomb (the primary) is physically
separated from the H-bomb’s (the secondary’s) fusion fuel, became
known as the “Teller-Ulam configuration.” All H-bombs are
cylindrical, with an atomic device at one end and the other components
filling the remaining space. Energy from the exploding primary
could be transported by X rays and would therefore affect the
fusion fuel at near light speed—before the arrival of the explosion.
Frederick de Hoffman’s work verified and enriched the new concept.
In the revised method, moderated X rays from the primary irradiate
a reactive plastic medium surrounding concentric and generally
cylindrical layers of fusion and fission fuel in the secondary.
Instantly, the plastic becomes a hot plasma that compresses and
heats the inner layer of fusion fuel, which in turn compresses a central
core of fissile plutonium to supercriticality. Thus compressed,
and bombarded by fusion-produced, high-energy neutrons, the fission
element expands rapidly in a chain reaction from the inside
out, further compressing and heating the surrounding fusion fuel,
releasing more energy and more neutrons that induce fission in a
fuel casing-tamper made of normally stable uranium 238.
With its equipment to refrigerate the hydrogen isotopes, the device
created to test Teller’s new concept weighed more than sixty
tons. During Operation Ivy, it was tested at Elugelab in the Marshall Islands on November 1, 1952. Exceeding the expectations of all concerned
and vaporizing the island, the explosion equaled 10.4 million
tons of trinitrotoluene (TNT), which meant that it was about
seven hundred times more powerful than the atomic bomb dropped
on Hiroshima, Japan, in 1945. A version of this device weighing
about 20 tons was prepared for delivery by specially modified Air
Force B-36 bombers in the event of an emergency during wartime.
In development at Los Alamos before the 1952 test was a device
weighing only about 4 tons, a “dry bomb” that did not require refrigeration
equipment or liquid fusion fuel; when sufficiently compressed
and heated in its molded-powder form, the new fusion fuel
component, lithium-6 deutride, instantly produced tritium, an isotope
of hydrogen. This concept was tested during Operation Castle
at Bikini atoll in 1954 and produced a yield of 15 million tons of TNT,
the largest-ever nuclear explosion created by the United States.
Consequences
Teller was not alone in believing that the world could produce
thermonuclear devices capable of causing great destruction. Months
before Fermi suggested to Teller the possibility of explosive thermonuclear
reactions on Earth, Japanese physicist Tokutaro Hagiwara
had proposed that a uranium 235 bomb could ignite significant fusion
reactions in hydrogen. The Soviet Union successfully tested an
H-bomb dropped from an airplane in 1955, one year before the
United States did so.
Teller became the scientific adviser on nuclear affairs of many
presidents, from Dwight D. Eisenhower to Ronald Reagan. The
widespread blast and fallout effects of H-bombs assured the mutual
destruction of the users of such weapons. During the Cold War
(from about 1947 to 1981), both the United States and the Soviet
Union possessed H-bombs. “Testing” these bombs made each side
aware of how powerful the other side was. Everyone wanted to
avoid nuclear war. It was thought that no one would try to start a
war that would end in the world’s destruction. This theory was
called deterrence: The United States wanted to let the Soviet Union
know that it had just as many bombs, or more, than it did, so that the
leaders of the Sovet Union would be deterred from starting a war.Teller knew that the availability of H-bombs on both sides was
not enough to guarantee that such weapons would never be used. It
was also necessary to make the Soviet Union aware of the existence
of the bombs through testing. He consistently advised against U.S.
participation with the Soviet Union in a moratorium (period of
waiting) on nuclear weapons testing. Largely based on Teller’s urging
that underground testing be continued, the United States rejected
a total moratorium in favor of the 1963 Atmospheric Test Ban
Treaty.
During the 1980’s, Teller, among others, convinced President
Reagan to embrace the Strategic Defense Initiative (SDI). Teller argued
that SDI components, such as the space-based “Excalibur,” a
nuclear bomb-powered X-ray laser weapon proposed by the Lawrence-
Livermore National Laboratory, would make thermonuclear
war not unimaginable, but theoretically impossible.

Sunday, July 19, 2009

Hovercraft





The invention: A vehicle requiring no surface contact for traction
that moves freely over a variety of surfaces—particularly
water—while supported on a self-generated cushion of air.
The people behind the invention:
Christopher Sydney Cockerell (1910- ), a British engineer
who built the first hovercraft
Ronald A. Shaw (1910- ), an early pioneer in aerodynamics
who experimented with hovercraft
Sir John Isaac Thornycroft (1843-1928), a Royal Navy architect
who was the first to experiment with air-cushion theory
Air-Cushion Travel
The air-cushion vehicle was first conceived by Sir John Isaac
Thornycroft of Great Britain in the 1870’s. He theorized that if a
ship had a plenum chamber (a box open at the bottom) for a hull
and it were pumped full of air, the ship would rise out of the water
and move faster, because there would be less drag. The main problem
was keeping the air from escaping from under the craft.
In the early 1950’s, Christopher Sydney Cockerell was experimenting
with ways to reduce both the wave-making and frictional
resistance that craft had to water. In 1953, he constructed a punt
with a fan that supplied air to the bottom of the craft, which could
thus glide over the surface with very little friction. The air was contained
under the craft by specially constructed side walls. In 1955,
the first true “hovercraft,” as Cockerell called it, was constructed of
balsa wood. It weighed only 127 grams and traveled over water at a
speed of 13 kilometers per hour.
On November 16, 1956, Cockerell successfully demonstrated
his model hovercraft at the patent agent’s office in London. It was
immediately placed on the “secret” list, and Saunders-Roe Ltd.
was given the first contract to build hovercraft in 1957. The first experimental
piloted hovercraft, the SR.N1, which had a weight of
3,400 kilograms and could carry three people at the speed of 25 knots, was completed on May 28, 1959, and publicly demonstrated
on June 11, 1959.
Ground Effect Phenomenon
In a hovercraft, a jet airstream is directed downward through a
hole in a metal disk, which forces the disk to rise. The jet of air has a
reverse effect of its own that forces the disk away from the surface.
Some of the air hitting the ground bounces back against the disk to
add further lift. This is called the “ground effect.” The ground effect
is such that the greater the under-surface area of the hovercraft, the
greater the reverse thrust of the air that bounces back. This makes
the hovercraft a mechanically efficient machine because it provides
three functions.
First, the ground effect reduces friction between the craft and the
earth’s surface. Second, it acts as a spring suspension to reduce
some of the vertical acceleration effects that arise from travel over
an uneven surface. Third, it provides a safe and comfortable ride at
high speed, whatever the operating environment. The air cushion
can distribute the weight of the hovercraft over almost its entire area
so that the cushion pressure is low.
The basic elements of the air-cushion vehicle are a hull, a propulsion
system, and a lift system. The hull, which accommodates the
crew, passengers, and freight, contains both the propulsion and lift
systems. The propulsion and lift systems can be driven by the same
power plant or by separate power plants. Early designs used only
one unit, but this proved to be a problem when adequate power was
not achieved for movement and lift. Better results are achieved
when two units are used, since far more power is used to lift the vehicle
than to propel it.
For lift, high-speed centrifugal fans are used to drive the air
through jets that are located under the craft. A redesigned aircraft
propeller is used for propulsion. Rudderlike fins and an air fan that
can be swiveled to provide direction are placed at the rear of the
craft.
Several different air systems can be used, depending on whether
a skirt system is used in the lift process. The plenum chamber system,
the peripheral jet system, and several types of recirculating air systems have all been successfully tried without skirting. Avariety
of rigid and flexible skirts have also proved to be satisfactory, depending
on the use of the vehicle.
Skirts are used to hold the air for lift. Skirts were once hung like curtains around hovercraft. Instead of simple curtains to contain the air,
there are now complicated designs that contain the cushion, duct the
air, and even provide a secondary suspension. The materials used in
the skirting have also changed from a rubberized fabric to pure rubber
and nylon and, finally, to neoprene, a lamination of nylon and plastic.
The three basic types of hovercraft are the amphibious, nonamphibious,
and semiamphibious models. The amphibious type can
travel over water and land, whereas the nonamphibious type is restricted
to water travel. The semiamphibious model is also restricted
to water travel but may terminate travel by nosing up on a prepared
ramp or beach. All hovercraft contain built-in buoyancy tanks in the
side skirting as a safety measure in the event that a hovercraft must
settle on the water. Most hovercraft are equipped with gas turbines
and use either propellers or water-jet propulsion.
Impact
Hovercraft are used primarily for short passenger ferry services.
Great Britain was the only nation to produce a large number of hovercraft.
The British built larger and faster craft and pioneered their
successful use as ferries across the English Channel, where they
could reach speeds of 111 kilometers per hour (160 knots) and carry
more than four hundred passengers and almost one hundred vehicles.
France and the former Soviet Union have also effectively demonstrated
hovercraft river travel, and the Soviets have experimented
with military applications as well.
The military adaptations of hovercraft have been more diversified.
Beach landings have been performed effectively, and the United
States used hovercraft for river patrols during the Vietnam War.
Other uses also exist for hovercraft. They can be used as harbor pilot
vessels and for patrolling shores in a variety of police-and customs-
related duties. Hovercraft can also serve as flood-rescue craft
and fire-fighting vehicles. Even a hoverfreighter is being considered.
The air-cushion theory in transport systems is rapidly developing.
It has spread to trains and smaller people movers in many
countries. Their smooth, rapid, clean, and efficient operation makes
hovercraft attractive to transportation designers around the world.

Thursday, July 16, 2009

Holography




The invention: A lensless system of three-dimensional photography
that was one of the most important developments in twentieth
century optical science.
The people behind the invention:
Dennis Gabor (1900-1979), a Hungarian-born inventor and
physicist who was awarded the 1971 Nobel Prize in Physics
Emmett Leith (1927- ), a radar researcher who, with Juris
Upatnieks, produced the first laser holograms
Juris Upatnieks (1936- ), a radar researcher who, with
Emmett Leith, produced the first laser holograms
Easter Inspiration
The development of photography in the early 1900’s made possible
the recording of events and information in ways unknown before
the twentieth century: the photographing of star clusters, the
recording of the emission spectra of heated elements, the storing of
data in the form of small recorded images (for example, microfilm),
and the photographing of microscopic specimens, among other
things. Because of its vast importance to the scientist, the science of
photography has developed steadily.
An understanding of the photographic and holographic processes
requires some knowledge of the wave behavior of light. Light is an
electromagnetic wave that, like a water wave, has an amplitude and a
phase. The amplitude corresponds to the wave height, while the
phase indicates which part of the wave is passing a given point at a
given time. A cork floating in a pond bobs up and down as waves
pass under it. The position of the cork at any time depends on both
amplitude and phase: The phase determines on which part of the
wave the cork is floating at any given time, and the amplitude determines
how high or low the cork can be moved. Waves from more
than one source arriving at the cork combine in ways that depend on
their relative phases. If the waves meet in the same phase, they add
and produce a large amplitude; if they arrive out of phase, they subtract and produce a small amplitude. The total amplitude, or intensity,
depends on the phases of the combining waves.
Dennis Gabor, the inventor of holography, was intrigued by the
way in which the photographic image of an object was stored by a
photographic plate but was unable to devote any consistent research
effort to the question until the 1940’s. At that time, Gabor was involved
in the development of the electron microscope. On Easter
morning in 1947, as Gabor was pondering the problem of how to
improve the electron microscope, the solution came to him. He
would attempt to take a poor electron picture and then correct it optically.
The process would require coherent electron beams—that is,
electron waves with a definite phase.
This two-stage method was inspired by the work of Lawrence
Bragg. Bragg had formed the image of a crystal lattice by diffracting
the photographic X-ray diffraction pattern of the original lattice.
This double diffraction process is the basis of the holographic process.
Bragg’s method was limited because of his inability to record
the phase information of the X-ray photograph. Therefore, he could
study only those crystals for which the phase relationship of the reflected
waves could be predicted.
Waiting for the Laser
Gabor devised a way of capturing the phase information after he
realized that adding coherent background to the wave reflected from
an object would make it possible to produce an interference pattern
on the photographic plate. When the phases of the two waves are
identical, a maximum intensity will be recorded; when they are out of
phase, a minimum intensity is recorded. Therefore, what is recorded
in a hologram is not an image of the object but rather the interference
pattern of the two coherent waves. This pattern looks like a collection
of swirls and blank spots. The hologram (or photograph) is then illuminated
by the reference beam, and part of the transmitted light is a
replica of the original object wave. When viewing this object wave,
one sees an exact replica of the original object.
The major impediment at the time in making holograms using
any form of radiation was a lack of coherent sources. For example,
the coherence of the mercury lamp used by Gabor and his assistant IvorWilliams was so short that they were able to make holograms of
only about a centimeter in diameter. The early results were rather
poor in terms of image quality and also had a double image. For this
reason, there was little interest in holography, and the subject lay almost
untouched for more than ten years.
Interest in the field was rekindled after the laser (light amplification
by stimulated emission of radiation) was developed in 1962.
Emmett Leith and Juris Upatnieks, who were conducting radar research
at the University of Michigan, published the first laser holographs
in 1963. The laser was an intense light source with a very
long coherence length. Its monochromatic nature improved the resolution
of the images greatly. Also, there was no longer any restriction
on the size of the object to be photographed.
The availability of the laser allowed Leith and Upatnieks to propose
another improvement in holographic technique. Before 1964,
holograms were made of only thin transparent objects. A small region
of the hologram bore a one-to-one correspondence to a region
of the object. Only a small portion of the image could be viewed at
one time without the aid of additional optical components. Illuminating
the transparency diffusely allowed the whole image to be
seen at one time. This development also made it possible to record
holograms of diffusely reflected three-dimensional objects. Gabor
had seen from the beginning that this should make it possible to create
three-dimensional images.
After the early 1960’s, the field of holography developed very
quickly. Because holography is different from conventional photography,
the two techniques often complement each other. Gabor saw
his idea blossom into a very important technique in optical science.
Impact
The development of the laser and the publication of the first laser
holograms in 1963 caused a blossoming of the new technique in
many fields. Soon, techniques were developed that allowed holograms
to be viewed with white light. It also became possible for holograms
to reconstruct multicolored images. Holographic methods
have been used to map terrain with radar waves and to conduct surveillance
in the fields of forestry, agriculture, and meteorology.By the 1990’s, holography had become a multimillion-dollar industry,
finding applications in advertising, as an art form, and in security
devices on credit cards, as well as in scientific fields. An alternate
form of holography, also suggested by Gabor, uses sound
waves. Acoustical imaging is useful whenever the medium around
the object to be viewed is opaque to light rays—for example, in
medical diagnosis. Holography has affected many areas of science,
technology, and culture.

Monday, July 13, 2009

Heat pump



The invention:Adevice that warms and cools buildings efficiently
and cheaply by moving heat from one area to another.
The people behind the invention:
T. G. N. Haldane, a British engineer
Lord Kelvin (William Thomson, 1824-1907), a British
mathematician, scientist, and engineer
Sadi Carnot (1796-1832), a French physicist and
thermodynamicist
The Heat Pump
A heat pump is a device that takes in heat at one temperature and
releases it at a higher temperature. When operated to provide heat (for
example, for space heating), the heat pump is said to operate in the
heating mode; when operated to remove heat (for example, for air conditioning),
it is said to operate in the cooling mode. Some type of work
must be done to drive the pump, no matter which mode is being used.
There are two general types of heat pumps: vapor compression
pumps and absorption pumps. The basic principle of vapor compression
cycle heat pumps is derived from the work of Sadi Carnot
in the early nineteenth century. Carnot’s work was published in
1824. It wasWilliam Thomson (later to become known as Lord Kelvin),
however, who first proposed a practical heat pump system, or
“heat multiplier,” as it was known then, and he also indicated that a
refrigerating machine could be used for heating.
Thomson’s heat pump used air as its working fluid. Thomson
claimed that his heat pump was able to produce heat by using only
3 percent of the energy that would be required for direct heating.
Absorption cycle machines have an even longer history. Refrigerators
based on the use of sulfuric acid and water date back to 1777.
Systems using this fluid combination, improved and modified by
Edmond Carré, were used extensively in Paris cafés in the late
1800’s. In 1849, a patent was filed by Ferdinand Carré for the working-
fluid pair of ammonia and water in absorption cycle machines.In the early nineteenth century, many people (including some
electrical engineers) believed that electrical energy could never be
used economically to produce large quantities of heat under ordinary
conditions. A few researchers, however, believed that it was
possible to produce heat by using electrical energy if that energy
was first converted to mechanical energy and if the Carnot principle
was then used to pump heat from a lower to a higher temperature.
In 1927, T. G. N. Haldane carried out detailed experiments showing
that the heat pump can be made to operate in either the heating
mode or the cooling mode. Aheat pump in the cooling mode works
like a refrigerator; a heat pump in the heating mode supplies heat
for heating. Haldane demonstrated that a refrigerator could be
modified to work as a heating unit. He used a vapor compression
cycle refrigerator for his demonstration.
In the design of a refrigerating device, the primary objective is
the production of cold rather than heat, but the two operations are
complementary. The process of producing cold is simply that of
pumping heat from a relatively cold to a relatively hot source, but in
the refrigeration process particular attention is paid to the prevention
of the leakage of heat into the cold source, whereas no attempt
is made to prevent the escape of heat from the hot source. If a refrigerating
device were treated as a heat pump in which the primary
product is the heat rejected to the hot source, the order of importance
would be reversed, and every opportunity would be taken to
allow heat to leak into the cold source and every precaution would
be taken against allowing heat to leak out of the hot source.
The components of a heat pump that operates on the principle of
vapor compression include an electric motor, a compressor, an evaporator,
and a condenser. The compressor sucks in gas from the evaporator
and compresses it to a pressure that corresponds to a saturation
temperature that is slightly higher than that of the required heat. From
the compressor, the compressed gas passes to the condenser, where it is
cooled and condensed, thereby giving up a large quantity of heat to the
water or other substance that it is intended to heat. The condensed gas
then passes through the expansion valve, where a sudden reduction of
pressure takes place. This reduction of pressure lowers the boiling point of the liquid, which therefore vaporizes and takes in heat from
the medium surrounding the evaporator. After evaporation, the gas
passes on to the compressor, and the cycle is complete.
Haldane was the first person in the United Kingdom to install a
heat pump. He was also the first person to install a domestic heat
pump to provide hot water and space heating.Impact
Since Haldane’s demonstration of the use of the heat pump, the
device has been highly successful in people’s homes, especially in
those regions where both heating and cooling are required for single-
and multifamily residences (for example, Australia, Japan, and
the United States). This is the case because the heat pump can provide
both heating and cooling; therefore, the cost of a heat pump
system can be spread over both heating and cooling seasons. Total
annual sales of heat pumps worldwide have risen to the millions,
with most sales being made in Japan and the United States.
The use of heat pumps can save energy. In addition, because they
are electric, they can save significant quantities of oil, especially in
the residential retrofit and replacement markets and when used as
add-on devices for existing heating systems. Some heat pumps are
now available that may compete cost-effectively with other heating
systems in meeting the heating demands of cooler regions.Technological developments by heat pump manufacturers are
continually improving the performance and cost-effectiveness of
heat pumps. The electric heat pump will continue to dominate the
residential market, although engine-driven systems are likely to
have a greater impact on the multifamily market.

Heart-lung machine




The invention: The first artificial device to oxygenate and circulate
blood during surgery, the heart-lung machine began the era of
open-heart surgery.
The people behind the invention:
John H. Gibbon, Jr. (1903-1974), a cardiovascular surgeon
Mary Hopkinson Gibbon (1905- ), a research technician
Thomas J. Watson (1874-1956), chairman of the board of IBM
T. L. Stokes and J. B. Flick, researchers in Gibbon’s laboratory
Bernard J. Miller (1918- ), a cardiovascular surgeon and
research associate
Cecelia Bavolek, the first human to undergo open-heart surgery
successfully using the heart-lung machine
A Young Woman’s Death
In the first half of the twentieth century, cardiovascular medicine
had many triumphs. Effective anesthesia, antiseptic conditions, and
antibiotics made surgery safer. Blood-typing, anti-clotting agents,
and blood preservatives made blood transfusion practical. Cardiac
catheterization (feeding a tube into the heart), electrocardiography,
and fluoroscopy (visualizing living tissues with an X-ray machine)
made the nonsurgical diagnosis of cardiovascular problems possible.
As of 1950, however, there was no safe way to treat damage or defects
within the heart. To make such a correction, this vital organ’s
function had to be interrupted. The problem was to keep the body’s
tissues alive while working on the heart. While some surgeons practiced
so-called blind surgery, in which they inserted a finger into the
heart through a small incision without observing what they were attempting
to correct, others tried to reduce the body’s need for circulation
by slowly chilling the patient until the heart stopped. Still other
surgeons used “cross-circulation,” in which the patient’s circulation
was connected to a donor’s circulation. All these approaches carried
profound risks of hemorrhage, tissue damage, and death.
In February of 1931, Gibbon witnessed the death of a young woman whose lung circulation was blocked by a blood clot. Because
her blood could not pass through her lungs, she slowly lost
consciousness from lack of oxygen. As he monitored her pulse and
breathing, Gibbon thought about ways to circumvent the obstructed
lungs and straining heart and provide the oxygen required. Because
surgery to remove such a blood clot was often fatal, the woman’s
surgeons operated only as a last resort. Though the surgery took
only six and one-half minutes, she never regained consciousness.
This experience prompted Gibbon to pursue what few people then
considered a practical line of research: a way to circulate and oxygenate
blood outside the body.
A Woman’s Life Restored
Gibbon began the project in earnest in 1934, when he returned to
the laboratory of Edward D. Churchill at Massachusetts General
Hospital for his second surgical research fellowship. He was assisted
by Mary Hopkinson Gibbon. Together, they developed, using
cats, a surgical technique for removing blood froma vein, supplying
the blood with oxygen, and returning it to an artery using tubes inserted
into the blood vessels. Their objective was to create a device
that would keep the blood moving, spread it over a very thin layer
to pick up oxygen efficiently and remove carbon dioxide, and avoid
both clotting and damaging blood cells. In 1939, they reported that
prolonged survival after heart-lung bypass was possible in experimental
animals.
WorldWar II (1939-1945) interrupted the progress of this work; it
was resumed by Gibbon at Jefferson Medical College in 1944. Shortly
thereafter, he attracted the interest of Thomas J.Watson, chairman of
the board of the International Business Machines (IBM) Corporation,
who provided the services of IBM’s experimental physics laboratory
and model machine shop as well as the assistance of staff engineers.
IBM constructed and modified two experimental machines
over the next seven years, and IBM engineers contributed significantly
to the evolution of a machine that would be practical in humans.
Gibbon’s first attempt to use the pump-oxygenator in a human
being was in a fifteen-month-old baby. This attempt failed, not because of a malfunction or a surgical mistake but because of a misdiagnosis.
The child died following surgery because the real problem
had not been corrected by the surgery.
On May 6, 1953, the heart-lung machine was first used successfully
on Cecelia Bavolek. In the six months before surgery, Bavolek
had been hospitalized three times for symptoms of heart failure
when she tried to engage in normal activity. While her circulation
was connected to the heart-lung machine for forty-five minutes, the
surgical team headed by Gibbon was able to close an opening between
her atria and establish normal heart function. Two months
later, an examination of the defect revealed that it was fully closed;
Bavolek resumed a normal life. The age of open-heart surgery had
begun.
Consequences
The heart-lung bypass technique alone could not make openheart
surgery truly practical. When it was possible to keep tissues
alive by diverting blood around the heart and oxygenating it, other
questions already under investigation became even more critical:
how to prolong the survival of bloodless organs, how to measure
oxygen and carbon dioxide levels in the blood, and how to prolong
anesthesia during complicated surgery. Thus, following the first
successful use of the heart-lung machine, surgeons continued to refine
the methods of open-heart surgery.
The heart-lung apparatus set the stage for the advent of “replacement
parts” for many types of cardiovascular problems. Cardiac
valve replacement was first successfully accomplished in 1960 by
placing an artificial ball valve between the left atrium and ventricle.
In 1957, doctors performed the first coronary bypass surgery, grafting
sections of a leg vein into the heart’s circulation system to divert
blood around clogged coronary arteries. Likewise, the first successful
heart transplant (1967) and the controversial Jarvik-7 artificial
heart implantation (1982) required the ability to stop the heart and
keep the body’s tissues alive during time-consuming and delicate
surgical procedures. Gibbon’s heart-lung machine paved the way
for all these developments.

Thursday, July 9, 2009

Hearing aid




The invention: Miniaturized electronic amplifier worn inside the
ears of hearing-impaired persons.
The organization behind the invention:
Bell Labs, the research and development arm of the American
Telephone and Telegraph Company
Trapped in Silence
Until the middle of the twentieth century, people who experienced
hearing loss had little hope of being able to hear sounds without the
use of large, awkward, heavy appliances. For many years, the only
hearing aids available were devices known as ear trumpets. The ear
trumpet tried to compensate for hearing loss by increasing the number
of sound waves funneled into the ear canal. A wide, bell-like
mouth similar to the bell of a musical trumpet narrowed to a tube that
the user placed in his or her ear. Ear trumpets helped a little, but they
could not truly increase the volume of the sounds heard.
Beginning in the nineteenth century, inventors tried to develop
electrical devices that would serve as hearing aids. The telephone
was actually a by-product of Alexander Graham Bell’s efforts to
make a hearing aid. Following the invention of the telephone, electrical
engineers designed hearing aids that employed telephone
technology, but those hearing aids were only a slight improvement
over the old ear trumpets. They required large, heavy battery packs
and used a carbon microphone similar to the receiver in a telephone.
More sensitive than purely physical devices such as the ear trumpet,
they could transmit a wider range of sounds but could not amplify
them as effectively as electronic hearing aids now do.
Transistors Make Miniaturization Possible
Two types of hearing aids exist: body-worn and head-worn.
Body-worn hearing aids permit the widest range of sounds to be
heard, but because of the devices’ larger size, many hearing impaired persons do not like to wear them. Head-worn hearing
aids, especially those worn completely in the ear, are much less conspicuous.
In addition to in-ear aids, the category of head-worn hearing
aids includes both hearing aids mounted in eyeglass frames and
those worn behind the ear.
All hearing aids, whether head-worn or body-worn, consist of
four parts: a microphone to pick up sounds, an amplifier, a receiver,
and a power source. The microphone gathers sound waves and converts
them to electrical signals; the amplifier boosts, or increases,
those signals; and the receiver then converts the signals back into
sound waves. In effect, the hearing aid is a miniature radio. After
the receiver converts the signals back to sound waves, those waves
are directed into the ear canal through an earpiece or ear mold. The
ear mold generally is made of plastic and is custom fitted from an
impression taken from the prospective user’s ear.
Effective head-worn hearing aids could not be built until the
electronic circuit was developed in the early 1950’s. The same invention—
the transistor—that led to small portable radios and tape
players allowed engineers to create miniaturized, inconspicuous
hearing aids. Depending on the degree of amplification required,
the amplifier in a hearing aid contains three or more transistors.
Transistors first replaced vacuum tubes in devices such as radios
and phonographs, and then engineers realized that they could be
used in devices for the hearing-impaired.
The research at Bell Labs that led to the invention of the transistor
rose out of military research duringWorldWar II. The vacuum tubes
used in, for example, radar installations to amplify the strength of electronic
signals were big, were fragile because they were made of
blown glass, and gave off high levels of heat when they were used.
Transistors, however, made it possible to build solid-state, integrated
circuits. These are made from crystals of metals such as germanium
or arsenic alloys and therefore are much less fragile than glass. They
are also extremely small (in fact, some integrated circuits are barely
visible to the naked eye) and give off no heat during use.
The number of transistors in a hearing aid varies depending upon
the amount of amplification required. The first transistor is the most
important for the listener in terms of the quality of sound heard. If the
frequency response is set too high—that is, if the device is too sensitive—the listener will be bothered by distracting background noise.
Theoretically, there is no limit on the amount of amplification that a
hearing aid can be designed to provide, but there are practical limits.
The higher the amplification, the more power is required to operate
the hearing aid. This is why body-worn hearing aids can convey a
wider range of sounds than head-worn devices can. It is the power
source—not the electronic components—that is the limiting factor. A
body-worn hearing aid includes a larger battery pack than can be
used with a head-worn device. Indeed, despite advances in battery
technology, the power requirements of a head-worn hearing aid are
such that a 1.4-volt battery that could power a wristwatch for several
years will last only a few days in a hearing aid.
Consequences
The invention of the electronic hearing aid made it possible for
many hearing-impaired persons to participate in a hearing world.
Prior to the invention of the hearing aid, hearing-impaired children
often were unable to participate in routine school activities or function
effectively in mainstream society. Instead of being able to live at
home with their families and enjoy the same experiences that were
available to other children their age, often they were forced to attend
special schools operated by the state or by charities.
Hearing-impaired people were singled out as being different and
were limited in their choice of occupations. Although not every
hearing-impaired person can be helped to hear with a hearing aid—
particularly in cases of total hearing loss—the electronic hearing aid
has ended restrictions for many hearing-impaired people. Hearingimpaired
children are now included in public school classes, and
hearing-impaired adults can now pursue occupations from which
they were once excluded.
Today, many deaf and hearing-impaired persons have chosen to
live without the help of a hearing aid. They believe that they are not
disabled but simply different, and they point out that their “disability”
often allows them to appreciate and participate in life in unique
and positive ways. For them, the use of hearing aids is a choice, not a
necessity. For those who choose, hearing aids make it possible to
participate in the hearing world.

Hard disk




The invention: A large-capacity, permanent magnetic storage device
built into most personal computers.
The people behind the invention:
Alan Shugart (1930- ), an engineer who first developed the
floppy disk
Philip D. Estridge (1938?-1985), the director of IBM’s product
development facility
Thomas J. Watson, Jr. (1914-1993), the chief executive officer of
IBM
The Personal Oddity
When the International Business Machines (IBM) Corporation
introduced its first microcomputer, called simply the IBM PC (for
“personal computer”), the occasion was less a dramatic invention
than the confirmation of a trend begun some years before. A number
of companies had introduced microcomputers before IBM; one
of the best known at that time was Apple Corporation’s Apple II, for
which software for business and scientific use was quickly developed.
Nevertheless, the microcomputer was quite expensive and
was often looked upon as an oddity, not as a useful tool.
Under the leadership of Thomas J. Watson, Jr., IBM, which had
previously focused on giant mainframe computers, decided to develop
the PC. A design team headed by Philip D. Estridge was assembled
in Boca Raton, Florida, and it quickly developed its first,
pacesetting product. It is an irony of history that IBM anticipated
selling only one hundred thousand or so of these machines, mostly
to scientists and technically inclined hobbyists. Instead, IBM’s product
sold exceedingly well, and its design parameters, as well as its
operating system, became standards.
The earliest microcomputers used a cassette recorder as a means
of mass storage; a floppy disk drive capable of storing approximately
160 kilobytes of data was initially offered only as an option.
While home hobbyists were accustomed to using a cassette recorder for storage purposes, such a system was far too slow and awkward
for use in business and science. As a result, virtually every IBM PC
sold was equipped with at least one 5.25-inch floppy disk drive.
Memory Requirements
All computers require memory of two sorts in order to carry out
their tasks. One type of memory is main memory, or random access
memory (RAM), which is used by the computer’s central processor
to store data it is using while operating. The type of memory used
for this function is built typically of silicon-based integrated circuits
that have the advantage of speed (to allow the processor to fetch or
store the data quickly), but the disadvantage of possibly losing or
“forgetting” data when the electric current is turned off. Further,
such memory generally is relatively expensive.
To reduce costs, another type of memory—long-term storage
memory, known also as “mass storage”—was developed. Mass
storage devices include magnetic media (tape or disk drives) and
optical media (such as the compact disc, read-only memory, or CDROM).
While the speed with which data may be retrieved from or
stored in such devices is rather slow compared to the central processor’s
speed, a disk drive—the most common form of mass storage
used in PCs—can store relatively large amounts of data quite inexpensively.
Early floppy disk drives (so called because the magnetically
treated material on which data are recorded is made of a very flexible
plastic) held 160 kilobytes of data using only one side of the
magnetically coated disk (about eighty pages of normal, doublespaced,
typewritten information). Later developments increased
storage capacities to 360 kilobytes by using both sides of the disk
and later, with increasing technological ability, 1.44 megabytes (millions
of bytes). In contrast, mainframe computers, which are typically
connected to large and expensive tape drive storage systems,
could store gigabytes (millions of megabytes) of information.
While such capacities seem large, the needs of business and scientific
users soon outstripped available space. Since even the mailing
list of a small business or a scientist’s mathematical model of a
chemical reaction easily could require greater storage potential than early PCs allowed, the need arose for a mass storage device that
could accommodate very large files of data.
The answer was the hard disk drive, also known as a “fixed disk
drive,” reflecting the fact that the disk itself is not only rigid but also
permanently installed inside the machine. In 1955, IBM had envisioned
the notion of a fixed, hard magnetic disk as a means of storing
computer data, and, under the direction of Alan Shugart in the
1960’s, the floppy disk was developed as well.
As the engineers of IBM’s facility in Boca Raton refined the idea
of the original PC to design the new IBM PC XT, it became clear that
chief among the needs of users was the availability of large-capability
storage devices. The decision was made to add a 10-megabyte
hard disk drive to the PC. On March 8, 1983, less than two years after
the introduction of its first PC, IBM introduced the PC XT. Like
the original, it was an evolutionary design, not a revolutionary one.
The inclusion of a hard disk drive, however, signaled that mass storage
devices in personal computers had arrived.
Consequences
Above all else, any computer provides a means for storing, ordering,
analyzing, and presenting information. If the personal computer
is to become the information appliance some have suggested
it will be, the ability to manipulate very large amounts of data will
be of paramount concern. Hard disk technology was greeted enthusiastically
in the marketplace, and the demand for hard drives has
seen their numbers increase as their quality increases and their
prices drop.
It is easy to understand one reason for such eager acceptance:
convenience. Floppy-bound computer users find themselves frequently
changing (or “swapping”) their disks in order to allow programs
to find the data they need. Moreover, there is a limit to how
much data a single floppy disk can hold. The advantage of a hard
drive is that it allows users to keep seemingly unlimited amounts of
data and programs stored in their machines and readily available.
Also, hard disk drives are capable of finding files and transferring
their contents to the processor much more quickly than a
floppy drive. A user may thus create exceedingly large files, keep them on hand at all times, and manipulate data more quickly than
with a floppy. Finally, while a hard drive is a slow substitute for
main memory, it allows users to enjoy the benefits of larger memories
at significantly lower cost.
The introduction of the PC XT with its 10-megabyte hard drive
was a milestone in the development of the PC. Over the next two decades,
the size of computer hard drives increased dramatically. By
2001, few personal computers were sold with hard drives with less
than three gigabytes of storage capacity, and hard drives with more
than thirty gigabytes were becoming the standard. Indeed, for less
money than a PC XT cost in the mid-1980’s, one could buy a fully
equipped computer with a hard drive holding sixty gigabytes—a
storage capacity equivalent to six thousand 10-megabyte hard drives.

Gyrocompass




The invention: The first practical navigational device that enabled
ships and submarines to stay on course without relying on the
earth’s unreliable magnetic poles.
The people behind the invention:
Hermann Anschütz-Kaempfe (1872-1931), a German inventor
and manufacturer
Jean-Bernard-Léon Foucault (1819-1868), a French experimental
physicist and inventor
Elmer Ambrose Sperry (1860-1930), an American engineer and
inventor
From Toys to Tools
A gyroscope consists of a rapidly spinning wheel mounted in a
frame that enables the wheel to tilt freely in any direction. The
amount of momentum allows the wheel to maintain its “attitude”
even when the whole device is turned or rotated.
These devices have been used to solve problems arising in such
areas as sailing and navigation. For example, a gyroscope aboard a
ship maintains its orientation even while the ship is rolling. Among
other things, this allows the extent of the roll to be measured accurately.
Moreover, the spin axis of a free gyroscope can be adjusted to
point toward true north. It will (with some exceptions) stay that
way despite changes in the direction of a vehicle in which it is
mounted. Gyroscopic effects were employed in the design of various
objects long before the theory behind them was formally
known. A classic example is a child’s top, which balances, seemingly
in defiance of gravity, as long as it continues to spin. Boomerangs
and flying disks derive stability and accuracy from the spin
imparted by the thrower. Likewise, the accuracy of rifles improved
when barrels were manufactured with internal spiral grooves that
caused the emerging bullet to spin.
In 1852, the French inventor Jean-Bernard-Léon Foucault built
the first gyroscope, a measuring device consisting of a rapidly spinning
wheel mounted within concentric rings that allowed the wheel to move freely about two axes. This device, like the Foucault pendulum,
was used to demonstrate the rotation of the earth around its
axis, since the spinning wheel, which is not fixed, retains its orientation
in space while the earth turns under it. The gyroscope had a related
interesting property: As it continued to spin, the force of the
earth’s rotation caused its axis to rotate gradually until it was oriented
parallel to the earth’s axis, that is, in a north-south direction. It
is this property that enables the gyroscope to be used as a compass.
When Magnets Fail
In 1904, Hermann Anschütz-Kaempfe, a German manufacturer
working in the Kiel shipyards, became interested in the navigation
problems of submarines used in exploration under the polar ice cap.
By 1905, efficient working submarines were a reality, and it was evident
to all major naval powers that submarines would play an increasingly
important role in naval strategy.
Submarine navigation posed problems, however, that could not
be solved by instruments designed for surface vessels. Asubmarine
needs to orient itself under water in three dimensions; it has no automatic
horizon with respect to which it can level itself. Navigation
by means of stars or landmarks is impossible when the submarine is
submerged. Furthermore, in an enclosed metal hull containing machinery
run by electricity, a magnetic compass is worthless. To a
lesser extent, increasing use of metal, massive moving parts, and
electrical equipment had also rendered the magnetic compass unreliable
in conventional surface battleships.
It made sense for Anschütz-Kaempfe to use the gyroscopic effect
to design an instrument that would enable a ship to maintain its
course while under water. Yet producing such a device would not be
easy. First, it needed to be suspended in such a way that it was free to
turn in any direction with as little mechanical resistance as possible.
At the same time, it had to be able to resist the inevitable pitching and
rolling of a vessel at sea. Finally, a continuous power supply was required
to keep the gyroscopic wheels spinning at high speed.
The original Anschütz-Kaempfe gyrocompass consisted of a pair
of spinning wheels driven by an electric motor. The device was connected
to a compass card visible to the ship’s navigator. Motor, gyroscope, and suspension system were mounted in a frame that allowed
the apparatus to remain stable despite the pitch and roll of the ship.
In 1906, the German navy installed a prototype of the Anschütz-
Kaempfe gyrocompass on the battleship Undine and subjected it to
exhaustive tests under simulated battle conditions, sailing the ship
under forced draft and suddenly reversing the engines, changing the
position of heavy turrets and other mechanisms, and firing heavy
guns. In conditions under which a magnetic compass would have
been worthless, the gyrocompass proved a satisfactory navigational
tool, and the results were impressive enough to convince the German
navy to undertake installation of gyrocompasses in submarines and
heavy battleships, including the battleship Deutschland.
Elmer Ambrose Sperry, a New York inventor intimately associated
with pioneer electrical development, was independently working on a design for a gyroscopic compass at about the same time.
In 1907, he patented a gyrocompass consisting of a single rotor
mounted within two concentric shells, suspended by fine piano
wire from a frame mounted on gimbals. The rotor of the Sperry
compass operated in a vacuum, which enabled it to rotate more
rapidly. The Sperry gyrocompass was in use on larger American
battleships and submarines on the eve ofWorldWar I (1914-1918).
Impact
The ability to navigate submerged submarines was of critical
strategic importance in World War I. Initially, the German navy
had an advantage both in the number of submarines at its disposal
and in their design and maneuverability. The German U-boat fleet
declared all-out war on Allied shipping, and, although their efforts
to blockade England and France were ultimately unsuccessful, the
tremendous toll they inflicted helped maintain the German position
and prolong the war. To a submarine fleet operating throughout
the Atlantic and in the Caribbean, as well as in near-shore European
waters, effective long-distance navigation was critical.
Gyrocompasses were standard equipment on submarines and
battleships and, increasingly, on larger commercial vessels during
World War I, World War II (1939-1945), and the period between the
wars. The devices also found their way into aircraft, rockets, and
guided missiles. Although the compasses were made more accurate
and easier to use, the fundamental design differed little from that invented
by Anschütz-Kaempfe.

Sunday, July 5, 2009

Geothermal power





The invention: Energy generated from the earth’s natural hot
springs.
The people behind the invention:
Prince Piero Ginori Conti (1865-1939), an Italian nobleman and
industrialist
Sir Charles Parsons (1854-1931), an English engineer
B. C. McCabe, an American businessman
Developing a Practical System
The first successful use of geothermal energy was at Larderello in
northern Italy. The Larderello geothermal field, located near the city
of Pisa about 240 kilometers northwest of Rome, contains many hot
springs and fumaroles (steam vents). In 1777, these springs were
found to be rich in boron, and in 1818, Francesco de Larderel began
extracting the useful mineral borax from them. Shortly after 1900,
Prince Piero Ginori Conti, director of the Larderello borax works,
conceived the idea of using the steam for power production. An experimental
electrical power plant was constructed at Larderello in
1904 to provide electric power to the borax plant. After this initial
experiment proved successful, a 250-kilowatt generating station
was installed in 1913 and commercial power production began.
As the Larderello field grew, additional geothermal sites throughout
the region were prospected and tapped for power. Power production
grew steadily until the 1940’s, when production reached
130 megawatts; however, the Larderello power plants were destroyed
late inWorldWar II (1939-1945). After the war, the generating
plants were rebuilt, and they were producing more than 400
megawatts by 1980.
The Larderello power plants encountered many of the technical
problems that were later to concern other geothermal facilities. For
example, hydrogen sulfide in the steam was highly corrosive to copper,
so the Larderello power plant used aluminum for electrical connections
much more than did conventional power plants of the time. Also, the low pressure of the steam in early wells at Larderello
presented problems. The first generators simply used steam to drive
a generator and vented the spent steam into the atmosphere. Asystem
of this sort, called a “noncondensing system,” is useful for small
generators but not efficient to produce large amounts of power.
Most steam engines derive power not only from the pressure of
the steam but also from the vacuum created when the steam is condensed
back to water. Geothermal systems that generate power
from condensation, as well as direct steam pressure, are called “condensing
systems.” Most large geothermal generators are of this
type. Condensation of geothermal steam presents special problems
not present in ordinary steam engines: There are other gases present
that do not condense. Instead of a vacuum, condensation of steam
contaminated with other gases would result in only a limited drop
in pressure and, consequently, very low efficiency.
Initially, the operators of Larderello tried to use the steam to heat
boilers that would, in turn, generate pure steam. Eventually, a device
was developed that removed most of the contaminating gases from
the steam. Although later wells at Larderello and other geothermal
fields produced steam at greater pressure, these engineering innovations
improved the efficiency of any geothermal power plant.
Expanding the Idea
In 1913, the English engineer Sir Charles Parsons proposed drilling
an extremely deep (12-kilometer) hole to tap the earth’s deep
heat. Power from such a deep hole would not come from natural
steam as at Larderello but would be generated by pumping fluid
into the hole and generating steam (as hot as 500 degrees Celsius) at
the bottom. In modern terms, Parsons proposed tapping “hot dryrock”
geothermal energy. (No such plant has been commercially operated
yet, but research is being actively pursued in several countries.)
The first use of geothermal energy in the United States was for direct
heating. In 1890, the municipal water company of Boise, Idaho,
began supplying hot water from a geothermal well. Water was
piped from the well to homes and businesses along appropriately
namedWarm Springs Avenue. At its peak, the system served more than four hundred customers, but as cheap natural gas became
available, the number declined.
Although Larderello was the first successful geothermal electric
power plant, the modern era of geothermal electric power began
with the opening of the Geysers Geothermal Field in California.
Early attempts began in the 1920’s, but it was not until 1955 that B.
C. McCabe, a Los Angeles businessman, leased 14.6 square kilometers
in the Geysers area and founded the Magma Power Company.
The first 12.5-megawatt generator was installed at the Geysers in
1960, and production increased steadily from then on. The Geysers
surpassed Larderello as the largest producing geothermal field in
the 1970’s, and more than 1,000 megawatts were being generated by
1980. By the end of 1980, geothermal plants had been installed in
thirteen countries, with a total capacity of almost 2,600 megawatts,
and projects with a total capacity of more than 15,000 megawatts
were being planned in more than twenty countries.
Impact
Geothermal power has many attractive features. Because the
steam is naturally heated and under pressure, generating equipment
can be simple, inexpensive, and quickly installed. Equipment
and installation costs are offset by savings in fuel. It is economically
practical to install small generators, a fact that makes geothermal
plants attractive in remote or underdeveloped areas. Most important
to a world faced with a variety of technical and environmental
problems connected with fossil fuels, geothermal power does not
deplete fossil fuel reserves, produces little pollution, and contributes
little to the greenhouse effect.
Despite its attractive features, geothermal power has some limitations.
Geologic settings suitable for easy geothermal power production
are rare; there must be a hot rock or magma body close to
the surface. Although it is technically possible to pump water from
an external source into a geothermal well to generate steam, most
geothermal sites require a plentiful supply of natural underground
water that can be tapped as a source of steam. In contrast, fossil-fuel
generating plants can be at any convenient location.

Genetically engineered insulin




The invention: Artificially manufactured human insulin (Humulin)
as a medication for people suffering from diabetes.
The people behind the invention:
Irving S. Johnson (1925- ), an American zoologist who was
vice president of research at Eli Lilly Research Laboratories
Ronald E. Chance (1934- ), an American biochemist at Eli
Lilly Research Laboratories
What Is Diabetes?
Carbohydrates (sugars and related chemicals) are the main food
and energy source for humans. In wealthy countries such as the
United States, more than 50 percent of the food people eat is made
up of carbohydrates, while in poorer countries the carbohydrate
content of diets is higher, from 70 to 90 percent.
Normally, most carbohydrates that a person eats are used (or metabolized)
quickly to produce energy. Carbohydrates not needed for
energy are either converted to fat or stored as a glucose polymer
called “glycogen.” Most adult humans carry about a pound of body
glycogen; this substance is broken down to produce energy when it
is needed.
Certain diseases prevent the proper metabolism and storage of
carbohydrates. The most common of these diseases is diabetes mellitus,
usually called simply “diabetes.” It is found in more than seventy
million people worldwide. Diabetic people cannot produce or
use enough insulin, a hormone secreted by the pancreas. When their
condition is not treated, the eyes may deteriorate to the point of
blindness. The kidneys may stop working properly, blood vessels
may be damaged, and the person may fall into a coma and die. In
fact, diabetes is the third most common killer in the United States.
Most of the problems surrounding diabetes are caused by high levels
of glucose in the blood. Cataracts often form in diabetics, as excess
glucose is deposited in the lens of the eye.
Important symptoms of diabetes include constant thirst, excessive urination, and large amounts of sugar in the blood and in the
urine. The glucose tolerance test (GTT) is the best way to find out
whether a person is suffering from diabetes. People given a GTT are
first told to fast overnight. In the morning their blood glucose level
is measured; then they are asked to drink about a fourth of a pound
of glucose dissolved in water. During the next four to six hours, the
blood glucose level is measured repeatedly. In nondiabetics, glucose
levels do not rise above a certain amount during a GTT, and the
level drops quickly as the glucose is assimilated by the body. In diabetics,
the blood glucose levels rise much higher and do not drop as
quickly. The extra glucose then shows up in the urine.
Treating Diabetes
Until the 1920’s, diabetes could be controlled only through a diet
very low in carbohydrates, and this treatment was not always successful.
Then Sir Frederick G. Banting and Charles H. Best found a
way to prepare purified insulin from animal pancreases and gave it
to patients. This gave diabetics their first chance to live a fairly normal
life. Banting and his coworkers won the 1923 Nobel Prize in
Physiology or Medicine for their work.
The usual treatment for diabetics became regular shots of insulin.
Drug companies took the insulin from the pancreases of cattle and
pigs slaughtered by the meat-packing industry. Unfortunately, animal
insulin has two disadvantages. First, about 5 percent of diabetics
are allergic to it and can have severe reactions. Second, the world
supply of animal pancreases goes up and down depending on how
much meat is being bought. Between 1970 and 1975, the supply of
insulin fell sharply as people began to eat less red meat, yet the
numbers of diabetics continued to increase. So researchers began to
look for a better way to supply insulin.
Studying pancreases of people who had donated their bodies to
science, researchers found that human insulin did not cause allergic
reactions. Scientists realized that it would be best to find a chemical
or biological way to prepare human insulin, and pharmaceutical
companies worked hard toward this goal. Eli Lilly and Company
was the first to succeed, and on May 14, 1982, it filed a new drug application
with the Food and Drug Administration (FDA) for the human insulin preparation it named “Humulin.”
Humulin is made by genetic engineering. Irving S. Johnson, who
worked on the development of Humulin, described Eli Lilly’s method
for producing Humulin. The common bacterium Escherichia coli
is used. Two strains of the bacterium are produced by genetic engineering:
The first strain is used to make a protein called an “A
chain,” and the second strain is used to make a “B chain.” After the
bacteria are harvested, the Aand B chains are removed and purified
separately. Then the two chains are combined chemically. When
they are purified once more, the result is Humulin, which has been
proved by Ronald E. Chance and his Eli Lilly coworkers to be chemically,
biologically, and physically identical to human insulin.
Consequences
The FDA and other regulatory agencies around the world approved
genetically engineered human insulin in 1982. Humulin
does not trigger allergic reactions, and its supply does not fluctuate.
It has brought an end to the fear that there would be a worldwide
shortage of insulin.
Humulin is important as well in being the first genetically engineered
industrial chemical. It began an era in which such advanced
technology could be a source for medical drugs, chemicals used in
farming, and other important industrial products. Researchers hope
that genetic engineering will help in the understanding of cancer
and other diseases, and that it will lead to ways to grow enough
food for a world whose population continues to rise.