Digital Cameras Essay, Research Paper
Overview
Digital
cameras capture images electronically and convert them into digital data
that can be stored and manipulated by a computer.
Like
conventional cameras, digital cameras have a lens, aperture, and shutter,
but they don’t use film. When light passes through the lens it is focused
on a photo-sensitive electronic chip called a charged coupling device
(CCD). The CCD converts light impulses into electrical impulses (also
called analog signal forms). The signals are fed into a microprocessor
and transformed into digital information. This process is called digitization.
Although
digital images do not yet match the quality of pictures produced on film,
they represent an enormously flexible medium. Photographers are no longer
limited by the physical properties of chemistry and optics. Computers
outfitted with the appropriate software can augment and transform images
in ways never before imagined.
History
The origins of digital cameras are intimately linked with
the evolution of television in the 1940s and 50s, and the development
of computer imaging by NASA in the
1960s.
Before the advent of the video tape recorder (VTR), television
images were optically displayed on monitors and then filmed by motion
picture cameras. Because film and television technologies were essentially
incompatible, Kinescopes, or "kinnys" as they were called, produced
inferior images.
A breakthrough occurred in 1951 when Bing Crosby Laboratories
introduced the VTR, a technology specifically designed to record television
images. Television cameras convert light waves into electronic impulses,
and the VTR records these impulses onto magnetic tape. Perfected in 1956
by the Ampex Corporation, video tape
recording produced clear, crisp and nearly flawless images. The use of
VTRs soon revolutionized the television industry.
The next great leap forward happened in the early 1960s
as NASA geared up for the Apollo Lunar Exploration Program. As a precursor
to landing humans on the moon, NASA sent out a series of probes to map
the lunar surface. The Ranger missions relied on video cameras outfitted
with transmitters that broadcast analog signals. These weak transmissions
were plagued by interference from natural radio sources like the Sun.
Conventional television receivers could not transform them into coherent
images.
Researchers at NASA’s Jet
Propulsion Laboratory (JPL) developed ways to "clean" and
enhance analog signals by processing them through computers. Signals were
analyzed by a computer and converted into numerical or digital information.
In this way, unwanted interference could be removed, while critical data
could be enhanced. By the time of the Ranger 7 mission, JPL was producing
crystal clear images of the moon’s surface. The age of digital imaging
had dawned.
Since that time, probes outfitted with digital imagers
have explored the boundaries of our solar system. The orbiting Hubble
telescope, a hybrid of optical and digital technology, maps the limits
of the known universe.
Here on earth, digital techniques gave rise to a host
of medical imaging devices, from improved X-ray imaging in the late 1960s,
to Magnetic Resonance Imaging and
Positron
Emission Tomography in the ’80s and ’90s.
A
Thousand Points of Light: How Digital Images Are Formed
Digital cameras come in several formats designed for the
specialized needs of photographers. They range from inexpensive snapshot
models to sophisticated scanner backs that fit on professional large format
film cameras. Regardless of their size or sophistication, all digital
cameras operate in much the same way.
All images we perceive are formed from optical light energy.
Even digital images created within a computer are eventually converted
into light energy that we can see. In order for a digital camera to store
an optical image, it must be converted into digital information.
A digital camera gathers light energy through a lens,
and focuses it on a CCD which converts it into electrical impulses. These
signals are fed into a microprocessor where they are sampled and transformed
into digital information. This numerical data is then stored, and usually
transferred later on to a computer where the image can be viewed and manipulated.
Black-and-White
Basics
A black-and-white photograph is composed of a wide range
of tonal variations. Like the spectrum of natural light it represents,
the photo’s tones are continuous and unbroken. By contrast, a black-and-white
digital image consists of myriad points of light sampled from the light
spectrum. A digital image’s range of tone is determined by the camera’s
capacity to sample and store different light values.
After the CCD converts light into an electrical signal,
it is sent to the image digitizer. The digitizer samples areas of light
and shadow from across the image, breaking them into points—or pixels.
The pixels are next quantized—assigned digital brightness values.
For black-and-white, this means placing the pixel on a numerical scale
that ranges from pure white to pure black. In color imaging, the process
includes scales for color resolution and chromatic intensity.
Spatial
Resolution:
Each pixel is assigned an x,y coordinate that corresponds to its place
and value in the optical image. The more pixels, the greater the image’s
range of tone. This quality is called spatial density, and is a
vital component of image quality. How good a picture looks is also affected
by optical resolution—meaning the camera’s optics and electronics.
Together, spatial density and optical resolution determine the image’s
spatial resolution; its tonal spectrum and clarity of detail.
In the end, spatial resolution is decided by the camera’s lesser most
quality: spatial density or optical resolution.
Spatial
Frequency
If crisp, clear pictures are the result of spatial density,
then a camera’s digitizer should sample an image as broadly and often
as possible. The digitizer’s ability to do this results in the image’s
spatial frequency.
Imagine a picture of a palm tree on a sandy beach. The
sky is bright blue with barely a cloud in the sky. The sand is golden,
and covered here and there by white breakers. The ocean is an unbroken
expanse of deep blue. The palm’s dark forest greens are broken by shafts
of filtered light.
When the digitizer scans this image it will find the sky,
beach and ocean fairly simple patterns of continuous tones. They vary
little in brightness or color; one sampled point of light is pretty much
the same as the next one. These areas have low spatial frequency.
The digitizer doesn’t need many samples to accurately read their tones.
The tree, however, with its deep shadows and brilliant
highlights, presents a greater challenge. Bright tones and dark tones
vary greatly from one pixel to the next. This rapid rate of tonal shifting
is called high spatial frequency. In order to build an accurate
representation, the digitizer needs many more samples than it does for
a low frequency area.
After determining the area of highest spatial frequency,
the digitizer calculates a sampling rate for the entire image. That speed
is double the rate of the image’s highest spatial frequency. In this way
the digitizer captures all of the scene’s subtle tonal nuance.
Of course, the camera’s sampling rate is not infinite,
especially in lower priced models. It’s ability to sample is limited by
its number of pixels. Pixel density depends on the amount of capacitors
on the CCD chip. This varies quite a bit between different makes and models
of cameras. Generally, cameras are assigned spatial frequency rates that
cover most situations photographers are likely to encounter.
Brightness
Resolution
The apparent brightness of an object in the real world
is quite different from its representation in a picture. Anyone who has
ever gazed at the sun instinctively knows the difference between the actual
object and a photograph of it. This may seem an academic distinction,
but it is a key concept in digital imaging.
The sun, the moon, the trees and flowers—everything
we see in our physical environment—possess radiant intensity.
They emit and reflect light energy. Paintings, photographs, and digital
images, on the other hand, possess luminous brightness. Though
they have radiant intensity, it is not the same intensity as the objects
they represent. The sun shown on a television or movie screen does not
have the radiant intensity of the actual celestial body. It is a representation.
In a digital photograph, each pixel has an assigned brightness
value—a luminous brightness—that corresponds to a radiant intensity
in the physical world. This value is determined by how many bits are in
the quantizer.
A 3-bit quantizer, for example, can only render a scale
of eight distinct tones ranging from pure white to pure black. If this
camera took a picture of our beach scene, it would create a high contrast
image with very few middle tones. This effect is called brightness
contouring, and is similar to the phenomenon of posterization in conventional
photography.
Brightness contouring has many pragmatic and creative
applications when an image is ready to be manipulated in a computer. However,
when capturing images with a camera, it’s best to preserve as wide a tonal
range as possible. Every bit added to a quantizer doubles its scale of
tones. Most modern digital cameras are equipped with 8-bit quantizers
capable of producing 256 different shades. Some professional quality cameras
have quantizers that can render well over a thousand tones.
Color
Resolution
Making digital images in color requires an additional
step. In black-and-white, the brightness resolution of a pixel is determined
by one gray value. In color, that value has three components, one for
each primary color, red, green or blue. This concept is called trichromacy.
Color digital cameras are outfitted with three different
sensors, each one sensitive to a primary waveband of light. After an image
is scanned and quantized, it is further broken down into color values.
Each pixel is assigned three color values which represent qualities of
red, green or blue. Color values are further distinguished by their hue
saturation and brightness.
Suppose, for example, a photographer snaps an image of
a pink balloon. The camera’s red sensor is stimulated and the quantizer
assigns the pixels that hue. Next, a saturation value is determined. Deep
red is a fully saturated color, while pink is much less saturated. It
is relatively faded and much closer to the white extreme of the scale.
Lastly, the brightness value determines the luminous intensity of the
color. Is this a pink balloon drifting through the shade of a forest?
Or does it float freely across a bright blue sky? These considerations
will compose the saturation and intensity of the image.
Digital
Imaging: From Camera to Computer
Most digital images form within a blink of the camera’s
shutter. In that fragmentary instant, an image made of light is transformed
into a stream of numerical data by a complex web of technologies. What’s
more, the image stored within the camera’s memory chip is only the beginning.
To be viewed and appreciated, the camera’s data must be uploaded into
a computer. Here, an imaginative photographer can alter and transform
the image in almost any way desired. With the proper software, even the
most mundane snapshot can evolve into a work of artistry.
The political, social and artistic ramifications of digital
imaging technology are yet to be ascertained. One thing is certain: the
way we create and perceive the fruits of human imagination will never
be the same.
Bibliography
Books
Baxes, Gregory, Digital Image Processing: Principles &
Applications, New York: John Wiley & Sons, Inc., 1994.
Brown, Les, Les Brown’s Encyclopedia of Television: Third
Edition, Detroit: Gale Research, 1992.
Grotta, Sally Wiener, and Grotta, Daniel, Digital Imaging
for Visual Artists, New York: Windcrest/McGraw-Hill, 1994.
Katz, Ephraim, The Film Encyclopedia:Second Edition, New
York: Harper Perennial, 1994.
Articles
Baig, Edward C., "Smile—You’re on Candid Computer,"
Business Week, 4 November 1996.
Diehl, Stanford, "Byte’s Video Workshop," Byte,
May 1995.
Joch, Alan, "Beyond Hollywood," Byte, May 1995.
Lu, Cary, "Digital Cameras on the Move," MacWorld,
June 1996.
McNamara, Michael J.,"New Imaging, Today & Tomorrow:
3 New Digital Cameras," Popular Photography, August 1996.
Wiener, Leonard, "Camcorders Go Pro," U.S. News
& World Report, 25 November 1996.
Zuckerman, Jim, "Digital Portraits," Petersen’s
Photographic, September 1996.
! |
Как писать рефераты Практические рекомендации по написанию студенческих рефератов. |
! | План реферата Краткий список разделов, отражающий структура и порядок работы над будующим рефератом. |
! | Введение реферата Вводная часть работы, в которой отражается цель и обозначается список задач. |
! | Заключение реферата В заключении подводятся итоги, описывается была ли достигнута поставленная цель, каковы результаты. |
! | Оформление рефератов Методические рекомендации по грамотному оформлению работы по ГОСТ. |
→ | Виды рефератов Какими бывают рефераты по своему назначению и структуре. |