How PC’s/Microprocessors Work1
The computer you are using to read this page uses a
microprocessor to do its work. The microprocessor is the heart of
any normal computer, whether it is a desktop machine, a server or a
laptop. The microprocessor you are using might be a Pentium, a K6,
a PowerPC, a Sparc or any of the many other brands and types of
microprocessors, but they all do approximately the same thing in
approximately the same way.
A microprocessor -- also known as a CPU or central processing unit -
- is a complete computation engine that is fabricated on a single chip.
The first microprocessor was the Intel 4004, introduced in 1971. The
4004 was not very powerful -- all it could do was add and subtract,
and it could only do that 4 bits at a time. But it was amazing that
everything was on one chip. Prior to the 4004, engineers built computers either from collections of chips
or from discrete components (transistors wired one at a time). The 4004 powered one of the first portable
electronic calculators.
If you have ever wondered what the microprocessor in your computer is doing, or if you have ever
wondered about the differences between types of microprocessors, then read on. In this article, you will
learn how fairly simple digital logic techniques allow a computer to do its job, whether it’s playing a game
or spell checking a document!
Microprocessor Progression: Intel
The first microprocessor to make it into a home computer was the Intel
8080, a complete 8-bit computer on one chip, introduced in 1974. The first
microprocessor to make a real splash in the market was the Intel 8088,
introduced in 1979 and incorporated into the IBM PC (which first appeared
around 1982). If you are familiar with the PC market and its history, you
know that the PC market moved from the 8088 to the 80286 to the 80386 to
the 80486 to the Pentium to the Pentium II to the Pentium III to the Pentium
4. All of these microprocessors are made by Intel and all of them are improvements on the basic design
of the 8088. The Pentium 4 can execute any piece of code that ran on the original 8088, but it does it
about 5,000 times faster!
Since 2004, Intel has introduced microprocessors with multiple cores and millions more transistors. But
even these microprocessors follow the same general rules as earlier chips.
The date is the year that the processor
was first introduced. Many processors are
re-introduced at higher clock speeds for
many years after the original release
date.
Transistors is the number of transistors
on the chip. You can see that the number
of transistors on a single chip has risen
steadily over the years.
Microns is the width, in microns, of the
smallest wire on the chip. For
comparison, a human hair is 100 microns
thick. As the feature size on the chip goes down, the number of transistors rises.
Clock speed is the maximum rate that the chip can be clocked at. Clock speed will make more sense in
the next section.
Data Width is the width of the ALU. An 8-bit ALU can add/subtract/multiply/etc. two 8-bit numbers, while
a 32-bit ALU can manipulate 32-bit numbers. An 8-bit ALU would have to execute four instructions to add
two 32-bit numbers, while a 32-bit ALU can do it in one instruction. In many cases, the external data bus
is the same width as the ALU, but not always. The 8088 had a 16-bit ALU and an 8-bit bus, while the
modern Pentiums fetch data 64 bits at a time for their 32-bit ALUs.
MIPS stands for "millions of instructions per second" and is a rough measure of the performance of a
CPU. Modern CPUs can do so many different things that MIPS ratings lose a lot of their meaning, but you
can get a general sense of the relative power of the CPUs from this column.
From this table you can see that, in general, there is a relationship between clock speed and MIPS. The
maximum clock speed is a function of the manufacturing process and delays within the chip. There is also
a relationship between the number of transistors and MIPS. For example, the 8088 clocked at 5 MHz but
only executed at 0.33 MIPS (about one instruction per 15 clock cycles). Modern processors can often
execute at a rate of two instructions per clock cycle. That improvement is directly related to the number of
transistors on the chip and will make more sense in the next section.
Microprocessor Memory
The previous section talked about the address and data buses, as well as the RD and WR lines. These
buses and lines connect either to RAM or ROM -- generally both. In our sample microprocessor, we have
an address bus 8 bits wide and a data bus 8 bits wide. That means that the microprocessor can address
(28) 256 bytes of memory, and it can read or write 8 bits of the memory at a time. Let's assume that this
simple microprocessor has 128 bytes of ROM starting at address 0 and 128 bytes of RAM starting at
address 128.
ROM stands for read-only memory. A ROM chip is programmed with a
permanent collection of pre-set bytes. The address bus tells the ROM chip
which byte to get and place on the data bus. When the RD line changes state, the ROM chip presents the
selected byte onto the data bus.
RAM stands for random-access memory. RAM contains bytes of information, and the microprocessor
can read or write to those bytes depending on whether the RD or WR line is signaled. One problem with
today's RAM chips is that they forget everything once the power goes off. That is why the computer needs
ROM.
By the way, nearly all computers contain some amount of ROM (it is possible to create a simple computer
that contains no RAM -- many microcontrollers do this by placing a handful of RAM bytes on the
processor chip itself -- but generally impossible to create one that contains no ROM). On a PC, the ROM
is called the BIOS (Basic Input/Output System). When the microprocessor starts, it begins executing
instructions it finds in the BIOS. The BIOS instructions do things like test the hardware in the machine,
and then it goes to the hard disk to fetch the boot sector (see How Hard Disks Work for details). This
boot sector is another small program, and the BIOS stores it in RAM after reading it off the disk. The
microprocessor then begins executing the boot sector's instructions from RAM. The boot sector program
will tell the microprocessor to fetch something else from the hard disk into RAM, which the
microprocessor then executes, and so on. This is how the microprocessor loads and executes the
entire operating system.
Processor History Details2 Pictures Only
1969: Four-Phase Systems AL1 The AL1 was an 8-bit bit slice chip containing eight registers and an ALU It was designed by Lee Boysel in 1969 At the time, it formed part of a nine-chip, 24-bit CPU with three AL1s, but it was later called a microprocessor when, in response to 1990s litigation by Texas Instruments, a demonstration system was constructed where a single AL1 formed part of a courtroom demonstration computer system, together with RAM, ROM, and an input-output device
1971: Pico/GI PICO1/GI250 In 1971 Pico Electronics and General Instrument (GI) introduced their first collaboration in ICs, a complete single chip calculator IC for the Monroe/Litton Royal Digital III calculator. This chip could also arguably lay claim to be one of the first microprocessors or microcontrollers having ROM, RAM and a RISC instruction set on-chip. The layout for the four layers of the PMOS process was hand drawn at x500 scale on mylar film, a significant task at the time given the complexity of the chip.
1971: Intel 4004 The 4004 was Intel's first microprocessor. This breakthrough invention powered the Busicom calculator and paved the way for embedding intelligence in inanimate objects as well as the personal computer.
1972: Intel 8008 The 8008 was twice as powerful as the 4004. A 1974 article in Radio Electronics referred to a device called the Mark-8 which used the 8008. The Mark-8 is known as one of the first computers for the home --one that by today's standards was difficult to build, maintain and operate.
1974: Intel 8080 The 8080 became the brains of the first personal computer--the Altair, allegedly named for a destination of the Starship Enterprise from the Star Trek television show. Computer hobbyists could purchase a kit for the Altair for $395. Within months, it
sold tens of thousands, creating the first PC back orders in history.
1974: Texas Instruments TMS1000
1974: National Semiconductor "Pace" 16Bit
1975: Motorola 6502
1976: Zilog Z80
1978: Intel 8086-8088 A pivotal sale to IBM's new personal computer division made the 8088 the brains of IBM's new hit product--the IBM PC. The 8088's success propelled Intel into the ranks of the Fortune 500, and Fortune magazine named the company one of the "Business Triumphs of the Seventies."
1979: Motorola 68000
1981: HP 32 bit
1982: Intel 80286 The 80286, originally known as the 80286, was the first Intel processor that could run all the software written for its predecessor. This software compatibility remains a hallmark of Intel's family of microprocessors. Within 6 years of its release, an estimated 15 million 80286-based personal computers were installed around the world.
1985: Intel 80386 The Intel386™ microprocessor featured 275,000 transistors--more than 100times as many as the original 4004. It was a 32-bit chip and was "multi tasking," meaning it could run multiple programs at the same time.
1989: Intel 80486 The 80486 processor generation really meant you go from a command-level computer into point-and-click computing. "I could have a color computer for the first time and do desktop publishing at a significant speed," recalls technology historian David K. Allison of the Smithsonian's National Museum of American History. The 80486 processor was the first to offer a built-in math coprocessor, which speeds up computing because it offloads complex math functions from the central processor.
1989: AMD486DX
1993: Intel Pentium The Pentium processor allowed computers to more easily incorporate "real world" data such as speech, sound, handwriting and photographic images. The Intel Pentium brand, mentioned in the comics and on television talk shows, became a household word soon after introduction.
1995: Intel Pentium Pro Released in the fall of 1995 the Pentium Pro processor is designed to fuel 32-bit server and workstation applications, enabling fast computer-aided design, mechanical engineering and scientific computation. Each Pentium Pro processor is packaged together with a second speed-enhancing cache memory chip. The powerful Pentium Pro processor boasts 5.5 million transistors.
1997: Intel Pentium II The 7.5 million-transistor Pentium II processor incorporates Intel MMX™ technology, which is designed specifically to process video, audio and graphics data efficiently. It was introduced in innovative Single Edge Contact (S.E.C) Cartridge that
also incorporated a high-speed cache memory chip. With this chip, PC users can capture, edit and share digital photos with friends and family via the Internet; edit and add text, music or between-scene transitions to home movies; and, with a video phone, send video over standard phone lines and the Internet.
1998: Intel Pentium II Xeon The Pentium II Xeon processors are designed to meet the performance requirements of mid-range and higher servers and workstations. Consistent with Intel's strategy to deliver unique processor products targeted for specific markets segments, the Intel® Pentium II Xeon processors feature technical innovations specifically designed for workstations and servers that utilize demanding business applications such as Internet services, corporate data warehousing, digital content creation, and electronic and mechanical design automation. Systems based on the processor can be configured to scale to four or eight processors and beyond.
1999: Intel Celeron Continuing Intel's strategy of developing processors for specific market segments, the Celeron processor is designed for the value PC market segment. It provides consumers great performance at an exceptional price, and it delivers excellent performance for uses such as gaming and educational software.
1999: Intel Pentium III The Pentium III processor features 70 new instructions--Internet Streaming SIMD extensions-- that dramatically enhance the performance of advanced imaging, 3-D, streaming audio, video and speech recognition applications. It was designed to significantly enhance Internet experiences, allowing users to do such things as browse through realistic online museums and stores and download high-quality video. The processor incorporates 9.5 million transistors, and was introduced using 0.25-micron technology.
1999: Intel Pentium III Xeon™ The Pentium III Xeon processor extends Intel's offerings to the workstation and server market segments, providing additional performance for e-Commerce applications and advanced business computing. The processors incorporate the Pentium III processor's 70 SIMD instructions, which enhance multimedia and streaming video applications. The Pentium III Xeon processor's advance cache technology speeds information from the system bus to the processor, significantly boosting performance. It is designed for systems with multiprocessor configurations.
2000: Intel Pentium 4 Users of Pentium 4 processor-based PCs can create professional-quality movies; deliver TV-like video via the Internet; communicate with real-time video and voice; render 3D graphics in real time; quickly encode music for MP3 players; and simultaneously run several multimedia applications while connected to the Internet. The processor debuted with 42 million transistors and circuit lines of 0.18 microns. Intel's first microprocessor, the 4004, ran at 108 kilohertz (108,000 hertz), compared to the Intel® Pentium® 4 processor's initial speed of 1.5 gigahertz (1.5 billion hertz). If automobile speed had increased similarly over the same period, you could now drive from San Francisco to New York in about 13 seconds.
2001: Intel Xeon The Xeon processor is targeted for high-performance and mid-range, dual-processor workstations, dual and multi-processor server configurations coming in the future. The platform offers customers a choice of operating systems and applications, along with high performance at affordable prices. Intel Xeon processor-based workstations are expected to achieve performance increases between 30 and 90 percent over systems featuring Pentium III Xeon™ processors depending on applications and configurations. The processor is based on the Intel NetBurst™ architecture, which is designed to deliver the processing power needed for video and audio applications, advanced Internet technologies, and complex 3-D graphics.
2001: Intel Itanium The Itanium processor is the first in a family of 64-bit products from Intel. Designed for high-end, enterprise-class servers and workstations, the processor was built from the ground up with an entirely new architecture based on Intel's Explicitly Parallel Instruction Computing (EPIC) design technology. The processor delivers world-class performance for the most demanding enterprise and high-performance computing applications, including e-Commerce security transactions, large databases, mechanical computer-aided engineering, and sophisticated scientific and engineering computing.
2002: Intel Itanium 2 The Itanium™ 2 processor is the second member of the Itanium processor family, a line of enterprise-class processors. The family brings outstanding performance and the volume economics of the Intel Architecture to the most data-intensive, business-critical and technical computing applications. It provides leading performance for databases, computer-aided engineering, secure online transactions, and more.
2003: Intel Pentium M The Pentium M processor, the Intel 855 chipset family, and the PRO/Wireless 2100 network connection are the three components of Centrino™ mobile technology. Intel Centrino mobile technology is designed specifically for portable computing, with built-in wireless LAN capability and breakthrough mobile performance. It enables extended battery life and thinner, lighter mobile computers.
The Computer Mouse3
Mice first broke onto the public stage with the introduction of the Apple Macintosh in 1984, and since then
they have helped to completely redefine the way we use computers.
Every day of your computing life, you reach out for your mouse whenever you want to move your cursor
or activate something. Your mouse senses your motion and your clicks and sends them to the computer
so it can respond appropriately.
In this article we'll take the cover off of this important part of the human-machine interface and see exactly
what makes it tick.
Evolution of the Computer Mouse
It is amazing how simple and effective a mouse is, and it is also amazing how long it took mice to become
a part of everyday life. Given that people naturally point at things -- usually before they speak -- it is
surprising that it took so long for a good pointing device to develop. Although originally conceived in the
1960s, a couple of decades passed before mice became mainstream.
In the beginning, there was no need to point because computers used crude interfaces like teletype
machines or punch cards for data entry. The early text terminals did nothing more than emulate a teletype
(using the screen to replace paper), so it was many years (well into the 1960s and early 1970s) before
arrow keys were found on most terminals. Full screen editors were the first things to take real advantage
of the cursor keys, and they offered humans the first way to point.
Light pens were used on a variety of machines as a pointing device for many years, and graphics
tablets, joy sticks and various other devices were also popular in the 1970s. None of these really took off
as the pointing device of choice, however.
When the mouse hit the scene -- attached to the Mac, it was an immediate success. There is something
about it that is completely natural. Compared to a graphics tablet, mice are extremely inexpensive and
they take up very little desk space. In the PC world, mice took longer to gain ground, mainly because of a
lack of support in the operating system. Once Windows 3.1 made Graphical User Interfaces (GUIs) a
standard, the mouse became the PC-human interface of choice very quickly.
Inside a Mouse
The main goal of any mouse is to translate the motion of your hand into signals that the computer can
use. Let's take a look inside a track-ball mouse to see how it works:
1. A ball inside the mouse touches the desktop and rolls when the mouse moves. The underside of the
mouse's logic board: The exposed portion of the ball touches the desktop.
2. Two rollers inside the mouse touch the ball. One of the rollers is oriented so that it detects motion in the
X direction, and the other is oriented 90 degrees to the first roller so it detects motion in the Y direction.
When the ball rotates, one or both of these rollers rotate as well. The following image shows the two white
rollers on this mouse: The rollers that touch the ball and detect X and Y motion
3. The rollers each connect to a shaft, and the shaft spins a disk with holes in it. When a roller rolls, its shaft
and disk spin. The following image shows the disk: A typical optical encoding disk: This disk has 36
holes around its outer edge.
4. On either side of the disk there is an infrared LED and an infrared sensor. The holes in the disk break
the beam of light coming from the LED so that the infrared sensor sees pulses of light. The rate of the
pulsing is directly related to the speed of the mouse and the distance it travels. A close-up of one of the
optical encoders that track mouse motion: There is an infrared LED (clear) on one side of the disk
and an infrared sensor (red) on the other.
5. An on-board processor chip reads the pulses from the infrared sensors and turns them into binary data
that the computer can understand. The chip sends the binary data to the computer through the mouse's
cord.
The logic section of a mouse is dominated by an encoder chip, a small processor that reads the pulses coming from the infrared sensors and turns them into bytes sent to the computer. You can also see the two buttons that detect clicks (on either side of the wire connector).
In this optomechanical arrangement, the disk moves mechanically, and an optical system counts pulses
of light. On this mouse, the ball is 21 mm in diameter. The roller is 7 mm in diameter. The encoding disk
has 36 holes. So if the mouse moves 25.4 mm (1 inch), the encoder chip detects 41 pulses of light.
You might have noticed that each encoder disk has two infrared LEDs and two infrared sensors, one on
each side of the disk (so there are four LED/sensor pairs inside a mouse). This arrangement allows the
processor to detect the disk's direction of rotation. There is a piece of plastic with a small, precisely
located hole that sits between the encoder disk and each infrared sensor. It is visible in this photo:
A close-up of one of the optical encoders that track mouse motion: Note the piece of plastic between the infrared sensor (red) and the encoding disk.
This piece of plastic provides a window through which the infrared sensor can "see." The window on one
side of the disk is located slightly higher than it is on the other -- one-half the height of one of the holes in
the encoder disk, to be exact. That difference causes the two infrared sensors to see pulses of light at
slightly different times. There are times when one of the sensors will see a pulse of light when the other
does not, and vice versa.
Optical Mice
Developed by Agilent Technologies and introduced to the world in late 1999, the optical mouse actually
uses a tiny camera to take thousands of pictures every second.
Able to work on almost any surface without a mouse pad, most optical mice use a small, red light-emitting
diode (LED) that bounces light off that surface onto a complimentary metal-oxide semiconductor (CMOS)
sensor. In addition to LEDs, a recent innovation are laser-based optical mice that detect more surface
details compared to LED technology. This results in the ability to use a laser-based optical mouse on
even more surfaces than an LED mouse.
Here's how the sensor and other parts of an optical mouse work together:
The CMOS sensor sends each image to a digital signal processor (DSP) for analysis.
The DSP detects patterns in the images and examines how the patterns have moved since the previous
image.
Based on the change in patterns over a sequence of images, the DSP determines how far the mouse has
moved and sends the corresponding coordinates to the computer.
The computer moves the cursor on the screen based on the coordinates received from the mouse. This
happens hundreds of times each second, making the cursor appear to move very smoothly.
Optical mice have several benefits over track-ball mice:
No moving parts means less wear and a lower chance of failure.
There's no way for dirt to get inside the mouse and interfere with the tracking sensors.
Increased tracking resolution means a smoother response.
They don't require a special surface, such as a mouse pad.
Connecting Computer Mice
Most mice on the market today use a USB connector to attach to your computer. USB is a standard way
to connect all kinds of peripherals to your computer, including printers, digital cameras, keyboards and
mice.
Monitor Display Technology4
Often referred to as a monitor when packaged in a separate case, the display is the most-used output
device on a computer. The display provides instant feedback by showing you text and graphic images as
you work or play.
Most desktop displays use liquid crystal display (LCD); some older monitors use cathode ray tube (CRT)
technology (they are big and bulky compared to the thin LCD monitors), while nearly all portable
computing devices such as laptops incorporate LCD technology. Because of their slimmer design and
lower energy consumption, monitors using LCD technology (also called flat panel or flat screen displays)
are replacing the venerable CRT on most desktops. A new technology which will replace the LCD
monitors is LED, or light-emitting diodes; these tend to be brighter and have better contrast – but are also
more expensive.
Resolution refers to the number of individual dots of color, known as pixels, contained on a display.
Resolution is expressed by identifying the number of pixels on the horizontal axis (rows) and the number
on the vertical axis (columns), such as 800x600. Resolution is affected by a number of factors, including
the size of the screen.
As monitor sizes have increased over the years, display standards and resolutions have changed.
Common Display Standards and Resolutions
XGA (Extended Graphics Array) = 1024x768
SXGA (Super XGA) = 1280x1024
UXGA (Ultra XGA) = 1600x1200
QXGA (Quad XGA) = 2048x1536
WXGA (Wide XGA) = 1280x800
WSXGA+ (Wide SXGA plus) = 1680x1050
WUXGA (Wide Ultra XGA) = 1920x1200
WQHD = 2560 x 1440
WQXGA = 2560 x 1600
QSXGA = 2560 x 2048
In addition to the screen size, display standards and resolutions are related to something called
the aspect ratio. Next, we'll discuss what an aspect ratio is and how screen size is measured.
Aspect Ratio and Viewable Area
Two measures describe the size of your display: the aspect ratio and the screen size. Historically,
computer displays, like most televisions, have had an aspect ratio of 4:3. This means that the ratio of the
width of the display screen to the height is 4 to 3.
For widescreen LCD monitors, the aspect ratio is 16:9 (or sometimes 16:10 or 15:9). Widescreen LCD
displays are useful for viewing DVDmovies in widescreen format, playing games and displaying multiple
windows side by side.High definition television (HDTV) also uses a widescreen aspect ratio.
All types of displays include a projection surface, commonly referred to as the screen. Screen sizes are
normally measured in inches from one corner to the corner diagonally across from it. This diagonal
measuring system actually came about because the early television manufacturers wanted to make the
screen size of their TVs sound more impressive.
Interestingly, the way in which the screen size is measured for CRT and LCD monitors is different. For
CRT monitors, screen size is measured diagonally from outside edges of the display casing. In other
words, the exterior casing is included in the measurement as seen below.
For LCD monitors, screen size is measured diagonally from the inside of the beveled edge. The
measurement does not include the casing as indicated in the image below.
LCD screen size
Because of the differences in how CRT and LCD monitors
are measured, a 17-inch LCD display is comparable to a
19-inch CRT display. For a more accurate representation of
a CRT's size, find out its viewable screen size. This is the
measurement of a CRT display without its outside casing.
Popular screen sizes are 15, 17, 19 and 21 inches.
Notebook screen sizes are smaller, typically ranging from
12 to 17 inches. As technologies improve in both desktop
and notebook displays, even larger screen sizes are
becoming available. For professional applications, such as
medical imaging or public information displays, some LCD
monitors are 40 inches or larger!
Obviously, the size of the display directly affects resolution. The same pixel resolution is sharper on a
smaller monitor and fuzzier on a larger monitor because the same number of pixels is spread out over a
larger number of inches. An image on a 21-inch monitor with an 800x600 resolution will not appear nearly
as sharp as it would on a 15-inch display at 800x600.
Analog and DVI Connections
To display information on a monitor, your computer sends the monitor a signal. The signal can be in
analog or digital format.
Analog (VGA) Connection
Because most CRT monitors require the signal
information in analog (continuous electrical signals or
waves) form and not digital (pulses equivalent to the
binary digits 0 and 1), they typically use an analog
connection.
However, computers work in a digital world. The
computer and video adapter convert digital data into
analog format. A video adapter is an expansion card or component that provides the ability to convert
display information into a signal that is sent to the monitor. It can also be called a graphics adapter, video
card or graphics card.
Once the display information is in analog form, it is sent to the monitor through a VGA cable. The cable
connects at the back of the computer to an analog connector (also known as a D-Sub connector) that has
15 pins in three rows (see the diagram above):
1. Red out
2. Green out
3. Blue out
4. Unused
5. Ground
6. Red return (ground)
7. Green return (ground)
8. Blue return (ground)
9. Unused
10. Sync return (ground
11. Monitor ID 0 in
12. Monitor ID 1 in or data from display
13. Horizontal Sync out
14. Vertical Sync
15. Monitor ID 3 in or data clock
You can see that a VGA connector like this has three separate lines for the red, green and blue color
signals, and two lines for horizontal and vertical sync signals. In a normal television, all of these signals
are combined into a single composite video signal. The separation of the signals is one reason why a
computer monitor can have so many more pixels than a TV set.
Because a VGA (analog) connector does not support the use of digital monitors, the Digital Video
Interface (DVI) standard was developed.
DVI Connection
DVI keeps data in digital form from the computer to the monitor. There's no need to convert data from
digital information to analog information. LCD monitors work in a digital mode and support the DVI format.
(Although, some also accept analog information, which is then converted to digital format.) At one time, a
digital signal offered better image quality compared to analog technology. However, analog signal
processing technology has improved over the years and the difference in quality is now minimal.
The DVI specification is based on Silicon Image's Transition Minimized Differential Signaling (TMDS)
and provides a high-speed digital interface. A transmitter on the video adapter sends the digital
information to a receiver in the monitor. TMDS takes the signal from the video adapter, determines
the resolution andrefresh rate that the monitor is using, and spreads the signal out over the available
bandwidth to optimize the data transfer from computer to monitor.
DVI cables can be a single link cable that uses one TMDS transmitter or a dual link cable with two
transmitters. A single link DVI cable and connection supports a 1920x1080 image, and a dual link
cable/connection supports up to a 2048x1536 image.
DVI-D connectors carry a digital-only signal and DVI-I adds four pins for analog capability. Both connectors can be used with a single-link or a dual-link cable, depending upon the requirements of the display.
There are two main types of DVI connections:
DVI-digital (DVI-D) is a digital-only format. It requires a video adapter with a DVI-D connection and a
monitor with a DVI-D input. The connector contains 24 pins/receptacles in 3 rows of 8 plus a grounding
slot for dual-link support. For single-link support, the connector contains 18 pins/receptacles.
DVI-integrated (DVI-I) supports both digital and analog transmissions. This gives you the option to
connect a monitor that accepts digital input or analog input. In addition to the pins/receptacles found on
the DVI-D connector for digital support, a DVI-I connector has 4 additional pins/receptacles to carry an
analog signal.
If you buy a monitor with only a DVI (digital) connection, make sure that you have a video adapter with a
DVI-D or DVI-I connection. If your video adapter has only an analog (VGA) connection, look for a monitor
that supports the analog format.
References
1 How Stuff Works website. “How PC’s Work”. http://www.howstuffworks.com/microprocessor.htm
2. Microprocessor Hall of Fame website.
http://www.tayloredge.com/museum/processor/processorhistory.html
3. How Stuff Works website. “How Mice Work”. http://www.howstuffworks.com/mouse.htm
4. How Stuff Works website. “How Monitors Work”. http://www.howstuffworks.com/monitor1.htm
Top Related