digital computer

digital computer
a computer that processes information in digital form. Cf. analog computer.

* * *

Computer capable of solving problems by processing information expressed in discrete form.

By manipulating combinations of binary digits (see binary code), it can perform mathematical calculations, organize and analyze data, control industrial and other processes, and simulate dynamic systems such as global weather patterns. See also analog computer.

* * *


      any of a class of devices capable of solving problems by processing information in discrete form. It operates on data, including magnitudes, letters, and symbols, that are expressed in binary form—i.e., using only the two digits 0 and 1. By counting, comparing, and manipulating these digits or their combinations according to a set of instructions held in its memory, a digital computer can perform such tasks as to control industrial processes and regulate the operations of machines; analyze and organize vast amounts of business data; and simulate the behaviour of dynamic systems (e.g., global weather patterns and chemical reactions) in scientific research.

      A brief treatment of digital computers follows. For full treatment, see computer science: Basic computer components (computer science).

Functional elements
      A typical digital computer system has four basic functional elements: (1) input-output (input/output device) equipment, (2) main memory, (3) control unit, and (4) arithmetic-logic unit. Any of a number of devices is used to enter data and program instructions into a computer and to gain access to the results of the processing operation. Common input devices include keyboards and optical scanners; output devices include printers and cathode-ray tube and liquid-crystal display monitors. The information received by a computer from its input unit is stored in the main memory or, if not for immediate use, in an auxiliary storage device. The control unit selects and calls up instructions from the memory in appropriate sequence and relays the proper commands to the appropriate unit. It also synchronizes the varied operating speeds of the input and output devices to that of the arithmetic-logic unit (ALU) so as to ensure the proper movement of data through the entire computer system. The ALU performs the arithmetic and logic algorithms selected to process the incoming data at extremely high speeds—in many cases in nanoseconds (billionths of a second). The main memory, control unit, and ALU together make up the central processing unit (CPU) of most digital computer systems, while the input-output devices and auxiliary storage units constitute peripheral equipment.

Development of the digital computer
      Blaise Pascal (Pascal, Blaise) of France and Gottfried Wilhelm Leibniz (Leibniz, Gottfried Wilhelm) of Germany invented mechanical digital calculating machines during the 17th century. The English inventor Charles Babbage (Babbage, Charles), however, is generally credited with having conceived the first automatic digital computer. During the 1830s Babbage devised his so-called Analytical Engine, a mechanical device designed to combine basic arithmetic operations with decisions based on its own computations. Babbage's plans embodied most of the fundamental elements of the modern digital computer. For example, they called for sequential control—i.e., program control that included branching, looping, and both arithmetic and storage units with automatic printout. Babbage's device, however, was never completed and was forgotten until his writings were rediscovered over a century later.

      Of great importance in the evolution of the digital computer was the work of the English mathematician and logician George Boole (Boole, George). In various essays written during the mid-1800s, Boole discussed the analogy between the symbols of algebra and those of logic as used to represent logical forms and syllogisms. His formalism, operating on only 0 and 1, became the basis of what is now called Boolean algebra, on which computer switching theory and procedures are grounded.

 John V. Atanasoff (Atanasoff, John V(incent)), an American mathematician and physicist, is credited with building the first electronic digital computer, which he constructed from 1939 to 1942 with the assistance of his graduate student Clifford E. Berry. Konrad Zuse, a German engineer acting in virtual isolation from developments elsewhere, completed construction in 1941 of the first operational program-controlled calculating machine (Z3). In 1944 Howard Aiken (Aiken, Howard Hathaway) and a group of engineers at International Business Machines Corporation completed work on the Harvard Mark I, a machine whose data-processing operations were controlled primarily by electric relays (switching devices).

      Since the development of the Harvard Mark I, the digital computer has evolved at a rapid pace. The succession of advances in computer equipment, principally in logic circuitry, is often divided into generations, with each generation comprising a group of machines that share a common technology.

 In 1946 J. Presper Eckert (Eckert, J. Presper, Jr.) and John W. Mauchly (Mauchly, John W.), both of the University of Pennsylvania, constructed ENIAC (an acronym for electronic numerical integrator and computer), a digital machine and the first general-purpose, electronic computer. Its computing features were derived from Atanasoff's machine; both computers included vacuum tubes instead of relays as their active logic elements, a feature that resulted in a significant increase in operating speed. The concept of a stored-program computer was introduced in the mid-1940s, and the idea of storing instruction codes as well as data in an electrically alterable memory was implemented in EDVAC (electronic discrete variable automatic computer).

      The second computer generation began in the late 1950s, when digital machines utilizing transistors (transistor) became commercially available. Although this type of semiconductor device had been invented in 1948, more than 10 years of developmental work was needed to render it a viable alternative to the vacuum tube. The small size of the transistor, its greater reliability, and its relatively low power consumption made it vastly superior to the tube. Its use in computer circuitry permitted the manufacture of digital systems that were considerably more efficient, smaller, and faster than their first-generation ancestors.

      The late 1960s and '70s witnessed further dramatic advances in computer hardware. The first was the fabrication of the integrated circuit, a solid-state device containing hundreds of transistors, diodes, and resistors on a tiny silicon chip. This microcircuit made possible the production of mainframe (large-scale) computers of higher operating speeds, capacity, and reliability at significantly lower cost. Another type of third-generation computer that developed as a result of microelectronics was the minicomputer, a machine appreciably smaller than the standard mainframe but powerful enough to control the instruments of an entire scientific laboratory.

      The development of large-scale integration (LSI) enabled hardware manufacturers to pack thousands of transistors and other related components on a single silicon chip about the size of a baby's fingernail. Such microcircuitry yielded two devices that revolutionized computer technology. The first of these was the microprocessor, which is an integrated circuit that contains all the arithmetic, logic, and control circuitry of a central processing unit. Its production resulted in the development of microcomputers (microcomputer), systems no larger than portable television sets yet with substantial computing power. The other important device to emerge from LSI circuitry was the semiconductor memory. Consisting of only a few chips, this compact storage device is well-suited for use in minicomputers and microcomputers. Moreover, it has found use in an increasing number of mainframes, particularly those designed for high-speed applications, because of its fast-access speed and large storage capacity.

      By the beginning of the 1980s integrated circuitry had advanced to very large-scale integration (VLSI). This design and manufacturing technology greatly increased the circuit density of microprocessor, memory, and support chips—i.e., those that serve to interface microprocessors with input-output devices. By the 1990s some VLSI circuits contained more than 3 million transistors on a silicon chip less than 0.3 square inch (2 square cm) in area.

      The digital computers of the 1980s and '90s employing LSI and VLSI technologies are frequently referred to as fourth-generation systems. Many of the microcomputers produced during the 1980s were equipped with a single chip on which circuits for processor, memory, and interface functions were integrated. See also supercomputer.

* * *

Universalium. 2010.

Игры ⚽ Поможем сделать НИР

Look at other dictionaries:

  • digital computer — Computer Com*put er (k[o^]m*p[=u]t [ e]r), n. 1. One who computes. 2. (Computers) an electronic device for performing calculations automatically. It consists of a clock to provide voltage pulses to synchronize the operations of the devices within …   The Collaborative International Dictionary of English

  • digital computer — n. 1. (Computers) a computer that represents information by numerical digits with a fixed number of values; most commonly each piece of information is internally represented in binary code, as an array of {bits}, which are information units each… …   The Collaborative International Dictionary of English

  • digital computer — ☆ digital computer n. a computer for processing data represented by discrete, localized physical signals, as the presence or absence of an electric current: the most commonly used kind of computer: cf. ANALOG COMPUTER …   English World dictionary

  • digital computer — noun a computer that represents information by numerical (binary) digits • Hypernyms: ↑computer, ↑computing machine, ↑computing device, ↑data processor, ↑electronic computer, ↑information processing system …   Useful english dictionary

  • digital computer — skaitmeninė skaičiavimo mašina statusas T sritis automatika atitikmenys: angl. digital computer vok. Digitalrechner, m rus. ЦВМ; цифровая вычислительная машина, f pranc. ordinateur digital, m; ordinateur numéral, m; ordinateur numérique, m …   Automatikos terminų žodynas

  • digital computer — skaitmeninis kompiuteris statusas T sritis fizika atitikmenys: angl. digital computer vok. digitaler Rechner, m; Digitalrechenmaschine, f; Digitalrechner, m rus. цифровой компьютер, m pranc. calculateur numéral, m; calculateur numérique, m;… …   Fizikos terminų žodynas

  • digital computer — skaitmeninis kompiuteris statusas T sritis informatika apibrėžtis Kompiuteris, atliekantis veiksmus su ↑skaitmeniniais duomenimis. Elementarus skaitmeninio kompiuterio informacijos vienetas yra ↑bitas, kuriuo patogu pavaizduoti ↑dvejetainį… …   Enciklopedinis kompiuterijos žodynas

  • Digital Computer Museum — Computer History Museum Gebäude des CHM in Mountain View Daten Ort Mountain View Art …   Deutsch Wikipedia

  • digital computer — computer whose operations are based on a binary system …   English contemporary dictionary

  • digital computer — dig′ital comput′er n. cmp a computer that processes information in digital form Compare analog computer • Etymology: 1940–45 …   From formal English to slang

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”