The evolution of electronic computers over a period of time can be traced effectively dividing this period into various generations. Each generation is characterized by a major technological development that fundamentally changed the way computers operated. These helped to develop smaller, cheaper, powerful, efficient and reliable devices. Today, life has become indispensable without a computer. You find computerization in almost every sphere and industry. Computer evolution has been a fascinating process as we find out here.
The generation of computers may be broadly classified into 5 stages :
1. First Generation ( 1940 – 1956 )
2. Second Generation ( 1956 – 1963 )
3. Third Generation ( 1964 – 1971 )
4. Fourth generation ( 1971 – Present )
5. Fifth Generation ( present and Beyond )
FIRST GENERATION ( 1940 – 1956 )
World War gave rise to numerous developments and started off the computer age. Electronic Numerical Integrator and Computer (ENIAC) was produced by a partnership between University of Pennsylvannia and the US government. It consisted of 18,000 vacuum tubes and 7000 resistors. It was developed by John Presper Eckert and John W. Mauchly and was a general purpose computer. "Von Neumann designed the Electronic Discrete Variable Automatic Computer (EDVAC) in 1945 with a memory to hold both a stored program as well as data." Von Neumann's computer allowed for all the computer functions to be controlled by a single source.
Then in 1951 came the Universal Automatic Computer(UNIVAC I), designed by Remington rand and collectively owned by US census bureau and General Electric. UNIVAC amazingly predicted the winner of 1952, presidential elections, Dwight D. Eisenhower.
In first generation computers, the operating instructions or programs were specifically built for the task for which computer was manufactured. The Machine language was the only way to tell these machines to perform the operations. There was great difficulty to program these computers ,and more when there were some malfunctions.
The first generation of computers used vacuum tubes for circuitry and magnetic drums for memory. They were large in size, occupied a lot of space and produced enormous heat.
They were very expensive to operate and consumed large amount of electricity. Input was based on punched cards and paper tape and output was displayed on print outs. First generati0n computers could solve only one problem at a time.
SECOND GENERATION(1956– 1963)
The second generation of computers witnessed the vacuum tubes being replaced by the “ TRANSISTORS “. The transistor was far superb, faster, cheaper, energy-efficient and more reliable than their First- generation counter parts. The transistors also generated considerable heat that sometimes caused the computer to malfunction. But it was a vast improvement over the vacuum tube. Second-generation computers used punched cards for input and print outs for output.
Second generation computers moved from the use of machine language to assembly language, which allowed programmers to specify instructions in words. Even though complex in itself Assembly language was much easier than the binary code. High-level programming languages were also developed at this time, such as early version of COBOL (Common Business-Oriented Language) and FORTRAN (Formula Translator). The computers stored their instruction in their memory, which moved from a magnetic drum to magnetic core technology. Throughout the early 1960's, there were a number of commercially successful second generation computers used in businesses, universities, and government from companies such as Burroughs, Control Data, Honeywell, IBM, Sperry-Rand, and others. These second generation computers were also of solid state design, and contained transistors in place of vacuum tubes. They also contained all the components we associate with the modern day computer: printers, tape storage, disk storage, memory, and stored programs. One important example was the IBM 1401, which was universally accepted throughout industry, and is considered by many to be the Model T of the computer industry. By 1965, most large business routinely processed financial information using second generation computers (Gersting 218).
THIRD GENERATION ( 1964 – 1971 )
The development of the Integrated Circuit in 1958 by Jack Kilbey left its mark in the third generation of computers. Transistors were made smaller in size and placed on silicon chips, which dramatically increased the speed and efficiency of computers. As a result, computers became ever smaller as more components were squeezed onto the chip. Another third-generation development included the use of an operating system that allowed machines to run many different programs at once with a central program that monitored and coordinated the computer's memory . Fairchild Camera and Instrument Corp. built the first standard metal oxide semiconductor product for data processing applications, an eight-bit arithmetic unit and accumulator. The fundamental components of this semiconductor laid the groundwork for the future discovery of the microprocessor in 1971. Another company that took advantage of the third generation advancements was IBM with the unveiling of the IBM System/360. The company was making a transition from discrete transistors to integrated circuits, and its major source of revenue moved from punched-card equipment to electronic computer systems.
UNIX : In 1969 AT&T Bell Laboratories programmers Kenneth Thompson and Dennis Ritchie developed the UNIX operating system on a spare DEC minicomputer. UNIX was the first modern operating system that provided a sound intermediary between software and hardware. The UNIX operating system quickly secured a wide following, particularly among engineers and scientists at universities and other computer science organizations.
FOURTH GENERATION
( 1971 – Present )
The microprocessor brought the fourth generation of computers, as thousands of integrated circuits we rebuilt onto a single silicon chip. After the invention of the integrated circuit, the next step in the computer design process was to reduce the overall size. Large scale integration (LSI) could fit hundreds of components onto one chip. By the 1980's, very large scale integration (VLSI) squeezed hundreds of thousands of components onto a chip. Ultra-large scale integration (ULSI) increased that number into the millions. The ability to fit so much onto an area about half the size of a U.S. dime helped diminish the size and price of computers. It also increased their power, efficiency and reliability. The Intel 4004 chip, developed in 1971, took the integrated circuit one step further by locating all the components of a computer (central processing unit, memory, and input and output controls) on a minute chip. Whereas previously the integrated circuit had had to be manufactured to fit a special purpose, now one microprocessor could be manufactured and then programmed to meet any number of demands. Soon everyday household items such as microwave ovens, television sets, and automobiles with electronic fuel injection incorporated microprocessors (Gersting 35 - 39). Such condensed power allowed everyday people to harness a computer's power. They were no longer developed exclusively for large business or government contracts. By the mid-1970's, computer manufacturers sought to bring computers to general consumers. These minicomputers came complete with user-friendly software packages that offered even non-technical users an array of applications, most popularly word processing and spreadsheet programs. Pioneers in this field were Commodore, Radio Shack and Apple Computers. In the early 1980's, arcade video games such as Pac Man and home video game systems such as the Atari 2600 ignited consumer interest for more sophisticated, programmable home computers. In 1981, IBM introduced its personal computer (PC) for use in the home, office and schools. The 1980's saw an expansion in computer use in all three arenas as clones of the IBM PC made the personal computer even more affordable. The number of personal computers in use more than doubled from 2 million in 1981 to 5.5 million in 1982. Ten years later, 65 million PCs were being used. Computers continued their trend toward a smaller size, working their way down from desktop to laptop computers to palmtop. In direct competition with IBM's PC was Apple's Macintosh line, introduced in 1984. Notable for its user-friendly design, the Macintosh offered an operating system that allowed users to move screen icons instead of typing instructions. Users controlled the screen cursor using a mouse, a device that mimicked the movement of one's hand on the computer screen. As computers became more widespread in the workplace, new ways to harness their potential developed. As smaller computers became more powerful, they could be linked together, or networked, to share memory space, software, information and communicate with each other. As opposed to a mainframe computer, which was one powerful computer that shared time with many terminals for many applications, networked computers allowed individual computers to form electronic gateways. Using either direct wiring, called a Local Area Network (LAN), or telephone lines, these networks could reach enormous proportions. A global web of computer circuitry, the Internet, for example, links computers worldwide into a single network of information. During the 1992 U.S. presidential election, vice-presidential candidate Al Gore promised to make the development of this so-called "information superhighway" an administrative priority. The ideals expressed by Gore and others are in usage everyday through email, web browsing, and e-commerce.
As these small computers became more powerful, they could be linked together to form networks, which eventually led to the development of the Internet. Fourth generation computers also saw the development of GUI's, the mouse and handheld devices. A new generation of computers will emerge with the use wireless communications and wide area networking.
FIFTH GENERATION
( PRESENT & BEYOND )
Fifth generation computing devices, based on artificial intelligence, are still in development, though there are some applications, such as voice recognition, that are being used today. Artificial Intelligence is the branch of computer science concerned with making computers behave like humans. The term was coined in 1956 by John McCarthy at the Massachusetts Institute of Technology.
They will be able to take commands in an audio visual way and carry out instructions. Many of the operations which requires low human intelligence will be performed by these computers Artificial intelligence includes
Expert Systems: programming computers to make decisions in real-life situations (for example, some expert systems help doctors diagnose diseases based on symptoms)
Neural Networks: Systems that simulate intelligence by attempting to reproduce the types of physical connections that occur in animal brains
Robotics: programming computers to see and hear and react to other sensory stimuli
Natural Language: programming computers to understand natural human languages
Games Playing: programming computers to play games such as chess and checkers
Currently, no computers exhibit full artificial intelligence. The greatest advances have occurred in the field of games playing. The best computer chess programs are now capable of beating humans. In May,1997, an IBM super-computer called Deep Blue defeated world chess champion Gary Kasparov in a chess match.
In the area of robotics, computers are now widely used in assembly plants, but they are capable only of very limited tasks. Robots have great difficulty identifying objects based on appearance or feel, and they still move and handle objects clumsily.
Today, the hottest area of artificial intelligence is neural networks, which are proving successful in an umber of disciplines such as voice recognition and natural-language processing. There are several programming languages that are known as AI languages because they are used almost exclusively for AI applications.
You May Also Find These Documents Helpful
-
Computers have become one of our necessities in our daily life which makes it so hard to imagine a time when they did not really exist. And the birth of first computer occurred in 1950. It was the Electrical Numerical Integrator and Calculator or ENIAC. It was made out of 18,000 vacuum tubes which made it to consume about 180,000 watts of electrical power. However, it was only capable to give function such as multiplying numbers rapidly. Due to rapid growth in population, Census Bureau of United States decided to have a machine to tabulate the data, hence,…
- 755 Words
- 4 Pages
Good Essays -
The ENIAC Story: Historical Account of the Creation of the World's First Electronic Digital Computer
The ‘ENIAC Story’ is a historical account of the creation of the world’s first electronic digital…
- 877 Words
- 4 Pages
Powerful Essays -
Today, computers are a part of our lifestyle, but the first computer that was used was developed at the University of Pennsylvania in the year 1946! It had an ENIAC (Electronic Numerical Integrator And Computer) processor. The reprogramming feature that is so extensively used today, was introduced by Alan Turing and John von Neumann with their teams. The von Neumann architecture is the basis of modern computers.…
- 1918 Words
- 8 Pages
Powerful Essays -
At the beginning of 1940, although the European war had been on for several months, extensive…
- 4817 Words
- 20 Pages
Good Essays -
People have been in awe of computers since they were first invented. At first scientist said that computers would only be for government usage only. "Then when the scientists saw the potential computers had, scientist then predicted that by 1990 computers may one day invade the home of just about ever citizen in the world" ("History" Internet), the scientists were slightly wrong, because by 1990 computers were just beginning to catch on. Then a few years later when scientists when to major corporations to get help with a special project, the corporations said no, because computers would just be a fad and they wouldn 't make much money off of it. "By definition Abacus is the first computer (the proper definition of a computer is one who or that which computes) ever invented" (Internet).…
- 1469 Words
- 6 Pages
Better Essays -
The first computers were made by individuals who were striving to create devices that have the ability to automate processes. Individuals such as Alan Turing innovated and transformed the computer industry by creating the blueprint for the algorithms and computations for computers called the Turing Model. The Turing Model paved the way for the first working programmable computer that is automatic was called the Z3 Computer made by Konrad Zuse in 1936. (Wikipedia.org). The next major computer with workable programming that was made was the IBM 701 EDPM Computer. IBM became one of the first computers to be marketed by a major company. From the creation of the IBM 701, the first successful programming language called FORTRAN was created by IBM and John Backus. The next major innovation was the computer mouse made by Douglas Englebart. The 1970’s was one of the most innovative decades for computers with the invention of the floppy disk, random access memory, microprocessors and the first computers available to consumers. All of these components made way for the…
- 1694 Words
- 7 Pages
Good Essays -
* The first electronic computers were developed during the World War II, with the earliest of those being the Colossus.…
- 895 Words
- 4 Pages
Good Essays -
Following shortly after Z3, Britain's Colossus in 1943 and two years later America came up with another system ENIAC (Electronic Numerical Integrator and Computer)…
- 1492 Words
- 6 Pages
Powerful Essays -
Computers have evolved from the fundamental principles of mathematical calculation and language processing. The computing systems of today are an outcome of the efforts and intellect of mathematicians, logicians, linguists and technologists from all over the world. Computer history comprises the evolutionary journey of computing systems, the timeline of operating systems as well as the history behind the development of a wide variety of computer software and hardware. Bellow is the computer history in brief and the evolution of computers.…
- 1982 Words
- 8 Pages
Good Essays -
These early computers had no need for an operating system and were operated directly from the operator's console by a computer programmer, who had immediate knowledge of the computers design.…
- 282 Words
- 2 Pages
Good Essays -
Okay so the ENIAC (Electronic Numerical integrator And Computer) was the first general-use electronic computer. It was able to solve a whole mess of computing problems. Computing problems such as military related stuff, like stuff for the United States Army Ballistic Research Laboratory. The American Military funded the research, since it would really help them; it was also to their gain as well. As with the UNIVAC (Universal Automatic Computer) the ENIAC was invented by the same two, Joseph Mauchly and John P. Eckert. John Mauchly was the chief consultant and John P. Eckert was the chief engineer of the project. With their team, Eckert and Mauchly took around a year or so to design the ENIAC. But it took about a year and a half to actually…
- 547 Words
- 3 Pages
Good Essays -
A report, with a timeline of historic events in the ISD and a description of each of the following approaches; O-O, Structured, Agile and Soft.…
- 5297 Words
- 22 Pages
Powerful Essays -
These first generation computers relied on ‘machine language’ (which is the most basic programming language that can be understood by computers). These computers were limited to solving one problem at a time. Input was based on punched cards and paper tape. Output came out on print-outs. The two notable machines of this era were the UNIVAC and ENIAC machines – the UNIVAC is the first every commercial computer which was purchased in 1951 by a business – the US Census Bureau.…
- 1073 Words
- 5 Pages
Powerful Essays -
With the start of the Second World War, the countries involved in the war machine tried to their potential strategic importance of developing computer misuse.…
- 2785 Words
- 12 Pages
Powerful Essays -
Many people do not know how, or even when computers were first made. Even before the first electronic computers were made, many people believe that computers started with the abacus, a simple counting device. The abacus is believed to have been built in Babylon in the fourth century B.C. The "First Generation" of computers started in the very late 1930's. These computers were grotesquely slow, colossal in size, created much heat, used hundreds of kilowatts of power, and were about as reliable as a used up match (www.pbs.org). As time went on, these computers evolved from using vacuum tubes to transistors. Transistors mark the beginning of the "Second Generation" of computers in 1947. Unfortunately, the 2nd Generation did not advance as much as most people hoped, but eventually, the "Third Generation" was brought about by the invention of integrated circuits in 1958. Integrated circuits replaced transistors, and many computer languages came out in this time. Many more computer companies were born during this time, and eventually this led to personal computers for everyday use. Microprocessors introduced the beginning of the "Fourth Generation" and a time where computers were in almost every house.…
- 1010 Words
- 5 Pages
Good Essays