Monday, February 7, 2011

History of computers - computers and technology


Translate Request has too much data
Parameter name: request
Error in deserializing body of reply message for operation 'Translate'. The maximum string content length quota (8192) has been exceeded while reading XML data. This quota may be increased by changing the MaxStringContentLength property on the XmlDictionaryReaderQuotas object used when creating the XML reader. Line 2, position 8645.

The volume and use of computers in the world are so great, they have become difficult to ignore anymore. Computers appear to us in so many ways that many times, we fail to see them as they actually are. People associated with a computer when they purchased their morning coffee at the vending machine. As they drove themselves to work, the traffic lights that so often hampered us are controlled by computers in an attempt to speed the journey. Accept it or not, the computer has invaded our life.

The origins and roots of computers started out as many other inventions and technologies have in the past. They evolved from a relatively simple idea or plan designed to help perform functions easier and quicker. The first basic type of computers were designed to do just that; compute!. They performed basic math functions such as multiplication and division and displayed the results in a variety of methods. Some computers displayed results in a binary representation of electronic lamps. Binary denotes using only ones and zeros thus, lit lamps represented ones and unlit lamps represented zeros. The irony of this is that people needed to perform another mathematical function to translate binary to decimal to make it readable to the user.

One of the first computers was called ENIAC. It was a huge, monstrous size nearly that of a standard railroad car. It contained electronic tubes, heavy gauge wiring, angle-iron, and knife switches just to name a few of the components. It has become difficult to believe that computers have evolved into suitcase sized micro-computers of the 1990's.

Computers eventually evolved into less archaic looking devices near the end of the 1960's. Their size had been reduced to that of a small automobile and they were processing segments of information at faster rates than older models. Most computers at this time were termed "mainframes" due to the fact that many computers were linked together to perform a given function. The primary user of these types of computers were military agencies and large corporations such as Bell, AT&T, General Electric, and Boeing. Organizations such as these had the funds to afford such technologies. However, operation of these computers required extensive intelligence and manpower resources. The average person could not have fathomed trying to operate and use these million dollar processors.

The United States was attributed the title of pioneering the computer. It was not until the early 1970's that nations such as Japan and the United Kingdom started utilizing technology of their own for the development of the computer. This resulted in newer components and smaller sized computers. The use and operation of computers had developed into a form that people of average intelligence could handle and manipulate without to much ado. When the economies of other nations started to compete with the United States, the computer industry expanded at a great rate. Prices dropped dramatically and computers became more affordable to the average household.

Like the invention of the wheel, the computer is here to stay.The operation and use of computers in our present era of the 1990's has become so easy and simple that perhaps we may have taken too much for granted. Almost everything of use in society requires some form of training or education. Many people say that the predecessor to the computer was the typewriter. The typewriter definitely required training and experience in order to operate it at a usable and efficient level. Children are being taught basic computer skills in the classroom in order to prepare them for the future evolution of the computer age.

The history of computers started out about 2000 years ago, at the birth of the abacus, a wooden rack holding two horizontal wires with beads strung on them. When these beads are moved around, according to programming rules memorized by the user, all regular arithmetic problems can be done. Another important invention around the same time was the Astrolabe, used for navigation.

Blaise Pascal is usually credited for building the first digital computer in 1642. It added numbers entered with dials and was made to help his father, a tax collector. In 1671, Gottfried Wilhelm von Leibniz invented a computer that was built in 1694. It could add, and, after changing some things around, multiply. Leibnitz invented a special stopped gear mechanism for introducing the addend digits, and this is still being used.

The prototypes made by Pascal and Leibnitz were not used in many places, and considered weird until a little more than a century later, when Thomas of Colmar (A.K.A. Charles Xavier Thomas) created the first successful mechanical calculator that could add, subtract, multiply, and divide. A lot of improved desktop calculators by many inventors followed, so that by about 1890, the range of improvements included: Accumulation of partial results, storage and automatic reentry of past results (A memory function), and printing of the results. Each of these required manual installation. These improvements were mainly made for commercial users, and not for the needs of science.

While Thomas of Colmar was developing the desktop calculator, a series of very interesting developments in computers was started in Cambridge, England, by Charles Babbage (of which the computer store "Babbages" is named), a mathematics professor. In 1812, Babbage realized that many long calculations, especially those needed to make mathematical tables, were really a series of predictable actions that were constantly repeated. From this he suspected that it should be possible to do these automatically. He began to design an automatic mechanical calculating machine, which he called a difference engine. By 1822, he had a working model to demonstrate. Financial help from the British Government was attained and Babbage started fabrication of a difference engine in 1823. It was intended to be steam powered and fully automatic, including the printing of the resulting tables, and commanded by a fixed instruction program.

The difference engine, although having limited adaptability and applicability, was really a great advance. Babbage continued to work on it for the next 10 years, but in 1833 he lost interest because he thought he had a better idea; the construction of what would now be called a general purpose, fully program-controlled, automatic mechanical digital computer. Babbage called this idea an Analytical Engine. The ideas of this design showed a lot of foresight, although this couldn't be appreciated until a full century later.

The plans for this engine required an identical decimal computer operating on numbers of 50 decimal digits (or words) and having a storage capacity (memory) of 1,000 such digits. The built-in operations were supposed to include everything that a modern general - purpose computer would need, even the all important Conditional Control Transfer Capability that would allow commands to be executed in any order, not just the order in which they were programmed.

As people can see, it took quite a large amount of intelligence and fortitude to come to the 1990's style and use of computers. People have assumed that computers are a natural development in society and take them for granted. Just as people have learned to drive an automobile, it also takes skill and learning to utilize a computer.

Computers in society have become difficult to understand. Exactly what they consisted of and what actions they performed were highly dependent upon the type of computer. To say a person had a typical computer doesn't necessarily narrow down just what the capabilities of that computer was. Computer styles and types covered so many different functions and actions, that it was difficult to name them all. The original computers of the 1940's were easy to define their purpose when they were first invented. They primarily performed mathematical functions many times faster than any person could have calculated. However, the evolution of the computer had created many styles and types that were greatly dependent on a well defined purpose.

The computers of the 1990's roughly fell into three groups consisting of mainframes, networking units, and personal computers. Mainframe computers were extremely large sized modules and had the capabilities of processing and storing massive amounts of data in the form of numbers and words. Mainframes were the first types of computers developed in the 1940's. Users of these types of computers ranged from banking firms, large corporations and government agencies. They usually were very expensive in cost but designed to last at least five to ten years. They also required well educated and experienced manpower to be operated and maintained. Larry Wulforst, in his book Breakthrough to the Computer Age, describes the old mainframes of the 1940's compared to those of the 1990's by speculating, "...the contrast to the sound of the sputtering motor powering the first flights of the Wright Brothers at Kitty Hawk and the roar of the mighty engines on a Cape Canaveral launching pad". End of part one.

Works Cited

Wulforst, Harry. Breakthrough to the Computer Age. New York: Charles Scribner's Sons, 1982.

Palferman, Jon and Doron Swade. The Dream Machine. London: BBC Books, 1991.

Campbell-Kelly, Martin and William Aspray. Computer, A History of the Information Machine. New York: BasicBooks, 1996.








You may visit http://www.TermPaperAdvisor.com and http://www.TermPapersMadeEasy.com for instant access to over 45,000 plus term papers and essays. You may have all of these quality papers for only $19.95.


Wednesday, February 2, 2011

The development of technology - the history of computers



While computer is now an important part of life of people, it was a time where computer does not exist. Knowing you can understand the history of computers and how much progress has been made, such as complicated and innovative creation of computers really is.

Unlike most devices, the computer is one of the few inventions that has a specific inventor. During development of the computer, many people have their creations of list required, a computer work added. Some of the inventions were different types of computers, and some of you were parts required so that computers can be further developed.

At the beginning

Perhaps is the most important date in the history of computers in 1936. It was this year that the first "computer" was developed. It has synchronized Konrad of Zuse's test and the Z1 computer. This computer is the first, as it was the first system to be fully programmable. There were devices before this, but none had the computing power that distinguishes it from other electronics.

It was not until 1942 that any business saw profit and opportunities in computers. This first company called ABC computers, owned and operated by John Atanasoff and Clifford Berry. Two years later the Harvard Mark I computer was developed promoting the science of computing.

Inventors around the world began to look more in the study of computers and how to improve in the next few years. The next ten years say the introduction of the transistor, which would become a vital part of the inner workings of the computer, the ENIAC-1 computer and many other types of systems. The ENIAC 1 is perhaps one of the most interesting to use required 20,000 vacuum tubes as it. It was a massive machine, and started the revolution to build smaller, faster computer.

At the age of computers was changed by the introduction of international business machines or IBM in the computer industry in 1953 forever. This company has been throughout the history of the computer, a key player in the development of new systems and servers for the public and private use. This tutorial brought the track first real signs of competition in the history of computing, faster and better development of computers helped. Their first contribution was the IBM 701 EDPM computer.

A programming language developed

Programming language a year later created the first successful high level. This was written a programming language that are not in 'Assembly' or binary, the very low level languages. FORTRAN was written so that more people could begin to program computers easily.

In 1955, the Bank of America, coupled with Stanford Research Institute and General Electric, saw the creation of the first computers for use in banks. The MICR or magnetic ink character recognition, coupled with the actual computer, the ERMA was a breakthrough for the banking sector. Only actual banks use 1959 where the couple systems were implemented.

1958 An of the most important breakthroughs in computer history the creation of the integrated circuit occurred. Device, also known as the chip is one of the base requirements for modern computer systems. Many chips are on every motherboard and card within a computer system containing information about what the cards and the cards do. Without these chips, systems, as we know them today may not work.

Gaming, mice, and Internet

Games are now a vital part of the computing experience for many computer users. 1962 saw the creation of the first video game was created by Steve Russel and MIT, which SpaceWar was synchronized.

The mouse one of the basic components of modern computers, 1964 created by Douglass Engelbart. It got out of the device leading its name from the "tail".

One of the most important aspects of computers today was invented in 1969. ARPA net was the original Internet, which provided the basis for the Internet we know today. This development would result in the evolution of knowledge and business across the entire planet.

It was not until 1970 that Intel the scene with the first dynamic RAM chip, entered, which resulted in an explosion of computer science innovation.

On the heels of the RAM chip was the first microprocessor designed by Intel. These two components in addition would become the chip in 1958, core components of modern computers count.

A year later was the floppy disk created wins its name from the flexibility of the storage unit. This was the first step is, that most people to transfer bits data between unconnected computers.

The first NIC was born 1973, so data transfers between connected computers. This is similar to the Internet, but leaves for computer without connecting Internet usage.

Budget of PCs emerge

The next three years were very important for computers. This is when companies began to develop systems for the average consumer. Scelbi mark 8 Altair, IBM 5100, Apple I and II TRS-80 and Commodore Pet computer pioneer in this field were. While expensive began the trend of computers within the common budget these machines.

One of the most major Breathroughs in computer software in 1978 with the release of the VisiCalc spreadsheet program has occurred. All development costs within a period of two weeks, this pays one of the most successful programs of computer history.

1979 was perhaps one of the most important years for the home computer user. This is the year the WordStar, the first word processing program was released to the public for sale. This dramatically changed the utility of computers for the everyday user.

The IBM home computer has helped to revolutionize the consumer market, as was standard affordable for homeowners and consumers in 1981. 1981 she also saw the scene with the MS DOS operating system, type mega giant Microsoft. This operating system changed completely computing forever, it was easy enough for anyone to learn.

The competition begins: Apple vs. Microsoft

Another major change saw computers during 1983. The computer Apple Lisa was the first with a graphical user interface or GUI. Most modern programs provide a graphical interface that allows you easy to use and easy on the eyes. This marked the beginning of dating for most text based out of programs.

On this point in the history of computers many changes and changes occurred from the Apple Microsoft war, is on the development of microcomputers and plenty of computer breakthroughs that have become an accepted part of everyday life. None of this would have been possible without the initial first steps of computer history.








About the author

Rebecca Blain's professional hobby writer enjoys your own computer takes your fish and educating people about how to build, click here: http://www.build-your-own-computer-tips.com