kidzsearch.com > wiki
A computer is a machine that is able to take information (input), do some work on or make changes to the information, to make new information (output). Computers have existed for much of human history. Examples of early computers are the astrolabe and the abacus.
Modern computers are very different from early computers. They are now very powerful machines that are able to do billions of calculations every second. Most people have used a personal computer in their home or at work. Computers are useful for many different jobs where automatic tasks are useful. Some examples are controlling traffic lights, vehicle computers, security systems, Washing machines and Digital Televisions.
A person (called a user) can control a computer by telling it to do things. Some ways of controlling a computer are with a keyboard, mouse, buttons, touch screen. Some very new computers can also be controlled with voice commands or hand gestures.
Computers can be designed to do anything with information. Computers are used to control factories, which in the past were controlled by humans. They are also in homes, where they are used for things such as listening to music, reading the news, and writing.
Modern computers are electronic machines. A computer is only useful if it has both hardware and software. Hardware is the physical parts the computer is made of - for example keyboard, mouse, screen, tower, and the circuits inside it. Software is the computer programs (mathematical instructions). The software uses the hardware, when the user gives it instructions and changing it in to useful output.
Many modern computers do billions of calculations each second. They do mathematical arithmetic very quickly but computers do not really "think". They only follow the instructions in their software programs.
Computer programs are designed or written by computer programmers. A few programmers write programs in the computer's own language called machine code. Machine Code has only a few instructions and is based on logic and Mathematical arithmetic. Using machine code is difficult for most human programmers.
Today, programs are commonly written using a programming language like C++, Java, and Objective-C. These programming languages are more like the language you talk and write with every day. A program called a compiler translates the user's instructions into binary code (machine code) and turns it into useful information that a user can easily understand, and then the computer will understand the instructions and do what is needed.
- 1 History of computers
- 2 Analog and Digital Computers
- 3 Kinds of computers
- 4 Common uses of home computers
- 5 Working methods
- 6 The Internet
- 7 Computers and waste
- 8 Main hardware
- 9 References
History of computers
A computer is an electronic machine which helps in solving problems quickly and easily.It solves problems according to instructions given to it by the computer user.
The computer was invented by Charles Babbage. It is a digital machine used in all fields.
Humans have a problem with maths. To show this, try doing 584 x 3,220 in your head. It is hard to remember all the steps! People made tools to help them remember where they were in a maths problem. The other problem people have is that they have to do the same problem over and over and over again. A cashier used to make change every day in her head or with a piece of paper. That took a lot of time and people made mistakes. So people made machines that did those same things over and over. This part of computer history is called the "history of automated calculation," which is a fancy phrase for "the history of machines that make it easy for me to do this same maths problem over and over without making mistakes."
Some people did not want a machine that would do the same thing over and over again. For example, a music box is a machine that plays the same music over and over again. Some people wanted to be able to tell their machine to do different things. For example, they wanted to tell the music box to play different music every time. They wanted to be able to program the music box- to order the music box to play different music. This part of computer history is called the "history of programmable machines" which is a fancy phrase for "The history of machines that I can order to do different things if I know how to speak their language."
One of the first examples of this was built by Hero of Alexandria (c. 10–70 AD). He built a mechanical theater which performed a play lasting 10 minutes and was operated by a complex system of ropes and drums. These ropes and drums were the language of the machine- they told what the machine did and when. Some people argue that this is the first programmable machine.
Most historians agree that the "castle clock", an astronomical clock invented by Al-Jazari in 1206, is the first known programmable analog computer. It showed the zodiac, the solar and lunar orbits, a crescent moon-shaped pointer travelling across a gateway that made some doors to open every hour, and five robotic musicians who play music when levers hit them. The length of day and night could be changed (AKA re-programmed) every day in order to account for the changing lengths of day and night throughout the year. Ada Lovelace is considered to be the first programmer.
The Computing Era
At the end of the Middle Ages, people in Europe thought math and engineering were more important. In 1623, Wilhelm Schickard made a mechanical calculator. Other Europeans made more calculators after him. They were not modern computers because they could only add, subtract, and multiply- you could not change what they did to make them do something like play Tetris. Because of this, we say they were not programmable.
In 1801, Joseph Marie Jacquard used punched paper cards to tell his textile loom what kind of pattern to weave. He could use punch cards to tell the loom what to do, and he could change the punch cards, which means he could program the loom to weave the pattern he wanted. This means the loom was programmable.
Modern computers were made when someone (Charles Babbage) had an idea. He wanted to make a machine that could do all the boring parts of math, (like the automated calculators) and could be told to do them different ways (like the programmable machines.) Charles Babbage was the first to make a design of a fully programmable mechanical computer. He called it "The Analytical Engine". Because Babbage did not have enough money and always changed his design when he had a better idea, he never built his Analytical Engine.
As time went on, computers got more and more popular. And that stands out at the beginning. This is because people get bored easily doing the same thing over and over. Imagine spending your life writing things down on index cards, storing them, and then having to go find them again. The U.S. Census Bureau in 1890 had hundreds of people doing just that. People got very bored and very frustrated, and would say, "There HAS to be an easier way to do this." Then a clever person worked out how to make machines do a lot of the work. Herman Hollerith figured out how to make a machine that would automatically add up information that the Census bureau collected. The Computing Tabulating Recording Corporation(which later became IBM) made his machines, and everyone was happy. At least, they were happy until their machines broke down, got jammed, and had to be repaired. This is when the Computing Tabulating Recording Corporation invented tech support.
Because of machines like this, new ways of talking to these machines were invented, and new types of machines were invented, and eventually the computer as we know it was born.
Analog and Digital Computers
In the first half of the 20th century, scientists started using computers, mostly because scientists had a lot of math to figure out and wanted to spend more of their time thinking about the secrets of the universe instead of spending hours adding numbers together. For example, if they had to launch a rocket ship, they needed to do a lot of math to make sure the rocket did not blow up.
Scientists figured out how to make and use digital computers in the 1930s and 1940s. Scientists made a lot of digital computers, and as they did, they figured out how to ask them the right sorts of questions to get the most out of them. Here are a few of the computers they built:
|Name||First operational||Numeral system||Computing mechanism||Programming||Turing complete|
|Zuse Z3 (Germany)||May 1941||Binary||Electro-mechanical||Program-controlled by punched film stock||Yes (1998)|
|Atanasoff–Berry Computer (US)||mid-1941||Binary||Electronic||Not programmable—single purpose||No|
|Colossus (UK)||January 1944||Binary||Electronic||Program-controlled by patch cables and switches||No|
|Harvard Mark I – IBM ASCC (US)||1944||Decimal||Electro-mechanical||Program-controlled by 24-channel punched paper tape (but no conditional branch)||No|
|ENIAC (US)||November 1945||Decimal||Electronic||Program-controlled by patch cables and switches||Yes|
|Manchester Small-Scale Experimental Machine (UK)||June 1948||Binary||Electronic||Stored-program in Williams cathode ray tube memory||Yes|
|Modified ENIAC (US)||September 1948||Decimal||Electronic||Program-controlled by patch cables and switches plus a primitive read-only stored programming mechanism using the Function Tables as program ROM||Yes|
|EDSAC (UK)||May 1949||Binary||Electronic||Stored-program in mercury delay line memory||Yes|
|Manchester Mark 1 (UK)||October 1949||Binary||Electronic||Stored-program in Williams cathode ray tube memory and magnetic drum memory||Yes|
|CSIRAC (Australia)||November 1949||Binary||Electronic||Stored-program in mercury delay line memory||Yes|
- Konrad Zuse's electromechanical "Z machines". The Z3 (1941) was the first working machine that used binary arithmetic. Binary arithmetic means using "Yes" and "No." to add numbers together. You could also program it. In 1998 the Z3 was proved to be Turing complete. Turing complete means that it is possible to tell this particular computer anything that it is mathematically possible to tell a computer. It is the world's first modern computer.
- The non-programmable Atanasoff–Berry Computer (1941) which used vacuum tubes to store "yes" and "no" answers, and regenerative capacitor memory.
- The Harvard Mark I (1944), A big computer that you could kind of program.
- The U.S. Army's Ballistics Research Laboratory ENIAC (1946), which could add numbers the way people do (using the numbers 0 through 9) and is sometimes called the first general purpose electronic computer (since Konrad Zuse's Z3 of 1941 used electromagnets instead of electronics). At first, however, the only way you could reprogram ENIAC was by rewiring it.
Several developers of ENIAC saw its problems. They invented a way to for a computer to remember what they had told it, and a way to change what it remembered. This is known as "stored program architecture" or von Neumann architecture. John von Neumann talked about this design in the paper First Draft of a Report on the EDVAC, distributed in 1945. A number of projects to develop computers based on the stored-program architecture started around this time. The first of these was completed in Great Britain. The first to be demonstrated working was the Manchester Small-Scale Experimental Machine (SSEM or "Baby"), while the EDSAC, completed a year after SSEM, was the first really useful computer that used the stored program design. Shortly afterwards, the machine originally described by von Neumann's paper—EDVAC—was completed but was not ready for two years.
Nearly all modern computers use the stored-program architecture in some form. It has become the main concept which defines a modern computer. Most of the technologies used to build computers have changed since the 1940s, but many current computers still use the von-Neumann architecture.
In the 1950's computers were built out of mostly vacuum tubes. Transistors replaced vacuum tubes in the 1960's because they were smaller and cheaper. They also need less power and do not break down as much as vacuum tubes. In the 1970s, technologies were based on integrated circuits. Microprocessors, such as the Intel 4004 made computers smaller and cheaper. They also made computers faster and more reliable. By the 1980s, computers became small and cheap enough to replace mechanical controls in things like washing machines. The 1980s also saw home computers and personal computer. With the evolution of the Internet, personal computers are becoming as common as the television and the telephone in the household.
In 2005 Nokia started to call some of its mobile phones (the N-series) "multimedia computers" and after the launch of the Apple iPhone in 2007, many are now starting to add the smartphone category among "real" computers. In 2008, if the category of smartphones are included in the numbers of computers in the world, the biggest computer maker by units sold, is no longer Hewlett-Packard, but rather Nokia.[source?]
Kinds of computers
There are six types of computers: personal computers, mainframes, laptop computers, mini computers, super computers and embedded computers.
A "desktop computer" is a small machine that has a screen (which is not part of the computer). Most people keep them on top of a desk, which is why they are called "desktop computers." "Laptop computers" are computers small enough to fit on your lap. This makes them easy to carry around. Both laptops and desktops are called personal computers, because one person at a time uses them for things like playing music, surfing the web, or playing video games.
There are bigger computers that many people at a time can use. These are called "Mainframes," and these computers do all the things that make things like the internet work. You can think of a personal computer like this: the personal computer is like your skin: you can see it, other people can see it, and through your skin you feel wind, water, air, and the rest of the world. A mainframe is more like your internal organs: you (hopefully) never see them, and you barely even think about them, but if they suddenly went missing, you would have some very big problems.
There is another type of computer, called an embedded computer. An embedded computer is a computer that does one thing and one thing only, and usually does it very well. For example, an alarm clock is a embedded computer: it tells the time. Unlike your personal computer, you cannot use your clock to play Tetris. Because of this, we say that embedded computers cannot be programmed, because you cannot install programs like Tetris on your clock. Some mobile phones, automatic teller machines, microwave ovens, CD players and cars are examples of embedded computers.
All-in-one computers are desktop computers that have the CPU and speaker in the same case as the monitor. Apple has made several popular examples of all-in-one computers, such as the original Macintosh of the mid-1980s and the iMac of the late 1990s and 2000s.
Common uses of home computers
- Playing computer games
- Solving mathematics
- Looking for things on the Internet
- Watching TV and movies
- Listening to music
- Communicating with other people
- Sending e-mail
- Taking Pictures (photography)
Computers store data and the instructions telling them what to do with the data as numbers, because computers can do things with numbers very quickly. These data are stored as binary symbols (1s and 0s). A 1 or a 0 symbol stored by a computer is called a bit, which comes from the words binary digit. Computers can use many bits together to represent instructions and the data that these instructions use. A list of instructions is called a program and is stored on the computer's hard disk. Computers work through the program by using a central processing unit, and they use fast memory called RAM as a space to store the instructions and data while they are doing this. When the computer wants to store the results of the program for later, it uses the hard disk because things stored on a hard disk can still be remembered after the computer is turned off.
An operating system tells the computer how to understand what jobs it has to do, how to do these jobs, and how to tell people the results. It tells the electronics inside the computer, or "hardware", how to work to get the results it needs. This lets most computers have the same operating system, or list of orders to tell it how to talk to the user, while each computer can have its own computer programs or list of jobs to do what its user needs. Having different programs and operating systems makes it easy to learn how to use computers for new things. When a user needs to use a computer for something different, the user can learn how to use a new program.
One of the most important jobs that computers do for people is helping with communication. Communication is how people share information. Computers have helped people move forward in science, medicine, business, and learning, because they let experts from anywhere in the world work with each other and share information. They also let other people communicate with each other, do their jobs almost anywhere, learn about almost anything, or share their opinions with each other. The Internet is the thing that lets people communicate between their computers.
Computers and waste
A computer is now almost always an electronic device. It usually contains materials that will become toxic waste when disposed of. When a new computer is bought in some places, laws require that the cost of its waste management must also be paid for. This is called product stewardship.
Computers can become obsolete quickly, depending on what programs the user runs. Very often, they are thrown away within two or three years, because newer programs require a more powerful computer. This makes the problem worse, so computer recycling happens a lot. Many projects try to send working computers to developing nations so they can be re-used and will not become waste as quickly, as most people do not need to run new programs. Some computer parts, such as hard drives, can break easily. When these parts end up in the landfill, they can put poisonous chemicals like lead into the ground water. Hard drives can also contain secret information like credit card numbers. If the hard drive is not erased before being thrown away, an identity thief can get the information off of the hard drive, even if the drive doesn't work, and use it to steal money from the previous owner's bank account.
Computers come in different forms, but most of them have a common design.
- All computers have a CPU.
- All computers have some kind of data bus which lets them get inputs or output things to the environment.
- All computers have some form of memory. These are usually chips (integrated circuits) which can hold information.
- Many computers have some kind of sensors, which lets them get input from their environment.
- Many computers have some kind of display device, which lets them show output. They may also have other peripheral devices connected.
A computer has several main parts. When comparing a computer to a human body, the CPU is like a brain. It does most of the 'thinking' and tells the rest of the computer how to work. The CPU is on the Motherboard, which is like the skeleton. It provides the basis for where the other parts go, and carries the nerves that connect them to each other and the CPU. The motherboard is connected to a power supply, which provides electricity to the entire computer. The various drives (CD drive, floppy drive, and on many newer computers, USB flash drive) act like eyes, ears, and fingers, and allow the computer to read different types of storage, in the same way that a human can read different types of books. The hard drive is like a human's memory, and keeps track of all the data stored on the computer. Most computers have a sound card or another method of making sound, which is like vocal cords, or a voice box. Connected to the sound card are speakers, which are like a mouth, and are where the sound comes out. Computers might also have a graphics card, which helps the computer to create visual effects, such as 3D environments, or more realistic colors, and more powerful graphics cards can make more realistic or more advanced images, in the same way a well trained artist can.
- "Heron of Alexandria". http://www.mlahanas.de/Greeks/HeronAlexandria2.htm. Retrieved 2008-01-15.
- Ancient Discoveries, Episode 11: Ancient Robots, History Channel, http://www.youtube.com/watch?v=rxjbaQl0ad8, retrieved 2008-09-06
- Howard R. Turner (1997), Science in Medieval Islam: An Illustrated Introduction, p. 184, University of Texas Press, ISBN 0-292-78149-0
- Donald Routledge Hill, "Mechanical Engineering in the Medieval Near East", Scientific American, May 1991, pp. 64-9 (cf. Donald Routledge Hill, Mechanical Engineering)
- Fuegi & Francis 2003, pp. 16–26.
- Phillips, Ana Lena (November–December 2011). "Crowdsourcing gender equity: Ada lovelace day, and its companion website, aims to raise the profile of women in science and technology". American Scientist 99 (6): 463.
- "Ada Lovelace honoured by Google doodle", The Guardian, 2012 Dec 10, http://www.guardian.co.uk/technology/2012/dec/10/ada-lovelace-honoured-google-doodle, retrieved 10 December 2012.
- Don't confuse the Analytical Engine with Babbage's difference engine which was a non-programmable mechanical calculator.
- a Kempf, Kar (1961). "Historical Monograph: Electronic Computers Within the Ordnance Corps". Aberdeen Proving Ground (United States Army).
- a Phillips, Tony (2000). "The Antikythera Mechanism I". American Mathematical Society. http://www.math.sunysb.edu/~tony/whatsnew/column/antikytheraI-0400/kyth1.html. Retrieved 2006-04-05.
- a Shannon, Claude Elwood (1940). "A symbolic analysis of relay and switching circuits". Massachusetts Institute of Technology.
- a Digital Equipment Corporation (1972) (PDF). PDP-11/40 Processor Handbook. Maynard, MA: Digital Equipment Corporation. http://bitsavers.vt100.net/dec/www.computer.museum.uq.edu.au_mirror/D-09-30_PDP11-40_Processor_Handbook.pdf.
- a Verma, G.; Mielke, N. (1988). "Reliability performance of ETOX based flash memories". IEEE International Reliability Physics Symposium.
- a Meuer, Hans; Strohmaier, Erich; Simon, Horst; Dongarra, Jack (2006-11-13). "Architectures Share Over Time". TOP500. http://www.top500.org/lists/2006/11/overtime/Architectures. Retrieved 2006-11-27.
- Stokes, Jon (2007). Inside the Machine: An Illustrated Introduction to Microprocessors and Computer Architecture. San Francisco: No Starch Press.