Duy Tan University Lecture Learning Objectives Upon completion of this lecture, students will be able to: Understand basic computing terms and definitions. Appreciate the evolution and history of the computer industry. Outcomes: Receive a broad view of the evolution of the computer and its applications. Demonstrate an understanding on how information technologies improve business in the 21st century..
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University What Is A Computer? An electronic device that has the ability to store, retrieve, and process data, and can be programmed with instructions to perform tasks. An electronic machine capable of performing calculations and other manipulations of various types of data, under the control of a stored set of instructions. A machine that accepts data in digital format and manipulates it for a specific result, based on a sequence of instructions that describes how the data is to be manipulated or processed.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Computer Hardware Computer hardware is the physical part of a computer, which is distinguished from the computer software that executes within the hardware. Most computer hardware is not clearly visible. It is usually embedded in machines or electronic devices. The computer hardware familiar to most people is the Personal Computer (PC).
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Computer Software Computer software is a general term used to describe a collection of computer programs that perform tasks on a computer hardware system. Computer software includes: Application software such as word processors that perform productive tasks for users, System software such as operating systems, that interface with hardware to provide the necessary services for application software, and Middleware which controls and co-ordinates distributed systems.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Computer Network A computer network is an interconnected group of computers. Networks can be classified by the network layer which operates according to standard reference models. For example: The four-layer Internet Protocol model. The seven-layer Open Systems Interconnection (OSI) model.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Computer Industry Computer industry is a term used to describe the whole range of businesses from designing computer hardware, building computer networking infrastructures, manufacturing computer components, developing computer software, and providing of computer technology services.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Software Industry Software industry is a collection of businesses involved in the development, maintenance and publication of computer software. The software industry started in the mid-1970s at the time of the introduction of the personal computer. The industry also includes software services, such as training and consultancy. The largest and most profitable software companies are located in the United States. Since 1980, the software industry is dominated by a few companies such as Microsoft, Oracle, and IBM.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers Who invented the computer? The answer is that many people and inventors contributed to the development of computers. A computer is a complex piece of machinery made up of many parts, each of which can be considered a separate invention.
The ENIAC: First electronic computer
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: BC Chinese inventors created the abacus around 500 BC. The oldest computer probably is the abacus, a mechanical device used to perform mathematical calculations. Similar devices are also found in Mesopotamia around 200 BC.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 18th Century
Year The Computer
John Napier invented a system called "Napiers Bones," made from horn, bone and ivory to multiply by adding 1617 numbers and divide by subtracting. The circular slide rule is invented by William Oughtred to 1622 perform arithmetic. The first known workable mechanical calculating machine is 1623 invented by Wilhelm Schickard Frances Blaise Pascal invented the Pascaline, a machine that 1642 can add, subtract, and carry between digits. Gottfried Wilhelm Leibnitz created a machine that could 1674 add, subtract, multiply and divide automatically.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University 18th Century Computers
Napiers Bones
Schickard Calculating Machine
A Modern Slide Rule
Professor : John Vu Instructor: Luong, Vo Van Duy Tan University History Of Computers: 19th Century Year The Computer Frances Joseph-Marie Jacquard created his fully automated 1804 loom that is programmed by punch cards. Thomas de Colmar created the first reliable, useful and 1820 commercially successful calculating machine. Charles Babbage invented the Difference Engine. Lady Ada Byron, countess of Lovelace became the first computer 1821 analysts, founder of scientific computing. American Thomas Edison discovered the Edison effect, in 1883 which an electric current flows through a vacuum. Herman Hollerith started the Tabulating Machine Company, the company later became the well-known International 1896 Business Machines (IBM). Professor : John Vu Instructor: Luong, Vo Van Duy Tan University 19th Century Computers Hollerith’s Tabulating Machine
Charles Babbage’s Difference Machine
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century Year The Computer ENIAC (Electronic Numerical Integrator and Computer), the first general-purpose electronic digital computer was built. electronic computer. This computer is considered to be the 1943 first The Von Neumann Architecture is introduced in John von 1945 Neumann's report of the EDVAC. John Bardeen, Walter Brattain and William Shockley invented 1947 the first transistor at the Bell Laboratories. The Harvard-MARK III, the first computer to use an internally stored program and indirect addressing, went 1949 into operations again under the direction of Howard Aiken.
1953 IBM introduced the first computer,
Professor : John Vu the IBM 701. Instructor: Luong, Vo Van Duy Tan University The Transistors The IBM 1400 Series was the first large-scale computer to utilize transistors, as well as take advantage of development in magnetic core memory ware. These computers also marked the onset of assembly language, allowing short programming codes to replace the long and tedious binary code. IBM introduced the first transistorized computer in 1959 and quickly became a commercial data-processing powerhouse.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Mainframe Computer The IBM System 360 was a computer family that shared a similar operating system and architecture that were compatible with each other. This allowed users to upgrade their machines following the IBM migration path without having to rewrite all of their software. A side effect of this benefit was that users became increasingly dependent on IBM for their data-processing needs. IBM became the largest computer company in the industry, with sales exceeding many multi-million dollars each year for more than 20 years.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Computer Family
The Paper Tape
The IBM 360 Series
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Early 20th Century Computers
The IBM Mainframe
IBM Computer Punch Card
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Early 20th Century Computer Industry
The acceptance of computers in business created a
computer industry with corporations like IBM, DEC, NCR, NEC, and Wang, who competed against each other. Each corporation developed their own computers with their own system software and applications to run on their systems, then sold to businesses.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Microchip
The next historical event in computer hardware
development was the invention of the integrated circuit used to produce memory chips. The Integrated Circuit, or chip, combined multiple transistors onto a small silicon disc. The silicon chip was the enabling technology behind the development of smaller systems - the Personal Computer. Professor : John Vu Instructor: Luong, Vo Van Duy Tan University Moore’s Law The computer chip was a revolutionary device due to its phenomenal rate of performance improvement over time. Gordon Moore, Intel's co-founder, quantified this rate in what is known today as Moore's Law: “The number of components in a single chip doubles every two years”. Later it was changed into: “The speed of a computer doubles every two years and computing knowledge changes every seven years”.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Mini-Computer In 1974, Popular Electronics magazine published an article describing the "World's First Minicomputer called Altair.” This story excited two students – Bill Gates and Paul Allen. Bill Gates recalled: "The event that started everything for us was when we found an article about a small computer using Intel 8800 microprocessor. I told Paul that this computer is going to get better and better - Now we understand how Moore's Law really worked -- that each generation of a microprocessor chip was basically twice as fast as the previous one and that they got cheaper too. So Paul and I went out and bought our own computer. We realized things were starting to happen. We had a vision where this computer could go, what it could mean and the industry was going to be big, so we started our own company to build software for the computer industry…”
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Apple Computer Two computer enthusiasts, Stephen Wozniak and Steve Jobs joined the Homebrew Computer Club to learn about building computers. Steve Wozniak recalled: "Everybody was interested in computers so I started building a small computer at home and let people type on the keyboard and I would explain what's in it.” Steve Jobs saw that people were interested in his computer and said, “Let's start selling it and let's make this into a company. That is why we started Apple Computer.”
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The First Personal Computer
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Application Software Apple computers were viewed as toys until it developed applications (software) that would make it useful. The first application for Apple computer was VisiCalc, the spreadsheet that could be used for accounting, where if any value in the worksheet changed, all other effected values would be automatically recalculated. The VisiCalc concept was further enhanced and developed into modern spreadsheet programs like Lotus 1-2-3 and Microsoft Excel.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Personal Computer
The first IBM PC
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The World Wide Web The World Wide Web (commonly called the Web) is a system of interlinked documents that can be accessed via the Internet. Users can view these Web pages containing text, images, videos, and other multimedia by using a Web browser to navigate between them.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Dotcom Phenomenon As the internet became popular and the use of home computers grew, many people found that they could do business through the internet and internet-based companies were called “dotcoms”. As investors in stock markets saw their value increase rapidly in this new sector, they pumped more money in and created an exuberant environment in which many businesses dismissed standard practices, focusing on increasing market share at the expense of the profit margin. The Federal Reserve began to increase interest rates to control the runaway economy. The dot-com bubble burst and most companies went bankrupt, marking the beginning of a relatively lengthy recession in the developed world.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University History Of Computers: 20th Century
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Computer Industry - 1 The Industrial revolution that created significant economic changes in the 19th and early 20th century, hit a saturation point in the mid 20th century where profits and economic growth began to slow down. Many businesses turned to information technology as a way to reduce costs, improve productivity and create competitive advantage. This fueled a new economic growth based on knowledge, stemmed largely from the computer industry. The result was so significant that two thirds of U.S. manufacturing productivity grew in efficiency, by applying information technology. The United States dominated the computer industry with 80 percent of the industry's revenues worldwide. Most of these revenues were produced by ten companies, with Microsoft, IBM, and Oracle as leaders.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Computer Industry - 2 Business is about competition and striving for advantage. The business that can sustain competitive advantage wins, those who do not… lose. Although some businesses began to use computers in the early 20th century the cost was very high, and only large companies could afford it. The personal computer allows businesses, large and small, to use computing technology at a fraction of the cost of large mainframe computers. The new computing technology enabled businesses to reduce costs, expedite decision making, improve productivity and provide significant competitive advantage. Thus, the information technology market has grown into a 3 trillion dollar industry worldwide within a very short time.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Information Technology The effect of entrepreneurial ingenuity and creativity coupled with the development of the microchip has fueled waves of innovation that favored significant economic growth. Beginning with the computer industry, the application of information technology in business had a far reaching impact across the entire U.S. economy. Today, information technology has stimulated the overall U.S. computer and electronics industry with total revenue exceeding the combined revenue of all U.S. automobile, steel, and chemical manufacturers that the industrial revolution brought. The value of information technology has changed the entire economic and market value of global business.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University Results American Express reported that it reduced its annual cost by over $1 billion through process automation. AT&T Communication Systems turned a billion dollar loss into a billion dollar profit by using computers to automate its process for manufacturing, services and order fulfillment. Texas Instruments reported that it reduced cycle time of its order fulfillment process for integrated circuits by 56% which achieved record financial results. Progressive Insurance reduced cycle time for claim processing from several weeks to one day, reduced costs, fraud, and litigation and increased revenue by 70%.
Professor : John Vu Instructor: Luong, Vo Van
Duy Tan University The Software Industry The invention of personal computers led to the creation of fourteen thousand new software companies and several hundred new computer hardware manufacturers in the 1970s. In the end, software was the dominant force, with millions of people turning to the personal computer with all their energy and ingenuity, completely changing the way business is conducted.
Professor : John Vu Kondriatiev Waves ofLuong,
Instructor: Innovation Vo Van Duy Tan University Questions & Answers