Date post: | 15-Dec-2015 |
Category: |
Documents |
Upload: | arjun-speck |
View: | 217 times |
Download: | 2 times |
History of Computing Chapter 1
Introduction to Social and Ethical Computing
Historical Development of Computing
Development of the Internet
Development of the World Wide Web
The Emergence of Social and Ethical Problems in Computing
The Case for Computer Ethics Education
Historical Development Before 1900AD
Man sought to improved life through the invention of gargets. First utility tools recorded dealt with numbers First recorded on bones – 20,000 to 30,000 B.C. First place-value number system in place – 1800 B.C. Abacus – Mother of Computers – 1000 B.C. and 500 B.C. Zero and Negative Numbers – 300 B.C. and 500 A.D.
1500AD and 1900AD lot of activities in the development of computing devices Driven by commerce 1500 Leonardo da Vinci invented mechanical calculator 1621 invention of the slide rule 1625 Wilhelm Schichard’s mechanical calculator in 1640 Blaise Pascal’s Arithmetic Machine
Major breakthrough in speed up 1800 AD with the invention of the punched card by Joseph-Marie
Jacquard Revolutionized computing Quickly spread in other fields Speed up computation and storage of information
Historical Development Before 1900AD
1830 AD exciting period
1830 - Charles Babbage’s Analytical Engine
George and Edward Schutz’s Difference Engine
Within a decade - major milestone
George Boole’s invention of Boolean Algebra
Opened fields of mathematics, engineering, & computing
Lead to the new frontiers in logic
Historical Development Before 1900AD Mid 1850 through the turn of the century
1857 - Sir Charles Wheatstone’s invention Paper tape to store information
Created new excitement in the computing community of the time.
Huge amounts of data could be entered & stored
1869AD - Logic Machine by William Stanley Jovons ~1874 - first Keyboard by Sholes 1881 - Rectangular Logic Diagrams by Allan Marquand
Historical Development Before 1900AD
Mid 1850 through the turn of the century
1886, Charles Pierce - first linked Boolean Algebra to circuits based on switches
Major break through in mathematics, engineering and computing science
1890 - John Venn invented the Venn diagrams
Used extensively in switching algebras in both hardware and software development
1890 - Herman Hollerith invented the Tabulating Machine
Utilized Jacquard’s punched card to read the presence or absence of holes.
The data read was to be collated using an automatic electrical tabulating machine
Large number of clock-like counters Summed up and accumulated the results in a number
of selected categories.
After 1900 AD Computing in infancy Century began with a major milestone
Vacuum tube by John Ambrose Fleming. Played a major role in computing for the next half
century. All digital computer in the first half century ran on
vacuum tubes. 1906 - triode by Lee de Forest in 1906. 1926 - first semiconductor transistor
Not used for several years Came to dominate the computing industry in late
years
1937 - Turing Machine by Alan Turing
Invention of an abstract computer Some problems do not lend themselves to algorithmic
representations, not computable 1942 - COLOSSUS, one of the first working programmable
digital computers
After 1900 AD
1942 – Turing designed COLOSSUS
One of the first working programmable digital computers
1939 – Vincent Atanasoff – 1st digital computer model
utilized capacitors to store electronic charge to represent Boolean numbers
0 and 1 used by the machine in calculations
Input and output data was on punched cards
Some doubt it ever worked
After 1900 AD
Howard Aiken – developed Harvard Mark I
1st large scale automatic digital computer.
also known as IBM automatic sequencer calculator- ASCC
1943, Alan Turing – COLOSSUS
Considered 1st programmable compute
designed to break the German ENIGMA code
used about 1800 vacuum tubes
execute a variety of routines.
After 1900
John William Mauchly & J. Presper Eckert Jr - ENICAC
Vacuum tube-based general purpose
10 feet high
Weighed 30 tons
Occupied 1000 square feet
70,000 resistors
10,000 capacitors
6000 switches
18,000 vacuum tube
No internal memory
Hard-wired
Consistently programmed by switches and diodes
After 1900 1944-1952 John William Mauchly & J. Presper
Eckert Jr – EDVAC
Electronic discrete variable automatic computer
1st truly general purpose digital computer
Stored program instruction concept
completed in 1956
4,000 vacuum tubes and 10,000 crystal diodes
1948 - UNIVAC I
1st commercially available computer.
After 1900
Many companies became involved International Business Machines (IBM), Honeywell,
and Control Data Corporation (CDC) in the USA, and International Computers Limited, (ICL) in UK
Built mainframe Hugh – took entire rooms Expensive – use limited to big corporations
Mid to late sixties Developed less expensive but smaller computer Minicomputer Timesharing concept Let to idea of networking
After 1900
1971 and 1976 - first microprocessor Built with integrated circuit with many
transistors on a single board Vacuum tubes and diodes no longer used Ted Hoff The 4004 4-bit data path
1972 – Intel - 8008 8-bit microprocessor based on the 4004 fIrst microprocessor to use a compiler Specific application microprocessors
Microprocessor
1974 -truly general purpose microprocessor 8080 -8-bit device - 4,500
transistors & astonishing 200,000 operations per second
After 1974, development exploded
Computer Software and Personal Computer (PC)
Until mid 1970s Development led by hardware Computers were designed and software was
designed to fit the hardware.
Personal computing industry began 1976 - Apple I and Apple II microcomputer
were unveiled 1981 - IBM joined the PC wars
3 Major Players IBM Gary Kildall - Developed the first PC
operating system Bill Gates - Developed the Disk Operating
System (DOS).
The Development of the Internet
Internet based on 4 technologies Telegraph Telephone Radio Computers
Originated from the early work of J.C.R. Licklider Conceptualized a global interconnected set of computers Concept for communication between network nodes
Packets instead of circuits
Enabled computers to talk to each other.
1961 - Kleinrock Published first work on packet switching theory
The Development of the Internet
Two additional important projects Donald Davies and Roger Scantleberg
Coining the term packet Connected computer in Boston with
one in Los Angels Low speed dial-up telephone line created the first working Wide Area
Network 1967 Roberts - publishing the first
plan for ARPNET 1968 - team, lead by Frank Heart and
included Bob Kahn, developed IMP
ARPNET
Began as tool for defense contractors
Universities added
Government joined
Other countries joined
ARPANET ceased to exist in 1989
Internet was an entity to itself
Development World Wide Web
Beginning concepts - Tim Berners-Lee’s 1989 Proposal called HyperText and CERN Enable collaboration between physicists &
researchers in the high energy physics research
Three new technologies were incorporated. HyperText Markup Language (HTML)
hypertext concepts- to be used to write web documents
HyperText Transfer Protocol (HTTP) a protocol Used to transmit web pages between hosts Web browser client software program to
receive and interpret data and display results.
Development World Wide Web
Proposal included a very important concept for the user interface Consistent across all types of computer
platforms Enable users to access information from
any computer. Line-mode interface was developed &
named at CERN in late 1989
Development World Wide Web
Growth
Central computer at CERN with few web pages in 1991
50 world wide by 1992
720,000 by 1999
Over 24 million by 2001
1993 - graphic user interface browser
Mosaic
Popularized and fueled growth of internet
Emergence of the Social & Ethical Problems in Computing
The Emergence of Computer Crimes Perhaps started with the invention of the computer
virus The term virus is derived from a Latin word virus
which means poison Computer virus
Self-propagating computer program
Designed to alter or destroy a computer system resource
Spreads in the new environment Attacks major system Weakens the capacity of resources to perform
1972 – virus used to describe piece of unwanted computer code
Growth of Computer Vulnerabilities
The Case for Computer Ethics Education What is Computer Ethics
James H. Moore First coined the phrase "computer ethics“
Computer ethics is the analysis of the nature and social impact of computer technology and the corresponding formulation and justification of policies for the ethical use of such technology .
Definition focuses on the human actions Study, an analysis of the values of human actions
influenced by computer technology. Computer influence on human actions is widespread
throughout the decision making process preceding the action
Education we study the factors that influence the decision making process
Why You Should Study Computer Ethics
Central task of computer ethics determine what should be done Especially whenever there is a policy vacuum
Vacuums caused by the ‘confusion’ between the known policies and what is presented
Professionals unprepared to deal effectively with the ethical issues
Can stop the vacuums Can prepare the professionals
Schools of Thought
Study computer ethics as remedial moral education
Computer ethics education not as a moral education but as a field worthy of study in its own right
Justification for First Thought
We should study computer ethics because doing so will make us behave like responsible professionals.
We should study computer ethics because doing so will teach us how to avoid computer abuse and catastrophes.
Material taken from Walter Manner in “Is Computer Ethics Unique?”
Justification for Second Thought
We should study computer ethics because the advance of computing technology will continue to create temporary policy vacuums.
We should study computer ethics because the use of computing permanently transforms certain ethical issues to the degree that their alterations require independent study.
We should study computer ethics because the use of computing technology creates, and will continue to create, novel ethical issues that require special study.
We should study computer ethics because the set of novel and transformed issues is large enough and coherent enough to define a new field
Material taken from Walter Manner in “Is Computer Ethics Unique?”