INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo Poesio LECTURE 1: Intro to the course, History of...

Post on 15-Jan-2016

240 views 0 download

Tags:

Transcript of INTRODUCTION TO ARTIFICIAL INTELLIGENCE Massimo Poesio LECTURE 1: Intro to the course, History of...

INTRODUCTION TO ARTIFICIAL INTELLIGENCE

Massimo Poesio

LECTURE 1: Intro to the course, History of AI

ARTIFICIAL INTELLIGENCE: A DEFINITION

• The branch of Computer Science and/or Cognitive Science whose aim is – ORIGINALLY: to develop machines able to display

intelligent behavior– NOW: to study and design INTELLIGENT AGENTS using

computational methods

• Strong AI: developing machines all but undistinguishable from human beings

• Weak AI: developing SIMULATIONS of human intelligence

A BRIEF HISTORY OF AI

• Forerunners, I: logic and ontologies• Forerunners, II: mechanical machines / robots• The beginning of AI: Turing, Dartmouth,

Games, Search• The role of Knowledge in Human Intelligence• The role of Learning• Modern AI

ARISTOTLE

• Aristotle developed the first theory of knowledge and reasoning – his ideas eventually developed into modern– LOGIC– ONTOLOGIES

ARISTOTLE: SYLLOGISM(Prior Analytics)

BARBARA:A: All animals are mortalA: All men are animals.A: Therefore, all men are mortal.

The first attempt to develop a precise method for reasoning about knowledge: identify VALID REASONING PATTERNS, or SYLLOGISMS

DARII:A: All students in Fil., Logica & Informatica take Intro to AII: Some students in Filosofia take Fil., Logica & Informatica.I: Therefore, some student in Filosofia takes Intro to AI.

FIRST ONTOLOGIES

LOGIC: BEYOND ARISTOTLE

• Ramon Lull (13th Century): first mechanical devices for automatic reasoning (Lull’s disks)

• Leibniz (17th Century): Encoding for syllogisms• Boole (19th Century): Boolean Algebra• Frege (1879): Predicate calculus

BOOLEAN ALGEBRA

FORERUNNERS, 2

THE BEGINNINGS OF AI: 1930-1955

• The development of the first electronic computing machines in the ‘30s also initiated a lot of theoretical thinking about what these machines could do

TURING

• Alan Turing is the father of both theoretical computer science and of AI

• He was the first to imagine machines able of intelligent behavior

• Devised an intelligence test, the TURING TEST: replace the question “Can a machine be endowed with intelligence” with the question:– Can a machine display such human-like behavior to

convince a human observer that it is a human being?

THE TURING TEST

• Alan M. Turing, “Computing Machinery and Intelligence”, 1950 (in Boden, Philosophy of AI)

• The question: what does it mean for a machine to be intelligent?

Turing Test

20

Objections to the Turing Test

• Objections are basically of two forms:– “No computer will ever be able to pass this test”– “Even if a computer passed this test, it wouldn’t

be intelligent”

• Chinese Room argument (Searle, 1980), responses, and counterresponses– Robot reply– Systems reply

21

“Machines can’t think”

• “It’s simply not possible, that’s all”• Arguments from incompleteness theorems– But people aren’t complete, are they?

• Machines can’t be conscious or feel emotions– Reductionism doesn’t really answer the question: why

can’t machines be conscious or feel emotions??• Machines don’t have Human Quality X• Machines just do what we tell them to do– Maybe people just do what their neurons tell them to do…

• Machines are digital; people are analog

DARTMOUTH

• In 1956 a group of researchers including J. McCarthy, M. Minsky, C. Shannon, N. Rochester organized a workshop at Dartmouth to study the possibility of developing machine intelligence

THE BEGINNINGS OF AI (1956-1966)

• Early AI researchers identified intelligence with the kind of behavior that would be considered intelligent when displayed by a human, and tried to develop programs that reproduced that behavior

• Examples:– Games: checkers (Samuel), chess– Theorem proving: Newell Shaw and Simon

HEURISTIC SEARCH

• This early research focused on the development of SEARCH ALGORITHMS (A*) that would allow computers to explore a huge number of alternatives very efficiently

THE SUCCESS OF EARLY AI

In 1997, the chess-playing program DEEP BLUE developed by IBM researchers led by Feng-hsiung Hsu, beat the chess world champion Gary Kasparov over six games

THE LOGIC THEORIST

• Developed by Newell, Shaw and Simon in 1956

• Managed to prove 38 of the first 52 theorems in the second chapter of Russell and Whitehead’s Principia Mathematica

• (More on theorem proving later)

EARLY AI RUNS INTO TROUBLE (1966-1973)

• Soon however researchers realized that these methods could not be applied to all problems requiring intelligence, and that there were a number of ‘simple’ problems that could not be handled with these methods at all– Example of the first: machine translation (the

ALPAC report)– Example of the second: natural language, vision

COMMONSENSE KNOWLEDGE IN LANGUAGE UNDERSTANDING

• Winograd (1974):– The city council refused the women a

permit because they feared violence.– The city council refused the women a

permit because they advocated violence

AI KEY DISCOVERIES, 1

• Performing even apparently simple tasks like understanding natural language requires lots of knowledge and reasoning

THE ‘KNOWLEDGE YEARS’ (1969-1985)

• Development of knowledge representation techniques

• Development of EXPERT SYSTEMS• Development of knowledge-based techniques

for – Natural Language Understanding– Vision

KNOWLEDGE REPRESENTATION METHODS

• Logic is the older formalization of reasoning• It was natural to think of logic as providing the

tools to develop theories of knowledge and its use in natural language comprehension and other tasks

• Great success in developing theorem provers• But AI researchers quickly realized that the

form of logic required was not valid deduction

FROM LOGIC TO AUTOMATED REASONING

• Starting from the ‘50s AI researchers developed techniques for automatic theorem proving

• These techniques are still being developed and have been used to prove non-trivial theorems

RESOLUTION THEOREM PROVING

All Cretans are islanders. All islanders are liars. Therefore all Cretans are liars.

∀X C(X) implies I(X) ∀X I(X) implies L(X) Therefore, X C(X) implies L(X) ∀

¬C(X) ∨ I(X) ¬I(Y) L(Y) ∨

¬C(X) L(Y) ∨

HIGH PERFORMANCE THEOREM PROVING

• There are now a number of very efficient theorem provers that can be used to demonstrate sophisticated mathematical theorems– Otter– Donner & Blitzen

THE FOUR-COLOR PROBLEM

• Conjecture: given a plane divided in regions, it is possible to color the regiones in such a way that two adjacent regions are always of different colors using no more than 4 colors

• This conjecture was demonstrated by an automatic theorem prover in 1997

AI KEY DISCOVERIES, 2

• Neither commonsense nor ‘expert’ reasoning involve only valid inferences from certain premisses:– Commonsense reasoning often involves jumping

to plausible conclusions– Expert reasoning involves making decisions with

uncertainty

COMMONSENSE KNOWLEDGE IN LANGUAGE UNDERSTANDING

• Winograd (1974):– The city council refused the women a

permit because they feared violence.– The city council refused the women a

permit because they advocated violence

DIAGNOSTIC REASONING IN EXPERT SYSTEMS

IF patient has loss of weight AND anorexia AND spleen is palpable AND feverTHEN patient MAY have Hodgkin’s disease

IF patient has loss of weight AND anorexia AND spleen is palpable AND feverTHEN patient MAY have Hodgkin’s disease

BEYOND VALID INFERENCE

• As a result of discovery number 2, AI researchers quickly got involved with developing better theories of commonsense reasoning and uncertain reasoning– Drawing inspiration from psychology (work on

semantic networks, scripts)– Developing new inference formalisms

SOME RESULTS FROM COGNITIVE PSYCHOLOGY: SEMANTIC NETWORKS

• Collins & Quillian, 1969: knowledge appears to be organized around objects and in a taxonomic way– A canary is yellow– A canary has feathers– A canary eats food

• Haviland & Clark 1974, Sanford & Garrod 1979: ‘associated’ knowledge available when concepts are mentioned– I looked around the house.– The lounge was very spacious.

AN EXAMPLE OF SEMANTIC NETWORK

COMMONSENSE REASONING: SCRIPTS (SCHANK & ABELSON, 1977)

John entered the restaurant.

The menu had changed, so he ordered one of the new dishes.

The bill was very steep.

AN EXAMPLE OF SCRIPT

‘KNOWLEDGE-BASED’ SYSTEMS FOR LANGUAGE UNDERSTANDING: SHRDLU

Winograd, 1974

A DIALOGUE WITH SHRDLU

Person: PICK UP A BIG RED BLOCK.Computer: OK.Person: GRASP THE PYRAMID.Computer: I DON'T UNDERSTAND WHICH PYRAMID YOU MEAN.Person (changing their mind): FIND A BLOCK WHICH IS TALLER THAN THE ONE YOU ARE HOLDING AND PUT IT INTO THE BOX.Computer: BY "IT", I ASSUME YOU MEAN THE BLOCK WHICH IS TALLER THAN THE ONE I AM HOLDING.Computer: OK.Person: WHAT DOES THE BOX CONTAIN?Computer: THE BLUE PYRAMID AND THE BLUE BLOCK.Person: WHAT IS THE PYRAMID SUPPORTED BY?Computer: THE BOX.Person: HOW MANY BLOCKS ARE NOT IN THE BOX?Computer: FOUR OF THEM.Person: IS AT LEAST ONE OF THEM NARROWER THAN THE ONE WHICH I TOLD YOU TO PICK UP?Computer: YES, THE RED CUBE.

EXPERT SYSTEMS

• Beginning with the ’70s, AI researchers developed systems that could be used to diagnose – Illnesses (MYCIN, DXPLAIN)– Mechanical problems• DELTA-CATS1, General Electric, Bonissone et al 1984• IDEA, Centro Ricerche Fiat, ~1993

– etc

EXPERT SYSTEMS

A COMMONSENSE ENCYCLOPEDIA: CYC

• A project initiated in 1984 by Doug Lenat. The goal: to encode all of commonsense knowledge

• Changed the underlying formalism several times. – These days: a logic-based representation

• Two versions available:– OpenCyc (http://www.opencyc.org/)

• 50 000 concepts, 300 000 facts• Can be downloaded / on the Web

– ResearchCyc (http://research.cyc.com/)• 300 000 concepts, 3 million facts

KNOWLEDGE IN CYC

"Bill Clinton belongs to the class of US Presidents“

(#$isa #$BillClinton #$UnitedStatesPresident)

“All trees are plants”

(#$genls #$Tree-ThePlant #$Plant)

“Paris is the capital of France".

(#$capitalCity #$France #$Paris)

COMMONSENSE REASONING

• Modelling commonsense inference required the development of entirely new paradigms for inference beyond classical logic

• Non-monotonic reasoning• Probabilistic models

AI RUNS INTO TROUBLE, AGAIN

• The CYC project started in 1984, and by common opinion is nowhere near finished– Hand-coding of commonsense FACTS is unfeasible – (We will get back to this point later when talking

about socially constructed knowledge)• Work on lower-level tasks such as speech

perception revealed the impossibility of hand-coding commonsense RULES and assigning them priorities

SPEECH

KEY AI DISCOVERIES, 3

• A theory of intelligence requires a theory of how commonsense knowledge and cognitive abilities are LEARNED

THE MACHINE LEARNING YEARS (1985-PRESENT)

• The development of methods for learning from evidence started even before Dartmouth

• But machine learning has now taken center stage in AI

CYBERNETICS

• McCulloch, Pitts (1943): first artificial neurons model (based on studies of real neurons)

KNOWLEDGE REPRESENTATION IN THE BRAIN

MODELS OF LEARNING BASED ON THE BRAIN: THE PERCEPTRON

LEARNING TO CLASSIFY OBJECTS

ARTIFICIAL INTELLIGENCE TODAY

• Artificial intelligence as a science:– Computer science vs. Cognitive Science

• Artificial Intelligence as a technology

COGNITION VS ENGINEERING

• Cognitive AI: use computer simulations to study human cognitive abilities– E.g., many neural net models of knowledge

• Engineering AI: develop machines able to carry out intelligent tasks (possibly in a non-human way)– E.g., Deep Blue

STRONG AI / WEAK AI

• Weak AI (Turing): developing machines that are EXTERNALLY UNDISTINGUISHABLE from human beings– But whose internal behavior may be rather

different

• Strong AI: developing machines undistinguishable from human beings IN ALL RESPECTS

Three Schools of Artificial Intelligence

• symbolic functionalism

– intelligence represented in symbols and mutual manipulations

– formal models of reasoning, knowledge based

• connectionism

– inspired by natural processes– emergency of intelligent behaviour

• robotics functionalism

– computational implementation of behaviourism

MAIN SUBFIELDS

• Reasoning and knowledge representation• Planning• Learning• Perception• Motion• Natural language processing• Social intelligence

MAIN SUBFIELDS

• Reasoning and knowledge representation• Planning• Learning• Perception and Vision• Motion• Natural language processing• Social intelligence

AI INDUSTRY: GOOGLE

AI INDUSTRY: ROBOTICS

CONTENTS OF THE COURSE• A very brief recap of knowledge

representation themes: logic, ontologies, etc– (Assumption: you all followed Language Resources

and Ontologies)• Machine Learning– A reminder about statistics– Supervised learning– Unsupervised learning– Applications

• Vision

PRACTICAL INFORMATION

• 42 hours / 6 credits• Evaluation– A PROJECT to be presented at the exam

• Web site: http://clic.cimec.unitn.it/massimo/Teach/AI

READING MATERIAL• Required:

– The course slides, available from the Web Site– Other material downloadable from the Website

• Recommended readings:– History of AI:

• J Copeland, Artificial Intelligence: A Philosophical Perspective, Blackwell’s• M Boden (ed), The Philosophy of AI, Oxford

– Russell & Norvig, Artificial Intelligence: A Modern Approach (2nd ed), Prentice-Hall– Bianchini, Gliozzo, Matteuzzi, Instrumentum vocale: intelligenza artificiale e linguaggio,

Bononia

• Supplementary on specific sub-areas of AI:– KR:

• John F. Sowa, Knowledge Representation, Brooks / Cole

– ML:• Mitchell – Machine Learning – Prentice-Hall• Witten, Frank, Hall – Data Mining – Morgan Kaufmann (3 rd edition)

– NL:• Blackburn & Bos, Representation and Inference for Natural language, CSLI• Jurafsky & Martin, Speech and Language Processing, Pearson

READINGS• This lecture:

– http://en.wikipedia.org/wiki/Timeline_of_artificial_intelligence– John F. Sowa, Knowledge Representation, Brooks / Cole, ch. 1