1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI...
-
Upload
neal-cleavenger -
Category
Documents
-
view
236 -
download
1
Transcript of 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI...
![Page 1: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/1.jpg)
1
Knowledge Representation 6 6.0 Issues in Knowledge
Representation
6.1 A Brief History of AIRepresentationalSystems
6.2 Conceptual Graphs: ANetwork Language
6.3 Alternatives to ExplicitRepresentation
6.4 Agent Based andDistributed ProblemSolving
6.5 Epilogue and References
6.6 Exercises
![Page 2: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/2.jpg)
2
Chapter Objectives
• Learn different formalisms for Knowledge Representation (KR)
• Learn about representing concepts in a canonical form
• Compare KR formalisms to predicate calculus
• The agent model: Transforms percepts and results of its own actions to an internal representation
![Page 3: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/3.jpg)
3
“Shortcomings” of logic
• Emphasis on truth-preserving operations rather than the nature of human reasoning (or natural language understanding)
• if-then relationships do not always reflect how humans would see it:
X (cardinal (X) red(X))
X( red (X) cardinal(X))
• Associations between concepts is not always clearsnow: cold, white, snowman, slippery, ice,
drift, blizzard
•Note however, that the issue here is clarity or ease of understanding rather than expressiveness.
![Page 4: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/4.jpg)
4
Semantic network developed by Collins and Quillian (Harmon and King 1985)
![Page 5: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/5.jpg)
5
Network representation of properties of snow and ice
![Page 6: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/6.jpg)
6
Three planes representing three definitions of the word “plant” (Quillian 1967)
![Page 7: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/7.jpg)
7
Intersection path between “cry” and “comfort” (Quillian 1967)
![Page 8: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/8.jpg)
8
“Case” oriented representation schemes
• Focus on the case structure of English verbs
• Case relationships include:agent locationobject timeinstrument
• Two approachescase frames: A sentence is represented as
a verb node, with various case links to nodes representing other participants in the action
conceptual dependency theory: The situation is classified as one of the standard action types. Actions have conceptual cases (e.g., actor, object).
![Page 9: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/9.jpg)
9
Case frame representation of “Sarah fixed the chair with glue.”
![Page 10: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/10.jpg)
10
Conceptual dependency theory
Four primitive conceptualizations:
• ACTs actions
• PPs objects (picture producers)
• AAs modifiers of actions (action aiders)
• PAs modifiers of objects (picture aiders)
![Page 11: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/11.jpg)
11
Conceptual dependency theory (cont’d)
Primitive acts:
• ATRANS transfer a relationship (give)
• PTRANS transfer of physical location of an object (go)
• PROPEL apply physical force to an object (push)
• MOVE move body part by owner (kick)
• GRASP grab an object by an actor (grasp)
• INGEST ingest an object by an animal (eat)
• EXPEL expel from an animal’s body (cry)
• MTRANS transfer mental information (tell)
• MBUILD mentally make new information (decide)
• CONC conceptualize or think about an idea (think)
• SPEAK produce sound (say)
• ATTEND focus sense organ (listen)
![Page 12: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/12.jpg)
12
“John hit the cat.”
ACT: [apply a force] or PROPELACTOR: johnOBJECT: cat
john PROPEL cato
![Page 13: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/13.jpg)
13
Basic conceptual dependencies
![Page 14: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/14.jpg)
14
Examples with the basic conceptual dependencies
![Page 15: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/15.jpg)
15
Examples with the basic conceptual dependencies (cont’d)
![Page 16: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/16.jpg)
16
“John ate the egg”
![Page 17: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/17.jpg)
17
“John prevented Mary from giving a book to Bill”
![Page 18: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/18.jpg)
18
Representing Picture Aiders (PAs)
thing <> state-type (state-value)
• “The ball is red” ball <> color (red)
• “John is 6 feet tall” john <> height (6 feet)
• “John is tall” john <> height (>average)
• “John is taller than Jane”john <> height (X)jane <> height (Y)X > Y
![Page 19: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/19.jpg)
19
More PA examples
• “John is angry.” john <> anger(5)
• “John is furious.” john <> anger(7)
• “John is irritated.” john <> anger (2)
• “John is ill.” john <> health (-3)
• “John is dead.” john <> health (-10)
![Page 20: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/20.jpg)
20
Variations on the story of the poor cat
“John applied a force to the cat by moving some object to come in contact with the cat”
John <> *PROPEL* cat
John <> *PTRANS* [ ]
o
loc(cat)
i o
![Page 21: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/21.jpg)
21
Variations on the cat story (cont’d)
“John kicked the cat.”
John <> *PROPEL* cat
John <> *PTRANS* foot
kick = hit with one’s foot
o
loc(cat)
i
o
![Page 22: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/22.jpg)
22
Variations on the cat story (cont’d)
“John hit the cat.”
John <> *PROPEL* cat
cat <
Hitting was detrimental to the cat’s health.
o
Health(-2)
<
![Page 23: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/23.jpg)
23
Causals
“John hurt Jane.”
John <> DO Jane
Jane <
John did something to cause Jane to become hurt.
o
Pain( > X)
<
Pain (X)
![Page 24: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/24.jpg)
24
Causals (cont’d)
“John hurt Jane by hitting her.”
John <> PROPEL Jane
Jane <
John hit Jane to cause Jane to become hurt.
o
Pain( > X)
<
Pain (X)
![Page 25: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/25.jpg)
25
How about?
“John killed Jane.”
“John frightened Jane.”
“John likes ice cream.”
![Page 26: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/26.jpg)
26
“John killed Jane.”
John <> *DO*
Jane <Health(-10)
<
Health(> -10)
![Page 27: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/27.jpg)
27
“John frightened Jane.”
John <> *DO*
Jane <Fear (> X)
<
Fear (X)
![Page 28: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/28.jpg)
28
“John likes ice cream.”
John <> *INGEST* IceCream
John <
o
Joy ( > X)
<
Joy ( X )
![Page 29: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/29.jpg)
29
Comments on CD theory
• Ambitious attempt to represent information in a language independent way
formal theory of natural language semantics, reduces problems of ambiguity
canonical form, internally syntactically identical
• The major problem is incompleteness no quantification
no hierarchy for objects
are those the right primitives?
how much should the inferences be carried?
fuzzy logic?
still not well studied/understood
![Page 30: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/30.jpg)
30
Understanding stories about restaurants
John went to a restaurant last night. He ordered steak. When he paid he noticed he was running out of money. He hurried home since it had started to rain.
Did John eat dinner?Did John pay by cash or credit card?What did John buy?Did he stop at the bank on the way home?
![Page 31: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/31.jpg)
31
Restaurant stories (cont’d)
She went out to lunch. She sat at a table and called a waitress, who brought her a menu. She ordered a sandwich.
Was Sue at a restaurant?Why did the waitress bring Sue a menu?Who does “she” refer to in the last sentence?
![Page 32: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/32.jpg)
32
Restaurant stories (cont’d)
Kate went to a restaurant. She was shown to a table and ordered steak from a waitress. She sat there and waited for a long time. Finally, she got mad and she left.
Who does “she” refer to in the third sentence?
Why did Kate wait?Why did she get mad? (might not be in the
“script”)
![Page 33: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/33.jpg)
33
Restaurant stories (cont’d)
John visited his favorite restaurant on the way to the concert. He was pleased by the bill because he liked Mozart.
Which bill? (which “script” to choose: restaurant or concert?)
![Page 34: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/34.jpg)
34
Scripts
• Entry conditions: conditions that must be true for the script to be called.
• Results: conditions that become true once the script terminates.
• Props: “things” that support the content of the script.
• Roles: the actions that the participants perform.
• Scenes: a presentation of a temporal aspect of a script.
![Page 35: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/35.jpg)
35
A RESTAURANT script
Script: RESTAURANT
Track: coffee shop
Props: Tables, Menu, F = food,Check, Money
Roles: S= CustomerW = WaiterC = CookM = CashierO = Owner
![Page 36: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/36.jpg)
36
A RESTAURANT script (cont’d)
Entry conditions: S is hungryS has money
Results: S has less moneyO has more moneyS is not hungryS is pleased (optional)
![Page 37: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/37.jpg)
37
A RESTAURANT script (cont’d)
![Page 38: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/38.jpg)
38
A RESTAURANT script (cont’d)
![Page 39: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/39.jpg)
39
A RESTAURANT script (cont’d)
![Page 40: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/40.jpg)
40
Frames
Frames are similar to scripts, they organize stereotypic situations.
Information in a frame:
• Frame identification
• Relationship to other frames
• Descriptors of the requirements
• Procedural information
• Default information
• New instance information
![Page 41: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/41.jpg)
41
Part of a frame description of a hotel room
![Page 42: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/42.jpg)
42
Conceptual graphs
A finite, connected, bipartite graph
Nodes: either concepts or conceptual relations
Arcs: no labels, they represent relations between concepts
Concepts: concrete (e.g., book, dog) orabstract (e.g., like)
![Page 43: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/43.jpg)
43
Conceptual relations of different arities
bird flies
dog color brown
child parents
mother
father
Fliesis a
unaryrelation
Coloris a
binaryrelation
Parentsis a
ternaryrelation
![Page 44: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/44.jpg)
44
“Mary gave John the book.”
![Page 45: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/45.jpg)
45
Conceptual graph indicating that a particular (but unnamed) dog is brown:
Conceptual graph indicating that a dog named emma is brown:
Conceptual graphs involving a brown dog
Conceptual graph indicating that the dog named emma dog is brown:
![Page 46: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/46.jpg)
46
Conceptual graph of a person with three names
![Page 47: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/47.jpg)
47
“The dog scratches its ear with its paw.”
![Page 48: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/48.jpg)
48
The type hierarchy
A partial ordering on the set of types:
t s
where, t is a subtype of s, s is a supertype of t.
If t s and t u, then t is a common subtype of s and u.
If s v and u v, then v is a common supertype of s and u.
Notions of: minimal common supertypemaximal common subtype
![Page 49: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/49.jpg)
49
A lattice of subtypes, supertypes, the universal type, and the absurd type
wr v
s u
t
![Page 50: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/50.jpg)
50
Four graph operations
• copy: exact copy of a graph
• restrict: replace a concept node with a node representing its specialization
• join: combines graph based on identical nodes
• simplify: delete duplicate relations
![Page 51: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/51.jpg)
51
Restriction
![Page 52: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/52.jpg)
52
Join
![Page 53: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/53.jpg)
53
Simplify
![Page 54: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/54.jpg)
54
Inheritance in conceptual graphs
![Page 55: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/55.jpg)
55
“Tom believes that Jane likes pizza.”
experiencer
likes
pizza
agentperson:jane
believe
proposition
object
object
person:tom
![Page 56: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/56.jpg)
56
“There are no pink dogs.”
![Page 57: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/57.jpg)
57
Translate into English
instrument
object
hand
person:john eat pizzaagent
part
![Page 58: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/58.jpg)
58
Translate into English
![Page 59: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/59.jpg)
59
Algorithm to convert a conceptual graph, g, to a predicate calculus expression
1. Assign a unique variable, x1, x2, …, xn, to each one of the n generic concepts in g.
2. Assign a unique constant to each individual constant in g. This constant may simply be the name or marker used to indicate the referent of the concept.
3. Represent each concept by a unary predicate with the same name as the type of that node and whose argument is the variable or constant given that node.
4. Represent each n-ary conceptual relation in g as an n-ary predicate whose name is the same as the relation. Let each argument of the predicate be the variable or constant assigned to the corresponding concept node linked to that relation.
5. Take the conjunction of all the atomic sentences formed under 3 and 4. This is the body of the predicate calculus expression. All the variables in the expression are existentially quantified.
![Page 60: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/60.jpg)
60
Example conversion
1. Assign variablesto generic concepts X1
2. Assign constantsto individual concepts emma
3. Represent each concept node dog(emma) brown(X1)
4. Represent eachn-ary relation color(emma, X1)
5. Take the conjunctionall the predicates from 3 and 4 dog(emma) color(emma, X1) brown(X1)
All the variables areexistentiallyquantified. X1 dog(emma) color(emma, X1) brown(X1)
![Page 61: 1 Knowledge Representation 6 6.0Issues in Knowledge Representation 6.1A Brief History of AI Representational Systems 6.2Conceptual Graphs: A Network Language.](https://reader036.fdocuments.us/reader036/viewer/2022081506/56649cc25503460f94989c53/html5/thumbnails/61.jpg)
61
Note
We will skip Section 6.3 and Section 6.4.