Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of...

10

Click here to load reader

Transcript of Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of...

Page 1: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

Procedia Computer Science 22 ( 2013 ) 211 – 220

1877-0509 © 2013 The Authors. Published by Elsevier B.V.Selection and peer-review under responsibility of KES Internationaldoi: 10.1016/j.procs.2013.09.097

ScienceDirect

17th International Conference in Knowledge Based and Intelligent Information and Engineering Systems -

KES2013

Integrating Robot Task Planner with Common-Sense Knowledge Base to Improve the Efficiency of Planning

Ahmed Al-Moadhen*, Renxi Qiu, Michael Packianather, Ze Ji, and Rossi Setchi School of Engineering, Cardiff University, Cardiff, CF24 3AA, UK

Abstract

This paper presents a developed approach for intelligently generating symbolic plans by mobile robots acting in domestic environments, such as offices and houses. The significance of the approach lies in developing a new framework that consists of the new modeling of high-level robot actions and then their integration with common-sense knowledge in order to support a robotic task planner. This framework will enable interactions between the task planner and the semantic knowledge base directly. By using common-sense domain knowledge, the task planner will take into consideration the properties and relations of objects and places in its environment, before creating semantically related actions that will represent a plan. This plan will accomplish the user order. The robot task planner will use the available domain knowledge to check the next related actions to the current one and the action's conditions met will be chosen. Then the robot will use the immediately available knowledge information to check whether the plan outcomes are met or violated. © 2013 The Authors. Published by Elsevier B.V. Selection and peer-review under the responsibility of KES International. Keywords: Task Planner; Common-Sense knowledge; Semantic Action Model; Semantic Network; ConceptNet.

1. Introduction

One of the most important key tools which enable the robots to work, for servicing in a domestic environment, is knowledge. This knowledge will support them in completing their service tasks in a real environment. A Semantic network, ConceptNet or Graph techniques can be used to represent this knowledge. These techniques are capable of representing knowledge as concepts and the semantic relations between these concepts. The description of concepts and the relationships between concepts is represented as an ontology. The semantic information can improve robot reasoning and knowledge inference.

* Corresponding author. Tel.: +44-743-851-6890. E-mail address: [email protected].

Available online at www.sciencedirect.com

© 2013 The Authors. Published by Elsevier B.V.Selection and peer-review under responsibility of KES International

Page 2: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

212 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

The actions that are used with the knowledge to change robot states are also important. This paper will present action models as an ontology that reflect its inputs, preconditions, outputs and effects, and then connects its arguments to a common-sense knowledge base to enrich the planning domain with semantic facts about the objects and places in the robot environment. Action models in this case are to some extent similar to web services [1].

The main contributions of this paper are: The robot high-level actions will be modeled in a new model called a Semantic Action Model (SAM). In

this model, the action model consists of two parts: one is the profile part and the second is the process part. Then the action inputs, outputs, preconditions and effects predicates will be represented based on the semantics of their parameters.

The common-sense knowledge will be integrated with a Semantic Action Model (SAM) to generate a planning domain that will be used to support the task planner and then improve the robot efficiency. This integration will make the robot more aware of facts and information that an ordinary person is expected to be.

A new algorithm has been devised to engage the information in an action model (SAM), with its related data in common-sense knowledge, and to produce a planning domain represented in PDDL format. Then the planning domain will be fed into an external planner to produce the plan for robot tasks.

The rest of this paper is organized as follows: in Section 2, a list of some related works are presented. In section 3, the architecture of the framework will be presented. The experiments that are used to verify the benefit of using the framework will be described in Section 4. In Section 5, conclusions and future works will be suggested.

2. Related Works

Using semantic information to support the robot task planner is a topic of great interests in this field. The semantic knowledge could be integrated with spatial knowledge to produce semantic maps [2]. These maps help a planner in reasoning and in inferring the types of objects from their places. Another benefit of using semantic information is to detect any violation of facts that are stored in the knowledge base [3]. So, when the robot detects, for example, that a tooth brush is in the living room, it will be register that the brush must be returned to the bathroom, since according to its knowledge base, the brush’s correct position is in the bathroom. The planner must first initiate a goal to recover this violation. In more recent work [4], the semantic information was used to build a robot map. The building process took in its consideration the semantics of the objects and places in the environment.

The task planner efficiency was improved in [5] through abstracting robot world definitions. The improvement was done in computational efficiency by starting the planning process in the high level then stepping down and refining the plan by discarding objects or places irrelevant to this plan.

The semantic information in [6] could be gained from a robot sensor, then used to build semantic maps. These maps could be organized in a multi-hierarchical fashion and used to support robot navigation in its environment. The linking process between spatial and semantic information is done by using the anchoring technique [7]. This framework gives the robot the ability to use and infer new semantic information from its environment and this leads to improvements in its operation.

In [8], the author explored a new benefit of semantic information. This makes robot task planning more efficient in complex tasks. So, this led the robot to deal with a great number of objects, places and actions. In the first, the researcher built a semantic plan consisting of the classification of objects, places and actions, and then presented a generalized version of the requested task. Afterwards, the plan can be used for removing redundant information to enforce the robot planning in an efficient manner.

Page 3: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

213 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

Other research focused on semantic knowledge as a source of information to monitor execution [9]. That paper presented an intelligent plan-execution monitoring method, which used semantic information in the robot domain to extract implied expectations that the robot can use to verify the correctness of executing its plan actions.

The author in [10] reported a new knowledge processing system called KNOWROB, designed for making personal robots more autonomous in their work. He described the necessary control routines and sequence of actions that automatically execute the particular requested task. The knowledge processing system integrates encyclopaedic knowledge, an environmental model, action-based reasoning and human observations, and has the ability to access all this information in a uniform, symbolic way. The map that is used as a source of information to the KNOWROB knowledge processing system is described in [11]. This map combines spatial knowledge about objects in the environment and their relationships, and then provides a robot with information about these objects.

The semantic information can be used with spatial thing categorization [12], and then a model is built that is used to give the robot high levels of communication with humans. This model is composed of layers, which represent maps, such as metric, navigation, topological and conceptual at different level of abstraction.

The work in [13] lists some challenges that faces robot operation in unstructured environments, and then proposes methods to solve them. For example, household robots that work in real-world environments are highly dynamic, so the robot needs specific methods to deal with these uncertain situations. The author’s method for dealing with these situations was to create probabilistic first order models by using statistical relational models to represent probabilistic knowledge.

A good action representation that integrates a good planning algorithm is important for the task performance of many intelligent systems. The work in [14] focused on that point and thus made integration between the hierarchy of action (STRIPS-style) with a plan decomposition hierarchy. This integration seems ideal for natural language processing and more expressive and efficient.

[15] proposes a framework for representing the knowledge of household service robots. The framework is divided into classes, axioms and rules to enable a robot to recognize objects and navigate while inferring localization-related knowledge, even if there was hidden and partial data due to noisy sensor data. The task planner in [16] was improved by optimizing the symbolic representation of the robot environment by using a genetic algorithm.

The next section will describe planning system architecture, which is used to improve the efficiency of task planning in a domestic environment.

3. Planning System Architecture

Figure 1 illustrates the main components of the planning system. It consists of the semantic action model (SAM), the ontology manager and the transformation algorithm, problem and domain definition, planner, and the common-sense knowledge base. The user provides the system with the task to be solved by the robot.

3.1 Planner

At the core of the system there is a planner. The planner takes as input the problem and domain definitions in PDDL format. This format became a standard for planning competitions since 1998. There are many planners in the world. In this paper, the planner that is used to produce the plan is known as Metric-FF [17]. This planner has become very popular during the last few years. The operation in this paper is not limited to this planner, but any STRIPS-style one can be used. The Metric-FF planner may already have some other

Page 4: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

214 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

techniques for improving its computational efficiency of planning, so the overall improvement of the proposed framework can be high.

3.2 Semantic Action Model (SAM)

The domain definition starts from the modelling of the actions, which can be done by the robot. The action model takes STRIPS-style to represent its preconditions and effects. This model is stored as an OWL file that represents it as ontology in order to connect action parameters to the main ontology (knowledge base). This main ontology is represented in the proposed framework as a common-sense knowledge base. This idea of SAM is clear in figure 2, which explains the model of MOVE action.

The main components of the proposed semantic action model are shown as a block diagram in figure (3) (left side). These components are related to each other in a semantic way and stored as an OWL file [18]. The components are:

1. Profile part: shows an action as a function of three basic types of information: what robot part provides this action, what function the action does, and what robot part hosts features that specify characteristics of the action. The profile describes the action inputs and outputs, and also the preconditions required by the action and the expected effects that result from the execution of the action. 2. Process part: is a specification of the ways a robot may interact with an action. There are two types of processes: the atomic process, which has one (or more) simple input(s) and one simple output, and the composite process, which takes more than one input and produces one or more outputs. Process part is the main part in the action model and the algorithm that works on it will be presented in next section. The process inputs and outputs and the world states are used to produce new information, while preconditions and effects can produce change in the robot world as a transition from one state to another state.

The process inputs represent information that is required for operating the process. The process outputs represent the information provided to the user or the next actions. The process preconditions should all be verified for the process to be successfully invoked. The process effects show the results of applying the actions.

Fig. 1. Planning System Architecture

Page 5: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

215 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

There are some properties or relations that describe process functions and their parameters. These properties are hasInput (used to describe action inputs), hasOutput (used to describe action outputs). Sometimes, when necessary, other properties are used to describe the action preconditions (hasPrecondition) and effects (hasEffect).

3.3 Semantic Action to Domain Definition Algorithm

The next step after semantic action representation is the algorithm that is used to convert the semantic action model represented in OWL file into pddl format. Figure 3 shows the mapping process between the action model and the planning model. Algorithm1 shows the pseudocode that is used to convert SAM to PDDL format.

.

Algorithm1 Procedure SAM to PDDL ( OWL_SAM set M): PDDL planning domain set P P=0 For each m M v= input parameter group of m preco=(and(conjunction of all inputs and preconditions defined in model m)) effec= (and (conjunction of all outputs and effects defined in model m)) Add (A(v) preco effec) to P End

Fig. 2. Move Semantic Action Model (SAM)

Page 6: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

216 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

3.4. Common-Sense Knowledge Base

This knowledge is built to represent the information that is clear for humans, but not explicitly clear for robots. This knowledge base will be used to support the task planner to get a more efficient and reasonable plan. The planner will take into consideration the semantics of objects and places that make up the robot world.

Common-sense knowledge has been collected and made publicly available by the Open Mind In-door Common Sense project (OMICS) [19] and it is intended to be used in indoor mobile robotics. This source of knowledge is complemented with openCyc ontology [11] in order to provide robots with a broad range of knowledge that is necessary to accomplish their tasks. Figure (4) shows an excerpt of the knowledge that is used in this paper. This knowledge base is stored as an OWL file to reflect the ontology of the environment that consists of rooms such as the kitchen and bathroom, and objects such as the fridge, TV, etc.

Figure 4 Robot Common-Sense Knowledge Base

Semantic Action Model

Fig. 3. Mapping between Semantic Action Model and Planning Domain

Planning Domain

Preconditions

Name

Effects

Action Profile Part

Action Process Part

Precondition

Outputs

Inputs

ID

Effects

Page 7: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

217 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

4. Experiments

This section will demonstrate the utility of the framework proposed in this paper in different situations by explaining some useful scenarios. There are two scenarios used to verify the supporting of the robot task planner with SAM and common-sense knowledge. The first scenario deals with a navigation situation, while the second deals with a manipulation situation.

4.1 Set up

In this paper, the experiments were done according to the programming languages and software available to model the robot environment. The SAM and knowledge base were built by using Protege [20], integrated with a Pellet inference system [21]. The task planner was done using Metric-FF [17]. The algorithm was implemented using Java interfaced with Pellet and Metric-FF planner through API for both Pellet and planner. The ontology in the common-sense knowledge base was coded by hand, whereas the domains in the planner were set automatically by algorithm1.

Figure (5) shows the environment that is used to test the benefits of using a semantic action model with common-sense knowledge to support the robot task planner. This environment is divided into five rooms, namely the kitchen, office, bedroom, bathroom and living room. Each room is divided into places (represented as small squares in each room that are numbered to be clear for the reader) that are used by the robot to navigate around its current room or to go out to other rooms. This environment also contains some objects related to each room, such as a microwave, table, TV, etc. The ontology representation of this environment was shown in figure 4. The following conventions are used in these scenarios: R means robot, Pi means position i.

15 16 17

18 19 20

1 2 3

4 5 6

9 10 11

12 13 14

21 22 23

24 25 26

7 8

Office Kitchen

Bedroom Bathroom

Living

Fig. 5. Robot Environment

Page 8: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

218 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

4.2 Navigation Scenario

Initially, the robot was placed in the kitchen in position 1 and the user gave orders to the robot to go to the bedroom at position 12. Assume that position 4 contains table and positions 5, 10, 19, and 22 are doors. There are two scenarios that are used to test the framework of this environment. The first one deals with navigation without using the knowledge base, whereas the second one allows support from the knowledge base.

4.2.1 Navigation Planning without Common-Sense Knowledge

The planning domain of this scenario contains the following definition of the move action that is gained after feeding SAM in figure 2 to the algorithm1.

(:action move :parameters (?r-robot ?from-place ?to-place) :precondition (and (robot ?r-robot) (place ?from-place)(place ?to-place)(at ?r-robot ?from-place) (isconnectedto ?from-place ?to-place) )

:effect (and (at?r-robot?to-place) (not (at?r-robot?from-place)))

The plan that is generated from this domain contains the following sequence of actions. (MOVE R P1 P4) (MOVE R P4 P5) (MOVE R P5 P7) (MOVE R P7 P10) (MOVE R P10 P9) (MOVE R P9 P12)

It is clear that the robot will hit the table in position 4, as the robot does not take into consideration that the table is an obstacle and must be avoided. This situation may lead to the destruction of the robot.

4.2.2 Navigation Planning with common-sense knowledge

The planning domain of this scenario is the same as above with the addition of a new definition of obstacle: (or(not(obstcle ?to-place))) The plan according to this domain is described below and it is clear that the robot will avoid hitting the table

by moving to position 2 then to position 5, before continuing on its path to position 12. This is proof that when the robot is aware of the semantics of things, it can avoid critical situations. So, according to this modification the generated plan is:

(MOVE R P1 P2) (MOVE R P2P5) (MOVE R P5 P7) (MOVE R P7 P10) (MOVE R P10 P9) (MOVE R P9 P12)

4.3 Manipulation Scenario

In this scenario, the environment consists of (in addition to what was explained above) some objects like a milk box, table, etc. This scenario represents the user order that the robot must bring the milk box from the

Page 9: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

219 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

kitchen to the living room and put it in any suitable place. Because of space limitations, this scenario follows the same steps as in the previous scenario to generate the planning domain, and only the generated plan will be displayed.

4.3.1 Manipulation Planning without Common-Sense Knowledge

The generated plan according to this sub-scenario is: (Take R Milk_box P1) (MOVE R P1 P2) (MOVE R P2P5) (MOVE R P5 P7) (Drop R Milk_box Tv) It is clear that the robot could put the milk box on the TV because the planner does not take into account that

the objects must be put on a flat surface. This information can be gleaned only from common-sense knowledge.

4.3.2 Manipulation Planning with Common-Sense Knowledge

In this sub-scenario, the planner get support from the knowledge base and it defines the object workspaces as any object that has a flat surface. So, the new plan generated according to this domain is:

(Take R Milk_box P1) (MOVE R P1 P2) (MOVE R P2P5) (MOVE R P5 P7) (Drop R Milk_box Table)

5. Conclusions and Future Works

This paper presents a new representation of high-level robot actions called SAM and then integrates them to the common-sense robot knowledge-base. This integration provides the robot with the abilities to be more aware of the semantic of the places and objects in it environment. This awareness makes the robot avoid hitting obstacles, losing its way to other rooms, putting objects in wrong workspaces, and detecting whether or not it is possible to carry objects. These features make the robot more flexible in completing its task.

This work needs modification in the path of combining this framework, presented in this paper, with a perception model. This will help the robot to add new objects and places to its knowledge base.

References

[1] “OWL-S 1.1 Release.” [Online]. Available: http://www.daml.org/services/owl-s/1.1/. [Accessed: 15-Mar-2013].

[2] C. Galindo, J.-A. Fernández-Madrigal, J. González, and A. Saffiotti, “Robot task planning using semantic maps,” Robotics and Autonomous Systems, vol. 56, no. 11, pp. 955–966, Nov. 2008.

[3] C. Galindo and A. Saffiotti, “Inferring robot goals from violations of semantic knowledge,” Robotics and Autonomous Systems, vol. null, no. null, Jan. 2013.

[4] N. Blodow, L. C. Goron, Z.-C. Marton, D. Pangercic, T. Ruhr, M. Tenorth, and M. Beetz, “Autonomous semantic mapping for robots performing everyday manipulation tasks in kitchen environments,” in 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2011, pp. 4263–4270.

Page 10: Integrating Robot Task Planner with Common-sense Knowledge Base to Improve the Efficiency of Planning

220 Ahmed Al-Moadhen et al. / Procedia Computer Science 22 ( 2013 ) 211 – 220

[5] C. Galindo, J.-A. Fernandez-Madrigal, and J. Gonzalez, “Improving Efficiency in Mobile Robot Task Planning Through World Abstraction,” IEEE Transactions on Robotics, vol. 20, no. 4, pp. 677–690, Aug. 2004.

[6] C. Galindo, A. Saffiotti, S. Coradeschi, P. Buschka, J. A. Fernandez-Madrigal, and J. Gonzalez, “Multi-hierarchical semantic maps for mobile robotics,” in 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2005, pp. 2278–2283.

[7] S. Coradeschi and A. Saffiotti, “An introduction to the anchoring problem,” Robotics and Autonomous Systems, vol. 43, no. 2–3, pp. 85–96, May 2003.

[8] J. -a. F. C. Galindo, “Using semantic information for improving efficiency of robot task planning,” in: ICRA Workshop: Semantic Information in Robotics, 2007.

[9] A. Bouguerra, L. Karlsson, and A. Saffiotti, “Semantic Knowledge-Based Execution Monitoring for Mobile Robots,” in Proceedings 2007 IEEE International Conference on Robotics and Automation, 2007, pp. 3693–3698.

[10] M. Tenorth and M. Beetz, “KNOWROB — knowledge processing for autonomous personal robots,” in 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2009, pp. 4261–4266.

[11] M. Tenorth, L. Kunze, D. Jain, and M. Beetz, “KNOWROB-MAP - knowledge-linked semantic object maps,” in 2010 10th IEEE-RAS International Conference on Humanoid Robots, 2010, pp. 430–435.

[12] Oscar Martinez Mozos and Patric Jensfelt and Hendrik Zender and Geert-Jan M. Kruijff and Wolfram Burgard, “From labels to semantics: An integrated system for conceptual spatial representations of indoor environments for mobile robots,” Proceedings of the IEEE ICRA Workshop: Semantic information in robotics, pp. 33–40, 2007.

[13] M. B. Moritz Tenorth, Dominik Jain, “Knowledge Representation for Cognitive Robots,” In Künstliche Intelligenz, Springer, vol. 24, no. 3, pp. 233–240, 2010.

[14] C. Kemke and E. Walker, “Planning with Action Abstraction and Plan Decomposition Hierarchies,” in 2006 IEEE/WIC/ACM International Conference on Intelligent Agent Technology, 2006, pp. 447–451.

[15] Y.-T. Lim, Gi Hyun; Hwang, Wonil; Suh, Hyo-Won W.; Choi, Jung-Hwa; Park, “Ontology-based multi-layered robot knowledge framework (OMRKF) for robot intelligence,” in 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2007, pp. 429–436.

[16] C. Galindo, J.-A. Fernandez-Madrigal, J. Gonzalez, A. Saffiotti, and Pä. Buschka, “Life-Long Optimization of the Symbolic Model of Indoor Environments for a Mobile Robot,” IEEE Transactions on Systems, Man and Cybernetics, Part B (Cybernetics), vol. 37, no. 5, pp. 1290–1304, Oct. 2007.

[17] J. Hoffmann and B. Nebel, “The FF Planning System: Fast Plan Generation Through Heuristic Search,” Journal of Artificial Intelligence Research, vol. 14, no. 1, pp. 253–302, 2001.

[18] “OWL Web Ontology Language Reference.” [Online]. Available: http://www.w3.org/TR/owl-ref/. [Accessed: 15-Mar-2013].

[19] R. Gupta and M. J. Kochenderfer, “Common sense data acquisition for indoor mobile robots,” Proceedings - Nineteenth National Conference on Artificial Intelligence (AAAI-2004): Sixteenth Innovative Applications of Artificial Intelligence Conference (IAAI-2004);San Jose, CA;25 July 2004through29 July 2004;Code63904, pp. 605–610, Jul. 2004.

[20] “The Protégé Ontology Editor and Knowledge Acquisition System.” [Online]. Available: http://protege.stanford.edu/. [Accessed: 06-Mar-2013].

[21] E. Sirin, B. Parsia, B. C. Grau, A. Kalyanpur, and Y. Katz, “Pellet: A practical OWL-DL reasoner,” Web Semantics: Science, Services and Agents on the World Wide Web, vol. 5, no. 2, pp. 51–53, Jun. 2007.