Domain Independent Ontology Based Intelligence Vetting using Multiple Virtual Private Networks and...
-
Upload
sherilyn-harmon -
Category
Documents
-
view
220 -
download
0
description
Transcript of Domain Independent Ontology Based Intelligence Vetting using Multiple Virtual Private Networks and...
Domain Independent
Ontology Based Intelligence Vetting
using
Multiple Virtual Private Networksand
Ontology Visualization
Dr. Paul Prueitt 1/29/2003
Alerting
Reporting
GeneratingOptions
UnderstandingPossible Outcomes
Producing and MatchingModels
DiscoveringRelationships
Detecting Factsand Events
Measurement and Instrumentation
Producing Actionable IntelligenceProducing Actionable IntelligenceIterative Process ModelIterative Process Model
Object Sciences Corporation 12/5/2002
PersistentOntology ServicesRepresentation and
Encoding
Human: Action-Perception cycle
Transaction Components
Enterprise Middleware
Synchronous & AsynchronousCollaboration ToolsVisualization
Data Schemas, OntologiesHeterogeneousDatabases
Analysis Tools andEducational Processes
Producing Actionable IntelligenceProducing Actionable Intelligence
Technology SupportTechnology Support
Object Sciences Corporation 12/5/2002
Presentation to user
Simple Reusable Architecture
HTTP (SOAP,XML)
Connecter Architecture
Persistence
Email andPacket Transfer
ApplicationsKnowledge Sharing Foundation
Data Collaboration
Visualization
Network of Virtual Private Networks
Education
Object Sciences Corporation 12/5/2002
Operational Architecture for Ontology Production
Categorizerand
InferenceEngines
Data Source
Schema-independent data
Schema-dependent data
Repository
Object Sciences Corporation 12/5/2002
Knowledge Sharing Foundation
Core Engines&
Educational Services
RepositoryKSF
Education
Vendor Tools
Data
Micro-transaction accounting system supporting outcome metrics and
revenue generation
Real time analysis
Distribution of revenue in compensation for use of tools or educational services
Industry
Universities
Diagram from Prueitt (2003)Important Innovation: Knowledge Sharing Foundation from The George Washington University
Compensation for use of data, tools, educational services, and work product
RepositoryEducation Vendor ToolsData
Embedded micro-transaction accounting system supporting outcome metrics and revenue generation
Important Innovation: IP protected micro-transaction accounting system available from Dr. Brad Cox
Advanced data mining and natural language processing
Making the case that new capabilities are within reach
Differential Ontology Framework
By the expression “Differential Ontology” we choose to mean the interchange of structural information between Implicit (machine-based) Ontology and Explicit (machine-based) Ontology
• by Implicit Ontology we mean an attractor neural network system or one of the variations of latent semantic indexing. These are continuum mathematics with only partial representation on the computer.
• by Explicit Ontology we mean an bag of ordered triples { < a , r, b > }, where a and b are locations and r is a relational type, organized into a graph structure, and perhaps accompanied by first order predicate logic (such as the Topic Maps or Cyc ontologies). This is a discrete formalism.
Implicit Explicit
Diagram from Prueitt (2002)Important Innovation: Differential Ontology from Dr. Paul Prueitt
We use LSI in a specific fashion to produce a cognitive science figure-ground corollary
Opponent-Ontologies based on
Latent Semantic Indexing
Ground text collection,
C ground
C ground produces LSI transform, T ground
Exemplar text collection, C exemplar
T ground ( C exemplar )
forms the figure-from-ground
The figure-ground then measures the real time flow of response passages
We have identified three collections of semi-structured data (natural language)
Measurement of Response I
• C ground produces a linear algebra type transform T ground . • This transform is defined from R m into R n (a transform between m and n dimensional Euclidean (vector) spaces).
• Exemplar sets, B i , I = 1, . . . , q, are made for each q categories of human-specified response.
• For example, response messages might be outputs from a harvester system that is measuring Arabic response to world events.
Ground text collection, C ground , Exemplar text collection, C exemplar , Response text collection, C response
C exemplar = B 1 B n B q
C ground = { set of documents that are designed to cover linguistic variation that needs to be picked up in a categorization process }
C response = { the output from a web Harvesting system }
Measurement of Response II
For text unit t C response , T ground ( t ) is a point in R n
Finding the correct structure and content of the exemplar set is the key
C exemplar = B 1 B 2 B q
T ground ( B i ) = N i is the set of points (neighborhood) in R n formed by the i th exemplar bin
Creating these neighborhoods is easy. However, validating that the neighborhoods have fidelity to the task at hand requires the use of an Ontology Lens that bring fidelity into focus.
The distance { T ( B i ) , T ( t ) } is now defined as the distance { centroid i , T ( t ) } where the centroid is a point in R^300 the “stands in for” the image of the i th exemplar bin. The “similarity” between a response passage and the linguistic contents of each of the exemplar bins is approximated as a single positive real number.
The Ontology Lens shows structural relationships between the categories of the exemplar set, and allows specialists to restructure the contents of the exemplar bins so that the categories exhibit a high degree of independence, as measured by the degree of relationship.
Production of Concept Metrics
ImplicitOntology
d C response
T ground (B 1)
T ground (B 2) T ground (B 3)
T ground (B 4) T ground (B 5)
Explicit
Ontology
For each response message, d, the implicit ontology produces a set of concept metrics, { mk }, and these concept metrics are used as the atoms of a logic. These atoms are used to produce an explicit ontology. The logic is then equipped with a set of inference rules. Evaluations rules are then added to produce an inference about opinions of the authors of the response set.
Diagram from Prueitt (2002)Important Innovation: Concept Metrics from Object Science Corporation
Inter-Role Collaboration using Ontology
Synchronous Collaboration
Periodic Update
Knowledge Repository
Role and Event Specific View
Views
Views
Knowledge Worker
Knowledge Worker
Object Sciences Corporation 12/5/2002
Tri-level Architecture
The Tri-level architecture is based on the study of natural systems that exist as transient stabilities far from equilibrium.
The most basic element of this study is the Process Compartment Hypothesis (PCH) that makes the observation that “systems” come into being, have a stable period (of autopoiesis) and then collapse.
Human cognition is modeled in exactly the same way. Human mental events are modeled as the aggregation of elements of memory shaped by anticipation. The tri-level architecture for machine
intelligence is developed to reflect the PCH. A set of basic event atoms are developed through observation and human analysis. Event structures are then expressed using these atoms, and only these atoms, and over time a theory of event chemistry is developed and reified.
Diagram from Prueitt (1995)Important Innovation: Tri-level architecture from Dr. Paul Prueitt
cA/eC
Neuroscience informs us that the physical process that brings the human experience of the past to the present moment involves three stages.
1) First, measured states of the world are parceled into substructural categories. 2) An accommodation process organizes substructural categories as a by-product of learning. 3) Finally, the substructural elements are evoked by the properties of real time stimulus to produce an emergent composition in which the memory is mixed with anticipation.
Each of these three processes involves the emergence of attractor points in physically distinct organizational strata. The study of Stratified Complexity appeals first to foundational work in quantum mechanics and then to disciplines such as cultural anthropology and social-biology.
categoricalAbstraction (cA) is the measurement of the invariance of data patterns using finite set of logical atoms derived from the measurement. eventChemistry (eC) is a theory of type that depends on having anticipatory processes modeled in the form of aggregation rules, where the aggregation is of the cA logical atoms.
Diagram from Prueitt (1995)Important Innovation: eventChemistry from Dr. Paul Prueitt
gF/cA/eC
We have projected a physical theory of structural constraint imposed on any formative processes, to a computational architecture based on frameworks.Various forms are conjectured to exist as part of emergent classes, and in each case each class of emergent types has a periodic table – like, in many ways, the atomic period table. The Sowa-Ballard Framework has 18 “semantic primitives”.
Zackman Framework
“According to Alvin Toffler, knowledge will become the central resource of the advanced economy, and because it reduces the need for other resources, its value will soar. (Alvin Toffler, Power Shift, 1990). Data warehousing concepts, supported by the technological advances which led to the client/.server environment and by architectural constructs such as the Zackman Framework, can prepare organizations to tap their inner banks of knowledge to improve their competitive positions in the twenty-first-century.
Evocative generalFramework (gF) theory constructs cA/eC knowledge bases directly in “conversation” with humans
Ballard/Sowa Framework
Diagram from Prueitt (2001)Important Innovation: Framework software from Drs. Paul Prueitt and Richard Ballard
Situational Logic Construction
A latent technology transform, T ground , is used to produce simple metrics on membership of documents from the response collection C response in the categories defined by the contents of the bins
C exemplar = B 1 B n2 B q
These bins are represented in the situational logic as the logical atoms A , from which a specific logic is constructed. These atoms are then endowed with a set of q real numbers that are passed to an Inference Processor.
The set of q real numbers are computed from a formal “evaluations of the structural relationship between logic atoms” using the Ontology Lens.
Atom a { r1 , r2 , . . . , rq }
The process of developing a situational logic is to be modeled after quasi axiomatic theory. In this model, new data structure are in-put as axioms, and then a process of reduction of axioms to logical atoms occurs.
The reduction also requires the Ontology Lens, (invented 2002 by Prueitt).
Diagram from Prueitt (2002)Important Innovation: Situational logics from Paul Prueitt
The Ontology Len (discovered by Prueitt, 2002) is a structural focus “instrument” that is designed to allow non-computer scientists to specify high quality exemplar sets. This is done with an Implicit Ontology to Explicit Ontology (IO-EO) loop.
• When the user puts a new unit into a bin or removes a unit from a bin, then the IO-EO loop will produce a different result.
• It is the human responsibility to govern the IO-EO loops so that the results have the properties that the human wants, mostly independence of categories, but perhaps some specific (and maybe interesting) “category entanglements”.
A graphic representation of what we call a “LSI structural similarity matrix”.
The similarity is called structural because the exact notion of semantic similarity is not known from this algorithmic computation by itself.
The paragraphs of a small exemplar set (see appendix A) are ordered as labels for the columns and rows.
One would expect that a paragraph would be structurally similar with itself, and this is in fact what one sees as a set of dots
(representing a value of 1)
down the diagonal. Diagram from SAIC (2002)
Minimal Work FlowProduction of the Explicit Ontology
ImplicitOntology
C response = C response = C response
T ground (B 1) T ground (B 2) T ground (B 3)
T ground (B 4) T ground (B 5)
Ontology
Lens
Schema-independent data
Schema-independent data is developed from the Ontology Lens, in the form of a set of syntagmatic units
{ < a, r , b > }
Where a and b are categories defined by the exemplar set, , and r is a measure of relationship.
Searching and FilteringStoring Analyzing EntitiesVisualizing LinksClusteringCategorizingResolving Cover TermsMatching Models / Detecting ChangesSimulationGenerating HypothesesGenerating Threat ScenariosStructured ArgumentationLearning PatternsUnderstanding IntentPerforming Risk AnalysisGenerating OptionsGenerating Plausible FuturesStorytellingCreating ExplanationsAlertingVisualizing GIS Data Understanding PoliciesPreparing Video SourcesProcessing Text SourcesProcessing SensorsProcessing Audio SourcesTranslating LanguagesIdentifying HumansSummarizing DataSummarizing TextSearching and FilteringCategorizingIndexingVisualizing SummariesCollaborationPresenting Recommendations Presenting Analysis ResultsPresenting Situation Status Presenting OptionsBuilding TeamsBuilding Context
Lessons Learned
Synchronous & AsynchronousCollaboration Tools
Visualization
Analysis Tools andEducational Processes
From the Intelligence Community and DoD
To the Center for Disease Control and University Research Centers
Alerting
Reporting
GeneratingOptions
UnderstandingPossible Outcomes
Producing and MatchingModels
DiscoveringRelationships
Detecting Factsand Events
Measurement and Instrumentation
Producing Actionable IntelligenceProducing Actionable IntelligenceIterative Process Model - decomposition of function/structureIterative Process Model - decomposition of function/structure
PersistentOntology ServicesRepresentation and
Encoding
Human: Action-Perception cycle
V: visualization,
D: data mining,
N: natural language processes,
S: support decision making,
R: structuring of reasoning
K: knowledge representation
T: technical support
V, N, S, R, K
V, D, N, S, R, K
V, D, N, S, R, K
V, D, N, S, R, K
V, D, N, S, R, K
V, S, R, K
V, S, R, K
V, S, R, K
S, R, K