The Knowledge Sharing Foundation
The knowledge sharing foundation concept was first developed (2003) as a suggestion supporting the US intelligence agency needs to develop information about event structure. Previous to this, a small group of scientists have talked about the need for a curriculum, K 12 and college, to support an advancement of cultural understanding of the complexity of natural science. By natural sciences, we mean social and cognitive science in the context of human communication.
The suggestion to support new intelligence technology deployments is predicated on the intelligence communitys responsible use and on the co-development of an open public understanding of the technologies employed.
Ontologystream Inc has developed an (fairly complete) understanding of the types of text understanding technologies available within the intelligence community.
Questions and Answers
September 16, 2003
Hyperlinks (click below)
A-1: The structure of data produced from measurement.
Real time web harvest of natural language discourse (Memetic measurement)
Global measurement of reports from medical professionals (Genetic measurement)
Measurement of social network relationships and dynamic boundaries of social systems
Cyber intrusion instrumentation and analysis (Cyber measurement)
Measurement of all NASA Earth Observation Data (see proposal to NASA)
With an aggregation of invariance in the structure of data.
With a separation of statistical and categorical artifacts in correspondence with human memory processes and human anticipatory responses
With the production of just in time machine-ontology formation as cognitive enhancement
With the development of event templates indicating meaningful constructs
A-3: Educational processes that allow users of an intelligence system to work within the limitations of machine and artificial intelligence.
University course credit
Tools expressed as un-encumbered capabilities
University certified educational support on all tools
Separation of all module services from vendor control, with the appropriate payment for actual use of intellectual property
Open-results competitive testing of all modules, with modules expressed as open source software so that code can be seen and understood
A-3: Community based compensation infrastructure
Commercial rights are protected with copyright and patents
Use-compensation based on software self-accounting to honor copyright and patents
Micro-transaction accounting and payment for services embedded in each software component.
A-1: Establish coherence within the market space
A-2: Advance the state of the art for information generation systems and open new markets
A-3: Establish a new basis for innovation
A-4: Intellectual Property mapping and patent evaluation will result in a reduction of uncertainty over ownership
Text Transformed into Structured Data
Unsupervised Pattern Mining
Situational Logic Development
Logical Inference (ΰ see current discussion about induction and deduction)
Event Detection from Data Invariance
Knowledge Flow Mapping
Social network and linguistic variation analysis
Single-algorithm Analytic Servers
Multiple User Domain
Ontology based Inference Engine
Knowledge Encoding and Propagation
Latent Semantic Technology
Concept-Based Document Indexing
Context-Free Grammar Parsing
Supervised Text Classification
Associative memory and top down expectation using neural networks
Social network theory and analysis
Invariance in the data is used to construct situational logic
Continuum mathematics methods are used to derive an "implicit ontology" from a body of documents or other data sources
An "explicit ontology" is provided by human beings, e.g. in the form of categorized sentences, and then refined using iteration
Human feedback and inference rules are used to further refine & process the derived classifications
A technique for searching datasets for signs of real world events
Takes abstract atoms of invariance observed in data, and forms interesting combinations of them
Requires a Human-in-the-loop cognitive acuity to provide interpretation of meaning
Works naturally with the output of semi-supervised text classification, clustering and categorization methodology
Fits naturally with "chemical compound" metaphor, where a period table of atomic elements are discovered and used in event detection
Post relational database technology, using new types of algorithms
(type:value) pair data constructions encode localization of information without schema
(type:value) pair data construction organizational processes has well delineated correspondence to human memory and anticipation
Referential bases support stratified processing so the ontology constructions can be formative and situational
A-1: The systemic development of educational processes involves
the development of consensus on what are the separated techniques in computational intelligence
The mapping of scholarly literature helps in comprehensive mapping of patent disclosure and copyright
A-2: As this consensus develops,
the description of general systems theory, cognitive and social science is made available within the academic community
a "liberal arts" education in the knowledge sciences is made available to intelligence analysts
A-1: Analytic features are to be replicated from existing software and implemented as separated components.
A mapping of all software based innovation in the area of computational intelligence is developed based on latent semantic technology indexing of patents and copyrights
In cases where the core technology has legitimate ownership, then licenses are arranged
In cases where the core technology is developed by the government then the core engines are made public domain
Each core technology component is rendered in binary with an internal accounting module that reports usage as part of a knowledge flow mapping and use compensation (when appropriate)
A-1: The system of core objects is open to innovation.
Negotiations to acquire a new innovation occur through Intellectual Property mapping processes and comprehensive testing of object inherit capability
Innovations targeted for acquisitions are studied in highly structured usability testing that includes deep education in the innovations' inherit capabilities.
These acquisition studies are conducted in the public view and are not governed by commercial processes.
Knowledge Sharing and SenseMaking diagram
The George Washington University
Founder: Ontology Stream Inc