Design
Document for Automated Suggestive Reasoning
September 10, 2000
Revised April 27, 2003
Internal Ontology Stream Inc Document
All Rights Reserved
1.1: Graph theory and a “Unifying Logical
Vision”
1.2: Suggestion Reasoning Object (SRO)
2.1: The methodology of descriptive
enumeration, use cases
2.2: A methodology for constructing
structures
3.1: Extending the ability to communicate
3.2: Automated suggestive reasoning as a
many-to-one projection
Design
Document for Automated Suggestive Reasoning
September 10, 2000
Revised April 27, 2003
All Rights Reserved
The purpose of this document is to bring into
one place our thoughts about innovation and computer software developed to
bring specific innovations into a market.
We then extend some fundamental principles to show a consistency between
routing and retrieval technology and our current intuitions.
The First Knowledge Sharing Core [1] is a
project that is bringing several patented innovations together to demonstrate
Information Production [2]
capability.
In the last century, foundational work on
information technology was extensive, and yet critically fragmented by
unsettled philosophical issues. The
fragmentation was not properly managed largely due to the world’s scholarly
community not having a standard formal understanding of what information is. Similar confusion exists in the scholarly
community over what human knowledge is.
Some scholars now express the view that, because
of these confusions, the world’s ecological-business system has yet to make a
full transition from a primarily industrial age to a pure knowledge age. This historical perspective brings light on
how the emergence of knowledge technology markets is being inhibited.
It is clear that deeply rooted information
control will continue to have a deep effect on economics and society. During the closing decades of the twentieth
century, computer science has dominated discussions with business process
re-engineering, expert systems, and knowledge management.
Promises have been marginally meet, and more is
expected than delivered. For many, this
observed impedance mismatch, between promised and expected, leads to the
question:
“What is the
relationship between the deeply rooted control of information and the
mal-performance of first generation information technology?”
BCNGroup founders [3] have developed an historical view over
this question. The first generation
information technology was built based on analogies to specific scientific
paradigms. These paradigms enforce a
viewpoint that requires complexity to be modeled in a purely reductionist
fashion. Reduce to rules. Reduce to processes. Reduce to knowledge. In spite of the low hanging fruit, this
foundational work is incorrect in the limit.
Taken as a limiting strategy the reduction of all things to a computer
program forms the basis for the artificial intelligence mythology [4].
Information Technology exists because of the
commercial success of Information Technology companies. The problems of the client are important to
the degree sufficient to meet market expectations, but expectations were
low.
We claim that the primary reason for the
development of the culture of Information Technology is the absence of a
standard understanding about information and about human knowledge.
Several technologies are positioned to achieve
the significant breakthrough we anticipate [1]. Some examples of how these technologies
integrate together are derived from representational theory and from new data
encoding patents. These technologies
are consistent with each other, and in line with other patents that provide
protection for our innovations.
For automated suggestive reasoning, the breakthrough
technology is the tri-level architecture [5]. Data encoding and representational theory is
core to this technology [6]. The architecture requires that events (the
middle level) be separated from context (the upper level) and dis-aggregated
into a memory core (a substructural level). The trilevel is used in the context
of a navigational aid within the computer addressable world [7] of locally defined informational
fragments and structures.
The First Knowledge Sharing Core is to be the
first example of the integration of these innovations in support of Information
Production [1].
The phrase “Ontology Stream” is used as a noun
to indicate a virtual communication space where taxonomies and concept maps are
moved about within an XML peer-to-peer environment. We are developing a new technology called Information Production
via ontology streaming.
We can regard a node as a location within the
taxonomy, or more generally a location within the structures of the ontology
(such as an generalized conceptual construct structure). The ontology appears to “stream” as the
user/program moves from one location to the next. The experience may have textual, symbolic or auditory form.
A user/program experiences representations of
knowledge during this motion. Underlying formalisms can express a calculus of
over encoded data structures as a system of metrics on knowledge
representational changes.
The formal tokens of a traversal path are a
measure of the history of the use of structures within the ontology. At any one
location, the user/program’s actual path within the ontology is a simple
connected graph, actually the concatenation of a number of line segments. However, the location itself has various
routes from that node to other locations in the ontology (see Figure 1b). The tokens of the path are indications of
the thought processes engaged by the user/program.
The CCM construct [6] is a
structure that frames the tokens of machine-readable ontology in a specific
fashion. This special structure has a
correspondence to the way human natural language is written. In the CCM generalized conceptual construct,
parts of speech are loosely organized into sections and subsections, and into
paragraphs and topics. The
correspondence between the structures of the generalized conceptual construct
and written language facilitates the projection of an automatically generated
report; written as natural language, or in some database format, or as a XML
stream.
Automated Suggestive Reasoning can be projected
because much of the structure of real time cognitive reasoning is represented
in computer addressable form.
Automated Suggestive Reasoning within a
structured taxonomy depends on having:
1) a usable representation
of a history of taxonomy traversals, and
2) special nodes in the
taxonomy where some judgment is anticipated.
The representational issue can be solved in
simple or more complex fashion.
Randomly assigned, semantically pure, tokens can be used for representation,
and still give the user the feeling that the software knows the needs of the
user/program.
Taxonomy traversal is one way to establish
context. The history gives the location
a temporary context. Different
histories can lead to the same location.
The location plus the context is changed, from one event to the
next. The location itself can be said
to “afford” moving from the location to one of a set of other locations. This affordance is represented well as a
Peircean existential graph (see Figure 1a).
The tri-level architecture is an extension of
the logics of C. S. Peirce (1839 – 1917).
This extension adds the notion of environmental, or contextual,
information. The logic of Peirce is
bi-level, substructural and compounds, and reflects his 19th century
training in chemistry. What has been
missing from his work and from work on computational emergence, are the notions
of social-affective information [8] that constrains
the aggregation of atoms into compounds. This constraint is mapped as bases of
attraction [2]
Simple traversal of the nodes of an ontology
using the tri-level architecture has been prototyped. The prototypes, done in 1997 and 1998, show an interesting
personalization of a path an individual takes while visiting a series of nodes
in a structured taxonomy. The
personalization is an aggregation of context, only partially captured by the
aggregation of representational tokens during the traversal. The human keeps
some of the personalization as awareness of situation.
We will not report the full results now, since
some work needs to be done to reproduce previously discovered traversal
characteristics and to discuss the science of interaction between an evolving
image of self and an evolving knowledge representation of context.
The bottom line is that a user gets the feeling
that the software knows the needs of the user/program. The 1997-1998 prototypes told us (Prueitt,
Murray and Kugler) that this bottom line could be achieved easily using the
trilevel architecture and very simple token representational algorithms. Specifically the members of the community of
knowledge scientists are inventing various means to representation that we feel
are “semantically pure” and simple to implement. The First Knowledge Sharing
Core [1] is
demonstrating some of these innovations.
The selection of the next location to visit is
made using the simple voting procedure described in Prueitt’s published work.
This selection process requires that each of the possible next locations have a
representation that is aggregated from a core memory set. This core memory is the atomic invariances
that are used across the entire enterprise to represent locations, histories
and categories. The core memory, like
human memory, is developed so that it has the property that any experience or
system state can be operationally represented by some composition, or
aggregation, of some subset of the set of all core memory elements. Data-morphic transformations of these
compositions are possible and quite simple.
The core memory elements can be thought of as
atomic, in a direct analogy to the atomic elements of the periodic table of
atoms. Some 90 years ago, C. S. Peirce
called this analogy the “unifying logical vision” or ULV. The ULV was then extended, specifically by a
segment of the Russian cybernetics community, and made richer with Mill’s logic
and situational semiotics. Other
citations of scholarship indicate that the ULV is a core understanding that is
really required to build knowledge technologies. The ULV is the minimal complexity that is required to establish
the scientific grounding for the trilevel architecture.
1. Represent in semantic
free form
2. Represent with
linguistic technology (Oracle ConText, Semio, VisualText)
3. Represent with
Bayesian/Shannon algorithms (Autonomy)
4. Represent with stratified vetting
(Tacit Knowledge Systems, SchemaLogic Inc.)
a. Set representational
algorithms (initially semantic free)
b. Develop fundamental set
of tokens
3) Recall
a. Make representation for
any history, event or category
b. Validate
representational structure (SchemaLogic Inc, Topic Maps)
4) Spectrum
a. Called by the container
object only in specific cases
b. Stores the category
representations for selected questions
5) Histories
a. XML file description of
histories
b. Show histories using
graphs
6) Management
a. Add/delete or modify
questions nodes to the set of nodes equipped with automated suggestive
reasoning
b. Make manual changes to
representations of events, locations or histories.
The tri-level algorithms then create a formative
ontology that is constrained by semantic valances and human inspection. The simplest of these algorithms is the
minimal voting procedure [11]. The
more sophisticated algorithms are SLIP [12] and quasi-axiomatic
theory [13].
This section focuses specifically on the
evolution of an objective methodology for developing generalized conceptual construct
structures. We envision this
methodology as the basis for certification programs and for consulting income
to the community of knowledge scientists.
Our Consulting Methodology has three components:
1) A methodology for
constructing generalized conceptual construct structures
2) A methodology designed
to open doors within properly identified markets, and to negotiate the initial
conditions for business enterprise implementation.
3) A Certification program
Currently we have work on the Methodology for
Implementation of Consulting Methodology. In addition to certification for the
Methodology for Implementation of Consulting Methodology, we have work on a
curriculum for training Information Technology professionals in all aspects of
product design, metrics and maintenance.
However, these components are not directly
addressed in this Design Document. We find it important now only to indicate
the relevance of these other two components when we frame the scope of
automated suggestive reasoning.
The central concept of the methodology is the
enumeration of topics. The enumeration
is fitted into a generalized conceptual construct topic hierarchy. In Section
2.1, we have some use cases that suggest how enumeration proceeds. In Section 2.2 we have a description of
issues that frame the development of a generalized conceptual construct within
the business environment.
Section 2.1 and 2.2 bring into the Design
Document the background necessary to understand how our innovation is viewed.
This background sets the stage for seeing automated suggestive reasoning in the
context of a technology extension for human communication.
2.1: The
methodology of descriptive enumeration, use cases
This section is modified from a June 27th,
2000 document, “Modeling within the Enterprise”. User/program and program must work together.
C-1:
Create generalized conceptual construct framework and containers (the dialog
tree)
C-1.1: Work top down to
create the generalized conceptual construct framework and containers
C-1.1.1: A user/program works top-down, completing all top sections
first and then all parts of the next level of organization.
C-1.1.2: At each level the user/program attempts to fulfill informal
completeness and independence conditions.
C-1.1.3: Finally the user/program develops topics under each of the
lowest level of the tree structure.
C-1.2: Work bottom up
to create the generalized conceptual construct framework and containers
C-1.2.1: The user/program develops a set of questions such as a questionnaire
or instructional type test.
C-1.2.2: For each question,
user/program develops topics and auxiliary information that is related to the
question.
C-1.2.3: A clustering algorithm
routs topics into bins (categories) and these bins identified with a concept
C-1.2.4: Category bins are allowed to dynamically evolve as new
topics/questions are introduced.
C-1.2.5: A clustering algorithm
clusters bin representation into higher levels of organization.
C-1.3: Mix working
bottom up and top down to create the generalized conceptual construct framework
and containers
C-1.3.1: The use develops the structure by working either top-down
(C-1.1) or bottom up (C-1.2)
C-1.3.2: Two partial
generalized conceptual construct frameworks with containers might be merged into
one.
C-1.4: During this
process (C-1.3) the user/program will specify some or all of the properties of
each of the containers as the tree is extended and populated.
C-1.5: after all of
the topics are developed, user/program attaches questions to each topic.
C-1.6: in some cases,
user/program re-contextualizes specific topics or question containers and these
containers are made accessible from different locations in the ending tree
structure.
The following issues characterize the
development of a generalized conceptual construct:
1) A survey and interview methodology produces an enumeration
of information flow events within an enterprise.
a. It is noted that an event
is not simply something that occurs, but something that occurs on a regular
basis
b. Regularly occurring
events are found via frequency analysis and co-occurrence metrics. This
principle is central to classical data-mining methods
c. Humans identify regularly
occurring events. The knowledge of
these events can be communicated using natural language or (now) a generalized
conceptual construct structure.
2) An information flow model is developed to include
both computer information exchanges as well as information exchanged during
human conversation.
a. The information model is
acquired through professional systems analysis
b. This analysis provides a
stand-alone value similar to Business Process Re-engineering consulting
services
c. The ability to conduct
survey and interviews and to do the systems analysis is an ability that we
enhance and then certify.
3) Systems analysis leads
to the identification of events where a generalized
conceptual construct provides
structured communication of knowledge between
stakeholders.
a. Structured generalized
conceptual construct communication brings reminders to stakeholders as to what
knowledge is useful in making judgments
b. The regular vetting of
judgments via a generalized conceptual construct knowledge structure provides
accountability when measured against outcome metrics
c. The generalized
conceptual construct reduces the time required to make a full reporting of
analysis and judgments
Section 3: An extension of natural language
Classes
of technologies exist that support the advanced management of knowledge
representation. Our intuition
anticipates integrating these technologies into a single enterprise system with
products as a core feature.
3.1: Extending the ability to communicate
We
take the position that the generalized conceptual construct internal structure
provides an extending technology to human communication. This is a simple claim.
Communication
that uses generalized conceptual construct structures is different from
conversation with spoken language in that generalized conceptual construct
communication can be one-to-many or many-to-one. One-to-many is accomplished sometimes, but many-to-one requires
some new capability. In our viewpoint,
many-to-one communication will be accomplished through concept aggregation and
automated vetting of community viewpoints.
In the following sections on suggestive reasoning we begin to unfold how
this is accomplished using semantically pure knowledge representation, the
trilevel architecture, and the voting procedure.
The
trilevel architecture is capable of pushing and pulling information about
“people, places and things” and to do so in a way that is computationally
simpler and, in theory, functionally superior to the push and pull technology
used by Tacit Knowledge Systems, Autonomy and Semio. The year 2000 value of this market space exceeds 6 Billon
dollars.
Our
current product does not, however, compete in the Autonomy market space. We are looking forward into a new space that
is characterized by knowledge discovery, knowledge use, and knowledge
sharing. This new space is just
becoming defined. We have the unique opportunity to establish an early
presence. Our current intellectual work
and patent policies are directed at establishing this presence.
Interpretation
and generation of human knowledge is facilitated by an automation of how humans
create the generalized conceptual construct structure and how questions are
answered. This creative process forces
human thought to conform within a form (the generalized conceptual construct
internal structure) that can then be algorithmically transformed. The structure, like human language, is not
knowledge unless perceived in the mind of an individual.
A difference between human conversation and
generalized conceptual construct knowledge sharing is due to “form-based”
computational processes. We theorize
these processes will be accomplished using data-morphosis.
The “technology” for data-morphosis is an
expression of our research and the work of certain scholars. Within the
formalization of the theory, data is seen to transform into knowledge
experience within a second order semantic control system. The scholarship (of Pribram, Bohm, Maturana,
Varela, etc) supports this notion.
3.2:
Automated suggestive reasoning as a many to one projection
Automated reasoning is now seen as an
aggregation of the representation of past events. The notion of time is not so important here. Thus we can aggregate over responses that
arrive from many people. Many-to-one
structural coupling flow from principles developed within the ontologyStream technology. Knowledge is projected from many-to-one, and
from the past-to-the-present.