Voronoi diagram is derived over points in feature
space which represents teachers’ input in order to
realize the desired classification. However, to reduce
the size of the neural network and make the learning
efficient, clustering procedure that enables the
subject to manage a number of teachers in a lump is
implemented (Kenji, Masakazu and Shigeru, 1999).
Our approaches, however, only utilize point-wise
cell-membership – as new information – by means of
nearest-neighbor queries and do not utilize further
geometric information about Voronoi cells since the
computation of Voronoi diagrams is prohibitively
expensive in high dimensions. Therefore, a Monte-
Carlo-Markov-Chain integration-based approach
(Polianskii and Pokorny, 2019) that computes a
weighted integral over the boundaries of Voronoi
cells, thus incorporates additional information – as
retroactive confirmation – about the Voronoi cell
structure is established. This dynamic proactive-
retroactive learning method predicts and prescribes
an action in “expected” response to an activity of
human (or interchangeably a machine), depending on
individual’s state, for one or more end-rewards, or
pay-offs at a given point in time.
Since most information related to immediate
relevance including dynamic active cognitive state
and/or active experiences, hence individuals apply a
certain set of rules that are associated with either
sequential monadic (e.g., individual’s state from a to
𝑎́ as self-improvement) or paired-comparison (e.g.,
individual’s state x compared with another
individual’s state y) with nearest neighbor or a group
where individual is a member. This, in imperfect or
asymmetric and incomplete information conditions,
creates “hidden” multi-layered combinations on
multi-dimensions – functional, non-functional, non-
discriminating and discriminating – features to
predict and determine the cognitive state (or “state”).
The group, where individual is a member, may also
apply a certain set of collective “hidden” information
associated with either linear-non-linear (e.g., a race-
car driver uses wind direction data while cornering at
speeds more than 200 mph without informing the
opponent) or paired comparison (e.g. race car the
team analyzes data of other racers’ degradation rates
on the tires and of the health of various mechanical
components, and recording the drivers’ steering,
braking and throttle inputs). In imperfect and
incomplete information conditions, this generates
aggregated “hidden” multi-layered combinations on
multi-dimensions features to predict and determine a
collective state. For example, a trading system
analyzes data to predict if the state of any trading
stock and its change with new features, conditions
and functions – the underlying latent variables –
affect the price, as an outcome, in the marketplace.
The hypotheses here are that the dynamic
proactive-retroactive learning method would derive
to be a better prediction on the individual’s current
action for future reward, as final pay-off, over a
broad range of time and information scales, including
(im)perfect and (in)complete information conditions.
For example, if the trading system predicts that the
state of the product (or service) and its change with
the underlying latent variables affect price in the
marketplace, then the expected response of the buyer
may also likely to change (either to buy immediately
or defer for the future price), thus may create a
different reward or pay-off outcome (revenue or
saving for the trader).
4 TRAINING DATA
A self-organized learning method, in accordance with
the dynamic proactive-retroactive learning method is
executed to segment a graph network data based on
bounded diffusion of collective individual
information interactions. The nearest-neighbor or
group data is determined from grouping of individual
transactional data for a group where individual is a
member. After a certain upper-bound number of
groups, the system applies a diffusion-limited
aggregation (“DLA”) – a formation process whereby
individuals in a group, as particles, and their signals
– defined as change or the first derivative in an
individual’s data – of a subject matter undergo a
stochastic process for clustering together to different
aggregates (“clusters”) of such individuals. These
signals and their changes – defined as the second
derivative in an individual’s data – are used for
predicting the group’s current state, as described
above, and applied sheafing method, (or group
theory) for “grouping” mechanism (Tennison, 2011)
– depending on the geometry of the growth, for
example, whether it be from a single point radially
outward or from a plane or line – of clusters where
the individual is a member, to determine the state.
The self-organized learning method presents
individual’s data, for example, as stimulus, at some
time t=0 and then presenting a response data at a
variable time post stimulus on the group. The
bounded diffusion in DLA, for example, may have
one additional parameter, the position of the decision
bound, say A. If at time t of the state data of the
individual (or subject matter e.g., search for an item)
is x, the distribution of the state at a future time may
be s > t, hence the term “forward” diffusion. The