-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathlecture.Rmd
773 lines (510 loc) · 24.8 KB
/
lecture.Rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
---
title: "Analogical Reasoning"
subtitle: "Module 6, Neuroscience 299, University of California, Berkeley"
author: "Ross W Gayler <font size=\"2\">ORCiD: 0000-0003-4679-585X</font> ross@rossgayler.com"
date: "2021-10-06"
output:
ioslides_presentation:
widescreen: true
incremental: false
editor_options:
markdown:
wrap: 72
# This is an ioslides presentation
# Present the html file by opening it in a web browser (e.g. Chrome)
# The presentation can be printed to PDF from the web browser
#
# Unfortunately:
# 1. The html presentation is fixed size, which means displaying it for screen
# sharing or projection can be a bit fiddly to get the slide to fill the
# view port (https://github.com/rstudio/rmarkdown/issues/185).
# 2. Even when the presentation is set to widescreen (16:9) the print to PDF
# aspect ratio is set to 4:3.
#
# Next time I will try to find a more modern presentation package that allows
# printing to PDF.
---
```{r LICENSE, include=FALSE}
# "Analogical Reasoning" (c) by Ross W. Gayler 2021
# This document is the source for the slides of the lecture
# "Analogical reasoning" presented 2021-10-06 as Module 6 of Neuroscience 299:
# Computing with High-Dimensional Vectors at the Redwood Center for Theoretical
# Neuroscience, University of California, Berkeley.
# https://redwood.berkeley.edu/courses/computing-with-high-dimensional-vectors
#
# This document is licensed under a Creative Commons Attribution 4.0 International License.
#
# You should have received a copy of the license along with this work.
# If not, see http://creativecommons.org/licenses/by/4.0/.
```
```{r setup, include=FALSE}
knitr::opts_chunk$set(echo = FALSE)
```
## Motivation & Objectives
Analogy is really cool and central to cognition
Analogy is a good use case for the unique properties of VSA/HDC
- What makes analogy hard for conventional computing?
- Which VSA/HDC features might help with analogy?
- Not a solved problem
Use a set of attempts at aspects of analogy to highlight some VSA design
issues
## Outline
What is analogy?
Why is analogy hard for conventional computing?
VSA design examples:
- Plate - Similarity of hand crafted vectors
- Mikolov et al - Similarity of learned word vectors
- Kanerva - Simple substitution
- Emruli et al - Substitution with lookup
- Gayler & Levy - Settling on substitution
## What is *ANALOGY*?
*ANALOGY* $\triangleq$ what analogy *really* is
Whatever it is, *ANALOGY* as a cognitive phenomenon is a complex,
nuanced thing
- Everybody presents a different partial view of *ANALOGY*
- Tendency to interpret the partial view as the whole thing
- Analogical reasoning
- Proportional analogies
- Grand analogy (analogy as a party trick)
- Please don't do that
*ANALOGY* is too big to fit in this lecture, so I will resort to
assertions and hand waving to explain enough of it for current purposes
## Analogy is the core of cognition
Quote from Hofstadter (2006):
analogy-making $\triangleq$\
the perception of common essence^1^\
between two things^2^
<font size="4"> ^1^ In one's current frame of mind\
^2^ Thing $\triangleq$ mental thing </font>
See also\
Gust et al (2008)\
Chalmers et al (1992)\
Blokpoel et al (2018)
I will jump off from Blokpoel:\
cognition as inference to the best explanation
## Inference to the Best Explanation
The cognitive loop:\
Given some inputs (evidence $e$) and a set of potential explanations
(hypotheses $H$) find the hypothesis ($h$) that best explains the
evidence
Evidence and hypotheses are represented relationally (trees/graphs)
- A bet that natural regularities are "best" captured as
transformations
"explains" is interpreted as graph structure matching - (sub)graph
isomorphism
- structural similarity = literal similarity \| optimal substitution
of literals
- analogical "common essence" = common relational structure
Partial structure matching enables inference by carrying structure from
one representation to another (pattern completion via autoassociative
memory)
## Where do the hypotheses come from?
Hypotheses are generated from *all* the agent's relevant knowledge
The hypothesis space must be open-ended, to allow for explaining novelty
- Hypotheses must be compositional
- Allows infinite productivity
- Allows novel compositions of familiar components
- Like a grammar for hypotheses
- Substitution enables composition (there may be other mechanisms)
## Example: Relational representation
solar system = base structure = hypothesis (on this slide)\
atom = target structure = evidence (on this slide)\
structural similarity = literal similarity \|
<font size="5">$\{sun \mapsto nucleus, planet \mapsto electron\}$</font>
![<font size="1">Chalmers et al
(1992)</font>](assets/chalmers-analogy_graphs.png){width="60%"}
## Example: Relational representation of evidence
![<font size="1">Blokpoel et al
(2018)</font>](assets/blokpoel-evidence.png){width="100%"}
## Example: Relational representation of knowledge
![<font size="1">Blokpoel et al
(2018)</font>](assets/blokpoel-knowledge.png){width="38%"}
## Example: Analogical augmentation
![<font size="1">Blokpoel et al
(2018)</font>](assets/blokpoel-augmentation_2.png){width="80%"}
## Example: Augmentation of evidence
![<font size="1">Blokpoel et al
(2018)</font>](assets/blokpoel-augmentation_1.png){width="70%"}
## Example: Explanation
![<font size="1">Blokpoel et al
(2018)</font>](assets/blokpoel-explanation_2.png){width="72%"}
## Why is analogy hard for conventional computing?
Subgraph isomorphism, of two graphs, is NP-complete (intractable)
- Considers all possible vertex mappings
- The "obvious" approach is brute-force exhaustive enumeration
- Each vertex mapping provides very little information about the
adequacy of the other vertex mappings
Considering all the base structures in the agent's knowledge is much
larger
Considering the transitive closure of analogical augmentations is much
larger
## Preview: Which VSA features might help?
<font size="5">
Hardware parallelism (elementwise operations with small fan-in)
Mathematical parallelism (avoids explicit enumeration)
- The hardware only "sees" the total vector
- Distributive parallelism
- (A + B + C) \* $\rho$(P + Q + R) = A\*$\rho$P + A\*$\rho$Q + ...
B\*$\rho$P + B\*$\rho$Q + ...
- Equational parallelism
- T = (A + B + C) = (P + Q + R + S) = (X \* Y \* Z) = ...
- Enables holistic transformations
Substitution is a primitive (via binding)
- Every value is potentially a variable
</font>
## Plate (1994) - Hand crafted similarity
Focus of Chapter 6 of Plate's thesis (1994) is the use of dot-product
similarity as a measure of structural similarity of representations
Reports experiments with hand-crafted representations aimed at
qualitatively reproducing the results of psychology research into human
judgement of analogical similarity under varying contributions of
component similarity to overall similarity.
My take,
- superficial similarity $\overset{\operatorname{very}}{\approx}$
similarity of arguments of relations
- structural similarity $\overset{\operatorname{very}}{\approx}$
similarity of pattern of relations
but researchers are free to suit the details of their definitions to
their needs
## Example stimuli
<font size="5">
**P** (Probe) "Spot bit Jane, causing Jane to flee from Spot"
**LS** (Literal Similarity) "Fido bit John, causing John to flee from
Fido." (Has both structural and superficial similarity to the probe P.)
**SF** (Surface features) "John fled from Fido, causing Fido to bite
John." (Has superficial but not structural similarity.)
**CM** (Cross-mapped analogy) "Fred bit Rover, causing Rover to flee
from Fred." (Has both structural and superficial similarity, but types
of corresponding objects are switched.)
**AN** (Analogy) "Mort bit Felix, causing Felix to flee from Mort." (Has
structural but not superficial similarity).
**FOR** (First-order-relations only) "Mort fled from Felix, causing
Felix to bite Mort." (Has neither structural nor superficial similarity,
other than shared predicates.)
</font>
<font size="1">Plate (2000) </font>
## Base and token vectors
![<font size="1">Plate
(1994)</font>](assets/plate-base_token.png){width="74%"}
## Stimulus episode representation construction
Probe episode (**P**): "Spot bit Jane, causing Jane to flee from Spot"
![<font size="1">Plate
(1994)</font>](assets/plate-probe_vector.png){width="100%"}
Note the addition of "lower level" components into the representations
- These are not strictly necessary for representing the structure
Construction of all the episode representations follows the same scheme
## Dot-product similarity with Probe
![<font size="1">Plate
(2000)</font>](assets/plate-similarities.png){width="100%"}
## Reminders of dot-product similarity properties
<font size="5">
$sim(A, A') = sim(A, (A + X)) \gt 0$
$sim(A, (A \otimes P)) \approx 0$
$sim((A \otimes P), (A' \otimes P)) = sim(A, A') \gt 0$
$sim((A \otimes P), (A' \otimes P')) = sim(A, A') \times sim(P, P') \gt 0$
$sim((A \otimes B \otimes \ldots \otimes X), (A' \otimes B \otimes \ldots \otimes X)) = sim(A, A') \gt 0$
$sim((A \otimes B \otimes \ldots \otimes X), (A' \otimes B \otimes \ldots \otimes X \otimes Y)) \approx 0$
If using self-inverse products:
$sim((A \otimes B \otimes \ldots \otimes X), (A' \otimes B \otimes \ldots \otimes X \otimes Y) \otimes Y^\dagger) = sim(A, A')$
$\dagger$ $. \otimes Y$ is equivalent to applying the substitution
$Y \mapsto \textbf{1}$
</font>
## My interpretation of Plate (1994) Chapter 6
<font size="5">
Representation of structure *requires* product operations,\
which destroys dot-product similarity of result to arguments
- structural similarity $\neq$ only the dot-product similarity of core
structure
- Needs something extra
Plate decorates the composite core structures with components to create
similarity
- Might be *ad hoc* (depends on whether it is natural for the
construction process)
- Might be missing necessary structure
- Predicates are not represented as unique instances,
- "Spot bit Fido causing Felix to bite John" is ambiguous
- but representing them as unique instances might destroy their
similarity
- $sim((bite_1 \otimes bite_{agt} \otimes Spot), (bite_2 \otimes bite_{agt} \otimes Spot)) \overset{?}{\approx} 0$
</font>
## Dot-product similarity is local
Dot-product similarity is at the heart of VSA system dynamics
Dot-product similarity is very "local"
- Almost all vectors are quasi-orthogonal to current state vector
- Only a tiny fraction of the vector space has nonzero similarity with
state
- Miraculous luck if all directions of interest are local to the state
- Similarity driven dynamics alone can't select between non-local
directions
Relational structure encoded by Multiply and Permute, which are
orthogonalising
- Something needs to be done to map relational structure into the
local space so that it can engage the similarity dynamics
## Mikolov et al (2013) - Learned word similarity
Proportional analogy with learned "semantic" vectors for words
- $a:a' :: b:b'$
- $man:woman :: king::queen$
- $V_{woman} - V_{man} + V_{king} \overset{?}{=} V_{queen}$
![<font size="1">Mikolov et al
(2013)</font>](assets/mikolov-offsets.png){width="70%"}
## Successful analogy? Not so much
Doesn't work as well as originally thought (Rogers et al, 2017)
- Relies on excluding $b$ from answer set (or using multiple choice)
- Works best when $a$, $a'$, and $b$ are relatively similar to each
other and to $b'$
- Poor at some classes of relations, e.g. synonymy, antonymy
*ANALOGY* enables proportional analogy, but proportional analogy $\neq$
*ANALOGY*
"semantic" vectors don't capture *SEMANTICS*
- Semantic vectors don't know how to change a flat tyre
- Captures a narrow subset of linguistic regularities induced by
*SEMANTICS*
- *ANALOGY* engages with *SEMANTICS*
## Vector semantics and dot-product similarity
Proportional analogy via semantic vectors implies semantics $\equiv$
additive features
$V_{woman} - V_{man} + V_{king}$\
$= (V_{person} + V_{FvsM}) - (V_{person} - V_{FvsM}) + (V_{person} - V_{FvsM} + V_{royal})$\
$= V_{person} + V_{FvsM} + V_{royal}$\
$= V_{queen}$\
Additive features can't capture *SEMANTICS*
*ANALOGY* is about structural relational similarity
- Representing relational structure requires product operators
- Static dot-product similarity structure is driven by additive
structure
- Dot-product similarity (by itself) can't fully capture structural
similarity
## Systematic substitution via binding
A binding can be used as a partial function for substitution\
The substitution is applied uniformly across the components of the
argument
With a commutative, self-inverse product operator, e.g. BSC, MAP\
(I won't discuss non-commutative or non-self-inverse products here):\
$A \otimes X \equiv \{ A \mapsto X, X \mapsto A \}$
Apply the substitution by binding it with the argument:\
$\mathbf{(A \otimes X)} \otimes (A + A \otimes B + X \otimes C + D)$\
$= A \otimes X \otimes A + A \otimes X \otimes A \otimes B + A \otimes X \otimes X \otimes C + A \otimes X \otimes D$\
$= (A \otimes A) \otimes X + (A \otimes A) \otimes X \otimes B + A \otimes (X \otimes X) \otimes C + A \otimes X \otimes D$\
$= \textbf{1} \otimes X + \textbf{1} \otimes X \otimes B + A \otimes \textbf{1} \otimes C + A \otimes X \otimes D$\
$= X + X \otimes B + A \otimes C + A \otimes X \otimes D$
$(A + A \otimes B + X \otimes C + D) \overset{(A \otimes X)}{\mapsto} (X + X \otimes B + A \otimes C + \mathbf{A \otimes X \otimes D})$
## Subtlety of binding
<font size="5">
Multiple interpretations of bindings ($A \otimes X)$ depending on how
they are used, e.g.:
- key:value pair in a dictionary (note key and value are treated
identically)
- variable:value pair (note variable and value are treated
identically)
- Virtual feature detector neuron ($A$ = pattern detected, $X$ =
identity of neuron)
- Inference rule ($A$ = antecedent, $B$ = consequent)
- Substitution pattern ($A$ = search pattern, $B$ = replacement
pattern)
$A$ and $X$ can be arbitrarily complex composites (everything is just a
vector), e.g.:
- $(A + B + C) \otimes X = A \otimes X + B \otimes X + C \otimes X$
- $A \otimes (X + Y + Z) = A \otimes X + A \otimes Y + A \otimes Z$
- $A \otimes B \otimes X \otimes Y = A \otimes (B \otimes X \otimes Y) = (A \otimes X \otimes Y) \otimes B = \ldots$
</font>
## Kanerva (2010) - What is the Dollar of Mexico?
$V_{ustates} = V_{name} \otimes V_{USA} + V_{capital} \otimes V_{WDC} + V_{currency} \otimes V_{USD}$\
$V_{mexico} = V_{name} \otimes V_{MEX} + V_{capital} \otimes V_{MXC} + V_{currency} \otimes V_{MXN}$\
$V_{U \otimes M} = V_{ustates} \otimes V_{mexico}$\
$= V_{USA} \otimes V_{MEX} + V_{WDC} \otimes V_{MXC} + V_{USD} \otimes V_{MXN} + noise_1$\
(a sum of filler substitutions - fillers occupying the same role have
been bound)
Warning: $noise$ is the sum of all terms you *know* will not be
important
Apply the USA/Mexico mapping, e.g. "What is the Dollar of Mexico?"\
$V_{U \otimes M} \otimes V_{USD}$\
$= (V_{USA} \otimes V_{MEX} + V_{WDC} \otimes V_{MXC} + V_{USD} \otimes V_{MXN} + noise_1) \otimes V_{USD}$\
$= noise_2 + noise_3 + V_{USD} \otimes V_{MXN} \otimes V_{USD} + noise_1 \otimes V_{USD}$\
$\approx V_{MXN}$\
## Hand crafted substitution
This approach works because of identical roles, e.g. $V_{currency}$
The role representations are static and enumerated in advance
Great if that's all you need
- Seriously, if that's all you need, it's the best way to do something
analogy-like
*ANALOGY* needs dynamic substitutions chosen in response to the context
## Emruli et al (2013) - Learned substitution
![<font size="1">Emruli et al
(2013)</font>](assets/emruli-AMU.png){width="70%"}
"The analogical mapping unit (AMU) which learns mappings of the type
$x_k \mapsto y_k$ from examples and uses bundled mapping vectors stored
in the SDM to calculate the output vector $y_k$" Emruli et al (2013)
## How it works
$x_k$ is used as the address for the Sparse Distributed Memory (SDM)\
The mapping $x_k \otimes y_k$ is used as the value to store in the SDM
Mappings are noncommutative because $x_k$ and $y_k$ are used differently
Write mode: mappings are written to SDM\
Read mode: retrieves average mapping corresponding to $x_k$ and applies
it to $x_k$
SDM does a sort of averaging memory over similar addresses\
Interpolates over mappings\
Can't generate completely novel mappings
Note the "circuit based" approach, including non-VSA components (SDM)\
There is usually an amount of plumbing and control to deal with\
A purist would make the control distributed (VSA-like) but it's usual to
make the control localist as an engineering hack
## Gayler & Levy (2009) - Settling on substitution
Graph isomorphism (not analogical mapping, but a proxy for necessary
process)
Find vertex mappings that make the graphs identical
- $\{ A \mapsto P, B \mapsto Q, C \mapsto S, D \mapsto R \}$
- $\{ A \mapsto P, B \mapsto Q, C \mapsto R, D \mapsto S \}$
![<font size="1">Gayler & Levy
(2009)</font>](assets/gayler-2_graphs.png){width="70%"}
## How it works: The long and winding road
The explanation is going to be long winded (sorry)
- What is a graph isomorphism (implementable definition)?
- Localist heuristic to find graph isomorphisms
- VSA distributed implementation of localist method
## Adjacency matrix of a graph
How to represent a graph with a matrix
- Row and column indices correspond to vertices
- Cell entries indicate edges
![<font size="1">Gayler & Levy
(2009)</font>](assets/gayler-adjacency.png){width="60%"}
## Association graph
<font size="5">
The association graph is a graph product of the two graphs\
- Vertices correspond to vertex mappings of the two graphs\
- Edges correspond to edge existence agreement of the vertex mappings\
- A maximal clique corresponds to a maximal isomorphism of the two
graphs
</font>
![<font size="1">Gayler (2009) Melbourne University presentation
(edited). Only a subset of vertices and edges
shown.</font>](assets/gayler-association2.png){width="59%"}
## How to find a maximal clique of a graph
Replicator equations (from evolutionary game theory)\
Also interpretable as Bayesian update\
$x(t)$ = prior distribution = support for each possible vertex mapping\
$x(t + 1)$ = posterior distribution = support for each possible vertex
mapping\
$w$ = adjacency matrix of association graph\
$\pi(t)$ = likelihood = multiplicative update to vertex mapping support
given $w$
![<font size="1">Gayler (2009) Melbourne University
presentation</font>](assets/gayler-replicator_eqns.png){width="20%"}
## Replicator equation circuit
Localist representation of mappings (potentially large)
$k$ number of vertices in each of the two graphs\
$dim(x) = dim(\pi) = k^2$\
$dim(w) = k^2 \times k^2$
![<font size="1">Gayler (2009) Melbourne University
presentation</font>](assets/gayler-circuit_localist.png){width="35%"}
$\land \triangleq$ elementwise product
## Settling of localist replicator equations
![<font size="1">Gayler (2009) Melbourne University
presentation</font>](assets/gayler-settling_localist.png){width="55%"}
## VSA representations for replicator equations
<font size="5">
Vertices: $A, B, C, D, P, Q, R, S$
Edges: $B \otimes C,B \otimes D, Q \otimes R, Q \otimes S$
Graph vertex sets: $(A + B + C + D), (P + Q + R + S)$
Graph edge sets:
$(B \otimes C + B \otimes D), (Q \otimes R + Q \otimes S)$
Initial potential vertex mappings = $x(t = 1)$ =\
$(A + B + C + D) \otimes (P + Q + R + S)$ =\
$A \otimes P + A \otimes Q + \ldots + B \otimes P + B \otimes Q + \ldots + D \otimes S$
Association graph edges (positive only) = potential edge mappings = $w$
=\
$(B \otimes C + B \otimes D) \otimes (Q \otimes R + Q \otimes S)$ =\
$(B \otimes C \otimes Q \otimes R) + (B \otimes C \otimes Q \otimes S) + (B \otimes D \otimes Q \otimes R) + (B \otimes D \otimes Q \otimes S)$
</font>
## VSA replicator equation circuit
Interpret all vectors as being of the form $kV$,\
where $k$ is the magnitude/support for $V$ (the unit magnitude
direction)
Analog computing: $V$ is the labelled wire, $k$ is the voltage on the
wire
![<font size="1">Gayler (2009) Melbourne University
presentation</font>](assets/gayler-circuit_distributed.png){width="70%"}
## Multiset intersection
$\land \triangleq$ multiset intersection
A multiset is a set with a nonnegative magnitude of membership for each
element (i.e. the magnitudes of the component vectors vary across
components)
$arg_1 = a V_1 + b V_2 + c V_3$\
$arg_1 = p V_1 + q V_2 + r V_4$\
$\land(arg_1, arg_2) = a p V_1 + b q V_2$
The elementwise multiplication of the magnitudes of the component
vectors
This corresponds to the elementwise multiplication of support for the
vertex mappings in the localist version
Won't go into the implementation here
## Evidence propagation
<font size="5">
Association graph edges have the form:
$B \otimes C \otimes Q \otimes R$\
and can be interpreted as mappings:\
$(B \otimes C) \otimes (Q \otimes R)$ \# interpret as mapping between
edges in the graphs\
$(B \otimes Q) \otimes (C \otimes R)$ \# interpret as mapping between
vertex mappings
Association graph edges applied as inference rules:\
$\pi = x \otimes w$\
= $(B \otimes Q + \ldots ) \otimes (B \otimes C \otimes Q \otimes R + \ldots )$\
= (vertex mappings) $\otimes$ (mappings between vertex mappings)\
= $(k(B \otimes Q) + \ldots ) \otimes ((B \otimes Q) \otimes (C \otimes R) + \ldots )$\
= $k(C \otimes R) + \ldots$
Interpret $(B \otimes Q) \otimes (C \otimes R)$ as the rule:\
"To the extent $k$ that $B \otimes Q$ is supported as part of the solution\
Increase the support for $C \otimes R$ as part of the solution by $k$"
</font>
## Settling of VSA replicator equation circuit
![<font size="1">Gayler (2009) Melbourne University
presentation</font>](assets/gayler-settling_distributed.png){width="63%"}
## References / Reading
M. Blokpoel, T. Wareham, P. Haselager, and I. van Rooij (2018) *Deep
Analogical Inference as the Origin of Hypotheses*. The Journal of
Problem Solving
D. J. Chalmers, R. M. French, and D. R. Hofstadter (1992) *High-level
perception, representation, and analogy: A critique of artificial
intelligence methodology*. Journal of Experimental & Theoretical
Artificial Intelligence
B. Emruli, R. W. Gayler, and F. Sandin (2013) *Analogical mapping and
inference with binary spatter codes and sparse distributed memory*. The
2013 International Joint Conference on Neural Networks (IJCNN)
B. Emruli and F. Sandin (2014) Analogical *Mapping with Sparse
Distributed Memory: A Simple Model that Learns to Generalize from
Examples*. Cognitive Computation
------------------------------------------------------------------------
R. W. Gayler and S. D. Levy (2009) *A Distributed Basis for Analogical
Mapping*. New Frontiers in Analogy Research, Proceedings of the Second
International Conference on Analogy, ANALOGY-2009
R. W. Gayler and R. Wales (1998) *Connections, Binding, Unification and
Analogical Promiscuity*. Advances In Analogy Research: Integration Of
Theory And Data From The Cognitive, Computational, And Neural Sciences
H. Gust, U. Krumnack, K.-U. Kühnberger, and A. Schwering (2008)
*Analogical Reasoning: A Core of Cognition*. KI - Künstliche Intelligenz
D. R. Hofstadter (2006) *Analogy as the core of cognition*. Stanford
Presidential Lecture
P. Kanerva (2000) *Large Patterns Make Great Symbols: An Example of
Learning from Example*. Hybrid Neural Systems
------------------------------------------------------------------------
P. Kanerva (2010) *What We Mean when We Say "What's the Dollar of
Mexico?": Prototypes and Mapping in Concept Space*. Quantum Informatics
2010: AAAI-Fall 2010 Symposium on Quantum Informatics for Cognitive,
Social, and Semantic Processes
S. D. Levy and R. W. Gayler (2009) *"Lateral inhibition" in a fully
distributed connectionist architecture*. In Proceedings of the Ninth
International Conference on Cognitive Modeling (ICCM 2009)
T. Mikolov, W. Yih, and G. Zweig (2013) *Linguistic Regularities in
Continuous Space Word Representations*. Proceedings of NAACL-HLT 2013
T. A. Plate (1994) *Distributed Representations and Nested Compositional
Structure*. PhD Thesis. University of Toronto
T. A. Plate (2000) *Analogy retrieval and processing with distributed
vector representations*. Expert Systems
------------------------------------------------------------------------
A. Rogers, A. Drozd, and L. Bofang (2017) *The (too Many) Problems of
Analogical Reasoning with Word Vectors*. Proceedings of the 6th Joint
Conference on Lexical and Computational Semantics (\*SEM 2017)
------------------------------------------------------------------------
[DOI: 10.5281/zenodo.5552219](http://doi.org/10.5281/zenodo.5552219)
<font size="2">This presentation is licensed under a Creative Commons
Attribution 4.0 International License</font>
![<font size="2"><https://creativecommons.org/licenses/by/4.0/></font>](assets/cc_by.png){width="28%"}