-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathkhoei17fle.tex
2609 lines (2488 loc) · 148 KB
/
khoei17fle.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
%!TEX TS-program = PdfLaTeX
%!TEX encoding = UTF-8 Unicode
%!TEX spellcheck = en-US
%!BIB TS-program = bibtex
\def\Draft{0}% draft=1 or no draft = 0
\def\Pace{0}% draft=1 or no draft = 0
% -------definitions-----
\newcommand{\AuthorA}{Khoei}%
\newcommand{\AuthorB}{Masson}%
\newcommand{\AuthorC}{Perrinet}%
\newcommand{\FirstNameA}{Mina A.}%
\newcommand{\FirstNameB}{Guillaume S.}%
\newcommand{\FirstNameC}{Laurent U.}%
\newcommand{\Institute}{Institut de Neurosciences de la Timone, UMR7289, CNRS / Aix-Marseille Universit\'e}%
\newcommand{\Address}{27, Bd. Jean Moulin, 13385 Marseille Cedex 5, France}%
\newcommand{\Website}{http://invibe.net/LaurentPerrinet}%
\newcommand{\EmailA}{m.khoiee@gmail.com}%Mina.aliakbari-khoei@univ-amu.fr}%
\newcommand{\EmailC}{Laurent.Perrinet@univ-amu.fr}%
\newcommand{\EmailB}{Guillaume.Masson@univ-amu.fr}%
\newcommand{\Title}{The flash-lag effect as a motion-based predictive shift}%
\newcommand{\Abstract}{%
Due to its inherent neural delays,
the visual system has an outdated access to sensory information
about the current position of moving objects.
In contrast, living organisms are remarkably able to track and intercept moving objects
under a large range of challenging environmental conditions.
Physiological, behavioral and psychophysical evidences strongly suggest
that position coding is extrapolated using an explicit and
reliable representation of object's motion
but it is still unclear how these two representations interact.
For instance, the so-called flash-lag effect supports the idea of
a differential processing of position between moving and static objects.
Although elucidating such mechanisms is crucial in our understanding of
the dynamics of visual processing,
a theory is still missing to explain the different facets of this visual illusion.
Here, we reconsider several of the key aspects of the flash-lag effect in order
to explore the role of motion upon neural coding of objects' position.
First, we formalize the problem using a Bayesian modeling framework
which includes a graded representation of the degree of belief about visual motion.
We introduce a motion-based prediction model
as a candidate explanation for the perception of coherent motion.
By including the knowledge of a fixed delay,
we can model the dynamics of sensory information integration
by extrapolating the information acquired at previous instants in time.
Next, we simulate the optimal estimation of object position
with and without delay compensation and
compared it with human perception
under a broad range of different psychophysical conditions.
Our computational study suggests that the explicit, probabilistic representation
of velocity information is crucial in explaining position coding,
and therefore the flash-lag effect.
We discuss these theoretical results in light of the putative corrective mechanisms
that can be used to cancel out the detrimental effects of neural delays and
illuminate the more general question of the dynamical representation
of spatial information
at the present time
in the visual pathways.
}%
\newcommand{\AuthorSummary}{%
Visual illusions are powerful tools to explore
the limits and constraints of human perception.
One of them has received considerable empirical and theoretical interests:
the so-called ``flash-lag effect''.
When a visual stimulus moves along a continuous trajectory,
it may be seen \emph{ahead} of its veridical position
with respect to an unpredictable event such as a punctuate flash.
This illusion tells us something important
about the visual system: contrary to classical computers,
neural activity travels at a relatively slow speed.
It is largely accepted that the resulting delays cause this perceived spatial lag of the flash.
Still, after three decades of debates, there is no consensus
regarding the underlying mechanisms.
Herein, we re-examine the original hypothesis that this effect may be caused by the extrapolation of the stimulus' motion that is
naturally generated in order to compensate for neural delays.
Contrary to classical models,
we propose a novel theoretical framework, called parodiction,
that optimizes this process by explicitly using
the precision of both sensory and predicted motion.
Using numerical simulations, we show that the parodiction theory subsumes
many of the previously proposed models and empirical studies.
More generally, the parodiction hypothesis proposes
that neural systems implement generic neural computations
that can systematically compensate the existing neural delays
in order to represent the predicted visual scene at the present time.
It calls for new experimental approaches to directly explore
the relationships between neural delays and predictive coding.
}%
\newcommand{\CopyRight}{
We confirm that Fig 1 contains three items from different sources which are compatible with your Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) with explicit permissions (CC-0 license):
- ball: see https://openclipart.org/detail/9618/football-soccer
- player : see https://pixabay.com/en/soccer-football-player-sport-307188/
- running player : see https://pixabay.com/en/soccer-football-playing-running-306925/.
}
\newcommand{\Keywords}{predictive coding, motion coherency, flash-lag effect, neural delays, diagonal model, motion extrapolation, probabilistic models}%
\newcommand{\Acknowledgments}{
LUP would like to thank Rick Adams and Karl Friston for fruitful discussions
and the Wellcome Trust for NeuroImaging for supporting this collaboration.
}\newcommand{\Funding}{
MK was funded by the FACETS-ITN Marie Curie Training Network
of the European Union (FP7-PEOPLE-ITN-2008-237955).
LUP and GSM were supported by the European Union (BrainScaleS, FP7-FET-2010-269921), the Agence Nationale de la Recherche (Speed, ANR-13-SHS2-0006) and CNRS.
} %
%--------------------------
%\documentclass[pdftex,a4paper,english]{article} % For LaTeX2e
%\usepackage{times}
\documentclass[10pt,letterpaper,english]{article}
\usepackage[top=0.85in,left=2.75in,footskip=0.75in]{geometry}
%% Use adjustwidth environment to exceed column width (see example table in text)
\usepackage{changepage}
%
%% Use Unicode characters when possible
\usepackage[utf8]{inputenc}
%
%% textcomp package and marvosym package for additional characters
\usepackage{textcomp,marvosym}
%
\usepackage{babel}
%\usepackage[protrusion=true,expansion=true]{microtype} % Better typography
% fixltx2e package for \textsubscript
%\usepackage{fixltx2e}
% amsmath and amssymb packages, useful for mathematical formulas and symbols
\usepackage{amsmath,amssymb}
%\usepackage{amsmath} % math fonts
%\usepackage{amsthm}
%\usepackage{amsfonts}
%
%% cite package, to clean up citations in the main text. Do not remove.
%\usepackage{cite}
%
%% Use nameref to cite supporting information files (see Supporting Information section for more info)
\usepackage{nameref}
%
%
%% ligatures disabled
\usepackage{microtype}
\DisableLigatures[f]{encoding = *, family = * }
%
%% rotating package for sideways tables
\usepackage{rotating}
\if 1\Draft
%% line numbers
\usepackage[right]{lineno}
\usepackage{setspace}
\doublespacing
%\usepackage{lineno}
\fi
% Remove comment for double spacing
%\usepackage{setspace}
%
% Text layout
\raggedright
\setlength{\parindent}{0.5cm}
\textwidth 5.25in
\textheight 8.75in
%
%
%% Bold the 'Figure #' in the caption and separate it from the title/caption with a period
%% Captions will be left justified
\usepackage[aboveskip=1pt,labelfont=bf,
labelsep=period,justification=raggedright,singlelinecheck=off]{caption}
% Remove brackets from numbering in List of References
\makeatletter
\renewcommand{\@biblabel}[1]{\quad#1.}
\makeatother
% Leave date blank
\date{}
%% Header and Footer with logo
\usepackage{graphicx}
% Enable pdflatex
\if 0\Pace
\graphicspath{{./figures/}}%
\DeclareGraphicsExtensions{.pdf}
\else
\graphicspath{{./figures/},{./}}%
\DeclareGraphicsExtensions{.eps,.tiff}
\fi
\usepackage{lastpage}
\usepackage{fancyhdr}
\usepackage{epstopdf}
\pagestyle{myheadings}
\pagestyle{fancy}
\fancyhf{}
\lhead{\includegraphics[width=2.0in]{PLOS-submission}}
\rfoot{\thepage/\pageref{LastPage}}
\renewcommand{\footrule}{\hrule height 2pt \vspace{2mm}}
\fancyheadoffset[L]{2.25in}
\fancyfootoffset[L]{2.25in}
\lfoot{\sf PLOS}
\usepackage{csquotes}
\usepackage{units}%
%\usepackage{float}
%\usepackage{placeins}
\usepackage{siunitx}
%\renewcommand{\cite}{\citep}%
\newcommand{\ms}{\si{\milli\second}}%
\newcommand{\m}{\si{\meter}}%
\newcommand{\s}{\si{\second}}%
\newcommand{\NN}{\mathbb{N}}
\newcommand{\RR}{\mathbb{R}}
\newcommand{\ZZ}{\mathbb{Z}}
%\usepackage[natbib=true,sortcites=true,backend=biber,doi=false,isbn=false,url=false]{biblatex}%block=space,
%\addbibresource{thesis.bib}
\usepackage[comma,compress,numbers,square]{natbib}%sort&
\newcommand{\citeA}{\citeauthor}%
\renewcommand{\cite}{\citep}%
%
%\graphicspath{{../../commons/},{./figures/}}%
%\DeclareGraphicsExtensions{.png,.pdf}%
\usepackage[pdfusetitle,colorlinks=false,pdfborder={0 0 0}]{hyperref}%
% \usepackage{authblk}%
\makeatletter
% http://tex.stackexchange.com/questions/43648/why-doesnt-lineno-number-a-paragraph-when-it-is-followed-by-an-align-equation
% Make a copy of macros responsible for entering display math mode
\let\start@align@nopar\start@align
\let\start@gather@nopar\start@gather
\let\start@multline@nopar\start@multline
% Add the "empty line" command to the macros
\long\def\start@align{\par\start@align@nopar}
\long\def\start@gather{\par\start@gather@nopar}
\long\def\start@multline{\par\start@multline@nopar}
\makeatother
\listfiles
\if 1\Draft
%\usepackage[nomarkers,tablesonly]{endfloat}
\usepackage[nolists]{endfloat}
\fi
\begin{document}%
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
\if 1\Draft
%\linenumbers
\doublespacing
\fi
\vspace*{0.35in}
% Title must be 250 characters or less.
% Please capitalize all terms in the title except conjunctions, prepositions, and articles.
\begin{flushleft}
{\Large
\textbf\newline{\Title}
}
\newline
% Insert author names, affiliations and corresponding author email (do not include titles, positions, or degrees).
\\
\FirstNameA\ \AuthorA , % \textsuperscript{1},
\FirstNameB\ \AuthorB , %\textsuperscript{1},
\FirstNameC\ \AuthorC \textsuperscript{*}
\\
\bigskip
%\bf{1}
\Institute\\\Address
%\\
%\bf{2} Affiliation Dept/Program/Center, Institution Name, City, State, Country
%\\
%\bf{3} Affiliation Dept/Program/Center, Institution Name, City, State, Country
\\
\bigskip
% Use the asterisk to denote corresponding authorship and provide email address in note below.
* \EmailC
\end{flushleft}
\if 1\Draft
\linenumbers
\fi
\section*{Abstract}
\Abstract
% Please keep the Author Summary between 150 and 200 words
% Use first person. PLOS ONE authors please skip this step.
% Author Summary not valid for PLOS ONE submissions.
\section*{Author Summary}
\AuthorSummary
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
\section{Introduction}
\subsection{Neural delays and motion-induced position shifts}%
Though it is barely noticeable in everyday life,
visual signals captured on the retina take a significant amount of time
before they can elicit even the simplest actions such as eye movements.
This neural delay is composed of two terms: a fixed delay caused
by the axonal transfer of sensory signals up to motor effectors
and a variable delay associated with the neural processing time occurring
at each computational step.
Moreover, different neural systems can lead to different delays,
even for the simplest feed-forward sensorimotor transformations
where most of the computational load occurs at sensory level.
Just to mention, a delay of $90~\ms$ is observed between
the onset of retinal image motion and
the first acceleration of tracking eye movements
in humans~\citep{Lisberger2010, Masson12, Montagnini15bicv}
while the exact same sensorimotor transformation
takes less than $60~\ms$ in monkeys~\citep{Masson12}.
Furthermore, increasing signal uncertainty
would further increase these delays~\citep{Masson12} illustrating the fact
that neural delays also vary with many environmental or contextual factors.
A mere consequence of these unavoidable neural delays should be
that we perceive sensory events
with a slight, but permanent lag~\citep{Nijhawan94,Inui06}.
This is well illustrated in a position estimation task such as the one faced
by a soccer referee. If a ball is shot at an unexpected instant by one fixed player,
in the direction of another running player,
he will generally perceive the moving player ``ahead'' of its actual position~\citep{Baldo02}
and signal an off-side position despite the fact that the players' physical positions
were strictly aligned to that of the referee (see Fig~\ref{fig:FLE_cartoon}).
As a general rule, if no mechanism would intervene to compensate for such neural delays, one would expect
severe inefficiencies in sensory computations as well as in goal-directed action control.
On the contrary, there are ample evidences
that animals can in fact cope with neural delays
in order to plan and execute timely goal-directed actions.
Thus, it seems evident that throughout natural evolution,
some sophisticated compensatory mechanisms based
on internal models have been selected~\citep{Franklin11}.
Thus, studying neural delays and how they may be compensated
is a critical question that needs to be resolved in order to decipher how basic neural computations
such as the dynamical processing of sensory information can be efficiently performed
(for a review, see~\citep{PerrinetAdamsFriston14}). Solving this enigma would have several theoretical
consequences such as, in particular, understanding
how neural activity can encode both space and time~\citep{Nijhawan10}.
% A platform to study neural delays: motion-induced position shifts
Although these neural delays are usually rather short,
they can easily be unveiled by psychophysical experiments.
This Flash-lag effect (FLE) is a well-studied perceptual illusion
which is intimately linked with the existence of neural delays~\citep{MacKay58}.
In a standard empirical variant of the FLE,
a first stimulus moves continuously along the central horizontal axis of the screen display.
At the time this moving stimulus reaches the center of the screen,
a second stimulus is flashed in its near vicinity but in perfect vertical alignment with it.
Despite the fact that horizontal positions of the two stimuli
are physically identical at the time of the flash, the moving stimulus is most often perceived \emph{ahead}
of the flashed one (see the square stimulus in Fig~\ref{fig:FLE_cartoon}).
The flash-lag effect falls in the vast category of motion-induced position shifts
(e.g. the Fröhlich effect or
the representational momentum effect~\citep{Musseler02, Eagleman07, Jancke09}),
in which the perceived position of an object is biased
by its own visual motion or by other motion signals from its visual surrounding.
How can we relate the FLE with the existence of the aforementioned neural delays?
Several experimental studies have suggested
that this visual illusion unveils predictive mechanisms
that could compensate for the existing neural delays by extrapolating
the object's motion~\citep{Nijhawan94, Berry99, Jancke04, Jancke09}.
Since in natural scenes smooth trajectories are more probable than jittered ones,
an internal representation may dynamically integrate information
along the trajectory in order to predict the most expected position
of the stimulus forward in time,
\emph{knowing} an average estimate of the different neural delays.
Though computationally simple,
this algorithmic solution requires that neural computations can build and
use an internal representation of position and velocity over time,
that is, that they can manipulate the dynamic representation of a variable.
% the method we propose
%<*ParoDiction>
The aim of our theoretical work is to better understand the interactions %
between position and motion coding that are based on predictive mechanisms and that could be %
implemented within the early visual system. %
To do so, we introduce a generic probabilistic model that was previously shown to efficiently solve %
other classical problems in sensory processing %
such as the aperture problem and motion extrapolation~\citep{Perrinet12pred, Khoei13}. %
This computational framework allows to quantify the relative efficiency %
of these different coding mechanisms and to explain the main empirical psychophysical observations. %
We propose a novel solution for introducing neural delays in the dynamics %
of probabilistic inference and discuss how this approach is related %
to previous models of motion diffusion and position coding. %
Taking the Flash-lag effect as a well-documented illustration of the generic problem %
of computing with delays, we show that our model can coalesce most %
of the cardinal perceptual aspects of FLE and thus, %
unite the previous models described below. %
More generally, such generic computational principles could be shared %
by other sensory modalities facing similar delays. %
%</ParoDiction>
%------------------------------------------------------------------------------------------------%
%: fig:FLE_cartoon
\begin{figure}%
\begin{center}
\if 0\Draft
\if 0\Pace
\includegraphics[width=\textwidth]{FLE_cartoon}
\else
\includegraphics[width=\textwidth]{Fig1}
\fi
\else
Figure 1 around here
\fi
\end{center}
\caption{
\textbf{The flash-lag effect (FLE) as a motion-induced predictive shift} :
%<*FleCartoon>
To follow the example given by~\citep{Baldo02}, a football (soccer) player that would run %
along a continuous path %
(the green path, where the gradient of color denotes the flow of time) %
is perceived to be ahead (the red position) of its actual position %
at the unexpected moment a ball is shot (red star) %
even if these positions are physically aligned. %
A referee would then signal an ``offside'' position. %
Similarly, such a flash-lag effect (FLE) is observed systematically in psychophysical experiments %
by showing a moving and a flashed stimuli (here, a square). %
By varying their characteristics (speed, relative position), %
one can explore the fundamental principles of the FLE. %
%</FleCartoon>
\label{fig:FLE_cartoon}}%
\end{figure}
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
\subsection{A brief overview of the Flash-lag effect}
\label{sec:intro}
% The basis of FLE
The Flash-lag effect was first described by~\citet{Metzger32}
and subsequently investigated by~\citet{MacKay58}.
After these early studies, the phenomenon did not attract much attention
until~\citeA{Nijhawan94} begun to study a similar question.
In his empirical approach, a moving and a static (flashed) stimuli
are presented with a perfect spatial and temporal alignment at the time of the flash but
most subjects perceive the moving object as leading in space (see Fig~\ref{fig:FLE_cartoon}).
Such perceptual effect was reproduced in other species, in particular in monkeys~\citep{Subramaniyan2013}.
Motion extrapolation is the correction of the object's position
based on an estimate of its own motion over the time period introduced by neural delays.
\citeA{Nijhawan94} proposed that such motion extrapolation
can explain this perceived spatial offset between the two stimuli.
In this theoretical framework, the visual system is predictive and takes advantage
of the available information about object's motion
in order to correct for the positional error caused by neural delays.
%the problems of flash-initiated and flash-terminated cycles
The seminal work of~\citeA{Nijhawan94} resurrected the interest for the FLE phenomenon.
Since then, the perceptual mechanisms underlying the FLE have been extensively explored
by the group of Nijhawan~\citep{Nijhawan04, Nijhawan09, Maus10b} and
others~\citep{Whitney98, Whitney00-1, Krekelberg_science, Schlag00, Eagleman00}
(for a review see~\citep{Nijhawan02, Hubbard14}). Different variants of the original experiment
were designed in order to challenge the different motion extrapolation models.
These studies revealed a flaw in \citeA{Nijhawan94}'s motion extrapolation theory
since it cannot account for the experimental observations made
with two specific variants of the FLE, often called half-cycle FLEs~\citep{Nijhawan02}.
Their common principle is to manipulate the position of the flash
relative to the trajectory of the moving object.
While in the standard FLE, the flash appears in the middle of the motion path,
the flash can now appear either at the beginning or at the end of the motion trajectory,
thus defining the flash-initiated and flash-terminated cycle FLEs, respectively.
The motion extrapolation hypothesis predicts that, at the beginning of the trajectory,
the flashed and moving objects are not likely to be differentiated. However, this prediction was contradicted by the psychophysical results
showing a comparable position shift in both the flash-initiated cycle and the standard FLE.
Furthermore, extrapolating a trajectory should impose an inertial component
even in the presence of sudden changes in the visual motion properties,
such as motion termination or reversal.
By consequence, the motion extrapolation hypothesis predicts
a perceptual overshoot that is similar in both flash-terminated and standard FLE.
Again, this prediction was contradicted by psychophysical evidence
demonstrating a lack of position shift
in the flash-terminated cycle FLE~\citep{Eagleman00}.
Lastly, several studies suggested that the motion extrapolation hypothesis
needs to be supplemented with complementary mechanisms
such as the a posteriori correction of the predicted position,
in order to account for the perceived position
after an abrupt change in the motion trajectory~\citep{Maus06, Nijhawan08, Maus09}.
% alternatives models: latency, persistence and postdiction
These new empirical evidences called for alternative hypotheses able to
unify all of these different aspects of FLE.
A first set of studies proposed that moving and static objects
are processed with different latencies in the early visual system. Hence,
the perceived lag in FLE could be explained by the faster processing of moving objects,
as compared to flashed inputs~\citep{Whitney98, Purushothaman98, Whitney00-1,Jancke04}.
There may exist multiple origins at retinal and cortical levels
for a more rapid processing of moving objects. Some authors reasoned that, since both flashed and moving stimuli are processed and
transmitted within a single (magno-cellular) retino-thalamo-cortical pathway,
any difference would be explained by intra-cortical mechanisms that would process
differently predictable and unpredictable events~\citep{Jancke04}.
However, there is still a lack of solid neurophysiological empirical evidences in support of this differential latency hypothesis.
A second hypothesis suggested that the FLE may be explained by the position persistence for the flashed visual input~\citep{Krekelberg01, Krekelberg00-1}.
The central idea is that motion information is averaged within a $500~\ms$ window. By consequence, the perceived position of the flash would persist,
while the averaged position for the moving object is perceived ahead of its actual position, along its motion path. The main flaw of this hypothesis is that
the supposed time constant ($500~\ms$) is unrealistically long with respect to the known dynamics of motion integration.
%<*WojtachAnalogies>
More recently,~\citet{Wojtach08} proposed that the FLE %
may be seen as a mere consequence of the distribution of speeds %
that are empirically observed during our visual experience. %
Using the perspective transform from the three-dimensional physical space %
to the two-dimensional plane of the retinotopic space, %
they assigned empirical probabilities of the observed retinal speeds %
from the mapping of objects' velocities in the physical world. %
By doing so, they defined an \emph{a priori} probability distribution of speeds %
which can be combined with sensory evidence. %
This solution proposes a probabilistic framework inferring an optical estimate of motion speed. %
Such estimate is then used in a motion extrapolation model %
compensating for the neural delay. %
The authors estimated the amplitude of the lag %
in respect to an extended speed range of object motion. %
Their model depicts a nonlinear relationship %
between motion speed and the perceptual lag, %
similar to the one observed with the standard flash-lag experiment. %
Thus, the model from~\citet{Wojtach08} provides an ingenious extension %
of the motion extrapolation model using inferred speed. %
However, this model was not tested against the aforementioned complete, and challenging set %
of empirical studies probing the FLE at different epochs of the motion trajectory. %
%</WojtachAnalogies>
One last approach is the \emph{postdiction} hypothesis~\citep{Eagleman00}
postulating that visual awareness attributes the position of a moving stimuli
at the instant of the flash appearance according to the information collected
within an $\approx 80 \ms$ time window following the flash.
In particular, the flash is considered as a reset for motion integration and,
as such, this would be sufficient in explaining why the FLE
is not perceived in the flash-terminated cycle.
The postdiction hypothesis relies on two main assumptions. First,
both moving and flashed inputs have the same neural delays. Second,
the flash acts as a resetting mechanism.
By consequence, the model predicts that observers shall perceive both
a spatial lag of the flash and a change in the speed of the moving object.
However, such a speed increment has never been reported in the context of FLE~\citep{Nijhawan02}.
The postdiction model is thus an elegant hypothesis that allows us
to understand a wide range of variants of the FLE but fails
to explain this later aspect of FLE.
In summary, the half-cycles variants of the FLE introduced by~\citet{Eagleman00}
remain challenging for all current theoretical approaches of the FLE,
despite the fact that they might reveal how the visual system processes
motion onset and offset and their impact on position coding.
%###############################################################################
\subsection{The parodiction hypothesis}
%###############################################################################
Overall, previous theoretical studies can be grouped according %
to two main hypotheses. %
On one hand, models based on latency difference or motion extrapolation rely %
on how the neural representation of position information is encoded. %
On the other hand, the postdiction hypothesis is based on %
how visual awareness decodes objects' positions from neural activity in a short temporal window. %
In the present theoretical study, we will propose a new hypothesis %
which subsumes both of these aspects. %
Our theoretical approach is based upon two major constraints faced by any neural system, %
in comparison to a conventional computer, %
when estimating the position of an object: %
First, there is no access to a central clock, that is, %
the present, exteroceptive, physical timing is hidden %
(or latent in machine learning terms) to the nervous system. %
Second, the internal representation encoded in the neural activity is distributed and dynamical. %
In particular, the system is confronted to non-stationary sources of noises and %
has to provide for an optimal estimate at any time for upstream neural networks. %
\\
Driven by these constraints, %
a biologically-realistic hypothesis is that a perceived position %
corresponds to the most likely position at the present time~\citep{Changizi01}. %
According to the probabilistic brain hypothesis %
(see~\citep{PerrinetAdamsFriston14} for a generic application to eye movements), %
an optimal solution is that the internal representation encodes beliefs %
in the form of probability distribution functions (pdf) and %
that the optimal estimate is computed knowing both the instantaneous sensory data %
and the internal representation. %
When the represented variable, such as the position, is predictable, %
this process involves that the internal representation %
uses a generative model of its dynamics to progressively refine the estimation. %
As a result, using a probabilistic formulation of predictive coding, %
it is possible to explicitly represent the instantaneous information %
about object's motion and its precision, %
coherently with the role played by perceptual precision in the FLE~\citep{Kanai04}. %
Consequently, we propose that a generic goal of these neural computations is %
to optimally align the position represented in the neural activity %
with that at the veridical, but hidden, physical time. %
We will call this approach the \emph{parodiction} hypothesis, %
from the ancient Greek $\pi\alpha\rho{\acute o}\nu$, the present time. %
\\
Herein, we will show that probabilistic motion extrapolation %
can efficiently compensate for the neural delays and %
explain the shift in perceived positions in the different variants of the FLE. %
The paper is organized as follows. %
First, we will define the probabilistic motion extrapolation model %
and we will describe how delays can be taken into account. %
This model extends a simple motion-based predictive model %
based on the temporal coherency of visual trajectories that we proposed earlier~\citep{Perrinet12pred} %
and is also a novel formulation of the original motion extrapolation model proposed by Nijhawan~\citep{Nijhawan94}. %
Second, we present the results of this model with the standard FLE %
and in particular we titrate the role of the velocity of the moving object. %
Then, we show that the model can account for both standard and half-cycle FLEs. %
In particular, we demonstrate that within this optimal integration scheme, %
the relative precision of sensory and internal information %
may modulate the gain of their interaction. %
This is first illustrated by challenging the model with a motion reversal experiment. %
To further investigate this behavior, we manipulated the contrast of the stimuli. %
This allowed us to dynamically switch the system from a purely feed-forward model, %
exhibiting differential latencies, %
to a model showing a pure motion extrapolation behavior. %
We will finally discuss the advantages and limitations %
of our \emph{parodiction} hypothesis, in comparison with the previously proposed models. %
%-------------------------------------------------------------------------------
%###############################################################################
%-------------------------------------------------------------------------------
%###############################################################################
%-------------------------------------------------------------------------------
\section{Model \& methods}%
\label{sec:met}
%###############################################################################
\subsection{Motion-based prediction and the diagonal models}
%###############################################################################
%###############################################################################
This computational study explores the potential role
of predictive coding in explaining the dynamics of position coding.
Similar to most predictive models, a natural choice for the representation
of information is to use probabilities.
Thus, the motion of an object is best described
by the probability distribution function (pdf)
of its instantaneous position $(x, y)$ and velocity $(u, v)$~\citep{Perrinet12pred}.
Note that these coordinates are defined in the planar visual space,
under the assumption that we model small displacements
in the vicinity of the visual axis.
The pdf $p(x, y, u, v)$ represents at a given time
the degree of belief among a set of possible positions and velocities.
In this framework, a Bayesian predictive model will
optimally integrate the sensory information available
from the sensory inputs (likelihood) with an internal model
(i.~e. an \emph{a priori} distribution) of state transition
in order to compute an \emph{a posteriori} pdf of motion.
Typically, the likelihood is computed using
a model of sensory noise,
an approach that fits well to the motion energy model
of the direction-selective cells in the cortex~\citep{Adelson85,Weiss01,Perrinet07}.
By sequentially combining at any given time $t$ the \emph{a posteriori} estimate
with the likelihood using the prior on state transition,
we implement a Markov chain forming a dynamical predictive system.
One novelty of motion-based prediction is to encapsulate
the coherency of motion trajectory in the internal model,
that is, in the prior of state transition. In particular,
this prior knowledge instantiates a preference
for smooth transitions of successive motion states, as expected
from the statistics of natural visual motion trajectories~\citep{Weiss01}.
Such an internal model was first proposed in a neural network
implementing the detection of a single moving dot embedded
in very high level of noise~\citep{Burgi00}.
More recently, we have shown that this motion-based prediction model
can explain the dynamics of the neural solution
to both the aperture problem~\citep{Perrinet12pred}
and motion extrapolation~\citep{Khoei13}.
In the present study, we will show that it can also be used to compensate
for known neural delays~\citep{PerrinetAdamsFriston14}.
It is important to recall that our model is reminiscent of the diagonal model
originally proposed by~\citep{Nijhawan09}
(called thereafter Nijhawan's diagonal model), but with one important distinction: motion information
is now represented by probabilities.
%############################
\subsection{A probabilistic implementation of the Nijhawan's diagonal model}
%###########################
%###########################
In order to define a predictive system, one can use a classical Markov chain
formed by sequentially combining, at any given time,
the likelihood with a prior on state transition. % (see Fig~\ref{fig:DiagonalMarkov}-A).
When describing visual motion (i.e., position and velocity) at time $t$
by the instantaneous state vector $z_t = (x_t, y_t, u_t, v_t)$,
the master equations of this Markov chain become:
%
\begin{align}%
\textit{estimation:~}& p(z_{t} | I_{0:t}) \propto p( I_{t-\delta t:t} | z_{t})
\cdot p(z_{t} | I_{0:t-\delta t}) \label{eq:mc1}\\%
\textit{prediction:~}& p(z_{t} | I_{0:t-\delta t}) = %\nonumber \\
\int {dz_{t-\delta t}} \cdot p( z_{t} | z_{t-\delta t})
\cdot p(z_{t-\delta t} | I_{0:t-\delta t}) \label{eq:mc2}%
\end{align}%
%
where $p( I_{t-\delta t:t} | z_{t})$ is the likelihood
computed over the infinitesimally small temporal window $[t-\delta t, t)$,
that is, the time window of width $\delta t$ before present time $t$.
%(see Fig~\ref{fig:DiagonalMarkov}-A).
By definition, the pdf $p(z_{t} | I_{0:t})$ corresponds
to the belief in the motion $z_t$ at time $t$,
knowing the sensory information $I$ being integrated over the temporal window starting
at the beginning of the observations ($t=0$) and extending to the current time $t$.
Notice that the probabilistic notations will allow us
to conveniently describe the current belief on the state vector
before observing a new sensory information as the pdf $p(z_{t} | I_{0:t-\delta t})$.
Importantly, this dynamical system is biologically realistic as it
describes the belief in a macroscopic time window $[0, t)$, based only
on the integration of the information available
at the present time $[t-\delta t, t)$.
Intuitively, this model combines the two basic operations of probability theory.
First, in the estimation stage, the multiplication corresponds
to the combination of two independent pieces of information:
one element is derived from the measurements
(i.e. the likelihood $p( I_{t-\delta t:t} | z_{t})$) and
the other is the current knowledge about the state before the measurements.
Information is assumed to be conditionally independent
because the source of noise in the likelihood (measurement noise)
is assumed to be independent from the internally generated
state estimation noise.
This first stage corresponds to a {\sc AND} operator in boolean logic.
Second, in the prediction state, the possible state $p(z_{t} | I_{0:t-\delta t})$
is inferred by an addition over all possible previous states,
given by the integral sign.
The integrals sum over the whole distribution of estimated positions
the possible state transitions at $t-\delta t$ that would yield $z_t$.
By consequence, very unlikely states (at time $t-\delta t$) and
state transitions (for instance, incoherent non-smooth motions)
will have little weight in this summation.
This computational step corresponds to a {\sc OR} operator in boolean logic.
These two steps implement the classical ``predict-update cycle''
of the Markov model and are sufficient
to define our dynamical predictive coding model.
%check the two sentences below
However, in this mathematical framework,
one needs to gain an immediate access to sensory information,
that is, to know the image $I$ at time $t$,
in order to compute $p( I_{t-\delta t:t} | z_{t})$.
This is impossible in the presence of neural delays.
Considering a known (fixed) neural delay $\tau$,
at time $t$ the system only had access to $I_{0:t-\tau}$
and thus one may only estimate $p(z_{t} | I_{0:t-\tau})$.
An essential property of motion-based predictive coding
is the ability to extrapolate motion when sensory information
is transiently absent~\citep{Khoei13}.
As illustrated in Fig~\ref{fig:DiagonalMarkov}-A,
the probability distribution function $p(z_{t} | I_{0:t-\tau})$
may be predicted by ``pushing forward'' the information $p(z_{t-\tau} | I_{0:t-\tau})$
such as to compensate for the delay, while being still recursively
computed in a way similar to a classical Markov chain model
(see the bottom part of Fig~\ref{fig:DiagonalMarkov}-A).
Thus, a classical Markov chain in the presence of a known delay
can be redrawn in a ``diagonal'' mode.
It is similar to the original suggestion
made by~\citet{Nijhawan09} in order to explain the detailed mechanism
of motion extrapolation in retinal ganglion cells.
Here, we generalize this diagonal mode as a probabilistic model
of predictive motion estimation. %
As a consequence, the master equations of this diagonal model can be written as:
%
\begin{align}%
\textit{estimation:~} & \; p(z_{t-\tau} | I_{0:t-\tau}) \propto
p( I_{t-\tau-\delta t:t-\tau} | z_{t-\tau})
\cdot p( z_{t-\tau} | I_{0:t-\tau-\delta t})\label{eq:diag-push1}\\
\textit{prediction:~} & \; p(z_{t-\tau} | I_{0:t- \tau-\delta t}) =
\int {dz_{t- \tau-\delta t}} \cdot p( z_{t-\tau} | z_{t-\tau-\delta t})
\cdot p(z_{t-\tau-\delta t} | I_{0:t-\tau-\delta t}) \label{eq:diag-push2}\\
\textit{extrapolation:~} & \; p(z_{t} | I_{0:t- \tau}) =
\int {dz_{t-\tau}} \cdot p( z_{t} | z_{t-\tau})
\cdot p(z_{t-\tau} | I_{0:t-\tau}) \label{eq:diag-push3}
\end{align}%
%
As evidenced by these equations,
Equations~\ref{eq:diag-push1} and~\ref{eq:diag-push2} are similar
to Equations~\ref{eq:mc1} and~\ref{eq:mc2},
except that these are now delayed by $\tau$, the known sensory delay.
This information $p(z_{t-\tau} | I_{0:t-\tau})$
is then ``pushed'' forward in time using the extrapolation step
(see Equation~\ref{eq:diag-push3}),
in a similar fashion to the predictive step
on the infinitesimal period (Equations~\ref{eq:mc2} and~\ref{eq:diag-push2})
but now on the possibly longer period of the sensory delay (in general $\tau >> \delta t$).
As a result, we obtain the estimate of motion at the current time,
knowing the information acquired until $t-\tau$,
that is, $p( z_{t} | I_{0:t- \tau})$.
Finally, the next states correspond to the integration
of the estimations at the actual current stimulus position and motion,
overcoming the restrictive effect of delay~\citep{PerrinetAdamsFriston14}.
Note that the earliest part of the trajectory is necessarily missed since
motion estimation begins integrating sensory information only
after the delay $\tau$, as there is no sensory input before.
Decisively, this model is now compatible
with our initial hypothesis that sensory information
is only available after a delay.
%------------------------------------------------------------------------------%
%: fig:DiagonalMarkov
\begin{figure}%[ht]
\centering{%
\if 0\Draft
\if 0\Pace %
\includegraphics[width=\textwidth]{FLE_DiagonalMarkov} %
\else %
\includegraphics[width=\textwidth]{Fig2} %
\fi %
\else
Figure 2 around here
\fi
}%
\caption{\textbf{Diagonal Markov chain.}
In the current study, the estimated state vector $z = \{x, y, u, v\}$
is composed of the 2D position ($x$ and $y$) and velocity ($u$ and $v$)
of a (moving) stimulus.
{\sf (A)}~First, we extend a classical Markov chain using
Nijhawan's diagonal model in order to take into account
the known neural delay $\tau$:
At time $t$, information is integrated until time $t-\tau$,
using a Markov chain and a model of state transitions $p(z_t | z_{t-\delta t})$
such that one can infer the state
until the last accessible information $p( z_{t-\tau} | I_{0:t-\tau})$.
This information can then be ``pushed'' forward in time by predicting
its trajectory from $t-\tau$ to $t$.
In particular $p( z_{t} | I_{0:t-\tau})$ can be predicted
by the same internal model by using the state transition
at the time scale of the delay, that is, $p(z_t | z_{t-\tau})$.
This is virtually equivalent to a motion extrapolation model
but without sensory measurements during the time window
between $t-\tau$ and $t$.
Note that both predictions in this model are based
on the same model of state transitions.
{\sf (B)}~One can write a second, equivalent ``pull'' mode for the diagonal model.
Now, the current state is directly estimated based on a Markov chain
on the sequence of delayed estimations.
While being equivalent to the push-mode described above,
such a direct computation allows to more easily combine information
from areas with different delays.
Such a model implements Nijhawan's ``diagonal model'', but now
motion information is probabilistic and therefore,
inferred motion may be modulated by the respective precisions
of the sensory and internal representations.
{\sf (C)}~Such a diagonal delay compensation can be demonstrated
in a two-layered neural network including
a source (input) and a target (predictive) layer~\citep{KaplanKhoei14}.
The source layer receives the delayed sensory information and
encodes both position and velocity topographically
within the different retinotopic maps of each layer.
For the sake of simplicity, we illustrate only one 2D map of the motions $(x, v)$.
The integration of coherent information can either be done
in the source layer (push mode) or in the target layer (pull mode).
Crucially, to implement a delay compensation in this motion-based prediction model,
one may simply connect each source neuron to a predictive neuron corresponding
to the corrected position of stimulus $(x+v\cdot \tau, v)$ in the target layer.
The precision of this anisotropic connectivity map can be tuned
by the width of convergence
from the source to the target populations.
Using such a simple mapping, we have previously shown that
the neuronal population activity
can infer the current position along the trajectory
despite the existence of neural delays~\citep{KaplanKhoei14}.
}\label{fig:DiagonalMarkov}
\end{figure}
%--------------------------------------------------------------------------------------
Although this first model (i.e. the ``pushing'' mode) is the easiest to understand
with respect to a Markov chain, it is less practical to consider within a
biological setting since it defines that, at time $t$,
the state is inferred from a representation
of a past state $p(z_{t-\tau} | I_{0:t-\tau})$.
For neural networks implementations,
the internal representation (as encoded by the neural activity)
is only accessible at the present time.
As a consequence, it may be more convenient
to derive a set of predictive steps that would directly act on the estimation
of the state at the current time $p( z_{t} | I_{0:t-\tau})$.
This question is particularly acute for complex architectures mimicking
the deep hierarchies of low-level visual cortical areas
where information should not be conditioned by the delays arising
at each processing layer but rather be based on a common temporal reference
such as the current time $t$.
In that objective, one notes that, by merging
the estimation and prediction steps in the master equation, we obtain:
%#################
\begin{align}%
p(z_{t} | I_{0:t-\tau}) =& \int {dz_{t-\tau}} \cdot p( z_{t} | z_{t-\tau})
\cdot p(z_{t-\tau} | I_{0:t-\tau}) \nonumber \\
\propto & \int {dz_{t-\tau}} \cdot p( z_{t} | z_{t-\tau})
\cdot [ p( I_{t-\tau-\delta t:t-\tau} | z_{t-\tau})
\cdot p( z_{t-\tau} | I_{0:t-\tau-\delta t}) ] \\
\propto & [\int {dz_{t-\tau}} \cdot p( z_{t} | z_{t-\tau})
\cdot p( I_{t-\tau-\delta t:t-\tau} | z_{t-\tau})]
\cdot \int {dz_{t- \tau-\delta t}}
\cdot p( z_{t-\tau} | z_{t-\tau-\delta t}) \cdot \nonumber \\
& p(z_{t-\tau-\delta t} | I_{0:t-\tau-\delta t})
\end{align}
Regrouping terms, it becomes:
\begin{align}%
p(z_{t} | I_{0:t-\tau}) \propto &
\int {dz_{t-\tau}}
\cdot p( z_{t} | z_{t-\tau}) \cdot
[ \int {dz_{t- \tau-\delta t}}
\cdot p( z_{t- \tau} | z_{t- \tau-\delta t})
\cdot p( I_{t-\tau-\delta t:t-\tau} | z_{t-\tau}) \cdot \nonumber \\
& p(z_{t-\tau-\delta t} | I_{0:t-\tau-\delta t})]
\end{align}
The term within brackets can be written as an argument of an extrapolation
from $t-\tau$ to $t$, yielding to:
\begin{align}%
p(z_{t} | I_{0:t-\tau}) \propto
\int {dz_{t-\delta t}}\cdot p( z_{t} | z_{t-\delta t})
\cdot p( I_{t-\tau-\delta t:t-\tau} | z_{t})
\cdot p( z_{t-\delta t} | I_{0:t-\tau-\delta t})
\end{align}
As frequently assumed, the transition matrix is stationary:
our prior assumption on the internal model
(here, the parameters with which we model the coherence of trajectories)
do not change over time. Finally, regrouping terms, we obtain:
\begin{align}%
p(z_{t} | I_{0:t-\tau}) \propto
p( I_{t-\tau-\delta t:t-\tau} | z_{t}) \cdot [ \int {dz_{t-\delta t}}
\cdot p( z_{t} | z_{t-\delta t}) \cdot p(z_{t-\delta t} | I_{0:t-\tau-\delta t}) ]
\end{align}
Therefore, the master equation to the ``push'' mode are equivalent to:
%#################
\begin{align}%
%\textit{pulling mode:} \nonumber \\
\textit{estimation:~} & \; p(z_{t} | I_{0:t-\tau}) \propto % \nonumber \\
p( I_{t-\tau-\delta t:t-\tau} | z_{t})
\cdot p( z_{t} | I_{0:t-\tau-\delta t})\label{eq:diag-pull1}\\
\textit{prediction:~} & \; p(z_{t} | I_{0:t- \tau-\delta t}) = % \nonumber \\
\int {dz_{t-\delta t}} \cdot p( z_{t} | z_{t-\delta t})
\cdot p(z_{t-\delta t} | I_{0:t-\tau-\delta t}) \label{eq:diag-pull2}\\
\textit{extrapolation:~} & \; p( I_{t-\tau-\delta t:t-\tau} | z_{t}) =
\int {dz_{t-\tau}} \cdot p( z_{t} | z_{t-\tau})
\cdot p( I_{t-\tau-\delta t:t-\tau} | z_{t-\tau}) \label{eq:diag-pull3}
\end{align}
We will call this second mode the ``pulling'' mode
and is illustrated in Fig~\ref{fig:DiagonalMarkov}-B.
The two modes that have been presented above share the same processing logic
but have different implications about the manner
with which both the internal model and the likelihood function might be implemented.
In the pushing mode, the motion state $z_{t-\tau}$ is estimated from both
a delayed sensory input $I_{t-\tau-\delta t:t-\tau}$ and the motion coherency.
Equation \ref{eq:diag-push1} calculates the probability of a desired motion state,
using the likelihood of that state
(measured from the sensory information with a delay $\tau$)
and the predicted belief given by equation \ref{eq:diag-push2}.
At the next step, the estimated motion is extrapolated
for a period of duration $\tau$, similar to a ``virtual blank''
during which there is no sensory measurements~\citep{Khoei13}.
% (see Fig~\ref{fig:DiagonalMarkov}-B).
Thus, the extrapolation step shown by equation \ref{eq:diag-push3}
is purely predictive, under the constraint of motion coherency (see Equation~\ref{eq:diag-pull2}) and
with the available information about the delay $\tau$.
In the pulling mode, the probabilistic representation is different
as the current state is directly estimated from the delayed measurements and
the extrapolative step is ``hidden''
in the probability $p(z_{t} | I_{0:t- \tau-\delta t})$.
Under the stationarity assumption,
both modes are mathematically equivalent and
produce the same probabilistic representation
of instantaneous states based on delayed measurements.
In summary, the information about the motion estimates (position, velocity)
at time $t$ knowing the sensory information observed between $0$ and $t-\tau$
is contained in the pdf $p(z_{t} | I_{0:t-\tau})$.
As we have seen above, it can be computed using the diagonal model in push mode and
summarized in the following master equations:
\begin{align}%
p(z_{t} | I_{0:t-\tau}) &\propto \int {dz_{t-\delta t}} \cdot p( z_{t} | z_{t-\delta t})
\cdot p( I_{t-\tau-\delta t:t-\tau} | z_{t}) %\nonumber \\
\cdot p(z_{t-\delta t} | I_{0:t-\tau-\delta t}) \label{eq:master-push1} \\
p(I_{t-\tau-\delta t:t-\tau} | z_{t}) &=
\int {dz_{t-\tau}} \cdot p( z_{t} | z_{t-\tau})
\cdot p(I_{t-\tau-\delta t:t-\tau} | z_{t-\tau}) \label{eq:master-push2}
\end{align}
Equations~\ref{eq:master-push1} and \ref{eq:master-push2} are the master equations
of Nijhawan's diagonal model when framing it in a probabilistic setting.
Importantly, the inferred motion may be modulated by the respective precisions
of the sensory ($p(I_{t-\tau-\delta t:t-\tau} | z_{t})$)
and internal ($p(z_{t-\delta t} | I_{0:t-\tau-\delta t})$) representations.
%
The model gives a probabilistic distribution of the estimated motion state $z_t$,
based on delayed motion measurements $I_{t-\tau}$.
In the next section, we will describe how
the transition probability distribution functions $p( z_{t} | z_{t-\tau})$ and
$p( z_{t} | z_{t-\delta t})$ are computed.
%
\subsection{Diagonal Motion-based prediction (dMBP)}
\label{subsec:mbp}
% todo : see Kowler for more complex trajectories
We have seen above that one needs to characterize the temporal coherency of motion
for different temporal steps, as represented by $p( z_{t} | z_{t-\Delta t})$
with $\Delta t = \delta t$ or $\Delta t = \tau$. Assuming that motion
is \emph{transported} in time during this time period of $\Delta t$
(with a drift similar to a Brownian motion and characterized
by some given diffusion parameters), we obtain~\citep{Perrinet12pred}:
%#############################
\begin{align}
& \left\{
\begin{array}{lll}
\ x_{t} &= x_{t-\Delta t} + u_{t-\Delta t} \cdot \Delta t + {\nu}_x \\%\nonumber \\%
\ y_{t} &= y_{t-\Delta t} + v_{t-\Delta t} \cdot \Delta t + {\nu}_y %\\%
\end{array}
\right.
\label{eq:smooth1}
\\
& \left\{
\begin{array}{lll}
\ u_{t} &= \gamma \cdot u_{t-\Delta t} + {\nu}_u \\%\nonumber \\%
\ v_{t} &= \gamma \cdot v_{t-\Delta t} + {\nu}_v %
%\ u_{t} &= u_{t-\Delta t} + {\nu}_u \\%\nonumber \\%
%\ v_{t} &= v_{t-\Delta t} + {\nu}_v %
\end{array}
\right.
\label{eq:smooth2}%
\end{align}%
Here, $\gamma=(1 + \frac{D_{V}}{\sigma_{p}^{2}})^{-1}$ is the damping factor
introduced by the prior on slowness of motion~\citep{Khoei13}.
As defined by~\citet{Weiss01}, this prior information
about slowness and smoothness of visual motion can be parameterized
by its variance $\sigma_{p}^{2}$ and
$\gamma \approx 1$ for a high value of $\sigma_p$.
The diffusion parameters characterize the precision of the temporal motion coherency and
are parameterized by the variance of the Gaussian distributions that
define the additive noise ${\nu}_x, {\nu}_y$ in the transport equations.
First, the variance $D_{X} \cdot |\Delta t|$ setting the blur in position define the noise distribution as:
%