pages/index.html

1594 lines
105 KiB
HTML
Raw Permalink Blame History

This file contains invisible Unicode characters!

This file contains invisible Unicode characters that may be processed differently from what appears below. If your use case is intentional and legitimate, you can safely ignore this warning. Use the Escape button to reveal hidden characters.

<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">
<html xmlns="http://www.w3.org/1999/xhtml">
<head>
<meta http-equiv="content-type" content="text/html;charset=iso-8859-1" />
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<link rel="stylesheet" href="mws.css" type="text/css" />
<script type="text/javascript" src="jquery.js"></script>
<script type="text/javascript">
jQuery(document).ready(
function($) {
$(".abstract").toggle();
$("div").click(function() {
//$(".abstract").hide();
if (!$(this).next("div.abstract").is(":visible"))
$(this).next("div.abstract").toggle(600);
else
$(this).next("div.abstract").toggle(400);
});
});
</script>
<title>Michael Spratling: Home Page</title>
</head>
<body>
<div class="navbar">
<div><a href="#top">Dr. Michael Spratling</a></div>
<a href=""></a>
<a href="https://codeberg.org/mwspratling">Code</a>
<a href="#publications">Publications</a>
<a href="#research_interests">Research</a>
</div>
<div id="top"></div><br><br>
<table cellspacing=0 cellpadding=10px border=0>
<tr>
<td valign=center width=50%>
<p>Reader in Computational Neuroscience and Visual Cognition<br />
Department of Informatics<br />
King's College London<br />
University of London</p>
<script type="text/javascript">
var email = 'michael.spratling';
var emailHost = 'kcl.ac.uk';
var contact= 'michael <i>dot</i> spratling <i>at </i>kcl <i>dot</i> ac <i>dot</i> uk';
document.write('<a href=' + 'ma' + 'il' + 'to:' + email + '@' +
emailHost + '>' + contact + '</a>');
</script><p>
</td>
<td bgcolor="#ffffff" align=center valign=center>
<img src="Logo/cnvc.jpg" alt="logo" width="150" style="margin-left:-10px"/>
</td>
<td bgcolor="#ffffff" align=left valign=top width=50%>
</td>
</tr>
</table>
<div id="research_interests"></div><br>
<h2 >Research Interests</h2>
<table cellspacing=0 cellpadding=5px border=0>
<tr>
<td bgcolor="#ffffff" align=right valign=top width=50%>
<h3>Neural Networks</h3>
deep learning; biologically-inspired neural networks; adversarial robustness; generalisation;
representation learning; dynamic and deformable neural networks
</td>
<td bgcolor="#ffffff" align=center valign=center>
<img src="Logo/neural.jpg" width="120" align=left/>
</td>
<td bgcolor="#ffffff" align=left valign=top width=50%>
<h3>Computational Neuroscience</h3>
predictive coding; neural information processing and coding; dendritic computation; cortical feedback connections; cortical region interactions
</td>
</tr>
<tr>
<td bgcolor="#ffffff" align=right valign=top width=50%>
<h3>Computer Vision</h3>
object recognition; viewpoint invariance; image segmentation; object detection; tracking; anomaly detection; open-set recognition
</td>
<td bgcolor="#ffffff" align=center valign=center>
<img src="Logo/vision.jpg" width="120"/>
</td>
<td bgcolor="#ffffff" align=left valign=top width=50%>
<h3>Visual Cognition</h3>
visual attention; biased competition; top-down and contextual influences; visual salience; perceptual inference
</td>
</tr>
<tr>
<td bgcolor="#ffffff" align=right valign=top width=50%>
<h3>Machine Learning</h3>
sparse coding; continuous learning; few-shot learning; unsupervised and self-supervised learning; competitive learning; classification; clustering; contrastive and deep-metric learning
<td bgcolor="#ffffff" align=center valign=center>
<img src="Logo/development.jpg" width="120"/>
</td>
<td bgcolor="#ffffff" align=left valign=top width=50%>
<h3>Learning and Development</h3>
cortical, behavioural and cognitive development; perceptual learning; perceptual and conceptual development; epigenetic and developmental robotics
</td>
</tr>
</table>
<div id="publications"></div><br>
<H2>Publications
<div style="float: right; font-size: small">
<A class="download" HREF = "http://scholar.google.com/citations?user=A2tJIE4AAAAJ&hl=en"><B>Google Scholar</B></A><A class="download" HREF = "https://www.scopus.com/authid/detail.uri?authorId=6603009678"><B>Scopus</B></A><A class="download" HREF = "http://www.researcherid.com/rid/G-7689-2011"><B>Researcher ID</B></A><A class="download" HREF = "http://orcid.org/0000-0001-9531-2813"><B>ORCID</B></A>
</div></H2>
<ul>
<!--
L. Li and M. W. Spratling (2023) <b>Improved adversarial training through adaptive instance-wise loss smoothing.</b> <I> <A HREF = ""></A></I>, .
<ul>
<A class="download" HREF = "https://arxiv.org/pdf/2303.14077">PDF</A><A
class="download" HREF = "https://github.com/TreeLLi/Instance-adaptive-Smoothness-Enhanced-AT">Code</A><div class="download">Abstract</div><div class="abstract"> eep neural networks can be easily fooled into making incorrect predictions through corruption of the input by adversarial perturbations: human-imperceptible artificial noise. So far adversarial training has been the most successful defense against such adversarial attacks. This work focuses on improving adversarial training to boost adversarial robustness. We first analyze, from an instance-wise perspective, how adversarial vulnerability evolves during adversarial training. We find that during training an overall reduction of adversarial loss is achieved by sacrificing a considerable proportion of training samples to be more vulnerable to adversarial attack, which results in an uneven distribution of adversarial vulnerability among data. Such "uneven vulnerability", is prevalent across several popular robust training methods and, more importantly, relates to overfitting in adversarial training. Motivated by this observation, we propose a new adversarial training method: Instance-adaptive Smoothness Enhanced Adversarial Training (ISEAT). It jointly smooths both input and weight loss landscapes in an adaptive, instance-specific, way to enhance robustness more for those samples with higher adversarial vulnerability. Extensive experiments demonstrate the superiority of our method over existing defense methods. Noticeably, our method, when combined with the latest data augmentation and semi-supervised learning techniques, achieves state-of-the-art robustness against ℓ∞-norm constrained attacks on CIFAR10 of 59.32% for Wide ResNet34-10 without extra data, and 61.55% for Wide ResNet28-10 with extra data.
</div></ul>
<br>-->
M. W. Spratling (2023) <B>Comprehensive assessment methods are key to progress in deep learning</B> [commentary]. <I><A HREF = "https://www.cambridge.org/core/journals/behavioral-and-brain-sciences">Behavioral and Brain Sciences</A></I>, in press.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/commentary_bowers.pdf">PDF</A>
</ul>
<br>
H. Guan and M. W. Spratling (2023) <b>Query semantic reconstruction for background in few-shot segmentation.</b> <I><A HREF = "https://doi.org/10.1007/s00371-023-02817-x">The Visual Computer</A></I>, in press.
<ul>
<A class="download" HREF = "https://arxiv.org/pdf/2210.12055">PDF</A><!--<A class="download" HREF = "">Code</A>--><div class="download">Abstract</div><div class="abstract">
Few-shot segmentation (FSS) aims to segment unseen classes using a few annotated samples.
Typically, a prototype representing the foreground class is extracted from annotated support image(s)
and is matched to features representing each pixel in the query image. However, models learnt in this
way are insufficiently discriminatory, and often produce false positives: misclassifying background
pixels as foreground. Some FSS methods try to address this issue by using the background in the
support image(s) to help identify the background in the query image. However, the backgrounds
of theses images is often quite distinct, and hence, the support image background information is
uninformative. This article proposes a method, QSR, that extracts the background from the query
image itself, and as a result is better able to discriminate between foreground and background features
in the query image. This is achieved by modifying the training process to associate prototypes with
class labels including known classes from the training data and latent classes representing unknown
background objects. This class information is then used to extract a background prototype from
the query image. To successfully associate prototypes with class labels and extract a background
prototype that is capable of predicting a mask for the background regions of the image, the machinery
for extracting and using foreground prototypes is induced to become more discriminative between
different classes. Experiments achieves state-of-the-art results for both 1-shot and 5-shot FSS on the
PASCAL-5i and COCO-20i dataset. As QSR operates only during training, results are produced with
no extra computational complexity during testing.
</div></ul>
<br>
L. Li and M. W. Spratling (2023) <b>Data augmentation alone can improve adversarial training.</b> <I>Proceedings of the 11th <A HREF = "https://openreview.net/forum?id=y4uc4NtTWaq">International Conference on Learning Representations (ICLR)</A></I>.
<ul>
<A class="download" HREF = "https://arxiv.org/pdf/2301.09879.pdf">PDF</A><A
class="download" HREF = "https://github.com/TreeLLi/DA-Alone-Improves-AT">Code</A><div class="download">Abstract</div><div class="abstract">Adversarial training suffers from the issue of robust overfitting, which seriously impairs its generalization performance. Data augmentation, which is effective at preventing overfitting in standard training, has been observed by many previous works to be ineffective in mitigating overfitting in adversarial training. This work proves that, contrary to previous findings, data augmentation alone can significantly boost accuracy and robustness in adversarial training. We find that the hardness and the diversity of data augmentation are important factors in combating robust overfitting. In general, diversity can improve both accuracy and robustness, while hardness can boost robustness at the cost of accuracy within a certain limit and degrade them both over that limit. To mitigate robust overfitting, we first propose a new crop transformation Cropshift with improved diversity compared to the conventional one (Padcrop). We then propose a new data augmentation scheme, based on Cropshift, with much improved diversity and well-balanced hardness. Empirically, our augmentation method achieves the state-of-the-art accuracy and robustness for data augmentations in adversarial training. Furthermore, it matches, or even exceeds when combined with weight averaging, the performance of the best contemporary regularization methods for alleviating robust overfitting.
</div></ul>
<br>
J. Ning, H. Guan and M. W. Spratling (2023) <b>Rethinking the backbone architecture for tiny object detection.</b> <I>Proceedings of the 18th <A HREF =
"https://doi.org/10.5220/0011643500003417">International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISAPP)</A></I>, Volume 5, pp. 103-114.
<ul>
<A class="download" HREF = "https://arxiv.org/pdf/2303.11267">PDF</A><!--<A
class="download" HREF = "">Code</A>--><div class="download">Abstract</div><div class="abstract">
Tiny object detection has become an active area of research because images with tiny targets are common in several important real-world scenarios. However, existing tiny object detection methods use standard deep neural networks as their backbone architecture. We argue that such backbones are inappropriate for detecting tiny objects as they are designed for the classification of larger objects, and do not have the spatial resolution to identify small targets. Specifically, such backbones use max-pooling or a large stride at early stages in the architecture. This produces lower resolution feature-maps that can be efficiently processed by subsequent layers. However, such low-resolution feature-maps do not contain information that can reliably discriminate tiny objects. To solve this problem we design ”bottom-heavy” versions of backbones that allocate more resources to processing higher-resolution features without introducing any additional computational burden overall. We also investigate if pre-training these backbones on images of appropriate size, using CIFAR100 and ImageNet32, can further improve performance on tiny object detection. Results on TinyPerson and WiderFace show that detectors with our proposed backbones achieve better results than the current state-of-the-art methods.
</div></ul>
<br>
L. Li and M. W. Spratling (2023) <b>Understanding and combating robust overfitting via input loss landscape analysis and regularization.</b> <I><A HREF = "https://doi.org/10.1016/j.patcog.2022.109229">Pattern Recognition</A></I>, 136 (109229).
<ul>
<A class="download" HREF = "https://www.sciencedirect.com/science/article/pii/S0031320322007087/pdfft?md5=793d3d60b7b6a89f014c27b8562c1276&pid=1-s2.0-S0031320322007087-main.pdf">PDF</A><A
class="download" HREF = "https://github.com/TreeLLi/Combating-RO-AdvLC">Code</A><div class="download">Abstract</div><div class="abstract">
Adversarial training is widely used to improve the robustness of deep neural networks to adversarial attack. However, adversarial training is prone to overfitting, and the cause is far from clear. This work sheds light on the mechanisms underlying overfitting through analyzing the loss landscape w.r.t. the input. We find that robust overfitting results from standard training, specifically the minimization of the clean loss, and can be mitigated by regularization of the loss gradients. Moreover, we find that robust overfitting turns severer during adversarial training partially because the gradient regularization effect of adversarial training becomes weaker due to the increase in the loss landscape's curvature. To improve robust generalization, we propose a new regularizer to smooth the loss landscape by penalizing the weighted logits variation along the adversarial direction. Our method significantly mitigates robust overfitting and achieves the highest robustness and efficiency compared to similar previous methods.
</div></ul>
<br>
B. Gao and M. W. Spratling (2023) <b>Explaining away results in more robust visual tracking.</b> <I><A HREF = "https://doi.org/10.1007/s00371-022-02466-6">The Visual Computer</A></I>, 39:2081-95.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_tracking.pdf">PDF</A><A
class="download" HREF = "https://github.com/iminfine/DIMtracking">Code</A><div class="download">Abstract</div><div class="abstract">
Many current trackers utilise an appearance model to localise the target object in each frame. However, such approaches often fail when there are similar looking distractor objects in the surrounding background, meaning that target appearance alone is insufficient for robust tracking. In contrast, humans consider the distractor objects as additional visual cues, in order to infer the position of the target. Inspired by this observation, this paper proposes a novel tracking architecture in which not only is the appearance of the tracked object, but also the appearance of the distractors detected in previous frames, taken into consideration using a form of probabilistic inference known as explaining away. This mechanism increases the robustness of tracking by making it more likely that the target appearance model is matched to the true target, rather than similar-looking regions of the current frame. The proposed method can be combined with many existing trackers. Combining it with SiamFC, DaSiamRPN, Super DiMP and ARSuper DiMP all resulted in an increase in the tracking accuracy compared to that achieved by the underlying tracker alone. When combined with Super DiMP and ARSuper DiMP the resulting trackers produce performance that is competitive with the state-of-the-art on seven popular benchmarks.
</div></ul>
<br>
B. Gao and M. W. Spratling (2022) <b>Shape-texture debiased training for robust template matching.</b> <I><A HREF = "https://doi.org/10.3390/s22176658">Sensors</A></I>, 22(17), 6658.
<ul>
<A class="download" HREF = "https://www.mdpi.com/1424-8220/22/17/6658/pdf?version=1662127786">PDF</A><A
class="download" HREF = "https://github.com/iminfine/Deep-DIM">Code</A><div class="download">Abstract</div><div class="abstract">
Finding a template in a search image is an important task underlying many computer vision applications. This is typical solved by calculating a similarity map using features extracted from the separate images. Recent approaches perform template matching in a deep feature-space, produced by a convolutional neural network (CNN), which is found to provide more tolerance to changes in appearance. Inspired by these findings, in this article we investigate if enhancing the CNN's encoding of shape information can produce more distinguishable features that improve the performance of template matching. By comparing features from a same CNN but trained by different shape-texture training methods, we determined a feature-space which improves the performance of most template matching algorithms. When combining the proposed method with the Divisive Input Modulation (DIM) template matching algorithm, its performance is greatly improved, and the resulting method produces state-of-the-art results on a standard benchmark. To confirm these results we also create a new benchmark and show that the proposed method also outperforms existing techniques on this new dataset.
</div></ul>
<br>
N. Manchev and M. W. Spratling (2022) <b>On the biological plausibility of orthogonal initialisation for solving gradient instability in deep neural networks.</b> <I>Proceedings of the 9th <A HREF = "https://doi.org/10.1109/ISCMI56532.2022.10068489">International Conference on Soft Computing and Machine Intelligence (ISCMI)</A></I>, pp. 47-55.
<ul>
<A class="download" HREF = "https://arxiv.org/pdf/2211.08408.pdf">PDF</A><!--<A class="download" HREF = "https://github.com/nmanchev/">Code</A>--><div class="download">Abstract</div><div class="abstract">
Initialising the synaptic weights of artificial neural
networks (ANNs) with orthogonal matrices is known to alleviate
vanishing and exploding gradient problems. A major objection
against such initialisation schemes is that they are deemed
biologically implausible as they mandate factorization techniques
that are difficult to attribute to a neurobiological process. This
paper presents two initialisation schemes that allow a network to
naturally evolve its weights to form orthogonal matrices, provides
theoretical analysis that pre-training orthogonalisation always
converges, and empirically confirms that the proposed schemes
outperform randomly initialised recurrent and feedforward networks.
</div></ul>
<br>
C. Huang, H. Guan, A. Jiang, Y. Zhang, M. W. Spratling and Y.-F. Wang (2022) <b>Registration based few-shot anomaly detection.</b> <I><A HREF = "https://doi.org/10.1007/978-3-031-20053-3_18">European Conference on Computer Vision (ECCV)</A></I>, In: Avidan, S., Brostow, G., Ciss&eacute, M., Farinella, G.M., Hassner, T. (eds) Proceedings part 24, Lecture Notes in Computer Science, Volume 13684, pp:303-19, Springer.
<ul>
<A class="download" HREF = "https://arxiv.org/pdf/2207.07361.pdf">PDF</A><A class="download" HREF = "https://github.com/MediaBrain-SJTU/RegAD">Code</A><div class="download">Abstract</div><div class="abstract">
This paper considers few-shot anomaly detection (FSAD), a practical yet under-studied setting for anomaly detection (AD), where only a limited number of normal images are provided for each category at training. So far, existing FSAD studies follow the one-model-per-category learning paradigm used for standard AD, and the inter-category commonality has not been explored. Inspired by how humans detect anomalies, i.e., comparing an image in question to normal images, we here leverage registration, an image alignment task that is inherently generalizable across categories, as the proxy task, to train a category-agnostic anomaly detection model. During testing, the anomalies are identified by comparing the registered features of the test image and its corresponding support (normal) images. As far as we know, this is the first FSAD method that trains a single generalizable model and requires no re-training or parameter fine-tuning for new categories. Experimental results have shown that the proposed method outperforms the state-of-the-art FSAD methods by 3%-8% in AUC on the MVTec and MPDD benchmarks. Source code will be publicly available.
</div></ul>
<br>
H. Guan and M. W. Spratling (2022) <b>CobNet: cross attention on object and background for few-shot segmentation.</b> <I>Proceedings of the 26th <A HREF = "https://doi.org/10.1109/ICPR56361.2022.9956070">International Conference on Pattern Recognition (ICPR)</A></I>, pp. 39-45.
<ul>
<A class="download" HREF = "https://arxiv.org/pdf/2210.11968.pdf">PDF</A><!--<A class="download" HREF = "">Code</A>--><div class="download">Abstract</div><div class="abstract">
Few-shot segmentation aims to segment images containing objects from previously unseen classes using only a few annotated samples. Most current methods focus on using object information extracted, with the aid of human annotations, from support images to identify the same objects in new query images. However, background information can also be useful to distinguish objects from their surroundings. Hence, some previous methods also extract background information from the support images. In this paper, we argue that such information is of limited utility, as the background in different images can vary widely. To overcome this issue, we propose CobNet which utilises information about the background that is extracted from the query images without annotations of those images. Experiments show that our method achieves a mean Intersection-over-Union score of 61.4% and 37.8% for 1-shot segmentation on PASCAL-5i and COCO-20i respectively, outperforming previous methods. It is also shown to produce state-of-the-art performances of 53.7% for weakly-supervised few-shot segmentation, where no annotations are provided for the support images.
</div></ul>
<br>
B. Gao and M. W. Spratling (2022) <b>More robust object tracking via shape and motion cue integration.</b> <I><A HREF = "https://doi.org/10.1016/j.sigpro.2022.108628">Signal Processing</A></I>, 199 (108628).
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/trackingmulticues.pdf">PDF</A><A
class="download" HREF = "https://github.com/iminfine/trackingmulticues">Code</A><div class="download">Abstract</div><div class="abstract">
Most current trackers utilise an appearance model to localise the target object in each frame. However, such approaches often fail when there are similar looking distractor objects in the surrounding background. This paper promotes an approach that can be combined with many existing trackers to tackle this issue and improve tracking robustness. The proposed approach makes use of two additional cues to target location: making the appearance model more sensitive to shape cues in offline training, and using the historical locations of the target to predict its future position during online inference. Combining these additional mechanisms with SiamFC, SiamFC++, Super DiMP and ARSuper DiMP all resulted in an increase in the tracking accuracy compared to that achieved by the corresponding underlying tracker alone. When combined with ARSuper DiMP the resulting tracker is shown to outperform all popular state-of-the-art trackers on three benchmark datasets (OTB-100, NFS, and LaSOT), and produce performance that is competitive with the state-of-the-art on the UAV123, Trackingnet, GOT-10K and VOT2020 datasets.
</div></ul>
<br>
B. Gao and M. W. Spratling (2021) <b>Robust template matching via hierarchical convolutional features from a shape biased CNN.</b> <I>Proceedings of the <A HREF = "https://doi.org/10.1007/978-981-16-6963-7_31">International Conference on Image, Vision and Intelligent Systems (ICIVIS)</A></I>, Lecture Notes in Electrical Engineering, Volume 813. Springer, Singapore.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_cnn_patchmatching.pdf">PDF</A><A
class="download" HREF = "https://github.com/iminfine/Deep-DIM">Code</A><div class="download">Abstract</div><div class="abstract">
Finding a template in a search image is an important task underlying many computer vision applications. Recent approaches perform template matching in a deep feature-space, produced by a convolutional neural network (CNN), which is found to provide more tolerance to changes in appearance. In this article we investigate if enhancing the CNN's encoding of shape information can produce more distinguishable features that improve the performance of template matching. This investigation results in a new template matching method that produces state-of-the-art results on a standard benchmark. To confirm these results we also create a new benchmark and show that the proposed method also outperforms existing techniques on this new dataset.
</div></ul>
<br>
M. W. Spratling (2020) <b>Explaining away results in accurate and tolerant template matching.</b> <I><A HREF = "https://doi.org/10.1016/j.patcog.2020.107337">Pattern Recognition</A></I>, 104 (107337).
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_patchmatching.pdf">PDF</A><A
class="download" HREF = "https://codeberg.org/mwspratling/PatchMatching">Code</A><div class="download">Abstract</div><div class="abstract">
Recognising and locating image patches or sets of image features is an important task underlying much work in computer vision. Traditionally this has been accomplished using template matching. However, template matching is notoriously brittle in the face of changes in appearance caused by, for example, variations in viewpoint, partial occlusion, and non-rigid deformations. This article tests a method of template matching that is more tolerant to such changes in appearance and that can, therefore, more accurately identify image patches. In traditional template matching the comparison between a template and the image is independent of the other templates. In contrast, the method advocated here takes into account the evidence provided by the image for the template at each location and the full range of alternative explanations represented by the same template at other locations and by other templates. Specifically, the proposed method of template matching is performed using a form of probabilistic inference known as "explaining away". The algorithm used to implement explaining away has previously been used to simulate several neurobiological mechanisms, and been applied to image contour detection and pattern recognition tasks. Here it is applied for the first time to image patch matching, and is shown to produce superior results in comparison to the current state-of-the-art methods.
</div></ul>
<br>
N. Manchev and M. W. Spratling (2020) <b>Target propagation in recurrent neural networks.</b> <I><A HREF = "http://www.jmlr.org/papers/v21/18-141.html">Journal of Machine Learning Research</A></I>, 21(7): 1-33.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/rnn_tptt.pdf">PDF</A><A
class="download" HREF = "https://github.com/nmanchev/tptt">Code</A><div class="download">Abstract</div><div class="abstract">
Recurrent Neural Networks have been widely used to process sequence data, but have
long been criticized for their biological implausibility and training difficulties related to
vanishing and exploding gradients. This paper presents a novel algorithm for training
recurrent networks, target propagation through time (TPTT), that outperforms standard
backpropagation through time (BPTT) on four out of the five problems used for testing.
The proposed algorithm is initially tested and compared to BPTT on four synthetic time
lag tasks, and its performance is also measured using the sequential MNIST data set. In
addition, as TPTT uses target propagation, it allows for discrete nonlinearities and could
potentially mitigate the credit assignment problem in more complex recurrent architectures.
</div></ul>
<br>
M. W. Spratling (2019) <b>Fitting predictive coding to the neurophysiological data.</b> <I><A HREF = "https://doi.org/10.1016/j.brainres.2019.146313">Brain Research</A></I>, 1720: 146313.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pc_biological_fit.pdf">PDF</A><A
class="download" HREF = "https://codeberg.org/mwspratling/BiologicalFit">Code</A><div class="download">Abstract</div><div class="abstract">
Recent neurophysiological data showing the effects of locomotion on neural
activity in mouse primary visual cortex has been interpreted as providing strong
support for the predictive coding account of cortical function. Specifically,
this work has been interpreted as providing direct evidence that
prediction-error, a distinguishing property of predictive coding, is encoded in
cortex. This article evaluates these claims and highlights some of the
discrepancies between the proposed predictive coding model and the
neuro-biology. Furthermore, it is shown that the model can be modified so as to
fit the empirical data more successfully.
</div></ul>
<br>
I. E. Kartoglu and M. W. Spratling (2018) <b>Two collaborative filtering recommender systems based on sparse dictionary coding.</b> <I><A HREF = "http://dx.doi.org/10.1007/s10115-018-1157-2">Knowledge and Information Systems</A></I>, 57(3): 709-20.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/sparse_recommender.pdf">PDF</A><A
class="download" HREF = "https://github.com/iemre/MRSR">Code</A><div class="download">Abstract</div><div class="abstract">
This paper proposes two types of recommender systems based on sparse dictionary
coding. Firstly, a novel predictive recommender system that attempts to predict
a user's future rating of a specific item. Secondly, a top-n recommender system
which finds a list of items predicted to be most relevant for a given user. The
proposed methods are assessed using a variety of different metrics and are shown
to be competitive with existing collaborative filtering recommender
systems. Specifically, the sparse dictionary-based predictive recommender has
advantages over existing methods in terms of a lower computational cost and not
requiring parameter tuning. The sparse dictionary-based top-n recommender system
has advantages over existing methods in terms of the accuracy of the predictions
it makes and not requiring parameter tuning. An open-source software implemented
and used for the evaluation in this paper is also provided for reproducibility.
</div></ul>
<br>
Q. Wang and M. W. Spratling (2018) <b>Contour detection refined by a sparse reconstruction-based discrimination method.</b> <I><A HREF = "http://dx.doi.org/10.1007/s11760-017-1147-y">Signal, Image and Video Processing</A></I>, 12(2): 207-14.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/SIVPRO-final.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
Sparse representations have been widely used for many image processing tasks. In
this paper, a sparse reconstruction-based discrimination (SRBD) method, which
was previously proposed for the classification of image patches, is utilized to
improve boundary detection in colour images. This method is applied to refining
the results generated by three different algorithms: a biologically-inspired
method, and two state-of-the-art algorithms for contour detection. All of the
contour detection results are evaluated by the BSDS300 and BSDS500 benchmarks
using the quantitative measures: F-score, ODS, OIS and AP. Evaluation results
show that the performance of each algorithm is improved using the proposed
method of refinement with at least one of the quantitative measures increased by
0.01. In particularly, even two state-of-the-art algorithms are slightly
improved by applying the SRBD method to refine their contour detection results.
</div></ul>
<br>
M. W. Spratling (2017) <b>A predictive coding model of gaze shifts and the
underlying neurophysiology.</b> <I><A HREF = "http://dx.doi.org/10.1080/13506285.2017.1336141">Visual Cognition</A></I>, 25(7-8): 770-801.
<ul> <A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_gaze.pdf">PDF</A><A
class="download" HREF = "https://codeberg.org/mwspratling/GazeSimulation">Code</A><div
class="download">Abstract</div><div class="abstract"> A comprehensive model of
gaze control must account for a number of empirical observations at both the
behavioural and neurophysiological levels. The computational model presented in
this article can simulate the coordinated movements of the eye, head, and body
required to perform horizontal gaze shifts. In doing so it reproduces the
predictable relationships between the movements performed by these different
degrees of freedom (DOFs) in the primate. The model also accounts for the
saccadic undershoot that accompanies large gaze shifts in the biological visual
system. It can also account for our perception of a stable external world
despite frequent gaze shifts and the ability to perform accurate memory-guided
and double-step saccades. The proposed model also simulates peri-saccadic
compression: the mis-localisation of a briefly presented visual stimulus towards
the location that is the target for a saccade. At the neurophysiological level,
the proposed model is consistent with the existence of cortical neurons tuned to
the retinal, head-centred, body-centred, and world-centred locations of visual
stimuli and cortical neurons that have gain-modulated responses to visual
stimuli. Finally, the model also successfully accounts for peri-saccadic
receptive field (RF) remapping which results in reduced responses to stimuli in
the current RF location and an increased sensitivity to stimuli appearing at the
location that will be occupied by the RF after the saccade. The proposed model
thus offers a unified explanation for this seemingly diverse range of
phenomena. Furthermore, as the proposed model is an implementation of the
predictive coding theory, it offers a single computational explanation for these
phenomena and relates gaze shifts to a wider framework for understanding
cortical function.
</div>
</ul>
<br>
W. Muhammad and M. W. Spratling (2017) <b>A neural model for eye-head-arm coordination.</b> <I><A HREF = "http://dx.doi.org/10.1080/01691864.2017.1315319">Advanced Robotics</A></I>, 31(12): 650-63.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_eye_head_arm_control.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
The coordinated movement of the eyes, the head and
the arm is an important ability in both animals and humanoid
robots. To achieve this the brain and the robot control system
need to be able to perform complex non-linear sensory-motor
transformations in the forward and inverse directions between
many degrees of freedom. In this article we apply an omni-
directional basis function neural network to this task. The
proposed network can perform 3-D coordinated gaze shifts and
3-D arm reach movements to a visual target. Particularly, it can
perform direct sensory-motor transformations to shift gaze and
to execute arm reach movements and can also perform inverse
sensory-motor transformations in order to shift gaze to view the
hand. </div>
</ul>
<br>
M. W. Spratling (2017) <b>A hierarchical predictive coding model of object recognition in natural images.</b> <I><A HREF =
"http://dx.doi.org/10.1007/s12559-016-9445-1">Cognitive Computation</A></I>, 9(2): 151-67.
<ul>
<A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_image_recognition.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/ImageRecognition">Code</A><div
class="download">Abstract</div><div class="abstract">
Predictive coding has been proposed as a model of the hierarchical
perceptual inference process performed in the cortex. However, results
demonstrating that predictive coding is capable of performing the complex
inference required to recognise objects in natural images have not
previously been presented.
This article proposes a hierarchical neural network based on predictive
coding for performing visual object recognition.
This network is applied to the tasks of categorising hand-written digits,
identifying faces, and locating cars in images of street scenes. It is shown
that image recognition can be performed with tolerance to position,
illumination, size, partial occlusion and within-category variation.
The current results, therefore, provide the first practical demonstration
that predictive coding (at least the particular implementation of predictive
coding used here; the PC/BC-DIM) is capable of performing accurate
visual object recognition.
</div>
</ul>
<br>
D. Re, A. Gibaldi, S. P. Sabatini and M. W. Spratling (2017) <b>An integrated system based on binocular learned receptive fields for saccade-vergence on visually salient targets.</b> <I>Proceedings of the 12th <A HREF =
"http://www.scitepress.org/DigitalLibrary/PublicationsDetail.aspx?ID=TyxypQ+XiJc=&t=1">International Conference on Computer Vision Theory and Applications (VISAPP)</A></I>, Volume 6, pp:204-15.
<ul>
<A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/re_visapp17.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
The human visual system uses saccadic and vergence eyes movements to foveate interesting objects with both eyes, and thus exploring the visual scene. To mimic this biological behavior in active vision, we proposed a bio-inspired integrated system able to learn a functional sensory representation of the environment, together with the motor commands for binocular eye coordination, directly by interacting with the environment itself. The proposed architecture, rather than sequentially combining different functionalities, is a robust integration of different modules that rely on a front-end of learned binocular receptive fields to specialize on different sub-tasks. The resulting modular architecture is able to detect salient targets in the scene and perform precise binocular saccadic and vergence movement on it. The performances of the proposed approach has been tested on the iCub Simulator, providing a quantitative evaluation of the computational potentiality of the learned sensory and motor resources.
</div>
</ul>
<br>
M. W. Spratling (2017) <b>A review of predictive coding algorithms.</b> <I><A HREF =
"http://dx.doi.org/10.1016/j.bandc.2015.11.003">Brain and Cognition</A></I>, 112: 92-7.
<ul>
<A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/taxonomy_pc.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">Predictive coding is a
leading theory of how the brain performs probabilistic inference. However, there
are a number of distinct algorithms which are described by the term "predictive
coding". This article provides a concise review of these different predictive
coding algorithms, highlighting their similarities and differences. Five
algorithms are covered: linear predictive coding which has a long and
influential history in the signal processing literature; the first
neuroscience-related application of predictive coding to explaining the function
of the retina; and three versions of predictive coding that have been proposed
to model cortical function. While all these algorithms aim to fit a generative
model to sensory data, they differ in the type of generative model they employ,
in the process used to optimise the fit between the model and sensory data, and
in the way that they are related to neurobiology.</div>
</ul>
<br>
W. Muhammad and M. W. Spratling (2017) <b>A neural model of coordinated head and eye movement control.</b> <I><A HREF = "http://dx.doi.org/10.1007/s10846-016-0410-8">Journal of Intelligent & Robotic Systems</A></I>, 85(1):107-26.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_head_control.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">Gaze shifts require the coordinated movement of both the eyes and the head in both animals and humanoid robots. To achieve this the brain and the robot control system needs to be able to perform complex non-linear sensory-motor transformations between many degrees of freedom and resolve the redundancy in such a system. In this article we propose a hierarchical neural network model for performing 3-D coordinated gaze shifts. The network is based on the PC/BC-DIM (Predictive Coding/Biased Competition with Divisive Input Modulation) basis function model. The proposed model consists of independent eyes and head controlled circuits with mutual interactions for the appropriate adjustment of coordination behaviour. Based on the initial eyes and head positions the network resolves redundancies involved in 3-D gaze shifts and produces accurate gaze control without any kinematic analysis or imposing any constraints. Furthermore the behaviour of the proposed model is consistent with coordinated eye and head movements observed in primates.</div>
</ul>
<br>
Q. Wang and M. W. Spratling (2016) <b>Contour detection in colour images using a neurophysiologically inspired model.</b> <I><A HREF = "http://dx.doi.org/10.1007/s12559-016-9432-6">Cognitive Computation</A></I>, 8(6):1027-35.
<ul>
<A class="download" HREF = "http://rdcu.be/kDd8">PDF</A><div class="download">Abstract</div><div class="abstract">
The predictive coding/biased competition (PC/BC) model of V1 has previously been applied to locate boundaries defined by local discontinuities in intensity within an image. Here it is extended to perform contour detection for colour images. The proposed extensions are inspired by neurophysiological data from single neurons in macaque primary visual cortex (V1), and the behaviour of this extended model is consistent with the neurophysiological experimental results. Furthermore, when compared to methods used for contour detection in computer vision, the colour PC/BC model of V1 slightly outperforms some recently proposed algorithms which use more cues and/or require a complicated training procedure.
</ul>
<br>
M. W. Spratling (2016) <b>A neural implementation of Bayesian inference based on predictive coding.</b> <I><A HREF =
"http://dx.doi.org/10.1080/09540091.2016.1243655">Connection Science</A></I>, 28(4):346-83.
<ul>
<A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_prob.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/ProbabilisticInference">Code</A><div
class="download">Abstract</div><div class="abstract">
Predictive coding is a leading theory of cortical function that
has previously been shown to explain a great deal of neurophysiological and
psychophysical data. Here it is shown that predictive coding can perform almost
exact Bayesian inference when applied to computing with population codes. It is
demonstrated that the proposed algorithm, based on predictive coding, can:
decode probability distributions encoded as noisy population codes; combine
priors with likelihoods to calculate posteriors; perform cue integration and cue
segregation; perform function approximation; be extended to perform hierarchical
inference; simultaneously represent and reason about multiple stimuli; and
perform inference with multi-modal and non-Gaussian probability distributions.
Predictive coding thus provides a neural network based method for performing
probabilistic computation and provides a simple, yet comprehensive, theory of
how the cerebral cortex performs Bayesian inference.
</div>
</ul>
<br>
M. W. Spratling (2016) <b>Predictive coding as a model of cognition.</b> <I><A HREF =
"http://dx.doi.org/10.1007/s10339-016-0765-6">Cognitive Processing</A></I>, 17(3): 279-305.
<ul>
<A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_cognition.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/Cognition">Code</A><div
class="download">Abstract</div><div class="abstract">
Previous work has shown that predictive coding can provide a detailed
explanation of a very wide range of low-level perceptual processes. It is also
widely believed that predictive coding can account for high-level, cognitive,
abilities. This article provides support for this view by showing that
predictive coding can simulate phenomena such as categorisation, the influence
of abstract knowledge on perception, recall and reasoning about conceptual
knowledge, context-dependent behavioural control, and naive physics. The
particular implementation of predictive coding used here (PC/BC-DIM) has previously
been used to simulate low-level perceptual behaviour and the neural mechanisms
that underlie them. This algorithm thus provides a single framework for
modelling both perceptual and cognitive brain function.
</div>
</ul>
<br>
M. W. Spratling (2016) <b>A neural implementation of the Hough transform and the advantages of explaining away.</b> <I><A HREF ="http://dx.doi.org/10.1016/j.imavis.2016.05.001">Image and Vision Computing</A></I>, 52:15-24.
<ul>
<A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_hough.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/HoughTransform">Code</A><div
class="download">Abstract</div><div class="abstract">
The Hough Transform (HT) is widely used for feature extraction and object
detection. However, during the HT individual image elements vote for many
possible parameter values. This results in a dense accumulator array and
problems identifying the parameter values that correspond to image
features. This article proposes a new method for implementing the voting process
in the HT. This method employs a competitive neural network algorithm to perform
a form of probabilistic inference known as "explaining away". This results in a
sparse accumulator array in which the parameter values of image features can be
more accurately identified. The proposed method is initially demonstrated using
the simple, prototypical, task of straight line detection in synthetic
images. In this task it is shown to more accurately identify straight lines, and
the parameter of those lines, compared to the standard Hough voting process. The
proposed method is further assessed using a version of the implicit shape model
(ISM) algorithm applied to car detection in natural images. In this application
it is shown to more accurately identify cars, compared to using the standard
Hough voting process in the same algorithm, and compared to the original ISM
algorithm.
</div>
</ul>
<br>
Q. Wang and M. W. Spratling (2016) <b>A simplified texture gradient method for improved image segmentation.</b> <I><A HREF = "http://dx.doi.org/10.1007/s11760-015-0794-0">Signal, Image and Video Processing</A></I>, 10(4):679-86.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/wang_sivpro15.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
Inspired by the probability of boundary (Pb) algorithm, a simplified texture gradient method has been developed to locate texture boundaries within grayscale images. Despite considerable simplification, the proposed algorithm's ability to locate texture boundaries is comparable with Pb's texture boundary method. The proposed texture gradient method is also integrated with a biologically inspired model, to enable boundaries defined by discontinuities in both intensity and texture to be located. The combined algorithm outperforms the current state-of-art image segmentation method (Pb) when this method is also restricted to using only local cues of intensity and texture at a single scale.</div>
</ul>
<br>W. Muhammad and M. W. Spratling (2015) <b>A neural model of binocular saccade planning and
vergence control.</b> <I><A HREF = "http://dx.doi.org/10.1177/1059712315607363">Adaptive Behavior</A></I>, 23(5):265-82.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_eye_control.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract"> The human visual system
uses saccadic and vergence eye movements to foveate visual targets. To mimic
this aspect of the biological visual system the PC/BC-DIM neural network is used
as an omni-directional basis function network for learning and performing
sensory-sensory and sensory-motor transformations without using any hard-coded
geometric information. A hierarchical PC/BC-DIM network is used to learn a
head-centred representation of visual targets by dividing the whole problem into
independent subtasks. The learnt head- centred representation is then used to
generate saccade and vergence motor commands. The performance of the proposed
system is tested using the iCub humanoid robot simulator.</div>
</ul>
<br>
M. W. Spratling (2014) <b>Classification using sparse
representations: a biologically plausible approach.</b> <I><A HREF =
"http://dx.doi.org/10.1007/s00422-013-0579-x">Biological Cybernetics</A></I>,
108(1):61-73.
<ul>
<A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/sparse_classification.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/SparseClassification">Code</A><div
class="download">Abstract</div><div class="abstract"> Representing signals as
linear combinations of basis vectors sparsely selected from an overcomplete
dictionary has proven to be advantageous for many applications in pattern
recognition, machine learning, signal processing, and computer vision. While
this approach was originally inspired by insights into cortical information
processing, biologically-plausible approaches have been limited to exploring the
functionality of early sensory processing in the brain, while more practical
application have employed non-biologically-plausible sparse-coding
algorithms. Here, a biologically-plausible algorithm is proposed that can be
applied to practical problems. This algorithm is evaluated using standard
benchmark tasks in the domain of pattern classification, and its performance is
compared to a wide range of alternative algorithms that are widely used in
signal and image processing. The results show that, for the classification
tasks performed here, the proposed method is very competitive with the best of
the alternative algorithms that have been evaluated. This demonstrates that
classification using sparse representations can be performed in a
neurally-plausible manner, and hence, that this mechanism of classification
might be exploited by the brain. </div>
</ul>
<br>M. W. Spratling (2014) <b>A single functional model of drivers and modulators in cortex.</b> <I><A HREF = "http://dx.doi.org/10.1007/s10827-013-0471-7">Journal of Computational Neuroscience</A></I>, 36(1): 97-118.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/drivers_modulators.pdf">PDF</A><A class="download" HREF = "https://codeberg.org/mwspratling/DriversModulators">Code</A><div class="download">Abstract</div><div class="abstract">
A distinction is commonly made between synaptic connections capable of evoking a
response ("drivers") and those that can alter ongoing activity but not
initiate it ("modulators"). Here it is proposed that, in cortex, both drivers
and modulators are an emergent property of the perceptual inference performed by
cortical circuits. Hence, it is proposed that there is a single underlying
computational explanation for both forms of synaptic connection. This idea is
illustrated using a predictive coding model of cortical perceptual inference.
In this model all synaptic inputs are treated identically. However,
functionally, certain synaptic inputs drive neural responses while others have a
modulatory influence. This model is shown to account for driving and modulatory
influences in bottom-up, lateral, and top-down pathways, and is used to simulate
a wide range of neurophysiological phenomena including surround suppression,
contour integration, gain modulation, spatio-temporal prediction, and attention.
The proposed computational model thus provides a single functional explanation
for drivers and modulators and a unified account of a diverse range of
neurophysiological data. </div>
</ul>
<br>M. W. Spratling (2013) <b>Predictive coding.</b> In <I><A HREF = "http://dx.doi.org/10.1007/978-1-4614-7320-6_509-6">Encyclopedia of Computational Neuroscience,</A></I> D. Jaeger and R. Jung (Eds.), Springer, New York.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pc_encyc.pdf">PDF</A>
</ul>
<br>M. W. Spratling (2013) <B>Distinguishing theory from implementation in predictive coding accounts of brain function</B> [commentary]. <I><A HREF = "http://dx.doi.org/10.1017/S0140525X12002178">Behavioral and Brain Sciences</A></I>, 36(3):231-2.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/commentary_clark.pdf">PDF</A>
</ul>
<br>M. W. Spratling (2013) <b>Image segmentation using a sparse coding model of cortical area V1.</b> <I><A HREF = "http://dx.doi.org/10.1109/TIP.2012.2235850">IEEE Transactions on Image Processing</A></I>, 22(4):1631-43.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/edge_detection.pdf">PDF</A><A class="download" HREF = "https://codeberg.org/mwspratling/EdgeDetection">Code</A><div class="download">Abstract</div><div class="abstract">
Algorithms that encode images using a sparse set of basis
functions have previously been shown to explain aspects of the physiology of
primary visual cortex (V1), and have been used for applications such as image
compression, restoration, and classification. Here, a sparse coding algorithm,
that has previously been used to account of the response properties of
orientation tuned cells in primary visual cortex, is applied to the task of
perceptually salient boundary detection. The proposed algorithm is currently
limited to using only intensity information at a single scale. However, it is
shown to out-perform the current state-of-the-art image segmentation method (Pb)
when this method is also restricted to using the same information. </div>
</ul>
<br>K. De Meyer and M. W. Spratling (2013) <b>A model of partial reference frame transforms through pooling of gain-modulated responses.</b> <I><A HREF =
"http://dx.doi.org/10.1093/cercor/bhs117">Cerebral Cortex</A></I>, 23(5):1230-9.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/partial_transforms.pdf">PDF</A><A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Code/partial_transforms.zip">Code</A><div class="download">Abstract</div><div class="abstract">
In multimodal integration and sensorimotor transformation areas of posterior
parietal cortex (PPC), neural responses often appear encoded in spatial
reference frames that are intermediate to intrinsic sensory reference frames,
e.g., eye-centred for visual or head-centred for auditory stimulation. Many
sensory responses in these areas are also modulated by direction of gaze. We
demonstrate that certain types of mixed-frame responses can be generated by
pooling gain-modulated responses - similarly to how complex cells in visual
cortex are thought to pool the responses of simple cells. The proposed model
simulates two types of mixed-frame responses observed in PPC: in particular,
sensory responses that shift differentially with gaze in horizontal and vertical
dimensions; and sensory responses that shift differentially for different start
and end points along a single dimension of gaze. We distinguish these two types
of mixed-frame responses from a third type in which sensory responses shift a
partial yet approximately equal amount with each gaze shift. We argue that the
empirical data on mixed-frame responses may be caused by multiple mechanisms,
and we adapt existing reference-frame measures to distinguish between the
different types. Finally, we discuss how mixed-frame responses may be revealing
of the local organisation of presynaptic responses.
</div>
</ul>
<br>M. W. Spratling (2012) <B>Predictive coding accounts for V1
response properties recorded using reverse correlation.</B> <I><A HREF =
"http://dx.doi.org/10.1007/s00422-012-0477-7">Biological Cybernetics</A></I>, 106(1):37-49.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_reverse_correl.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/V1_ReverseCorrelation">Code</A><div
class="download">Abstract</div><div class="abstract">
PC/BC ("Predictive Coding/Biased Competition") is a simple computational model that has previously been shown to explain a very wide range of V1 response properties. This article extends work on the PC/BC model of V1 by showing that it can also account for V1 response properties measured using the reverse correlation methodology. Reverse correlation employs an experimental procedure that is significantly different from that used in more typical neurophysiological experiments, and measures some distinctly different response properties in V1. Despite these differences PC/BC successfully accounts for the data. The current results thus provide additional support for the PC/BC model of V1 and further demonstrate that PC/BC offers a unified explanation for the seemingly diverse range of behaviours observed in primary visual cortex.
</div>
</ul>
<br>M. W. Spratling (2012) <B>Predictive coding as a model of the V1
saliency map hypothesis.</B> <I><A HREF = "http://dx.doi.org/10.1016/j.neunet.2011.10.002">Neural Networks</A></I>, 26:7-28.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_saliency.pdf">PDF</A><A
class="download" HREF = "https://codeberg.org/mwspratling/V1_Saliency">Code</A><div
class="download">Abstract</div><div class="abstract">
The predictive coding/biased competition (PC/BC) model is a specific implementation of predictive coding theory that has previously been shown to provide a detailed account of the response properties of orientation tuned cells in primary visual cortex (V1). Here it is shown that the same model can successfully simulate psychophysical data relating to the saliency of unique items in search arrays, of contours embedded in random texture, and of borders between textured regions. This model thus provides a possible implementation of the hypothesis that V1 generates a bottom-up saliency map. However, PC/BC is very different from previous models of visual salience, in that it proposes that saliency results from the failure of an internal model of simple elementary image components to accurately predict the visual input. Saliency can therefore be interpreted as a mechanism by which prediction errors attract attention in an attempt to improve the accuracy of the brain's internal representation of the world.
</div>
</ul>
<br>M. W. Spratling (2012) <B>Unsupervised learning of generative and discriminative
weights encoding elementary image components in a predictive coding model of
cortical function.</B> <A HREF = "http://dx.doi.org/10.1162/NECO_a_00222">
<I>Neural Computation</I></A>, 24(1): 60-103.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_learn_recip_weights.pdf">PDF</A><A
class="download" HREF = "https://codeberg.org/mwspratling/LearnReciprocalWeights">Code</A><div
class="download">Abstract</div><div class="abstract">
A method is presented for learning the reciprocal feedforward and feedback
connections required by the predictive coding model of cortical
function. Using this method feedforward and feedback connections are learnt
simultaneously and independently in a biologically plausible manner. The
performance of the proposed algorithm is evaluated by applying it to learning
the elementary components of artificial images and of natural images. For
artificial images the bars problem is employed and the proposed algorithm is
shown to produce state-of-the-art performance on this task. For natural
images, components resembling Gabor functions are learnt in the first
processing stage and neurons responsive to corners are learnt in the second
processing stage. The properties of these learnt representations are in good
agreement with neurophysiological data from V1 and V2. The proposed algorithm
demonstrates for the first time that a single computational theory can explain
the formation of cortical RFs, and also the response properties of cortical
neurons once those RFs have been learnt.
</div>
</ul>
<br>K. De Meyer and M. W. Spratling (2011) <b>Multiplicative gain
modulation arises through unsupervised learning in a predictive coding model of
cortical function.</b> <A HREF="http://dx.doi.org/10.1162/NECO_a_00130"><I>Neural Computation</I></A>, 23(6):1536-67.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/gain_modulation.pdf">PDF</A><A
class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Code/gain_modulation.zip">Code</A><div
class="download">Abstract</div><div class="abstract">
The combination of two or more population-coded signals in a
neural model of predictive coding can give rise to multiplicative
gain modulation in the response properties of individual neurons.
Synaptic weights generating these multiplicative response
properties can be learned using an unsupervised, Hebbian, learning
rule. The behaviour of the model is compared to empirical data on
gaze-dependent gain modulation of cortical cells, and found to be
in good agreement with a range of physiological observations.
Furthermore, it is demonstrated that the model can learn to
represent a set of basis functions. The current paper thus
connects an often-observed neurophysiological phenomenon and
important neurocomputational principle (gain modulation) with an
influential theory of brain operation (predictive coding).
</div>
</ul>
<br>M. W. Spratling (2011) <B>A single functional model accounts for the
distinct properties of suppression in cortical area V1.</B> <A
HREF="http://dx.doi.org/10.1016/j.visres.2011.01.017"><I>Vision
Research</I></A>, 51(6):563-76.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_suppression_mechanisms.pdf">PDF</A><A
class="download" HREF = "https://codeberg.org/mwspratling/V1_SuppressionMechanisms">Code</A><div
class="download">Abstract</div><div class="abstract">
Cross-orientation suppression and surround suppression have been extensively
studied in primary visual cortex (V1). These two forms of suppression have some
distinct properties which has led to the suggestion that they are generated by
different underlying mechanisms. Furthermore, it has been suggested that
mechanisms other than intracortical inhibition may be central to both forms of
suppression. A simple computational model (PC/BC), in which intracortical
inhibition is fundamental, is shown to simulate the distinct properties of
cross-orientation and surround suppression. The same model has previously been
shown to account for a large range of V1 response properties including
orientation-tuning, spatial and temporal frequency tuning, facilitation and
inhibition by flankers and textured surrounds as well as a range of other
experimental results on cross-orientation suppression and surround
suppression. The current results thus provide additional support for the PC/BC
model of V1 and for the proposal that the diverse range of response properties
observed in V1 neurons have a single computational explanation. Furthermore,
these results demonstrate that current neurophysiological evidence is
insufficient to discount intracortical inhibition as a central mechanism
underlying both forms of suppression. </div>
</ul>
<br>M. W. Spratling (2010) <B>Predictive coding as a model of response
properties in cortical area V1.</B> <I><A HREF =
"http://dx.doi.org/10.1523/JNEUROSCI.4911-09.2010">Journal of
Neuroscience</A></I>, 30(9):3531-43.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_response_properties.pdf">PDF</A><A
class="download" HREF = "https://codeberg.org/mwspratling/V1_ResponseProperties">Code</A><div
class="download">Abstract</div><div class="abstract">
A simple model is shown to account for a large range of V1 classical, and
non-classical, receptive field properties including orientation-tuning,
spatial and temporal frequency tuning, cross-orientation suppression, surround
suppression, and facilitation and inhibition by flankers and textured
surrounds. The model is an implementation of the predictive coding theory of
cortical function and thus provides a single computational explanation for a
diverse range of neurophysiological findings. Furthermore, since predictive
coding can be related to the biased competition theory and is a specific
example of more general theories of hierarchical perceptual inference the
current results relate V1 response properties to a wider, more unified,
framework for understanding cortical function.
</div>
</ul>
<br>M. W. Spratling (2009) <B>Learning posture invariant spatial
representations through temporal correlations.</B> <I><A HREF =
"http://dx.doi.org/10.1109/TAMD.2009.2038494"> IEEE Transactions on Autonomous
Mental Development</A></I>, 1(4):253-63.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/tamd09_preprint.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
A hierarchical neural network model is used to learn, without supervision,
sensory-sensory coordinate transformations like those believed to be encoded
in the dorsal pathway of the cerebral cortex. The resulting representations of
visual space are invariant to eye orientation, neck orientation, or posture in
general. These posture invariant spatial representations are learned using the
same mechanisms that have previously been proposed to operate in the cortical
ventral pathway to learn object representation that are invariant to
translation, scale, orientation, or viewpoint in general. This model thus
suggests that the same mechanisms of learning and development operate across
multiple cortical hierarchies.
</div>
</ul>
<br>K. De Meyer and M. W. Spratling (2009) <B> A model of non-linear
interactions between cortical top-down and horizontal connections explains the
attentional gating of collinear facilitation.</B> <I><A HREF =
"http://dx.doi.org/10.1016/j.visres.2008.12.017">Vision Research</A></I>, 49(5):553-68.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/visres09_preprint.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
Past physiological and psychophysical experiments have shown that attention can
modulate the effects of contextual information appearing outside the classical
receptive field of a cortical neuron. Specifically, it has been suggested that
attention, operating via cortical feedback connections, gates the effects of
long-range horizontal connections underlying collinear facilitation in cortical
area V1. This article proposes a novel mechanism, based on the computations
performed within the dendrites of cortical pyramidal cells, that can account for
these observations. Furthermore, it is shown that the top-down gating signal
into V1 can result from a process of biased competition occurring in
extrastriate cortex. A model based on these two assumptions is used to replicate
the results of physiological and psychophysical experiments on collinear
facilitation and attentional modulation.
</div>
</ul>
<br>M. W. Spratling, K. De Meyer and R. Kompass (2009)
<B>Unsupervised learning of overlapping image components using divisive input modulation.</B>
<I><A HREF = "http://www.hindawi.com/journals/cin/aip.381457.html">
Computational Intelligence and Neuroscience</A></I>, 2009(381457):1-19.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/DivisiveInputModulation">Code</A><div class="download">Abstract</div><div
class="abstract">
This paper demonstrates that non-negative matrix factorisation is mathematically
related to a class of neural networks that employ negative feedback as a
mechanism of competition. This observation inspires a novel learning algorithm
which we call Divisive Input Modulation (DIM). The proposed algorithm provides a
mathematically simple and computationally efficient method for the unsupervised
learning of image components, even in conditions where these elementary features
overlap considerably. To test the proposed algorithm, a novel artificial task is
introduced which is similar to the frequently-used bars problem but employs
squares rather than bars to increase the degree of overlap between
components. Using this task, we investigate how the proposed method performs on
the parsing of artificial images composed of overlapping features, given the
correct representation of the individual components; and secondly, we
investigate how well it can learn the elementary components from artificial
training images. We compare the performance of the proposed algorithm with its
predecessors including variations on these algorithms that have produced
state-of-the-art performance on the bars problem. The proposed algorithm is more
successful than its predecessors in dealing with overlap and occlusion in the
artificial task that has been used to assess performance.
</div>
</ul>
<br>M. W. Spratling (2008)
<B>Predictive coding as a model of biased competition in visual attention.</B>
<I><A HREF = "http://dx.doi.org/10.1016/j.visres.2008.03.009">Vision Research</A></I>, 48(12):1391-408.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/visres08.pdf">PDF</A><A class="download" HREF =
"https://codeberg.org/mwspratling/PredictiveCodingAttention">Code</A><div class="download">Abstract</div><div
class="abstract">
Attention acts, through cortical feedback pathways, to enhance the response of
cells encoding expected or predicted information. Such observations are
inconsistent with the predictive coding theory of cortical function which
proposes that feedback acts to suppress information predicted by higher-level
cortical regions. Despite this discrepancy, this article demonstrates that the
predictive coding model can be used to simulate a number of the effects of
attention. This is achieved via a simple mathematical rearrangement of the
predictive coding model, which allows it to be interpreted as a form of biased
competition model. Nonlinear extensions to the model are proposed that enable it
to explain a wider range of data.
</div>
</ul>
<br>M. W. Spratling (2008)
<B>Reconciling predictive coding and biased competition models of cortical function.</B>
<I><A HREF = "http://www.frontiersin.org/Journal/Abstract.aspx?s=237&name=Computational_Neuroscience&ART_DOI=10.3389/neuro.10.004.2008">Frontiers in Computational Neuroscience</A></I>, 2(4):1-8.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pc_equals_bc.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
A simple variation of the standard biased competition model is shown, via some
trivial mathematical manipulations, to be identical to predictive coding.
Specifically, it is shown that a particular implementation of the biased
competition model, in which nodes compete via inhibition that targets the inputs
to a cortical region, is mathematically equivalent to the linear predictive
coding model. This observation demonstrates that these two important and
influential rival theories of cortical function are minor variations on the same
underlying mathematical model.
</div>
</ul>
<br>M. S. C. Thomas, G. Westermann, D. Mareschal M. H. Johnson, S. Siros and M. W. Spratling (2008)
<B>Studying development in the 21st century</B> [response to commentaries].
<I><A HREF = "http://dx.doi.org/10.1017/S0140525X08004202">Behavioral and Brain Sciences</A></I>, 31(3):345-56.
<ul><div class="download">Abstract</div><div class="abstract">
In this response, we consider four main issues arising
from the commentaries to the target article. These include further
details of the theory of interactive specialization, the relationship
between neuroconstructivism and selectionism, the implications
of neuroconstructivism for the notion of representation, and the
role of genetics in theories of development. We conclude by
stressing the importance of multidisciplinary approaches in the
future study of cognitive development and by identifying
the directions in which neuroconstructivism can expand in the
Twenty-first Century.
</div>
</ul>
<br>S. Siros, M. W. Spratling, M. S. C. Thomas, G. Westermann, D. Mareschal and
M. H. Johnson (2008)
<B>Pr&eacute;cis of Neuroconstructivism: how the brain constructs cognition.</B>
<I><A HREF = "http://dx.doi.org/10.1017/S0140525X0800407X">Behavioral and Brain Sciences</A></I>, 31(3):321-31.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/bbs08_preprint.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
Neuroconstructivism proposes a unifying framework for the study of development
that brings together (1) constructivism (which views development as the
progressive elaboration of increasingly complex structures), (2) cognitive
neuroscience (which aims to understand the neural mechanisms underlying
behaviour), and (3) computational modelling (which proposes formal and explicit
specifications of information processing). The guiding principle of our approach
is context dependence, within and (in contrast to Marr) between levels of
organization. We propose that three mechanisms guide the emergence of
representations: competition, cooperation, and chronotopy, which themselves
allow for two central processes: proactivity and progressive specialization. We
suggest that the main outcome of development is partial representations,
distributed across distinct functional circuits. This framework is derived by
examining development at the level of single neurons, brain systems, and whole
organisms. We use the terms encellment, embrainment, and embodiment to describe
the higher-level contextual influences that act at each of these levels of
organization. To illustrate these mechanisms in operation we provide case
studies in early visual perception, infant habituation, phonological
development, and object representations in infancy. Three further case studies
are concerned with interactions between levels of explanation: social
development, atypical development and within that, the development of
dyslexia. We conclude that cognitive development arises from a dynamic,
contextual change in neural structures leading to partial representations across
multiple brain regions and timescales.
</div>
</ul>
<br>X. Zhang and M. W. Spratling (2008) <B>Automated learning of coordinate
transformations.</B> <I>Proceedings of the Eighth International Conference on <A
HREF = "http://www.lucs.lu.se/epirob/proceedings-epigenetic-robotics-2008/">Epigenetic Robotics: Modeling
Cognitive Development in Robotic Systems</A> (EPIROB08).</I>
<br><br>G. Westermann, D. Mareschal, M. H. Johnson, S. Siros, M. W. Spratling and M. S. C. Thomas (2007)
<B>Neuroconstructivism.</B><I> <A HREF = "http://dx.doi.org/10.1111/j.1467-7687.2007.00567.x">Developmental Science</A></I>, 10(1):75-83.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dev_sci07.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
Neuroconstructivism is a theoretical framework focusing on the construction of
representation in the developing brain. Cognitive development is explained as
emerging from the experience-dependent development of neural structures
supporting mental representations. Neural development occurs in the context of
multiple interacting constraints acting on different levels, from the individual
cell to the external environment of the developing child. Cognitive development
can thus be understood as a trajectory originating from the constraints on the
underlying neural structures. This perspective offers an integrated view of
normal and abnormal development as well as of development and adult processing,
and it stands apart from traditional cognitive approaches in taking seriously
the constraints on cognition inherent by the substrate that delivers it.
</div>
</ul>
<br>L. A. Watling, M. W. Spratling, K. De Meyer and M. Johnson
(2007) <B>The role of feedback in the determination of figure and ground: a
combined behavioral and modeling study.</B> <I>Proceedings of the 29th Meeting of
the <A HREF = "http://www.cognitivesciencesociety.org/">Cognitive Science
Society</A> (CogSci07).</I>
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/watling_cogsci07.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
Object knowledge can exert on important influence on even the earliest stages of
visual processing. This study demonstrates how a familiarity bias, acquired only
briefly before testing, can affect the subsequent segmentation of an otherwise
ambiguous figure-ground array, in favor of perceiving the familiar shape as
figure. The behavioral data are then replicated using a biologically plausible
neural network model that employs feedback connections to implement the
demonstrated familiarity bias.
</div>
</ul>
<br>D. Mareschal, M. H. Johnson, S. Siros, M. W. Spratling, M. S. C. Thomas
and G. Westermann (2007) <B><I><A HREF =
"http://www.oup.com/uk/catalogue/?ci=9780198529910"> Neuroconstructivism: How
the Brain Constructs Cognition</A></I></B>, Oxford University Press: Oxford,
UK.
<br><br>M. W. Spratling (2006)
<B>Learning image components for object recognition.</B>
<I><A HREF = "http://jmlr.csail.mit.edu/papers/v7/spratling06a.html">
Journal of Machine Learning Research</A></I>, 7:793-815.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/jmlr06.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
In order to perform object recognition it is necessary to learn representations
of the underlying components of images. Such components correspond to objects,
object-parts, or features. Non-negative matrix factorisation is a generative
model that has been specifically proposed for finding such meaningful
representations of image data, through the use of non-negativity constraints on
the factors. This article reports on an empirical investigation of the
performance of non-negative matrix factorisation algorithms. It is found that
such algorithms need to impose additional constraints on the sparseness of the
factors in order to successfully deal with occlusion. However, these constraints
can themselves result in these algorithms failing to identify image components
under certain conditions. In contrast, a recognition model (a competitive
learning neural network algorithm) reliably and accurately learns
representations of elementary image features without such constraints.
</div>
</ul>
<br>M. W. Spratling and M. H. Johnson (2006)
<B>A feedback model of perceptual learning and categorisation.</B>
<I><A HREF = "http://dx.doi.org/10.1080/13506280500168562">
Visual Cognition</A></I>, 13(2):129-65.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/vis_cog06.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
Top-down, feedback, influences are known to have significant effects on visual
information processing. Such influences are also likely to affect perceptual
learning. This article employs a computational model of the cortical region
interactions underlying visual perception to investigate possible influences of
top-down information on learning. The results suggest that feedback could bias
the way in which perceptual stimuli are categorised and could also facilitate
the learning of sub-ordinate level representations suitable for object
identification and perceptual expertise.
</div>
</ul>
<br>M. W. Spratling (2005)
<B>Learning viewpoint invariant perceptual representations from cluttered images.</B>
<I><A HREF = "http://doi.ieeecomputersociety.org/10.1109/TPAMI.2005.105">
IEEE Transactions on Pattern Analysis and Machine Intelligence</A></I>, 27(5):753-61.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/tpami05.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
In order to perform object recognition, it is necessary to form perceptual
representations that are sufficiently specific to distinguish between objects,
but that are also sufficiently flexible to generalise across changes in
location, rotation and scale. A standard method for learning perceptual
representations that are invariant to viewpoint is to form temporal associations
across image sequences showing object transformations. However, this method
requires that individual stimuli are presented in isolation and is therefore
unlikely to succeed in real-world applications where multiple objects can
co-occur in the visual input. This article proposes a simple modification to the
learning method, that can overcome this limitation, and results in more robust
learning of invariant representations.
</div>
</ul>
<br>M. W. Spratling (2004)
<B>Local versus distributed: a poor taxonomy of neural coding strategies</B> [commentary].
<I><A HREF="http://journals.cambridge.org/action/displayAbstract?aid=287732">
Behavioral and Brain Sciences</A></I>, 27(5):700-2.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/bbs_page.pdf">PDF</A>
</ul>
<br>M. W. Spratling and M. H. Johnson (2004)
<B>Neural coding strategies and mechanisms of competition.</B>
<I><A HREF = "http://dx.doi.org/10.1016/j.cogsys.2003.11.002">Cognitive Systems
Research</A></I>, 5(2):93-117.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cogsysres.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
A long running debate has concerned the question of whether neural
representations are encoded using a distributed or a local coding scheme. In
both schemes individual neurons respond to certain specific patterns of
pre-synaptic activity. Hence, rather than being dichotomous, both coding
schemes are based on the same representational mechanism. We argue that a
population of neurons needs to be capable of learning both local and distributed
representations, as appropriate to the task, and should be capable of generating
both local and distributed codes in response to different stimuli. Many neural
network algorithms, which are often employed as models of cognitive processes,
fail to meet all these requirements. In contrast, we present a neural network
architecture which enables a single algorithm to efficiently learn, and respond
using, both types of coding scheme.
</div>
</ul>
<br>M. W. Spratling and M. H. Johnson (2004)
<B>A feedback model of visual attention.</B>
<I><A HREF="http://dx.doi.org/10.1162/089892904322984526">
Journal of Cognitive Neuroscience</A></I>, 16(2):219-37.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/jocn04.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
Feedback connections are a prominent feature of cortical anatomy and are likely
to have a significant functional role in neural information processing. We
present a neural network model of cortical feedback that successfully simulates
neurophysiological data associated with attention. In this domain our model can
be considered a more detailed, and biologically plausible, implementation of the
biased competition model of attention. However, our model is more general as it
can also explain a variety of other top-down processes in vision, such as
figure/ground segmentation and contextual cueing. This model thus suggests that
a common mechanism, involving cortical feedback pathways, is responsible for a
range of phenomena and provides a unified account of currently disparate areas
of research.
</div>
</ul>
<!--</ul><br>R. Fukuda, M. W. Spratling, D. Mareschal and M. Johnson
(2003) <B>Modelling cortico basal-ganglionic loops and the development of
sequential information encoding.</B> <I>Proceedings of the 3rd International
Workshop on <A HREF = "http://www.epigenetic-robotics.org">Epigenetic
Robotics: Modeling Cognitive Development in Robotic Systems</A> (EPIROB03).</I>-->
<br>M. W. Spratling and M. H. Johnson (2003)
<B>Exploring the functional significance of dendritic inhibition in cortical pyramidal cells.</B>
<I><A HREF = "http://dx.doi.org/10.1016/S0925-2312(02)00847-0">Neurocomputing</A></I>, 52-54:389-95.
<!--Proceedings of the 11th <A HREF ="http://www.neuroinf.org/CNS.shtml">Computational Neuroscience Meeting</A> (CNS02)</I>.-->
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cns02.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
Inhibitory synapses contacting the soma and axon initial segment are commonly
presumed to participate in shaping the response properties of cortical pyramidal
cells. Such an inhibitory mechanism has been explored in numerous computational
models. However, the majority of inhibitory synapses target the dendrites of
pyramidal cells, and recent physiological data suggests that this dendritic
inhibition affects tuning properties. We describe a model that can be used to
investigate the role of dendritic inhibition in the competition between
neurons. With this model we demonstrate that dendritic inhibition significantly
enhances the computational and representational properties of neural networks.
</div>
</ul>
<br>M. W. Spratling and M. H. Johnson (2002) <B>Exploring the functional
significance of dendritic inhibition in cortical pyramidal cells.</B>
<I>Proceedings of the 11th <A HREF ="http://www.neuroinf.org/CNS.shtml">Computational Neuroscience Meeting</A> (CNS02).</I> (Reprinted in the journal Neurocomputing, 2003; see above)
<br><br>M. W. Spratling (2002)
<B>Cortical region interactions and the functional role of apical dendrites.</B>
<I><A HREF = "http://dx.doi.org/10.1177/1534582302001003003">
Behavioral and Cognitive Neuroscience Reviews</A></I>, 1(3):219-28.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/apical_review.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
The basal and distal apical dendrites of pyramidal cells occupy distinct
cortical layers and are targeted by axons originating in different cortical
regions. Hence, apical and basal dendrites receive information from distinct
sources. Physiological evidence suggests that this anatomically observed
segregation of input sources may have functional significance. This possibility
has been explored in various connectionist models that employ neurons with
functionally distinct apical and basal compartments. A neuron in which separate
sets of inputs can be integrated independently has the potential to operate in a
variety of ways which are not possible for the conventional model of a neuron in
which all inputs are treated equally. This article thus considers how
functionally distinct apical and basal dendrites can contribute to the
information processing capacities of single neurons and, in particular, how
information from different cortical regions could have disparate affects on
neural activity and learning.
</div>
</ul>
<br>M. W. Spratling and M. H. Johnson (2002)
<B>Pre-integration lateral inhibition enhances unsupervised learning.</B>
<I><A HREF = "http://dx.doi.org/10.1162/089976602320264033">Neural
Computation</A></I>, 14(9):2157-79.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/neurocomp.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
A large and influential class of neural network architectures use
post-integration lateral inhibition as a mechanism for competition. We argue
that these algorithms are computationally deficient in that they fail to
generate, or learn, appropriate perceptual representations under certain
circumstances. An alternative neural network architecture is presented in which
nodes compete for the right to receive inputs rather than for the right to
generate outputs. This form of competition, implemented through pre-integration
lateral inhibition, does provide appropriate coding properties and can be used
to efficiently learn such representations. Furthermore, this architecture is
consistent with both neuro-anatomical and neuro-physiological data. We thus
argue that pre-integration lateral inhibition has computational advantages over
conventional neural network architectures while remaining equally biologically
plausible.
</div>
</ul>
<br>M. W. Spratling and M. H. Johnson (2001)
<B>Dendritic inhibition enhances neural coding properties.</B>
<I><A HREF = "https://doi.org/10.1093/cercor/11.12.1144">Cerebral Cortex</A></I>, 11(12):1144-9.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cereb_cort.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
The presence of a large number of inhibitory contacts at the soma and axon
initial segment of cortical pyramidal cells has inspired a large and influential
class of neural network model which use post-integration lateral inhibition as a
mechanism for competition between nodes. However, inhibitory synapses also
target the dendrites of pyramidal cells. The role of this dendritic inhibition
in competition between neurons has not previously been addressed. We
demonstrate, using a simple computational model, that such pre-integration
lateral inhibition provides networks of neurons with useful representational and
computational properties which are not provided by post-integration
inhibition.
</div>
</ul>
<br>S. J. Grice, M. W. Spratling, A. Karmiloff-Smith, H. Halit, G. Csibra, M. de Haan and M. H. Johnson (2001) <B>Disordered visual processing and oscillatory brain activity in autism and Williams Syndrome.</B>
<I><A HREF ="http://www.neuroreport.com/pt/re/neuroreport/abstract.00001756-200108280-00021.htm">NeuroReport</A></I>, 12(12):2697-700.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/disordered_gamma.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
Two developmental disorders, autism and Williams Syndrome, are both commonly
described as having difficulties in integrating perceptual features, i.e.,
binding spatially separate elements into a whole. It is already known that
healthy adults and infants display electroencephalographic (EEG) gamma band
bursts (around 40Hz) when the brain is required to achieve such binding . Here
we explore gamma band EEG in autism and Williams Syndrome and demonstrate
differential abnormalities in the two phenotypes. We show that despite putative
processing similarities at the cognitive level, binding in Williams Syndrome and
autism can be dissociated at the neurophysiological level by different
abnormalities in underlying brain oscillatory activity. Our study is the first
to identify that binding related gamma EEG can be disordered in humans.
</div>
</ul>
<br>M. W. Spratling and M. H. Johnson (2001) <B>Activity-dependent processes in regional cortical specialization</B> [commentary]. <I><A HREF="http://dx.doi.org/10.1111/1467-7687.00159">
Developmental Science</A></I>, 4(2):153-4.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/kingsbury_finlay.html">HTML</A>
</ul>
<br>G. Csibra, G. Davis, M. W. Spratling and M. H. Johnson (2000)
<B>Gamma oscillations and object processing in the infant brain.</B>
<I><A HREF = "http://dx.doi.org/10.1126/science.290.5496.1582">Science</A></I>, 290(5496):1582-5.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/infant_gamma.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
An enduring controversy in neuroscience concerns how the brain binds
together separately coded stimulus features to form unitary
representations of objects. Recent evidence has indicated a close link
between this binding process and 40Hz (gamma-band) oscillations
generated by localized neural circuits (1). In a separate line of
research, the ability of young infants to perceive objects as unitary
and bounded has become a central focus for debates about the
mechanisms of perceptual development (2). However, to date these
infant studies have been behavioural, and there have been few, if any,
paradigms involving direct measures of neural function. Here we
demonstrate for the first time that binding-related 40Hz oscillations
are evident in the infant brain around 8 months of age, the same age
as some behavioral studies indicate the onset of perceptual binding of
spatially separated static visual features. The discovery of
binding-related gamma in infants opens up a new vista for experiments
on postnatal functional brain development in infants.
</div>
</ul>
<br>M. W. Spratling and G. M. Hayes (2000)
<B>Learning synaptic clusters for non-linear dendritic processing.</B>
<I><A HREF = "http://dx.doi.org/10.1023/A:1009634821039">
Neural Processing Letters</A></I>, 11(1):17-27.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/sigma_pi.pdf">PDF</A ><A class="download" HREF =
"https://nms.kcl.ac.uk/michael.spratling/Doc/sigma_pi.ps.gz">Gzipped Postscript</A><div class="download">Abstract</div><div class="abstract">
Nonlinear dendritic processing appears to be a feature of biological neurons and
would also be of use in many applications of artificial neural networks. This
paper presents a model of an initially standard linear unit which uses
unsupervised learning to find clusters of inputs within which inactivity at one
synapse can occlude the activity at the other synapses.
</div>
</ul>
<br>M. W. Spratling (1999)
<B>Artificial Ontogenesis: A Connectionist Model of Development.</B>
PhD Thesis, <A HREF = "http://www.ed.ac.uk/">University of Edinburgh</A>.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/phd.pdf">PDF</A><div
class="download">Abstract</div><div class="abstract">
This thesis suggests that ontogenetic adaptive processes are important for
generating intelligent behaviour. It is thus proposed that such processes, as
they occur in nature, need to be modelled and that such a model could be used
for generating artificial intelligence, and specifically robotic
intelligence. Hence, this thesis focuses on how mechanisms of intelligence are
specified.<br><br>
A major problem in robotics is the need to predefine the behaviour to be
followed by the robot. This makes design intractable for all but the simplest
tasks and results in controllers that are specific to that particular task and
are brittle when faced with unforeseen circumstances. These problems can be
resolved by providing the robot with the ability to adapt the rules it follows
and to autonomously create new rules for controlling behaviour. This solution
thus depends on the predefinition of how rules to control behaviour are to be
learnt rather than the predefinition of rules for behaviour themselves.<br><br>
Learning new rules for behaviour occurs during the developmental process in
biology. Changes in the structure of the cerebral cortex underly behavioural and
cognitive development throughout infancy and beyond. The uniformity of the
neocortex suggests that there is significant computational uniformity across the
cortex resulting from uniform mechanisms of development, and holds out the
possibility of a general model of development. Development is an interactive
process between genetic predefinition and environmental influences. This
interactive process is constructive: qualitatively new behaviours are learnt by
using simple abilities as a basis for learning more complex ones. The
progressive increase in competence, provided by development, may be essential to
make tractable the process of acquiring higher-level abilities.<br><br>
While simple behaviours can be triggered by direct sensory cues, more complex
behaviours require the use of more abstract representations. There is thus a
need to find representations at the correct level of abstraction appropriate to
controlling each ability. In addition, finding the correct level of abstraction
makes tractable the task of associating sensory representations with motor
actions. Hence, finding appropriate representations is important both for
learning behaviours and for controlling behaviours. Representations can be
found by recording regularities in the world or by discovering re-occurring
patterns through repeated sensory-motor interactions. By recording regularities
within the representations thus formed, more abstract representations can be
found. Simple, non-abstract, representations thus provide the basis for learning
more complex, abstract, representations.<br><br>
A modular neural network architecture is presented as a basis for a model of
development. The pattern of activity of the neurons in an individual network
constitutes a representation of the input to that network. This representation
is formed through a novel, unsupervised, learning algorithm which adjusts the
synaptic weights to improve the representation of the input data.
Representations are formed by neurons learning to respond to correlated sets of
inputs. Neurons thus became feature detectors or pattern recognisers. Because
the nodes respond to patterns of inputs they encode more abstract features of
the input than are explicitly encoded in the input data itself. In this way
simple representations provide the basis for learning more complex
representations. The algorithm allows both more abstract representations to be
formed by associating correlated, coincident, features together, and invariant
representations to be formed by associating correlated, sequential, features
together.<br><br>
The algorithm robustly learns accurate and stable representations, in a format
most appropriate to the structure of the input data received: it can represent
both single and multiple input features in both the discrete and continuous
domains, using either topologically or non-topologically organised nodes. The
output of one neural network is used to provide inputs for other networks. The
robustness of the algorithm enables each neural network to be implemented using
an identical algorithm. This allows a modular `assembly' of neural networks to
be used for learning more complex abilities: the output activations of a network
can be used as the input to other networks which can then find representations
of more abstract information within the same input data; and, by defining the
output activations of neurons in certain networks to have behavioural
consequences it is possible to learn sensory-motor associations, to enable
sensory representations to be used to control behaviour.
</div>
</ul>
<br>M. W. Spratling (1999)
<B>Pre-synaptic lateral inhibition provides a better architecture for self-organising neural networks.</B> <I><A HREF = "http://dx.doi.org/10.1088/0954-898X_10_4_301">Network: Computation in Neural Systems</A></I>, 10(4):285-301.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/factopol.pdf">PDF</A ><A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/factopol.ps.gz">Gzipped Postscript</A><div
class="download">Abstract</div><div class="abstract">
Unsupervised learning is an important property of the brain and of
many artificial neural networks. A large variety of unsupervised
learning algorithms have been proposed. This paper takes a different
approach in considering the architecture of the neural network rather
than the learning algorithm. It is shown that a self-organising neural
network architecture using pre-synaptic lateral inhibition enables a
single learning algorithm to find distributed, local, and topological
representations as appropriate to the structure of the input data
received. It is argued that such an architecture not only has
computational advantages but is a better model of cortical
self-organisation.
</div>
</ul>
<br>M. W. Spratling and G. M. Hayes (1998)
<B>Learning sensory-motor cortical mappings without training.</B>
<I>Proceedings of the <A HREF = "https://www.elen.ucl.ac.be/esann/proceedings/papers.php?ann=1998#ES1998-20">6th European Symposium on Artificial Neural Networks (ESANN)</A></I>.
M. Verleysen (ed.) pp. 339-44. D-facto Publications.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cort_map.ps.gz">Gzipped Postscript</A><div
class="download">Abstract</div><div class="abstract">
This paper shows how the relationship between two arrays of artificial
neurons, representing different cortical regions, can be learned. The
algorithm enables each neural network to self-organise into a topological map
of the domain it represents at the same time as the relationship between
these maps is found. Unlike previous methods learning is achieved without a
separate training phase; the algorithm which learns the mapping is also that
which performs the mapping.
</div>
</ul>
<br>M. W. Spratling and G. M. Hayes (1998)
<B>A self-organising neural network for modelling cortical development.</B>
<I>Proceedings of the <A HREF = "https://www.elen.ucl.ac.be/esann/proceedings/papers.php?ann=1998#ES1998-19">6th European Symposium on Artificial Neural Networks (ESANN)</A></I>.
M. Verleysen (ed.) pp. 333-8. D-facto Publications.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/map_alg.ps.gz">Gzipped Postscript</A><div
class="download">Abstract</div><div class="abstract">
This paper presents a novel self-organising neural network. It has been
developed for use as a simplified model of cortical development. Unlike
many other models of topological map formation all synaptic weights start at
zero strength (so that synaptogenesis might be modelled). In addition, the
algorithm works with the same format of encoding for both inputs to and
outputs from the network (so that the transfer and recoding of information
between cortical regions might be modelled).
</div>
</ul>
<br>M. W. Spratling (1997)
<B>Artificial Ontogenesis: Cognitive and Behavioural Development for Robots.</B>
Unpublished Departmental Discussion Paper,
<A HREF = "http://www.dai.ed.ac.uk/">Department of Artificial Intelligence</A>,
<A HREF = "http://www.ed.ac.uk/">University of Edinburgh</A>.
<ul>
<div class="download">Abstract</div><div class="abstract">
There are three classes of adaptive process (structural definition,
structural adjustment, and parameter adjustment) which appear to underly the
development of intelligence in nature. In artificial intelligence only two
of these processes are used; AI ignores development (structural adjustment).
While AI attempts to predefine explicit rules for behaviour, nature's
success in building complex creatures depends on predefining how rules to
control behaviour can be learned. It is the developmental processes in
biology through which such rules are learned. This proposal is to apply
mechanisms similar to those used in biological development to robots. This
will move robotics from `development' meaning design and production, towards
`development' in its biological sense meaning a process of growth and
progressive change. Defining the rules for development is design at a
meta-level to that currently used. It is proposed that the long process of
evolution used by nature to define these developmental processes might be
supplanted by another adaptive process, that of engineering, to more quickly
enable study of ontogenetic development.<br><br>
This project thus aims to apply techniques inspired by animal development to
engineering robot control systems. Specifically it is proposed that a
hierarchical control system, based on the cerebral cortex, is used and that
this develops through constructivist learning algorithms (ones in which the
interaction of a situated agent with its environment guides the creation of
cognitive machinery appropriate for representing and acting in that
environment). Such a robot would be provided with some innate, low-level,
behavioural abilities and through experience develop more complex behaviour.
</div>
</ul>
<br>M. W. Spratling and R. Cipolla (1996)
<B>Uncalibrated visual servoing.</B>
<I>Proceedings of the <A HREF = "http://dx.doi.org/10.5244/C.10.52">
7th British Machine Vision Conference (BMVC)</A></I>.
R. B. Fisher and E. Trucco (eds.) pp. 545-54. BMVA.
<ul>
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/vis_servo.ps.gz">Gzipped Postscript</A><div
class="download">Abstract</div><div class="abstract">
Visual servoing is a process to enable a robot to position a camera with
respect to known landmarks using the visual data obtained by the camera
itself to guide camera motion. A solution is described which requires very
little a priori information freeing it from being specific to a particular
configuration of robot and camera. The solution is based on closed loop
control together with deliberate perturbations of the trajectory to provide
calibration movements for refining that trajectory. Results from
experiments in simulation and on a physical robot arm (camera-in-hand
configuration) are presented.
</div>
</ul>
<br>M. W. Spratling (1994)
<B>Learning the Mapping Between Sensor and Motor Spaces to Produce Hand-Eye Coordination.</B>
MSc Dissertation,
<A HREF = "http://www.dai.ed.ac.uk/">Department of Artificial Intelligence</A>,
<A HREF = "http://www.ed.ac.uk/">University of Edinburgh</A>.
<ul>
<div class="download">Abstract</div><div class="abstract">
Coordination between sensory inputs and motor actions is essential for
intelligent robotics. This dissertation considers the control of a
simple manipulator using sensory information to locate the target
position for the end-effector. The control mechanisms investigated all
form topographic maps of possible configurations of the manipulator
joints (the motor space) and the values of the sensor inputs (the
sensor space). Various methods are considered for learning to relate a
location on the sensor space map (which represents the target position
in the world) and the location in the motor space map which will
configure the manipulator to reach this target position. These methods
are analysed using a computer simulation and a suitable algorithm to
solve the hand-eye coordination problem is presented.
</div>
</ul>
<br>
M. M. Ross, M. W. Spratling, C. B. Kirkland and P. S. Story (1994)
<B>Measurement of microfog wetness in a model steam turbine using a miniature optical spectral extinction probe.</B>
<I>IMechE International Symposium on Optical Methods and Data Processing In Heat and Fluid Flow.</I> <!--14-15 April 1994, City University, London.-->
</ul>
</div>
<P>
<HR>
<div id="footer">
(c) M. W. Spratling
</div>
</body>
</html>