1594 lines
105 KiB
HTML
1594 lines
105 KiB
HTML
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">
|
||
<html xmlns="http://www.w3.org/1999/xhtml">
|
||
<head>
|
||
<meta http-equiv="content-type" content="text/html;charset=iso-8859-1" />
|
||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||
<link rel="stylesheet" href="mws.css" type="text/css" />
|
||
<script type="text/javascript" src="jquery.js"></script>
|
||
<script type="text/javascript">
|
||
jQuery(document).ready(
|
||
function($) {
|
||
$(".abstract").toggle();
|
||
$("div").click(function() {
|
||
//$(".abstract").hide();
|
||
if (!$(this).next("div.abstract").is(":visible"))
|
||
$(this).next("div.abstract").toggle(600);
|
||
else
|
||
$(this).next("div.abstract").toggle(400);
|
||
});
|
||
});
|
||
</script>
|
||
<title>Michael Spratling: Home Page</title>
|
||
</head>
|
||
<body>
|
||
|
||
<div class="navbar">
|
||
<div><a href="#top">Dr. Michael Spratling</a></div>
|
||
<a href=""></a>
|
||
<a href="https://codeberg.org/mwspratling">Code</a>
|
||
<a href="#publications">Publications</a>
|
||
<a href="#research_interests">Research</a>
|
||
</div>
|
||
|
||
<div id="top"></div><br><br>
|
||
<table cellspacing=0 cellpadding=10px border=0>
|
||
<tr>
|
||
<td valign=center width=50%>
|
||
<p>Reader in Computational Neuroscience and Visual Cognition<br />
|
||
Department of Informatics<br />
|
||
King's College London<br />
|
||
University of London</p>
|
||
|
||
<script type="text/javascript">
|
||
var email = 'michael.spratling';
|
||
var emailHost = 'kcl.ac.uk';
|
||
var contact= 'michael <i>dot</i> spratling <i>at </i>kcl <i>dot</i> ac <i>dot</i> uk';
|
||
document.write('<a href=' + 'ma' + 'il' + 'to:' + email + '@' +
|
||
emailHost + '>' + contact + '</a>');
|
||
</script><p>
|
||
</td>
|
||
<td bgcolor="#ffffff" align=center valign=center>
|
||
<img src="Logo/cnvc.jpg" alt="logo" width="150" style="margin-left:-10px"/>
|
||
</td>
|
||
<td bgcolor="#ffffff" align=left valign=top width=50%>
|
||
</td>
|
||
</tr>
|
||
</table>
|
||
|
||
<div id="research_interests"></div><br>
|
||
<h2 >Research Interests</h2>
|
||
<table cellspacing=0 cellpadding=5px border=0>
|
||
<tr>
|
||
<td bgcolor="#ffffff" align=right valign=top width=50%>
|
||
<h3>Neural Networks</h3>
|
||
deep learning; biologically-inspired neural networks; adversarial robustness; generalisation;
|
||
representation learning; dynamic and deformable neural networks
|
||
</td>
|
||
<td bgcolor="#ffffff" align=center valign=center>
|
||
<img src="Logo/neural.jpg" width="120" align=left/>
|
||
</td>
|
||
<td bgcolor="#ffffff" align=left valign=top width=50%>
|
||
<h3>Computational Neuroscience</h3>
|
||
predictive coding; neural information processing and coding; dendritic computation; cortical feedback connections; cortical region interactions
|
||
</td>
|
||
</tr>
|
||
<tr>
|
||
<td bgcolor="#ffffff" align=right valign=top width=50%>
|
||
<h3>Computer Vision</h3>
|
||
object recognition; viewpoint invariance; image segmentation; object detection; tracking; anomaly detection; open-set recognition
|
||
</td>
|
||
<td bgcolor="#ffffff" align=center valign=center>
|
||
<img src="Logo/vision.jpg" width="120"/>
|
||
</td>
|
||
<td bgcolor="#ffffff" align=left valign=top width=50%>
|
||
<h3>Visual Cognition</h3>
|
||
visual attention; biased competition; top-down and contextual influences; visual salience; perceptual inference
|
||
</td>
|
||
</tr>
|
||
<tr>
|
||
<td bgcolor="#ffffff" align=right valign=top width=50%>
|
||
<h3>Machine Learning</h3>
|
||
sparse coding; continuous learning; few-shot learning; unsupervised and self-supervised learning; competitive learning; classification; clustering; contrastive and deep-metric learning
|
||
<td bgcolor="#ffffff" align=center valign=center>
|
||
<img src="Logo/development.jpg" width="120"/>
|
||
</td>
|
||
<td bgcolor="#ffffff" align=left valign=top width=50%>
|
||
<h3>Learning and Development</h3>
|
||
cortical, behavioural and cognitive development; perceptual learning; perceptual and conceptual development; epigenetic and developmental robotics
|
||
</td>
|
||
</tr>
|
||
</table>
|
||
|
||
|
||
<div id="publications"></div><br>
|
||
<H2>Publications
|
||
<div style="float: right; font-size: small">
|
||
<A class="download" HREF = "http://scholar.google.com/citations?user=A2tJIE4AAAAJ&hl=en"><B>Google Scholar</B></A><A class="download" HREF = "https://www.scopus.com/authid/detail.uri?authorId=6603009678"><B>Scopus</B></A><A class="download" HREF = "http://www.researcherid.com/rid/G-7689-2011"><B>Researcher ID</B></A><A class="download" HREF = "http://orcid.org/0000-0001-9531-2813"><B>ORCID</B></A>
|
||
</div></H2>
|
||
<ul>
|
||
|
||
<!--
|
||
L. Li and M. W. Spratling (2023) <b>Improved adversarial training through adaptive instance-wise loss smoothing.</b> <I> <A HREF = ""></A></I>, .
|
||
<ul>
|
||
<A class="download" HREF = "https://arxiv.org/pdf/2303.14077">PDF</A><A
|
||
class="download" HREF = "https://github.com/TreeLLi/Instance-adaptive-Smoothness-Enhanced-AT">Code</A><div class="download">Abstract</div><div class="abstract"> eep neural networks can be easily fooled into making incorrect predictions through corruption of the input by adversarial perturbations: human-imperceptible artificial noise. So far adversarial training has been the most successful defense against such adversarial attacks. This work focuses on improving adversarial training to boost adversarial robustness. We first analyze, from an instance-wise perspective, how adversarial vulnerability evolves during adversarial training. We find that during training an overall reduction of adversarial loss is achieved by sacrificing a considerable proportion of training samples to be more vulnerable to adversarial attack, which results in an uneven distribution of adversarial vulnerability among data. Such "uneven vulnerability", is prevalent across several popular robust training methods and, more importantly, relates to overfitting in adversarial training. Motivated by this observation, we propose a new adversarial training method: Instance-adaptive Smoothness Enhanced Adversarial Training (ISEAT). It jointly smooths both input and weight loss landscapes in an adaptive, instance-specific, way to enhance robustness more for those samples with higher adversarial vulnerability. Extensive experiments demonstrate the superiority of our method over existing defense methods. Noticeably, our method, when combined with the latest data augmentation and semi-supervised learning techniques, achieves state-of-the-art robustness against ℓ∞-norm constrained attacks on CIFAR10 of 59.32% for Wide ResNet34-10 without extra data, and 61.55% for Wide ResNet28-10 with extra data.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>-->
|
||
M. W. Spratling (2023) <B>Comprehensive assessment methods are key to progress in deep learning</B> [commentary]. <I><A HREF = "https://www.cambridge.org/core/journals/behavioral-and-brain-sciences">Behavioral and Brain Sciences</A></I>, in press.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/commentary_bowers.pdf">PDF</A>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>
|
||
H. Guan and M. W. Spratling (2023) <b>Query semantic reconstruction for background in few-shot segmentation.</b> <I><A HREF = "https://doi.org/10.1007/s00371-023-02817-x">The Visual Computer</A></I>, in press.
|
||
<ul>
|
||
<A class="download" HREF = "https://arxiv.org/pdf/2210.12055">PDF</A><!--<A class="download" HREF = "">Code</A>--><div class="download">Abstract</div><div class="abstract">
|
||
Few-shot segmentation (FSS) aims to segment unseen classes using a few annotated samples.
|
||
Typically, a prototype representing the foreground class is extracted from annotated support image(s)
|
||
and is matched to features representing each pixel in the query image. However, models learnt in this
|
||
way are insufficiently discriminatory, and often produce false positives: misclassifying background
|
||
pixels as foreground. Some FSS methods try to address this issue by using the background in the
|
||
support image(s) to help identify the background in the query image. However, the backgrounds
|
||
of theses images is often quite distinct, and hence, the support image background information is
|
||
uninformative. This article proposes a method, QSR, that extracts the background from the query
|
||
image itself, and as a result is better able to discriminate between foreground and background features
|
||
in the query image. This is achieved by modifying the training process to associate prototypes with
|
||
class labels including known classes from the training data and latent classes representing unknown
|
||
background objects. This class information is then used to extract a background prototype from
|
||
the query image. To successfully associate prototypes with class labels and extract a background
|
||
prototype that is capable of predicting a mask for the background regions of the image, the machinery
|
||
for extracting and using foreground prototypes is induced to become more discriminative between
|
||
different classes. Experiments achieves state-of-the-art results for both 1-shot and 5-shot FSS on the
|
||
PASCAL-5i and COCO-20i dataset. As QSR operates only during training, results are produced with
|
||
no extra computational complexity during testing.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
L. Li and M. W. Spratling (2023) <b>Data augmentation alone can improve adversarial training.</b> <I>Proceedings of the 11th <A HREF = "https://openreview.net/forum?id=y4uc4NtTWaq">International Conference on Learning Representations (ICLR)</A></I>.
|
||
<ul>
|
||
<A class="download" HREF = "https://arxiv.org/pdf/2301.09879.pdf">PDF</A><A
|
||
class="download" HREF = "https://github.com/TreeLLi/DA-Alone-Improves-AT">Code</A><div class="download">Abstract</div><div class="abstract">Adversarial training suffers from the issue of robust overfitting, which seriously impairs its generalization performance. Data augmentation, which is effective at preventing overfitting in standard training, has been observed by many previous works to be ineffective in mitigating overfitting in adversarial training. This work proves that, contrary to previous findings, data augmentation alone can significantly boost accuracy and robustness in adversarial training. We find that the hardness and the diversity of data augmentation are important factors in combating robust overfitting. In general, diversity can improve both accuracy and robustness, while hardness can boost robustness at the cost of accuracy within a certain limit and degrade them both over that limit. To mitigate robust overfitting, we first propose a new crop transformation Cropshift with improved diversity compared to the conventional one (Padcrop). We then propose a new data augmentation scheme, based on Cropshift, with much improved diversity and well-balanced hardness. Empirically, our augmentation method achieves the state-of-the-art accuracy and robustness for data augmentations in adversarial training. Furthermore, it matches, or even exceeds when combined with weight averaging, the performance of the best contemporary regularization methods for alleviating robust overfitting.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
|
||
J. Ning, H. Guan and M. W. Spratling (2023) <b>Rethinking the backbone architecture for tiny object detection.</b> <I>Proceedings of the 18th <A HREF =
|
||
"https://doi.org/10.5220/0011643500003417">International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISAPP)</A></I>, Volume 5, pp. 103-114.
|
||
<ul>
|
||
<A class="download" HREF = "https://arxiv.org/pdf/2303.11267">PDF</A><!--<A
|
||
class="download" HREF = "">Code</A>--><div class="download">Abstract</div><div class="abstract">
|
||
Tiny object detection has become an active area of research because images with tiny targets are common in several important real-world scenarios. However, existing tiny object detection methods use standard deep neural networks as their backbone architecture. We argue that such backbones are inappropriate for detecting tiny objects as they are designed for the classification of larger objects, and do not have the spatial resolution to identify small targets. Specifically, such backbones use max-pooling or a large stride at early stages in the architecture. This produces lower resolution feature-maps that can be efficiently processed by subsequent layers. However, such low-resolution feature-maps do not contain information that can reliably discriminate tiny objects. To solve this problem we design âÂÂbottom-heavyâ versions of backbones that allocate more resources to processing higher-resolution features without introducing any additional computational burden overall. We also investigate if pre-training these backbones on images of appropriate size, using CIFAR100 and ImageNet32, can further improve performance on tiny object detection. Results on TinyPerson and WiderFace show that detectors with our proposed backbones achieve better results than the current state-of-the-art methods.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
L. Li and M. W. Spratling (2023) <b>Understanding and combating robust overfitting via input loss landscape analysis and regularization.</b> <I><A HREF = "https://doi.org/10.1016/j.patcog.2022.109229">Pattern Recognition</A></I>, 136 (109229).
|
||
<ul>
|
||
<A class="download" HREF = "https://www.sciencedirect.com/science/article/pii/S0031320322007087/pdfft?md5=793d3d60b7b6a89f014c27b8562c1276&pid=1-s2.0-S0031320322007087-main.pdf">PDF</A><A
|
||
class="download" HREF = "https://github.com/TreeLLi/Combating-RO-AdvLC">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Adversarial training is widely used to improve the robustness of deep neural networks to adversarial attack. However, adversarial training is prone to overfitting, and the cause is far from clear. This work sheds light on the mechanisms underlying overfitting through analyzing the loss landscape w.r.t. the input. We find that robust overfitting results from standard training, specifically the minimization of the clean loss, and can be mitigated by regularization of the loss gradients. Moreover, we find that robust overfitting turns severer during adversarial training partially because the gradient regularization effect of adversarial training becomes weaker due to the increase in the loss landscape's curvature. To improve robust generalization, we propose a new regularizer to smooth the loss landscape by penalizing the weighted logits variation along the adversarial direction. Our method significantly mitigates robust overfitting and achieves the highest robustness and efficiency compared to similar previous methods.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
B. Gao and M. W. Spratling (2023) <b>Explaining away results in more robust visual tracking.</b> <I><A HREF = "https://doi.org/10.1007/s00371-022-02466-6">The Visual Computer</A></I>, 39:2081-95.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_tracking.pdf">PDF</A><A
|
||
class="download" HREF = "https://github.com/iminfine/DIMtracking">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Many current trackers utilise an appearance model to localise the target object in each frame. However, such approaches often fail when there are similar looking distractor objects in the surrounding background, meaning that target appearance alone is insufficient for robust tracking. In contrast, humans consider the distractor objects as additional visual cues, in order to infer the position of the target. Inspired by this observation, this paper proposes a novel tracking architecture in which not only is the appearance of the tracked object, but also the appearance of the distractors detected in previous frames, taken into consideration using a form of probabilistic inference known as explaining away. This mechanism increases the robustness of tracking by making it more likely that the target appearance model is matched to the true target, rather than similar-looking regions of the current frame. The proposed method can be combined with many existing trackers. Combining it with SiamFC, DaSiamRPN, Super DiMP and ARSuper DiMP all resulted in an increase in the tracking accuracy compared to that achieved by the underlying tracker alone. When combined with Super DiMP and ARSuper DiMP the resulting trackers produce performance that is competitive with the state-of-the-art on seven popular benchmarks.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
B. Gao and M. W. Spratling (2022) <b>Shape-texture debiased training for robust template matching.</b> <I><A HREF = "https://doi.org/10.3390/s22176658">Sensors</A></I>, 22(17), 6658.
|
||
<ul>
|
||
<A class="download" HREF = "https://www.mdpi.com/1424-8220/22/17/6658/pdf?version=1662127786">PDF</A><A
|
||
class="download" HREF = "https://github.com/iminfine/Deep-DIM">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Finding a template in a search image is an important task underlying many computer vision applications. This is typical solved by calculating a similarity map using features extracted from the separate images. Recent approaches perform template matching in a deep feature-space, produced by a convolutional neural network (CNN), which is found to provide more tolerance to changes in appearance. Inspired by these findings, in this article we investigate if enhancing the CNN's encoding of shape information can produce more distinguishable features that improve the performance of template matching. By comparing features from a same CNN but trained by different shape-texture training methods, we determined a feature-space which improves the performance of most template matching algorithms. When combining the proposed method with the Divisive Input Modulation (DIM) template matching algorithm, its performance is greatly improved, and the resulting method produces state-of-the-art results on a standard benchmark. To confirm these results we also create a new benchmark and show that the proposed method also outperforms existing techniques on this new dataset.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
N. Manchev and M. W. Spratling (2022) <b>On the biological plausibility of orthogonal initialisation for solving gradient instability in deep neural networks.</b> <I>Proceedings of the 9th <A HREF = "https://doi.org/10.1109/ISCMI56532.2022.10068489">International Conference on Soft Computing and Machine Intelligence (ISCMI)</A></I>, pp. 47-55.
|
||
<ul>
|
||
<A class="download" HREF = "https://arxiv.org/pdf/2211.08408.pdf">PDF</A><!--<A class="download" HREF = "https://github.com/nmanchev/">Code</A>--><div class="download">Abstract</div><div class="abstract">
|
||
Initialising the synaptic weights of artificial neural
|
||
networks (ANNs) with orthogonal matrices is known to alleviate
|
||
vanishing and exploding gradient problems. A major objection
|
||
against such initialisation schemes is that they are deemed
|
||
biologically implausible as they mandate factorization techniques
|
||
that are difficult to attribute to a neurobiological process. This
|
||
paper presents two initialisation schemes that allow a network to
|
||
naturally evolve its weights to form orthogonal matrices, provides
|
||
theoretical analysis that pre-training orthogonalisation always
|
||
converges, and empirically confirms that the proposed schemes
|
||
outperform randomly initialised recurrent and feedforward networks.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
C. Huang, H. Guan, A. Jiang, Y. Zhang, M. W. Spratling and Y.-F. Wang (2022) <b>Registration based few-shot anomaly detection.</b> <I><A HREF = "https://doi.org/10.1007/978-3-031-20053-3_18">European Conference on Computer Vision (ECCV)</A></I>, In: Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T. (eds) Proceedings part 24, Lecture Notes in Computer Science, Volume 13684, pp:303-19, Springer.
|
||
<ul>
|
||
<A class="download" HREF = "https://arxiv.org/pdf/2207.07361.pdf">PDF</A><A class="download" HREF = "https://github.com/MediaBrain-SJTU/RegAD">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
This paper considers few-shot anomaly detection (FSAD), a practical yet under-studied setting for anomaly detection (AD), where only a limited number of normal images are provided for each category at training. So far, existing FSAD studies follow the one-model-per-category learning paradigm used for standard AD, and the inter-category commonality has not been explored. Inspired by how humans detect anomalies, i.e., comparing an image in question to normal images, we here leverage registration, an image alignment task that is inherently generalizable across categories, as the proxy task, to train a category-agnostic anomaly detection model. During testing, the anomalies are identified by comparing the registered features of the test image and its corresponding support (normal) images. As far as we know, this is the first FSAD method that trains a single generalizable model and requires no re-training or parameter fine-tuning for new categories. Experimental results have shown that the proposed method outperforms the state-of-the-art FSAD methods by 3%-8% in AUC on the MVTec and MPDD benchmarks. Source code will be publicly available.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
H. Guan and M. W. Spratling (2022) <b>CobNet: cross attention on object and background for few-shot segmentation.</b> <I>Proceedings of the 26th <A HREF = "https://doi.org/10.1109/ICPR56361.2022.9956070">International Conference on Pattern Recognition (ICPR)</A></I>, pp. 39-45.
|
||
<ul>
|
||
<A class="download" HREF = "https://arxiv.org/pdf/2210.11968.pdf">PDF</A><!--<A class="download" HREF = "">Code</A>--><div class="download">Abstract</div><div class="abstract">
|
||
Few-shot segmentation aims to segment images containing objects from previously unseen classes using only a few annotated samples. Most current methods focus on using object information extracted, with the aid of human annotations, from support images to identify the same objects in new query images. However, background information can also be useful to distinguish objects from their surroundings. Hence, some previous methods also extract background information from the support images. In this paper, we argue that such information is of limited utility, as the background in different images can vary widely. To overcome this issue, we propose CobNet which utilises information about the background that is extracted from the query images without annotations of those images. Experiments show that our method achieves a mean Intersection-over-Union score of 61.4% and 37.8% for 1-shot segmentation on PASCAL-5i and COCO-20i respectively, outperforming previous methods. It is also shown to produce state-of-the-art performances of 53.7% for weakly-supervised few-shot segmentation, where no annotations are provided for the support images.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
B. Gao and M. W. Spratling (2022) <b>More robust object tracking via shape and motion cue integration.</b> <I><A HREF = "https://doi.org/10.1016/j.sigpro.2022.108628">Signal Processing</A></I>, 199 (108628).
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/trackingmulticues.pdf">PDF</A><A
|
||
class="download" HREF = "https://github.com/iminfine/trackingmulticues">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Most current trackers utilise an appearance model to localise the target object in each frame. However, such approaches often fail when there are similar looking distractor objects in the surrounding background. This paper promotes an approach that can be combined with many existing trackers to tackle this issue and improve tracking robustness. The proposed approach makes use of two additional cues to target location: making the appearance model more sensitive to shape cues in offline training, and using the historical locations of the target to predict its future position during online inference. Combining these additional mechanisms with SiamFC, SiamFC++, Super DiMP and ARSuper DiMP all resulted in an increase in the tracking accuracy compared to that achieved by the corresponding underlying tracker alone. When combined with ARSuper DiMP the resulting tracker is shown to outperform all popular state-of-the-art trackers on three benchmark datasets (OTB-100, NFS, and LaSOT), and produce performance that is competitive with the state-of-the-art on the UAV123, Trackingnet, GOT-10K and VOT2020 datasets.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
B. Gao and M. W. Spratling (2021) <b>Robust template matching via hierarchical convolutional features from a shape biased CNN.</b> <I>Proceedings of the <A HREF = "https://doi.org/10.1007/978-981-16-6963-7_31">International Conference on Image, Vision and Intelligent Systems (ICIVIS)</A></I>, Lecture Notes in Electrical Engineering, Volume 813. Springer, Singapore.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_cnn_patchmatching.pdf">PDF</A><A
|
||
class="download" HREF = "https://github.com/iminfine/Deep-DIM">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Finding a template in a search image is an important task underlying many computer vision applications. Recent approaches perform template matching in a deep feature-space, produced by a convolutional neural network (CNN), which is found to provide more tolerance to changes in appearance. In this article we investigate if enhancing the CNN's encoding of shape information can produce more distinguishable features that improve the performance of template matching. This investigation results in a new template matching method that produces state-of-the-art results on a standard benchmark. To confirm these results we also create a new benchmark and show that the proposed method also outperforms existing techniques on this new dataset.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2020) <b>Explaining away results in accurate and tolerant template matching.</b> <I><A HREF = "https://doi.org/10.1016/j.patcog.2020.107337">Pattern Recognition</A></I>, 104 (107337).
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_patchmatching.pdf">PDF</A><A
|
||
class="download" HREF = "https://codeberg.org/mwspratling/PatchMatching">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Recognising and locating image patches or sets of image features is an important task underlying much work in computer vision. Traditionally this has been accomplished using template matching. However, template matching is notoriously brittle in the face of changes in appearance caused by, for example, variations in viewpoint, partial occlusion, and non-rigid deformations. This article tests a method of template matching that is more tolerant to such changes in appearance and that can, therefore, more accurately identify image patches. In traditional template matching the comparison between a template and the image is independent of the other templates. In contrast, the method advocated here takes into account the evidence provided by the image for the template at each location and the full range of alternative explanations represented by the same template at other locations and by other templates. Specifically, the proposed method of template matching is performed using a form of probabilistic inference known as "explaining away". The algorithm used to implement explaining away has previously been used to simulate several neurobiological mechanisms, and been applied to image contour detection and pattern recognition tasks. Here it is applied for the first time to image patch matching, and is shown to produce superior results in comparison to the current state-of-the-art methods.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
N. Manchev and M. W. Spratling (2020) <b>Target propagation in recurrent neural networks.</b> <I><A HREF = "http://www.jmlr.org/papers/v21/18-141.html">Journal of Machine Learning Research</A></I>, 21(7): 1-33.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/rnn_tptt.pdf">PDF</A><A
|
||
class="download" HREF = "https://github.com/nmanchev/tptt">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Recurrent Neural Networks have been widely used to process sequence data, but have
|
||
long been criticized for their biological implausibility and training difficulties related to
|
||
vanishing and exploding gradients. This paper presents a novel algorithm for training
|
||
recurrent networks, target propagation through time (TPTT), that outperforms standard
|
||
backpropagation through time (BPTT) on four out of the five problems used for testing.
|
||
The proposed algorithm is initially tested and compared to BPTT on four synthetic time
|
||
lag tasks, and its performance is also measured using the sequential MNIST data set. In
|
||
addition, as TPTT uses target propagation, it allows for discrete nonlinearities and could
|
||
potentially mitigate the credit assignment problem in more complex recurrent architectures.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2019) <b>Fitting predictive coding to the neurophysiological data.</b> <I><A HREF = "https://doi.org/10.1016/j.brainres.2019.146313">Brain Research</A></I>, 1720: 146313.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pc_biological_fit.pdf">PDF</A><A
|
||
class="download" HREF = "https://codeberg.org/mwspratling/BiologicalFit">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Recent neurophysiological data showing the effects of locomotion on neural
|
||
activity in mouse primary visual cortex has been interpreted as providing strong
|
||
support for the predictive coding account of cortical function. Specifically,
|
||
this work has been interpreted as providing direct evidence that
|
||
prediction-error, a distinguishing property of predictive coding, is encoded in
|
||
cortex. This article evaluates these claims and highlights some of the
|
||
discrepancies between the proposed predictive coding model and the
|
||
neuro-biology. Furthermore, it is shown that the model can be modified so as to
|
||
fit the empirical data more successfully.
|
||
</div></ul>
|
||
|
||
|
||
|
||
<br>
|
||
I. E. Kartoglu and M. W. Spratling (2018) <b>Two collaborative filtering recommender systems based on sparse dictionary coding.</b> <I><A HREF = "http://dx.doi.org/10.1007/s10115-018-1157-2">Knowledge and Information Systems</A></I>, 57(3): 709-20.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/sparse_recommender.pdf">PDF</A><A
|
||
class="download" HREF = "https://github.com/iemre/MRSR">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
This paper proposes two types of recommender systems based on sparse dictionary
|
||
coding. Firstly, a novel predictive recommender system that attempts to predict
|
||
a user's future rating of a specific item. Secondly, a top-n recommender system
|
||
which finds a list of items predicted to be most relevant for a given user. The
|
||
proposed methods are assessed using a variety of different metrics and are shown
|
||
to be competitive with existing collaborative filtering recommender
|
||
systems. Specifically, the sparse dictionary-based predictive recommender has
|
||
advantages over existing methods in terms of a lower computational cost and not
|
||
requiring parameter tuning. The sparse dictionary-based top-n recommender system
|
||
has advantages over existing methods in terms of the accuracy of the predictions
|
||
it makes and not requiring parameter tuning. An open-source software implemented
|
||
and used for the evaluation in this paper is also provided for reproducibility.
|
||
</div></ul>
|
||
|
||
|
||
<br>
|
||
Q. Wang and M. W. Spratling (2018) <b>Contour detection refined by a sparse reconstruction-based discrimination method.</b> <I><A HREF = "http://dx.doi.org/10.1007/s11760-017-1147-y">Signal, Image and Video Processing</A></I>, 12(2): 207-14.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/SIVPRO-final.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
|
||
Sparse representations have been widely used for many image processing tasks. In
|
||
this paper, a sparse reconstruction-based discrimination (SRBD) method, which
|
||
was previously proposed for the classification of image patches, is utilized to
|
||
improve boundary detection in colour images. This method is applied to refining
|
||
the results generated by three different algorithms: a biologically-inspired
|
||
method, and two state-of-the-art algorithms for contour detection. All of the
|
||
contour detection results are evaluated by the BSDS300 and BSDS500 benchmarks
|
||
using the quantitative measures: F-score, ODS, OIS and AP. Evaluation results
|
||
show that the performance of each algorithm is improved using the proposed
|
||
method of refinement with at least one of the quantitative measures increased by
|
||
0.01. In particularly, even two state-of-the-art algorithms are slightly
|
||
improved by applying the SRBD method to refine their contour detection results.
|
||
</div></ul>
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2017) <b>A predictive coding model of gaze shifts and the
|
||
underlying neurophysiology.</b> <I><A HREF = "http://dx.doi.org/10.1080/13506285.2017.1336141">Visual Cognition</A></I>, 25(7-8): 770-801.
|
||
<ul> <A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_gaze.pdf">PDF</A><A
|
||
class="download" HREF = "https://codeberg.org/mwspratling/GazeSimulation">Code</A><div
|
||
class="download">Abstract</div><div class="abstract"> A comprehensive model of
|
||
gaze control must account for a number of empirical observations at both the
|
||
behavioural and neurophysiological levels. The computational model presented in
|
||
this article can simulate the coordinated movements of the eye, head, and body
|
||
required to perform horizontal gaze shifts. In doing so it reproduces the
|
||
predictable relationships between the movements performed by these different
|
||
degrees of freedom (DOFs) in the primate. The model also accounts for the
|
||
saccadic undershoot that accompanies large gaze shifts in the biological visual
|
||
system. It can also account for our perception of a stable external world
|
||
despite frequent gaze shifts and the ability to perform accurate memory-guided
|
||
and double-step saccades. The proposed model also simulates peri-saccadic
|
||
compression: the mis-localisation of a briefly presented visual stimulus towards
|
||
the location that is the target for a saccade. At the neurophysiological level,
|
||
the proposed model is consistent with the existence of cortical neurons tuned to
|
||
the retinal, head-centred, body-centred, and world-centred locations of visual
|
||
stimuli and cortical neurons that have gain-modulated responses to visual
|
||
stimuli. Finally, the model also successfully accounts for peri-saccadic
|
||
receptive field (RF) remapping which results in reduced responses to stimuli in
|
||
the current RF location and an increased sensitivity to stimuli appearing at the
|
||
location that will be occupied by the RF after the saccade. The proposed model
|
||
thus offers a unified explanation for this seemingly diverse range of
|
||
phenomena. Furthermore, as the proposed model is an implementation of the
|
||
predictive coding theory, it offers a single computational explanation for these
|
||
phenomena and relates gaze shifts to a wider framework for understanding
|
||
cortical function.
|
||
</div>
|
||
</ul>
|
||
|
||
<br>
|
||
W. Muhammad and M. W. Spratling (2017) <b>A neural model for eye-head-arm coordination.</b> <I><A HREF = "http://dx.doi.org/10.1080/01691864.2017.1315319">Advanced Robotics</A></I>, 31(12): 650-63.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_eye_head_arm_control.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
The coordinated movement of the eyes, the head and
|
||
the arm is an important ability in both animals and humanoid
|
||
robots. To achieve this the brain and the robot control system
|
||
need to be able to perform complex non-linear sensory-motor
|
||
transformations in the forward and inverse directions between
|
||
many degrees of freedom. In this article we apply an omni-
|
||
directional basis function neural network to this task. The
|
||
proposed network can perform 3-D coordinated gaze shifts and
|
||
3-D arm reach movements to a visual target. Particularly, it can
|
||
perform direct sensory-motor transformations to shift gaze and
|
||
to execute arm reach movements and can also perform inverse
|
||
sensory-motor transformations in order to shift gaze to view the
|
||
hand. </div>
|
||
</ul>
|
||
|
||
<br>
|
||
M. W. Spratling (2017) <b>A hierarchical predictive coding model of object recognition in natural images.</b> <I><A HREF =
|
||
"http://dx.doi.org/10.1007/s12559-016-9445-1">Cognitive Computation</A></I>, 9(2): 151-67.
|
||
<ul>
|
||
<A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_image_recognition.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/ImageRecognition">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Predictive coding has been proposed as a model of the hierarchical
|
||
perceptual inference process performed in the cortex. However, results
|
||
demonstrating that predictive coding is capable of performing the complex
|
||
inference required to recognise objects in natural images have not
|
||
previously been presented.
|
||
This article proposes a hierarchical neural network based on predictive
|
||
coding for performing visual object recognition.
|
||
This network is applied to the tasks of categorising hand-written digits,
|
||
identifying faces, and locating cars in images of street scenes. It is shown
|
||
that image recognition can be performed with tolerance to position,
|
||
illumination, size, partial occlusion and within-category variation.
|
||
The current results, therefore, provide the first practical demonstration
|
||
that predictive coding (at least the particular implementation of predictive
|
||
coding used here; the PC/BC-DIM) is capable of performing accurate
|
||
visual object recognition.
|
||
</div>
|
||
</ul>
|
||
|
||
<br>
|
||
D. Re, A. Gibaldi, S. P. Sabatini and M. W. Spratling (2017) <b>An integrated system based on binocular learned receptive fields for saccade-vergence on visually salient targets.</b> <I>Proceedings of the 12th <A HREF =
|
||
"http://www.scitepress.org/DigitalLibrary/PublicationsDetail.aspx?ID=TyxypQ+XiJc=&t=1">International Conference on Computer Vision Theory and Applications (VISAPP)</A></I>, Volume 6, pp:204-15.
|
||
<ul>
|
||
<A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/re_visapp17.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
The human visual system uses saccadic and vergence eyes movements to foveate interesting objects with both eyes, and thus exploring the visual scene. To mimic this biological behavior in active vision, we proposed a bio-inspired integrated system able to learn a functional sensory representation of the environment, together with the motor commands for binocular eye coordination, directly by interacting with the environment itself. The proposed architecture, rather than sequentially combining different functionalities, is a robust integration of different modules that rely on a front-end of learned binocular receptive fields to specialize on different sub-tasks. The resulting modular architecture is able to detect salient targets in the scene and perform precise binocular saccadic and vergence movement on it. The performances of the proposed approach has been tested on the iCub Simulator, providing a quantitative evaluation of the computational potentiality of the learned sensory and motor resources.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2017) <b>A review of predictive coding algorithms.</b> <I><A HREF =
|
||
"http://dx.doi.org/10.1016/j.bandc.2015.11.003">Brain and Cognition</A></I>, 112: 92-7.
|
||
<ul>
|
||
<A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/taxonomy_pc.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">Predictive coding is a
|
||
leading theory of how the brain performs probabilistic inference. However, there
|
||
are a number of distinct algorithms which are described by the term "predictive
|
||
coding". This article provides a concise review of these different predictive
|
||
coding algorithms, highlighting their similarities and differences. Five
|
||
algorithms are covered: linear predictive coding which has a long and
|
||
influential history in the signal processing literature; the first
|
||
neuroscience-related application of predictive coding to explaining the function
|
||
of the retina; and three versions of predictive coding that have been proposed
|
||
to model cortical function. While all these algorithms aim to fit a generative
|
||
model to sensory data, they differ in the type of generative model they employ,
|
||
in the process used to optimise the fit between the model and sensory data, and
|
||
in the way that they are related to neurobiology.</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>
|
||
W. Muhammad and M. W. Spratling (2017) <b>A neural model of coordinated head and eye movement control.</b> <I><A HREF = "http://dx.doi.org/10.1007/s10846-016-0410-8">Journal of Intelligent & Robotic Systems</A></I>, 85(1):107-26.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_head_control.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">Gaze shifts require the coordinated movement of both the eyes and the head in both animals and humanoid robots. To achieve this the brain and the robot control system needs to be able to perform complex non-linear sensory-motor transformations between many degrees of freedom and resolve the redundancy in such a system. In this article we propose a hierarchical neural network model for performing 3-D coordinated gaze shifts. The network is based on the PC/BC-DIM (Predictive Coding/Biased Competition with Divisive Input Modulation) basis function model. The proposed model consists of independent eyes and head controlled circuits with mutual interactions for the appropriate adjustment of coordination behaviour. Based on the initial eyes and head positions the network resolves redundancies involved in 3-D gaze shifts and produces accurate gaze control without any kinematic analysis or imposing any constraints. Furthermore the behaviour of the proposed model is consistent with coordinated eye and head movements observed in primates.</div>
|
||
</ul>
|
||
|
||
|
||
<br>
|
||
Q. Wang and M. W. Spratling (2016) <b>Contour detection in colour images using a neurophysiologically inspired model.</b> <I><A HREF = "http://dx.doi.org/10.1007/s12559-016-9432-6">Cognitive Computation</A></I>, 8(6):1027-35.
|
||
<ul>
|
||
<A class="download" HREF = "http://rdcu.be/kDd8">PDF</A><div class="download">Abstract</div><div class="abstract">
|
||
The predictive coding/biased competition (PC/BC) model of V1 has previously been applied to locate boundaries defined by local discontinuities in intensity within an image. Here it is extended to perform contour detection for colour images. The proposed extensions are inspired by neurophysiological data from single neurons in macaque primary visual cortex (V1), and the behaviour of this extended model is consistent with the neurophysiological experimental results. Furthermore, when compared to methods used for contour detection in computer vision, the colour PC/BC model of V1 slightly outperforms some recently proposed algorithms which use more cues and/or require a complicated training procedure.
|
||
</ul>
|
||
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2016) <b>A neural implementation of Bayesian inference based on predictive coding.</b> <I><A HREF =
|
||
"http://dx.doi.org/10.1080/09540091.2016.1243655">Connection Science</A></I>, 28(4):346-83.
|
||
<ul>
|
||
<A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_prob.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/ProbabilisticInference">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Predictive coding is a leading theory of cortical function that
|
||
has previously been shown to explain a great deal of neurophysiological and
|
||
psychophysical data. Here it is shown that predictive coding can perform almost
|
||
exact Bayesian inference when applied to computing with population codes. It is
|
||
demonstrated that the proposed algorithm, based on predictive coding, can:
|
||
decode probability distributions encoded as noisy population codes; combine
|
||
priors with likelihoods to calculate posteriors; perform cue integration and cue
|
||
segregation; perform function approximation; be extended to perform hierarchical
|
||
inference; simultaneously represent and reason about multiple stimuli; and
|
||
perform inference with multi-modal and non-Gaussian probability distributions.
|
||
Predictive coding thus provides a neural network based method for performing
|
||
probabilistic computation and provides a simple, yet comprehensive, theory of
|
||
how the cerebral cortex performs Bayesian inference.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2016) <b>Predictive coding as a model of cognition.</b> <I><A HREF =
|
||
"http://dx.doi.org/10.1007/s10339-016-0765-6">Cognitive Processing</A></I>, 17(3): 279-305.
|
||
<ul>
|
||
<A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_cognition.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/Cognition">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Previous work has shown that predictive coding can provide a detailed
|
||
explanation of a very wide range of low-level perceptual processes. It is also
|
||
widely believed that predictive coding can account for high-level, cognitive,
|
||
abilities. This article provides support for this view by showing that
|
||
predictive coding can simulate phenomena such as categorisation, the influence
|
||
of abstract knowledge on perception, recall and reasoning about conceptual
|
||
knowledge, context-dependent behavioural control, and naive physics. The
|
||
particular implementation of predictive coding used here (PC/BC-DIM) has previously
|
||
been used to simulate low-level perceptual behaviour and the neural mechanisms
|
||
that underlie them. This algorithm thus provides a single framework for
|
||
modelling both perceptual and cognitive brain function.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2016) <b>A neural implementation of the Hough transform and the advantages of explaining away.</b> <I><A HREF ="http://dx.doi.org/10.1016/j.imavis.2016.05.001">Image and Vision Computing</A></I>, 52:15-24.
|
||
<ul>
|
||
<A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_hough.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/HoughTransform">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
The Hough Transform (HT) is widely used for feature extraction and object
|
||
detection. However, during the HT individual image elements vote for many
|
||
possible parameter values. This results in a dense accumulator array and
|
||
problems identifying the parameter values that correspond to image
|
||
features. This article proposes a new method for implementing the voting process
|
||
in the HT. This method employs a competitive neural network algorithm to perform
|
||
a form of probabilistic inference known as "explaining away". This results in a
|
||
sparse accumulator array in which the parameter values of image features can be
|
||
more accurately identified. The proposed method is initially demonstrated using
|
||
the simple, prototypical, task of straight line detection in synthetic
|
||
images. In this task it is shown to more accurately identify straight lines, and
|
||
the parameter of those lines, compared to the standard Hough voting process. The
|
||
proposed method is further assessed using a version of the implicit shape model
|
||
(ISM) algorithm applied to car detection in natural images. In this application
|
||
it is shown to more accurately identify cars, compared to using the standard
|
||
Hough voting process in the same algorithm, and compared to the original ISM
|
||
algorithm.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
|
||
<br>
|
||
Q. Wang and M. W. Spratling (2016) <b>A simplified texture gradient method for improved image segmentation.</b> <I><A HREF = "http://dx.doi.org/10.1007/s11760-015-0794-0">Signal, Image and Video Processing</A></I>, 10(4):679-86.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/wang_sivpro15.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
|
||
Inspired by the probability of boundary (Pb) algorithm, a simplified texture gradient method has been developed to locate texture boundaries within grayscale images. Despite considerable simplification, the proposed algorithm's ability to locate texture boundaries is comparable with Pb's texture boundary method. The proposed texture gradient method is also integrated with a biologically inspired model, to enable boundaries defined by discontinuities in both intensity and texture to be located. The combined algorithm outperforms the current state-of-art image segmentation method (Pb) when this method is also restricted to using only local cues of intensity and texture at a single scale.</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>W. Muhammad and M. W. Spratling (2015) <b>A neural model of binocular saccade planning and
|
||
vergence control.</b> <I><A HREF = "http://dx.doi.org/10.1177/1059712315607363">Adaptive Behavior</A></I>, 23(5):265-82.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pcbc_eye_control.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract"> The human visual system
|
||
uses saccadic and vergence eye movements to foveate visual targets. To mimic
|
||
this aspect of the biological visual system the PC/BC-DIM neural network is used
|
||
as an omni-directional basis function network for learning and performing
|
||
sensory-sensory and sensory-motor transformations without using any hard-coded
|
||
geometric information. A hierarchical PC/BC-DIM network is used to learn a
|
||
head-centred representation of visual targets by dividing the whole problem into
|
||
independent subtasks. The learnt head- centred representation is then used to
|
||
generate saccade and vergence motor commands. The performance of the proposed
|
||
system is tested using the iCub humanoid robot simulator.</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>
|
||
M. W. Spratling (2014) <b>Classification using sparse
|
||
representations: a biologically plausible approach.</b> <I><A HREF =
|
||
"http://dx.doi.org/10.1007/s00422-013-0579-x">Biological Cybernetics</A></I>,
|
||
108(1):61-73.
|
||
<ul>
|
||
<A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/sparse_classification.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/SparseClassification">Code</A><div
|
||
class="download">Abstract</div><div class="abstract"> Representing signals as
|
||
linear combinations of basis vectors sparsely selected from an overcomplete
|
||
dictionary has proven to be advantageous for many applications in pattern
|
||
recognition, machine learning, signal processing, and computer vision. While
|
||
this approach was originally inspired by insights into cortical information
|
||
processing, biologically-plausible approaches have been limited to exploring the
|
||
functionality of early sensory processing in the brain, while more practical
|
||
application have employed non-biologically-plausible sparse-coding
|
||
algorithms. Here, a biologically-plausible algorithm is proposed that can be
|
||
applied to practical problems. This algorithm is evaluated using standard
|
||
benchmark tasks in the domain of pattern classification, and its performance is
|
||
compared to a wide range of alternative algorithms that are widely used in
|
||
signal and image processing. The results show that, for the classification
|
||
tasks performed here, the proposed method is very competitive with the best of
|
||
the alternative algorithms that have been evaluated. This demonstrates that
|
||
classification using sparse representations can be performed in a
|
||
neurally-plausible manner, and hence, that this mechanism of classification
|
||
might be exploited by the brain. </div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2014) <b>A single functional model of drivers and modulators in cortex.</b> <I><A HREF = "http://dx.doi.org/10.1007/s10827-013-0471-7">Journal of Computational Neuroscience</A></I>, 36(1): 97-118.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/drivers_modulators.pdf">PDF</A><A class="download" HREF = "https://codeberg.org/mwspratling/DriversModulators">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
A distinction is commonly made between synaptic connections capable of evoking a
|
||
response ("drivers") and those that can alter ongoing activity but not
|
||
initiate it ("modulators"). Here it is proposed that, in cortex, both drivers
|
||
and modulators are an emergent property of the perceptual inference performed by
|
||
cortical circuits. Hence, it is proposed that there is a single underlying
|
||
computational explanation for both forms of synaptic connection. This idea is
|
||
illustrated using a predictive coding model of cortical perceptual inference.
|
||
In this model all synaptic inputs are treated identically. However,
|
||
functionally, certain synaptic inputs drive neural responses while others have a
|
||
modulatory influence. This model is shown to account for driving and modulatory
|
||
influences in bottom-up, lateral, and top-down pathways, and is used to simulate
|
||
a wide range of neurophysiological phenomena including surround suppression,
|
||
contour integration, gain modulation, spatio-temporal prediction, and attention.
|
||
The proposed computational model thus provides a single functional explanation
|
||
for drivers and modulators and a unified account of a diverse range of
|
||
neurophysiological data. </div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2013) <b>Predictive coding.</b> In <I><A HREF = "http://dx.doi.org/10.1007/978-1-4614-7320-6_509-6">Encyclopedia of Computational Neuroscience,</A></I> D. Jaeger and R. Jung (Eds.), Springer, New York.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pc_encyc.pdf">PDF</A>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2013) <B>Distinguishing theory from implementation in predictive coding accounts of brain function</B> [commentary]. <I><A HREF = "http://dx.doi.org/10.1017/S0140525X12002178">Behavioral and Brain Sciences</A></I>, 36(3):231-2.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/commentary_clark.pdf">PDF</A>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2013) <b>Image segmentation using a sparse coding model of cortical area V1.</b> <I><A HREF = "http://dx.doi.org/10.1109/TIP.2012.2235850">IEEE Transactions on Image Processing</A></I>, 22(4):1631-43.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/edge_detection.pdf">PDF</A><A class="download" HREF = "https://codeberg.org/mwspratling/EdgeDetection">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
Algorithms that encode images using a sparse set of basis
|
||
functions have previously been shown to explain aspects of the physiology of
|
||
primary visual cortex (V1), and have been used for applications such as image
|
||
compression, restoration, and classification. Here, a sparse coding algorithm,
|
||
that has previously been used to account of the response properties of
|
||
orientation tuned cells in primary visual cortex, is applied to the task of
|
||
perceptually salient boundary detection. The proposed algorithm is currently
|
||
limited to using only intensity information at a single scale. However, it is
|
||
shown to out-perform the current state-of-the-art image segmentation method (Pb)
|
||
when this method is also restricted to using the same information. </div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>K. De Meyer and M. W. Spratling (2013) <b>A model of partial reference frame transforms through pooling of gain-modulated responses.</b> <I><A HREF =
|
||
"http://dx.doi.org/10.1093/cercor/bhs117">Cerebral Cortex</A></I>, 23(5):1230-9.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/partial_transforms.pdf">PDF</A><A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Code/partial_transforms.zip">Code</A><div class="download">Abstract</div><div class="abstract">
|
||
In multimodal integration and sensorimotor transformation areas of posterior
|
||
parietal cortex (PPC), neural responses often appear encoded in spatial
|
||
reference frames that are intermediate to intrinsic sensory reference frames,
|
||
e.g., eye-centred for visual or head-centred for auditory stimulation. Many
|
||
sensory responses in these areas are also modulated by direction of gaze. We
|
||
demonstrate that certain types of mixed-frame responses can be generated by
|
||
pooling gain-modulated responses - similarly to how complex cells in visual
|
||
cortex are thought to pool the responses of simple cells. The proposed model
|
||
simulates two types of mixed-frame responses observed in PPC: in particular,
|
||
sensory responses that shift differentially with gaze in horizontal and vertical
|
||
dimensions; and sensory responses that shift differentially for different start
|
||
and end points along a single dimension of gaze. We distinguish these two types
|
||
of mixed-frame responses from a third type in which sensory responses shift a
|
||
partial yet approximately equal amount with each gaze shift. We argue that the
|
||
empirical data on mixed-frame responses may be caused by multiple mechanisms,
|
||
and we adapt existing reference-frame measures to distinguish between the
|
||
different types. Finally, we discuss how mixed-frame responses may be revealing
|
||
of the local organisation of presynaptic responses.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2012) <B>Predictive coding accounts for V1
|
||
response properties recorded using reverse correlation.</B> <I><A HREF =
|
||
"http://dx.doi.org/10.1007/s00422-012-0477-7">Biological Cybernetics</A></I>, 106(1):37-49.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_reverse_correl.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/V1_ReverseCorrelation">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
PC/BC ("Predictive Coding/Biased Competition") is a simple computational model that has previously been shown to explain a very wide range of V1 response properties. This article extends work on the PC/BC model of V1 by showing that it can also account for V1 response properties measured using the reverse correlation methodology. Reverse correlation employs an experimental procedure that is significantly different from that used in more typical neurophysiological experiments, and measures some distinctly different response properties in V1. Despite these differences PC/BC successfully accounts for the data. The current results thus provide additional support for the PC/BC model of V1 and further demonstrate that PC/BC offers a unified explanation for the seemingly diverse range of behaviours observed in primary visual cortex.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2012) <B>Predictive coding as a model of the V1
|
||
saliency map hypothesis.</B> <I><A HREF = "http://dx.doi.org/10.1016/j.neunet.2011.10.002">Neural Networks</A></I>, 26:7-28.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_saliency.pdf">PDF</A><A
|
||
class="download" HREF = "https://codeberg.org/mwspratling/V1_Saliency">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
The predictive coding/biased competition (PC/BC) model is a specific implementation of predictive coding theory that has previously been shown to provide a detailed account of the response properties of orientation tuned cells in primary visual cortex (V1). Here it is shown that the same model can successfully simulate psychophysical data relating to the saliency of unique items in search arrays, of contours embedded in random texture, and of borders between textured regions. This model thus provides a possible implementation of the hypothesis that V1 generates a bottom-up saliency map. However, PC/BC is very different from previous models of visual salience, in that it proposes that saliency results from the failure of an internal model of simple elementary image components to accurately predict the visual input. Saliency can therefore be interpreted as a mechanism by which prediction errors attract attention in an attempt to improve the accuracy of the brain's internal representation of the world.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2012) <B>Unsupervised learning of generative and discriminative
|
||
weights encoding elementary image components in a predictive coding model of
|
||
cortical function.</B> <A HREF = "http://dx.doi.org/10.1162/NECO_a_00222">
|
||
<I>Neural Computation</I></A>, 24(1): 60-103.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim_learn_recip_weights.pdf">PDF</A><A
|
||
class="download" HREF = "https://codeberg.org/mwspratling/LearnReciprocalWeights">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
A method is presented for learning the reciprocal feedforward and feedback
|
||
connections required by the predictive coding model of cortical
|
||
function. Using this method feedforward and feedback connections are learnt
|
||
simultaneously and independently in a biologically plausible manner. The
|
||
performance of the proposed algorithm is evaluated by applying it to learning
|
||
the elementary components of artificial images and of natural images. For
|
||
artificial images the bars problem is employed and the proposed algorithm is
|
||
shown to produce state-of-the-art performance on this task. For natural
|
||
images, components resembling Gabor functions are learnt in the first
|
||
processing stage and neurons responsive to corners are learnt in the second
|
||
processing stage. The properties of these learnt representations are in good
|
||
agreement with neurophysiological data from V1 and V2. The proposed algorithm
|
||
demonstrates for the first time that a single computational theory can explain
|
||
the formation of cortical RFs, and also the response properties of cortical
|
||
neurons once those RFs have been learnt.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>K. De Meyer and M. W. Spratling (2011) <b>Multiplicative gain
|
||
modulation arises through unsupervised learning in a predictive coding model of
|
||
cortical function.</b> <A HREF="http://dx.doi.org/10.1162/NECO_a_00130"><I>Neural Computation</I></A>, 23(6):1536-67.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/gain_modulation.pdf">PDF</A><A
|
||
class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Code/gain_modulation.zip">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
The combination of two or more population-coded signals in a
|
||
neural model of predictive coding can give rise to multiplicative
|
||
gain modulation in the response properties of individual neurons.
|
||
Synaptic weights generating these multiplicative response
|
||
properties can be learned using an unsupervised, Hebbian, learning
|
||
rule. The behaviour of the model is compared to empirical data on
|
||
gaze-dependent gain modulation of cortical cells, and found to be
|
||
in good agreement with a range of physiological observations.
|
||
Furthermore, it is demonstrated that the model can learn to
|
||
represent a set of basis functions. The current paper thus
|
||
connects an often-observed neurophysiological phenomenon and
|
||
important neurocomputational principle (gain modulation) with an
|
||
influential theory of brain operation (predictive coding).
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2011) <B>A single functional model accounts for the
|
||
distinct properties of suppression in cortical area V1.</B> <A
|
||
HREF="http://dx.doi.org/10.1016/j.visres.2011.01.017"><I>Vision
|
||
Research</I></A>, 51(6):563-76.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_suppression_mechanisms.pdf">PDF</A><A
|
||
class="download" HREF = "https://codeberg.org/mwspratling/V1_SuppressionMechanisms">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Cross-orientation suppression and surround suppression have been extensively
|
||
studied in primary visual cortex (V1). These two forms of suppression have some
|
||
distinct properties which has led to the suggestion that they are generated by
|
||
different underlying mechanisms. Furthermore, it has been suggested that
|
||
mechanisms other than intracortical inhibition may be central to both forms of
|
||
suppression. A simple computational model (PC/BC), in which intracortical
|
||
inhibition is fundamental, is shown to simulate the distinct properties of
|
||
cross-orientation and surround suppression. The same model has previously been
|
||
shown to account for a large range of V1 response properties including
|
||
orientation-tuning, spatial and temporal frequency tuning, facilitation and
|
||
inhibition by flankers and textured surrounds as well as a range of other
|
||
experimental results on cross-orientation suppression and surround
|
||
suppression. The current results thus provide additional support for the PC/BC
|
||
model of V1 and for the proposal that the diverse range of response properties
|
||
observed in V1 neurons have a single computational explanation. Furthermore,
|
||
these results demonstrate that current neurophysiological evidence is
|
||
insufficient to discount intracortical inhibition as a central mechanism
|
||
underlying both forms of suppression. </div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2010) <B>Predictive coding as a model of response
|
||
properties in cortical area V1.</B> <I><A HREF =
|
||
"http://dx.doi.org/10.1523/JNEUROSCI.4911-09.2010">Journal of
|
||
Neuroscience</A></I>, 30(9):3531-43.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/v1_response_properties.pdf">PDF</A><A
|
||
class="download" HREF = "https://codeberg.org/mwspratling/V1_ResponseProperties">Code</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
A simple model is shown to account for a large range of V1 classical, and
|
||
non-classical, receptive field properties including orientation-tuning,
|
||
spatial and temporal frequency tuning, cross-orientation suppression, surround
|
||
suppression, and facilitation and inhibition by flankers and textured
|
||
surrounds. The model is an implementation of the predictive coding theory of
|
||
cortical function and thus provides a single computational explanation for a
|
||
diverse range of neurophysiological findings. Furthermore, since predictive
|
||
coding can be related to the biased competition theory and is a specific
|
||
example of more general theories of hierarchical perceptual inference the
|
||
current results relate V1 response properties to a wider, more unified,
|
||
framework for understanding cortical function.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2009) <B>Learning posture invariant spatial
|
||
representations through temporal correlations.</B> <I><A HREF =
|
||
"http://dx.doi.org/10.1109/TAMD.2009.2038494"> IEEE Transactions on Autonomous
|
||
Mental Development</A></I>, 1(4):253-63.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/tamd09_preprint.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
A hierarchical neural network model is used to learn, without supervision,
|
||
sensory-sensory coordinate transformations like those believed to be encoded
|
||
in the dorsal pathway of the cerebral cortex. The resulting representations of
|
||
visual space are invariant to eye orientation, neck orientation, or posture in
|
||
general. These posture invariant spatial representations are learned using the
|
||
same mechanisms that have previously been proposed to operate in the cortical
|
||
ventral pathway to learn object representation that are invariant to
|
||
translation, scale, orientation, or viewpoint in general. This model thus
|
||
suggests that the same mechanisms of learning and development operate across
|
||
multiple cortical hierarchies.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>K. De Meyer and M. W. Spratling (2009) <B> A model of non-linear
|
||
interactions between cortical top-down and horizontal connections explains the
|
||
attentional gating of collinear facilitation.</B> <I><A HREF =
|
||
"http://dx.doi.org/10.1016/j.visres.2008.12.017">Vision Research</A></I>, 49(5):553-68.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/visres09_preprint.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Past physiological and psychophysical experiments have shown that attention can
|
||
modulate the effects of contextual information appearing outside the classical
|
||
receptive field of a cortical neuron. Specifically, it has been suggested that
|
||
attention, operating via cortical feedback connections, gates the effects of
|
||
long-range horizontal connections underlying collinear facilitation in cortical
|
||
area V1. This article proposes a novel mechanism, based on the computations
|
||
performed within the dendrites of cortical pyramidal cells, that can account for
|
||
these observations. Furthermore, it is shown that the top-down gating signal
|
||
into V1 can result from a process of biased competition occurring in
|
||
extrastriate cortex. A model based on these two assumptions is used to replicate
|
||
the results of physiological and psychophysical experiments on collinear
|
||
facilitation and attentional modulation.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling, K. De Meyer and R. Kompass (2009)
|
||
<B>Unsupervised learning of overlapping image components using divisive input modulation.</B>
|
||
<I><A HREF = "http://www.hindawi.com/journals/cin/aip.381457.html">
|
||
Computational Intelligence and Neuroscience</A></I>, 2009(381457):1-19.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dim.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/DivisiveInputModulation">Code</A><div class="download">Abstract</div><div
|
||
class="abstract">
|
||
This paper demonstrates that non-negative matrix factorisation is mathematically
|
||
related to a class of neural networks that employ negative feedback as a
|
||
mechanism of competition. This observation inspires a novel learning algorithm
|
||
which we call Divisive Input Modulation (DIM). The proposed algorithm provides a
|
||
mathematically simple and computationally efficient method for the unsupervised
|
||
learning of image components, even in conditions where these elementary features
|
||
overlap considerably. To test the proposed algorithm, a novel artificial task is
|
||
introduced which is similar to the frequently-used bars problem but employs
|
||
squares rather than bars to increase the degree of overlap between
|
||
components. Using this task, we investigate how the proposed method performs on
|
||
the parsing of artificial images composed of overlapping features, given the
|
||
correct representation of the individual components; and secondly, we
|
||
investigate how well it can learn the elementary components from artificial
|
||
training images. We compare the performance of the proposed algorithm with its
|
||
predecessors including variations on these algorithms that have produced
|
||
state-of-the-art performance on the bars problem. The proposed algorithm is more
|
||
successful than its predecessors in dealing with overlap and occlusion in the
|
||
artificial task that has been used to assess performance.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2008)
|
||
<B>Predictive coding as a model of biased competition in visual attention.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1016/j.visres.2008.03.009">Vision Research</A></I>, 48(12):1391-408.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/visres08.pdf">PDF</A><A class="download" HREF =
|
||
"https://codeberg.org/mwspratling/PredictiveCodingAttention">Code</A><div class="download">Abstract</div><div
|
||
class="abstract">
|
||
Attention acts, through cortical feedback pathways, to enhance the response of
|
||
cells encoding expected or predicted information. Such observations are
|
||
inconsistent with the predictive coding theory of cortical function which
|
||
proposes that feedback acts to suppress information predicted by higher-level
|
||
cortical regions. Despite this discrepancy, this article demonstrates that the
|
||
predictive coding model can be used to simulate a number of the effects of
|
||
attention. This is achieved via a simple mathematical rearrangement of the
|
||
predictive coding model, which allows it to be interpreted as a form of biased
|
||
competition model. Nonlinear extensions to the model are proposed that enable it
|
||
to explain a wider range of data.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2008)
|
||
<B>Reconciling predictive coding and biased competition models of cortical function.</B>
|
||
<I><A HREF = "http://www.frontiersin.org/Journal/Abstract.aspx?s=237&name=Computational_Neuroscience&ART_DOI=10.3389/neuro.10.004.2008">Frontiers in Computational Neuroscience</A></I>, 2(4):1-8.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/pc_equals_bc.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
A simple variation of the standard biased competition model is shown, via some
|
||
trivial mathematical manipulations, to be identical to predictive coding.
|
||
Specifically, it is shown that a particular implementation of the biased
|
||
competition model, in which nodes compete via inhibition that targets the inputs
|
||
to a cortical region, is mathematically equivalent to the linear predictive
|
||
coding model. This observation demonstrates that these two important and
|
||
influential rival theories of cortical function are minor variations on the same
|
||
underlying mathematical model.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. S. C. Thomas, G. Westermann, D. Mareschal M. H. Johnson, S. Siros and M. W. Spratling (2008)
|
||
<B>Studying development in the 21st century</B> [response to commentaries].
|
||
<I><A HREF = "http://dx.doi.org/10.1017/S0140525X08004202">Behavioral and Brain Sciences</A></I>, 31(3):345-56.
|
||
<ul><div class="download">Abstract</div><div class="abstract">
|
||
In this response, we consider four main issues arising
|
||
from the commentaries to the target article. These include further
|
||
details of the theory of interactive specialization, the relationship
|
||
between neuroconstructivism and selectionism, the implications
|
||
of neuroconstructivism for the notion of representation, and the
|
||
role of genetics in theories of development. We conclude by
|
||
stressing the importance of multidisciplinary approaches in the
|
||
future study of cognitive development and by identifying
|
||
the directions in which neuroconstructivism can expand in the
|
||
Twenty-first Century.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>S. Siros, M. W. Spratling, M. S. C. Thomas, G. Westermann, D. Mareschal and
|
||
M. H. Johnson (2008)
|
||
<B>Précis of Neuroconstructivism: how the brain constructs cognition.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1017/S0140525X0800407X">Behavioral and Brain Sciences</A></I>, 31(3):321-31.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/bbs08_preprint.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Neuroconstructivism proposes a unifying framework for the study of development
|
||
that brings together (1) constructivism (which views development as the
|
||
progressive elaboration of increasingly complex structures), (2) cognitive
|
||
neuroscience (which aims to understand the neural mechanisms underlying
|
||
behaviour), and (3) computational modelling (which proposes formal and explicit
|
||
specifications of information processing). The guiding principle of our approach
|
||
is context dependence, within and (in contrast to Marr) between levels of
|
||
organization. We propose that three mechanisms guide the emergence of
|
||
representations: competition, cooperation, and chronotopy, which themselves
|
||
allow for two central processes: proactivity and progressive specialization. We
|
||
suggest that the main outcome of development is partial representations,
|
||
distributed across distinct functional circuits. This framework is derived by
|
||
examining development at the level of single neurons, brain systems, and whole
|
||
organisms. We use the terms encellment, embrainment, and embodiment to describe
|
||
the higher-level contextual influences that act at each of these levels of
|
||
organization. To illustrate these mechanisms in operation we provide case
|
||
studies in early visual perception, infant habituation, phonological
|
||
development, and object representations in infancy. Three further case studies
|
||
are concerned with interactions between levels of explanation: social
|
||
development, atypical development and within that, the development of
|
||
dyslexia. We conclude that cognitive development arises from a dynamic,
|
||
contextual change in neural structures leading to partial representations across
|
||
multiple brain regions and timescales.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>X. Zhang and M. W. Spratling (2008) <B>Automated learning of coordinate
|
||
transformations.</B> <I>Proceedings of the Eighth International Conference on <A
|
||
HREF = "http://www.lucs.lu.se/epirob/proceedings-epigenetic-robotics-2008/">Epigenetic Robotics: Modeling
|
||
Cognitive Development in Robotic Systems</A> (EPIROB08).</I>
|
||
|
||
|
||
|
||
<br><br>G. Westermann, D. Mareschal, M. H. Johnson, S. Siros, M. W. Spratling and M. S. C. Thomas (2007)
|
||
<B>Neuroconstructivism.</B><I> <A HREF = "http://dx.doi.org/10.1111/j.1467-7687.2007.00567.x">Developmental Science</A></I>, 10(1):75-83.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/dev_sci07.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
|
||
Neuroconstructivism is a theoretical framework focusing on the construction of
|
||
representation in the developing brain. Cognitive development is explained as
|
||
emerging from the experience-dependent development of neural structures
|
||
supporting mental representations. Neural development occurs in the context of
|
||
multiple interacting constraints acting on different levels, from the individual
|
||
cell to the external environment of the developing child. Cognitive development
|
||
can thus be understood as a trajectory originating from the constraints on the
|
||
underlying neural structures. This perspective offers an integrated view of
|
||
normal and abnormal development as well as of development and adult processing,
|
||
and it stands apart from traditional cognitive approaches in taking seriously
|
||
the constraints on cognition inherent by the substrate that delivers it.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
<br>L. A. Watling, M. W. Spratling, K. De Meyer and M. Johnson
|
||
(2007) <B>The role of feedback in the determination of figure and ground: a
|
||
combined behavioral and modeling study.</B> <I>Proceedings of the 29th Meeting of
|
||
the <A HREF = "http://www.cognitivesciencesociety.org/">Cognitive Science
|
||
Society</A> (CogSci07).</I>
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/watling_cogsci07.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
|
||
Object knowledge can exert on important influence on even the earliest stages of
|
||
visual processing. This study demonstrates how a familiarity bias, acquired only
|
||
briefly before testing, can affect the subsequent segmentation of an otherwise
|
||
ambiguous figure-ground array, in favor of perceiving the familiar shape as
|
||
figure. The behavioral data are then replicated using a biologically plausible
|
||
neural network model that employs feedback connections to implement the
|
||
demonstrated familiarity bias.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>D. Mareschal, M. H. Johnson, S. Siros, M. W. Spratling, M. S. C. Thomas
|
||
and G. Westermann (2007) <B><I><A HREF =
|
||
"http://www.oup.com/uk/catalogue/?ci=9780198529910"> Neuroconstructivism: How
|
||
the Brain Constructs Cognition</A></I></B>, Oxford University Press: Oxford,
|
||
UK.
|
||
|
||
|
||
<br><br>M. W. Spratling (2006)
|
||
<B>Learning image components for object recognition.</B>
|
||
<I><A HREF = "http://jmlr.csail.mit.edu/papers/v7/spratling06a.html">
|
||
Journal of Machine Learning Research</A></I>, 7:793-815.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/jmlr06.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
In order to perform object recognition it is necessary to learn representations
|
||
of the underlying components of images. Such components correspond to objects,
|
||
object-parts, or features. Non-negative matrix factorisation is a generative
|
||
model that has been specifically proposed for finding such meaningful
|
||
representations of image data, through the use of non-negativity constraints on
|
||
the factors. This article reports on an empirical investigation of the
|
||
performance of non-negative matrix factorisation algorithms. It is found that
|
||
such algorithms need to impose additional constraints on the sparseness of the
|
||
factors in order to successfully deal with occlusion. However, these constraints
|
||
can themselves result in these algorithms failing to identify image components
|
||
under certain conditions. In contrast, a recognition model (a competitive
|
||
learning neural network algorithm) reliably and accurately learns
|
||
representations of elementary image features without such constraints.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2006)
|
||
<B>A feedback model of perceptual learning and categorisation.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1080/13506280500168562">
|
||
Visual Cognition</A></I>, 13(2):129-65.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/vis_cog06.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Top-down, feedback, influences are known to have significant effects on visual
|
||
information processing. Such influences are also likely to affect perceptual
|
||
learning. This article employs a computational model of the cortical region
|
||
interactions underlying visual perception to investigate possible influences of
|
||
top-down information on learning. The results suggest that feedback could bias
|
||
the way in which perceptual stimuli are categorised and could also facilitate
|
||
the learning of sub-ordinate level representations suitable for object
|
||
identification and perceptual expertise.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2005)
|
||
<B>Learning viewpoint invariant perceptual representations from cluttered images.</B>
|
||
<I><A HREF = "http://doi.ieeecomputersociety.org/10.1109/TPAMI.2005.105">
|
||
IEEE Transactions on Pattern Analysis and Machine Intelligence</A></I>, 27(5):753-61.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/tpami05.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
In order to perform object recognition, it is necessary to form perceptual
|
||
representations that are sufficiently specific to distinguish between objects,
|
||
but that are also sufficiently flexible to generalise across changes in
|
||
location, rotation and scale. A standard method for learning perceptual
|
||
representations that are invariant to viewpoint is to form temporal associations
|
||
across image sequences showing object transformations. However, this method
|
||
requires that individual stimuli are presented in isolation and is therefore
|
||
unlikely to succeed in real-world applications where multiple objects can
|
||
co-occur in the visual input. This article proposes a simple modification to the
|
||
learning method, that can overcome this limitation, and results in more robust
|
||
learning of invariant representations.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (2004)
|
||
<B>Local versus distributed: a poor taxonomy of neural coding strategies</B> [commentary].
|
||
<I><A HREF="http://journals.cambridge.org/action/displayAbstract?aid=287732">
|
||
Behavioral and Brain Sciences</A></I>, 27(5):700-2.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/bbs_page.pdf">PDF</A>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2004)
|
||
<B>Neural coding strategies and mechanisms of competition.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1016/j.cogsys.2003.11.002">Cognitive Systems
|
||
Research</A></I>, 5(2):93-117.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cogsysres.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
A long running debate has concerned the question of whether neural
|
||
representations are encoded using a distributed or a local coding scheme. In
|
||
both schemes individual neurons respond to certain specific patterns of
|
||
pre-synaptic activity. Hence, rather than being dichotomous, both coding
|
||
schemes are based on the same representational mechanism. We argue that a
|
||
population of neurons needs to be capable of learning both local and distributed
|
||
representations, as appropriate to the task, and should be capable of generating
|
||
both local and distributed codes in response to different stimuli. Many neural
|
||
network algorithms, which are often employed as models of cognitive processes,
|
||
fail to meet all these requirements. In contrast, we present a neural network
|
||
architecture which enables a single algorithm to efficiently learn, and respond
|
||
using, both types of coding scheme.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2004)
|
||
<B>A feedback model of visual attention.</B>
|
||
<I><A HREF="http://dx.doi.org/10.1162/089892904322984526">
|
||
Journal of Cognitive Neuroscience</A></I>, 16(2):219-37.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/jocn04.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Feedback connections are a prominent feature of cortical anatomy and are likely
|
||
to have a significant functional role in neural information processing. We
|
||
present a neural network model of cortical feedback that successfully simulates
|
||
neurophysiological data associated with attention. In this domain our model can
|
||
be considered a more detailed, and biologically plausible, implementation of the
|
||
biased competition model of attention. However, our model is more general as it
|
||
can also explain a variety of other top-down processes in vision, such as
|
||
figure/ground segmentation and contextual cueing. This model thus suggests that
|
||
a common mechanism, involving cortical feedback pathways, is responsible for a
|
||
range of phenomena and provides a unified account of currently disparate areas
|
||
of research.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<!--</ul><br>R. Fukuda, M. W. Spratling, D. Mareschal and M. Johnson
|
||
(2003) <B>Modelling cortico basal-ganglionic loops and the development of
|
||
sequential information encoding.</B> <I>Proceedings of the 3rd International
|
||
Workshop on <A HREF = "http://www.epigenetic-robotics.org">Epigenetic
|
||
Robotics: Modeling Cognitive Development in Robotic Systems</A> (EPIROB03).</I>-->
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2003)
|
||
<B>Exploring the functional significance of dendritic inhibition in cortical pyramidal cells.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1016/S0925-2312(02)00847-0">Neurocomputing</A></I>, 52-54:389-95.
|
||
<!--Proceedings of the 11th <A HREF ="http://www.neuroinf.org/CNS.shtml">Computational Neuroscience Meeting</A> (CNS02)</I>.-->
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cns02.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Inhibitory synapses contacting the soma and axon initial segment are commonly
|
||
presumed to participate in shaping the response properties of cortical pyramidal
|
||
cells. Such an inhibitory mechanism has been explored in numerous computational
|
||
models. However, the majority of inhibitory synapses target the dendrites of
|
||
pyramidal cells, and recent physiological data suggests that this dendritic
|
||
inhibition affects tuning properties. We describe a model that can be used to
|
||
investigate the role of dendritic inhibition in the competition between
|
||
neurons. With this model we demonstrate that dendritic inhibition significantly
|
||
enhances the computational and representational properties of neural networks.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2002) <B>Exploring the functional
|
||
significance of dendritic inhibition in cortical pyramidal cells.</B>
|
||
<I>Proceedings of the 11th <A HREF ="http://www.neuroinf.org/CNS.shtml">Computational Neuroscience Meeting</A> (CNS02).</I> (Reprinted in the journal Neurocomputing, 2003; see above)
|
||
|
||
|
||
|
||
<br><br>M. W. Spratling (2002)
|
||
<B>Cortical region interactions and the functional role of apical dendrites.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1177/1534582302001003003">
|
||
Behavioral and Cognitive Neuroscience Reviews</A></I>, 1(3):219-28.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/apical_review.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
The basal and distal apical dendrites of pyramidal cells occupy distinct
|
||
cortical layers and are targeted by axons originating in different cortical
|
||
regions. Hence, apical and basal dendrites receive information from distinct
|
||
sources. Physiological evidence suggests that this anatomically observed
|
||
segregation of input sources may have functional significance. This possibility
|
||
has been explored in various connectionist models that employ neurons with
|
||
functionally distinct apical and basal compartments. A neuron in which separate
|
||
sets of inputs can be integrated independently has the potential to operate in a
|
||
variety of ways which are not possible for the conventional model of a neuron in
|
||
which all inputs are treated equally. This article thus considers how
|
||
functionally distinct apical and basal dendrites can contribute to the
|
||
information processing capacities of single neurons and, in particular, how
|
||
information from different cortical regions could have disparate affects on
|
||
neural activity and learning.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2002)
|
||
<B>Pre-integration lateral inhibition enhances unsupervised learning.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1162/089976602320264033">Neural
|
||
Computation</A></I>, 14(9):2157-79.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/neurocomp.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
|
||
A large and influential class of neural network architectures use
|
||
post-integration lateral inhibition as a mechanism for competition. We argue
|
||
that these algorithms are computationally deficient in that they fail to
|
||
generate, or learn, appropriate perceptual representations under certain
|
||
circumstances. An alternative neural network architecture is presented in which
|
||
nodes compete for the right to receive inputs rather than for the right to
|
||
generate outputs. This form of competition, implemented through pre-integration
|
||
lateral inhibition, does provide appropriate coding properties and can be used
|
||
to efficiently learn such representations. Furthermore, this architecture is
|
||
consistent with both neuro-anatomical and neuro-physiological data. We thus
|
||
argue that pre-integration lateral inhibition has computational advantages over
|
||
conventional neural network architectures while remaining equally biologically
|
||
plausible.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2001)
|
||
<B>Dendritic inhibition enhances neural coding properties.</B>
|
||
<I><A HREF = "https://doi.org/10.1093/cercor/11.12.1144">Cerebral Cortex</A></I>, 11(12):1144-9.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cereb_cort.pdf">PDF</A><div class="download">Abstract</div><div class="abstract">
|
||
The presence of a large number of inhibitory contacts at the soma and axon
|
||
initial segment of cortical pyramidal cells has inspired a large and influential
|
||
class of neural network model which use post-integration lateral inhibition as a
|
||
mechanism for competition between nodes. However, inhibitory synapses also
|
||
target the dendrites of pyramidal cells. The role of this dendritic inhibition
|
||
in competition between neurons has not previously been addressed. We
|
||
demonstrate, using a simple computational model, that such pre-integration
|
||
lateral inhibition provides networks of neurons with useful representational and
|
||
computational properties which are not provided by post-integration
|
||
inhibition.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>S. J. Grice, M. W. Spratling, A. Karmiloff-Smith, H. Halit, G. Csibra, M. de Haan and M. H. Johnson (2001) <B>Disordered visual processing and oscillatory brain activity in autism and Williams Syndrome.</B>
|
||
<I><A HREF ="http://www.neuroreport.com/pt/re/neuroreport/abstract.00001756-200108280-00021.htm">NeuroReport</A></I>, 12(12):2697-700.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/disordered_gamma.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Two developmental disorders, autism and Williams Syndrome, are both commonly
|
||
described as having difficulties in integrating perceptual features, i.e.,
|
||
binding spatially separate elements into a whole. It is already known that
|
||
healthy adults and infants display electroencephalographic (EEG) gamma band
|
||
bursts (around 40Hz) when the brain is required to achieve such binding . Here
|
||
we explore gamma band EEG in autism and Williams Syndrome and demonstrate
|
||
differential abnormalities in the two phenotypes. We show that despite putative
|
||
processing similarities at the cognitive level, binding in Williams Syndrome and
|
||
autism can be dissociated at the neurophysiological level by different
|
||
abnormalities in underlying brain oscillatory activity. Our study is the first
|
||
to identify that binding related gamma EEG can be disordered in humans.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and M. H. Johnson (2001) <B>Activity-dependent processes in regional cortical specialization</B> [commentary]. <I><A HREF="http://dx.doi.org/10.1111/1467-7687.00159">
|
||
Developmental Science</A></I>, 4(2):153-4.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/kingsbury_finlay.html">HTML</A>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>G. Csibra, G. Davis, M. W. Spratling and M. H. Johnson (2000)
|
||
<B>Gamma oscillations and object processing in the infant brain.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1126/science.290.5496.1582">Science</A></I>, 290(5496):1582-5.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/infant_gamma.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
An enduring controversy in neuroscience concerns how the brain binds
|
||
together separately coded stimulus features to form unitary
|
||
representations of objects. Recent evidence has indicated a close link
|
||
between this binding process and 40Hz (gamma-band) oscillations
|
||
generated by localized neural circuits (1). In a separate line of
|
||
research, the ability of young infants to perceive objects as unitary
|
||
and bounded has become a central focus for debates about the
|
||
mechanisms of perceptual development (2). However, to date these
|
||
infant studies have been behavioural, and there have been few, if any,
|
||
paradigms involving direct measures of neural function. Here we
|
||
demonstrate for the first time that binding-related 40Hz oscillations
|
||
are evident in the infant brain around 8 months of age, the same age
|
||
as some behavioral studies indicate the onset of perceptual binding of
|
||
spatially separated static visual features. The discovery of
|
||
binding-related gamma in infants opens up a new vista for experiments
|
||
on postnatal functional brain development in infants.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and G. M. Hayes (2000)
|
||
<B>Learning synaptic clusters for non-linear dendritic processing.</B>
|
||
<I><A HREF = "http://dx.doi.org/10.1023/A:1009634821039">
|
||
Neural Processing Letters</A></I>, 11(1):17-27.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/sigma_pi.pdf">PDF</A ><A class="download" HREF =
|
||
"https://nms.kcl.ac.uk/michael.spratling/Doc/sigma_pi.ps.gz">Gzipped Postscript</A><div class="download">Abstract</div><div class="abstract">
|
||
Nonlinear dendritic processing appears to be a feature of biological neurons and
|
||
would also be of use in many applications of artificial neural networks. This
|
||
paper presents a model of an initially standard linear unit which uses
|
||
unsupervised learning to find clusters of inputs within which inactivity at one
|
||
synapse can occlude the activity at the other synapses.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (1999)
|
||
<B>Artificial Ontogenesis: A Connectionist Model of Development.</B>
|
||
PhD Thesis, <A HREF = "http://www.ed.ac.uk/">University of Edinburgh</A>.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/phd.pdf">PDF</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
This thesis suggests that ontogenetic adaptive processes are important for
|
||
generating intelligent behaviour. It is thus proposed that such processes, as
|
||
they occur in nature, need to be modelled and that such a model could be used
|
||
for generating artificial intelligence, and specifically robotic
|
||
intelligence. Hence, this thesis focuses on how mechanisms of intelligence are
|
||
specified.<br><br>
|
||
|
||
A major problem in robotics is the need to predefine the behaviour to be
|
||
followed by the robot. This makes design intractable for all but the simplest
|
||
tasks and results in controllers that are specific to that particular task and
|
||
are brittle when faced with unforeseen circumstances. These problems can be
|
||
resolved by providing the robot with the ability to adapt the rules it follows
|
||
and to autonomously create new rules for controlling behaviour. This solution
|
||
thus depends on the predefinition of how rules to control behaviour are to be
|
||
learnt rather than the predefinition of rules for behaviour themselves.<br><br>
|
||
|
||
Learning new rules for behaviour occurs during the developmental process in
|
||
biology. Changes in the structure of the cerebral cortex underly behavioural and
|
||
cognitive development throughout infancy and beyond. The uniformity of the
|
||
neocortex suggests that there is significant computational uniformity across the
|
||
cortex resulting from uniform mechanisms of development, and holds out the
|
||
possibility of a general model of development. Development is an interactive
|
||
process between genetic predefinition and environmental influences. This
|
||
interactive process is constructive: qualitatively new behaviours are learnt by
|
||
using simple abilities as a basis for learning more complex ones. The
|
||
progressive increase in competence, provided by development, may be essential to
|
||
make tractable the process of acquiring higher-level abilities.<br><br>
|
||
|
||
While simple behaviours can be triggered by direct sensory cues, more complex
|
||
behaviours require the use of more abstract representations. There is thus a
|
||
need to find representations at the correct level of abstraction appropriate to
|
||
controlling each ability. In addition, finding the correct level of abstraction
|
||
makes tractable the task of associating sensory representations with motor
|
||
actions. Hence, finding appropriate representations is important both for
|
||
learning behaviours and for controlling behaviours. Representations can be
|
||
found by recording regularities in the world or by discovering re-occurring
|
||
patterns through repeated sensory-motor interactions. By recording regularities
|
||
within the representations thus formed, more abstract representations can be
|
||
found. Simple, non-abstract, representations thus provide the basis for learning
|
||
more complex, abstract, representations.<br><br>
|
||
|
||
A modular neural network architecture is presented as a basis for a model of
|
||
development. The pattern of activity of the neurons in an individual network
|
||
constitutes a representation of the input to that network. This representation
|
||
is formed through a novel, unsupervised, learning algorithm which adjusts the
|
||
synaptic weights to improve the representation of the input data.
|
||
Representations are formed by neurons learning to respond to correlated sets of
|
||
inputs. Neurons thus became feature detectors or pattern recognisers. Because
|
||
the nodes respond to patterns of inputs they encode more abstract features of
|
||
the input than are explicitly encoded in the input data itself. In this way
|
||
simple representations provide the basis for learning more complex
|
||
representations. The algorithm allows both more abstract representations to be
|
||
formed by associating correlated, coincident, features together, and invariant
|
||
representations to be formed by associating correlated, sequential, features
|
||
together.<br><br>
|
||
|
||
The algorithm robustly learns accurate and stable representations, in a format
|
||
most appropriate to the structure of the input data received: it can represent
|
||
both single and multiple input features in both the discrete and continuous
|
||
domains, using either topologically or non-topologically organised nodes. The
|
||
output of one neural network is used to provide inputs for other networks. The
|
||
robustness of the algorithm enables each neural network to be implemented using
|
||
an identical algorithm. This allows a modular `assembly' of neural networks to
|
||
be used for learning more complex abilities: the output activations of a network
|
||
can be used as the input to other networks which can then find representations
|
||
of more abstract information within the same input data; and, by defining the
|
||
output activations of neurons in certain networks to have behavioural
|
||
consequences it is possible to learn sensory-motor associations, to enable
|
||
sensory representations to be used to control behaviour.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (1999)
|
||
<B>Pre-synaptic lateral inhibition provides a better architecture for self-organising neural networks.</B> <I><A HREF = "http://dx.doi.org/10.1088/0954-898X_10_4_301">Network: Computation in Neural Systems</A></I>, 10(4):285-301.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/factopol.pdf">PDF</A ><A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/factopol.ps.gz">Gzipped Postscript</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Unsupervised learning is an important property of the brain and of
|
||
many artificial neural networks. A large variety of unsupervised
|
||
learning algorithms have been proposed. This paper takes a different
|
||
approach in considering the architecture of the neural network rather
|
||
than the learning algorithm. It is shown that a self-organising neural
|
||
network architecture using pre-synaptic lateral inhibition enables a
|
||
single learning algorithm to find distributed, local, and topological
|
||
representations as appropriate to the structure of the input data
|
||
received. It is argued that such an architecture not only has
|
||
computational advantages but is a better model of cortical
|
||
self-organisation.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and G. M. Hayes (1998)
|
||
<B>Learning sensory-motor cortical mappings without training.</B>
|
||
<I>Proceedings of the <A HREF = "https://www.elen.ucl.ac.be/esann/proceedings/papers.php?ann=1998#ES1998-20">6th European Symposium on Artificial Neural Networks (ESANN)</A></I>.
|
||
M. Verleysen (ed.) pp. 339-44. D-facto Publications.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/cort_map.ps.gz">Gzipped Postscript</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
This paper shows how the relationship between two arrays of artificial
|
||
neurons, representing different cortical regions, can be learned. The
|
||
algorithm enables each neural network to self-organise into a topological map
|
||
of the domain it represents at the same time as the relationship between
|
||
these maps is found. Unlike previous methods learning is achieved without a
|
||
separate training phase; the algorithm which learns the mapping is also that
|
||
which performs the mapping.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and G. M. Hayes (1998)
|
||
<B>A self-organising neural network for modelling cortical development.</B>
|
||
<I>Proceedings of the <A HREF = "https://www.elen.ucl.ac.be/esann/proceedings/papers.php?ann=1998#ES1998-19">6th European Symposium on Artificial Neural Networks (ESANN)</A></I>.
|
||
M. Verleysen (ed.) pp. 333-8. D-facto Publications.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/map_alg.ps.gz">Gzipped Postscript</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
This paper presents a novel self-organising neural network. It has been
|
||
developed for use as a simplified model of cortical development. Unlike
|
||
many other models of topological map formation all synaptic weights start at
|
||
zero strength (so that synaptogenesis might be modelled). In addition, the
|
||
algorithm works with the same format of encoding for both inputs to and
|
||
outputs from the network (so that the transfer and recoding of information
|
||
between cortical regions might be modelled).
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (1997)
|
||
<B>Artificial Ontogenesis: Cognitive and Behavioural Development for Robots.</B>
|
||
Unpublished Departmental Discussion Paper,
|
||
<A HREF = "http://www.dai.ed.ac.uk/">Department of Artificial Intelligence</A>,
|
||
<A HREF = "http://www.ed.ac.uk/">University of Edinburgh</A>.
|
||
<ul>
|
||
<div class="download">Abstract</div><div class="abstract">
|
||
There are three classes of adaptive process (structural definition,
|
||
structural adjustment, and parameter adjustment) which appear to underly the
|
||
development of intelligence in nature. In artificial intelligence only two
|
||
of these processes are used; AI ignores development (structural adjustment).
|
||
While AI attempts to predefine explicit rules for behaviour, nature's
|
||
success in building complex creatures depends on predefining how rules to
|
||
control behaviour can be learned. It is the developmental processes in
|
||
biology through which such rules are learned. This proposal is to apply
|
||
mechanisms similar to those used in biological development to robots. This
|
||
will move robotics from `development' meaning design and production, towards
|
||
`development' in its biological sense meaning a process of growth and
|
||
progressive change. Defining the rules for development is design at a
|
||
meta-level to that currently used. It is proposed that the long process of
|
||
evolution used by nature to define these developmental processes might be
|
||
supplanted by another adaptive process, that of engineering, to more quickly
|
||
enable study of ontogenetic development.<br><br>
|
||
|
||
This project thus aims to apply techniques inspired by animal development to
|
||
engineering robot control systems. Specifically it is proposed that a
|
||
hierarchical control system, based on the cerebral cortex, is used and that
|
||
this develops through constructivist learning algorithms (ones in which the
|
||
interaction of a situated agent with its environment guides the creation of
|
||
cognitive machinery appropriate for representing and acting in that
|
||
environment). Such a robot would be provided with some innate, low-level,
|
||
behavioural abilities and through experience develop more complex behaviour.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling and R. Cipolla (1996)
|
||
<B>Uncalibrated visual servoing.</B>
|
||
<I>Proceedings of the <A HREF = "http://dx.doi.org/10.5244/C.10.52">
|
||
7th British Machine Vision Conference (BMVC)</A></I>.
|
||
R. B. Fisher and E. Trucco (eds.) pp. 545-54. BMVA.
|
||
<ul>
|
||
<A class="download" HREF = "https://nms.kcl.ac.uk/michael.spratling/Doc/vis_servo.ps.gz">Gzipped Postscript</A><div
|
||
class="download">Abstract</div><div class="abstract">
|
||
Visual servoing is a process to enable a robot to position a camera with
|
||
respect to known landmarks using the visual data obtained by the camera
|
||
itself to guide camera motion. A solution is described which requires very
|
||
little a priori information freeing it from being specific to a particular
|
||
configuration of robot and camera. The solution is based on closed loop
|
||
control together with deliberate perturbations of the trajectory to provide
|
||
calibration movements for refining that trajectory. Results from
|
||
experiments in simulation and on a physical robot arm (camera-in-hand
|
||
configuration) are presented.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>M. W. Spratling (1994)
|
||
<B>Learning the Mapping Between Sensor and Motor Spaces to Produce Hand-Eye Coordination.</B>
|
||
MSc Dissertation,
|
||
<A HREF = "http://www.dai.ed.ac.uk/">Department of Artificial Intelligence</A>,
|
||
<A HREF = "http://www.ed.ac.uk/">University of Edinburgh</A>.
|
||
<ul>
|
||
<div class="download">Abstract</div><div class="abstract">
|
||
Coordination between sensory inputs and motor actions is essential for
|
||
intelligent robotics. This dissertation considers the control of a
|
||
simple manipulator using sensory information to locate the target
|
||
position for the end-effector. The control mechanisms investigated all
|
||
form topographic maps of possible configurations of the manipulator
|
||
joints (the motor space) and the values of the sensor inputs (the
|
||
sensor space). Various methods are considered for learning to relate a
|
||
location on the sensor space map (which represents the target position
|
||
in the world) and the location in the motor space map which will
|
||
configure the manipulator to reach this target position. These methods
|
||
are analysed using a computer simulation and a suitable algorithm to
|
||
solve the hand-eye coordination problem is presented.
|
||
</div>
|
||
</ul>
|
||
|
||
|
||
|
||
<br>
|
||
M. M. Ross, M. W. Spratling, C. B. Kirkland and P. S. Story (1994)
|
||
<B>Measurement of microfog wetness in a model steam turbine using a miniature optical spectral extinction probe.</B>
|
||
<I>IMechE International Symposium on Optical Methods and Data Processing In Heat and Fluid Flow.</I> <!--14-15 April 1994, City University, London.-->
|
||
|
||
|
||
</ul>
|
||
</div>
|
||
<P>
|
||
<HR>
|
||
<div id="footer">
|
||
(c) M. W. Spratling
|
||
</div>
|
||
</body>
|
||
</html>
|