Show simple item record

Files in this item

Thumbnail

Item metadata

dc.contributor.advisorFoldiak, Peter
dc.contributor.authorEndres, Dominik M.
dc.coverage.spatial200en
dc.date.accessioned2007-02-09T11:28:26Z
dc.date.available2007-02-09T11:28:26Z
dc.date.issued2006-11-30
dc.identifier.urihttps://hdl.handle.net/10023/162
dc.description.abstractThe overarching purpose of the studies presented in this report is the exploration of the uses of information theory and Bayesian inference applied to neural codes. Two approaches were taken: Starting from first principles, a coding mechanism is proposed, the results are compared to a biological neural code. Secondly, tools from information theory are used to measure the information contained in a biological neural code. Chapter 3: The REC model proposed by Harpur and Prager codes inputs into a sparse, factorial representation, maintaining reconstruction accuracy. Here I propose a modification of the REC model to determine the optimal network dimensionality. The resulting code for unfiltered natural images is accurate, highly sparse and a large fraction of the code elements show localized features. Furthermore, I propose an activation algorithm for the network that is faster and more accurate than a gradient descent based activation method. Moreover, it is demonstrated that asymmetric noise promotes sparseness. Chapter 4: A fast, exact alternative to Bayesian classification is introduced. Computational time is quadratic in both the number of observed data points and the number of degrees of freedom of the underlying model. As an example application, responses of single neurons from high-level visual cortex (area STSa) to rapid sequences of complex visual stimuli are analyzed. Chapter 5: I present an exact Bayesian treatment of a simple, yet sufficiently general probability distribution model. The model complexity, exact values of the expectations of entropies and their variances can be computed with polynomial effort given the data. The expectation of the mutual information becomes thus available, too, and a strict upper bound on its variance. The resulting algorithm is first tested on artificial data. To that end, an information theoretic similarity measure is derived. Second, the algorithm is demonstrated to be useful in neuroscience by studying the information content of the neural responses analyzed in the previous chapter. It is shown that the information throughput of STS neurons is maximized for stimulus durations of approx. 60ms.en
dc.format.extent1337112 bytes
dc.format.mimetypeapplication/pdf
dc.language.isoenen
dc.publisherUniversity of St Andrews
dc.rightsCreative Commons Attribution-NonCommercial-ShareAlike 2.5 Generic
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/2.5/
dc.subjectInformation theoryen
dc.subjectBayesian methodsen
dc.subjectComputational neuroscienceen
dc.subject.lccQP356.E6
dc.subject.lcshNeurosciencesen
dc.subject.lcshInformation theoryen
dc.subject.lcshBayesian statistical decision theoryen
dc.subject.lcshVision--Mathematical modelsen
dc.titleBayesian and information-theoretic tools for neuroscienceen
dc.typeThesisen
dc.contributor.sponsorITAS-SYS Gbren
dc.type.qualificationlevelDoctoralen
dc.type.qualificationnamePhD Doctor of Philosophyen
dc.publisher.institutionThe University of St Andrewsen


The following licence files are associated with this item:

  • Creative Commons

This item appears in the following Collection(s)

Show simple item record

Creative Commons Attribution-NonCommercial-ShareAlike 2.5 Generic
Except where otherwise noted within the work, this item's licence for re-use is described as Creative Commons Attribution-NonCommercial-ShareAlike 2.5 Generic