dc.contributor.author | Ceni, A | |
dc.contributor.author | Ashwin, P | |
dc.contributor.author | Livi, L | |
dc.date.accessioned | 2019-03-04T11:02:35Z | |
dc.date.issued | 2019-03-23 | |
dc.description.abstract | Introduction: Machine learning provides fundamental tools both for scientific research and for the
development of technologies with significant impact on society. It provides methods that facilitate the
discovery of regularities in data and that give predictions without explicit knowledge of the rules governing
a system. However, a price is paid for exploiting such flexibility: machine learning methods are typically
black-boxes where it is difficult to fully understand what the machine is doing or how it is operating.
This poses constraints on the applicability and explainability of such methods. Methods: Our research
aims to open the black-box of recurrent neural networks, an important family of neural networks used for
processing sequential data. We propose a novel methodology that provides a mechanistic interpretation
of behaviour when solving a computational task. Our methodology uses mathematical constructs called
excitable network attractors, which are invariant sets in phase space composed of stable attractors and
excitable connections between them. Results and Discussion: As the behaviour of recurrent neural
networks depends both on training and on inputs to the system, we introduce an algorithm to extract
network attractors directly from the trajectory of a neural network while solving tasks. Simulations
conducted on a controlled benchmark task confirm the relevance of these attractors for interpreting the
behaviour of recurrent neural networks, at least for tasks that involve learning a finite number of stable
states and transitions between them. | en_GB |
dc.description.sponsorship | Engineering and Physical Sciences Research Council (EPSRC) | en_GB |
dc.identifier.citation | Published online 23 March 2019. | en_GB |
dc.identifier.doi | 10.1007/s12559-019-09634-2 | |
dc.identifier.grantnumber | EP/N014391/1 | en_GB |
dc.identifier.uri | http://hdl.handle.net/10871/36221 | |
dc.language.iso | en | en_GB |
dc.publisher | Springer (part of Springer Nature) | en_GB |
dc.rights | © The Author(s) 2019. Open Access. This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. | |
dc.subject | Recurrent neural networks | en_GB |
dc.subject | Dynamical systems | en_GB |
dc.subject | Network attractors | en_GB |
dc.subject | Bifurcations | en_GB |
dc.title | Interpreting recurrent neural networks behaviour via excitable network attractors | en_GB |
dc.type | Article | en_GB |
dc.date.available | 2019-03-04T11:02:35Z | |
dc.identifier.issn | 1866-9956 | |
dc.description | This is the author accepted manuscript. The final version is available from Springer via the DOI in this record. | en_GB |
dc.identifier.journal | Cognitive Computation | en_GB |
dc.rights.uri | http://www.rioxx.net/licenses/all-rights-reserved | en_GB |
dcterms.dateAccepted | 2019-03-02 | |
exeter.funder | ::Engineering and Physical Sciences Research Council (EPSRC) | en_GB |
rioxxterms.version | AM | en_GB |
rioxxterms.licenseref.startdate | 2019-03-02 | |
rioxxterms.type | Journal Article/Review | en_GB |
refterms.dateFCD | 2019-03-02T16:33:30Z | |
refterms.versionFCD | AM | |
refterms.dateFOA | 2019-04-04T13:47:33Z | |
refterms.panel | B | en_GB |