Please use this identifier to cite or link to this item:
|Title:||Echo state Gaussian process||Authors:||Demiris, Yiannis
Chatzis, Sotirios P.
|Keywords:||Communication, Networking & Broadcasting
Computing & Processing (Hardware/Software)
|Issue Date:||2011||Publisher:||IEEE Xplore||Source:||IEEE transactions on neural networks, 2011, Volume 22, Issue 9, Pages 1435-1445||Abstract:||Echo state networks (ESNs) constitute a novel approach to recurrent neural network (RNN) training, with an RNN (the reservoir) being generated randomly, and only a readout being trained using a simple computationally efficient algorithm. ESNs have greatly facilitated the practical application of RNNs, outperforming classical approaches on a number of benchmark tasks. In this paper, we introduce a novel Bayesian approach toward ESNs, the echo state Gaussian process (ESGP). The ESGP combines the merits of ESNs and Gaussian processes to provide a more robust alternative to conventional reservoir computing networks while also offering a measure of confidence on the generated predictions (in the form of a predictive distribution). We exhibit the merits of our approach in a number of applications, considering both benchmark datasets and real-world applications, where we show that our method offers a significant enhancement in the dynamical data modeling capabilities of ESNs. Additionally, we also show that our method is orders of magnitude more computationally efficient compared to existing Gaussian process-based methods for dynamical data modeling, without compromises in the obtained predictive performance||URI:||http://ktisis.cut.ac.cy/handle/10488/7224||ISSN:||1045-9227||DOI:||10.1109/TNN.2011.2162109||Rights:||© 2006 IEEE|
|Appears in Collections:||Άρθρα/Articles|
Show full item record
checked on Mar 9, 2017
WEB OF SCIENCETM
checked on Dec 5, 2016
Page view(s) 5012
checked on Mar 27, 2017
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.