Home | About | Journals | Submit | Contact Us | Français |

**|**NPG Open Access**|**PMC3195233

Formats

Article sections

- Abstract
- Results
- Discussion
- Methods
- Author contributions
- Additional information
- Supplementary Material
- References

Authors

Related links

Nature Communications

Nat Commun. 2011 September 13; 2: 468.

doi: 10.1038/ncomms1476

PMCID: PMC3195233

L. Appeltant,^{1} M.C. Soriano,^{2} G. Van der Sande,^{1} J. Danckaert,^{1} S. Massar,^{3} J. Dambre,^{4} B. Schrauwen,^{4} C.R. Mirasso,^{2} and I. Fischer^{a,}^{2}

Received 2010 December 23; Accepted 2011 August 12.

Copyright © 2011, Nature Publishing Group, a division of Macmillan Publishers Limited. All Rights Reserved.

This work is licensed under a Creative Commons Attribution-NonCommercial-Share Alike 3.0 Unported License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-sa/3.0/

This article has been cited by other articles in PMC.

Novel methods for information processing are highly desired in our information-driven society. Inspired by the brain's ability to process information, the recently introduced paradigm known as 'reservoir computing' shows that complex networks can efficiently perform computation. Here we introduce a novel architecture that reduces the usually required large number of elements to a single nonlinear node with delayed feedback. Through an electronic implementation, we experimentally and numerically demonstrate excellent performance in a speech recognition benchmark. Complementary numerical studies also show excellent performance for a time series prediction benchmark. These results prove that delay-dynamical systems, even in their simplest manifestation, can perform efficient information processing. This finding paves the way to feasible and resource-efficient technological implementations of reservoir computing.

Nonlinear systems with delayed feedback and/or delayed coupling, often simply put as 'delay systems', are a class of dynamical systems that have attracted considerable attention, both because of their fundamental interest and because they arise in a variety of real-life systems1. It has been shown that delay has an ambivalent impact on the dynamical behaviour of systems, either stabilizing or destabilizing them2. Often it is sufficient to tune a single parameter (for example, the feedback strength) to access a variety of behaviours, ranging from stable via periodic and quasi-periodic oscillations to deterministic chaos3. From the point of view of applications, the dynamics of delay systems is gaining more and more interest. While initially it was considered more as a nuisance, it is now viewed as a resource that can be beneficially exploited. One of the simplest possible delay systems consists of a single nonlinear node whose dynamics is influenced by its own output a time *τ* in the past. Such a system is easy to implement, because it comprises only two elements, a nonlinear node and a delay loop. A well-studied example is found in optics: a semiconductor laser whose output light is fed back to the laser by an external mirror at a certain distance4. In this article, we demonstrate how the rich dynamical properties of delay systems can be beneficially employed for processing time-dependent signals, by appropriately modifying the concept of reservoir computing.

Reservoir computing (RC)5,6,7,8,9,10 is a recently introduced, bio-inspired, machine-learning paradigm that exhibits state-of-the-art performance for processing empirical data. Tasks, which are deemed computationally hard, such as chaotic time series prediction7, or speech recognition11,12, amongst others, can be successfully performed. The main inspiration underlying RC is the insight that the brain processes information generating patterns of transient neuronal activity excited by input sensory signals13. Therefore, RC is mimicking neuronal networks.

Traditional RC implementations are generally composed of three distinct parts: an input layer, the reservoir and an output layer, as illustrated in Figure 1a. The input layer feeds the input signals to the reservoir via fixed random weight connections. The reservoir usually consists of a large number of randomly interconnected nonlinear nodes, constituting a recurrent network, that is, a network that has internal feedback loops. Under the influence of input signals, the network exhibits transient responses. These transient responses are read out at the output layer via a linear weighted sum of the individual node states. The objective of RC is to implement a specific nonlinear transformation of the input signal or to classify the inputs. Classification involves the discrimination between a set of input data, for example, identifying features of images, voices, time series and so on. To perform its task, RC requires a training procedure. As recurrent networks are notoriously difficult to train, they were not widely used until the advent of RC. In RC, this problem is resolved by keeping the connections fixed. The only part of the system that is trained are the output layer weights. Thus, the training does not affect the dynamics of the reservoir itself. As a result of this training procedure, the system is capable to generalize, that is, process unseen inputs or attribute them to previously learned classes.

To efficiently solve its tasks, a reservoir should satisfy several key properties. First, it should nonlinearly transform the input signal into a high-dimensional state space in which the signal is represented. This is achieved through the use of a large number of reservoir nodes that are connected to each other through the recurrent nonlinear dynamics of the reservoir. In practice, traditional RC architectures employ several hundreds/thousands of nonlinear reservoir nodes to obtain good performance. In Figure 2, we illustrate how such a nonlinear mapping to a high-dimensional state space facilitates separation (classification) of states14. Second, the dynamics of the reservoir should be such that it exhibits a fading memory (that is, a short-term memory): the reservoir state is influenced by inputs from the recent past, but independent of the inputs from the far past. This property is essential for processing temporal sequences (such as speech) for which only the recent history of the signal is important. Additionally, the results of RC computations must be reproducible and robust against noise. For this, the reservoir should exhibit sufficiently different dynamical responses to inputs belonging to different classes. At the same time, the reservoir should not be too sensitive: similar inputs should not be associated to different classes. These competing requirements define when a reservoir performs well. Typically, reservoirs depend on a few parameters (such as the feedback gain and so on) that must be adjusted to satisfy the above constraints. Experience shows that these requirements are satisfied when the reservoir operates (in the absence of input) in a stable regime, but not too far from a bifurcation point. Further introduction to RC, and in particular its connection with other approaches to machine learning, can be found in the Supplementary Discussion.

In this article, we propose to implement a reservoir computer in which the usual structure of multiple connected nodes is replaced by a dynamical system comprising a nonlinear node subjected to delayed feedback. Mathematically, a key feature of time-continuous delay systems is that their state space becomes infinite dimensional. This is because their state at time *t* depends on the output of the nonlinear node during the continuous time interval [*t*–*τ*, *t*[, with *τ* being the delay time. The dynamics of the delay system remains finite dimensional in practice15, but exhibits the properties of high dimensionality and short-term memory. Therefore, delay systems fulfil the demands required of reservoirs for proper operation. Moreover, they seem very attractive systems to implement RC experimentally, as only few components are required to build them. Here we show that this intuition is correct. Excellent performance on benchmark tasks is obtained when the RC paradigm is adapted to delay systems. This shows that very simple dynamical systems have high-level information-processing capabilities.

In this section, we present the conceptual basis of our scheme, followed by the main results obtained for the two tasks we considered: spoken digit recognition and dynamical system modelling. We start by presenting in Figure 1b the basic principle of our scheme. Within one delay interval of length *τ*, we define *N* equidistant points separated in time by *θ*=*τ*/*N*. We denote these *N* equidistant points as 'virtual nodes', as they have a role analogous to the one of the nodes in a traditional reservoir. The values of the delayed variable at each of the *N* points define the states of the virtual nodes. These states characterize the transient response of our reservoir to a certain input at a given time. The separation time *θ* among virtual nodes has an important role and can be used to optimize the reservoir performance. We chose *θ*<*T*, with *T* being the characteristic time scale of the nonlinear node. With this choice, the states of the virtual nodes become dependent on the states of neighbouring nodes. Interconnected in this way, the virtual nodes emulate a network serving as reservoir (Supplementary Discussion). We demonstrate in the following that the single nonlinear node with delayed feedback performs comparably to traditional reservoirs.

The virtual nodes are subjected to the time-continuous input stream **u**(*t*) or time-discrete input **u**(*k*) which can be a time-varying scalar variable or vector of any dimension *Q*. The feeding to the individual virtual nodes is achieved by serializing the input using time multiplexing. For this, the input stream **u**(*t*) or **u**(*k*) undergoes a sample and hold operation to define a stream **I**(*t*) that is constant during one delay interval *τ*, before it is updated. Thus, in our approach, the input to the reservoir is always discretized in time, no matter whether it stems from a time-continuous or time-discrete input stream. To define the coupling weights from the stream **I**(*t*) to the virtual nodes we introduce a random (NxQ) matrix **M**, in the following called mask (we recall that *N* is the number of virtual nodes). On carrying out the multiplication **J**(*t*_{0})=**M**×**I**(*t*_{0}) at a certain time *t*_{0}, we obtain a *N* dimensional vector **J**(*t*_{0}) that represents the temporal input sequence within the interval [*t*_{0}, *t*_{0}+*τ*]. Each virtual node is updated using the corresponding component of **J**(*t*_{0}). Alternatively, one can view *J*(*t*) as a continuous time scalar function that is constant over periods, corresponding to the separation *θ* of the virtual nodes. Figure 1c illustrates the above input preparation and masking procedure. After a period *τ*, the states of all virtual nodes have been updated and the new reservoir state is obtained. Subsequently, **I**(*t*_{0}) is updated to drive the reservoir during the next period of duration *τ*. For each period *τ*, the reservoir state is read out for further processing. A training algorithm assigns an output weight to each virtual node, such that the weighted sum of the states approximates the desired target value as closely as possible (see Supplementary discussion). The training of the read-out follows the standard procedure for RC5,7. The testing is then performed using previously unseen input data of the same kind of those used for training.

To demonstrate our concept, we have chosen the widely studied Mackey–Glass oscillator16. The model of this dynamical system with its characteristic delayed feedback term, extended to include the external input **J**(*t*), reads

1

with *X* denoting the dynamical variable, its derivative with respect to a dimensionless time *t*, and *τ* the delay in the feedback loop. The characteristic time scale of the oscillator, determining the decay of *X* in the absence of the delayed feedback term, has been normalized to *T*=1. The parameters *η* and *γ* represent feedback strength and input scaling, respectively. The value of *η* we use guarantees that the system operates in a stable fixed point in the absence of external input (*γ*=0). With input, however, the system might exhibit complex dynamics. The choice of the nonlinearity in equation (1) has two main advantages. First, it can be easily implemented by an analogue electronic circuit17, which allows for fine parameter tuning18. This is the physical realization we have employed for the experimental demonstration (Fig. 3). Second, the exponent *p* can be used to tune the nonlinearity if needed. However, we expect other nonlinear functions to perform equally well.

We now demonstrate by experiment and simulation, the capability of our single nonlinear delay-dynamical node to perform isolated spoken digits recognition. The isolated spoken digit recognition task, as introduced by Doddington and Schalk19, is generally accepted as a benchmark speech recognition task. Our experimental realization has a Mackey–Glass nonlinearity with an exponent of *p*=7 that is easily obtained with standard electronic components. We have verified through numerical simulations that a broad range of values of *p* yields similar results. The virtual node separation is set at *θ*=0.2 that offers optimal performance (Methods section), whereas the total number of virtual nodes is *N*=400. The specifics of the training procedure are detailed in the Supplementary Discussion and the Methods section. As an example, Figure 4 depicts the experimentally and numerically obtained classification performance of unknown samples as a function of *η* for *γ*=0.5, which has been chosen such that input and feedback signals are of the same order of magnitude. The classification performance is expressed in two ways: the word error rate (WER) that shows the percentage of words that have been wrongly classified, and the margin (distance) between the reservoir's best guess of the target and the closest competitor. It can be seen that an increase in margin corresponds to a decrease in WER. Our results show that there is a broad parameter range in *η* with good performance, with an optimum for both, margin and WER, around *η*=0.8. Note that the performance breaks down when *η* approaches 1. This is expected as it corresponds to the threshold of instability of the MG oscillator when there is no input (*γ*=0). At the optimum value of *η*, we obtain a WER as low as 0.2% in experiments and 0.14% in numerical simulations. This corresponds to only one misclassification in 500 words. These performance levels are comparable to or even better than those obtained with traditional RC composed of more than 1,200 nodes for which a WER of 4.3% was reported11, with a reservoir of 308 nodes for which more recently a WER of 0.2% was obtained12 and also with alternative approaches based on hidden Markov models that achieved a WER of 0.55% (ref. 20).

We now present results from numerical simulations demonstrating the computational capabilities of the single nonlinear delay-dynamical node for a second task, commonly used in RC literature: dynamical system modelling, in particular training the readout to reproduce a certain signal21,22. Specifically, we train the system to model the response to white noise of a discrete-time tenth order nonlinear auto-regressive moving average (NARMA) system, originally introduced in reference 23. The details of this benchmark are described in the Supplementary Discussion and in the Methods section. To quantify the performance of the reservoir, the normalized root mean square error (NRMSE) of the predicted value versus the value obtained from the NARMA model is used. Up to now, the best performance reported in traditional RC for a reservoir of *N*=100 nodes, is NRMSE=0.18 (ref. 12). If the reservoir is replaced by a shift register that contains the input, the minimal NRMSE is 0.4. NRMSE values below this level require a nonlinear reservoir. We have chosen a nonlinear exponent of *p*=1, resulting in a weak nonlinearity that allows for longer memory as compared with the previously used value of *p*=7 (Supplementary Discussion). Figure 5 depicts our numerical results in the *γ*−*η* plane. A large region with NRMSE<0.2 (dark green) has been obtained. In Figure 6, we depict the NRMSE as a function of the virtual node separation *θ*. The optimal value lies around *θ*=0.2. When *θ* is larger, there is not enough coupling between virtual nodes, and performance decreases. When *θ* is smaller, too much averaging causes decreased performance. The minimal normalized root mean square error is as low as NRMSE=0.15. We therefore achieved comparable performance to conventional RC, but with a much simpler architecture.

We have demonstrated, both in experiment and simulation, that a simple nonlinear dynamical system subject to delayed feedback can efficiently perform information processing. As a consequence, our simple scheme can replace the complex networks used in traditional RC. Moreover, to the best of our knowledge, this experiment represents the first hardware implementation of RC with results comparable to those obtained with state-of-the-art digital realizations (Supplementary Discussion).

To get good performance with our system, a number of parameters need to be adjusted. These include the feedback gain *η*, the input gain *γ*, the delay time *τ*, the separation of virtual nodes in the delay line *θ*, the type of nonlinearity (in our case the exponent *p* of the MG system), and the choice of input mask. These parameters have analogues with similar parameters used in traditional RC: feedback gain and input gain have similar roles to spectral radius and input scaling; the delay time *τ* is related to the number of nodes and the separation of the virtual nodes to the sparsity of the interconnection matrix; the type of nonlinearity can also be varied in traditional RC and so on. As demonstrated here, some of the parameters can vary significantly around certain optimal values and still yield very good results. As experience with systems such as ours grows, we expect–as in traditional RC–that good heuristics on what parameter values to use will emerge. We expect that delay reservoirs can be realized that are, within some restrictions, versatile for different tasks. Moreover, owing to their much simpler hardware implementation, specifically optimized solutions for certain tasks could make sense. From a fundamental point of view, the simplicity of our architecture should facilitate gaining a deeper understanding of the interplay of dynamical properties and reservoir performance.

Besides the fundamental aspect of understanding information processing capabilities of dynamical systems, our architecture also offers practical advantages. The reduction of a complex network to a single hardware node facilitates implementations enormously, because only few components are needed. Nevertheless, the use of delay dynamical systems implies certain constraints, because the feeding of the virtual nodes is carried out serially, in contrast to the parallel feeding of the nodes in traditional RC. This serial feeding procedure results in a slow-down of the information processing, compared with parallel feeding. This potential slow-down is compensated for by the much simpler hardware architecture of the reservoir, and by the fact that the read-out can be taken at a single point of the delay line. These simplifications will enable ultra-high-speed implementations, using high-speed components that would be too demanding or expensive to be used for many nodes. In particular, realizations based on electronics or photonics systems should be feasible using this simple scheme, including real-time processing capabilities. Moreover, we expect that compromises can be found concerning speed, performance and memory capacity by extending the concept to network motifs of delay-coupled elements. Ultimately, a novel information-processing paradigm might emerge.

In the spoken digit recognition task, the input dataset consists of a subset of the NIST TI-46 corpus24 with ten spoken digits (0–9), each one recorded ten times by five different female speakers. Hence, we have 500 spoken words, all sampled at 12.5 kHz. The input **u**(*k*) (with *k* the discretized time) for the reservoir is, in this case, a set of 86-dimensional state vectors (channels) with up to 130 time steps. Each of these inputs represents a spoken digit, preprocessed using a standard cochlear ear model25. To construct an appropriate target function, ten linear classifiers are trained, each representing a different digit of the dataset. The target function is −1 if the spoken word does not correspond to the sought digit, and +1 if it does. For every target, the time trace is averaged in time and a winner-takes-all approach is applied to select the actual digit.

To eliminate the impact of the specific division of the available data samples between regularization, training and testing, we used *n*-fold cross-validation. This means that the entire process of regularization, training and testing is repeated *n* times (with *n*=20) on the same data, but each time with a different assignment of data samples to each of the three stages. The reported performances are the mean across these *n* runs.

For the spoken digit recognition task, the mask consists of a random assignment of three values: 0.59, 0.41 and 0. The first two values have equal probability of being selected, whereas the third one is more likely to be selected. Using a zero mask value implies that some nodes are insensitive to certain channels, thus avoiding averaging of all the channels.

The NARMA10 task is one of the most widely used benchmarks in RC. It was introduced in21, and used in many other publications in the context of RC, for instance in references 23,26. For the NARMA10 task, the input *u*(*k*) of the system consists of scalar random numbers, drawn from a uniform distribution in the interval [0, 0.5] and the target *y*(*k*+1) is given by the following recursive formula

2

The input stream *J*(*t*) for the NARMA10 test is obtained from *u*_{k} according to the procedure discussed in the manuscript. For the regularization, training and test of the dynamical system modelling task, we used 4 samples with a length of 800 points each and twofold cross-validation. The input scaling for the mask consists of a random series of amplitudes of 0.1 and −0.1. The input signal, multiplied with the mask and the input scaling factor γ, feeds the Mackey-Glass node as in equation (1). The importance of the parameters γ, the feedback strength η, and the separation between virtual nodes *θ*, has been discussed above. In particular, *θ* has a crucial role. In Figure 6, we show the numerically obtained performance of the Mackey–Glass system for the NARMA10 test when scanning *θ*. The optimal point is found for virtual node separations of *θ*=0.2, in units of the characteristic time scale of the nonlinear node. For shorter separations, too much averaging takes place and the Mackey–Glass system does not respond to the external input. For larger separations, the connectivity among virtual nodes is lost and, consequently, also the memory with respect to previous input. For node separations *θ*>3, the NRMSE reaches a level of 0.4, which is the performance of a shift-register.

I.F., C.R.M., S.M., J.Dan., B.S., G.VdS., M.C.S. and J.Dam. have contributed to development and/or implementation of the concept. M.C.S. performed the experiments, partly assisted by L.A. and supervised by C.R.M. and I.F. L.A. performed the numerical simulations, supervised by G.VdS. and J.Dan. All authors contributed to the discussion of the results and to the writing of the manuscript.

**How to cite this article:** Appeltant, L. *et al*. Information processing using a single dynamical node as complex system. *Nat. Commun.* 2:468 doi: 10.1038/ncomms1476 (2011).

Supplementary Figures S1-S9, Supplementary Discussion and Supplementary References.

Click here to view.^{(338K, pdf)}

We thank the members of the PHOCUS consortium and the members of the IAP Photonics@Be for helpful discussions and I. Veretennicoff and G. Verschaffelt for careful reading of our manuscript. Stimulating discussions are acknowledged with Jan Van Campenhout by I.F. and J.Dan., and with Yvan Paquot and Marc Haelterman by S.M. This research was partially supported by the Belgian Science Policy Office, under grant IAP P6-10 'photonics@be', by FWO and FRS–FNRS (Belgium), MICINN (Spain) under projects FISICOS (FIS2007-60327) and DeCoDicA (TEC2009-14101) and by the European project PHOCUS (EU FET-Open grant: 240763). L.A. and G.VdS. are a PhD Fellow and a Postdoctoral Fellow of the Research Foundation-Flanders (FWO).

- Erneux T. Applied Delayed Differential Equations (Springer Science + Business Media, 2009).
- Schöll E. & Schuster H. G. Handbook of Chaos Control (2nd edn , Wiley-VCH, 2008).
- Ikeda K. & Matsumoto K. High-dimensional chaotic behavior in systems with time-delayed feedback. Physica D29, 223–235 (1987).
- Fischer I., Hess O., Elsässer W. & Göbel E. High-dimensional chaotic dynamics of an external cavity semiconductor laser. Phys. Rev. Lett. 73, 2188–2191 (1994). [PubMed]
- Jaeger H. The 'echo state' approach to analyzing and training recurrent neural networks. Technical Report GMD Report 148, German National Research Center for Information Technology (2001).
- Maass W., Natschläger T. & Markram H. Real-time computing without stable states: a new framework for neural computation based on perturbations. Neural Comput. 14, 2531–2560 (2002). [PubMed]
- Jaeger H. & Haas H. Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication. Science 304, 78–80 (2004). [PubMed]
- Verstraeten D., Schrauwen B., D'Haene M. & Stroobandt D. An experimental unification of reservoir computing methods. Neural Netw. 20, 391–403 (2007). [PubMed]
- Buonomano D. V. & Maass W. State-dependent computations: spatiotemporal processing in cortical network. Nat. Rev. Neurosci. 10, 113–125 (2009). [PubMed]
- Maass W., Joshi P. & Sontag E. Computational aspects of feedback in neural circuits. PLOS Comput. Biol. 3, 1–20 (2007). [PubMed]
- Verstraeten D., Schrauwen B., Stroobandt D. & Van Campenhout J. Isolated word recognition with the liquid state machine: a case study. Inform. Process. Lett. 95, 521–528 (2005).
- Verstraeten D., Schrauwen B. & Stroobandt D. Reservoir-based techniques for speech recognition. In Proceedings of IJCNN06, International Joint Conference on Neural Networks, 1050–1053 (2006).
- Rabinovich M., Huerta R. & Laurent G. Transient dynamics for neural processing. Science 321, 48–50 (2008). [PubMed]
- Cover T. Geometrical and statistical properties of systems of linear inequalities with applications in pattern recognition. IEEE Trans. Electron. 14, 326–334 (1965).
- Le Berre M., Ressayre E., Tallet A., Gibbs H. M., Kaplan D. L. & Rose M. H. Conjecture on the dimensions of chaotic attractors of delayed-feedback dynamical systems. Phys. Rev. A 35, 4020–4022 (1987). [PubMed]
- Mackey M. C. & Glass L. Oscillation and chaos in physiological control systems. Science 197, 287–289 (1977). [PubMed]
- Namajunas A., Pyragas K. & Tamasevicius A. An electronic analog of the Mackey-Glass system. Phys. Lett. A 201, 42–46 (1995).
- Sano S., Uchida A., Yoshimori S. & Roy R. Dual synchronization of chaos in Mackey-Glass electronic circuits with time-delayed feedback. Phys. Rev. E 75, 016207–. (2007). [PubMed]
- Doddington G. & Schalk T. Speech recognition: turning theory to practice. IEEE Spectrum 18, 26–32 (1981).
- Walker W. et al. . Sphinx-4: A Flexible Open Source Framework for Speech Recognition (Technical Report, Sun Microsystems, 2004).
- Jaeger H. Adaptive nonlinear system identification with echo state networks: in Advance in Neural Information Processing Systems (eds Becker, S., Thrun, S., Obermayer, K.) Vol 15, 593–600 (MIT Press, 2003).
- Steil J. In Lecture Notes in Computer Science Vol
**3697**, 43–48 (Springer, 2005). - Atiya A. F. & Parlos A. G. New results on recurrent network training: unifying the algorithms and accelerating convergence. IEEE T. Neural Netw. 11, 697–709 (2000). [PubMed]
- Texas Instruments-Developed 46-Word Speaker-Dependent Isolated Word Corpus (TI46), September 1991, NIST Speech Disc 7-1.1 (1 disc).
- Lyon R. F. A computational model of filtering, detection, and compression in the cochlea. Proc. IEEE-ICASSP'82, vol. 7, 1282–1285 (1982).
- Rodan A. & Tino P. Minimum complexity echo state network. IEEE T. Neural Netw. 22, 131–144 (2011). [PubMed]

Articles from NPG Open Access are provided here courtesy of **Nature Publishing Group**

PubMed Central Canada is a service of the Canadian Institutes of Health Research (CIHR) working in partnership with the National Research Council's national science library in cooperation with the National Center for Biotechnology Information at the U.S. National Library of Medicine(NCBI/NLM). It includes content provided to the PubMed Central International archive by participating publishers. |