Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Role of delay-times in delay-based photonic reservoir computing [Invited]

Open Access Open Access

Abstract

Delay-based reservoir computing has gained a lot of attention due to the relative simplicity with which this concept can be implemented in hardware. However, unnecessary constraints are commonly placed on the relationship between the delay-time and the input clock-cycle, which can have a detrimental effect on the performance. We review the existing literature on this subject and introduce the concept of delay-based reservoir computing in a manner that demonstrates that no predefined relationship between the delay-time and the input clock-cycle is required for this computing concept to work. Choosing the delay-times independent of the input clock-cycle, one gains an important degree of freedom. Consequently, we discuss ways to improve the computing performance of a reservoir formed by delay-coupled oscillators and show the impact of delay-time tuning in such systems.

© 2022 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

Corrections

Antonio Hurtado, Bruno Romeira, Bhavin Shastri, Zengguang Cheng, and Sonia Buckley, "Emerging Optical Materials, Devices and Systems for Photonic Neuromorphic Computing feature issue: publisher’s note," Opt. Mater. Express 12, 1945-1945 (2022)
https://opg.optica.org/ome/abstract.cfm?uri=ome-12-5-1945

13 April 2022: A typographical correction was made to the title.

1. Introduction

Reservoir Computing is a human brain-inspired machine learning scheme that is versatile, has fast training times, and utilises the intrinsic information-processing capacities of dynamical systems [1,2]. The simple training scheme that is employed in reservoir computing, which involves only training the output layer, avoids difficulties that typically arise in the training of recurrent neural networks, such as the vanishing gradient in time [3]. This makes reservoir computing particularly suited for hardware implementation [48]. In the context of optimising computation speeds and power consumption, optical implementations are of particular interest and were made feasible by the introduction of so-called delay-based reservoir computing in [9]. In [9] the authors drew the connection between time-delayed systems and networks to show that a single non-linear node with time-delayed feedback could serve as a reservoir in which the reservoir responses are multiplexed in time rather than in space. Various experimental implementations of this delay-based scheme have been realised, including in opto-electronic [913], optical [6,8,1421] and electromechanical [22] systems. Delay-based reservoir computers have demonstrated good performance in applications including time-series-predictions [18,20,2325], fast word recognition [26], hand-written digit recognition [19] and signal conditioning [8,21,27,28], while new proposals emphasize the future potential for applications in cyber-security, wireless communications [29] or telecommunication [30].

Despite the extensive research that has been carried out on delay-based reservoir computers since their inception in 2011 [9], there has been very little focus on the influence of the delay-times on the reservoir computing performance. In most studies one of two fixed values are chosen for the delay-time; either resonance with the input clock-cycle [9,18,31,32] or slightly larger than the input clock-cycle, commonly referred to as the desynchronised regime [11,15]. Furthermore, it is often stated that it is necessary for the delay-based reservoir computing concept, that the delay-time is chosen as one of these values. However, this is an unnecessary constraint that stems from viewing delay-based reservoirs as networks. In this article, we will introduce the concept of delay-based reservoir computing in a manner that demonstrates that there is no predefined relationship between the delay-time of the reservoir and the input clock-cycle. We will review delay-based reservoir computing in the context of the influence of the delay-time, highlighting the effect of resonances between the reservoir delay-time and the input clock-cycle on the computing performance. We start by reviewing the effect of tuning the delay in single-delay reservoir computing systems and subsequently show, how adding more delays allows for an even better tunability of the reservoir properties and can thus lead to performance improvement.

This work is structured as follows. First, we will give a brief introduction to the general reservoir computing concept in Sec. 2. In Sec. 3.1 we will introduce delay-based reservoir computing by first introducing the concept of time-multiplexed reservoir computing and then explaining why time-delay systems serve as suitable time-multiplexed reservoirs. Following this, we will review the existing literature on aspects of this topic in Sec. 3.23.4, Sec. 4 and 5. Then, in Sec. 6, we present new results for the two delay-coupled oscillators, where we show how the memory capacity can be tuned, and how choosing the correct delays can lead to a substantial performance improvement for the NARMA10 task.

2. Reservoir computing concept

There are many articles and reviews that give good and thorough introductions to the reservoir computing concept. Here we will give a brief introduction and refer the reader to [1,3336] for more in-depth discussions of the concepts and mathematical foundations.

A reservoir computer consists of three parts; the input layer, the reservoir, and the output layer. Figure 1 shows a sketch of this layout for the case of one-dimensional input and output layers (the concept can easily be extended to high-dimensional inputs and outputs, for example, as shown in [35]). In this sketch, we have depicted the reservoir as a recurrent neural network, as this was the original system for which the reservoir computing concept was conceived and allows for a direct comparison between reservoir computing and more traditional machine learning algorithms. However, we would like to emphasize that the reservoir does not need to be a network for reservoir computing to work. It is the training method that sets reservoir computing apart from other machine learning methods. Specifically, only the weights connecting the reservoir with the output layer are trained (connections highlighted in green in Fig. 1). The input weights (connections between the input layer and the reservoir) and any internal properties of the reservoir are kept fixed during training. Typically, the input weights are chosen randomly, as are the internal weights when the reservoir is a recurrent neural network.

 figure: Fig. 1.

Fig. 1. Sketch of a reservoir computer based on a recurrent neural network (echo state network). The green lines represent the connections that need to be trained, while the rest of the network remains unchanged. Concept taken from [1].

Download Full Size | PDF

In the training phase, the reservoir is fed $K_\textrm {tr}$ successive inputs and the response of the reservoir is sampled $N$ times for each input. These responses (node states in the case of a neural network reservoir) are written into an $K_\textrm {tr}\times \left (N+1\right )$-dimensional state matrix $\underline {{\underline{\mathbf S}}}$, the last column of which is filled with a bias term of one. The training step is then to find the weights ${\underline {\mathbf W}}^{\textrm{out}} \in \mathbb {R}^{N+1}$ that minimise the difference between the output $\underline{{\mathbf o}}=\underline{{\underline{\mathbf S}}}^{~}{\underline{\mathbf W}}^{\textrm{out}} \in \mathbb {R}^{K_\textrm {tr}}$ and the target output $\hat{{\underline{\mathbf o}}}\in \mathbb {R}^{K_\textrm {tr}}$, i.e. the ${\underline{\mathbf W}}^{\textrm{out}}$ that solves

$$\min_{{\underline{\mathbf W}}^{\textrm{out}}}\left(||\underline{{\underline{\mathbf S}}}^{~} {\underline{\mathbf W}}^{\textrm{out}} - \hat{{\underline{\mathbf o}}}||^{2}_2 +\lambda||{\underline{\mathbf W}}^{\textrm{out}}||^{2}_2\right),$$
where $\lambda$ is the Tikhonov regularisation parameter and $||\cdot ||_2$ is the Euclidean norm. The solution to this problem can be found using the Moore-Penrose pseudoinverse and is given by
$${\underline{\mathbf W}}^{\textrm{out}}=(\underline{{\underline{\mathbf S}}}^{T}\underline{{\underline{\mathbf S}}}+\lambda \underline{{\underline{\mathbf I}}})^{{-}1} \underline{{\underline{\mathbf S}}}^{T} \hat{{\underline{\mathbf o}}}.$$
For the case of a multi-dimensional output, the target sequence and the output weight vectors become matrices and Eq. (2) is adjusted accordingly [35].

2.1 Performance measure

To quantify the quality of the prediction, $\underline{{\mathbf o}}=\underline{{\underline{\mathbf S}}}^{~}{\underline{\mathbf W}}^{\textrm{out}}$, there are various error measures (see for example [37,38]). One commonly used measure is the normalised root-mean-square error (NRMSE), defined as

$$\textrm{NRMSE}=\sqrt{\frac{\sum_{k'=1}^{K_{o}}\left( \hat{o}_{k'}-o_{k'}\right)^{2}}{K_{o} \textrm{var}\left( \underline{{ \hat{\mathbf o}}}\right)}},$$
where $\hat {o}_{k'}$ are the target values, $o_{k'}$ are the outputs produced by the reservoir computer, $K_{o}$ is the length of the vector $\underline {{\hat{\mathbf o}}}$ and $\textrm {var}\left ( \underline {{\hat{\mathbf o}}} \right )$ is the variance of the target sequence.

3. Delay-based reservoir computing

3.1 Concept

We will introduce the concept of delay-based reservoir computing in a different manner to the usual approach. We do this to clarify a misconception about the relationship between the delay-time $\tau$ of the reservoir and the input clock-cycle $T$, the misconception being that these quantities must be equal, or more generally, that there is any predefined relationship between these quantities. In order to do this, we will first introduce the concept of time-multiplexed reservoir computing independently of delay-based systems and will subsequently show why time-delayed systems are suitable reservoirs for time-multiplexed reservoir computing.

3.1.1 Time-multiplexed reservoir computing

One of the main concepts underlying delay-based reservoir computing is that rather than multiplexing in space, as is done in conventional reservoir computing, one multiplexes in time. This means that the response of the reservoir is sampled multiple times as a function of time rather than as a function of space. In such a scheme, each piece of input data must be fed into the reservoir for a certain time interval, which we will refer to as the clock-cycle $T$. During that interval, the response of the system must be sampled a number of times. A sketch of an example input sequence and the corresponding reservoir responses are shown in Fig. 2(a) and Fig. 2(c), respectively. The number of times that the system is sampled (blue circles in Fig. 2(a)) gives the output dimension, commonly referred to as the virtual nodes $N_v$ in delay-based reservoir computing. It is analogous to the output dimension $N$ in the general reservoir computing scheme described in Section 2.

 figure: Fig. 2.

Fig. 2. Visualisation of time-multiplexing (sequential sampling indicated by blue dots) and masking. Left and right panel show unmasked and masked input (top row) while the respective system response is shown below. Time is normalized to the input cycle $T$.

Download Full Size | PDF

In principle, any dynamical system can be used as the reservoir for such a time-multiplexed scheme, however, the dynamics and characteristic time-scales of the system play an important role in the performance that can be achieved. It has been shown that systems generally perform well when the dynamics of the underlying autonomous system (i.e. the system not driven by the input data) are steady state dynamics but the system is close to a bifurcation leading to chaotic or other high-dimensional dynamics ([39,40]). As indicated in Fig. 2(c), if the reservoir relaxes back to its steady state within an input clock-cycle, multiple identical system responses are sampled. As these responses are not linearly independent, the effective output dimension of the reservoir is reduced. One way to mitigate this problem is to apply a mask to the input data. An example of a masked input is shown in Fig. 2(b). By disturbing the system within one input cycle $T$, the response can be diversified. If the length of the mask steps, typically labelled $\theta$, is chosen appropriately, the system can be prevented from relaxing to its steady state.

Typically, a sequence of step functions with randomly selected step heights, that repeat for each clock-cycle, are used for the masking procedure, as depicted in the example shown in Fig. 2(b). However, work has also been done on the performance of chaotic, multilevel, binary, or pattered masks [24,4143]. In the example shown in Fig. 2(b), the system is sampled at the end of each mask step. The choice of sampling position is arbitrary, however, it must be the same in each clock-cycle. It has also been shown that it can be beneficial to vary the lengths of the mask step and the sampling positions within one clock-cycle [4446]. Which type of mask results in the best performance is both system and task-dependent.

3.1.2 Time-multiplexed RC with a delay system

Having introduced the concept of time-multiplexed reservoir computing, we will now address why time-delayed systems are suitable as reservoirs in this scheme. For reservoir computing to yield good results, the responses of the reservoir to sufficiently different inputs must be linearly separable. For this to be possible, the state-space of the reservoir must be sufficiently high-dimensional. This can be achieved by having a large network of coupled nodes or by using one nonlinear node with time-delayed feedback. For delay-based reservoir computing, delay differential equations of the following form are often considered (see for example [9,26,35,47,48]):

$$\frac{dx\left(t\right)}{dt} ={-}x\left( t\right)+f \left[ x\left( t-\tau\right)+\eta J\left( t \right) \right],$$
where $f \left [ \cdot \right ]$ is a non-linear function, $J\left ( t \right )$ is the input function and $\eta$ is the input scaling. In a strict mathematical sense, such a time-delayed continuous system is infinitely-dimensional [49] as the initial conditions of the system must be defined over the entire time interval $[-\tau,0]$. In a practical sense, time-delayed systems do not behave as infinite-dimensional systems, but they can exhibit complex dynamics such as quasi-periodicity and chaos. Close to bifurcations leading to these dynamics, time-delayed systems can also exhibit complex and high-dimensional transient dynamics and can therefore perform well as time-multiplexed reservoirs. Furthermore, the delay term introduces an additional, tunable, timescale that directly influences the memory capabilities of the reservoir (see Sec. 6). Due to the relative simplicity with which time-delayed feedback can be implemented experimentally, especially in photonic systems, time-multiplexed reservoir computing using delayed-systems has been extensively researched since its introduction in [9].

In Fig. 3, we show a sketch of a time-delayed reservoir. The system is fed a masked input sequence and the response of the system is sampled at predefined time intervals. Here we would like to emphasise that the time-multiplexing procedure can be viewed completely independently of the time-delayed system, and there is no predefined relationship between the delay time $\tau$ and the clock-cycle $T$. However, resonances between $\tau$ and $T$ can have a significant impact on the performance that can be achieved, as will be discussed in Section 3.3.

 figure: Fig. 3.

Fig. 3. Sketch of a delay-based reservoir computer realized with a laser subjected to optical feedback, masked input, and time-multiplexed output.

Download Full Size | PDF

3.2 Viewing delay-based reservoirs as networks

When the delay-based reservoir computing scheme was introduced in [9] the authors introduced the concept of virtual nodes equidistantly distributed along the delay-line. They chose $T=\tau =\theta N_v$, where $N_v$ is the number of virtual nodes, and described how this system can be considered as a network. They considered the two cases $\theta < T'$ and $\theta >>T'$, where $T'$ is the characteristic timescale of the nonlinear node. In the first case, $\theta < T'$, the system is perturbed by each subsequent input step before it can relax to a steady state. This means that the transient state of the system for one input interval depends on its state in the previous intervals. This can be viewed as a unidirectional coupling between neighbouring virtual nodes. Figure 4(a) shows a sketch of the network that such a system emulates. In addition to the coupling via the nonlinear node dynamics, choosing $\tau =\theta N_v$ means that in such a network view of the system, the virtual nodes are also coupled with their state in the previous clock-cycle, as indicated by the self-feedback loops in Fig. 4(a).

 figure: Fig. 4.

Fig. 4. Network topologies as they arise in delay-based reservoir computing systems when the delay is restricted to $\tau =n \theta$ ($n\in \mathbb {N}$). In (a,b) the input cycle $T$ is chosen resonant to the delay, $\tau =T$, while $\tau =T+\theta$ in (c). The characteristic system time $T'$ is larger than $\theta$ in (a) while it is smaller than $\theta$ in (b,c).

Download Full Size | PDF

In the second case, $\theta >>T'$, the nonlinear node relaxes to its steady state within one $\theta$ mask interval, this means that there is no dependence on the state of the system during the previous mask step and the system can be described by a set of independent nodes with self-feedback (see Fig. 4(b)). In this scenario, for reservoirs of the form given by Eq. (4), the delay-differential equation describing the reservoir can be reduced to a set of recursive equations that describe the state of the virtual nodes. Let $x_k$ be $x\left ( t\right )$ at the end of the $k^{\textrm{th}}$ $\theta$ interval of the piece-wise constant masked input function $J\left ( t\right )$, i.e. $x_k=x\left ( t\right )$ for $t=\left ( k+1\right )\theta$ with $J\left ( t\right )=J_k$ for $k\theta < t \leq \left ( k+1\right )\theta$. Then, if the dynamics of the underlying autonomous system are steady state dynamics, for $\theta >>T'$, Eq. (4) can be written as

$$0 ={-}x_k+f\left[ x_{k-N_v}+\eta J_k\right].$$
Rewriting $J_k$ in terms of the input data sequence $I_{k'}$ and the mask values $M_n$, and relabelling $x_k$ in terms of the virtual nodes number $n\in [0,N_v)$ and the input index $k'$, Eq. (5) becomes
$$x_{k'}^{n}=f\left[ x_{k'-1}^{n}+\eta M_n I_{k'} \right].$$
If, instead of $\tau =\theta N_v$, the delay is chosen as $\tau =\theta \left (N_v+1\right )$, as was introduced in [11], then Eq. (6) becomes
$$x_{k'}^{n}=f\left[ x_{k'-1}^{n-1}+\eta M_n I_{k'} \right] \quad \textrm{for} \quad n\in \left[1,N_v-1\right], ~~\textrm{and}\quad x_{k'}^{0}=f\left[ x_{k'-2}^{N_v-1}+\eta M_0 I_{k'} \right] .$$
In this case, the network representation is equivalent to a unidirectional ring, as sketched in Fig. 4(c).

The equivalences between time-delayed systems and networks, and more generally the now well-known equivalence between time-delayed and spatially extended systems [50], lead to the inception of delay-based reservoir computing. However, continuing to view delay-based reservoirs as networks puts unnecessary constraints on the relationship between the delay-time $\tau$ and the clock-cycle $T$. Mainly because the direct equivalence is only applicable when $\tau =n\theta$ for $n\in \mathbb {N}$ and $\theta =T/N_v$ [48]. It is not necessary to view delay-based systems as networks for reservoir computing purposes and letting go of this idea could lead to improved performance. This is because the delay-time plays an important role in determining the memory of the reservoir and allowing this parameter to be tuned more freely could lead to better fulfilment of task specific memory requirements [47,51,52].

3.3 Delay and clock-cycle resonances

It is generally known that resonances between various characteristic time-scales can have an important influence on the dynamics of a system. This is a phenomenon that is observed in a multitude of systems, including neural oscillators [53,54], mode-locked lasers [55] and chemical oscillators [56], among many more. Sometimes these resonances are beneficial to the desired outcome and other times they have a detrimental effect. In the case of delay-based reservoir computing, resonances between the delay-time of the reservoir and the clock-cycle generally have a detrimental effect on the performance. This has been demonstrated in a number of publications [17,47,48,51,52,5759]. An example of $\tau$-$T$ resonances is shown in Fig. 5, wherein the total memory capacity and the NARMA10 NRMSE (see the Supplemental Material for the definitions of these two commonly used benchmarking tasks) are plotted as a function of $\tau$ and $T$. In Fig. 5(a), dips in the memory capacity are found at the main resonance $\tau =T$ and at higher order resonances $p\tau =qT$ for $p,q\in \mathbb {N}$. In Fig. 5(b) increases in the NRMSE are found at some of these resonances for the NARMA10 task. By inspecting the evolution of the performance in Fig. 5(b), it also becomes clear that the optimal NARMA10 performance is not achieved at the commonly used relation of $\tau =T+\theta$. Note that the detrimental effect of resonances affects different parts of the memory capacity differently. Since memory requirements are different for each task, the performance for some tasks may therefore not be reduced by the resonant case.

 figure: Fig. 5.

Fig. 5. Total memory capacity (a) and performance of NARMA10 prediction (NRMSE) (b) plotted as a function of clock-cycle $T$ and delay-time $\tau$. Dashed lines indicate resonances of rational ratio between the two timescales, i.e. $n T = m \tau$, with $n,m \in \mathbb {N}$. System and parameters as in [51].

Download Full Size | PDF

Two explanations for this resonance phenomenon were given in [48] and [52]. The first, taken from [48], translates the delay-based reservoir computer via an Euler-step method into an equivalent spatially extended network. This method works for a constrained set of delay differential equations. Looking at the connections of this equivalent network system, one sees a reduction of node interconnection for resonant setups. Thus, a resonant case yields a badly connected network, lowering the reservoir computation capabilities.

The second explanation, given in [52], uses an eigenvalue analysis and describes how the reservoir reacts to the inputs in the regime where the input is a small perturbation. They show that resonances between the delay-time $\tau$ and clock-cycle $T$ yield a multitude of overlapping half-circle rotations in the many complex planes of the delay-based reservoir computers infinite-dimensional phase space. As mentioned above, practically, the reservoir only has a finite number of accessible eigendirections, as all but a finite number of eigendirections decay infinitely fast. Thus, the remaining finite eigendirections yield the usable phase space of the reservoir to project the input information into. In the case of a long delay-time $\tau$ (long in comparison to the local timescale of the reservoir, which is the case in typical reservoir computing settings), the imaginary parts of all eigenvalues $\mu _{k}$ are approximately given by

$$\mu_{k} \approx \frac{\pi}{\tau}(2k - \nu),$$
with $k \in \mathbb {N}$ being the index of the $k$-th eigenvalue and $\nu \in \{0,1\}$ being a constant phase shift. The rotation of a trajectory in one of the $k$ complex planes during one clock-cycle $T$, yields $T \mu _{k} \approx \frac {T \pi }{\tau }(2k - \nu )$. This is the traversed angle of rotation during one input clock-cycle $T$. Setting the clock-cycle $T$ to some rational part of the delay-time $T= n \tau$ (i.e. in resonance with $n \in \mathbb {Q}$), the traversed angle will always be a rational multiple of $\pi$, with $T \mu _{k} \approx n \pi (2k - \nu )$. Thus, in this resonant case, the rotation is always a multiple of a half-circle rotation, which reduces the usable phase space (and therefore the reservoir computing performance) and leads to the resonance lines of lower performance found in Fig. 5.

3.4 Hardware implementation

There have been various hardware implementations of photonic delay-based reservoir computing showing good performance on a range of benchmarking tasks. We refer the reader to [34] for a general review of photonic reservoir computing, [35] for a review specifically about delay-based photonic reservoir computing, and [36] for a general review on the hardware implementation of reservoir computing, including time-delayed systems. Here we shall mention only selected papers.

Opto-electronic and all-optical implementations of delay-based reservoir computing have demonstrated good performance in benchmarking tasks such as spoken digit recognition and non-linear channel equalisation [11,14,15], at data processing rates up to 1 Gbyte/s [14]. However, in these experimental setups, results have only been reported for fixed feedback delay-times that are either equal to the clock-cycle [14] or chosen according to the so-called desynchronised setup with $\tau =T+\theta$ [11,15]. The results of the theoretical studies discussed in Section 3.3 indicate that further improvement in the performance could be achieved by freely tuning the feedback delay-time.

In [18] a reservoir consisting of a semiconductor laser with feedback from a short external cavity is studied. In this work, the authors aim to increase the number of virtual nodes under the constraint of viewing the reservoir as a network, meaning that they restrict the clock-cycle to $T=n\tau$ for $n\in \mathbb {N}$. This restriction, which is unnecessary for the time-multiplexed reservoir computing concept, is detrimental to the memory capacity of the reservoir. There are two factors that lead to decreased memory capacity when $T=n\tau$ for $n\in \mathbb {N}$. Firstly, the resonance between $T$ and $\tau$ decreases the memory capacity, as discussed in Section 3.3 and shown in [48]. Secondly, with each roundtrip in the feedback cavity the influence of past inputs decreases. Therefore, as $n$ is increased, the memory of previous clock-cycles decreases, as is demonstrated by the low memory capacities above the main resonance ($T=\tau$) in Fig. 5(a). The authors of [18] mitigate the low memory of their setup by including multiple past inputs in the input signal. A similar approach of using multiple delayed inputs was recently investigated theoretically in [59], showing promising results for task-specific performance enhancement in hardware implemented reservoirs. The authors of [20] experimentally realized an integrated setup of a delay-based reservoir computer, also with a short feedback delay which restricts the number of virtual nodes, and post-processing methods are investigated to improve the performance. In this study there is a small mismatch between clock-cycle $T$ and delay-time $\tau$.

4. Reservoir computing with multiple delays

An extension of the delay-based reservoir computing setup reviewed in the previous section is to include additional delay-lines. So far, there have only been a few works on this topic, which are mostly theoretical. However, one of the earliest works was an experimental study using an opto-electronic reservoir [60]. In [60] 15 time-delayed feedback lines were included. The delay-times were chosen as integer multiples of the virtual node separation, $\tau _n=n\theta$ for $n\in \mathbb {N}$, with the largest delay being equal to the clock-cycle, $\tau _{\textrm {max}}=T=N_v\theta$. The feedback weights were randomly selected and the equivalent network had sparse connectivity. The authors described their approach as a method of providing enhanced dynamical connectivity within a network view of the reservoir. Once again, the restrictions placed on the feedback delay-times in the setup used in [60] do not allow for tasks specific delay-tuning of the memory capacity.

In [57] a theoretical investigation of the influence of a second time-delayed feedback term is done. The first delay is chosen resonant with the clock-cycle, $\tau _1=T$, and the second is given by $\tau _2=n\theta$ for $n\in \mathbb {N}$, where $n$ is varied to find task specific optimal values. At resonances between $\tau _1$ and $\tau _2$, the authors find poor performance for the NARMA10 task. This holds with the results of [48] since $\tau _1=T$ in [57].

In [17] a semiconductor laser with dual optical feedback is simulated. For the first delay-line the desynchronised scheme, $\tau _1=T+\theta$, is chosen and the second delay is tuned. In this work, a strong dependence on $\tau _2$ is shown for the Santa Fe time series prediction task, as well as resonance effects between $\tau _1$ and $\tau _2$. Similarly, in [12] an opto-electronic dual feedback scheme is simulated, also with $\tau _1=T+\theta$ and tuned $\tau _2$. Here, a strong $\tau _2$ dependence and delay resonance effects are also demonstrated.

In each of the studies mentioned above, at least one of the delay-times is restricted to $\tau =T$ or $\tau =T+\theta$. Furthermore, the ranges over which the second delay-time is tuned do not exceed $4T$ in these studies. However, task specific memory requirements could demand much larger delays, as will be demonstrated for the NARMA10 task in Section 6. and as is indicated by the task-specific dependence of the performance on delayed inputs recently shown in [59].

5. Reservoir computing with multiple delay-coupled nodes

So far we have considered delay-based reservoirs consisting of one non-linear node with feedback, with the multiplexing done in time. Hybrid spatial- and time-multiplexing approaches can also be realised, for example by delay-coupling multiple non-linear nodes. The advantage of such an approach is that the output dimension can be increased without the need for increasing the clock-cycle ($\theta$ can not be made arbitrarily small due to the finite time needed for a dynamical system to respond to the input). There are a number of theoretical studies on this topic [31,32,6164], however in none of these studies is the influence of the coupling delay-times considered as a tunable parameter, rather the focus is on increasing the output dimension or investigating the influence of system parameters other than the delay-times.

In [31] a comparative study is carried out on the performance of two delay-coupled non-linear nodes and two uncoupled non-linear nodes with self-feedback where the outputs are collected into the same state vector. Various parameters of their chosen nonlinearity, as well as input and feedback scaling parameters, are scanned to find areas of optimal performance for various benchmarking tasks. However, the delay-times are kept fixed and equal to the clock-cycle. The authors find that it is both parameter and task dependent as to whether the performance is better in the coupled or the uncoupled system.

Variants of two semiconductor lasers delay-coupled in a ring configuration are considered in [6264], with either resonant delay-times ($\tau =T$) or the desynchronised case ($\tau =T+\theta$). In [63] the influence of slightly varying one of the delay-times is considered, however, only over a small range with the aim of showing that the performance does not degrade if it is not possible to make the two delays exactly equal in an experimental setting.

In [32] the authors study the performance of multiple uncoupled semiconductor lasers with self-feedback, where, as in [31], the outputs are collected into one state vector. The purpose of this work is to demonstrate a method of increasing the output dimension. All delay-times are chosen equal and resonant to the clock-cycle in this study. An experimental realisation of such a setup, consisting of two uncoupled semiconductor lasers, is demonstrated in [65]. Here the delay-times are also kept fixed; one equal to the clock-cycle and the other slightly larger. The authors report improved performance compared with a single laser configuration.

Finally, in [61] the effect of mixing spatial- and time-multiplexing was investigated. The authors set the total output dimension to a fixed value and varied the ratio of real and virtual nodes. They considered various delayed ring-coupling configurations for the real nodes. For all configurations, they found optimal performance for the NARMA10 task when the number of virtual nodes is larger than the number of real nodes. In this study the number of delay-lines and the length of the delay-lines varied with the proportion of virtual to real nodes, however, this was done to keep the ratio of the number of virtual nodes and the delay-times constant rather than to investigate the influence of varying delay-times.

As of yet, the potential performance enhancement that could be achieved by introducing multiple time scales via the delay-times, in multi-node systems, has not been thoroughly investigated.

6. Memory capacity tuning in systems with multiple time-delays

We will now motivate a promising direction for future research in delay-based reservoir computing which involves more than one delay-line and their respective tuning. Our results show that this tuning crucially influences the memory capacity and the performance for certain tasks. For demonstrating the impact of the delay-times we use a system of two coupled Stuart-Landau oscillators. We look at the influence of the delay-times introduced via the coupling of the oscillators, as well as via self-feedback. The results pertaining to the additional self-feedback are also applicable for a reservoir consisting of a single non-linear node with multiple feedback terms.

6.1 Coupled Stuart-Landau oscillators - model and simulation details

To investigate the influence of multiple delays and coupled nodes, we use a system of two delay-coupled Stuart-Landau oscillators. The Stuart-Landau oscillator is a generic system describing the behaviour near a Hopf bifurcation and is therefore suitable for investigating the influence of delayed coupling and delayed feedback in a general context. The corresponding equations of motion are

$$\begin{aligned} \dot{z}_{j}=\left(\lambda_{j} + \eta_{j} g_{j}(t)u(t) + i\omega_{j} + (\gamma_{j} + i\alpha_{j})|z_{j}|^{2}\right) z_{j} ~~ + ~~\kappa_{j}~ e^{i\phi_{j}} ~z_{j+1}\left(t-\tau_{j}\right) ~+ ~~~~~~~~~~~~~~~~\\ \kappa^{self}_{j} e^{i\phi^{self}_{j}}z_{j}\left(t-\tau^{self}_{j}\right) +~ D\xi(t). \end{aligned}$$
Here, the index $j$ labels the oscillator $j \in \left [1,2\right ]$, $\lambda _j$ denotes the pump rate, $\omega _j$ the frequency, $\gamma _j$ the nonlinearity parameter, $\alpha _j$ the re-scaled sheer, $\kappa _j$ the feedback strength, $\phi _j$ the feedback phase and $\tau _j$ the delay-time. Optional self-feedback is included with the self-feedback strength, phase and delay-time given by $\kappa ^{self}_j$, $\phi ^{self}_j$ and $\tau ^{self}_j$, respectively. $\xi (t)$ models Gaussian white noise with amplitude $D=10^{-8}$. The input is denoted as $\eta _j g_j(t)u(t)$, where $u(t)$ is the original (unmasked) input sequence, $g_j(t)$ denotes the $j^{\textrm{th}}$ mask function and $\eta _j$ the input strength. The mask functions were chosen to be piecewise constant random binary. It is important to note that each oscillator $j \in \left [1,2\right ]$ has a distinct mask. Figure 6(a) illustrates the delayed, ring-coupling topology, and Fig. 6(b) shows the ring-coupling with self-feedback.

 figure: Fig. 6.

Fig. 6. Scheme of two oscillators (blue circles) delay-coupled in a ring-topology without (a) and with (b) self-feedback. The coupling is unidirectional, with coupling strength $\kappa _j$ and delay $\tau _j$ ($j \in \left [1,2\right ]$). Both nodes are externally driven with constant pump rate $\lambda _j$ and time dependent input $I_j(t)= g_j(t)u(t)$ scaled with $\eta _j$.

Download Full Size | PDF

Simulations were done in C++ using standard libraries and the linear algebra library "Armadillo" [66]. The delayed Stuart Landau equations Eq. (9) were integrated using a Runge-Kutta 4th order algorithm with $\Delta t=0.01$. The matrix inverse was calculated via the Moore-Penrose-Pseudoinverse from the C++ linear algebra library "Armadillo". For all simulations, the reservoir was initialised by first simulating without input for an integration time of 1000$\tau _1$ and then simulating with input for a buffer time of 1000 input clock-cycles. In the training stage, 5000 input values were used for the NARMA10 task and 40000 for the memory capacity. In both cases, the testing procedure involved 5000 inputs. Before feeding the input into the reservoir, it was normalised via a linear transformation, such that $u(t)\in [-1,1]$. This increases comparability between the memory capacity calculations and NARMA10, and ensures that inputs do not become too large. To avoid over-fitting, for the NARMA10 task regularisation by noise was included. This means that the Tikhonov regularisation parameter introduced in Eq. (2) was set to $\lambda =0$ and instead all entries of the state matrix were disturbed by additive Gaussian white noise with standard deviation $D=10^{-8}$. Regularisation by noise has been shown to be equivalent to Tikhonov regularisation [67]. For the masks, random binary values in $\{0,1\}$ were chosen. The default parameters used for simulation are given in Table 1.

Tables Icon

Table 1. Default Dimensionless Parameter Values for the Numerical Simulations of the Reservoir Computing Performance of Two Oscillators (Index $j$) Described by Eq. (9).

6.2 Impact of coupling-delay and clock-cycle in ring-coupled systems

In this section, we will investigate the influence of a second feedback delay-time in a system of two ring-coupled nodes by allowing the delay-times $\tau _1$ and $\tau _2$ to be different. This setup corresponds to Fig. 6(a), i.e. Equation (9) with $\kappa ^{self}_{1,2}=0$.

As discussed in Section 3.3, previous works show that resonances between the delay-time $\tau$ and the clock-cycle $T$ decrease the memory capacity in systems with one delay [48,51]. These resonances occur at $m\tau =nT$ for $n,m\in \mathbb {N}$ (see Fig. 5(a)). Introducing a second delay via the ring-coupling scheme gives the same resonance structure when $\tau _1=\tau _2$. However, if we introduce a mismatch between the delays, i.e. $\tau _1\neq \tau _2$, resonances are found at rational multiples between the mean delay, $\tau _\textrm {mean}=\left (\tau _1+\tau _2\right )/2$, and the clock-cycle: $m\tau _\textrm {mean}=nT$ for $n,m\in \mathbb {N}$. This resonance structure is shown in Fig. 7. The resonances are especially pronounced at the points, where both delays are in resonance with the clock-cycle (intersection points of horizontal and vertical lines in Fig. 7), with the largest losses in the memory capacity being found at the resonances points where both delays are equal, i.e. $\tau _1=\tau _2=nT$ for $n\in \mathbb {N}$.

 figure: Fig. 7.

Fig. 7. Total memory capacity of two mutually-coupled oscillators (see Fig. 6(a) and Eq. (9) with $\kappa _j^{self}=0$), color coded as a function of the ring-delays $\tau _1$ and $\tau _2$. Dashed lines indicate $1/2(\tau _1+\tau _2)=n/m T$ (mean delay resonance-lines) with $n,m \in \mathbb {N}$, $\tau _{1,2}=n/m T$ and $\tau _1=\tau _2$, respectively. Resonances are especially pronounced at the intersections where both delays are in resonance.

Download Full Size | PDF

In order to understand why the resonances occur at the mean delay, it is useful to show how the delay-mismatched system can be rewritten in terms of a system with two identical delays. A system of $N$ oscillators in a delayed ring-topology can generically be described by

$$\dot{z}_j=f(z_j(t),I_j(t))+ g(z_{j+1}(t-\tau_j)),$$
where $I_j(t)=\eta _j g_j(t)u(t)$ is the masked injection function, $u(t)$ is the original (unmasked) input, $f(\cdot )$ is the nonlinearity and $g(\cdot )$ is the coupling term. For the Stuart-Landau oscillator system, one would have $f(z_j,I_j)=\left (\lambda _j + I_j (t) + i\omega _j + (\gamma _j + i\alpha _j)|z_j|^{2}\right ) z_j$ and $g(z_{j+1}(t-\tau _j))=k_j e^{i\phi _j}g(z_{j+1}(t-\tau _j))$. Note, that the ring-topology implies $z_{N+1} \equiv z_1$. Equation (10) can be transformed into a system of $N$ delay-coupled oscillators with identical delays via the following transformation to a new coordinate $v_j$ [68]:
$$v_j(t)=z_j(t-t'_j),~~\textrm{with}~~t'_{j}=(N-j)\cdot\tau_\textrm{mean}-\sum^{N-j}_{i=1}\tau_{N-i},$$
where $\tau _\textrm {mean}=\sum _i \tau _i/N$ denotes the mean delay. This leads to a system of delay-differential equations for the transformed coordinate $v_j(t)$ with identical time-delays:
$$\dot{v}_j=f(v_j,I(t-t'_j)) + g(v_{j+1}(t-\tau_\textrm{mean})).$$
This transformation shows that the ring-coupled system with mismatched delays is equivalent to a ring with identical delays, equal to the mean delay, however, with delayed input $I(t-t'_j)$. In the equivalent system, we expect resonances with the clock-cycle $T$ for $\tau _\textrm {mean}=\frac {\sum _i\tau _i}{N}=\frac {m}{n} T;~m,n \in N$. Consequently, these are the resonances found in the original system (see Fig. 7).

For the equivalent system, Eq. (12), the input term is shifted from $I_j(t)$ to $I_j(t-t'_j)$. This implies a general correspondence between ring-coupled systems with multiple delays and systems with delayed inputs. In [59] it was recently shown that an appropriately chosen additional delayed input can improve the reservoir computing performance of a delay-based reservoir for various benchmark tasks [59]. In order to show how a similar effect can be achieved in the ring-coupled system, it is necessary to look at the dependence on the delay-times of the individual contributions to the total memory capacity.

Figure 8 shows the linear memory capacity as a function of the mean delay and the number of steps into the past. In agreement with previous work, increasing the delay-time of a reservoir while keeping the clock-cycle constant, increases the long-time memory capabilities [51]. However, as the delay-times become much larger than the clock-cycle and the characteristic timescale of the non-linearity, gaps in the memory capacity begin to form. The input regions that can be remembered appear as oblique stripes in Fig. 8. If the delays are equal (Fig. 8(a)), their position is given by multiples of the ratio between delay and clock-cycle. More precisely, if $\tau \in [nT,(n+1)T[,n\in \mathbb {N}$, then the first areas with high memory (dark blue regions in Fig. 8) are found at $u_{-n}$, since each input is fed into the system for a time interval of length $T$. The width of those areas is determined by the internal timescale of the system.

 figure: Fig. 8.

Fig. 8. Impact of delay-mismatch: Linear memory capacity $C_{u_{-steps}}$, color coded as a function of mean delay and steps into the past, averaged over 10 masks. (a) Identical delays $\tau _1=\tau _2$, (b) Mismatched delays $\tau _1=\tau _2+400$,

Download Full Size | PDF

The gaps can be reduced by mismatching the delay-times, as can be seen by comparing Fig. 8(a), where $\tau _1=\tau _2$, with Fig. 8(b), where $\tau _1=\tau _2+400$. For increasing mean delays, more long-time memory contributions are present, with smaller gaps in the mismatched case. This phenomenon is valid for the linear and non-linear capacities (corresponding scans for the quadratic terms are shown in the supplemental material). In [69] a similar reduction of gaps in the memory capacity was achieved by coupling systems with multiple delays, where the authors used a so-called deep reservoir computing scheme. With our setup, we are able to achieve similar results in the less complex ring-coupling scheme.

The filling of the gaps in the memory capacity for mismatched delays can be understood qualitatively. We first note that the system without delay ($\tau _\textrm {mean}=0$) has some capability to memorise inputs for short distances into the past, due to the non-zero relaxation time of the dynamical nonlinearity. For the sake of simplicity, this contribution is neglected in the following argumentation. When a delay is introduced, each delay line $\tau _j$ "stores" information about previous inputs and possibly nonlinear transformations of them. For instance, if we consider a ring of two oscillators with $\tau _1=\tau _2=3T$, then the ring will store information about inputs $u(t-3T),u(t-6T),u(t-9T)$ and so on. Since $T$ is the input clock-cycle, this corresponds to inputs $u_{-3},u_{-6},u_{-9},\ldots$, clearly leaving some gaps of previous inputs, that cannot be remembered. On the other hand, if we choose a system with equal mean delay, but a mismatch in the delay, e.g. $\tau _1=2T,\tau _2=4T$, we obtain information about $u_{-2},u_{-4},u_{-6},u_{-8}$ and so forth, filling some of the gaps in the equal delay case.

Aside from being able to fill gaps in the memory capacity by mismatching the delays, Fig. 8 shows that the delays can be chosen such that the system fulfils particular memory requirements that could be beneficial for specific tasks. However, in the ring-topology, varying one of the delay-times always influences the mean delay, meaning that the effect of large differences in the two delays can not be fully utilised. Therefore, in Section 6.3 we include an additional self-feedback delay-line that allows for more tunability of the memory of the system.

6.3 Manipulating the range of accessible past inputs via additional delay-lines

In this section, we show how an additional self-feedback term can be used to tune the memory capacity of our system of two ring-coupled nodes and how this can be used to improve the performance of the NARMA10 task. The setup corresponds to Fig. 6(b), i.e. Equation (9) with $\kappa ^{self}_{1,2}\neq 0$. For simplicity, we choose parameters identical to the default ring-topology parameters for all nodes and set the self-feedback coupling parameters to the values of the ring-coupling. The self-delays are chosen equal, $\tau ^{self}=\tau _1^{self}=\tau _2^{self}$, and $\tau ^{self}$ is treated as a free parameter. By further choosing $\tau _1=\tau _2$, there are effectively two independent time-scales introduced by the delays. This setup is mathematically equivalent to a single nonlinear node with two delayed-feedback terms. Therefore, all delay-dependent effects shown below are also applicable to such a system.

Figure 9(a) shows the linear memory capacities, as a function of the number of steps into the past and the self delay-time $\tau ^{self}$ (the mean-time of the ring-coupling is constant). Dark regions indicate high memory. First, for small steps into the past, the memory capabilities do not change with $\tau ^{self}$ and appear as a vertical stripe with a width given by the ring-delay time (see Fig. 8(a)). Second, there are tilted regions in the Fig. 9(a) that are found at multiples of the self-delay time. Their position is completely determined by the value of the self-delay. Note that in the supplemental material, the quadratic memory capacities are presented, which show similar trends.

 figure: Fig. 9.

Fig. 9. Impact of self-delay: Linear memory capacity $C_{u_{-steps}}$ color coded as a function of the self-delay $\tau ^{self}$ and the steps into the past, averaged over 10 mask realisations for (a) $\tau _{1}=\tau _{2}=210$, and (b) $\tau _{1}=\tau _{2}=620$.

Download Full Size | PDF

The width of the regions with high memory can be tuned via the value of the ring-delay (higher ring-delays yield larger width). Nevertheless, the manipulation of the width is limited, because as we have seen in the previous section, high ring-delays lead to gaps in the memory capacity (see Fig. 8(a)). Such a case is depicted in Fig. 9(b), where a ring-delay above the threshold, where gaps in the memory capacity begin to occur, was chosen. Note, that for fundamental reasons, a reservoir computer cannot remember all previous inputs (fading memory property) but it is possible to tune which ones contribute to the total memory capacity.

Now, even if the memory capacity of a system is completely known, it still depends crucially on the task if high or low performance is achieved. As an example for a demonstration of the complex connection between both, we use the NARMA10 time series prediction task. In Fig. 10(a) the NARMA10 normalised root-mean-square error (NRMSE) is shown as a function of the self-delay. As can be seen, if the delays are chosen equal, $\tau ^{self}=\tau _{1}=\tau _2$, the NRMSE starts with moderate values of around 0.3. Increasing the self-delay, the NRMSE decreases to slightly above 0.1. This is to our knowledge unprecedented in the reservoir computing literature for this number of virtual nodes (200) and demonstrates that the delay-times are an important tuning parameter for improving task specific performance.

 figure: Fig. 10.

Fig. 10. (a) NARMA10 computing error (NRMSE) as a function of the self-delay time $\tau ^{self}$ in the all-to-all coupled topology displayed in Fig. 6(b), averaged over 10 mask realisations (error bars indicate standard deviation). (b) Memory capacity corresponding to input terms $u_{-i}u_{-i-9}$ as they occur in the NARMA10 series Eq. (13) as a function of the self delay-time and the recall step.

Download Full Size | PDF

Further, increasing the self-delay above $\tau ^{self}=2000$, which is 10 times the clock-cycle, leads to a sudden performance drop-off which we can explain by the specific memory requirement of the NARMA10 task where the input terms $u_{-i-9},u_{-i}$ have a dominant role in the defining series:

$$A_{i+1}=0.3 A_i + 0.05 A_i\left(\sum^{9}_{j=0} A_{i-j}\right) + 1.5 u_{i-9}u_i + 0.1.$$
In Fig. 10(b) we underline the importance of these specific memory contributions by plotting the capacities $C_{u_{-i-9},u_{-i}}$. It can be seen that the capacity corresponding to these terms first increases with increasing $\tau ^{self}$. Then, for too high self-delays, the system cannot remember the first of the relevant terms as the gaps have been shifted to this position. Consequently, we obtain the performance drop-off at these too high self-delays.

7. Conclusion

In this article, we have reviewed delay-based reservoir computing with a focus on the influence of feedback and coupling delay-times. We aim to bring to the attention of the wider reservoir computing community that the delay-times have the potential to play a far more important role in improving the performance of delay-based reservoir computers than they have up until now. Furthermore, we highlight that, although it was important for the conception of delay-based reservoir computing, there is no need to view delay-based reservoirs as networks. Viewing delay-based reservoirs as networks has, in fact, led to misconceptions about the relationship between the clock-cycle and the delay-times of the reservoir. By introducing the concept of time-multiplexed reservoir computing independent of time-delayed systems, we demonstrate that there is no predetermined relationship between the clock-cycle and the reservoir delay-times.

Various recent studies have shown that choosing the delay-time resonant with the clock-cycle is generally detrimental to the computing capabilities of the reservoir. The commonly used desynchronised case, where delay and clock-cycle only differ by one virtual node, also does not lead to task specific optimal performance. If instead, the delay-times of the reservoir are freely tuned, after fixing the input clock-cycle, delay-times can be chosen to match the memory requirements of a particular task. This does not affect the computational speed and could be of particular interest, when designing experiments targeting tasks with specific memory requirements.

Additionally, we have demonstrated the potential of a second delay line for optimizing the performance by considering a system of two delay-coupled oscillators. We have shown how varying these delay-times influences the linear memory capacity by filling in gaps in the linear memory capacity. By adding a self-feedback term, the memory properties can be tuned even more freely to achieve both long- and short-term memory. For a task such as NARMA10, which has very specific memory requirements, correctly tuning the delay-times leads to a significant improvement in the performance.

Funding

Deutsche Forschungsgemeinschaft (LU 1729/3-1, SFB910).

Acknowledgments

We acknowledge support by the Deutsche Forschungsgemeinschaft(DFG) in the framework of the SFB910 project B9 and grant number LU 1729/3-1.

Disclosures

The authors declare no conflicts of interest.

Data Availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Supplemental document

See Supplement 1 for supporting content.

References

1. H. Jaeger, “The ’echo state’ approach to analysing and training recurrent neural networks,” GMD Report 148, GMD - German National Research Institute for Computer Science (2001).

2. W. Maass, T. Natschläger, and H. Markram, “Real-time computing without stable states: A new framework for neural computation based on perturbations,” Neural Comput. 14(11), 2531–2560 (2002). [CrossRef]  

3. S. Hochreiter, “The vanishing gradient problem during learning recurrent neural nets and problem solutions,” Int. J. Unc. Fuzz. Knowl. Based Syst. 06(02), 107–116 (1998). [CrossRef]  

4. K. Dockendorf, I. Park, P. He, J. C. Principe, and T. B. DeMarse, “Liquid state machines and cultured cortical networks: The separation property,” BioSystems 95(2), 90–97 (2009). [CrossRef]  

5. P. Antonik, F. Duport, M. Hermans, A. Smerieri, M. Haelterman, and S. Massar, “Online training of an opto-electronic reservoir computer applied to real-time channel equalization,” IEEE Trans. Neural Netw. Learn. Syst. 28(11), 2686–2698 (2017). [CrossRef]  

6. K. Vandoorne, P. Mechet, T. Van Vaerenbergh, M. Fiers, G. Morthier, D. Verstraeten, B. Schrauwen, J. Dambre, and P. Bienstman, “Experimental demonstration of reservoir computing on a silicon photonics chip,” Nat. Commun. 5(1), 3541 (2014). [CrossRef]  

7. C. Fernando and S. Sojakka, “Pattern recognition in a bucket,” in Advances in Artificial Life, (2003), pp. 588–597.

8. S. Sackesyn, C. Ma, J. Dambre, and P. Bienstman, “Experimental realization of integrated photonic reservoir computing for nonlinear fiber distortion compensation,” Opt. Express 29(20), 30991–30997 (2021). [CrossRef]  

9. L. Appeltant, M. C. Soriano, G. Van der Sande, J. Danckaert, S. Massar, J. Dambre, B. Schrauwen, C. R. Mirasso, and I. Fischer, “Information processing using a single dynamical node as complex system,” Nat. Commun. 2(1), 468 (2011). [CrossRef]  

10. L. Larger, M. C. Soriano, D. Brunner, L. Appeltant, J. M. Gutierrez, L. Pesquera, C. R. Mirasso, and I. Fischer, “Photonic information processing beyond turing: an optoelectronic implementation of reservoir computing,” Opt. Express 20(3), 3241–3249 (2012). [CrossRef]  

11. Y. Paquot, F. Duport, A. Smerieri, J. Dambre, B. Schrauwen, M. Haelterman, and S. Massar, “Optoelectronic reservoir computing,” Sci. Rep. 2(1), 287 (2012). [CrossRef]  

12. Y. Chen, L. Yi, J. Ke, Z. Yang, Y. Yang, L. Huang, Q. Zhuge, and W. Hu, “Reservoir computing system with double optoelectronic feedback loops,” Opt. Express 27(20), 27431–27440 (2019). [CrossRef]  

13. F. Duport, A. Smerieri, A. Akrout, M. Haelterman, and S. Massar, “Fully analogue photonic reservoir computer,” Sci. Rep. 6(1), 22381 (2016). [CrossRef]  

14. D. Brunner, M. C. Soriano, C. R. Mirasso, and I. Fischer, “Parallel photonic information processing at gigabyte per second data rates using transient states,” Nat. Commun. 4(1), 1364 (2013). [CrossRef]  

15. A. Dejonckheere, F. Duport, A. Smerieri, L. Fang, J. L. Oudar, M. Haelterman, and S. Massar, “All-optical reservoir computer based on saturation of absorption,” Opt. Express 22(9), 10868 (2014). [CrossRef]  

16. Q. Vinckier, F. Duport, A. Smerieri, K. Vandoorne, P. Bienstman, M. Haelterman, and S. Massar, “High-performance photonic reservoir computer based on a coherently driven passive cavity,” Optica 2(5), 438–446 (2015). [CrossRef]  

17. Y. S. Hou, G. Q. Xia, W. Y. Yang, D. Wang, E. Jayaprasath, Z. Jiang, C. X. Hu, and Z. M. Wu, “Prediction performance of reservoir computing system based on a semiconductor laser subject to double optical feedback and optical injection,” Opt. Express 26(8), 10211–10219 (2018). [CrossRef]  

18. K. Takano, C. Sugano, M. Inubushi, K. Yoshimura, S. Sunada, K. Kanno, and A. Uchida, “Compact reservoir computing with a photonic integrated circuit,” Opt. Express 26(22), 29424–29439 (2018). [CrossRef]  

19. M. Nakajima, K. Tanaka, and T. Hashimoto, “Scalable reservoir computing on coherent linear photonic processor,” Commun. Phys. 4(1), 20 (2021). [CrossRef]  

20. K. Harkhoe, G. Verschaffelt, A. Katumba, P. Bienstman, and G. Van der Sande, “Demonstrating delay-based reservoir computing using a compact photonic integrated chip,” Opt. Express 28(3), 3086 (2020). [CrossRef]  

21. F. Duport, B. Schneider, A. Smerieri, M. Haelterman, and S. Massar, “All-optical reservoir computing,” Opt. Express 20(20), 22783–22795 (2012). [CrossRef]  

22. G. Dion, S. Mejaouri, and J. Sylvestre, “Reservoir computing with a single delay-coupled non-linear mechanical oscillator,” J. Appl. Phys. 124(15), 152132 (2018). [CrossRef]  

23. J. Bueno, D. Brunner, M. C. Soriano, and I. Fischer, “Conditions for reservoir computing performance using semiconductor lasers with delayed optical feedback,” Opt. Express 25(3), 2401–2412 (2017). [CrossRef]  

24. Y. Kuriki, J. Nakayama, K. Takano, and A. Uchida, “Impact of input mask signals on delay-based photonic reservoir computing with semiconductor lasers,” Opt. Express 26(5), 5777–5788 (2018). [CrossRef]  

25. R. M. Nguimdo, G. Verschaffelt, J. Danckaert, and G. Van der Sande, “Fast photonic information processing using semiconductor lasers with delayed optical feedback: Role of phase dynamics,” Opt. Express 22(7), 8672–8686 (2014). [CrossRef]  

26. L. Larger, A. Baylón-Fuentes, R. Martinenghi, V. S. Udaltsov, Y. K. Chembo, and M. Jacquot, “High-speed photonic reservoir computing using a time-delay-based architecture: Million words per second classification,” Phys. Rev. X 7(1), 011015 (2017). [CrossRef]  

27. A. Argyris, J. Cantero, M. Galletero, E. Pereda, C. R. Mirasso, I. Fischer, and M. C. Soriano, “Comparison of photonic reservoir computing systems for fiber transmission equalization,” IEEE J. Sel. Top. Quantum Electron. 26(1), 1–9 (2020). [CrossRef]  

28. I. Estébanez, S. Li, J. Schwind, I. Fischer, S. Pachnicke, and A. Argyris, “56 gbaud pam-4 100 km transmission system with photonic processing schemes,” ArXiv 2105.07990 (2021).

29. K. Hamedani, Z. Zhou, K. Bai, and L. Liu, “The Novel Applications of Deep Reservoir Computing in Cyber-Security and Wireless Communication,” (2019).

30. M. Sorokina, S. Sergeyev, and S. Turitsyn, “Fiber echo state network analogue for high-bandwidth dual-quadrature signal processing,” Opt. Express 27(3), 2387–2395 (2019). [CrossRef]  

31. S. Ortín and L. Pesquera, “Reservoir computing with an ensemble of time-delay reservoirs,” Cognitive Comput. 9(3), 327–336 (2017). [CrossRef]  

32. C. Sugano, K. Kanno, and A. Uchida, “Reservoir computing using multiple lasers with feedback on a photonic integrated circuit,” IEEE J. Sel. Top. Quantum Electron. 26(1), 1–9 (2020). [CrossRef]  

33. M. Lukosevicius and H. Jaeger, “Reservoir computing approaches to recurrent neural network training,” Comput. Sci. Rev. 3(3), 127–149 (2009). [CrossRef]  

34. G. Van der Sande, D. Brunner, and M. C. Soriano, “Advances in photonic reservoir computing,” Nanophotonics 6(3), 561–576 (2017). [CrossRef]  

35. D. Brunner, B. Penkovsky, B. A. Marquez, M. Jacquot, I. Fischer, and L. Larger, “Tutorial: Photonic neural networks in delay systems,” J. Appl. Phys. 124(15), 152004 (2018). [CrossRef]  

36. G. Tanaka, T. Yamane, J. B. Héroux, R. Nakane, N. Kanazawa, S. Takeda, H. Numata, D. Nakano, and A. Hirose, “Recent advances in physical reservoir computing: A review,” Neural Netw. 115, 100–123 (2019). [CrossRef]  

37. A. Botchkarev, “A new typology design of performance metrics to measure errors in machine learning regression algorithms,” Interdiscip. J. Inf. Knowl. Manag. 14, 045–076 (2019). [CrossRef]  

38. P. Flach, “Performance evaluation in machine learning: The good, the bad, the ugly, and the way forward,” AAAI-19 Proc. 33, 9808–9814 (2019). [CrossRef]  

39. K. Nakajima and I. Fischer, “Reservoir Computing: Theory, Physical Implementations, and Applications,” (2021).

40. T. L. Carroll, “Do reservoir computers work best at the edge of chaos?” Chaos 30(12), 121109 (2020). [CrossRef]  

41. L. Appeltant, G. Van der Sande, J. Danckaert, and I. Fischer, “Constructing optimized binary masks for reservoir computing with delay systems,” Sci. Rep. 4(1), 3629 (2015). [CrossRef]  

42. M. C. Soriano, S. Ortín, D. Brunner, L. Larger, C. R. Mirasso, I. Fischer, and L. Pesquera, “Optoelectronic reservoir computing: tackling noise-induced performance degradation,” Opt. Express 21(1), 12–20 (2013). [CrossRef]  

43. A. Argyris, J. Schwind, and I. Fischer, “Fast physical repetitive patterns generation for masking in time-delay reservoir computing,” Sci. Rep. 11(1), 6701 (2021). [CrossRef]  

44. M. Goldmann, C. R. Mirasso, I. Fischer, and M. C. Soriano, “Exploiting transient dynamics of a time-multiplexed reservoir to boost the system performance,” in 2021 International Joint Conference on Neural Networks (IJCNN), (2021), pp. 1–8.

45. D. Z. Yue, Z. M. Wu, Y. S. Hou, B. Cui, Y. Jin, M. Dai, and G. Q. Xia, “Performance optimization research of reservoir computing system based on an optical feedback semiconductor laser under electrical information injection,” Opt. Express 27(14), 19931–19939 (2019). [CrossRef]  

46. H. Toutounji, S. Schumacher, and G. Pipa, “Homeostatic plasticity for single node delay-coupled reservoir computing,” Neural Comput. 27(6), 1159–1185 (2015). [CrossRef]  

47. A. Röhm, L. C. Jaurigue, and K. Lüdge, “Reservoir computing using laser networks,” IEEE J. Sel. Top. Quantum Electron. 26(1), 1–8 (2020). [CrossRef]  

48. F. Stelzer, A. Röhm, K. Lüdge, and S. Yanchuk, “Performance boost of time-delay reservoir computing by non-resonant clock cycle,” Neural Netw. 124, 158–169 (2020). [CrossRef]  

49. J. K. Hale and S. M. Verduyn Lunel, Introduction to Functional Differential Equations (Springer, New York, 1993).

50. F. T. Arecchi, G. Giacomelli, A. Lapucci, and R. Meucci, “Two-dimensional representation of a delayed dynamical system,” Phys. Rev. A 45(7), R4225–R4228 (1992). [CrossRef]  

51. F. Köster, D. Ehlert, and K. Lüdge, “Limitations of the recall capabilities in delay based reservoir computing systems,” Cogn. Comput. 2020 (2020). Springer, ISSN 1866–9956.

52. F. Köster, S. Yanchuk, and K. Lüdge, “Insight into delay based reservoir computing via eigenvalue analysis,” JPhys Photonics 3(2), 024011 (2021). [CrossRef]  

53. A. Panchuk, D. P. Rosin, P. Hövel, and E. Schöll, “Synchronization of coupled neural oscillators with heterogeneous delays,” Int. J. Bifurc. Chaos 23(12), 1330039 (2013). [CrossRef]  

54. R. C. Calleja, A. R. Humphries, and B. Krauskopf, “Resonance phenomena in a scalar delay differential equation with two state-dependent delays,” SIAM J. Appl. Dyn. Syst. 16(3), 1474–1513 (2017). [CrossRef]  

55. L. C. Jaurigue, E. Schöll, and K. Lüdge, “Suppression of noise-induced modulations in multidelay systems,” Phys. Rev. Lett. 117(15), 154101 (2016). [CrossRef]  

56. I. S. Proskurkin and V. K. Vanag, “Dynamical regimes of two frequency different chemical oscillators coupled via pulse inhibitory coupling with time delay,” Russ. J. Phys. Chem. 89(2), 331–335 (2015). [CrossRef]  

57. P. Nieters, J. Leugering, and G. Pipa, “Neuromorphic computation in multi-delay coupled models,” IBM J. Res. Dev. 61(2/3), 7–9 (2017). [CrossRef]  

58. S. Ortín and L. Pesquera, “Tackling the trade-off between information processing capacity and rate in delay-based reservoir computers,” Front. Phys. 7, 210 (2019). [CrossRef]  

59. L. C. Jaurigue, E. Robertson, J. Wolters, and K. Lüdge, “Reservoir computing with delayed input for fast and easy optimization,” Entropy 23(12), 1560 (2021). [CrossRef]  

60. R. Martinenghi, S. Rybalko, M. Jacquot, Y. K. Chembo, and L. Larger, “Photonic Nonlinear Transient Computing with Multiple-Delay Wavelength Dynamics,” Phys. Rev. Lett. 108(24), 244101 (2012). [CrossRef]  

61. A. Röhm and K. Lüdge, “Multiplexed networks: reservoir computing with virtual and real nodes,” J. Phys. Commun. 2(8), 085007 (2018). [CrossRef]  

62. X. X. Guo, S. Y. Xiang, Y. H. Zhang, L. Lin, A. J. Wen, and Y. Hao, “Four-channels reservoir computing based on polarization dynamics in mutually coupled vcsels system,” Opt. Express 27(16), 23293–23306 (2019). [CrossRef]  

63. Y. S. Hou, G. Q. Xia, E. Jayaprasath, D. Z. Yue, W. Y. Yang, and Z. M. Wu, “Prediction and classification performance of reservoir computing system using mutually delay-coupled semiconductor lasers,” Opt. Commun. 433, 215–220 (2019). [CrossRef]  

64. W. Y. Liang, S. R. Xu, L. Jiang, X. H. Jia, J. B. Lin, Y. L. Yang, L. M. Liu, and X. Zhang, “Design of parallel reservoir computing by mutually-coupled semiconductor lasers with optoelectronic feedback,” Opt. Commun. 495, 127120 (2021). [CrossRef]  

65. D. Z. Yue, Y. S. Hou, Z. M. Wu, C. X. Hu, Z. Z. Xiao, and G. Q. Xia, “Experimental investigation of an optical reservoir computing system based on two parallel time-delay reservoirs,” IEEE Photonics J. 13(3), 1–11 (2021). [CrossRef]  

66. C. Sanderson and R. Curtin, “Armadillo: a template-based C library for linear algebra,” JOSS 1(2), 26 (2016). [CrossRef]  

67. C. M. Bishop, “Training with noise is equivalent to tikhonov regularization,” Neural Comput. 7(1), 108–116 (1995). [CrossRef]  

68. P. Perlikowski, S. Yanchuk, O. V. Popovych, and P. A. Tass, “Periodic patterns in a ring of delay-coupled oscillators,” Phys. Rev. E 82(3), 036208 (2010). [CrossRef]  

69. M. Goldmann, F. Köster, K. Lüdge, and S. Yanchuk, “Deep time-delay reservoir computing: Dynamics and memory capacity,” Chaos 30(9), 093124 (2020). [CrossRef]  

Supplementary Material (1)

NameDescription
Supplement 1       Supplementary Document

Data Availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (10)

Fig. 1.
Fig. 1. Sketch of a reservoir computer based on a recurrent neural network (echo state network). The green lines represent the connections that need to be trained, while the rest of the network remains unchanged. Concept taken from [1].
Fig. 2.
Fig. 2. Visualisation of time-multiplexing (sequential sampling indicated by blue dots) and masking. Left and right panel show unmasked and masked input (top row) while the respective system response is shown below. Time is normalized to the input cycle $T$.
Fig. 3.
Fig. 3. Sketch of a delay-based reservoir computer realized with a laser subjected to optical feedback, masked input, and time-multiplexed output.
Fig. 4.
Fig. 4. Network topologies as they arise in delay-based reservoir computing systems when the delay is restricted to $\tau =n \theta$ ($n\in \mathbb {N}$). In (a,b) the input cycle $T$ is chosen resonant to the delay, $\tau =T$, while $\tau =T+\theta$ in (c). The characteristic system time $T'$ is larger than $\theta$ in (a) while it is smaller than $\theta$ in (b,c).
Fig. 5.
Fig. 5. Total memory capacity (a) and performance of NARMA10 prediction (NRMSE) (b) plotted as a function of clock-cycle $T$ and delay-time $\tau$. Dashed lines indicate resonances of rational ratio between the two timescales, i.e. $n T = m \tau$, with $n,m \in \mathbb {N}$. System and parameters as in [51].
Fig. 6.
Fig. 6. Scheme of two oscillators (blue circles) delay-coupled in a ring-topology without (a) and with (b) self-feedback. The coupling is unidirectional, with coupling strength $\kappa _j$ and delay $\tau _j$ ($j \in \left [1,2\right ]$). Both nodes are externally driven with constant pump rate $\lambda _j$ and time dependent input $I_j(t)= g_j(t)u(t)$ scaled with $\eta _j$.
Fig. 7.
Fig. 7. Total memory capacity of two mutually-coupled oscillators (see Fig. 6(a) and Eq. (9) with $\kappa _j^{self}=0$), color coded as a function of the ring-delays $\tau _1$ and $\tau _2$. Dashed lines indicate $1/2(\tau _1+\tau _2)=n/m T$ (mean delay resonance-lines) with $n,m \in \mathbb {N}$, $\tau _{1,2}=n/m T$ and $\tau _1=\tau _2$, respectively. Resonances are especially pronounced at the intersections where both delays are in resonance.
Fig. 8.
Fig. 8. Impact of delay-mismatch: Linear memory capacity $C_{u_{-steps}}$, color coded as a function of mean delay and steps into the past, averaged over 10 masks. (a) Identical delays $\tau _1=\tau _2$, (b) Mismatched delays $\tau _1=\tau _2+400$,
Fig. 9.
Fig. 9. Impact of self-delay: Linear memory capacity $C_{u_{-steps}}$ color coded as a function of the self-delay $\tau ^{self}$ and the steps into the past, averaged over 10 mask realisations for (a) $\tau _{1}=\tau _{2}=210$, and (b) $\tau _{1}=\tau _{2}=620$.
Fig. 10.
Fig. 10. (a) NARMA10 computing error (NRMSE) as a function of the self-delay time $\tau ^{self}$ in the all-to-all coupled topology displayed in Fig. 6(b), averaged over 10 mask realisations (error bars indicate standard deviation). (b) Memory capacity corresponding to input terms $u_{-i}u_{-i-9}$ as they occur in the NARMA10 series Eq. (13) as a function of the self delay-time and the recall step.

Tables (1)

Tables Icon

Table 1. Default Dimensionless Parameter Values for the Numerical Simulations of the Reservoir Computing Performance of Two Oscillators (Index j ) Described by Eq. (9).

Equations (13)

Equations on this page are rendered with MathJax. Learn more.

min W _ out ( | | S _ _   W _ out o _ ^ | | 2 2 + λ | | W _ out | | 2 2 ) ,
W _ out = ( S _ _ T S _ _ + λ I _ _ ) 1 S _ _ T o _ ^ .
NRMSE = k = 1 K o ( o ^ k o k ) 2 K o var ( o ^ _ ) ,
d x ( t ) d t = x ( t ) + f [ x ( t τ ) + η J ( t ) ] ,
0 = x k + f [ x k N v + η J k ] .
x k n = f [ x k 1 n + η M n I k ] .
x k n = f [ x k 1 n 1 + η M n I k ] for n [ 1 , N v 1 ] ,     and x k 0 = f [ x k 2 N v 1 + η M 0 I k ] .
μ k π τ ( 2 k ν ) ,
z ˙ j = ( λ j + η j g j ( t ) u ( t ) + i ω j + ( γ j + i α j ) | z j | 2 ) z j     +     κ j   e i ϕ j   z j + 1 ( t τ j )   +                                 κ j s e l f e i ϕ j s e l f z j ( t τ j s e l f ) +   D ξ ( t ) .
z ˙ j = f ( z j ( t ) , I j ( t ) ) + g ( z j + 1 ( t τ j ) ) ,
v j ( t ) = z j ( t t j ) ,     with     t j = ( N j ) τ mean i = 1 N j τ N i ,
v ˙ j = f ( v j , I ( t t j ) ) + g ( v j + 1 ( t τ mean ) ) .
A i + 1 = 0.3 A i + 0.05 A i ( j = 0 9 A i j ) + 1.5 u i 9 u i + 0.1.
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.