You are on page 1of 12

RESEARCH ARTICLE

Photonic Spiking Neural Networks with Highly


Efficient Training Protocols for Ultrafast Citation: Owen-Newns D, Robertson J,
Neuromorphic Computing Systems Hejda M, Hurtado A. Photonic Spiking
Neural Networks with Highly Efficient
Training Protocols for Ultrafast
Dafydd Owen-Newns*, Joshua Robertson, Matěj Hejda, Neuromorphic Computing Systems.
and Antonio Hurtado Intell. Comput. 2023;2:Article
0031. https://doi.org/10.34133/
Institute of Photonics, SUPA Department of Physics, University of Strathclyde, Glasgow, UK. icomputing.0031

Submitted 11 November 2022


*Address correspondence to: dafydd.owen-newns@strath.ac.uk Accepted 9 May 2023
Published 29 August 2023
Photonic technologies offer great prospects for novel, ultrafast, energy-efficient, and hardware-friendly
neuromorphic (brain-like) computing platforms. Moreover, neuromorphic photonic approaches based on Copyright © 2023 Dafydd Owen-
ubiquitous, technology-mature, and low-cost vertical-cavity surface-emitting lasers (VCSELs) (devices Newns et al. Exclusive licensee
Zhejiang Lab. No claim to original U.S.
found in fiber-optic transmitters, mobile phones, and automotive sensors) are of particular interest. Given

Downloaded from https://spj.science.org on February 20, 2024


Government Works. Distributed under
that VCSELs have shown the ability to realize neuronal optical spiking responses (at ultrafast GHz rates), a Creative Commons Attribution
their use in spike-based information-processing systems has been proposed. In this study, spiking neural License 4.0 (CC BY 4.0).
network (SNN) operation, based on a hardware-friendly photonic system of just one VCSEL, is reported
alongside a novel binary weight “significance” training scheme that fully capitalizes on the discrete nature
of the optical spikes used by the SNN to process input information. The VCSEL-based photonic SNN was
tested with a highly complex multivariate classification task (MADELON) before its performance was
compared using a traditional least-squares training method and an alternative novel binary weighting
scheme. Excellent classification accuracies of >94% were achieved by both training methods, exceeding
the benchmark performance of the dataset in a fraction of the processing time. The newly reported
training scheme also dramatically reduces the training set size requirements and the number of trained
nodes (≤1% of the total network node count). This VCSEL-based photonic SNN, in combination with the
reported “significance” weighting scheme, therefore grants ultrafast spike-based optical processing highly
reduced training requirements and hardware complexity for potential application in future neuromorphic
systems and artificial intelligence applications.

Introduction of lights to achieve highly parallel division-multiplexing appli-


cations [6]. More importantly, photonic platforms can allow
Photonic ANNs and spiking VCSELs NNs to operate at low power consumption and ultrafast speed,
Artificial neural networks (ANNs) appear more frequently in overcoming the squeeze of power efficiency and clock speed
the fast-developing field of artificial intelligence (AI), owing to faced by state-of-the-art chip-scale electronic components [7,8].
their proven ability to provide high-level performance in The realization of light-based NNs, or the so-called photonic
numerous complex information processing tasks [1,2] such as neural network (PNN) systems, is therefore of key interest to
computer vision and data classification. Neural networks (NNs), all future AI applications, where the benefits of increased oper-
which draw inspiration from networks of biological neurons in ation speed and power efficiency are directly felt in the training
the brain, are formed of highly parallel structures of nodes (neu- and processing of large data volumes.
rons) that realize numerous nonlinear transformations to Despite the relative infancy of brain-inspired (neuromorphic)
achieve efficient processing and decision-making. The desire to photonic systems, the field is advancing rapidly because of the
move toward a more-than-Moore era of computing and diverge attractive properties of optics that have inspired the production
from traditional computing paradigms has also spurred reports of multiple PNN accelerators [9–16]. These systems, based on
of brain-inspired computing, where multiple systems of technologies such as phase-change materials [9,10], optical mod-
large-scale electronic artificial NN systems, such as Loihi [3], ulators [11,16], and micro-ring weight banks [12–15], have been
TrueNorth [4], and BrainScaleS [5], exist. However, the number used to demonstrate different information processing tasks with
of studies demonstrating optical approaches to beyond-von photonic components. Neuromorphic photonic systems built
Neumann computing is also growing given the appeal of the using semiconductor laser (SL) technologies have also demon-
inherent properties of optical media. Optical-based systems can strated intriguing neuron-like nonlinear dynamics that are key
operate with very low component crosstalk, perform across large to the operation of PNN systems [17,18]. Vertical-cavity surface-­
data bandwidth ranges, and utilize non-interacting frequencies emitting lasers (VCSELs), which are well-established, commercially

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 1


available, and increasingly ubiquitous devices (found in mobile revealed that by using a time-multiplexing technique and
phones, data centers, supermarket barcode scanners, etc.), are masking inputs, we could interpret a single VCSEL neuron as
a type of neuromorphic SL technology. The ability of VCSELs a fixed-weight interconnected virtual SNN. This photonic sys-
to imitate the behavior of biological neurons was first proposed tem realizes both the reduced training requirements of an RC
in Ref. [19], where it was shown that controllable spike activa- system and the sparse, all-or-nothing, binary spike-based rep-
tion could be achieved with incoming optical signals by exploit- resentation of a full SNN. The VCSEL-based SNN operated
ing the nonlinear dynamics surrounding the injection locking with short temporal nodes (250 ps), fast optical spikes (100 ps),
condition. It has also been shown that electronic modulation and a configurable network node number, which allowed it to
of a VCSEL can cause similar spiking behavior [20]. Neuron- successfully operate on the benchmark iris flower classification
like spiking responses were achieved at ultrafast (GHz) rates task [47] with excellent performance.
with spike widths of approximately 100 ps, making the spiking In this study, we show that we not only can use time multi-
responses >6 orders of magnitude faster than those of biolog- plexing to create a VCSEL-based photonic SNN capable of a
ical neurons and multiple orders of magnitude faster than those highly complex processing task, but more importantly, we
of some electronic spiking systems. VCSELs have exhibited can achieve high performance with substantially reduced
several common neuronal behaviors (such as integration and training requirements (i.e., reduced training set sizes and num-
fire spiking [21,22], refractoriness, and rate coding [23]) in ber of trained nodes). This paper adopts the following structure:
experimental realizations. Further, early demonstrations of neu- In the Photonic SNN setup and operation section, we introduce
romorphic information processing functionalities with VCSELs the experimental setup used in this study to create a photonic
have been achieved using the spiking neuronal behaviors of SNN and discuss its operating principles. In the Dataset prepa-
VCSELs both experimentally, including image processing (edge ration section, we introduce the experimental classification task
feature detection) [24,25], pattern recognition [22], and exclu- and explain the preparation of the data for injection into the
sive OR operation [26], as well as theoretically through simu- VCSEL-based SNN. In the Ordinary least squares section,

Downloaded from https://spj.science.org on February 20, 2024


lations of the Yamada and Spin-flip models (see Ref. [27] for a we discuss the results of the classification task and explore
review). Moreover, single VCSEL devices have exhibited process- the training using a standard linear least-squares regression
ing functionality when operated in combination with software-­ method. In the Binary weight training section, we introduce
based spiking neural networks (SNNs), achieving high classification an alternative binary weight training scheme and discuss its
accuracy in the Modified National Institute of Standards and performance with respect to various training set sizes and
Technology (MNIST) handwritten digit recognition task [28]. node requirements. Finally, we present our conclusions in the
VCSELs therefore represent an increasingly promising techno- Conclusion section.
logical approach to spike-based PNNs that operate with ultrafast
telecommunication wavelengths and optical signals, and are Photonic SNN setup and operation
hardware-friendly, low-power, and low-cost. In this paper, we propose a hardware-friendly photonic SNN
Beyond the photonic imitation of spiking neuronal behav- based on a single time-multiplexed VCSEL neuron. The exper-
iors, reservoir computing (RC) has been demonstrated to be a imental setup used to build and investigate the photonic SNN
powerful method for creating PNNs, yielding excellent perfor- architecture is illustrated in Fig. 1A. The photonic SNN was
mance in complex tasks, while also benefiting from hardware-­ built using a single telecom-wavelength VCSEL and commer-
friendly architectures. First developed in the early 2000s [29,30], cially sourced fiber-optic telecom components. The selected
it was shown that in RC architectures (NN architectures with VCSEL exhibited wavelength emission within the key optical
fixed-weight hidden layer connections), only the output layer telecommunication O-band window (centered at approxi-
requires training to achieve successful performance. In these mately 1,300 nm) to showcase the full compatibility of the sys-
RC architectures, nodes within the hidden layer are referred tem with optical networking and data center technology. In the
to as the “reservoir”. Therefore, the reservoir is composed of experimental setup, external light injection was provided by a
unknown, fixed-strength, interconnected, nonlinear elements tunable laser (TL) source, which was intensity-modulated using
coupled to an output layer. Given the fixed nature of connec- a 10-GHz Mach Zehnder modulator (MZ). The MZ is respon-
tions, RC limits the training requirements of NNs and reduces sible for encoding the input information in the intensity of the
the resources (computational power and time) required to optical injection before it enters the VCSEL neuron. The MZ is
train large networks of nodes for successful operations [29]. driven by a DC power supply and a fast 12-GSa/s 5-GHz arbi-
Numerous theoretical [31–33] and experimental [34–36] reports trary waveform generator (AWG) that generates the required
have been published on reservoir computers based on photon- input information. A 10-dB electrical amplifier (AMP) was
ics platforms with VCSELs, which frequently feature [37–42]. used to increase the signal from the AWG, and a polarization
VCSELs are photonic devices in which both time-delay reser- controller (PC) was used to maximize TL light coupling into
voirs [39–42], which multiplex nodes over time to construct the MZ. The modulated injection was split by a fiber coupler
virtual networks and create memory through feedback connec- to measure the injection power before the TL light entered the
tions, and spatial–temporal reservoirs [43,44], which spatially circulator and VCSEL. The output of the VCSEL was captured
multiplex nodes, have demonstrated successful operations. by the circulator and fed into the analysis, where it was mea­
These systems have demonstrated impressive performance in sured using an optical spectrum analyzer and a fast 40-GSa/s
numerous complex processing tasks, utilizing both off-the-shelf 16-GHz real-time oscilloscope. The threshold and spectral
[42] and bespoke large-area VCSEL designs [43] to realize pho- characteristics of the O-band VCSEL used in the photonic SNN
tonic computing with continuous (nonspiking) VCSEL signals system are shown in Fig. 1B and C. When stabilized at room
(see Ref. [45] for a review). More recently, we demonstrated for temperature (293 K), the VCSEL exhibited a threshold current
the first time a combination of an RC technique with a neu- of 1.423 mA and produced 2 orthogonal polarization peaks in
romorphic spiking VCSEL neuron [46]. In that report, we the fundamental transverse mode. In this study, with the

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 2


Downloaded from https://spj.science.org on February 20, 2024
Fig. 1. (A) Experimental setup used to create the photonic SNN with a single VCSEL device. (B) The lasing threshold (light–current) characteristics of the experimental VCSEL
neuron reveals a threshold current of 1.423 mA. (C) The spectra of the experimental VCSEL neuron at different bias currents. At the operating current (3.5 mA, red), the device
exhibited 2 orthogonal polarization peaks at 1,287.44 nm and 1,287.59 nm. Measurements were taken with a room temperature (293 K) stabilized device.

selected operating current of 3.5 mA, these orthogonal polar- output of several nodes by sampling at set times (θ). Here,
ization peaks occurred at 1,287.44 nm and 1,287.59 nm, pro- using the VCSEL neuron, we apply the same concepts to inter-
ducing a total output power of ∼220 μW. Using a PC, the optical pret the output of a spiking VCSEL as an entire photonic SNN
injection was polarization-matched to the subsidiary (1,287.59 nm) (see Fig. 2A). The output of the VCSEL was time-multiplexed
polarization mode of the VCSEL, with injection made using neg- at intervals of θ to produce a controllable number of network
ative frequency detuning to induce polarization switching and nodes (Nv). The coupling of network nodes is achieved when
injection locking. The modulating intensity is responsible for trig- the duration of the nodes (θ) is less than the timescale of the
gering high-speed (approximately 100-ps-long) neuron-like spik- neuron-like spiking nonlinearity in the VCSEL (spikes can
ing responses near the injection-locking boundary. typically be triggered by strong modulations as short as 200
In this study, as in [46], we combined the concepts of RC ps, with a refractory period of ∼1 ns [23]). This condition
with a photonic spiking VCSEL neuron to create a fully pho- allows the leaky integrate-and-fire and refractory behaviors
tonic SNN using reduced hardware (a single laser device). As of the VCSEL neuron to link multiple nodes coupled in time,
discussed in the Photonic ANNs and spiking VCSELs section, thereby forming a complex interconnected virtual network
reservoir computers are a type of artificial NN that hosts a structure. As in Ref. [46] (and other VCSEL-based reservoir
hidden layer of interconnected nodes whose connections and computers [42]), the input data were prepared by randomly
weights are fixed. This means that in reservoir computers, only masking the data point features and injecting them into the
the output layer of the network must be trained to achieve high VCSEL as continuously modulated light. Following injection,
performance [29]. Furthermore, reservoir computers can use each θ-long node delivers an all-or-nothing binary output, deter-
time multiplexing to interpret the output of one node as the mined by the presence (or absence) of a fast 150-ps optical

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 3


Downloaded from https://spj.science.org on February 20, 2024
Fig. 2. (A) Photonic SNN operating principle. Masked input data are injected into the VCSEL forming the photonic SNN. The input information is time-multiplexed with each θ
time slot representing a virtual node (neuron) in the SNN. The time-multiplexed output of the VCSEL is interpreted as a binary node output, either spiking (1) or non-spiking
(0). The output layer nodes are weighted offline to provide the photonic SNN readout. (B) The architecture of the photonic SNN. A total of 500 features per data point are
input into the SNN system. In this experimental demonstration, the hidden layer was selected to be made up of either 2,048 or 4,096 virtual nodes. The output layer is trained
offline to provide a data point classification (either +1 or −1). (C) The MADELON classification task. Data points are placed on the corners of a 5D hypercube and randomly
labeled −1 or +1. Each artificial data point has 5 informative features (each corresponding to 1 of 5 dimensions). Fifteen linear combinations of the original features are made
creating 20 total features. Four hundred and eighty non-informative features (that carry no information of the label) are then added to the feature set. The task requires all
500 data features to be used to classify the initial data point label −1 or +1.

spike, effectively realizing the output of a fast photonic SNN. training scheme whereby the discrete nature of binary all-or-
In this system, the output weights (the weights applied to each nothing spiking responses can be used to improve the speed
node) are trained and calculated offline to provide a readout of and efficiency of training. In the following sections, we intro-
the system for the prescribed processing task (the task and duce a new, high-complexity, nonlinear classification task
training schemes are discussed in the Dataset preparation sec- (MADELON) [48] and demonstrate the successful perfor-
tion and the Results section). mance of the photonic SNN with 2 training methods: the afore-
In Ref. [46], we recently experimentally demonstrated that mentioned least-squares regression training method and an
a VCSEL neuron system can successfully realize SNN operation alternative node-importance training method, to highlight the
with fast optical spiking signals. We highlight the benchmark attainment of notable improvements in both training speed
nonlinear iris flower classification task, comprising 150 non- and resource requirements.
linearly separable data points (flowers), 4 features per data point
(sepal and petal length/height), and 3 flower classes. However, Dataset preparation
using our VCSEL-based photonic SNN, we demonstrated that We further demonstrate the capability of the VCSEL-based
a very high overall average classification accuracy of >97% photonic SNN by applying it to a highly complex classification
could be achieved, demonstrating the powerful flexibility and task. The task used in this study is an artificial dataset named
performance of photonic SNNs. Importantly, in that study, the “MADELON”, which was created to test feature selection meth-
training of the output weights was achieved using an ordinary ods as part of a NeurIPS 2003 challenge [48]. A simplified sche-
least squares (OLS) regression method; hence, the node outputs matic of the MADELON data creation process is presented in
were treated as if they were continuous (spiking nodes were Fig. 2C. The dataset comprises 32 clusters of data points placed
assigned the value of 1.0, and others were assigned 0.0). This on the vertices of a 5-dimensional (5D) hypercube. Two classes
result required the offline calculation of all node weights as real of data are created by randomly labeling each data point with
number (floating point) values. In this report, we focus on solv- the value −1 or +1. Each data point had 5 informative features
ing a much more complex problem and applying an alternative (one for each dimension) that could be used to identify a data

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 4


class. Using these 5 informative features, 15 linear combinations 293 K. The injected light signal, encoded with the input data,
were created, increasing the feature set size to 20. Finally, 20 had a mean optical power of 161 μW, and was injected with an
features were randomly scattered among 480 non-informative initial frequency detuning of −4.6 GHz (with respect to the
(no predictive power) features, creating a total of 500 features resonance of the VCSEL’s subsidiary polarization mode).
per data point. Therefore, this task is a multivariate, nonlinear, For a given data point, the output of the photonic SNN is a
2-class problem with continuous input variables. The task is vector S with length Nv + 8. The 8 additional virtual nodes were
highly complex, given that a classification must be made using the result of zero padding at the end of the input data sequence,
a feature set in which no feature by itself is informative. This equivalent to a 2-ns reset of the system between inputs.
problem is further complicated by the addition of non-­ Resetting the system between consecutive data point inputs
informative features, which require the system to eliminate and prevented the previous data point from influencing the new
ignore distractions. data point and generating undesired spike activations. In the
The MADELON dataset contains 2,000 data points intended SNN output vector S, the ith element was read as 1 if the ith
for training. However, to demonstrate the operation of the node spiked within the 250-ps-long time-multiplexed segment,
VCSEL-based photonic SNN, 300 data points (150 for each and was assigned a value of 0 otherwise.
class) were used. Limiting the number of input points reduced Figure 3 presents the experimentally measured time series
the length of the time-multiplexed output generated during the depicting the optical input (Fig. 3A and B) and output signals
experimental runs, thereby allowing us to test different network (Fig. 3C) of the photonic SNN. The example shown in Fig. 3 is
architectures with diverse node numbers while remaining for a network architecture with 2,048 nodes. Time multiplexing
within the memory of our experimental equipment. As the was used to inject each masked data point into all (virtual)
MADELON task is of much higher complexity than the pre- nodes of the photonic SNN (at a rate of 250 ps per virtual node).
liminary iris flower classification task [46], a higher number of This created a time-varying signal for each data point, which
virtual nodes (Nv) were selected, namely, 2,048 and 4,096, to was applied as the intensity modulation of the light injected into

Downloaded from https://spj.science.org on February 20, 2024


test the classification performance of the photonic SNN. the SNN. During the experiment, all 300 masked data points
The photonic SNN architecture used for this task is illus- (time-varying signals) were injected into the SNN sequentially,
trated in Fig. 2B. Each of the 500 features was fed into the Nv creating a large optical input time series. Figure 3A depicts a
input layer nodes via a random mask. The input data were part of the large optical input time series, showing 50 sequen-
masked by multiplying a column vector of all 500 features by tially encoded masked data points from the MADELON data-
a 500 × Nv random matrix, resulting in a vector of Nv compo- set. Figure 3B shows the first 512 ns of the optical input time
nents. The masked data were fed into the hidden layer of the series in Fig. 3A, which corresponds to the input signal encod-
photonic SNN, where fixed connections and weights were cre- ing the first MADELON data point. As the amplitude of the
ated by the nonlinear temporal dynamic behavior of the optical input signal varies, it triggers the firing of optical spikes
VCSEL. Experimentally, randomly masked data were generated from the VCSEL at specific times (virtual nodes). Therefore,
by the AWG and modulated by the MZ for optical injection of different input signals (masked data points) entering a photonic
the VCSEL. The output layer of Nv nodes is then provided by SNN produce different optical spike trains in the network. The
the segmentation of the VCSEL spiking output time series at θ optical spiking time series measured at the output of the SNN
intervals. In this study, a node separation (θ) of 250 ps was used in response to the first injected MADELON dataset point (input
for the time-multiplex network nodes, creating multiple inter- signal in Fig. 3B) is shown in Fig. 3C.
acting nodes within the 1-ns timescale of nonlinearity in the Figure 4 shows a 2D temporal map that collects in one plot the
VCSEL [22,23]. The final classification layer of the SNN then spiking output from all virtual nodes of the photonic SNN for all
provides a −1 or 1 label to the data following the offline training 300 consecutively injected data points in the MADELON task. In
and weighting of the output nodes. The following section dis- Fig. 4, a yellow (blue) dot indicates a spiking (non-spiking)
cusses the 2 training methods used in this study. response from the VCSEL-based SNN, and the resulting 1 (0)
binary output of the virtual node in vector S. The output vector S
Results was plotted horizontally for each input data point, revealing the
trends in node activation across inputs and classes. The results for
First, the MADELON task was executed in a photonic VCSEL- photonic SNNs configured with 2,048 and 4,096 virtual (spiking)
based SNN using a network node count of Nv = 2,048. This nodes are plotted in Fig. 4A and B, respectively. The red line
resulted in a total processing time of 512 ns per data point in through the center of the map separates the results for each dis-
the photonic SNN. The VCSEL was driven with a current of tinct class (−1 and +1) of the input data point. Figure 4A and B
3.5 mA (∼2.5 times the lasing threshold current) at a temper- show that each input data point of the same class elicits a similar
ature of 293 K and the injected light had a mean power of spike train at the output of the photonic SNN. These spike trains
142.6 μW and was injected into the VCSEL with an initial fre- were sufficiently distinct from those obtained from other classes
quency detuning of −5.4 GHz with respect to the resonant of data points to allow for a classification operation.
wavelength of the VCSEL’s subsidiary polarization mode peak. For the SNN with a 2,048-node architecture (Fig. 4A), a total
These conditions put the VCSEL in a state in which it was injec- of 298 data points (149 per class) were used in the experimental
tion-locked to an externally injected signal. Near the injec- run. This is because the first data point in class −1 was affected
tion-locking boundary, optical spikes could be triggered by by an experimental artifact and was removed from the analysis.
varying the intensity of the modulated optical injection light. Therefore, one point from class +1 was removed to keep the
In the second experimental run, the number of virtual (spik- size of each set equal, resulting in 298 points. For the network
ing) nodes in the network was increased to Nv = 4,096 (1,024 ns architecture with Nv = 4,096 (Fig. 4B), all 300 data points
of processing time per data point). In the second run, the VCSEL (150 per class) of the MADELON task were used to test the
was biased with a current of 3.45 mA at a set temperature of network operation.

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 5


Downloaded from https://spj.science.org on February 20, 2024
Fig. 4. Temporal maps showing the optical spiking patterns produced by the SNN in
response to the input dataset using (A) 2,048 virtual nodes and (B) 4,096 virtual nodes.
Fig. 3. Time series showing the measured input (blue) and output (red) of the photonic
SNN for a network architecture of 2,048 virtual nodes. (A) A section of injection
waveform showing the time-multiplexed intensity variations of 50 consecutive
Ordinary least squares
masked data points of the MADELON dataset. (B) A 512-ns segment of the input time In this study, we first used the OLS training method to test the
series corresponding to the first MADELON data point. Each of the 2,048 virtual nodes performance of the photonic SNN in the MADELON task, a
has a duration of 250 ps, creating the 512-ns input time series. (C) The corresponding standard training method applied in other photonic RC systems
measurement of the SNN output when injected with the first MADELON data point. [46,49]. Here, the output spiking patterns from the VCSEL (and
corresponding vector S) are treated as a binary sequence of
floating-point 1.0s and 0.0s. Then, the output layer weights are
Training was completed using a randomly selected subset found via linear regression where the values of each node (1.0
of input data points and their corresponding spiking responses. or 0.0) are arranged into matrix X (where the rows of X are the
The randomly selected subset contains an equal number of output vectors S). The labels (expected values) of each data point
points (Nt) for each class for a total training set size of 2Nt. In form the second-column vector Y. The weight calculation then
our photonic SNN, training is used to calculate the weights involves determining the weight matrix W that minimizes Eq. 1.
applied to the output layer nodes (i.e., the temporal spiking
patterns produced by the network and the corresponding vec- Error (W) = |Y − XW|2 . (1)
tor S). The output weights were determined to achieve the best
performance on this set with the assumption that the training Equation 1 is solved using W = (XTX)−1XTY (the Penrose
set is a good representation of the entire task dataset. The input inverse of X multiplied by Y). For this task, there are 2 output
data points not used in the training process were subsequently classes (−1 and +1). Therefore, the labels are 2-element vectors
used to test the performance of the MADELON task. The out- in which the nth element is set to 1 to denote the nth class, with
put layer weights calculated during training were applied to the the other elements being zero. This also implies that the result-
output layer to infer the class of a given test data point. The ing weight matrix W has the dimensions (Nv × 2). Using W,
inferred data point class was then compared to the true class the prediction of the class of a data point (producing a spike
label, and the performance was measured using classification pattern S) can be obtained by considering argmax(S ∗ W).
accuracy, which is the fraction of correctly classified data points Applying weights to the spike pattern vector yielded a score for
across all tested data points. The training/testing process was each class label, where the class with the highest score was the
repeated several times, and the training set was randomly inferred label (prediction) of the photonic SNN system.
selected each time to find an average value for the performance Using only 15 training data points per class (10% of the total
(a process known as random cross-validation). 300 input data points), the OLS training method achieved a

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 6


very high performance with a peak accuracy of 91% (for a ∑
2,048-node architecture) and 94.5% (for a 4,096-node archi- 2
zn,i = sn,i ∕ sn,i (2)
tecture). Example confusion matrices for Nv = 2,048 and Nv = i
4,096 trained using this method and Nt = 15 are shown in Fig. 5.
As expected, the network architecture with a higher number of 3. For each class, choose the top Nn scoring nodes and set
nodes (Nv) provided a better overall classification performance. their weights toward that class (Wn,i) to 1, and all remaining
A larger number of nodes created more spiking responses node weights to zero.
from the VCSEL, allowing for more class-specific responses Applying the aforementioned algorithm provides the scores
and better classification. The capability to adjust the node num- for all network nodes, with higher scores obtained not only for
ber at will is an inherent advantage of the photonic SNN used nodes that contain more spikes but also for nodes that have
in this study, as no physical or hardware changes are required their spikes more concentrated in one specific class. The result-
to increase the network architecture. Instead, all that is required ing output weight matrix W contains only nonzero entries on
is the alteration of the input mask and a longer time series nodes that are strongly indicative of a specific class. The 2D
measurement. This makes the reported VCSEL-based photonic temporal maps in Fig. 6 reveal the spike trains of a subset of
SNN an attractive and flexible system for fast, optical spike- network nodes, for the case when the photonic SNN was oper-
based implementation of numerous processing tasks. ated with a total node count of 2,048. This plot shows the scor-
ing system applied to selected nodes. For example, Fig. 6 shows
Binary weight training that node 844 (marked with an amber arrow) elicited an optical
As discussed in the previous section (Ordinary least squares spike for 57 of the total 298 input data points used in this exper-
section), the standard OLS training method delivers a high clas- imental run. In this case, the spiking responses from n = 844
sification performance during the MADELON task. How­ever, are evenly distributed across each class (31 spikes for class −1
the OLS method does not take full advantage of the discrete and 26 for class +1). In this specific case, the scores for this

Downloaded from https://spj.science.org on February 20, 2024


binary nature of the fast optical spiking responses in the photonic node are z844,−1 = 16.9 (class −1) and z844,+1 = 11.9 (class +1).
SNN. For example, it can be observed that in the ideal case of an In parallel, Fig. 6 shows that node 885 (marked with a red
SNN response, spiking patterns would be perfectly consistent arrow) fired fewer optical spikes overall. A total of 30 responses
between data of the same class, as well as completely distinct were elicited. However, because they were less evenly distrib-
from data of alternative classes. In this ideal case, 2k classes would uted (and hence more indicative of a particular class), the node
only require the use of k nodes to be completely separate and scores z885,−1 = 2.1 (class −1) and z885,+1 = 16.1 (class +1), a
identify data points. In the ideal case of the MADELON task, only class +1 score similar to that of node 844. Finally, when the
one node per class is required for identification. In the case of appearance of fast optical spikes is much more biased toward
experimental data, initially identifying nodes that have more pre- one specific class and is more likely to occur overall, the result-
dictive power (i.e., spike only for a particular class) prior to train- ing score z is much higher. For example, in node 853 (marked
ing would allow for the creation of weight matrices that retain with a green arrow), 103 spikes occurred, with 88 of them
high classification accuracy, despite the use and calculation of occurring in class −1, yielding scores of z853,−1 = 75.1 and
only a few nodes (a fraction of the total node count). z853,+1 = 2.1. With this method, the weight W853, − 1 of node 853
The following process is introduced to determine the weights is set to 1, as if a spike was detected in node 853, and it is likely
for this SNN: that the data point is of class −1. In this example, the weights
1. Count the number of spikes that occurred for each class of the other highlighted nodes (W844,−1 and W885,−1) are set
in each node, denoted sn,i (# spikes from data of class i occurring to zero.
in node n). For a data point to be classified, the spike-output vector S
2. Calculate the significance score of each node toward each must be multiplied by the weight matrix Wi for each class. The
class according to Eq. 2: class with the highest sum value was the prediction class of the

Fig. 5. Example confusion matrices depicting the performance of a single instance of the photonic SNN when trained using the ordinary least squares method. A training set
size of 15 (Nt = 15) was used to achieve accuracies of 91% and 94.4% for architectures of 2,048 and 4,096 nodes, respectively.

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 7


Downloaded from https://spj.science.org on February 20, 2024
Fig. 6. Section of a 2D temporal map for a network architecture of 2,048 nodes (between nodes 842 and 890). The arrows mark nodes 844 (orange), 853 (green), and 885 (red).
In node 853, 88 spikes occurred for class −1, and 15 for class +1, giving it a high score of 75.1 (for class −1) and 2.1 (for class +1). Node 844 has scores of 16.9 (for class −1) and
11.9 (for class +1). Node 885 has scores of 2.1 (for class −1) and 16.1 (for class +1). Node 853 has the highest score meaning it would likely be used to perform a classification.

data point. Because the output layer weights are binary, this is accuracy of 95.2% is achieved. Again, as expected, the architec-
equivalent to applying a filter to the optical spike trains and ture with a larger number of nodes provided a higher classifi-
counting the number of remaining spikes. Using this method, cation performance. Comparing this performance to the OLS
the performance can be tuned by selecting the number of nodes training method, also run at Nt = 15, we found that significance-
(Nn) for training (training node number). Using a fraction of training shows a small improvement in peak accuracy in both
the total number of available nodes can result in equal (if not cases (2,048- and 4,096-node architectures). This improvement
better) performance, while keeping the weight matrix and in performance shows that despite the same training set size,
number of weight calculations sparse. by considering only nodes indicative of specific classes, we
Figure 7 shows the resulting confusion matrices of the can achieve a similar or higher classification performance.
MADELON task when the significance-training algorithm was Furthermore, it is important to highlight that the training node
run with 15 training data points (Nt = 15). For the photonic number used in the significance-scoring method was sub-
SNN architecture with 2,048 nodes, a peak accuracy of 92.5% stantially lower than that used in the OLS method. In the
is obtained, whereas for the 4,096-node architecture, a peak novel training method, only the top 20 most significant nodes

Fig. 7. Example confusion matrices depicting the performance of a single instance of the photonic SNN when trained using the significance score method. A training set size
of 15 (Nt = 15) was used to achieve accuracies of 92.5% and 95.2% for architectures of 2,048 and 4,096 nodes, respectively. Only the 20 most significant nodes (Nn = 20)
were used to train the SNN.

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 8


(Nn = 20) were used in both reported SNN architecture cases Both Figs. 8 and 9 reveal 2 key results. First, the photonic
(2,048 and 4,096). Therefore, the novel training method uses SNN achieved very high accuracy levels, despite the complexity
only 1% and 0.5% (for 2,048 and 4,096 nodes, respectively) of of the MADELON task and its large number of features. Spe­
the total nodes used for OLS training, where every node in the cifically, a maximum accuracy of 94.4% was achieved for the
network is trained (Nn = Nv). This result indicates that training 2,048-node architecture, and a maximum accuracy of 95.7%
fewer nodes overall is not only computationally efficient, but was achieved for the 4,096-node architecture. The performance
does not directly hinder the performance of the system (and of this photonic SNN system was higher than that of previous
in this case, increases classification performance). However, software-based machine learning algorithms reported follow-
these newly presented classification performance values are not ing the NIPS 2003 challenge [48], which reached accuracy
within the limits of the photonic SNN with this training scheme. levels of up to 93.78%. Importantly, a key difference is that a
In Figs. 8 and 9, we consider a larger range of training set sizes photonic SNN is applied to the MADELON task with a hardware-­
(Nt), and compare their optimal performance as well as their friendly implementation (using only one VCSEL), low-power
optimal training node number (Nn). operation (sub-pJ energy per spike, ∼150 μW average optical
Increasing the number of nodes leads to better performance powers, and 3.5 mA of applied bias current), and ultrafast per-
because of the increased likelihood of a node spiking more formance (250 ps/node yielding a total processing time of 512 ns
strongly for one class than for another. This was caused by the and 1,024 ns per data point for 2,048 and 4,096 nodes, respec-
longer-range connections formed in the SNN owing to the tively). Second, the performance of the photonic SNN increases
refractory behavior of the spiking dynamics in the VCSEL. with the training set size (associated with the best-fit training),
Figures 8A and 9A show the peak accuracy achieved in the and the optimal value of Nn may slightly decrease with the
MADELON classification task as a function of the training set size training set size. This result was somewhat expected, as training
used for the 2,048 and 4,096 network architectures, respectively. with more data points typically improves the system perfor-
Figures 8B and 9B show the optimal training node number (Nn) mance, resulting in a lower dependence on high node numbers.

Downloaded from https://spj.science.org on February 20, 2024


used to attain the optimal accuracy. Thirdly, Figs. 8 and 9 show that the optimal training node

Fig. 8. Photonic SNN performance versus the training set size, using a 2,048-node architecture. (A) The optimal classification accuracy (averaged over 10 random selections
of the training set) is plotted against training set size (Nt, blue). The inset highlights the consistent high performance over various training set sizes (Nt ranges from 1 to 100).
(B) The number of training nodes (Nn) used to achieve the corresponding optimal accuracy, plotted against increasing training set size (red). The peak accuracy, 94.4%,
occurs when Nn = 6 and Nt = 76 (circled).

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 9


Downloaded from https://spj.science.org on February 20, 2024
Fig. 9. Photonic SNN performance versus the training set size, using a 4,096-node architecture. (A) The optimal classification accuracy (averaged over 10 random selections
of the training set) is plotted against training set size (Nt, blue). The inset highlights the consistent high performance over various training set sizes (Nt ranges from 1 to 100).
(B) The number of training nodes (Nn) used to achieve the corresponding optimal accuracy, plotted against increasing training set size (red). The peak accuracy, 95.7%, occurs
when Nn = 18 and Nt = 91 (circled).

number (Nn) remains consistently low while the system retains electronic storage, and as the nonzero elements are only 1 (no
very high performance, across all training set sizes (Nt) (from floating-­point numbers are used), the weights could potentially
1 to 100). Only small node numbers, as low as <10 (out of the be directly applied in optical hardware without the need for
total node count Nv), must be considered to successfully achieve precision modulation of the signal. Compared to traditional
high performance in the MADELON task. This means that, training methods, the algorithm for obtaining this weight matrix
independent of the training set size, the highly reduced training is much simpler: typical training methods use costly operations,
requirements of the significance scoring method remain a key such as matrix inversions, to calculate floating-point weights,
benefit of the alternative training scheme. Finally, as illustrated whereas this method simply counts the number of spikes occur-
in Figs. 8 and 9, very small training set sizes, Nt ≤ 10 (≤3% of ring at each node using low-cost computations.
the total dataset inputs), are capable of effectively training the
photonic SNN system and achieving high accuracy (>90%) in Conclusion
the MADELON task. Comparing the previous OLS results for
a network architecture of 4,096 (from 2.1) with those in Fig. 9, This study demonstrates the high classification performance of
we find that at training set sizes <15, the significance scoring a novel laser-based photonic SNN for tackling a highly com-
method achieves similar levels of classification accuracy. Making plex, multivariate, nonlinear classification task (MADELON)
use of the novel training approach can, therefore, enable a sim- with 500 data point features. Importantly, this study also intro-
ilar performance to the OLS method, but with the remarkable duces a novel “significance” training approach that makes use
benefit of being more flexible in terms of the number of training of binary weights (0 or 1), and that leverages on the advantages
data points. of the discrete optical spiking signals found in the photonic
The resulting (binary) weight matrix has several advantages SNN. The experimental approach to an SNN combines the
over those produced by typical methods: it is very sparse-all spiking dynamics of a VCSEL with a novel network architecture
but a few elements are nonzero, making this require very little inspired by the RC paradigm to process data entirely optically

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 10


at extremely high speeds (GHz rates). The SNN uses all-optical Data Availability
neuron-like spikes to create a time-multiplexed feed-forward
SNN in which the values of each time-multiplexed (virtual) All data underpinning this publication are openly available
node are linked through the VCSEL’s nonlinear temporal from the University of Strathclyde KnowledgeBase at https://
dy­namics. The computational power of the photonic SNN doi.org/10.15129/fffe7873-ee5a-45fa-b6f4-2214ce2383d7. For the
was demonstrated using an OLS method of weight training. purpose of Open Access, the author has applied a CC BY public
We demonstrated that, by training the output layer weights copyright license to any author-accepted manuscript (AAM)
with this OLS approach, we could achieve high accuracies version arising from this submission.
of up to 91% and 94.4% for the MADELON classification
task using SNN architectures with 2,048 and 4,096 nodes, References
respectively.
Additionally, we introduced a new “significance” training 1. Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y,
approach that assigns binary weights to (optical spiking) nodes Al-Shamma O, Santamaria J, Fadhel MA, Al-Amidie M,
according to their overall usefulness and significance score. In Farhan L. Review of deep learning: Concepts, CNN architectures,
this approach, only high-significance scoring nodes, that is, challenges, applications, future directions. J Big Data.
nodes that spike frequently for one class but not others, were 2021;8(1):Article 53.
considered and used for network training and performance 2. Kazanskiy NL, Butt MA, Khonina SN. Optical computing: Status
evaluation. We showed that only a very small fraction (≤1% in and perspectives. Nanomaterials (Basel). 2022;12(13):Article 2171.
the presented case) of the total number of nodes (in the output 3. Davies M, Srinivasa N, Lin T-H, Chinya G, Cao Y, Choday SH,
layer) were required to classify the data successfully. We Dimou G, Joshi P, Imam N, Jain S, et al. Loihi: A
demonstrated that classification accuracies of 94.4% and 95.7% neuromorphic manycore processor with on-chip learning.
could be achieved using this new training method. The accu- IEEE Micro. 2018;38(1):82–99.

Downloaded from https://spj.science.org on February 20, 2024


racies provided by the significance-training approach showed 4. DeBole MV, Taba B, Amir A, Akopyan F, Andreopoulos A,
an improvement over those achieved by the OLS method, while Risk WP, Kusnitz J, Ortega Otero C, Nayak TK, Appuswamy R,
also significantly reducing the number of training nodes. et al. TrueNorth: Accelerating from zero to 64 million neurons
Moreover, the photonic SNN demonstrated a classification in 10 years. Computer. 2019;52(5):20–29.
performance that improved upon the benchmark accuracy 5. Pehle C, Billaudelle S, Cramer B, Kaiser J, Schreiber K,
(93.78%) achieved by software-implemented NNs reported by Stradmann Y, Weis J, Leibfried A, Müller E, Schemmel J. The
the dataset authors [48]. Additionally, we demonstrated that a BrainScaleS-2 accelerated neuromorphic system with hybrid
photonic SNN trained with the new significance method could plasticity. Front Neurosci. 2022;16:Article 795876.
realize high-level performance with small training set sizes 6. Brackett CA. Dense wavelength division multiplexing networks:
(<10 data points), further reducing the overall resources nec- Principles and applications. IEEE J Selec Areas Commun.
essary for training the optical system. 1990;8(6):948–964.
Finally, the proposed photonic SNN offers several inherent 7. Prucnal PR, Shastri BJ, Teich MC. In: Prucnal R, Shastri BJ, editors.
physical and computational benefits over traditional digital Neuromorphic photonics. Boca Raton (FL): CRC Press; 2017.
semiconductor processing systems, notably ultrafast perfor- 8. Miller DAB. Attojoule optoelectronics for low-energy information
mance (250 ps/node), low-power usage (∼150 μW average processing and communications. J Lightwave Technol.
optical power and 3.5 mA of applied bias current), and hardware-­ 2017;35(3):346–396.
friendly implementation (using just one VCSEL to process all 9. Feldmann J, Youngblood N, Wright CD, Bhaskaran H,
virtual nodes). Furthermore, a VCSEL-based photonic SNN Pernice WHP. All-optical spiking neurosynaptic networks with
can adjust the performance and processing rate by changing self-learning capabilities. Nature. 2019;569(7755):208–214.
the number of virtual nodes used in the system, which can be 10. Feldmann J, Youngblood N, Karpov M, Gehring H, Li X,
performed arbitrarily and on the fly during pre-processing. In Stappers M, le Gallo M, Fu X, Lukashchuk A, Raja AS, et al.
conclusion, we believe these results open possibilities for fur- Parallel convolutional processing using an integrated photonic
ther photonics-based processing systems that run and operate tensor core. Nature. 2021;589(7840):52–58.
entirely on optical hardware, and that are capable of solving 11. Xu S, Wang J, Wang R, Chen J, Zou W. High-accuracy optical
highly complex tasks with high accuracy and ultrafast, energy-­ convolution unit architecture for convolutional neural
efficient operation. networks by cascaded acousto-optical modulator arrays. Optics
Express. 2019;27(14):19778–19787.
12. Tait AN, de Lima TF, Zhou E, Wu AX, Nahmias MA, Shastri BJ,
Acknowledgments Prucnal PR. Neuromorphic photonic networks using silicon
Funding: The authors acknowledge that this work was sup- photonic weight banks. Sci Rep. 2017;7(1):Article 7430.
ported by the UKRI Turing AI Acceleration Fellowships Prog­ 13. de Lima TF, Peng HT, Tait AN, Nahmias MA, Miller HB,
ramme (EP/V025198/1), the European Commission (Grant Shastri BJ, Prucnal PR. Machine learning with neuromorphic
828841-ChipAI-H2020-FETOPEN-2018-2020), and UK EPSRC photonics. J Lightwave Technol. 2019;37(5):1515–1534.
(EP/N509760/1, EP/P006973/1). Author contributions: D.O.-N. 14. Mehrabian A, Al-Kabani Y, Sorger VJ, El-Ghazawi T. PCNNA:
performed pre- and post-processing of all data. D.O.-N. and A photonic convolutional neural network accelerator. Paper
J.R. performed experimental runs on a photonic system. M.H. presented at: 31st IEEE International System-on-Chip
contributed to data postprocessing and visualization. A.H. Conference; 2018 Sep 4–7; Arlington, USA.
supervised all research efforts. All the authors contributed 15. Ashtiani F, Geers AJ, Aflatouni F. An on-chip photonic
equally to the writing of this manuscript. Competing interests: deep neural network for image classification. Nature.
The authors declare that they have no competing interests. 2022;606(7914):501–506.

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 11


16. Zhang H, Gu M, Jiang XD, Thompson J, Cai H, Paesani S, 32. Rohm A, Jaurigue L, Ludge K. Reservoir computing using laser
Santagati R, Laing A, Zhang Y, Yung MH, et al. An optical networks. IEEE J Sel Top Quantum Electron. 2020;26(1):1–8.
neural chip for implementing complex-valued neural network. 33. Huang Y, Zhou P, Yang YG, Cai DY, Li NQ. Enhanced performance
Nat Commun. 2021;12(1):Article 457. of reservoir computing using multiple self-injection and mutual
17. Chen Z, Sludds A, Davis R, Christen I, Bernstein L, Heuser T, injection VCSELs. IEEE J Sel Top Quantum Electron. 2023;29(2):1–9.
Heermerier N, Lott JA, Reitzenstein S, Hamerly R, et al. Deep 34. Vandoorne K, Mechet P, van Vaerenbergh T, Fiers M, Morthier G,
learning with coherent VCSEL neural networks. ArXiv. 2022. Verstraeten D, Schrauwen B, Dambre J, Bienstman P.
https://doi.org/10.48550/arXiv.2207.05329 Experimental demonstration of reservoir computing on a
18. Prucnal PR, Shastri BJ, Ferreira de Lima T, Nahmias MA, silicon photonics chip. Nat Commun. 2014;5(1):Article 3541.
Tait AN. Recent progress in semiconductor excitable lasers for 35. Brunner D, Soriano MC, Mirasso CR, Fischer I. Parallel photonic
photonic spike processing. Adv Opt Photon. 2016;8(2):Article 228. information processing at gigabyte per second data rates using
19. Hurtado A, Javaloyes J. Controllable spiking patterns in transient states. Nat Commun. 2013;4(1):Article 1364.
long-wavelength vertical cavity surface emitting lasers 36. Vinckier Q, Duport F, Smerieri A, Vandoorne K, Bienstman P,
for neuromorphic photonics systems. Appl Phys Lett. Haelterman M, Massar S. High-performance photonic
2015;107(24):Article 241103. reservoir computer based on a coherently driven passive cavity.
20. Robertson J, Wade E, Hurtado A. Electrically controlled Optica. 2015;2(5):Article 438.
neuron-like spiking regimes in vertical-cavity surface-emitting 37. Bueno J, Brunner D, Soriano MC, Fischer I. Conditions for
lasers at ultrafast rates. IEEE J Sel Top Quantum Electron. reservoir computing performance using semiconductor lasers with
2019;25(6):Article 5100307. delayed optical feedback. Opt Express. 2017;25(3):Article 2401.
21. Robertson J, Wade E, Kopp Y, Bueno J, Hurtado A. Towards 38. Argyris A, Bueno J, Fischer I. PAM-4 transmission at 1550 nm
neuromorphic photonic networks of ultrafast spiking laser neurons. using photonic reservoir computing post-processing. IEEE
IEEE J Sel Top Quantum Electron. 2019;26(1):Article 7700715. Access. 2019;7:37017–37025.

Downloaded from https://spj.science.org on February 20, 2024


22. Robertson J, Hejda M, Bueno J, Hurtado A. Ultrafast optical 39. Vatin J, Rontani D, Sciamanna M. Enhanced performance of
integration and pattern classification for neuromorphic a reservoir computer using polarization dynamics in VCSELs.
photonics based on spiking VCSEL neurons. Sci Rep. Opt Lett. 2018;43(18):Article 4497.
2020;10(1):Article 6098. 40. Vatin J, Rontani D, Sciamanna M. Experimental reservoir
23. Hejda M, Robertson J, Bueno J, Hurtado A. Spike-based computing using VCSEL polarization dynamics. Optics
information encoding in vertical cavity surface emitting Express. 2019;27(13):Article 18579.
lasers for neuromorphic photonic systems. J Phys Photonics. 41. Vatin J, Rontani, Sciamanna M. Experimental realization of
2020;2(4):Article 44001. dual task processing with a photonic reservoir computer. APL
24. Robertson J, Zhang Y, Hejda M, Bueno J, Xiang S, Hurtado A. Photonics. 2020;5(8):Article 086105.
Image edge detection with a photonic spiking VCSEL-neuron. 42. Bueno J, Robertson J, Hejda M, Hurtado A. Comprehensive
Opt Express. 2020;28(25):37526–37537. performance analysis of a VCSEL-based photonic reservoir
25. Zhang Y, Robertson J, Xiang S, Hejda M, Bueno J, Hurtado A. computer. IEEE Photon Technol Lett. 2021;33(16):920–923.
All-optical neuromorphic binary convolution with a spiking 43. Porte X, Skalli A, Haghighi N, Reitzenstein S, Lott JA,
VCSEL neuron for image gradient magnitudes. Photon Res. Brunner D. A complete, parallel and autonomous photonic
2021;9(5):B201–B209. neural network in a semiconductor multimode laser. J Phys
26. Zhang Y, Xiang S, Cao X, Zhao S, Guo X, Wen A, Hao Y. Photonics. 2021;3(2):Article 024017.
Experimental demonstration of pyramidal neuron-like 44. Skalli A, Porte X, Haghighi N, Reitzenstein S, Lott JA, Brinner D.
dynamics dominated by dendritic action potentials based on Computational metrics and parameters of an injection-locked
a VCSEL for all-optical XOR classification task. Photonics Res. large area semiconductor laser for neural network computing.
2021;9(6):Article 1055. ArXiv. 2021. https://doi.org/10.48550/arXiv.2112.08947
27. Xiang S, Han Y, Somng Z, Guo X, Zhang Y, Ren Z, Wang S, 45. Skalli A, Robertson J, Owen-Newns D, Hejda M, Porte X,
Ma Y, Zou W, Ma B. A review: Photonics devices, Reitzenstein S, Hurtado A, Brunner D. Photonic
architectures, and algorithms for optical neural computing. neuromorphic computing using vertical cavity semiconductor
J Semiconduct. 2021;42(2):Article 023105. lasers. Opt Mater Express. 2022;12(6):Article 2395.
28. Robertson J, Kirkland P, Alanis JA, Hejda M, Bueno J, 46. Owen-Newns D, Robertson J, Hejda M, Hurtado A. GHz rate
di Caterina G, Hurtado A. Ultrafast neuromorphic neuromorphic photonic spiking neural network with a single
photonic image processing with a VCSEL neuron. Sci Rep. vertical-cavity surface-emitting laser (VCSEL). IEEE J Sel Top
2022;12(1):Article 4874. Quantum Electron. 2023;29(2):1–10.
29. Jaeger H. The “echo state” approach to analysing and training 47. Fisher RA. The use of multiple measurements in taxonomic
recurrent neural networks—with an erratum note. Bonn, problems. Ann Eugenics. 1936;7(2):179–188.
Germany: German National Research Center for Information 48. Guyon I, Li J, Mader T, Pletscher PA, Schneider G, Uhr M.
Technology GMD Technical Report; 2001. Competitive baseline methods set new standards for the
30. Maass W, Natschläger T, Markram H. Real-time computing nips 2003 feature selection benchmark. Pattern Recogn Lett.
without stable states: A new framework for neural 2007;28(12):1438–1444.
computation based on perturbations. Neural Comput. 49. Bueno J, Robertson J, Hejda M, Hurtado A. Experimental
2002;14(11):2531–2560. implementation of a photonic neural network with a
31. Hülser T, Köster F, Jaurigue L, Lüdge K. Role of delay-times in 1550nm-VCSEL subject to optical injection and delayed
delay-based photonic reservoir computing invited. Opt Mater optical feedback. Paper presented at: 2020 IEEE Photonics
Express. 2022;12(3):Article 1214. Conference, IPC 2020—Proceedings; Sep 2020; online.

Owen-Newns et al. 2023 | https://doi.org/10.34133/icomputing.0031 12

You might also like