The explosive volume growth of deep-learning (DL) applications has triggered an era in computing, with neuromorphic photonic platforms promising to merge ultra-high speed and energy efficiency credentials with the brain-inspired computing primitives. The transfer of deep neural networks (DNNs) onto silicon photonic (SiPho) architectures requires, however, an analog computing engine that can perform tiled matrix multiplication (TMM) at line rate to support DL applications with a large number of trainable parameters, similar to the approach followed by state-of-the-art electronic graphics processing units. Herein, we demonstrate an analog SiPho computing engine that relies on a coherent architecture and can perform optical TMM at the record-high speed of 50 GHz. Its potential to support DL applications, where the number of trainable parameters exceeds the available hardware dimensions, is highlighted through a photonic DNN that can reliably detect distributed denial-of-service attacks within a data center with a Cohen’s kappa score-based accuracy of 0.636. |
1.IntroductionDuring the last decade, deep neural networks (DNNs) have become increasingly important for the resolution of numerous practical problems.1 With the amount of computing power required to train such DNNs doubling every 3.5 months,2 academic and industrial researchers started gravitating toward new technologies and hardware accelerators to keep pace with this growth. Highly parallelized computing solutions, including graphic processing units (GPUs),3 field programmable gate arrays,4 tensor processing units (TPUs),5 and application-specific integrated circuits,6–8 have been developed to accelerate the matrix-vector multiplication (MVM) operations, which form the most time- and power-consuming computational task in DNNs.9 Yet, as transistor scaling is stagnating,10 a high number of alternative emerging technologies have been investigated toward boosting energy efficiency and performance scaling, e.g., optoelectronic memristors,11–15 nanophotonics,16,17 and spintronics,18,19 with brain-inspired photonic accelerators forming one of the key candidate platforms for future AI computing engines due to their inherent credentials to support time-of-flight latencies and terahertz bandwidths.20,21 Remarkable progress has been witnessed during the last five years in the field of neuromorphic photonics across all necessary constituent technology blocks, including MVM photonic architectures,17,22–28 individual photonic computational elements,29–32 nonlinear activations,33–36 and photonic hardware-aware training models.37,38 All these demonstrations have highlighted the potential for energy-efficient and high-speed DNNs by utilizing low-speed weight encoding technologies and a rather small amount of neurons, validating their credentials to support inference within small scale neural network (NN) topologies that can fit in a practical silicon photonic (SiPho) chip. However, typical NN layouts used for benchmarking purposes, such as ResNet152 and AlexNet,39 require a total number of 25 and 62 million trainable parameters, respectively, that can hardly fit as hardware-coded information even into the available number of computational elements supported by current top-class GPU and TPU platforms. This has turned tiled matrix multiplication (TMM) into the mainstream processing paradigm in today’s AI engines,40,41 where both the input and the weighting values have to be updated at line rate through time division multiplexing (TDM) approaches until all matrix tiles are processed. To this end, the upgrade of neuromorphic photonics into a versatile AI processing platform has to proceed along the paradigm of today’s TPU and GPU computational engines, where a limited amount of hardware resources can execute DNNs with significantly higher dimensions. This would necessitate, however, the use of photonic architectures and technologies that can support dynamic reconfiguration of both the NN input and weight parameters, with the existing demonstrations being incapable of meeting these requirements, as they mostly rely on low-speed weight encoding technology, such as thermooptic (TO) devices17,26 and phase change materials.22 In this paper, we present a compact SiPho computing engine that supports both input and weight update rates at a record-high 50 GHz clock frequency, reporting for the first time, to the best of our knowledge, on high-speed TMM directly in the optical domain that allows for DNN implementations over limited-scale photonic hardware. The photonic accelerator comprises a two-input coherent linear neuron (COLN) layout with high-speed SiGe electro-absorption modulators (EAMs) used both for input and weight imprinting. We experimentally demonstrate its credentials to implement TMM and support DNNs with higher dimensions through its deployment in data center (DC) traffic inspection for network security applications, employing the photonic engine for the identification of distributed denial-of-service (DDoS) attack patterns via the classification of reconnaissance attacks (RAs). The DNN comprises 10 neurons and 64 trainable parameters and was successfully executed via the COLN, revealing high experimental accuracy values with a Cohen’s kappa score (-score)42 of 0.638 at 50 GHz. Finally, the scaling perspectives of the EAM-based two-input COLN into a higher dimension coherent photonic crossbar (Xbar) are presented, providing the practical framework for the deployment of optical TMM operations in a layer–scale layout and for higher-dimension tiles. 2.Neuromorphic Processor for Tiled Matrix MultiplicationThe TMM concept is illustrated in Figs. 1(a)–1(c), showing an example where three different steps are required for calculating the products between two rows of a matrix and a six-element input vector, when matrix tiles are used. The matrix tile starts from the top-left position of the matrix and gets multiplied with the first two input vector values, with the respective products being stored at the first two entries of the six-element output vector, as shown in Fig. 1(a). Then, the matrix tile shifts to the right and the two-element input vector tile shifts down [Fig. 1(b)] to incorporate the next entries of the first two matrix rows and the input vector, respectively, producing in this way two new partial weighted input sums through the multiplication of the tile with the corresponding values of the input vector. This process continues with the tile shifting to the right until the whole horizontal dimension of the matrix has been scanned, as illustrated in Fig. 1(c). The realization of TMM in the optical domain can be accomplished through a photonic MVM engine where inputs and weights can be updated at line rate, supported by an electronic circuitry for storing the matrix values, loading the necessary tiles to the photonic MVM and storing the partial output sums. This visionary architecture is pictorially represented by Fig. 1(d), showcasing all key building blocks of a neuromorphic photonic processor. The MVM linear operations are executed via the photonic MVM processor in the analog domain, utilizing an integrated or external laser source for “lighting up” the processor. The input and weight values are stored at an electronic memory unit and are loaded onto the photonic MVM processor using digital-to-analog converters. The photonic MVM output is connected to an array of photodiodes that transforms the computed signals back to the electronic domain, exploiting an analog-to-digital converter array for the digitization of the data so that they can be stored at the electronic memory. Additionally, an electronic control circuit is needed for data flow synchronization, orchestration, and communication between the memory block and the photonic MVM unit. Yet, with the NN depth and size increasing with problem complexity, the total number of the NN-trainable parameters will reach values well beyond the matrix dimensions supported by the photonic MVM engine, implying that the photonic MVM hardware has to be shared among a larger number of parameters through inter/intralayer or intraneuron TDM techniques (see Supplementary Material). The implementation of the above requires either the continuous update of the emerging partial sums for the calculation of the multiplication of a whole input vector with a weight matrix tile, as shown in Figs. 1(a)–1(c), or by storing all partial sums at different registers and then forcing them again for further addition via the photonic MVM engine. A pictorial example of the intraneuron TDM approach for TMM can be visualized in Fig. 1(e), illustrating how an elementary 2:1 neuron can carry out the linear summations of a five-axon neuron. This corresponds to the dot product between a row vector that contains the weights of the neuron and a column vector that includes the input values, executed through the use of and row and column vector tiles, respectively. The five-input neuron is unrolled into four 2:1 virtual neurons whose linear summation operations can be performed within three phases. More specifically, during the first phase, the 2:1 hardware is utilized in three time slots, calculating the linear summations , , and , with and being zero and one padded input and weight values, respectively. Afterwards, these three partial weighted input sums need to be summed in order to provide the required weighted summation of the five inputs of the neuron. Considering that the addition of the partial sums is carried out again in the optical domain, the summation operation can be performed on-chip by applying weight values equal to 1. Because the hardware can imprint two input values at a time slot, the remaining summations would be performed in two more phases, as depicted in Fig. 1(e). It can be derived that a photonic neuron of number of axons is capable of calculating the linear operations of a layer that comprises neurons of number of inputs each, in phases. The execution of the MVM product required for an entire neural layer via the same 2:1 NN hardware requires both intraneuron and intralayer TDM, with intraneuron and intralayer TDM corresponding to the use of multiple tiles across a single matrix row and a matrix column, respectively. Assuming, for example, a fully connected layer with number of neurons and number of inputs per neuron, Fig. 1(f) illustrates the MVM between a weight matrix with dimensions and an input vector with dimensions , with being equal to the inference batch size . The 2:1 NN hardware encodes two elements of the weight matrix and two elements of the input vectors at a time slot, highlighted in the colored red, green, and yellow rectangles. This is depicted in more detail in Fig. 1(g), where the NN hardware hosts the weight values of the rectangle A in its weighting modules and , along with the input values and , in the respective input modules and during the first time slot. Both input and weight values will be updated during the second time slot, with the weight values of the matrix tile B being loaded onto the weighting modules and and the input values and onto the respective input and stages. This process continues until all partial weighted input sums of the entire matrix and the first-sample input vector are calculated, i.e., until the last matrix row that comprises matrix tiles , , and gets also multiplied with the input vector from the first sample, completing in this way phase #1 of the process. Subsequently, the partial sums will be sequentially employed at the input modules and for their addition until they form the complete weighted input sum that corresponds to the product between a matrix row and the input vector of the first sample. This process is completed within the subsequent phases #2 and #3 of Fig. 1(g), utilizing weighting values equal to one. After completing phase #3 of the first sample, the multiplication of the weight matrix with the input vector in the second sample is initiated, following again the same TMM scheme and repeating all three phases. The MVM operation will be completed once the entire inference batch size has been processed. 3.Silicon Photonic Coherent Linear Neuron ArchitectureFigure 2(a) depicts the SiPho processor that was fabricated for direct on-chip and high-speed mapping of both the input and the weight elements of an NN, following the COLN architecture43 that can implement a dot-product operation. The SiPho chip comprises a coherent neuromorphic architecture that implements a two-input COLN capable of executing multiple-accumulate (MAC) operations, i.e., the weighted summation of its input data. It exploits the interference capabilities of Mach–Zehnder interferometers (MZIs), complemented by a bias branch that safeguards the retention of the sign of the weighted summation (see Supplementary Material). A visualization of the SiPho COLN and the experimental setup established for its evaluation are depicted in Fig. 2(b). Specifically, the SiPho processor comprises five compact and high-bandwidth SiGe EAMs (orange boxes), with two EAMs used in cascade at each MZI branch for on-chip input data and weight imprinting, respectively, and one EAM employed in the bias branch. The selection of the SiGe EAMs allows for a high compute rate, while retaining the energy consumption and the footprint at low values.44,45 The normalized electrooptic response of an EAM biased at is presented in Fig. 2(c), revealing a 1-dB bandwidth higher than 50 GHz. Finally, 3 TO phase shifters (PSs) [blue cylinders in Fig. 2(b)], one at the bias branch and one at each MZI arm, are employed for the application of the sign of the weighted inputs. Figure 2(d) illustrates the optical loss of with respect to the applied driving power, showcasing that is required for π phase shift. Similar behavior was observed in and . 4.Experimental Classification of Benign and Malicious Reconnaissance Attacks Using Photonic TMM at 50 GHzThe NN trained for the RAs classification follows the topology shown in Fig. 3(a). The six features of the port scanned traffic comprise the six inputs of the NN, followed by a fully connected hidden layer (Layer #1) of eight neurons and a two-neuron output layer. The Sigmoid and the SoftMax activation functions were applied to the hidden and output layers, respectively. During the experimental validation of the NN operation, both input and the weight values were updated at the rate of 16 and 50 GHz. Tiled MM was utilized for applying the 64 weight values of the 6:8:2 NN onto the 2:1 SiPho processor, adopting an intralayer and intraneuron TDM approach for calculating the MVM product within Layer #1 before repeating the same process for calculating the MVM products in Layer #2. The accuracy performance of the proposed SiPho neuromorphic prototype was evaluated via the classification of real traffic generated for RA identification (see Supplementary Material). Figures 3(b)–3(g) illustrate the time multiplexed sample traces obtained at the output of each phase during the experimental inference, with the black dashed lines representing the software-obtained traces, and the orange and blue solid lines corresponding to the experimental traces at 16 and 50 Gbaud, respectively. In particular, Figs. 3(b)–3(d) depict the time multiplexed sample traces that were obtained in the three phases of the first layer, respectively, whereas Figs. 3(e)–3(g) illustrate the time-multiplexed sample traces obtained at each phase of the output layer. Figure 3(h) presents the normalized mean squared errors (MSEs) of the experimentally captured signals per inference phase and per layer. The MSEs of the 16 and 50 Gbaud summations at the last phase of Layer #1 equal and , respectively, while the respective MSE values after the summations of the first phase of Layer#2 are reduced to and , respectively. The MSE is always higher at 50 GHz compared with the 16 GHz operational mode and increases as the process moves from the first to the last phase within the same layer, being the result of the noise accumulation that is associated with the reuse of the photonic processor and the higher noise bandwidth. Yet, the interlayer transition reduces the deviation between the experimental and the reference waveform, decreasing the amount of noise upon entering Layer #2 compared with the noise that was accumulated through all Layer #1 phases. This is the result of the Sigmoid activation function employed at Layer #1 output, which takes advantage of its high nonlinearity at its boundaries to compress the edge values of the samples. The inference classification performance of our SiPho prototype when performing with 500 samples of the generated traffic was quantified by calculating the -score, which comprises a statistical metric for the evaluation of the inference accuracy when imbalanced data sets are classified (see Supplementary Material), with their confusion matrices depicted in Figs. 4(a)–4(c). The software acquired -score was calculated to be equal to 0.70, and the respective values of the experimental classification at 16 and 50 Gbaud were measured equal to 0.688 and 0.636, as depicted in Fig. 4(d). Finally, the signal-to-noise ratio (SNR) values of the linear summations emerging from the photonic NN (PNN) were measured equal to 14.1 and 11.2 dB, respectively. 5.DiscussionThe successful proof-of-principle experimental validation of the optical TMM at 16 and 50 GHz and the classification of the RA traffic requires a total number of six TDM phases, which is the result of the 2:1 PNN that, inevitably, necessitates the use of 2:1 matrix tiles. The number of TDM phases required and the associated time overhead can be reduced by scaling the PNN chip into a higher-dimension layout that can host a higher amount of on-chip input and weight modulation elements. Figure 5(a) illustrates how the 2:1 COLN can scale to an neuron by following the layout that has been already mathematically validated and simulated in Ref. 43 and experimentally validated in Refs. 26 and 46. This is based on the introduction of an splitter followed by a stack of parallel waveguides, where every waveguide incorporates a high-speed amplitude modulator for the input signal, followed by a high-speed PS for sign update and a high-speed amplitude modulator for weight update. All these parallel waveguides recombine via an combiner, forming in this way a multibranch interferometer. Extending this architecture into a 2D matrix that can support matrix tiles and further reduce the MVM latency can be realized by adopting an coherent linear photonic Xbar architecture that follows the principles reported in Refs. 47 and 48 and uses EAMs as its input and weight modulation circuitry. The Xbar layout is depicted in Fig. 5(b), with the green rectangle illustrating that the 2:1 COLN utilized in our PNN chip [Fig. 2(a)] comprises a subblock within the design. The Xbar architecture can host neurons with axons per neuron simultaneously, allowing for the use of matrix tiles within the TMM process. The credentials of this architecture to support high-dimension matrix tiles within practical total loss values can be verified through a quantitative theoretical insertion loss (IL) analysis using experimentally measured specifications, assuming a symmetric Xbar that employs SiGe EAMs both for the input and the weight values (see Supplementary Material). As shown in Fig. 6, the total IL of the Xbar architecture increases with increasing matrix dimensions but retains a reasonable value of less than 30 dB, even for a layout, which supports a total amount of 1024 MAC operations. This can scale to higher total MAC capabilities by combining wavelength division multiplexing (WDM) with the coherent Xbar scheme, following the design reported in Refs. 47 and 49, and can support tensor tiles, with representing the number of wavelengths employed. The extension of the 2:1 COLN into an coherent Xbar design retains all its additional benefits with respect to flexibility, robustness, and energy and footprint efficiency50,51 (see also Supplementary Material), as it allows for one-to-one and high-fidelity single-step mapping of the NN parameters onto the PNN hardware48 and the deployment of high-speed nodes’ technology. 6.ConclusionRecent advances in SiPhos have enabled the exploitation of light for computing by accelerating the execution of deep-learning algorithms. In this work, we demonstrate an SiPho processor that is capable of performing linear algebra operations of any-dimensioned NN layer towards classifying, at record-high speeds, DDOS attacks within DC server packets. Specifically, by employing the TMM method, we were able to accelerate the MAC operations that take place into the AI processor up to the rate of 50 GHz, detecting successfully benign and malicious attacks with a -score of 0.636. Finally, towards minimizing the computing steps and maximizing the classification speeds, we provide a dimension scaling analysis of the demonstrated prototype into a space division multiplexed Xbar architecture capable of supporting layer-scale linear algebra operations. AcknowledgmentsThe authors would like to thank Dr. Frederick Gardes, Teerapat Rutirawut, and the University of Southampton for the wire bonding of the SiPho chip. Additionally, the authors would like to thank Nikos Bamiedakis for fruitful discussions. The work was in part funded by the EU-projects PlasmoniAC (Grant No. 871391), SIPHO-G (Grant No. 101017194) and by Hellenic Foundation for Research and Innovation (H.F.R.I.) under the “First Call for H.F.R.I. Research Projects to Support Faculty Members and Researchers and the Procurement of High-cost Research Equipment Grant” (Grant No. 4233, DeepLight). Author ContributionsGeorge Giamougiannis, Apostolos Tsakyridis, Miltiadis Moralis-Pegios, George Dabos, David Lazovsky, and Nikos Pleros conceived the experiment. Dimitrios Kalavrouziotis, Dimitris Syrivelis, Paraskevas Bakopoulos, and Elad Mentovich performed the SYN port scanning and the DDoS identification NN modeling. Nikolaos Passalis, Manos Kirtas, and Anastasios Tefas performed the NN training. George Giamougiannis, Apostolos Tsakyridis, George Mourgias-Alexandris, and Miltiadis Moralis-Pegios deployed the experimental setup, performed the experiment, and processed the experimental results. George Giamougiannis, Apostolos Tsakyridis, Angelina Totovic, and Nikos Pleros conceived the coherent linear Xbar scaling and performed the simulation analyses. All authors discussed the results and wrote the manuscript. Data AvailabilityThe data that support the plots within this paper and other findings of this study are available from the corresponding author upon reasonable request. ReferencesY. LeCun, Y. Bengio and G. Hinton,
“Deep learning,”
Nature, 521
(7553), 436
–444 https://doi.org/10.1038/nature14539
(2015).
Google Scholar
P. R. Prucnal et al.,
“Multiwavelength neuromorphic photonics,”
in Conf. Lasers and Electro-Opt. (CLEO),
(2019). Google Scholar
Z. Chen et al.,
“A fast deep learning system using GPU,”
1552
–1555
(2014). https://doi.org/10.1109/ISCAS.2014.6865444 Google Scholar
C. Zhang et al.,
“Optimizing FPGA-based accelerator design for deep convolutional neural networks,”
in Proc. 2015 ACM/SIGDA Int. Symp. Field-Programm. Gate Arrays, ser. FPGA ’15,
161
–170
(2015). Google Scholar
“Cloud Tensor Processing Units (TPUs),”
(4 January 2023). https://cloud.google.com/tpu/docs/tpus Google Scholar
L. Cavigelli and L. Benini,
“Origami: a 803-GOp/s/W convolutional network accelerator,”
IEEE Trans. Circuits Syst. Video Technol., 27
(11), 2461
–2475 https://doi.org/10.1109/TCSVT.2016.2592330 ITCTEM 1051-8215
(2017).
Google Scholar
Y. Chen et al.,
“A survey of accelerator architectures for deep neural networks,”
Engineering, 6
(3), 264
–274 https://doi.org/10.1016/j.eng.2020.01.007 ENGNA2 0013-7782
(2020).
Google Scholar
T. N. Theis and H.-S. P. Wong,
“The end of Moore’s law: a new beginning for information technology,”
Comput. Sci. Eng., 19
(2), 41
–50 https://doi.org/10.1109/MCSE.2017.29
(2017).
Google Scholar
L. X. Hu et al.,
“All-optically controlled memristor for optoelectronic neuromorphic computing,”
Adv. Funct. Mater., 31
(4), 2005582 https://doi.org/10.1002/adfm.202005582 AFMDC6 1616-301X
(2021).
Google Scholar
J. Yanh et al.,
“Optically driven intelligent computing with ZnO memristor,”
Fundam. Res.,
(2022).
Google Scholar
F. Brückerhoff-Plückelmann et al.,
“A large scale photonic matrix processor enabled by charge accumulation,”
Nanophotonics, https://doi.org/10.1515/nanoph-2022-0441
(2022).
Google Scholar
X. Li et al.,
“On-chip phase change optical matrix multiplication core,”
7.5.1
–7.5.4
(2020). https://doi.org/10.1109/IEDM13553.2020.9372052 Google Scholar
K. Portner et al.,
“Analog nanoscale electro-optical synapses for neuromorphic computing applications,”
ACS Nano, 15
(9), 14776
–14785 https://doi.org/10.1021/acsnano.1c04654 ANCAC3 1936-0851
(2021).
Google Scholar
V. J. Sorger et al.,
“Ultra-compact silicon nanophotonic modulator with broadband response,”
Nanophotonics, 1
(1), 17
–22 https://doi.org/10.1515/nanoph-2012-0009
(2012).
Google Scholar
Y. Shen et al.,
“Deep learning with coherent nanophotonic circuits,”
Nat. Photonics, 11
(7), 441
–446 https://doi.org/10.1038/nphoton.2017.93 NPAHBY 1749-4885
(2017).
Google Scholar
J. Grollier et al.,
“Neuromorphic spintronics,”
Nat. Electron., 3
(7), 360
–370 https://doi.org/10.1038/s41928-019-0360-9 NEREBX 0305-2257
(2020).
Google Scholar
Z. Guo et al.,
“Spintronics for energy-efficient computing: an overview and outlook,”
Proc. IEEE, 109
(8), 1398
–1417 https://doi.org/10.1109/JPROC.2021.3084997 IEEPAD 0018-9219
(2021).
Google Scholar
M. A. Nahmias et al.,
“Photonic multiply-accumulate operations for neural networks,”
IEEE J. Sel. Top. Quantum Electron., 26
(1), 7701518 https://doi.org/10.1109/JSTQE.2019.2941485 IJSQEN 1077-260X
(2020).
Google Scholar
A. Totovic et al.,
“Femtojoule per MAC neuromorphic photonics: an energy and technology roadmap,”
IEEE J. Sel. Top. Quantum Electron., 26
(5), 8800115 https://doi.org/10.1109/JSTQE.2020.2975579 IJSQEN 1077-260X
(2020).
Google Scholar
J. Feldmann et al.,
“Parallel convolutional processing using an integrated photonic tensor core,”
Nature, 589
(7840), 52
–58 https://doi.org/10.1038/s41586-020-03070-1
(2021).
Google Scholar
G. Giamougiannis et al.,
“Silicon-integrated coherent neurons with 32GMAC/sec/axon compute line-rates using EAM-based input and weighting cells,”
in Eur. Conf. Commun.,
(2021). Google Scholar
S. Ohno et al.,
“Si microring resonator crossbar array for on-chip inference and training of optical neural network,”
(2021). Google Scholar
C. Huang et al.,
“Demonstration of scalable microring weight bank control for large-scale photonic integrated circuits,”
APL Photonics, 5
(4), 040803 https://doi.org/10.1063/1.5144121
(2020).
Google Scholar
G. Mourgias-Alexandris et al.,
“A silicon photonic coherent neuron with 10GMAC/sec processing line-rate,”
in Opt. Fiber Comm. Conf.,
Tu5H.1
(2021). Google Scholar
B. Shi, N. Calabretta and R. Stabile,
“Deep neural network through an InP SOA-based photonic integrated cross-connect,”
IEEE J. Sel. Top. Quantum Electron., 26
(1), 7701111 https://doi.org/10.1109/JSTQE.2019.2945548 IJSQEN 1077-260X
(2020).
Google Scholar
H. Zhang et al.,
“An optical neural chip for implementing complex-valued neural network,”
Nat. Commun., 12 457 https://doi.org/10.1038/s41467-020-20719-7
(2021).
Google Scholar
W. Zhang et al.,
“Microring weight banks control beyond 8.5-bits accuracy,”
(2021). Google Scholar
J. K. George,
“Neuromorphic photonics with electro-absorption modulators,”
Opt. Express, 27
(4), 5181
–5191 https://doi.org/10.1364/OE.27.005181
(2019).
Google Scholar
C. Wu et al.,
“Programmable phase-change metasurfaces on waveguides for multimode photonic convolutional neural network,”
Nat. Commun., 12
(1), 96 https://doi.org/10.1038/s41467-020-20365-z NCAOBW 2041-1723
(2021).
Google Scholar
G. Dabos et al.,
“Neuromorphic photonic technologies and architectures: scaling opportunities and performance frontiers [Invited],”
Opt. Mater. Express, 12
(6), 2343
–2367 https://doi.org/10.1364/OME.452138
(2022).
Google Scholar
G. Mourgias-Alexandris et al.,
“An all-optical neuron with sigmoid activation function,”
Opt. Express, 27
(7), 9620
–9630 https://doi.org/10.1364/OE.27.009620 OPEXFF 1094-4087
(2019).
Google Scholar
C. Huang et al.,
“Programmable silicon photonic optical thresholder,”
IEEE Photonics Technol. Lett., 31
(22), 1834
–1837 https://doi.org/10.1109/LPT.2019.2948903 IPTLEL 1041-1135
(2019).
Google Scholar
R. Amin et al.,
“ITO-based electro-absorption modulator for photonic neural activation function,”
APL Mater., 7
(8), 081112 https://doi.org/10.1063/1.5109039
(2019).
Google Scholar
I. A. D. Williamson et al.,
“Reprogrammable electro-optic nonlinear activation functions for optical neural networks,”
IEEE J. Sel. Top. Quantum Electron., 26 7700412 https://doi.org/10.1109/JSTQE.2019.2930455 IJSQEN 1077-260X
(2019).
Google Scholar
M. Moralis-Pegios et al.,
“Neuromorphic silicon photonics and hardware-aware deep learning for high-speed inference,”
J. Lightwave Technol., 40
(10), 3243
–3254 https://doi.org/10.1109/JLT.2022.3171831 JLTEDG 0733-8724
(2022).
Google Scholar
G. Mourgias-Alexandris et al.,
“Channel response-aware photonic neural network accelerators for high-speed inference through bandwidth-limited optics,”
Opt. Express, 30
(7), 10664
–10671 https://doi.org/10.1364/OE.452803 OPEXFF 1094-4087
(2022).
Google Scholar
I. Kandel and M. Castelli,
“Transfer learning with convolutional neural networks for diabetic retinopathy image classification: a review,”
Appl. Sci., 10
(6), 2021 https://doi.org/10.3390/app10062021
(2020).
Google Scholar
G. E. Moon et al.,
“Evaluating spatial accelerator architectures with tiled matrix-matrix multiplication,”
IEEE Trans. Parallel Distrib. Syst., 33
(4), 1002
–1014 https://doi.org/10.1109/TPDS.2021.3104240
(2022).
Google Scholar
“NVIDIA app. Note,”
https://docs.nvidia.com/deeplearning/performance/dl-performance-matrix-multiplication/index.html
(4 January 2023).
Google Scholar
M. L. McHugh,
“Interrater reliability: the kappa statistic,”
Biochem. Med., 22
(3), 276
–282 https://doi.org/10.11613/BM.2012.031
(2012).
Google Scholar
G. Mourgias-Alexandris et al.,
“Neuromorphic photonics with coherent linear neurons using dual-IQ modulation cells,”
J. Lightwave Technol., 38
(4), 811
–819 https://doi.org/10.1109/JLT.2019.2949133 JLTEDG 0733-8724
(2020).
Google Scholar
M. Pantouvaki et al.,
“Active components for 50 Gb/s NRZ-OOK optical interconnects in a silicon photonics platform,”
J. Lightwave Technol., 35
(4), 631
–638 https://doi.org/10.1109/JLT.2016.2604839 JLTEDG 0733-8724
(2017).
Google Scholar
G. Sinatkas et al.,
“Electro-optic modulation in integrated photonics,”
J. Appl. Phys., 130
(1), 010901 https://doi.org/10.1063/5.0048712 JAPIAU 0021-8979
(2021).
Google Scholar
H. Zhou et al.,
“Photonic matrix multiplication lights up photonic accelerator and beyond,”
Light Sci. Appl., 11
(1), 30 https://doi.org/10.1038/s41377-022-00717-8
(2022).
Google Scholar
A. Totovic et al.,
“WDM equipped universal linear optics for programmable neuromorphic photonic processors,”
Neuromorph. Comput. Eng., 2
(2), 024010 https://doi.org/10.1088/2634-4386/ac724d
(2022).
Google Scholar
G. Giamougiannis et al.,
“A coherent photonic crossbar for scalable universal linear optics,”
J. Lightwave Technol., https://ieeexplore.ieee.org/document/10007926 JLTEDG 0733-8724
(2022).
Google Scholar
A. Totovic et al.,
“Programmable photonic neural networks combining WDM with coherent linear optics,”
Sci Rep., 12
(1), 5605 https://doi.org/10.1038/s41598-022-09370-y
(2022).
Google Scholar
A. Tsakyridis et al.,
“Universal linear optics for ultra-fast neuromorphic silicon photonics towards Fj/MAC and TMAC/sec/mm2 engines,”
IEEE J. Sel. Top. Quantum Electron., 28
(6), 8300815 https://doi.org/10.1109/JSTQE.2022.3219288 IJSQEN 1077-260X
(2022).
Google Scholar
G. Giamougiannis et al.,
“Universal linear optics revisited: new perspectives for neuromorphic computing with silicon photonics,”
IEEE J. Sel. Top. Quantum Electron., 28 6200116 https://doi.org/10.1109/JSTQE.2022.3228318 IJSQEN 1077-260X
(2022).
Google Scholar
J. Zhang et al.,
“Deep learning based attack detection for cyber-physical system cybersecurity: a survey,”
IEEE/CAA J. Autom. Sin., 9
(3), 377
–391 https://doi.org/10.1109/JAS.2021.1004261
(2022).
Google Scholar
M. Ge et al.,
“Deep learning-based intrusion detection for IoT networks,”
256
–25609
(2019). https://doi.org/10.1109/PRDC47002.2019.00056 Google Scholar
Z. Sheng et al.,
“A compact and low-loss MMI coupler fabricated with CMOS technology,”
IEEE Photonics J., 4
(6), 2272
–2277 https://doi.org/10.1109/JPHOT.2012.2230320
(2012).
Google Scholar
B. Sharma et al.,
“Design and simulation of ultra-low loss triple tapered asymmetric directional coupler at 1330 nm,”
Microelectron. J., 107 104957 https://doi.org/10.1016/j.mejo.2020.104957 MICEB9 0026-2692
(2021).
Google Scholar
Y. Ma et al.,
“Ultralow loss single layer submicron silicon waveguide crossing for SOI optical interconnect,”
Opt. Express, 21
(24), 29374
–29382 https://doi.org/10.1364/OE.21.029374 OPEXFF 1094-4087
(2013).
Google Scholar
BiographyGeorge Giamougiannis received his Diploma degree in “Electrical and Computer Engineering” from the Aristotle University of Thessaloniki, Greece, in 2017. At the beginning of 2020 he received his MSc in “Communication Networks and Systems Security” from the same department. Since September 2019, he is a member of the WinPhos research group pursuing his PhD with Prof. Nikos Pleros as his supervisor. His research interests focus on linear optics and neuromorphic photonics and he has (co-)authored 13 journal papers, 15 papers in international conferences, and 3 US patents. |