Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci. , 14 February 2025

Sec. Neuromorphic Engineering

Volume 19 - 2025 | https://doi.org/10.3389/fnins.2025.1511371

Evaluation of fluxon synapse device based on superconducting loops for energy efficient neuromorphic computing

  • 1Department of Electrical and Computer Engineering, University of California San Diego, San Diego, CA, United States
  • 2Department of Physics, University of California San Diego, San Diego, CA, United States
  • 3Department of Chemical and Nano Engineering, University of California San Diego, San Diego, CA, United States

With Moore’s law nearing its end due to the physical scaling limitations of CMOS technology, alternative computing approaches have gained considerable attention as ways to improve computing performance. Here, we evaluate performance prospects of a new approach based on disordered superconducting loops with Josephson-junctions for energy efficient neuromorphic computing. Synaptic weights can be stored as internal trapped fluxon states of three superconducting loops connected with multiple Josephson-junctions (JJ) and modulated by input signals applied in the form of discrete fluxons (quantized flux) in a controlled manner. The stable trapped fluxon state directs the incoming flux through different pathways with the flow statistics representing different synaptic weights. We explore implementation of matrix–vector-multiplication (MVM) operations using arrays of these fluxon synapse devices. We investigate the energy efficiency of online-learning of MNIST dataset. Our results suggest that the fluxon synapse array can provide ~100× reduction in energy consumption compared to other state-of-the-art synaptic devices. This work presents a proof-of-concept that will pave the way for development of high-speed and highly energy efficient neuromorphic computing systems based on superconducting materials.

1 Introduction

In the era of artificial intelligence (AI), artificial neural networks (ANNs) are at the forefront of the modern computing paradigms used in a wide range of applications, including image and video classification, pattern recognition, natural language processing (NLP), autonomous vehicles, robotics, gaming, virtual reality, and bioinformatics (Heaton, 2018; LeCun et al., 2015; O'Kelly et al., 2020; Schmidhuber, 2015; Vamathevan et al., 2019; Vaswani et al., 2017) The complexity and scaling requirement of targeted applications push AI models towards more complex and larger architectures. This results in significant increase in the overall energy requirement while performing massive training exercises for such large and complex models, leading to severe environmental issues in the future (Boahen, 2022). Specifically, multiply and accumulate (MAC) operations in ANNs contribute ∼ 70–90% to the total operational cost (Jouppi et al., 2017). Therefore, developing energy-efficient neuromorphic hardware solutions has become one of the most critical challenges for future computing systems.

Hardware accelerators based on various types of nanoelectronic devices have already been proposed to improve performance of various neuromorphic applications and algorithms (Dutta et al., 2021; Dutta et al., 2020; Ge et al., 2018; Kumar and Bezugam, 2024; Oh et al., 2021; Oh et al., 2018; Park et al., 2024; Rafiq et al., 2023; Selcuk et al., 2024; Seo et al., 2020; Shi et al., 2018). For large and energy efficient neural network implementations, a synaptic device must show some of the main attributes, i.e., low energy consumption for read and write operations, scalability to achieve high synaptic density, high reliability, in-situ nonvolatile storage and computation, and linear and symmetric synaptic weight updates, i.e., upward and downward (Kuzum et al., 2012; Park et al., 2024). However, many synaptic device candidates still suffer from limited precision, large variations, and high energy consumption to achieve the required conductance values representing the weights of a network. There has been recent interest in exploring disordered physical systems which exhibit a complex energy landscape with a finite number of local minima exhibiting synaptic memory behavior. These have long been considered as models to describe emergent computational behavior displayed by neural networks including our brain from the perspective of statistical mechanics (Hopfield, 1982; Little, 1974; Niazi et al., 2024). Such systems ‘collectively’ host an almost continuum of states that can be used to represent synaptic memory configurations, in an alternative approach when compared to distributed synaptic memory states.

In this context, fluxon synapses based on superconducting loops combined with Josephson Junctions (JJs) arise as a promising technology that can offer several advantages, including low power consumption, high-speed operation, indefinitely large endurance and scalability.

JJ translate input excitations to flux flow (fluxons). JJs are the only switching elements in the system and superconducting loops are responsible for storage of trapped fluxon as circulating superconducting currents. These JJs can be switched without any degradation and cycled indefinitely without change (Duzer, 1989). It has also been validated that the switching current ‘IC’ of a well fabricated JJ is independent of repetition rate and applied magnetic field (or input) from one test to another (Schroen, 1968). Russek et al. (2016) proposed a fluxon based neuromorphic computing for large scale neuromorphic system with each JJ spiking at the rate of ~1010 per second. This indicates the indefinite reuse of fluxon based synaptic device if used under optimum operating conditions. These Josephson memory cells store information as persistent circulating currents and equivalent to fluxons in superconducting loops. No refreshing/rewriting is necessary because currents can be maintained indefinitely in ideal lossless superconducting loops (Zmpe, 1980). This method of storage is nonvolatile and has no power consumption after storing the state. In a superconducting loop system, the magnetic flux is quantized as fluxons (Φ0 = 2.065×10−15 T/m2), and cells have been built which operate either with only one flux quantum, or with many of them with non-destructive read (Wolf, 1978; Zmpe, 1980; Goteti et al., 2022). JJs work as bridges between superconducting loops and can operate at high speeds up to a few THz. JJs are interconnected due to macroscopic coherence with long-range interactions in superconductor loops and display a rich spectrum of memory states while having zero-static power dissipation (Goteti et al., 2022; Goteti and Dynes, 2021; Jué et al., 2022; Schneider et al., 2022). The memory states trapped in the form of fluxons in the loops result in stable flux flow pathways when excited with input signals. The flow pathways can be characterized as synaptic weights from statistical correlations of flux between Josephson Junctions.

Recently, a few studies have proposed disordered networks consisting of several superconducting loops with Josephson’s junction showing stable memory configurations of trapped fluxons in loops and movement of spike signals (e.g., neuronal activity) in small-scale disordered networks of superconducting loops (Goteti et al., 2022; Goteti et al., 2024; Goteti and Dynes, 2021). The study of physics of collective behavior of these randomly connected superconducting loops showed a great promise to perform some neuromorphic computations at small-scale (Goteti et al., 2022; Goteti and Dynes, 2021). However, to date such disordered superconducting loops have not been studied at the network-level to perform complex computations more relevant to modern AI models or to implement learning and inference tasks using a standard dataset. In this work, we explore crossbar architecture based on a previously studied configuration with three superconducting loops as individual synaptic elements and evaluate its performance using standard models and datasets. We experimentally characterize the fluxon synapses to investigate stable memory states when arranged into a crossbar structure to implement MVM operations in a 2-layer MLP neural network. We simulate fluxon synapse crossbar based synaptic core to perform learning and inference tasks on the MNIST dataset. Finally, we benchmark the performance and energy efficiency of the proposed superconducting (disordered) loop based synaptic device against other state-of-the-art synaptic device technologies.

2 Fabrication of Josephson Junction (JJ) and three superconducting loop device

A Josephson Junction (JJ) is a superconductor-insulator-superconductor (SIS) structure. A helium ion microscope can be used to selectively create a tunnel barrier (insulator region) in high-Tc (85 K) YBa2Cu3O7 (YBCO) superconductors to form a JJ by exposure to focused He+ ions. The detailed fabrication steps are provided in the methods section. A JJ generates quantized flux, also called fluxons when a current greater than a critical current IC passes through the JJ (Fulton et al., 1973; Likharev and Semenov, 1991). A superconducting YBCO loop consisting of JJs can trap such quantized flux in the form of either clockwise or counterclockwise circulating supercurrents.

A structure consisting of three YBCO superconducting loops has been fabricated with JJ tunnel barriers as shown in Figures 1A,B. The JJs connect loops together and allow fluxon movement in and out of the loops. The fluxons in the form of voltage spikes are fed in and out of the network of loops through JJs at input and output nodes as designated in Figure 1B. Loops labeled 1 to 3 can trap multiples of fluxons while an additional loop is used to induce individual fluxons through the input junction J1 using input current I1 as shown in Figure 1B. A constant flow of fluxons induced at the input can be measured by the voltage V 1 and the resulting output flow of fluxons can be measured as V 2 as illustrated in Figure 1B. Fluxons are injected at a frequency of V Φ 0 given by the Josephson Equations 13, where each individual fluxon is represented by a voltage spike and Φ 0 represents a single fluxon as shown in Figure 1C.

I = I c sin φ + Φ 0 2 π R d φ d t     (1)
v = Φ 0 2 π d φ d t     (2)
V d t = Φ ο ; Φ 0 = 2.065 × 10 15 ; T / m 2     (3)
Figure 1
www.frontiersin.org

Figure 1. (A) Illustrates the fabrication of Josephson junction in between two YBCO superconducting regions. (B) Optical microscope image of a YBCO-based fabricated three superconducting loop circuit illustrating loops’ structure, all junctions from input (J1) to output junction (J6). All Josephson junctions lie within a square of a 200 μm x 200 μm of fabricated superconducting loop structure which is exposed to chemically remove the gold layer while maintaining the YBCO thin film. (C) Shows the high-frequency simulation of a few individual input spikes (fluxons) entering superconducting loops through input junction J1 and exiting through one of output-junction J6 (shown in B). At these timescales, the output spiking activity appears stochastic with a constant frequency activation at the input. Fluxon statistics is averaged over many spikes and results in a steady flux flow pattern and fixed output number.

Here, φ is the superconducting phase difference across the junction and Φ 0 is the flux quantum.

A schematic of a simple three loop superconducting fluxon synapse circuit illustrating the loops, JJs, and input & output spikes is shown in Figure 2A. Input and output of fluxon synapses are spikes, represented as equivalent voltages ( 0 t V d t =nΦ0). When a constant flow of spikes (as VIN) is applied to the input (J1) of a three loop fluxon synapse, it produces a flow of spikes (as VOUT) at the output junction (J6) as illustrated in Figure 2A. The connected superconducting loops exhibit macroscopic coherence across all the connected loops. However, JJs can be strongly or weakly connected to each other, and the connection strength can be systematically programmed using the trapped fluxon configurations (Goteti et al., 2024; Goteti and Dynes, 2021).

Figure 2
www.frontiersin.org

Figure 2. (A) Schematic of a fluxon synapse device based on three loop superconducting circuit with JJs. Spikes represented as equivalent voltages VIN and VOUT at J1 and J6. (B) A possible connectivity schematic of involved JJs in the superconducting loop network. (C) Schematics of superconducting loops with different fluxon configurations with circulating currents (clockwise and anticlockwise) representing non-volatile synaptic states.

Figure 2B shows this network level connectivity of JJs in the three superconducting loops consisting of 6 JJs. Two distinct signal flow pathways can be identified through J1-J4-J5-J6 and J1-J2-J3 in the superconducting structure as illustrated in Figure 2B. In this structure, fluxons can be stabilized into different flux configurations representing different nonvolatile synaptic weights as illustrated in Figure 2C. Each superconducting loop can trap multiples of fluxons (Φ0) and internally results in either clockwise or counterclockwise circulating supercurrents around each loops as illustrated in Figure 2C. The arrow width represents number (n) of trapped fluxons (Φ0) and direction represents a resultant circulating current in the loop either in clockwise or counterclockwise. A typical trapped fluxon configuration with circulating currents around an individual loops can be shown as nΦ0. The input flux diverges into these two pathways and one of these pathway signals can be read across the output node J6 as V 2 . The strength/weight of the pathway between input and output is defined as Equation 4.

p = V 2 V 1 = # output fluxons # input fluxons     (4)

Several fluxon configurations are possible for this superconducting loop array, which result in the two flux (signal) flow pathways with different weights between input and output nodes. The amount of flux in individual loops together with the direction of the resulting input or output flux can be used to represent different weights. The output spikes individually mimic a stochastic pattern in time when observed at the pico-second scale as illustrated in Figure 1C using an exemplary high-frequency simulation of few individual input spikes/fluxons entering at input junction. However, when the spiking statistics are averaged for longer times (i.e., 500 ps) over many spikes it results in a steady flux flow pattern leading to stable synaptic weights between the pairs J1-J4, J4-J5, and J5-J6.

The measured I-V characteristics at input (J1) and output (J6) junctions are shown in Figures 3A,B. The fluxon states as a function of voltage and the corresponding input spiking frequency derived from the I-V characteristics. Figure 3C contains the connection weights (fluxon states) between input J1 and only selected output node J6. Figure 3C shows a larger number of weights (ratioed value) due to a small input step size (~1uV) and larger sweep range up to 1 mV. However, adding an additional bias current (denoted by I2 in Figures 1B, 2A) enables accessing different energy minima (fluxon states) by reconfiguring the energy configuration space inside the three loops fluxon synapse device. The obtained synaptic weights of different fluxon states as a function of input voltage in Figure 3C corresponds to zero I2 bias current. With longer integration times the superconducting loops are subjected to steady flow input patterns and the resulting synaptic weights as a function of different input spiking frequencies is shown in Figure 3C. The fluxon flow rate at each junction is quantified by the number of discrete fluxons traversing the junction over a fixed period, which can be characterized by the constant average frequency or voltage. Adjusting the number of input fluxons within a fixed period can be interpreted as either a change in voltage (with fixed duration) or a change in fluxon frequency (with fixed amplitude). This relationship can be simply defined by considering the fluxons amplitude (Φ0) over a fixed integration time (i.e., f = V/(nΦ0)). Figure 3D shows the stable relaxed energy states when input V 1 is varied systematically from 0 to 1.0 mV. Figure 3D also shows the physical significance of fluxon storage in the superconducting loops as a change in synaptic strength. The energy of states in three loop circuit can be estimated corresponds to the excitation input voltage V1 (i.e., average flow rate ‘n’ is V10) for a fixed duration of t1 = 1 ns. Discrete energy states get modulated using different pulse heights of input V1 as shown in Figure 3D. The value of the energy state is estimated using Equation (5) where, n is the number of input spikes, IC – junction critical current (~100 μA), N - number of junctions between input and output, P6 – output spike ratio at J6, P3- output spike ratio at J3.

E 3loop = n I C Φ 0 + n I C Φ 0 N J 1 J 6 P 6 + n I C Φ 0 N J 1 J 3 P 3     (5)
Figure 3
www.frontiersin.org

Figure 3. (A) Applied input current and measured voltage characteristics corresponding to fluxons (spikes) generated at input junction/node J1. (B) Applied input current and measured voltage characteristics corresponding to fluxons (spikes) generated at output junction J6. (C) Synaptic weights of different fluxon states as a function of input voltage and the corresponding input spiking frequency, derived from the I-V characteristics at zero bias current between input J1 and only selected output node J6. The calculated spiking ratio (using Equation 4) from 0 to 0.4 represents the connection weight between input and output. The calculated ratios are mapped to the synaptic weight from the lowest to the highest value. (D) Different states (denoted by their potential energy due to the trapped fluxons in the loops) achieved after relaxation as the input V1 flows systematically in the loops. Inset shows representative stable states.

3 Superconducting loop array for neuromorphic computing tasks

Performing MVM operation in a crossbar structure is a widely used approach to accelerating neural networks towards achieving massive parallelism through in-memory computing (Gokmen and Vlasov, 2016; Hu et al., 2018; Li et al., 2018). In this section, we investigate crossbar arrays of fluxon synapses to perform MVM operations. Different numbers of synaptic weight states, i.e., 32, 64, 128, and 256 are obtained using fluxon synapses based on three superconducting loops in the form of energy states as shown in Figure 4A. These different numbers of states are achieved by varying input voltages in ~100 μV to 500 μV range with different step sizes (i.e., ~12 μV, 6 μV, 3 μV, and 1.5 μV). The write operation involves the application of high frequency spike input (equivalent to large write voltage) to reach the desired number of output fluxons (synaptic weight) whereas, the read operation involves application of a fixed lower frequency spike input (equivalent to smaller read voltage) to accumulate the output fluxon statistics without changing the internal energy state of the superconducting loops. The write operation excites the superconducting loops with a number of high frequency fluxons (e.g., 100 GHz) generated by a higher input voltage (≥100 μV) to induce a change in the energy state. Synaptic weight is modulated by changing the fluxon excitation frequencies (or equivalent input voltages, Figure 3C). During read operations, we use the low frequency fluxons (e.g., 1 GHz) generated by a smaller input voltage (2.065 μV). Therefore, for the MVM implementation with crossbars utilize low fluxon excitation frequency (~1 GHz), which could be understood as a low amplitude read signal that will not disturb the programmed weights. Figure 4B shows upward and downward changes in energy states during implementation of the nonvolatile synaptic weight update operation in a neural network with input voltage step size change of 6 μV to obtain 64 states. In the fluxon synapse device, there is no distinction between LTP and LTD states. When the device is properly configured, a particular synaptic state (energy minima) corresponds directly to the absolute input signal, regardless of LTP or LTD operation. These nonvolatile changes in the internal energy states represent weight-matrix elements in the synaptic core as shown in Figure 5A. In the system-level simulation framework, we employed an incremental/decremental pulse scheme with identical specifications to achieve the different states in the synaptic core.

Figure 4
www.frontiersin.org

Figure 4. Number of energy states representing synaptic levels. (A) Different number of states (i.e., 64, 128, and 256) obtained using three superconducting loops under the controlled application of input fluxons. (as V1 in Figure 3) with different input steps. These different numbers of energy states correspond to input voltages varied in between ~100 μV and 500 μV with different step sizes (i.e., ~6 μV, 3 μV, and 1.5 μV). (B) Gradual modulation of energy states in both upward and downward directions with input voltages representing the number of applied input fluxons or rate of fluxons for the 64 states.

Figure 5
www.frontiersin.org

Figure 5. (A) Crossbar structure of the synaptic core implemented with three superconducting loops acting individual synaptic element. (B) Schematic of implemented MLP neural architecture and used MNIST Dataset for image classification.

3.1 Neural network implementation and system-level performance benchmarking

The fluxon synapse device can be used to store and update/train the weights of multilayer perceptron (MLP) neural network implementation. We performed system level assessment using experimental data from fluxon synapse devices for classification of the handwritten digits from MNIST dataset (LeCun, 1998) consisting of 60,000 training and 10,000 test images. The implemented 2-layer MLP network of sized 400(input)-100(hidden)-10(output) nodes, where the edge cropped 20×20 MNIST images are used for 400 input nodes as illustrated in Figure 5B. Two different sizes of weight matrices, WIH (40,000) and WHO (1,000), corresponding to input-hidden and hidden-output layers respectively, are implemented (Figure 5B).

To evaluate the system level hardware performance for online-learning, we used the circuit-level macro-model NeuroSim platform (Chen et al., 2018). We emulated the synaptic core hardware for MLP network using the fluxon synaptic devices. For benchmarking purposes, we performed the online learning within NeuroSim framework for fluxon synaptic core and various other synaptic devices. For emulating the proper crossbar level operation, we assumed superconducting interconnections with extremely low resistivity. These interconnects are used inside the crossbar architecture to connect fluxon synapses and propagate the required signals in and out of the array. For benchmarking, we focused on the metrics for the crossbar arrays and did not include the periphery for a fair comparison between different synaptic device technologies.

For on-line learning with the MLP network, we used low precision step activation function for simpler hardware implementation and Adam optimizer. We utilized backpropagation as the weight update algorithm. The network is trained over 125 epochs with Adam optimizer to obtain classification accuracy results using 64 synaptic/energy states per crosspoint, as shown in Figure 6A. In case of high C2C variations Adam optimizer with lower learning rates compared to in case of stochastic gradient dissent (SGD) optimizer shows better accuracy with an increase in energy consumption. SGD shows relatively higher variance in results and low accuracy due to stochastic gradient choice between application of gradient function while minimizing the training error. However, it converges faster than other optimizers and provides the optimal solution in case of large number of training/test cases with small C2C variations. SGD also has smaller memory footprints and learns quickly (Gupta et al., 2021; Ruder, 2016; Zhou et al., 2020). Finally, we benchmarked the fluxon synaptic core implementation with different synaptic crossbars using other state of the art device technologies, i.e., RRAM, PCM, FeFET etc. The fluxon synaptic core reduces the overall on-chip learning energy requirement significantly while achieving comparable learning accuracy using only 64 synaptic states as shown in Figure 6B and Table 1.

Figure 6
www.frontiersin.org

Figure 6. (A) On-chip learning accuracy of implemented MLP network for MNIST image classification over 125 epochs with 64 synaptic states using the fluxon synaptic core. (B) Online learning energy consumption of different synaptic cores using MNIST dataset over 125 epochs.

Table 1
www.frontiersin.org

Table 1. System level benchmark results for state-of-art synaptic devices.

4 Discussion

In this work, we investigated and evaluated potential performance of a synaptic core made of fluxon synapse devices for highly energy efficient neuromorphic computing. The fluxon synapse device exhibits nonvolatile states as well as gradual modulation of states by the application of varied fluxon excitation frequency. Our experimental results show that the fluxon synaptic devices have the capability to provide a significantly large number of synaptic states, which can be leveraged for implementation of low energy on-chip learning with high precision weights. We performed system-level simulations for a hardware implementation of MLP network with the fluxon synaptic core. We benchmarked the superconducting loop synaptic core against the state-of-art synaptic devices, i.e., RRAM, PCM, EpiRAM, and FeFET. Our results suggest up to ~100x potential improvement in energy consumption for online learning over other technologies. It is important to mention that our analysis currently does not involve the cooling costs, which are difficult to estimate. Today, cooling costs are no longer specific to low temperature computing systems such as quantum computers or cryogenic CMOS. Data centers invest more than a third of their power budget on cooling costs, the aim of which is simply to prevent servers from shutting down induced by overheating (Saligram et al., 2024). Further work is needed for an accurate assessment of cooling costs for all new technologies. In addition, there is an increasing number of application areas for specialized low temperature computing, including cryogenic CMOS (Saligram et al., 2024) and quantum computing (Riel, 2021). Hardware accelerators based on fluxon synaptic arrays can be operated at liquid nitrogen temperature (~77 K) (Murduck, 2001). Parallel research and development of quantum computing and cryogenic CMOS may result in cheaper cooling, and it may provide a new and unique application for the superconducting loop devices for high performance AI applications.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

AK: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Visualization, Writing – original draft, Writing – review & editing. UG: Conceptualization, Data curation, Investigation, Methodology, Writing – review & editing. EC: Conceptualization, Writing – review & editing. RD: Conceptualization, Funding acquisition, Supervision, Writing – review & editing. DK: Conceptualization, Funding acquisition, Supervision, Visualization, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research was primarily supported as part of the “Quantum Materials for Energy Efficient Neuromorphic Computing” (Q-MEEN-C), an Energy Frontier Research Center funded by the U.S. Department of Energy, Office of Science, Basic Energy Sciences under Award No. DE-SC0019273 (project conception, device fabrication, experimental measurements) and by the Office of Naval Research (N00014-24-1-2127) (simulations, computations, analysis).

Acknowledgments

The authors thank Prof. Shane Cybart for fruitful discussions.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnins.2025.1511371/full#supplementary-material

References

Boahen, K. (2022). Dendrocentric learning for synthetic intelligence. Nature 612, 43–50. doi: 10.1038/s41586-022-05340-6

Crossref Full Text | Google Scholar

Chen, P.-Y., Peng, X., and Yu, S. (2018). Neuro Sim: a circuit-level macro model for benchmarking neuro-inspired architectures in online learning. IEEE Trans. Comput. Aided Design Integr. Circuits Syst. 37, 3067–3080. doi: 10.1109/TCAD.2018.2789723

Crossref Full Text | Google Scholar

Choi, S., Tan, S. H., Li, Z., Kim, Y., Choi, C., Chen, P.-Y., et al. (2018). SiGe epitaxial memory for neuromorphic computing with reproducible high performance based on engineered dislocations. Nat. Mater. 17, 335–340. doi: 10.1038/s41563-017-0001-5

Crossref Full Text | Google Scholar

Dutta, S., Khanna, A., Ye, H., Sharifi, M., Kazemi, A., San Jose, M., et al. (2021). Lifelong learning with monolithic 3D ferroelectric ternary content-addressable memory. 2021 IEEE international Electron devices meeting (IEDM).

Google Scholar

Dutta, S., Schafer, C., Gomez, J., Ni, K., Joshi, S., and Datta, S. (2020). Supervised learning in all FeFET-based spiking neural network: opportunities and challenges. Front. Neurosci. 14:634. doi: 10.3389/fnins.2020.00634

Crossref Full Text | Google Scholar

Duzer, T. V. (1989). Superconductor electronic device application. IEEE J. Quantum Electron. 25:11.

Google Scholar

Fulton, T., Dynes, R., and Anderson, P. (1973). The flux shuttle—a Josephson junction shift register employing single flux quanta. Proc. IEEE 61, 28–35. doi: 10.1109/PROC.1973.8966

Crossref Full Text | Google Scholar

Ge, R., Wu, X., Kim, M., Shi, J., Sonde, S., Tao, L., et al. (2018). Atomristor: nonvolatile resistance switching in atomic sheets of transition metal dichalcogenides. Nano Lett. 18, 434–441. doi: 10.1021/acs.nanolett.7b04342

Crossref Full Text | Google Scholar

Gokmen, T., and Vlasov, Y. (2016). Acceleration of deep neural network training with resistive cross-point devices: design considerations. Front. Neurosci. 10:333. doi: 10.3389/fnins.2016.00333

Crossref Full Text | Google Scholar

Goteti, U. S., Cai, H., LeFebvre, J. C., Cybart, S. A., and Dynes, R. C. (2022). Superconducting disordered neural networks for neuromorphic processing with fluxons. Sci. Adv. 8:eabn4485. doi: 10.1126/sciadv.abn4485

Crossref Full Text | Google Scholar

Goteti, U. S., Cybart, S. A., and Dynes, R. C. (2024). Collective neural network behavior in a dynamically driven disordered system of superconducting loops. Proc. Natl. Acad. Sci. 121:e2314995121. doi: 10.1073/pnas.2314995121

Crossref Full Text | Google Scholar

Goteti, U. S., and Dynes, R. C. (2021). Superconducting neural networks with disordered Josephson junction array synaptic networks and leaky integrate-and-fire loop neurons. J. Appl. Phys. 129:97. doi: 10.1063/5.0027997

Crossref Full Text | Google Scholar

Gupta, A., Ramanath, R., Shi, J., and Keerthi, S. S. (2021). Adam vs. sgd: Closing the generalization gap on image classification. OPT2021: 13th annual workshop on optimization for machine learning.

Google Scholar

Heaton, J. (2018). Ian goodfellow, yoshua bengio, and aaron courville: deep learning: the mit press, 2016, 800 pp, isbn: 0262035618. Genet. Program Evolvable Mach. 19, 305–307. doi: 10.1007/s10710-017-9314-z

Crossref Full Text | Google Scholar

Hopfield, J. J. (1982). Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. 79, 2554–2558. doi: 10.1073/pnas.79.8.2554

Crossref Full Text | Google Scholar

Hu, M., Graves, C. E., Li, C., Li, Y., Ge, N., Montgomery, E., et al. (2018). Memristor-based analog computation and neural network classification with a dot product engine. Adv. Mater. 30:1705914. doi: 10.1002/adma.201705914

Crossref Full Text | Google Scholar

Jerry, M., Chen, P.-Y., Zhang, J., Sharma, P., Ni, K., Yu, S., et al. (2017). Ferroelectric FET analog synapse for acceleration of deep neural network training. 2017 IEEE international electron devices meeting (IEDM),

Google Scholar

Jo, S. H., Chang, T., Ebong, I., Bhadviya, B. B., Mazumder, P., and Lu, W. (2010). Nanoscale memristor device as synapse in neuromorphic systems. Nano Lett. 10, 1297–1301. doi: 10.1021/nl904092h

Crossref Full Text | Google Scholar

Jouppi, N. P., Young, C., Patil, N., Patterson, D., Agrawal, G., Bajwa, R., et al. (2017). In-datacenter performance analysis of a tensor processing unit. Proceedings of the 44th annual international symposium on computer architecture.

Google Scholar

Jué, E., Pufall, M. R., Haygood, I. W., Rippard, W. H., and Schneider, M. L. (2022). Perspectives on nanoclustered magnetic Josephson junctions as artificial synapses. Appl. Phys. Lett. 121:287. doi: 10.1063/5.0118287

Crossref Full Text | Google Scholar

Kumar, A., and Bezugam, S. S. (2024). RRAM based processing-in-memory for efficient intelligent vision tasks at the edge. Mem. Mater. Devic. Circuits Syst. 8:100115. doi: 10.1016/j.memori.2024.100115

Crossref Full Text | Google Scholar

Kuzum, D., Jeyasingh, R. G., Lee, B., and Wong, H.-S. P. (2012). Nanoelectronic programmable synapses based on phase change materials for brain-inspired computing. Nano Lett. 12, 2179–2186. doi: 10.1021/nl201040y

Crossref Full Text | Google Scholar

LeCun, Y. (1998). The MNIST database of handwritten digits. Available at: http://yann.lecun.com/exdb/mnist/ (Accessed July 25, 2024).

Google Scholar

LeCun, Y., Bengio, Y., and Hinton, G. (2015). Deep learning. Nature 521, 436–444. doi: 10.1038/nature14539

Crossref Full Text | Google Scholar

Li, C., Hu, M., Li, Y., Jiang, H., Ge, N., Montgomery, E., et al. (2018). Analogue signal and image processing with large memristor crossbars. Nat. Electr. 1, 52–59. doi: 10.1038/s41928-017-0002-z

Crossref Full Text | Google Scholar

Likharev, K. K., and Semenov, V. K. (1991). RSFQ logic/memory family: a new Josephson-junction technology for sub-terahertz-clock-frequency digital systems. IEEE Trans. Appl. Supercond. 1, 3–28. doi: 10.1109/77.80745

Crossref Full Text | Google Scholar

Little, W. A. (1974). The existence of persistent states in the brain. Math. Biosci. 19, 101–120. doi: 10.1016/0025-5564(74)90031-5

Crossref Full Text | Google Scholar

Murduck, J. M. (2001). Fabrication of superconducting devices and circuits. Thin Films 28, 271–317. doi: 10.1016/S1079-4050(01)80022-3

Crossref Full Text | Google Scholar

Niazi, S., Chowdhury, S., Aadit, N. A., Mohseni, M., Qin, Y., and Camsari, K. Y. (2024). Training deep Boltzmann networks with sparse Ising machines. Nat. Electr. 7, 610–619. doi: 10.1038/s41928-024-01182-4

Crossref Full Text | Google Scholar

Oh, S., Shi, Y., Del Valle, J., Salev, P., Lu, Y., Huang, Z., et al. (2021). Energy-efficient Mott activation neuron for full-hardware implementation of neural networks. Nat. Nanotechnol. 16, 680–687. doi: 10.1038/s41565-021-00874-8

Crossref Full Text | Google Scholar

Oh, S., Shi, Y., Liu, X., Song, J., and Kuzum, D. (2018). Drift-enhanced unsupervised learning of handwritten digits in spiking neural network with PCM synapses. IEEE Electr. Device Lett. 39, 1768–1771. doi: 10.1109/LED.2018.2872434

Crossref Full Text | Google Scholar

O'Kelly, M., Zheng, H., Karthik, D., and Mangharam, R. (2020). F1tenth: an open-source evaluation environment for continuous control and reinforcement learning. Proceedings of Machine Learning Research.

Google Scholar

Park, J., Kumar, A., Zhou, Y., Oh, S., Kim, J.-H., Shi, Y., et al. (2024). Multi-level, forming and filament free, bulk switching trilayer RRAM for neuromorphic computing at the edge. Nat. Commun. 15:3492. doi: 10.1038/s41467-024-46682-1

Crossref Full Text | Google Scholar

Rafiq, M., Kaur, T., Gaidhane, A., Chauhan, Y. S., and Sahay, S. (2023). Ferroelectric FET-based time-mode multiply-accumulate accelerator: design and analysis. IEEE Trans. Electr. Devic. 70, 6613–6621. doi: 10.1109/TED.2023.3323261

Crossref Full Text | Google Scholar

Riel, H. (2021). Quantum computing technology. 2021 IEEE international Electron devices meeting (IEDM).

Google Scholar

Ruder, S. (2016). Optimizing gradient descent. Available at: https://www.ruder.io/optimizing-gradient-descent (Accessed July 25, 2024).

Google Scholar

Russek, S. E., Donnelly, C. A., Schneider, M. L., Baek, B., and Pufall, M. R. (2016). “Stochastic single flux quantum neuromorphic computing using magnetically tunable Josephson junctions” in IEEE international conference on rebooting computing (ICRC) (San Diego, CA, USA), 1–5.

Google Scholar

Saligram, R., Raychowdhury, A., and Datta, S. (2024). The future is frozen: cryogenic CMOS for high-performance computing. Chip 3:100082. doi: 10.1016/j.chip.2023.100082

Crossref Full Text | Google Scholar

Schmidhuber, J. (2015). Deep learning in neural networks: an overview. Neural Netw. 61, 85–117. doi: 10.1016/j.neunet.2014.09.003

Crossref Full Text | Google Scholar

Schneider, M., Toomey, E., Rowlands, G., Shainline, J., Tschirhart, P., and Segall, K. (2022). SuperMind: a survey of the potential of superconducting electronics for neuromorphic computing. Supercond. Sci. Technol. 35:053001. doi: 10.1088/1361-6668/ac4cd2

Crossref Full Text | Google Scholar

Schroen, W. (1968). Physics of preparation of Josephson barriers. J. Appl. Phys. 39, 2671–2678. doi: 10.1063/1.1656651

Crossref Full Text | Google Scholar

Selcuk, K., Kanai, S., Ota, R., Ohno, H., Fukami, S., and Camsari, K. Y. (2024). Double-free-layer stochastic magnetic tunnel junctions with synthetic antiferromagnets. Phys. Rev. Appl. 21:054002. doi: 10.1103/PhysRevApplied.21.054002

Crossref Full Text | Google Scholar

Seo, S., Kang, B.-S., Lee, J.-J., Ryu, H.-J., Kim, S., Kim, H., et al. (2020). Artificial van der Waals hybrid synapse and its application to acoustic pattern recognition. Nat. Commun. 11:3936. doi: 10.1038/s41467-020-17849-3

Crossref Full Text | Google Scholar

Shi, Y., Nguyen, L., Oh, S., Liu, X., Koushan, F., Jameson, J. R., et al. (2018). Neuroinspired unsupervised learning and pruning with subquantum CBRAM arrays. Nat. Commun. 9:5312. doi: 10.1038/s41467-018-07682-0

Crossref Full Text | Google Scholar

Vamathevan, J., Clark, D., Czodrowski, P., Dunham, I., Ferran, E., Lee, G., et al. (2019). Applications of machine learning in drug discovery and development. Nat. Rev. Drug Discov. 18, 463–477. doi: 10.1038/s41573-019-0024-5

Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., et al. (2017). Attention is all you need. Adv. Neural Inform. Process. Syst. 30, 6000–6010. doi: 10.5555/3295222.3295349

Crossref Full Text | Google Scholar

Wolf, P. (1978). “Application of the Josephson effect for digital storage” in Cyclecy Digital memory and storage. ed. W. E. Proebster. Springer-Verlag.

Google Scholar

Wu, W., Wu, H., Gao, B., Yao, P., Zhang, X., et al. (2018). A methodology to improve linearity of analog RRAM for neuromorphic computing. 2018 IEEE symposium on VLSI technology.

Google Scholar

Zhou, P., Feng, J., Ma, C., Xiong, C., and Hoi, S. C. H. (2020). Towards theoretically understanding why sgd generalizes better than Adam in deep learning. Adv. Neural Inf. Proces. Syst. 33, 21285–21296. doi: 10.1109/VLSIT.2018.8510690

Crossref Full Text | Google Scholar

Zmpe, H. H. (1980). Memory-cell Design in Josephson Technology. IEEE Transaction on Electron Devices 27, 1870–1882. doi: 10.1109/T-ED.1980.20124

Crossref Full Text | Google Scholar

Keywords: neuromorphic computing, superconducting loops, Josephson junctions, deep learning, image classification, energy efficient hardware

Citation: Kumar A, Goteti US, Cubukcu E, Dynes RC and Kuzum D (2025) Evaluation of fluxon synapse device based on superconducting loops for energy efficient neuromorphic computing. Front. Neurosci. 19:1511371. doi: 10.3389/fnins.2025.1511371

Received: 14 October 2024; Accepted: 27 January 2025;
Published: 14 February 2025.

Edited by:

Dashan Shang, Chinese Academy of Sciences (CAS), China

Reviewed by:

Xumeng Zhang, Fudan University, China
Zuheng Wu, Anhui University, China
Teng Zhang, Peking University, China

Copyright © 2025 Kumar, Goteti, Cubukcu, Dynes and Kuzum. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Duygu Kuzum, ZGt1enVtQHVjc2QuZWR1

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

Man ultramarathon runner in the mountains he trains at sunset

94% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more