Abstract
Spiking neural networks (SNNs) promise to bridge the gap between artificial neural networks (ANNs) and biological neural networks (BNNs) by exploiting biologically plausible neurons that offer faster inference, lower energy expenditure, and event-driven information processing capabilities. However, implementation of SNNs in future neuromorphic hardware requires hardware encoders analogous to the sensory neurons, which convert external/internal stimulus into spike trains based on specific neural algorithm along with inherent stochasticity. Unfortunately, conventional solid-state transducers are inadequate for this purpose necessitating the development of neural encoders to serve the growing need of neuromorphic computing. Here, we demonstrate a biomimetic device based on a dual gated MoS2 field effect transistor (FET) capable of encoding analog signals into stochastic spike trains following various neural encoding algorithms such as rate-based encoding, spike timing-based encoding, and spike count-based encoding. Two important aspects of neural encoding, namely, dynamic range and encoding precision are also captured in our demonstration. Furthermore, the encoding energy was found to be as frugal as ≈1–5 pJ/spike. Finally, we show fast (≈200 timesteps) encoding of the MNIST data set using our biomimetic device followed by more than 91% accurate inference using a trained SNN.
Similar content being viewed by others
Introduction
Billions of neurons coupled through trillions of synapses form the complex computational unit of the brain, that can simultaneously process a massive amount of information received from external/internal stimuli through various sensory organs. These neurons use action potential or spikes as the common language to speak with each other and to compute for learning and decision-making. Spikes are stereotypical electrical impulses or all-or-none (digital) point events in time, that allow long-distance neural communication and energy-efficient neural computation. However, external stimuli such as light, sound, smell, temperature, etc., and internal stimuli such as blood pressure, oxygen levels, feeling of pain and hunger, etc. that are primarily analog continuous variables in time. It requires specialized sensory neurons, also known as afferent neurons, to transform the specific type of analog stimulus into corresponding spike trains following one or more neural encoding algorithms, and subsequently relay the spike-encoded information to the central nervous system for processing. This process is referred to as sensory transduction. For example, in the auditory neural pathways, mechanoelectrical transduction is mediated by the hair cells within the ear1. Similarly, gustatory receptors in taste buds interact with chemicals in food to produce action potentials2,3. Phototransduction in the retina is mediated by rods and cones and eventually converted to spikes by the ganglion cells4,5. Sensory transduction also exhibits inherent stochasticity, which allows neurons to process information with better noise tolerance and energy efficiency6,7.
The diversity of neurobiological architectures and neural computational algorithms found inside even the simplest of animal brains continue to fascinate computer scientists and electronic device engineers. Neuromorphic computing pioneered by Carver Mead and colleagues is a branch of research that aims to mimic the computational power of the brain on a chip8,9. Unfortunately, the initial growth in neuromorphic computing was rather slow owing to the contemporary dominance of von Neuman architecture, and the success of the complementary metal-oxide-semiconductor (CMOS) technology. However, the recent demise in scaling and fundamental limitations of von Neuman computing is fueling the resurgence of bio-inspired neuromorphic hardware10,11,12. Artificial neural networks (ANNs) are the most prevalent form of neuromorphic computing that have already demonstrated breakthrough progress in many fields13. ANNs consist of multiple layers with each layer comprising of collection of computational units, called artificial neurons, which are connected through artificial synapses. While the models of cortical hierarchies from biological neural networks (BNNs) have been mimicked through deep learning14 in ANNs with a massive number of computational layers, only marginal similarities with brain-like computing can be recognized at the implementation level. The most obvious difference is that artificial neurons receive, process, and transmit analog information in continuous time, whereas biological neurons use action potential or spikes. Also, stochasticity is an inherent neural phenomenon, which is typically ignored by most ANNs. Spiking neural network (SNN) promises to bridge this gap by adopting a new computing paradigm based on biologically plausible neurons15,16. In fact, the past few years have seen tremendous progress in the development of SNNs offering unprecedented energy efficiency and faster inference owing to event-driven computation17. However, hardware realization of SNNs necessitates the development of neural encoders since conventional sensors are incapable of converting sensory input into spike trains.
Here, we report a biomimetic device based on a dual gated MoS2 field effect transistor (FET) with a stochastic sampling terminal capable of encoding analog signals, for example illuminance levels of a light emitting diode (LED), into corresponding spike trains. We are also able to implement various neural encoding algorithms, such as rate-based encoding, spike timing-based encoding, and spike count-based encoding. Two key features of neural encoding, namely, dynamic range and encoding precision are also captured in our demonstration. Finally, we show a fast and accurate inference of spike encoded MNIST data set using a trained spiking neural network (SNN) with inference accuracy of more than 91%. Remarkably, energy consumption by our biomimetic neural encoder was found to be as frugal as ≈1–5 pJ/spike.
Results
The overall philosophy of biomimetic neuromorphic computing is shown in Fig. 1. Figure 1a shows the schematic of a biological neural network that involves sensory transduction of analog stimulus to corresponding spike trains by specialized neurons and subsequent processing by the central nervous system. For example, external optical stimuli are converted into corresponding graded potentials by the photoreceptor cells (rods and cones) in the human eyes followed by neural encoding into spike trains using the ganglion cells, and eventual processing of the encoded visual stimulus by the visual cortex. Figure 1b shows the corresponding neuromorphic hardware comprising of neuromorphic sensors, neuromorphic encoders, and neuromorphic processors. Figure 1c shows the schematic of our experimental demonstration with a white light-emitting diode (LED) as the visual stimulus, a silicon (Si) photodiode (PD) as the sensor, a dual gated MoS2-based FET as the neuromorphic encoder, and a trained spiking neural network (SNN) as the neuromorphic processor.
Biomimetic neural encoder and neuromorphic transducer
We have used multilayer exfoliated MoS2 that belongs to the family of two-dimensional (2D) layered materials18,19,20 as the semiconducting channel, 285 nm SiO2 on p++-Si as the back-gate stack, and 120 nm of hydrogen silsesquioxane (HSQ) as the top-gate dielectric for the fabrication of the biomimetic neural encoder as shown schematically in Fig. 2a. The thickness of the MoS2 flake is ≈5 nm. The optical image of the device is shown in Fig. 2b. Source, drain, and gate metal stacks were patterned using electron-beam lithography followed by the deposition of 40 nm of Nickel (Ni) and 30 nm of gold (Au) using electron-beam evaporation. More details on device fabrication can be found in the “Methods” section and in our prior work21,22. The channel length and width were ≈1 μm and ≈2.8 μm, respectively. Note that the use of ultrathin body MoS2 as the semiconducting channel material is motivated by the growing interest in 2D materials, as a successor to Si as well as their promising use in neuromorphic and biomimetic devices12,22,23,24,25. Furthermore, various types of sensors such as photodetectors26, chemical sensors27, biological sensors27, touch sensors28, and radiation sensors29 have been demonstrated using MoS2 based devices, which allow direct integration of sensors and encoders in future neuromorphic hardware. The presynaptic signal obtained from the neuromorphic sensors such as the Si PD is applied to the back-gate terminal as analog voltage (VPSV), whereas, encoded information in form of postsynaptic current spikes (IPSC) is obtained at the drain terminal. The top-gate voltage (VTG) is applied as a sequence of sampling pulses, with a pulse duration (tp) of 10 ms and amplitude determined based on the desired encoding algorithm.
Figure 2c shows the transfer function of the neural encoder i.e., IPSC vs. VPSV measured at a drain bias, VDS = 1 V, for different VTG. The n-type unipolar characteristics is common for MoS2 FETs30,31. The monotonic positive shift in the transfer function with decreasing VTG can be explained form the principle of charge balance, i.e., the inversion charge induced by positive VPSV is compensated by the negative VTG and vice versa. Figure 2d shows the spiking threshold (VST), which we define as VPSV required to invoke a current spike, i.e., IPSC > IST as a function of VTG, and for different thresholding current, IST. As expected, the spiking threshold is higher for more negative VTG and higher IST. Note that the slope of VST vs. VTG is constant, irrespective of IST and is proportional to the ratio of back-gate capacitance to top-gate capacitance, i.e., CTG/CBG which was found to be ≈2.2, consistent with the thicknesses of ≈120 nm and ≈285 nm and dielectric constants of ≈3.2 and ≈3.9 of HSQ and SiO2, respectively. Also, note that the presynaptic terminal (back-gate) and encoding terminal (top-gate) can be interchanged (see Supplementary Fig. 1 and Supplementary Note 1). Figure 2e shows the circuit schematic for phototransduction comprising of a Si PD and a load resistor (RL). Figure 2f shows the current (IPD) vs. voltage (VPD) characteristics of the neuromorphic sensor, i.e. Si PD for different intensities of the visual stimulus, i.e., LED illuminance (PLED). Finally, Fig. 2g shows the phototransduction characteristics of the Si PD and corresponding VPSV applied to the neuromorphic encoder as a function of PLED obtained using RL.
Hardware acceleration of various neural encoding algorithms
Next, we implement various neural encoding algorithms, found in sensory neurobiology, using our biomimetic encoder for translating analog VPSV values obtained for different LED illuminations shown in Fig. 3a into corresponding spike trains. The most popular encoding principle is rate-based encoding, originally demonstrated by Adrian and Zotterman32 using an electrophysiological experiment in sensory nerve fibers of frog muscles. In rate encoding, it is postulated that the information about the stimulus is contained in the firing rate of the neuron, and not in individual spikes. This is more so because the sequence of spikes generated by the neurons in response to a given stimulus varies from trial to trial and over time owing to the inherent stochasticity in sensory transduction, whereas the mean firing rate, i.e., inverse of interspike interval remains practically constant. Numerous studies in sensory and motor systems of various species have validated the spike rate encoding hypothesis. Based on these observations, rate encoding is widely used for SNNs. For rate-based encoding using our biomimetic encoder, the magnitude of VTG pulses are randomly sampled from a Gaussian distribution with mean, μTG = −2.5 V, and standard deviation σTG = 0.8 V as shown in Fig. 3b and the responses corresponding to each VPSV value with IST = 500 pA is displayed in Fig. 3c (see “Methods” section for discussion on current-sampling method). During each trial, VTG pulses were sampled for N number of times (=32) and a total of 16 trials were conducted resulting in 512 sampling points for each VPSV. See Supplementary Fig. 2 for the complete circuit used to obtain neural encoding for different illumination levels. Also, see Supplementary Movie 1 for real-time encoding of different LED intensities into stochastic spike trains. Figure 3d shows the encoding transfer function, i.e., mean firing rate (inverse of the mean interspike interval) as a function of VPSV (see Supplementary Fig. 3 for distribution of interspike interval). Clearly, the firing rate increases monotonically with increasing stimulus intensity, indicating that our biomimetic encoder is capable of rate-based encoding. Finally, Fig. 3e shows the encoding energy per spike (Een) for rate-based encoding, computed based on Eq. (1). The monotonic increase in Een with increasing VPSV is consistent with increasing firing rate, i.e., more spiking in the postsynaptic neuron.
Typical energy consumption is around 1–5 pJ/spike. Note that the second term in Eq. (1) dominates in our demonstration since the first term contributes ≈100 fJ. Therefore, one obvious way to reduce the power consumption is through VDS scaling. Note that the neural encoder exploits subthreshold device characteristics and does not impose any requirement on the current device. Hence it is possible to operate the neural encoder with ultra-low VDS. Another alternative to reduce the power dissipation is to increase the sampling rate i.e., reduce tp. However, beyond a certain point, the first term will start to dominate, which can be scaled by scaling the oxide thickness to achieve encoding at scaled VTG values. Note that oxide thickness scaling increases CTG, but the square term involving VTG will determine the energy scaling.
Another encoding principle found in sensory neurobiology is spike count-based encoding. For example, rats show remarkable texture discriminations using their facial whiskers. It is found that the trigeminal ganglion cells that innerte the sensory receptor from each whisker use spike count to distinguish the stimuli33. Similar spike count encoding is observed for frequency discrimination of vibrotactile stimuli in the primary somatosensory cortex of trained monkeys34. Figure 3f shows the VTG pulse profile used to achieve spike count-based encoding using our biomimetic encoder. In this case, the magnitude of VTG pulses increases with added zero mean Gaussian noise of standard deviation σTG = 0.2 V. Each trial consists of 32 pulses, and 16 trials were recorded for each VPSV. The corresponding responses of the neural encoder are displayed in Fig. 3g. Figure 3h shows the encoding transfer function i.e., the mean spike count as a function of VPSV (see Supplementary Fig. 4 for total spike counts for all 16 trials for different σTG). Clearly, the mean spike count increases monotonically with increasing stimulus intensity, indicating that our biomimetic encoder is capable of spike count-based encoding. Note that, the implementation of spike count-based encoding does not necessarily require stochasticity, i.e., similar results could be obtained using σTG = 0 V. However, in the context of SNN, stochasticity can aid as hardware realization of integrate and fire (IF) neuron can be challenging. A more realistic neuron is leaky integrate and fire (LIF) neuron, where random spiking can compensate for the loss in information due to capacitive discharging between spikes. Figure 3i shows Een for spike count-based encoding, which shows a monotonic increase with VPSV since the spike count increases accordingly. The energy expenditure was found to be ≈1–3.5 pJ/spike.
Whereas, rate-based encoding and spike count-based encoding are the most broadly accepted view of neural computation, these approaches ignore the information possibly contained in the exact timing of the spikes. In fact, recent studies suggest that a straightforward firing rate or spike count-based encoding may be too simplistic to describe brain activity in its entirety. For example, neurophysiological experiments show that visual neurons in rhesus monkeys can recognize faces within ≈80–160 ms35. Anatomically, it involves more than ten synaptic stages between the photoreceptors of the retina and visually responsive neurons in the temporal cortex implying that each layer has, on average, only 10 ms of processing time. Since the firing rates of cortical neurons are in the range 0–100 spikes per second, a neuron in any given layer can only generate one spike before neurons in the next layer have to respond. This puts severe constraints on the way information is encoded in visual pathways. Firing-rate or spike count-based encoding seems inappropriate and evidence suggests that analog information is encoded by the relative arrival times of spikes36,37,38,39. Such an encoding scheme also referred to as the spike timing-based encoding, not only allows very rapid information processing but also offers tremendous energy benefits for future SNNs. Figure 3j shows the VTG pulse profile used for achieving spike timing-based encoding using our biomimetic encoder. In this case, the magnitude of VTG pulses decreases over time with added zero-mean Gaussian noise of standard deviation σTG = 0.2 V. Each trial consists of 32 pulses for each VPSV. The corresponding responses of the neural encoder are displayed in Fig. 3k. A sense amplifier is used to sense the arrival of the first spike triggering the deactivation of the VTG pulse sampling for the rest of the trial. Figure 3l shows the encoding transfer function i.e., mean spike-timing as a function of VPSV (see Supplementary Fig. 5 for distribution of spike timing over multiple trials). Clearly, high-intensity stimuli invoke early spiking and vice versa indicating that our biomimetic encoder is capable of spike timing-based encoding as well. Note that, the implementation of spike timing-based encoding does not require stochasticity, i.e., similar results could be obtained by using σTG = 0 V. However, the flexibility of noise adjustment makes our neural encoder more bio-realistic. Figure 3m shows Een for spike timing-based encoding. Unlike rate-based and count-based encoding, timing-based encoding shows a monotonic decrease with increasing VPSV. This is owing to the fact the spiking occurs earlier for higher VPSV deactivating the encoder and minimizing the energy consumption per spike. The fact that the encoding energy can be significantly lower for timing-based encoding compared to rate-based or count-based encoding is appealing for ultra-low-power neuromorphic computing using SNN (See Supplementary Note 2 showing the comparison of our neural encoder with other types of spike encoders).
Finally, Fig. 3n–q, respectively, shows the original Cameraman image and the corresponding spike rate, spike count, and spike timing-based encoding. The pixel values ranging from 0 to 255 were mapped linearly to the VPSV range of 0–5 V (see the “Methods” section for details on image encoding). Clearly, the Cameraman image is accurately encoded, irrespective of the encoding algorithm. Note that the contrast of the image in Fig. 3q is reversed compared to the original image, which is expected for spike-time based encoding since the higher pixel values should spike earlier than the lower pixel values. Supplementary movie files 2 show the time evolution of encoded images over time for rate-based, count-based, and timing-based encoding. Supplementary Fig. 6 shows the time evolution of the correlation coefficient (CC) between the original image and the encoded image. The CC reaches ≈1 at the end of encoding for all three encoding algorithms.
Dynamic range and encoding precision for rate-based encoding
Now, we focus on two key aspects of neural encoding, namely, dynamic range and encoding precision. A high dynamic range (HDR) allows neurons to respond to more extreme stimuli. For example, photoreceptors in human eyes can identify objects in starlight as well as in bright sunlight despite of illumination levels differing by ≈9 orders of magnitude, i.e., over a dynamic range of 90 dB40. Similarly, the dynamic range of human hearing is roughly 140 dB41. However, HDR does not necessarily guarantee high precision (HP). For example, a whisper cannot be heard in loud surroundings. Similarly, eyes take time to adapt to different illumination levels. In fact, most sensory neurons adjust their spike encoding based on the environment42,43,44. Figure 4a–f shows how our biomimetic encoder achieves similar functionality by adjusting σTG and μTG of the Gaussian distribution used for sampling VTG as well as and the thresholding current, IST for the spike rate-based encoding algorithm presented earlier. For numerical simulations, we have used the virtual source (VS) model described elsewhere22,23. Clearly, HDR can be achieved by using higher values of σTG, whereas smaller values of σTG allow HP (Fig. 4a, b). This is because the encoding transfer function follows the cumulative probability distribution of a random Gaussian variable since for a given VPSV stimulus, there will always be a postsynaptic spike if the magnitude of the VTG pulse is more positive than the one corresponding to the spiking threshold, VST, as shown in Fig. 2c. For a higher value of σTG, the cumulative probability distribution follows a linear trend allowing a larger VPSV range to be encoded, whereas a lower value of σTG results in a non-linear cumulative probability distribution that restricts the encoding range but improves the encoding precision. However, both feats cannot be achieved at the same time by adjusting σTG. However, by adjusting μTG it is possible to achieve HP for different ranges of stimulus intensity similar to the sensory neurons (Fig. 4c, d). The spiking rate can also be tuned by adjusting the IST (Fig. 4e, f). Lower values of IST allow more spiking events for any given VPSV, whereas, higher values of IST restrict spiking even for higher VPSV. Figure 4g and h show the original and scaled Cameraman images, and Fig. 4i–l, respectively, show the corresponding spike rate-based linear and non-linear encoding using our biomimetic encoder. The original Cameraman image necessitates linear encoding since the pixel values have a large dynamic range, whereas, the scaled Cameraman image is better encoded using high precision non-linear encoding.
MNIST digit classification using our neural encoder device
Finally, we exploit our biomimetic encoder for encoding MNIST data set on digit-classification into spike trains and infer using a trained SNN. For training the SNN we have used an approach described by Sengupta et al.45. This approach overcomes the lower accuracy of unsupervised learning rules such as the spike-time dependent plasticity (STDP) used for training SNNs46,47,48,49,50. The lower accuracy is due to the lack of efficient algorithms to make use of the spiking neurons. To bridge this gap, ANN-SNN conversion schemes are used, where an ANN is trained using the traditional back-propagation algorithm, followed by the conversion of the ANN to SNN45,51,52. This approach yields higher inference accuracy owing to near-lossless ANN-SNN conversion45. Here, we train a fully connected two-layered artificial neural network with 100 neurons in the hidden layer and 10 neurons in the output layer for digit-classification using the MNIST dataset as shown in architecture in Fig. 5a. MNIST dataset with a size of 28 × 28 pixels is flattened to obtain 784 pixels, which is fed to the input layer. The ten output neurons correspond to digits from 0 to 9. During training, for every input image, the network is trained through gradient descent to ensure that the output matches the expected label. Here, the ANN is trained with a learning rate of 0.0001 to ensure high convergence accuracy. Further, the following restrictions are incorporated while training the ANN to allow smooth ANN-SNN transition: rectified linear unit (ReLU) is used as the activation function due to its functional equivalence to IF spiking neuron used in SNN, bias terms are eliminated to ensure a smaller parameter space which enables easier ANN-SNN conversion, and no regularization is used. Sixty thousand images from the MNIST data set were used to train the ANN to achieve a training accuracy of 91.5% over 100 epochs. Following this, a testing accuracy of 92.7% was achieved using the remaining 10,000 images.
As discussed, SNNs use binary spikes in time which are representative of the action potential in BNNs. This requires the conversion of the analog image pixel intensities to digital spike trains. To accomplish the analog to spike conversion using our biomimetic encoder, first, the pixel intensity values ranging from 0 to 255 are mapped onto VPSV range of 0–5 V. Next, we record IPSC, for VPSV values corresponding to each pixel over a time window, TW, by applying stochastic VTG pulses. Each IPSC value subsequently undergoes a thresholding function with IST to generate binary X in time. We adopt rate-based encoding by applying VTG pulses with the pulse magnitude determined using a random Gaussian distribution, as described earlier. Figure 5b shows examples of spike encoded digit using μTG = −5.5 V, and σTG = 1 for the Gaussian distribution for the VTG pulses, and IST = 200 pA. The resultant X is fed into the SNN, as shown in the input layer of Fig. 5a. For ANN-SNN conversion, ReLU activation functions are replaced by IF neuron as shown in Fig. 5c following Eq. (2).
Here, the IF neuron is represented as the function of timesteps (t). Vmean(t) is the membrane potential, and w denotes the weights obtained from the trained ANN. In the IF neuron when the membrane potential crosses a certain threshold (Vth), the neuron spikes, propagating spike to the next layer, and it resets back to its resting potential which is set as zero. To optimize the IF neuron threshold, threshold-balancing is used to set the threshold as the maximum neuron activation for the corresponding layer obtained by the dot product of the weights and spike-train at an instance t45. The SNN is used to classify the set of 10,000 test images. Figure 5d shows the inference error versus the number of timesteps. Increasing the number of timesteps is important to allow sufficient firing, to effectively encode the pixel intensities. But remarkably, even with 200 timesteps, we achieve a low error of 9.5%. This is further improved as the timesteps are increased with the minimum error of 8.6% at 300 timesteps. Hence a maximum accuracy of 91.4% is achieved when the SNN is simulated with our biomimetic neural encoder. Additionally, similar test accuracies are obtained from both ANN and SNN, indicating a successful ANN-SNN transformation with a minimal loss of 1.3%.
Finally, we explore the dependence of inference accuracy on the dynamic range and the firing rate, parameters that can be adjusted in our biomimetic spike encoder by adjusting σTG and IST, respectively. As shown in Fig. 5e, minimum error of ≈8.6% is achieved at σTG of 0.8 and 1, with higher errors for lower and higher σTG. As described earlier in Fig. 4a, b, for lower σTG, the dynamic range is low to capture the variation in pixel intensities, whereas for very high σTG, there is insufficient difference between t firing rates corresponding to different pixel intensities resulting in a large error. A similar non-monotonic trend is seen in inference error with respect to IST in Fig. 5f, with the minimum error of 8.6% obtained at 200 pA. For higher IST, the spiking is minimal resulting in an inadequate representation of image pixels, while for lower IST, any pixel intensity results in excessive firing as seen in Fig. 4e, f. Nevertheless, by optimizing these parameters it is possible to ensure efficient encoding of the MNIST images, and thereby achieve a maximum accuracy of 91.4%.
Discussion
In conclusion, we have developed a neural encoder based on a dual gated MoS2 FET with a stochastic sampling terminal capable of encoding analog signals into spike trains. We also implemented three encoding algorithms, namely, spike rate-based encoding, spike count-based encoding, and spike timing-based encoding found in sensory neurobiology. As a prototype demonstration, we show the direct conversion of analog light intensities to corresponding rate-based spike trains analogous to phototransduction mechanisms in visual pathways. We also show frugal encoding energy expenditure in the range of few pico Joules per spike. Our biomimetic encoder also allows flexibility in terms of adjusting the encoding range and encoding precision, two key features found in biological sensory transduction to enable seamless adaption to different environmental conditions. Finally, we encoded the MNIST data set for digit classification using our spike encoder and achieved an inference accuracy of 91.4% by using a trained SNN. In brief, our demonstration of the biomimetic neural encoder is a leap forward towards achieving energy-efficient and bio-realistic neuromorphic hardware.
Methods
Device fabrication
The dual-gated devices were fabricated using micromechanically exfoliated MoS2 flakes on 285 nm thermally grown SiO2 substrates with highly doped-Si as the back-gate electrode. The source/drain contacts were defined using electron-beam lithography (Vistec EBPG5200). Ni (40 nm) followed by Au (30 nm) were deposited using electron-beam evaporation for the contacts. For fabricating the top-gate, hydrogen silsesquioxane (HSQ) was used as the dielectric. It was deposited by spin coating 6% HSQ in methyl isobutyl ketone (MIBK) (Dow Corning XR-1541-006) at 4000 rpm for 45 s and baked at 80 °C for 4 min. The HSQ was patterned using an e-beam dose of 2000 µC/cm2 and was developed at room temperature using 25% tetramethylammonium hydroxide (TMAH) for 30 s following a 90 s rinse in deionized water (DI). Next, it was cured in the air at 180 °C and then 250 °C for 2 min and 3 min, respectively. The top-gate electrode was patterned using electron-beam lithography followed by the deposition of Ni/Au using electron-beam evaporation as the contact.
Device measurements
Electrical characterization was performed at room temperature in high vacuum (≈10–6 Torr) on a Lake Shore CRX-VF probe station and using a Keysight B1500A parameter analyzer. We observed none to minimal hysteresis in the device characteristics for both top-gate and back-gate sweeps indicating high quality of MoS2/SiO2 and MoS2/HSQ interfaces (See Supplementary Fig. 7). For current sampling, when a sampling delay is set, (for example, T = 10 ms) the tool determines a time width (W) for integration based on the current value i.e., lower current values need larger W and vice versa. If W < T the tool measures the current at any point within the delay period. Else if W > T, irrespective of the time delay, the tool measures the current every time width (W) it sets for the current value range. The 10 ms delay we chose is large enough for the current value ranges we are operating to integrate. So, the tool will measure at any point within the 10 ms delay.
Image encoding
Note that the gray scale pixel values in an 8-bit cameraman image range from 0–255, which are mapped to VPSV range of 0–5 V. This would require a VPSV precision of 5/255 = 0.0196 ≈ 0.02 V. However, our experimental VPSV step size was 0.5 V. Therefore, we quantized the cameraman image into 11 distinct levels and correlate those levels to VPSV = 0, 0.5, 1.0, 1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, and 5.0 V. Note that for spike-rate based encoding, our maximum frequency is 100 Hz and the average standard deviation for the encoded frequencies i.e., the encoding error bar is ≈10 Hz, allowing distinct encoding of 11 levels. One way to improve the mapping precision for spike-rate based encoding is to increase the maximum encoding frequency to 1000 Hz through faster sampling if the encoding error bar remains unaltered. Similarly, for spike-count based encoding, the maximum count is 16 and the average standard deviation for the encoded count or the encoding error bar is ≈1.5, again allowing the distinct encoding of 11 levels. Finally, for spike-timing based encoding, the maximum number of time-steps is 30 and the average standard deviation for the encoded time-step or the encoding error bar is ≈4, allowing the distinct encoding of eight levels. The mapping precision for spike-count and spike-timing based encoding can be increased by increasing the total number of sampling points.
Data availability
The datasets generated during and/or analyzed during the current study are available from the corresponding author on reasonable request.
Code availability
The codes used for plotting the data are available from the corresponding authors on reasonable request.
References
Rudnicki, M., Schoppe, O., Isik, M., Völk, F. & Hemmert, W. Modeling auditory coding: from sound to spikes. Cell Tissue Res. 361, 159–175 (2015).
Reiter, S., Rodriguez, C. C., Sun, K. & Stopfer, M. Spatiotemporal coding of individual chemicals by the gustatory system. J. Neurosci. 35, 12309–12321 (2015).
Hallock, R. M. & Di Lorenzo, P. M. Temporal coding in the gustatory system. Neurosci. Biobehav. Rev. 30, 1145–1160 (2006).
Meister, M., Lagnado, L. & Baylor, D. A. Concerted signaling by retinal ganglion cells. Science 270, 1207–1210 (1995).
Choi, S.-Y. et al. Encoding light intensity by the cone photoreceptor synapse. Neuron 48, 555–562 (2005).
McDonnell, M. D. & Ward, L. M. The benefits of noise in neural systems: bridging theory and experiment. Nat. Rev. Neurosci. 12, 415–425 (2011).
Jaramillo, F. & Wiesenfeld, K. Mechanoelectrical transduction assisted by Brownian motion: a role for noise in the auditory system. Nat. Neurosci. 1, 384–388 (1998).
Mead, C. Neuromorphic electronic systems. Proc. IEEE 78, 1629–1636 (1990).
Douglas, R., Mahowald, M. & Mead, C. Neuromorphic analogue VLSI. Annu. Rev. Neurosci. 18, 255–281 (1995).
Furber, S. Large-scale neuromorphic computing systems. J. Neural Eng. 13, 051001 (2016).
Boybat, I. et al. Neuromorphic computing with multi-memristive synapses. Nat. Commun. 9, 1–12 (2018).
Sangwan, V. K. & Hersam, M. C. Neuromorphic nanoelectronic materials. Nat. Nanotechnol. 15, 1–12 (2020).
Hassoun, M. H. Fundamentals of Artificial Neural Networks (MIT Press, 1995).
Schmidhuber, J. Deep learning in neural networks: an overview. Neural Netw. 61, 85–117 (2015).
Ghosh-Dastidar, S. & Adeli, H. Spiking neural networks. Int. J. Neural Syst. 19, 295–308 (2009).
Ponulak, F. & Kasinski, A. Introduction to spiking neural networks: Information processing, learning and applications. Acta Neurobiol. Exp. 71, 409–433 (2011).
Han, B., Sengupta, A. & Roy, K. On the energy benefits of spiking deep neural networks: a case study. International Joint Conference on Neural Networks (IJCNN), 971–976 Vancouver, Canada (2016).
Butler, S. Z. et al. Progress, challenges, and opportunities in two-dimensional materials beyond graphene. ACS Nano 7, 2898–2926 (2013).
Bhimanapati, G. R. et al. Recent advances in two-dimensional materials beyond graphene. ACS Nano 9, 11509–11539 (2015).
Das, S., Robinson, J. A., Dubey, M., Terrones, H. & Terrones, M. Beyond graphene: progress in novel two-dimensional materials and van der Waals solids. Annu. Rev. Mater. Res. 45, 1–27 (2015).
Nasr, J. R. & Das, S. Seamless fabrication and threshold engineering in monolayer MoS2 dual-gated transistors via hydrogen silsesquioxane. Adv. Electron. Mater. 5, 1800888 (2019).
Das, S., Dodda, A. & Das, S. A biomimetic 2D transistor for audiomorphic computing. Nat. Commun. 10, 3450 (2019).
Sebastian, A., Pannone, A., Radhakrishnan, S. S. & Das, S. Gaussian synapses for probabilistic neural networks. Nat. Commun. 10, 1–11 (2019).
Dodda, A. et al. Stochastic resonance in MoS2 photodetector. Nat. Commun. 11, 4406 (2020).
Schranghamer, T. F., Oberoi, A. & Das, S. Graphene memristive synapses for high precision neuromorphic computing. Nat. Commun. 11, 5474 (2020).
Yin, Z. et al. Single-layer MoS2 phototransistors. ACS Nano 6, 74–80 (2012).
Wang, L. et al. Functionalized MoS2 nanosheet‐based field‐effect biosensor for label‐free sensitive detection of cancer marker proteins in solution. Small 10, 1101–1105 (2014).
Park, M. et al. MoS2‐based tactile sensor for electronic skin applications. Adv. Mater. 28, 2556–2562 (2016).
Arnold, A. J., Shi, T., Jovanovic, I. & Das, S. Extraordinary radiation hardness of atomically thin MoS2. ACS Appl. Mater. Interfaces 11, 8391–8399 (2019).
Das, S., Chen, H. Y., Penumatcha, A. V. & Appenzeller, J. High performance multilayer MoS2 transistors with scandium contacts. Nano Lett. 13, 100–105 (2013).
Schulman, D. S., Arnold, A. J. & Das, S. Contact engineering for 2D materials and devices. Chem. Soc. Rev. 47, 3037–3058 (2018).
Adrian, E. D. & Zotterman, Y. The impulses produced by sensory nerve-endings: Part II. The response of a single end-organ. J. Physiol. 61, 151 (1926).
Arabzadeh, E., Panzeri, S. & Diamond, M. E. Deciphering the spike train of a sensory neuron: counts and temporal patterns in the rat whisker pathway. J. Neurosci. 26, 9216–9226 (2006).
Luna, R., Hernández, A., Brody, C. D. & Romo, R. Neural codes for perceptual discrimination in primary somatosensory cortex. Nat. Neurosci. 8, 1210–1219 (2005).
Perrett, D. I., Rolls, E. T. & Caan, W. Visual neurones responsive to faces in the monkey temporal cortex. Exp. Brain Res. 47, 329–342 (1982).
Thorpe, S. J. Spike arrival times: A highly efficient coding scheme for neural networks. Parallel Process. Neural Syst. 91–94 (1990).
Butts, D. A. et al. Temporal precision in the neural code and the timescales of natural vision. Nature 449, 92–95 (2007).
Chase, S. M. & Young, E. D. Spike-timing codes enhance the representation of multiple simultaneous sound-localization cues in the inferior colliculus. J. Neurosci. 26, 3889–3898 (2006).
Reich, D. S., Mechler, F. & Victor, J. D. Temporal coding of contrast in primary visual cortex: when, what, and why. J. Neurophysiol. 85, 1039–1050 (2001).
Zaghloul, K. A., Boahen, K. & Demb, J. B. Contrast adaptation in subthreshold and spiking responses of mammalian Y-type retinal ganglion cells. J. Neurosci. 25, 860–868 (2005).
Zeng, F.-G., Fu, Q.-J. & Morse, R. Human hearing enhanced by noise. Brain Res. 869, 251–255 (2000).
Wen, B., Wang, G. I., Dean, I. & Delgutte, B. Dynamic range adaptation to sound level statistics in the auditory nerve. J. Neurosci. 29, 13797–13808 (2009).
Viana, R. et al. Dynamic range in a neuron network with electrical and chemical synapses. Commun. Nonlinear Sci. Numer. Simul. 19, 164–172 (2014).
Wachowiak, M. & Cohen, L. B. Representation of odorants by receptor neuron input to the mouse olfactory bulb. Neuron 32, 723–735 (2001).
Sengupta, A., Ye, Y., Wang, R., Liu, C. & Roy, K. Going deeper in spiking neural networks: VGG and residual architectures. Front. Neurosci. 13, 95 (2019).
Diehl, P. U. & Cook, M. Unsupervised learning of digit recognition using spike-timing-dependent plasticity. Front. Comput. Neurosci. 9, 99 (2015).
Zhang, W. & Li, P. Spike-train level backpropagation for training deep recurrent spiking neural networks. https://arxiv.org/1908.06378 (2019).
Brader, J. M., Senn, W. & Fusi, S. Learning real-world stimuli in a neural network with spike-driven synaptic dynamics. Neural Comput. 19, 2881–2912 (2007).
Querlioz, D., Bichler, O., Dollfus, P. & Gamrat, C. Immunity to device variations in a spiking neural network with memristive nanodevices. IEEE Trans. Nanotechnol. 12, 288–295 (2013).
Pfeiffer, M. & Pfeil, T. Deep learning with spiking neurons: opportunities and challenges. Front. Neurosci. 12, 774 (2018).
Diehl, P. U. et al. Fast-classifying, high-accuracy spiking deep networks through weight and threshold balancing. International Joint Conference on Neural Networks (IJCNN), 1–8 Killarney, Ireland (2015).
Cao, Y., Chen, Y. & Khosla, D. Spiking deep convolutional neural networks for energy-efficient object recognition. Int. J. Comput. Vis. 113, 54–66 (2014).
Acknowledgements
The work was supported by Army Research Office (ARO) through Contract Number W911NF1920338. Figure 1a was designed using resources from Freepik.com “image: Freepik.com”.
Author information
Authors and Affiliations
Contributions
Saptarshi Das and S.S. conceived the idea and designed the experiments. Saptarshi Das wrote the paper. Sarbashis Das fabricated the devices. A.S. performed SNN simulations. Saptarshi Das, S.S., A.S., A.O., and Sarbashis Das analyzed the data, discussed the results, agreed on their implications, and contributed to the preparation of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Peer review information Nature Communications thanks the anonymous reviewer(s) for their contribution to the peer review of this work. Peer reviewer reports are available.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Subbulakshmi Radhakrishnan, S., Sebastian, A., Oberoi, A. et al. A biomimetic neural encoder for spiking neural network. Nat Commun 12, 2143 (2021). https://doi.org/10.1038/s41467-021-22332-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-021-22332-8
- Springer Nature Limited
This article is cited by
-
Ultrafast readout, crosstalk suppression iontronic array enabled by frequency-coding architecture
npj Flexible Electronics (2024)
-
An artificial visual neuron with multiplexed rate and time-to-first-spike coding
Nature Communications (2024)
-
Skin-inspired, sensory robots for electronic implants
Nature Communications (2024)
-
A two-dimensional mid-infrared optoelectronic retina enabling simultaneous perception and encoding
Nature Communications (2023)
-
Optimization of the structural complexity of artificial neural network for hardware-driven neuromorphic computing application
Applied Intelligence (2023)