Spatio-temporal Learning with Arrays of Analog Nanosynapses

Spatio-temporal Learning with Arrays of Analog Nanosynapses


Emerging nanodevices such as resistive memories are being considered for hardware realizations of a variety of artificial neural networks (ANNs), including highly promising online variants of the learning approaches known as reservoir computing (RC) and the extreme learning machine (ELM). We propose an RC/ELM inspired learning system built with nanosynapses that performs both on-chip projection and regression operations. To address time-dynamic tasks, the hidden neurons of our system perform spatio-temporal integration and can be further enhanced with variable sampling or multiple activation windows. We detail the system and show its use in conjunction with a highly analog nanosynapse device on a standard task with intrinsic timing dynamics- the TI-46 battery of spoken digits. The system achieves nearly perfect () accuracy at sufficient hidden layer size, which compares favorably with software results. In addition, the model is extended to a larger dataset, the MNIST database of handwritten digits. By translating the database into the time domain and using variable integration windows, up to classification accuracy is achieved. In addition to an intrinsically low-power programming style, the proposed architecture learns very quickly and can easily be converted into a spiking system with negligible loss in performance- all features that confer significant energy efficiency.

on-chip learning, spatiotemporal classification,reservoir computing, extreme learning machine, memristive devices, nanosynapses

I Introduction

Reservoir computing (RC) systems refer to learning models which use the combination of random weights and recurrent projection spaces- often called the liquid state- to generate rich outputs that can be used in spatiotemporal classification or processing tasks. Previously referred to as echo state networks (ESN) as well as liquid state machines (LSM) [1, 2], these systems use linear regression or ridge regression on the output of the reservoir to analytically obtain output weights. Once these weights are set, the system can be used for on-line classification or clustering tasks. An experimental demonstration of the non-linear mapping abilities of this approach has been made in optoelectronic systems [3], amongst other media; however, the optimal realization of reservoir and read-out weights is seldom discussed in the RC literature. Meanwhile, the advent of non-linear nanoscale electronic elements with volatile and non-volatile modes, often called memristive devices, has paved the way towards dynamic weights with intrinsic plasticity (adaption) features [4]. The use of said devices, hereafter “nanosynapses”, to construct and enhance reservoir computers was first proposed discussed in [5]; here, topology was arbitrary (randomly generated) and training employed a genetic algorithm. In [6], the performance of an ensemble of such random reservoirs was discussed, yet an on-chip friendly read-out and regression scheme was lacking. A completely on-chip feasible design for RC was proposed using a ring topology in [7], yet nanosynapses were only used in the output or regression layer (reservoir synapses were built with CMOS). Our proposed scheme uses memristive synapses in both layers by coupling two crossbar arrays; this regular topology is currently being experimentally realized in both two (planar) and three (stacked) dimensions, and is straight-forward to read and write [8]. In contrast with RC systems, and more like a complementary approach known as the Extreme Learning Machine (ELM) [9], our system is feed-forward; however, even with such a simplification, strong performance on tasks with time-dynamic can be achieved by exploiting time-dynamics within the hidden layer. This approach is reminiscent of a dendritically-inspired architecture proposed by Tapson [10], who used a variety of synaptic kernels to achieve promising results on spatio-temporal tasks. Ours uses simpler spatio-temporal integration schemes specifically designed for on-chip learning with emerging, highly analog nanosynapse models. Given well-known advantages of the RC/ELM paradigm (fast training, online operation) and the spatio-temporal data processing challenges inherent in an upcoming era of distributed computing, we expect this proposal should be of great interest to neuromorphic designers.

Ii Architecture

Ii-a Conceptual Depiction

A conventional RC architecture, visible in Fig. 1(a) is built with a sparsely connected graph of excitatory and inhibitory neurons in the liquid, some or all of which are recurrent; this graph is excited along a set of input channels connected to and feeding out along weights . As is static, can be solved as the pseudo-inverse of all collected output activations (), given a matrix of labels or expected values (), as depicted in the bottom of Fig.  1(a).

Our proposed scheme, introduced in Fig. 1(b), also sets input weights and linearly regresses to achieve output weights, but now has a feed-forward fully-connected architecture to make it crossbar-compatible. To achieve the spatio-temporal features needed to emulate reservoir performance, the hidden neurons each possess a stateful value which corresponds to the past activations they have received within the present example. Symbolically, given example presented over total time frames , neuron computes output as


where is the input dimensionality (total number of channels) and is the index of the input channel.

The function means that, like an RC system, neurons have binary states: excitatory (+) or inhibitory (-). The scheme is built to perform well on tasks presented frame-by-frame such as audio, video, or other tasks mapped to a time domain.

Two variations of the architecture are possible. In the ’uniform’ case, each hidden neuron integrates over every frame presented. In the ’variable’ case, each hidden neuron integrates only a subset of the time frames, specific to the hidden neuron: the sum over is limited to a subset of values.

This scheme is illustrated in Fig.  1(b), where the corresponding lettered hidden layer neurons integrate over the frame areas pictured below them.

Fig. 1: Conceptual illustrations of (a) the standard RC/LSM paradigm and the on-chip, (b) RC-inspired scheme proposed here.

Ii-B Nanoelectric Implementation

Our architecture implementation with nanoelectronics is presented in Fig. 2. It is constituted by two nanodevices crossbars implementing a “projection layer” and a “readout layer” connected by circuits implementing the hidden neurons.

Input and projection layer

Inputs are presented to the first layer as consecutive frames of voltage pulses to input channels. Weights in this layer are set randomly at the initialization of operation and unchanged. Two modes were considered: analog and spiking. Following dissemination through the crossbar, output currents at the other side are subtracted in pairs (so as to achieve negative weights) and translated through standard op-amp circuitry into the voltage domain. Only the binary value of this voltage, or its sign are relevant to the following cell.

Hidden neurons activation

Different technological options are possible for implementing the time-integration performed by hidden neurons circuits. It can be achieved digitally using counters, or in an analog fashion using capacitors, as is often done in the neuromorphic electronics field [11]. It could also potentially performed using analog nanosynapses.

As mentioned above, each hidden neuron integrates activation over either all frames (uniform scheme) or smaller, random subsets of frames (variable scheme). To realize the variable scheme a small memory needs to be associated and configured with each hidden neuron. In addition to standard memory components, another crossbar of nanosynapses set randomly on or off could be used to constitute the physical reference for a sparsity matrix; in this case, dot-product operations between layers of a 3D crossbar array would significantly reduce required overhead.

Readout layer

The readout layer uses the weight space of pairs of analog nanosynapses to provide on-chip regression corresponding to the activations from the hidden neurons. The scheme proposed for the readout layer is described in detail in [12]. It consists in a conditional pulse programming scheme which uses the combination of pulses (from the hidden neurons) and pulses (from a training cell) to implement an on-chip friendly (digital) version of stochastic gradient descent, reducing error on all device pairs in the readout layer. As pictured in Fig. 2, this scheme simply checks the sign of the sample’s label against the sign of the output for that class (crossbar column). This scheme is generic to several classes of nanosynapses and highly resistant to device imperfections [13]. The training cell can use stateful memristive devices to correctly route programming pulses to output neurons with the corresponding error case in the readout crossbar layer [14]; the chosen nanosynapse has been directly integrated in such a scheme [15].

Fig. 2: Schematic illustrating the simulated nano-system used to obtain results. Each square icon in both cartoons represents a nano-synapse. In the first layer (projection) either binary device and/or natural dispersion around the maximal or minimal values of the device can be used; in the second layer (readout), full weight range is exploited in order to achieve on-chip regression using a simplified version of the Widrow-Hoff or delta learning rule.

Ii-C Nanosynapse

The nanosynapse model is extrapolated from experimental data obtained from the hardware learning of a highly analog polyermic nanosynapse with several hundred writable/readable states. The device physically functions via active filamentary redox mechanisms occurring within an electrografted polymer thin-film of tris-bipyridine iron complexes (TBFe) [12]. Our generic model evaluates the evolution of the device between and ; conductance increases if and only if it is biased at a voltage above a critical threshold but below and decreases above , as a function of its total addressable states and the write range , , as follows :


This case is significantly simplified by using a constant programming scheme where for reset, and programming (pulse) length , such that can be approximated in a discrete time-step simulator during each writing (learning) phase. Two varieties of this model were considered on the network level: perfect devices, for which every device has identical extrema ( ) and identical thresholds; and imperfect devices, which have extrema distributed around a Gaussian spread from those means, and with first and second thresholds distributed around , . For all the following simulations, 7 bits or addressable states were considered for the analog readout layer. For the projection layer () which is unadaptive, the devices were either in ’binary mode’ and programmed randomly at one extrema (perfect devices); or were all programmed randomly around the same extrema (ON or OFF), taking advantage of the natural Gaussian curve as proposed in [16] (imperfect devices).

Iii Simulation Methodology

First, we considered a sub-set of 500 examples from the TI-46 corpus of spoken digits. These spoken digits were encoded using an artificial cochleagram; pre-processing was based on the passive ear model or filter bank designed by R.F. Lyon [17]. Full details on this pre-processing and database are available in [18]. The processed data we used in our tests shows 77 channels with between 50 and 100 time frames or steps during each example digit. The dataset was imported from and used in the context of a generic software simulator for reservoir computing written in Python (Oger) [19]. The identical database and pre-processing methods were used to present this set as training and testing examples to a custom discrete-time crossbar simulator which tracks the evolution of all hidden neurons activation values and (second layer) device weights; this software was also written in Python.

The 500 samples were divided into a training set of 350 samples, and a test set of 150 samples. The training set is labeled and used either to teach the readout layer or to build an output matrix for ridge regression; conversely, testing examples are presented label-less, and the prediction or inference is compared externally to the system. Each full training procedure consisted of 10 subsequent presentations (epochs) of the sample set (5k iterations); examples were always shuffled between each epoch.

Although this task has intrinsic time dynamic, the number of total samples is very small. To further estimate the generalization abilities of our architectures, we also tested with the well known MNIST database of handwritten digits [20], which consists of 70,000 samples (10,000 tests, 60,000 training examples). In order to lend time-dynamic to the task we presented all individual examples (from training and test) as 28 frames presented subsequently to the input layer, as in [21]. As in the TI-46 database, training examples were always shuffled (presented in a different order). Training with this database, due to its large size, consists of only a single epoch; no distortions or pre-processing were made before presentation; reported scores are always based on classification on the entire test database.

For analog mode presentation: pixel intensity or channel’s analog value for cochleagram were input directly as the corresponding voltage pulses. For spiking mode, all negative cochleagram values (range ) were converted to 0 (no spike), and all positive to 1 (spike); for MNIST, pixel luminosity greater than 0.5 spiked, and that below did not (range ). When applying noise to spike channels, the noted percent of pixels or channels at each moment flipped from one bit value to the other (on average).

Iv Learning Performances

Iv-a Effect of Liquid/ Hidden Layer size

As visible in Fig. 3 (a), both the software reservoir and our proposed crossbar interpretation are strongly dependent on a large enough number of nodes in the reservoir or hidden layer to reach robust performance. However, while the software reservoir achieves already by neurons, a dimensionality of at least is required for the crossbar based version when it is built with perfect nanosynapses. As visible, the variable sampling method creates a noticeable but not dramatic improvement.

The case of the MNIST challenge shows the same strong dependence of performance on hidden layer size, but in this case, the performance improvement between the uniform integration and variable schemes is dramatic. The uniform case only approaches at large number of hidden neurons; this is an inferior result for a two layer system, given that a single layer (perceptron) system can achieve in software [22] and a one-layer crossbar system using the same device can approximate this result [12]. This suggests that the simple, or uniform integration scheme that was used earlier is not adequate to differentiate between many presented MNIST samples in the training phase. Two factors may be at play: first, when artificially converted into the time domain few input channels are available in general; second, integration (summation) across the entire time slice cannot enhance dimensionality but only reduce it. This would be especially punishing in the MNIST case due to the complexity of the training data (which is not, in general, linearly separable). In contrast, adding hidden layer variability results in a substantial boost in performance, with ultimate performance at 95% at large hidden size. To obtain these results a random subset of one fourth of the frames from each image was selected by each neuron; this preference is consistent over all presentations and between training and testing.

Fig. 3: Performance of the considered architecture on the two chosen tasks. (a) shows performance on Lyon for software, on-chip simple integration, and on-chip sparse integration; (c) shows both considered integration schemes for the MNIST challenge where mild variability () is showed in the ’imperfect’ series; (c) shows deterioration in performance as dispersion the , parameters is increased; (d) shows response of variably integrating systems to the degree of spareness in the first layer, where 1.0 is the default (uniform) case; (e) shows convergence speed of the considered systems as a fraction of the total given samples in other graphs; (f) shows degradation in performance as the input channels to the first layer are increasingly simplified (analog to digital) and corrupted (noise).

Iv-B Effect of sparsity on variable scheme

Next, we explored whether optimal sparse connection levels exist. As visible in Fig. 3 (c), the optimal sparsity greatly depends upon the task; the optimal sparsity for MNIST is quite high, with best performances coming when between 40-20% of frames are active per neuron. For the Lyon task, 85-65% connection produces the best results. While outcome for MNIST devices with 10% variability is mostly equivalent to perfect devices, the same cannot be said for Lyon imperfect devices which trail performance relative to perfect ones increasingly as sparsity increases.

Iv-C Effect of readout device variability

While projection layer variability can easily be taken in stride by using normal distributions as an asset, imperfect devices in the adaptive (readout) layer can confound the learning rule’s performance. Fig. 3 (d) shows performance on each of the two considered challenges as a function of increasing standard deviation or dispersion of maximal, minimal conductances and thresholds at for the Lyon task, and for the MNIST task. As visible, slight variability is not very harmful, while greater than starts to substantively effect performance. Notably, the rate of deterioration on the Lyon task is much higher than the MNIST task; in the far smaller readout crossbar for the Lyon task, each device is more important to successfully resolving the problem. Reducing the impact of device imperfections in general could be realized via better engineering or through architectural tweaks such as using more than two nanodevices per single nanosynapse element.

Iv-D Fast Training Capabilities

One of the signature benefits of the RC/ELM paradigm is that such systems generally require only a fraction of the quantity of training examples needed to solve the problem in more computationally demanding methods; e.g. multi-layer backpropagation or convolutional neural networks often require dozens of epochs (millions of individual updates) to properly converge. As illustrated in Fig. 3 (e), the Lyon and MNIST tasks achieve within 10% of the maximum performance in as few as 20% (12k) training examples, in the former case, and 60% (2.1k) training examples, in the latter case. For even more approximate applications (less stringent accuracy requirements), 80% of performance can be reached in only 5k samples, for the case of the MNIST, and 1.4k samples, for the Lyon case. As small error rates lead to substantial energy savings, these results suggest the system is an attractive approximate computer.

Iv-E Performance with Spiking Input

In neuromorphic applications, spike coding bestows extreme energy efficiency [23]. We tested the capability of our system to still provide strong solutions when the problem was presented in either in perfectly presented spike form, or imperfect (increasingly noisy) spiking channel input, as described previously. As visible in Fig. 3 (f), the initial conversion from analog to digital (spiking) presentation has a slim effect on ultimate performance (between 1.1% and 0.4% drop from ideal analog case). With increasing noise, performance deteriorates slowly; on a slightly noisy channel ( 5% bit flips) performance drops only 4-8%, and even with nearly one third of the channel corrupted performance is still within 80% of the maximum. The system’s performance degradation on the Lyon task is less than that on the MNIST challenge; this shows that encoding on the latter on only 28 channels is intrinsically difficult and with added noise many of the features are lost.

Iv-F Energy Estimates

The energy benefits of programming devices in a crossbar array are substantial relative to the alternatives such as SRAM [24]. Even better, this scheme keeps the first layer weights static and trains the second layer using a conditional programming pulse scheme (not every output neuron is programmed at every training step). Due to high voltage levels for the simulated TBFe polymeric device (), we also include energy estimates using an alternate polymeric electrochemical nanosynapse called ENODe [25] where . However, there is a trade-off in programming speed, since for TBFe and for ENODe. The energy cost per elementary write operation for the two devices is computed finally as and , respectively. Total programming expenditure for the Lyon task and the MNIST task for both polymeric nanosynapses at full training (full performance), within 10 % of max performance, and within 20% of max performance following the analysis in Fig. 3 (e) are visible in Table I; system (hidden layer) size for Lyon is , and for MNIST. While this estimation is only raw programming cost, power draw for sum/ integration operations in the hidden layer, and checking/routing of error should be far less, especially if they use nanosynapses.

Device/Task Full Performance 10% Loss 20% Loss
ENODe-Lyon 0.65mJ 0.39mJ 0.182mJ
TBFe-Lyon 154mJ 92.4mJ 43.12mJ
TBFe-MNIST 11.09J 2.22J 0.924J
ENODe-MNIST 46.8 mJ 9.36mJ 3.9mJ
TABLE I: Programming cost for two possible nanosynapses

Iv-G Benchmarks

Our best obtained results on the Lyon task match not only RC software results, but the best obtained results with complex Long-short term memory cell (LSTM) networks [26]. Our results are very slightly inferior to physically realized reservoir computers; [3] obtained only 0.4% error. Finally, to test how critical the two layer design is to the high performance we achieved, we simulated a one-layer system with the training data directly presented (only the readout layer). Using this approach, we achieved only 56% using the uniform and 62% using the variable spatio-temporal integration scheme. This suggests that direct integration is insufficient for effective spatio-temporal learning. As suggested in [10], the critical operation expressed by synaptic (in our system, hidden layer) kernels is to project inputs into a higher dimensional space.

Many benchmarks are available for MNIST, but of more relevance to our case are those using nanosynapse weights in part or whole. In comparison to the results obtained in [21], which again presented that database similarly to memristive crossbars, we obtained superior results while also avoiding off-chip computational requirements. Our top obtained result is superior to the standard ELM result (91.5%) and similar to the ’enhanced ELM’ result (94.1%) showed in [27]. But unlike the latter case, our system requires no special priming/pre-training step in the first layer. A similar combination of unsupervised and supervised learning was used in [28] to achieved accuracy on the MNIST test set; in contrast, our system avoids relying on any first layer plasticity effects. For many, although not all, nanosynapses, plasticity learning rules such as STDP rely on carefully designed waveforms.

V Discussion

While the proposed design has strong similarities to the RC paradigm: the first layer weights (synaptic attractions) are fixed, standard regression can be used to extract the value of this processing network, and neurons show both positive (excitatory) and negative (inhibitory) activations, like ELM, the system is feedforward, not recurrent. As recurrence bestows a higher memory capacity on a reservoir, it is not surprising that more hidden neurons are required in our case than a true reservoir. The computational capabilities of RCs and ELMs have previously been compared in [29], which noted that while RCs imply an unfortunate trade-off between non-linearity (mapping performance) and memory capacity (ability to retain said map beyond the fading time window), a synthesis of ELM and RC approaches might overcome this dilemma. Our system implicitly addresses this trade-off by forcing non-linearity at a large set of hidden neurons, and preserving capacity through stateful variables.

Our results with the frame-by-frame MNIST task suggest that achieving strong performance on a high-dimensional task requires more than simple spatio-temporal integration. When hidden neurons time variability is included so as to add diversity to activation functions, the computational power of the system amplifies substantially as neurons specialize on their preferred frames from the database. Effectively, this approach forces greater sparsity (breaks all-to-all connectivity). Software RC systems intrinsically exploit sparsity, while ELM systems exploiting this effect are rarer; different local receptive fields were suggested in [30] and used to achieve good performance on the NORB database, but the scheme is too complex to be easily implemented on-chip. In contrast, our proposed scheme is hardware-friendly, and has been designed with future 3D crossbar topologies in mind.

Vi Conclusion

By integrating two standard nanoelectric structures (crossbars) with hidden neurons performing spatio-temporal activation/integration, we have emulated many of the features of RC/ELM software systems: high performance, online learning, and resilience to imperfect dynamics or inputs. 99% on the T1-46 battery matches the performance of a software reservoir computer with floating point weights; while the same equivalence cannot be stated about our results on the MNIST task, 95% is nonetheless a strong result for a crossbar based learning system. Immediate next steps in translating the proposed design into a prototype include study on the energy draw and operation of optimal CMOS associated circuits, especially those involving hidden layer functions, electrical simulation of the proposed system to explore non-ideal effects, and design of toy problems capable of being implemented on far smaller crossbars than those used in the present study.


This work was supported by the Nanodesign Paris-Saclay lidex.


  1. W. Maass, T. Natschläger, and H. Markram, “Real-time computing without stable states: A new framework for neural computation based on perturbations,” Neural computation, vol. 14, no. 11, pp. 2531–2560, 2002.
  2. H. Jaeger, Tutorial on training recurrent neural networks, covering BPPT, RTRL, EKF and the” echo state network” approach.   GMD-Forschungszentrum Informationstechnik, 2002, vol. 5.
  3. Y. Paquot, F. Duport, A. Smerieri, J. Dambre, B. Schrauwen, M. Haelterman, and S. Massar, “Optoelectronic reservoir computing,” Scientific reports, vol. 2, p. 287, 2012.
  4. J. J. Yang, D. B. Strukov, and D. R. Stewart, “Memristive devices for computing,” Nature nanotechnology, vol. 8, no. 1, pp. 13–24, 2013.
  5. M. S. Kulkarni and C. Teuscher, “Memristor-based reservoir computing,” in Nanoscale Architectures (NANOARCH), 2012 IEEE/ACM International Symposium on.   IEEE, 2012, pp. 226–232.
  6. J. Bürger, A. Goudarzi, D. Stefanovic, and C. Teuscher, “Hierarchical composition of memristive networks for real-time computing,” in Nanoscale Architectures (NANOARCH), 2015 IEEE/ACM International Symposium on.   IEEE, 2015, pp. 33–38.
  7. D. Kudithipudi, Q. Saleh, C. Merkel, J. Thesing, and B. Wysocki, “Design and analysis of a neuromemristive reservoir computing architecture for biosignal processing,” Frontiers in neuroscience, vol. 9, 2015.
  8. G. C. Adam, B. D. Hoskins, M. Prezioso, F. Merrikh-Bayat, B. Chakrabarti, and D. B. Strukov, “3-d memristor crossbars for analog and neuromorphic computing applications,” IEEE Transactions on Electron Devices, vol. 64, no. 1, pp. pp–312, 2017.
  9. G.-B. Huang, Q.-Y. Zhu, and C.-K. Siew, “Extreme learning machine: theory and applications,” Neurocomputing, vol. 70, no. 1, pp. 489–501, 2006.
  10. J. Tapson, G. Cohen, S. Afshar, K. Stiefel, Y. Buskila, R. Wang, T. J. Hamilton, and A. van Schaik, “Synthesis of neural networks for spatio-temporal spike pattern recognition and processing,” arXiv preprint arXiv:1304.7118, 2013.
  11. G. Indiveri, B. Linares-Barranco, T. J. Hamilton, A. Van Schaik, R. Etienne-Cummings, T. Delbruck, S.-C. Liu, P. Dudek, P. Häfliger, S. Renaud et al., “Neuromorphic silicon neuron circuits,” Frontiers in neuroscience, vol. 5, p. 73, 2011.
  12. Y.-P. Lin, C. H. Bennett, T. Cabaret, D. Vodenicarevic, D. Chabi, D. Querlioz, B. Jousselme, V. Derycke, and J.-O. Klein, “Physical realization of a supervised learning system built with organic memristive synapses,” Scientific Reports, vol. 6, 2016.
  13. D. Chabi, W. Zhao, D. Querlioz, and J.-O. Klein, “On-chip universal supervised learning methods for neuro-inspired block of memristive nanodevices,” ACM Journal on Emerging Technologies in Computing Systems (JETC), vol. 11, no. 4, p. 34, 2015.
  14. D. Chabi, Z. Wang, C. Bennett, J.-O. Klein, and W. Zhao, “Ultrahigh density memristor neural crossbar for on-chip supervised learning,” IEEE Transactions on Nanotechnology, vol. 14, no. 6, pp. 954–962, 2015.
  15. C. Bennett, D. Chabi, T. Cabaret, B. Jousselme, V. Derycke, D. Querlioz, and J.-O. Klein, “Supervised learning with organic memristor devices and prospects for neural crossbar arrays,” in Nanoscale Architectures (NANOARCH), 2015 IEEE/ACM International Symposium on, no. Proceedings of IEEE.   IEEE, 2015, pp. 181–186.
  16. M. Suri and V. Parmar, “Exploiting intrinsic variability of filamentary resistive memory for extreme learning machine architectures,” IEEE Trans. Nanotechnol., vol. 14, no. 6, pp. 963–968, 2015.
  17. R. Lyon, “A computational model of filtering, detection, and compression in the cochlea,” in Acoustics, Speech, and Signal Processing, IEEE International Conference on ICASSP’82., vol. 7.   IEEE, 1982, pp. 1282–1285.
  18. D. Verstraeten, B. Schrauwen, D. Stroobandt, and J. Van Campenhout, “Isolated word recognition with the liquid state machine: a case study,” Information Processing Letters, vol. 95, no. 6, pp. 521–528, 2005.
  19. D. Verstraeten, B. Schrauwen, S. Dieleman, P. Brakel, P. Buteneers, and D. Pecevski, “Oger: modular learning architectures for large-scale sequential processing,” Journal of Machine Learning Research, vol. 13, no. Oct, pp. 2995–2998, 2012.
  20. Y. LeCun, C. Cortes, and C. J. Burges, “Mnist handwritten digit database,” AT&T Labs [Online]. Available: http://yann. lecun. com/exdb/mnist, vol. 2, 2010.
  21. J. Bürger and C. Teuscher, “Volatile memristive devices as short-term memory in a neuromorphic learning architecture,” in Proceedings of the 2014 IEEE/ACM International Symposium on Nanoscale Architectures.   ACM, 2014, pp. 104–109.
  22. Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998.
  23. P. A. Merolla, J. V. Arthur, R. Alvarez-Icaza, A. S. Cassidy, J. Sawada, F. Akopyan, B. L. Jackson, N. Imam, C. Guo, Y. Nakamura et al., “A million spiking-neuron integrated circuit with a scalable communication network and interface,” Science, vol. 345, no. 6197, pp. 668–673, 2014.
  24. S. Agarwal, T.-T. Quach, O. Parekh, A. H. Hsia, E. P. DeBenedictis, C. D. James, M. J. Marinella, and J. B. Aimone, “Energy scaling advantages of resistive memory crossbar based computation and its application to sparse coding,” Frontiers in neuroscience, vol. 9, 2015.
  25. Y. van de Burgt, E. Lubberman, E. J. Fuller, S. T. Keene, G. C. Faria, S. Agarwal, M. J. Marinella, A. A. Talin, and A. Salleo, “A non-volatile organic electrochemical device as a low-voltage artificial synapse for neuromorphic computing,” Nature Materials, vol. 16, no. 4, pp. 414–418, 2017.
  26. A. Graves, D. Eck, N. Beringer, and J. Schmidhuber, “Biologically plausible speech recognition with lstm neural nets,” in International Workshop on Biologically Inspired Approaches to Advanced Information Technology.   Springer, 2004, pp. 127–136.
  27. C. H. Bennett, S. La Barbera, A. F. Vincent, J.-O. Klein, F. Alibart, and D. Querlioz, “Exploiting the short-term to long-term plasticity transition in memristive nanodevice learning architectures,” in Neural Networks (IJCNN), 2016 International Joint Conference on.   IEEE, 2016, pp. 947–954.
  28. D. Querlioz, W. Zhao, P. Dollfus, J.-O. Klein, O. Bichler, and C. Gamrat, “Bioinspired networks with nanoscale memristive devices that combine the unsupervised and supervised learning approaches,” in Proceedings of the 2012 IEEE/ACM International Symposium on Nanoscale Architectures.   ACM, 2012, pp. 203–210.
  29. J. Butcher, D. Verstraeten, B. Schrauwen, C. Day, and P. Haycock, “Reservoir computing and extreme learning machines for non-linear time-series data analysis,” Neural networks, vol. 38, pp. 76–89, 2013.
  30. G.-B. Huang, Z. Bai, L. L. C. Kasun, and C. M. Vong, “Local receptive fields based extreme learning machine,” IEEE Computational Intelligence Magazine, vol. 10, no. 2, pp. 18–29, 2015.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description