Do we know the operating principles of our computers better than those of our brain?

by   János Végh, et al.

The increasing interest in understanding the behavior of the biological neural networks, and the increasing utilization of artificial neural networks in different fields and scales, both require a thorough understanding of how neuromorphic computing works. On the one side, the need to program those artificial neuron-like elements, and, on the other side, the necessity for a large number of such elements to cooperate, communicate and compute during tasks, need to be scrutinized to determine how efficiently conventional computing can assist in implementing such systems. Some electronic components bear a surprising resemblance to some biological structures. However, combining them with components that work using different principles can result in systems with very poor efficacy. The paper discusses how the conventional principles, components and thinking about computing limit mimicking the biological systems. We describe what changes will be necessary in the computing paradigms to get closer to the marvelously efficient operation of biological neural networks.



There are no comments yet.


page 1

page 2

page 3

page 4


On the spatiotemporal behavior in biology-mimicking computing systems

The payload performance of conventional computing systems, from single p...

Simulation of neural function in an artificial Hebbian network

Artificial neural networks have diverged far from their early inspiratio...

Neuromorphic Control

Neuromorphic engineering is a rapidly developing field that aims to take...

Characterizing Self-Developing Biological Neural Networks: A First Step Towards their Application To Computing Systems

Carbon nanotubes are often seen as the only alternative technology to si...

Biologically Inspired Design Principles for Scalable, Robust, Adaptive, Decentralized Search and Automated Response (RADAR)

Distributed search problems are ubiquitous in Artificial Life (ALife). M...

Which scaling rule applies to Artificial Neural Networks

Although Artificial Neural Networks are biology-mimicking systems, they ...

Connectionism, Complexity, and Living Systems: a comparison of Artificial and Biological Neural Networks

While Artificial Neural Networks (ANNs) have yielded impressive results ...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Today we have the ”golden age” of neuromorphic (brain-inspired, artificial intelligence) architectures and computing. However, the meaning of the word has changed considerably since Carver Mead 


coined the wording. Today practically every single solution that borrows at least one single operating principle from the biology and mimics some of its functionality in a more or less successful way deserves this name. As always, to grasp out some single aspect and implement it in an environment and from components based on entirely different principles, is dangerous. Historically, ’neuromorphic’ architectures were suggested to be based on different principles and components, such as mechanics, pneumatics, telephones, analog and digital electronics, computing. Some initial resemblance surely exists, and even some straightforward systems can demonstrate more or less successfully functionality in some aspects similar to that of the nervous system. There is a noteworthy analogy between the deep learning of neuronal nodes and the long-term potentiation found in synapses.

However, when scrutinizing the scalability (i.e., how those systems shall work when used under real-life conditions in which a vast number of similar subsystems shall work and cooperate), the picture is not favorable at all. ”Successfully addressing these challenges [of neuromorphic computing] will lead to a new class of computers and systems architectures” NeuromorphicComputing:2015 has been targeted. However, as noticed by the judges of the Gordon Bell Prize, ”surprisingly, [among the winners,] there have been no brain-inspired massively parallel specialized computers” GordonBellPrize:2017 . Despite the vast need and investments, furthermore the concentrated and coordinated efforts, just because of mimicking the biological systems with computing inadequately.

Given ”that the quest to build an electronic computer based on the operational principles of biological brains has attracted attention over many years” FurberNeuralEngineering:2007 , modeling the neuronal operation became a well-known field in electronics. At the same time, more and more details come to light about the computational operations of the brain. However, it would appear, that the ’wet’ neuroscience is miles ahead of the ’silicon’ neuroscience. There are projects and exaggerated claims about extremely large computing systems, even about targeting the simulation of the brain of some animals and eventually even the human brain. Often these claims are followed by a long silence, or some rather slim or no results. As that the operating principles of the large computer systems tend to deviate from the operating principles of a single processor, it is worth reopening the discussion on a decade-old question ”Do computer engineers have something to contribute. . . to the understanding of brain and mind?”  FurberNeuralEngineering:2007 . Maybe, and they surely have something to contribute to the understanding of computing itself. There is no doubt that the brain does computing, the key question is how?

Section 2 presents some computing systems, having, as we point out: as a consequence of the computing paradigm, enormously high energy consumption. Section 3 discuses the primary reasons for the issues and failures: the computing paradigm and their consequences such as the serial bus and the effect of the physical size. The timely behavior is especially important in the biological objects, so their fair imitation in the computing systems is of crucial importance, as section 4 discusses it. The neuromorphic computing is a special type of workloads that have a dominating role in forming the computational efficiency of the computing systems, as section 5 discusses it. Section 6 presents some further limitations, rooting in the classical paradigm; furthermore, it draws parallels with classic versus modern science and classic versus modern computing. Section 7 provides examples, why is of limited validity to consider the role of a grasped-out component: a neuromorphic system is not a simple sum of its components. In section 8, the paper attempts to make a clear pointer where we can continue using the classic computing and where we shall base the systems on the new principles.

2 Issues with the large scale computing

The worst limiting factor in conventional computing is the method of communication between processors, which increases exponentially with increasing complexity/number. Historically in the model of computing proposed by von Neumann, there is one single entity, an isolated (non-communicating) processor, whereas in the bio-inspired models, billions of entities, organized into specific assemblies, cooperate via communication. (The communication here means not only sending data, but also sending/receiving signals, including synchronization of the operation of the entities.) Neuromorphic systems, expected to perform tasks in one paradigm, but assembled from components manufactured using principles of (and implemented by experts trained in) the other paradigm are unable to perform at the required speed and efficacy for real-world solutions. The larger the system, the higher the communication load and the performance debt.

To get nearer to the marvelously efficient operation of the biological brain, other features must also be mimicked from the biology. Only a little portion of the neurons are working simultaneously in solving the actual task; there is a massive number of very simple (’very thin’) processors rather than a ’fat’ processor111One should scrutinize whether it is worth to implement accelerators (such as pipelines, branch predictors) intended to be used in large computing systems, to achieve just a couple of times higher processing speed at the price of using several hundred times more transistors; only a portion of the functionality and connection are pre-wired, the rest is mobile; there is an inherent redundancy, replacing a faulty neuron may be possible via systematic training. The conventional processors can only either run or halt, but not to make a little break. The biology uses purely event-driven (asynchronous) computing, while modern electronics uses clock-driven systems; for the catastrophic consequences of attempting to simulate a neuromorphic system, such as the human brain, using components prepared for conventional computing, see the case of SpiNNaker, discussed below.

The large computing systems can cope with the tasks with growing difficulty, enormously decreasing computing efficiency, and enormously growing energy consumption. Being not aware of that the collaboration between processors needs a different approach (another paradigm), resulted in demonstrative failures already known (such as the supercomputers Gyoukou and Aurora’18, or the brain simulator SpiNNaker)222The explanations are quite different: Gyoukou is simply withdrawn; Aurora is practically withdrawn: retargeted and delayed; Despite the failure of SpiNNaker1, the SpiNNaker2 is also under construction SpiNNaker2:2018 ; ”Chinese decision-makers decided to withhold the country’s newest Shuguang supercomputers even though they operate more than 50 percent faster than the best current US machines”. and many more (all they intend to deliver 0.13-0.2 Eflops) may follow: such as Aurora’21 DOEAurora:2017 , the China mystic supercomputers333 and the EU planned supercomputers444 doc_id =60156. Systems having ”only” millions of processors already show the issues, and the brain-like systems want to comprise four orders of magnitude higher number of computing elements. Besides, the scaling is strongly nonlinear VeghReevaluate:2020 ; VeghScalingANN:2020 . When targeting neuromorphic features such as ”deep learning training”, the issues start to manifest at just a couple of dozens of processors DeepNeuralNetworkTraining:2016 VeghAIperformance:2020 .

3 Limitations due to the Single Processor Approach

As suspected by many experts, the computing paradigm itself, ”the implicit hardware/software contract AsanovicParallelCACM:2009 ”, is responsible for the experienced issues: ”No current programming model is able to cope with this development [of processors], though, as they essentially still follow the classical van Neumann model” SoOS:2010 . When thinking about ”advances beyond 2020”, the solution was expected from the ”more efficient implementation of the von Neumann architecture” DeBenedictis_supercomputing:2014 , however. Even when speaking about building up computing from scratch (”rebooting the model” RebootingComputingModels:2019 ), only implementing different gating technology for the same computing model is meant. However, the paradigm prevents building large neuromorphic systems, too.


Figure 1: Implementing neuronal communication in different technical approaches. A: the parallel bus; B and C: the shared serial bus, before and after reaching the communication ”roofline” WilliamsRoofline:2009

The bottleneck is essentially the ”technical implementation” of the communication, stemming from the SPA, as illustrated in Fig. 1. The inset shows a simple neuromorphic use case: one input neuron and one output neuron communicating through a hidden layer, comprising only two neurons. Fig. 1.A mostly shows the biological implementation: all neurons are directly wired to their partners, i.e., a system of ”parallel buses” (the axons) exists. Notice that the operating time also comprises two non-payload times: the data input and data output, which coincide with the non-payload time of the other communication party. The diagram displays the logical and temporal dependencies of the neuronal functionality.

The payload operation (”the computing”) can only start after the data is delivered (by the, from this point of view, non-payload functionality: input-side communication), and the output communication can only begin when the computing finished. Importantly, the communication and calculation mutually block each other. Two important points that neuromorphic systems must mimic noticed immediately: i/ the communication time is an integral part of the total execution time, and ii/ the ability to communicate is a native functionality of the system. In such a parallel implementation, the performance of the system, measured as the resulting total time (processing + transmitting), scales linearly with increasing both the non-payload communication speed and the payload processing speed.

The present technical approaches assume a similar linearity of the performance of the computing systems as ”Gustafson’s formulation Gustafson:1988 gives an illusion that as if N [the number of the processors] can increase indefinitely” AmdalVsGustafson96 . The fact that ”in practice, for several applications, the fraction of the serial part happens to be very, very small thus leading to near-linear speedups” AmdalVsGustafson96 , however, misled the researchers. Gustafson’s ’linear scaling’ neglects the communication entirely (which is not the case, especially not in neuromorphic computing). He established his conclusions on only several hundred processors, and the interplay of the improving parallelization and the general HW development (including the non-determinism of the modern HW PerformanceCounter2013 ) covered for decades that the scaling was used far outside of its range of validity VeghReevaluate:2020 ; VeghScalingANN:2020 . Not considering the effect of the time of communication (i.e., the timely behavior), means not considering a vital feature of the biological system. Essentially the same effect (the vastly increased number of idle cycles due to the physical size of supercomputers) leads to the failures of supercomputer projects (for a detailed discussion, see VeghHowMany:2020 ). The ’real scaling’ is strongly nonlinear, with nature-defined bound.

Fig. 1.B shows a technical implementation of a high-speed shared bus for communication. To the right of the grid, the activity that loads the bus at the given time is shown. A double arrow illustrates the communication bandwidth, the length of which is proportional to the number of packages the bus can deliver in a given time unit. The high-speed bus is only very slightly loaded. We assume that the input neuron can send its information in a single message to the hidden layer furthermore that the processing by the neurons in the hidden layer both starts and ends at the same time. However, the neurons must compete for accessing the bus, and only one of them can send its message immediately, the other(s) must wait until the bus gets released. The output neuron can only receive the message when the first neuron completes it. Furthermore, the output neuron must first acquire the second message from the bus, and the processing can only begin after having both input arguments. This constraint results in sequential bus delays both during non-payload processing in the hidden layer and the payload processing in the output neuron. Adding one more neuron to the layer introduces one more delay, which explains why ”shallow networks with many neurons per layer …scale worse than deep networks with less neuronsDeepNeuralNetworkTraining:2016 : the system sends them at different times in the different layers (and even they may have independent buses between the layers), although the shared bus persists in limiting the communication.

Figure 2: The effect of the SPA implementation of the computing performance. The surface displays how the efficiency of the SPA systems depends on the goodness of the parallelization (aka the non-payload/payload ratio) and the number of cores. The figure marks show at what efficiency values can the top supercomputers run different workloads (for a discussion see section 5): the ’best load’ benchmark HPL, and the ’real-life load’ HPCG. The right bottom part displays the expected efficiency of running neuromorphic calculations on SPA computers.

The dependence of the performance is strongly nonlinear at higher performance values (implemented using a large number of processors). The effect is especially disadvantageous for the systems, such as the neuromorphic ones, that need much more communication, thus making the non-payload to payload ratio very wrong. The linear dependence at low nominal performance values explains why the initial successes of any new technology, material or method in the field, using the classic computing model, can be misleading: in simple cases, the classic paradigm performs tolerably well thanks to that compared to biological neural networks, current neuron/dendrite models are simple, the networks small and learning models appear to be rather basic. Recall that for artificial neuronal networks the saturation is reached at just dozens of processors DeepNeuralNetworkTraining:2016 , because of the extreme high proportion of communication.

Fig. 2 depicts how the SPA paradigm also defines the computational performance of the parallelized sequential system. Given that the task defines how many computations it needs, and the computational time is inversely proportional with the efficiency, one can trivially conclude that the decreasing computational efficiency leads to increasing energy consumption, just because of the SPA. ”This decay in performance is not a fault of the architecture, but is dictated by the limited parallelism” ScalingParallel:1993 .

4 The importance of imitating the timely behavior

In both biological and electronic systems, both the distance between the entities of the network, and the signal propagation speed is finite. Because of this, in the physically large-sized systems the ’idle time’ of the processors defines the final performance a parallelized sequential system can achieve. In the conventional computing systems also the ’data dependence’ limits the available parallelism: we must compute the data before we can use it as an argument for another computation. Although of course in the conventional computing the data must be delivered to the place of the second utilization, thanks to the ’weak scaling’ Gustafson:1988 , this ’communication time’ is neglected.

In neuromorphic computing, however, as discussed in connection with Fig. 1, the transfer time is a vital part of information processing. A biological brain must deploy a ”speed accelerator” to ensure that the control signals arrive at the target destination before the arrival of the controlled messages, despite that the former derived from a distant part of the brain BuzsakiGammaOscillations:2012 . This aspect is so vital in biology that the brain deploys many cells with the associated energy investment to keep the communication speed higher for the control signal. Computer technology cannot speed up the communication selectively, as in biology, and it is not worth to keep part of the system for a lower speed selectively.

Extending on our previous work [22], here we introduce the Explicitly Many-Processor Approach (EMPA) which is a new method using clusters of computational units (‘neurons’ Fig. 3B) to mimic the timely behaviour of ‘biological brains’. The clusters are built using the following key novel ideas: 1) implementing directly-wired connections between physically neighbouring cells; 2) creating a special hierarchical bus system; 3) placing a special communication unit, the (ICCB, Fig. 3B, purple) between the computer cores mimicking neurons ( Fig. 3B, green); 4) creating a specialized ‘cluster head’ ( Fig. 3B) with the extra abilities to access the local and far memories ( Fig. 3 M) and to forward messages via the gateway ( Fig. 3  G) to other ‘clusters’ (similar gateways can be implemented for the inter-processor communication, and higher organizational levels; providing access to different levels of the hierarchical buses). The cluster members are denoted by their relative position (the addressing mode enables using virtual cores, mapped to physical cores at runtime), and they can access the memory and other clusters only through the head of the cluster. This enables both easy sharing of locally important state variables, keeps local traffic away from the bus(es) and reduces wiring inside the chip. The ICCBs can forward messages via direct wired connections with up to 2 ’hops’ to the immediate neighbors and the second neighbors (even if they belong to another cluster). This solution enables billions of ’neurons’ to communicate at the same time, without delay, although the distant neuron must use one (or more) of the hierarchical buses, in function or their location. The resemblance between Fig. 3A and Fig. 7 in reference BuzsakiGammaOscillations:2012 underlines the importance of making a clear distinction between handling ’near’ and ’far’ signals, and accounting for relative signal timing. Although the ICCB blocks can adequately represent ’locally connected’ interneurons and the ’G’ gateway the ’long-range interneurons’ BuzsakiGammaOscillations:2012 , in biological systems conduction time must be separately maintained by the neurons in biology-mimicking computing systems. Making time-stamps and relying on the computer network delivery principles is not sufficient for maintaining correct relative timing. The timely behavior is a vital feature of the biology-mimicking systems, cannot not be replaced it with the synchronization principles of computing. Ignoring this requirement massively contributes to the failures of biology-mimicking computing systems. Communication time is less vital for using neurons in AI, but even in that case, one must consider the communication time explicitly.

Figure 3: Subfigure A (compare to Fig.7 in BuzsakiGammaOscillations:2012 ) shows a proposal VeghSPAEMPA:2020 of how to reduce the limiting effect of the SPA, via mimicking the communication between local neurons using direct-wired inter-core communication and the communication between the farther neurons via using the inter-cluster communication bus, in the cluster head. Subfigure B suggests a possible implementation of the principle: the Inter-Core Communication Blocks represent a ”local bus” (directly wired, with no contention), while the cores can communicate with the cores in other clusters through the ’G’ gateway as well as the ’M’ (local and global) memory.

Of course, computing works through having time quanta: what happens within a clock period of the processor, happens ”at the same time”. Given that the clock period of computers is in the range of nanoseconds, in the classic computing good approximation is that computing time is continuous. Simulating many-neuron systems in SPA, however, one faces a lack of the cooperative behavior. As the computing time in the artificial neurons is not proportional with the biological time they simulate, these different time scales must be scaled to each other.

One possible way is to put a ”time grid” on the processes simulating biology: within a time slot, the artificial neurons would be free to compute, but at the time boundary they would send the results of their calculation to each other. This results in the neurons continuing their calculation periodically from some concerted state. Such a method of synchronization introduces a ”biological clock period” that is million-fold longer than the clock period of the processor: what happens in this ”grid time”, happens ”at the same time”. Although this effect drastically reduces the achievable computing temporal performance VeghBrainAmdahl:2019 , the synchronization principle is so common that also the special-purpose neuromorphic chips TrueNorth:2016 ; IntelLoihi:2018 use it as a built-in feature. In their case the speed of neuronal functionality is hundreds of times higher than that of the competing solutions, and the communication principles are slightly different (i.e., the non-payload/payload ratio is vastly different), the performance-limiting effect of the ”quantal nature of computing time” persists when used in extensive systems.

5 The role of the workload on the computing efficiency

As was very early predicted AmdahlSingleProcessor67 and decades later experimentally confirmed ScalingParallel:1993 , the scaling of the parallelized computing is not linear. Even, ”there comes a point when using more processors …actually increases the execution time rather than reducing it” ScalingParallel:1993 . Where that point comes, depends on the workload. Paper VeghHowMany:2020 discusses first/second order approaches to explain the issue. The first order approach explains the experienced saturation, and the second order the predicted decrease.

Figure 4: The performance gains of supercomputers modeled as ”roofline” WilliamsRoofline:2009 , as measured with the benchmarks HPL and HPCG (taken from the database TOP500 TOP500:2017 ), and the one for brain simulation is concluded from NeuralNetworkPerformance:2018 .

As VeghHowMany:2020 discusses, the different workloads, mainly due to their different communication-to-computation ratio, work with different efficiency on the same computer system DifferentBenchmarks:2017 . The neuromorphic operation on conventional architectures shows the same issues VeghAIperformance:2020 ; VeghReevaluate:2020 ; VeghScalingANN:2020 . Fig. 4 illustrates how the different workloads cause a saturation in the value of the performance gain. Compared to the benchmark HPL, the HPCG comprises much more communication because of the iterative nature of the task.

Fig. 4

also depicts an estimated efficiency for the case of simulating brain-like operation on a conventional architecture. Given that in 

NeuralNetworkPerformance:2018 , the power consumption efficiency was also investigated, one can presume that (to avoid obsolete energy consumption) the authors measured at the point where involving more cores increased the power consumption but did not increase the payload simulation performance. The performance gain of an AI workload on supercomputers can be estimated to be between those of HPCG and brain simulation; closer to the HPCG gain. As discussed experimentally in DeepNeuralNetworkTraining:2016 and theoretically in VeghAIperformance:2020 , in the case of neural networks (especially in the case of selecting improper layering depth) the efficiency can be much lower. Depending on the architecture, the performance gain reaches the saturation level by using just dozens of cores in the system, mimicking neuromorphic operation on a conventional system.

6 Limitations due to the classic computing paradigm

Figure 5: The limiting effect considered in the ’modern’ theories. One left side, the speed limit, as explained by the theory of relativity, is illustrated. The refractory index of the medium defines the value of the speed limit. On the right side, the payload performance limit of the parallelized sequential computing systems, as explained by the ”modern paradigm”, is illustrated. The ratio of the non-payload to payload processing defines the payload performance.

Mainly, the effect of the shared bus defines the payload performance of the computing systems assembled from components manufactured for SPA systems. The right subfigure in Fig. 5 displays the payload performance of a many-processor SPA system when executing different workloads (that define the non-payload to payload ratio); for the math details see VeghModernParadigm:2019 ; VeghHowMany:2020 . The top diagram lines represent the best payload performance that the supercomputers can achieve when running the benchmark HPL that represents the minimum communication a parallelized sequential system needs. The bottom diagram line represents the estimation of the payload performance that neuromorphic-type processing can achieve in SPA systems. Notice the similarity with the left subfigure: under extreme conditions, in the science, an environment-dependent speed limit exists, and in computing, a workload-dependent payload performance limit exists VeghModernParadigm:2019 .

The careful analysis discovers a remarkable parallel between the proposed ’modern computing’ VeghModernParadigm:2019 versus the classic computing and the modern science versus the classic science. ”Modern computing” does not invalidate the ”classic computing”. Instead, it draws the range of the validity of the classic approximation and sets up the rules of computing under extreme conditions. The ”modern computing” in its field leads to counter-intuitive and shocking conclusions at some extreme parameter values, as did the ”modern science” more than hundred years ago. The parallel can help to accept that what one can not experience in the every-day computing, can be true when using computing under extreme conditions. Fig. 5 depicts one such consequence. In the modern science, unlike in the classic science, a speed limit exists. In the modern computing, unlike in the classic computing, a payload performance limit exists. For further parallels, see VeghModernParadigm:2019 .

Using another computing theory is a must, especially when targeting neuromorphic computing. In the frames of ”classic computing”, as was bitterly admitted NeuralNetworkPerformance:2018 , ”any studies on processes like plasticity, learning, and development exhibited over hours and days of biological time are outside our reach”.

7 A system is not a simple sum of its components

Although it is valid for most systems, that one must not conclude from a feature of a component to the similar feature of the system: the non-linearity discussed above it is especially valid for the large-scale computing systems mimicking neuromorphic operation. We mention two prominent examples here. One can assume that if the time of the operation of a neuron can be shortened, the performance of the whole system gets proportionally better. Two distinct options are to use shorter operands (move less data and to perform less bit manipulations) and to mimic the operation of the neuron in an entirely different way: using quick analog signal processing rather than slow digital calculation.

The so-called HPL-AI benchmark used Mixed Precision555Both names are used rather inconsequentially. On one side, the test itself has not much to do with AI, just uses the operand length common in AI tasks; the benchmark HPL, similarly to AI, is a workload type. On the other side, the Mixed Precision is Half Precision: it is natural that for multiplication twice as long operands are used temporarily. It is a different question that the operations are contracted. MixedPrecisionHPL:2018 rather than Double Precision operands in benchmarking their supercomputer. The name suggests as if in solving AI tasks, the supercomputer can show that peak efficiency. When executing the HPL benchmark, this change resulted in a higher performance merit number. However, as correctly stated in the announcement, ”Achieving a 445 petaflops mixed-precision result on HPL (equivalent to our 148.6 petaflops DP result)”, i.e. the peak DP performance did not change.

We expect that when using half-precision (FP16) rather than double precision (FP64) operands in the calculations, four times less data are transferred and manipulated by the system. The measured power consumption data underpin the statement. However, the computing performance is only 3 times higher than in the case of using 64-bit (FP64) operands. The non-linearity has its effect even in this simple case. In the benchmark, the housekeeping activity also takes time. Even the measured performance data enable us to estimate the execution time with zero precision (FP0) operands, see VeghHowMany:2020 . The performance corresponding to is slightly above 1 EFlops (when making no floating operations, i.e., rather Eops). Another peak performance reported666 when running genomics code on the same supercomputer (by using a mixture of operands with different numerical precision and mostly non-floating point instructions) is 1.88 Eops, corresponding to ; for the scaling of that type of calculations see Fig. 5. Given that those two values refer to a different mixture of instructions, the agreement is more than satisfactory.

Another plausible assumption is that if we use quick analog signal processing to replace the slow digital calculation, as proposed in RecipeMemristor:2020 ; NatureBuildingBrain:2020 , the system gets proportionally quicker. Adding analog components to a digital processor, however, has its price. Given that the digital processor cannot handle resources outside of its world, one must call the OS for help. That help, however, is rather expensive in terms of execution time. The required context switching takes time in the order of executing instructions armContextSwitching:2007 ; Tsafrir:2007 , which greatly increases the total execution time and makes the non-payload to payload ratio much worse.

Although these cases seem to be very different, they share at least the common feature, that they change not only one parameter: they also change the non-payload to payload ratio that defines the efficiency. They have different side-effects: changing the operand length has its effect on the cache behavior, using analog processing needs linking between the analog and the digital processing. However, even those one-parameter changes have a nonlinear effect on the efficiency of the system.

8 Summary

The authors have identified some critical bottlenecks in current computational systems/neuronal networks rendering the conventional computing architectures unadaptable to large (and even medium) sized neuromorphic computing. Built with the segregated processor (SPA, wording from Amdahl AmdahlSingleProcessor67 ), the current systems lack autonomous communication of processors and have an inefficient method of imitating biological systems. To overcome these limitations, the authors introduce a drastically different approach to computing, the Explicitly Many-Processor Approach (EMPA), which can serve as the basis for development as well as specific and practical solutions.


The authors thank Prof. Péter Somogyi for valuable comments on a previous version of the manuscript. Project no. 125547 should have been implemented with the support provided from the National Research, Development and Innovation Fund of Hungary, financed under the K funding scheme. However, at the time of writing the paper, the fund is in 20 month delay with providing the support. Because of this, temporarily, the project is supported by the Kalimános BT.