Computers That Look Like the Brain

Artificial intelligence applications have been developing rapidly over the past few years, allowing computers to perform complex actions, such as driving without a driver, making decisions, and recognizing faces. These applications require that many calculations be performed in parallel and immense amounts of information are needed. This article demonstrates how inefficient today’s computer structure is for performing artificial intelligence applications. To deal with this challenge and improve artificial intelligence applications, we will see how inspiration from the way the human brain works will allow us to build completely new computers, which will rock the way computers have been built for many years.


STANDARD COMPUTERS
Computers come in di erent shapes and sizes, from the tiny computers in cellphones to supercomputers in huge halls called server farms. Despite the big di erences in the usages and abilities of various computers, their basic structure is similar-they all contain a computation unit called a processor, and a data storage unit The computer is built from a control and computation unit (processor, CPU) and an data storage unit (memory). It also contains input and output units. (B) An example of a computer program adding two numbers. Commands and read both numbers from memory and bring them to the processor, command adds the two numbers in the processor, and command stores the result in memory. called memory. The computation unit performs a collection of MEMORY Memory is a device that is used to store information for immediate use in a computer.
arithmetic operations on numbers that are stored in the memory. In order to perform a complex operation, a collection of simpler operations is written as a computer program, where each operation is a command written as a line of wording in the program. For example, when we want to add two numbers, the computer will actually perform a number of operations including reading each number by bringing it from the memory to the processor, adding the numbers (performed by an electronic component that adds numbers, located inside the processor), and writing/storing the result into a specific location in the memory. The commands themselves are also stored in the memory.
This basic computer structure was first suggested by John von Neumann from Princeton University in , and therefore the basic structure is called von Neumann architecture (Figure ) [ ]. For many

VON NEUMANN ARCHITECTURE
The computer structure common today, which consists of a processor and a storage (memory) unit. The processor controls the system and performs logic and arithmetic operations. The memory stores the program and the data the program reads or writes. years, computers changed their shapes and improved their abilities, both for storage (the amount of information that can be saved), and computation (the number and complexity of operations that are performed in a certain time period). However, the basic structure and principles have not changed since . In today's computers, the processors perform arithmetic operations very quickly. A modern processor can perform more than one billion operations, such as addition and subtraction, in second. Accessing the memory for writing and reading is much slower than processing and requires a lot of energy. Therefore, the maximum e ciency of computers is achieved in applications in which the information is available and accessible to the processor.

ARTIFICIAL INTELLIGENCE IS SHUFFLING THE CARDS
In recent years, artificial intelligence (AI) applications have been

ARTIFICIAL INTELLIGENCE (AI)
"The ability to make a machine behave in a way that would be considered intelligent if a person would behave that way" (Marvin Minsky).
increasing and can be seen in almost every field, including medicine, automotive, security, and manufacturing. The ability to connect many computers together, allowing the storage and computation of huge amounts of information, has enabled the creation of some very impressive AI. In many cases, computers outperform humans. For  Figure Artificial intelligence based on artificial neural networks. The computer predicts whether the received image is of a cat. Each input (IN) receives a numerical value according the color of an associated pixel from the image. The inputs are connected to two outputs (OUT) through connections (called weights W ij ) with a numeric value representing the strength of the connection between each input-output pair.
In this example, output is the probability that the image is a cat and output is the probability that the image is a dog. Every output receives a value depending on the sum of all entrances and weights. Calculation shows that there is a . probability that the image is a cat and a . probability that it is a dog, therefore the prediction is that the image contains a cat. The learning process is the process of updating connection values. In complex AI systems, there are millions of weights and many thousands of inputs and outputs connected in various ways. example, object recognition programs are recognizing objects more accurately than the average person, and automated cars are driving more safely than the human driver.
AI applications are rocking the stability of von Neumann architecture and the computational model that has held on for so many years. AI applications require both enormous amounts of data and computational models that include a huge number of di erent parameters (numbers) corresponding to the data. For example, to perform object identification ( Figure ), millions of images are transferred to the computer, along with the desired tagging result identifying the image (for example, the type of animal). For each example, a training process is performed, in which the result predicted by the computer is compared to the tag. Based on the di erence between the desired result and the result obtained by the computer, the parameters of the computational model are updated according to mathematical rules that were defined in advance. Usually, a large error will result in a larger change in the parameters. After many examples, the parameters of the computational model will enable an accurate result.
We can think of a computer's memory as a warehouse or a library: its size will directly influence the time and energy required to access a specific location. Therefore, for AI applications, the memory must be extremely large, which means accessing information is slow and requires lots of energy. Additionally, predicting results and updating the parameters during training requires a huge number of arithmetic operations, on the scale of a billion addition and multiplication operations for the recognition of a single image. A normal computer requires the processor to perform arithmetic operations one after the other, and in between, the information must be moved between the processor and memory. Since AI applications must perform billions of calculations on millions of di erent number, standard computers are very ine cient and waste huge amounts of time and energy transferring the information, in addition to the slowness of performing many computations one after the other.

THE NEED FOR NEW COMPUTERS
The huge success of AI is driving computer companies to create bigger and stronger computers that can handle the huge amounts of data necessary for e ective AI. To do this, graphical processing units (GPUs) are being used in addition to normal processors. GPUs

GRAPHICAL PROCESSING UNIT
Graphical processing unit (GPU) is a special purpose processor, originally developed in a unique structure to accelerate the creation of images in display devices. Now they are widely used also to accelerate other applications such as AI.
were originally developed to help computers perform calculations on images. In most cases, this requires performing the same calculation operation many times, such as changing the value of all pixels in an image. In AI, GPUs are used to perform a large number (usually a few hundreds or thousands) of addition and multiplication operations in parallel (at the same time), therefore reducing the computation time compared to a standard processor that performs single operations one at a time. Many companies have started producing dedicated processors that can perform the unique operations needed by AI by adding many electronic circuits to perform addition and multiplication operations in parallel. Also, instead of using one very big memory, systems are being built with memories of di erent sizes and in di erent locations, so that data that is used many times will be located in a small, fast memory, and data that is used infrequently will be located in a big, slower memory.
Despite these attempts to modify classical computers to fit the age of AI, the computational method of AI applications is significantly di erent than that of the von Neumann computer. Although memories of di erent sizes are now being used, it is di cult to take advantage of the e ciency of small memories, so the big, slow memories are still often accessed. The new processors created for AI have significantly increased the number of computations performed within a certain time, but huge amounts of data must still be transferred to memory, so a lot of energy and time is still wasted.

NEUROMORPHIC COMPUTERS: COMPUTERS THAT WORK LIKE THE BRAIN
The tasks performed by AI (driving, identification, etc.) are performed in a completely di erent way by the human brain, and with amazing e ciency. The human brain requires less than one tenth the energy of a single processor to perform an operation that requires the connection of thousands of AI processors together. We should try to learn from the brain and build computational models and computers that are similar to the brain in the way they operate. In the brain, there is no separation between data storage and processing. All the information, expressed through electrical signals and chemical changes, is transferred between brain cells called neurons. This basically a change in the connection strength between various neurons. Synapses are very e cient-there is no separation between the calculation of the change and storage of the connection's strength, as there is in current AI (Figure ).

NEURON
To match the world of computing to the needs of AI, we must build computers with a structure inspired by the brain. These computers should have no separation between storage and computation, but instead use units capable of calculating and saving the values simultaneously. Electrical circuits inspired by the brain are called neuromorphic circuits, because they have the shape (morphic) of

NEUROMORPHIC CIRCUIT
Electronic circuits that mimic the behavior of neurons in the human body. Neuromorphic circuits can be combined to create neuromorphic computers.
a neuron (neuro) [ ]. In neuromorphic circuits, the same electronic components are used to both perform a calculation and to store the result locally. The structure of such a computer is completely di erent from a von Neumann machine, and it looks like a neural network in the brain.
Recently, several computers using neuromorphic circuits have been developed [ -], but there are still many challenges before these new computers can replace existing ones. Neuromorphic computers require new materials and devices that can combine calculation and memory, since the devices in standard computers are used only for a single purpose. Additionally, neuromorphic computers need new programs capable of giving the right instructions to the special structure. As von Neumann computer systems are increasingly combined with neuromorphic computers, AI applications will become more e cient, and computers will continue advancing toward human capabilities.

CONFLICT OF INTEREST:
The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
COPYRIGHT © Kvatinsky. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

YOUNG REVIEWERS
ORT EBIN, SCIENCE AND TECHNOLOGY RESERVE CLASS, AGES: -We are a Science and Technological reserve class at the ORT Ebin school in Ramat Gan. ORT Ebin is a high school of sciences and arts. The school promotes leading technological subjects: biotechnology, robotics and computer science-cyber.

SHAHAR KVATINSKY
Shahar Kvatinsky is an Associate Professor at the Andrew and Erna Viterbi Faculty of Electrical Engineering, Technion-Israel Institute of Technology. Shahar received the B.Sc. degree in Computer Engineering and Applied Physics and an MBA degree in and , respectively, both from the Hebrew University of Jerusalem.