In many ways computers today are nothing more than very fast number-crunchers and information manipulators. They can process lots of data, but they really don’t think. They all adhere to the Von Neumann architecture, largely unchanged in the last half-century, in which computers are constructed by separating memory and processing and operate by executing a series of pre-written “if X then do Y” equations.
With the advent of Big Data, which grows larger, faster and more diverse by the day, this type of computing model is inadequate to process and make sense of the volumes of information that people and organizations need to deal with.
The cognitive capabilities of the brain includes understanding the surrounding environment, dealing with ambiguity, acting in real time and within context – all while consuming less than power than a light bulb and occupying less space than a two-liter bottle of soda. These new silicon, neurosynaptic chips allow for computing systems that emulate the brain’s computing efficiency, size and power usage.
Each corelet has a particular function that can be put together in different configurations to create new applications. For example, a corelet could include all of the individual cores that perceive sound. The programmer could use that corelet in conjunction with others that represent edge detection and color identification to develop a new application that takes advantage of all those features.
Asst. Professor, ISE Dept
Brindavan College of Engineering