BRN Discussion Ongoing

Tothemoon24

Top 20

LEADERS

The Future of Generative AI Is the Edge​

mm

Published
October 19, 2023
By
Ravi Annavajjhala
edge-computing.png

The advent of ChatGPT, and Generative AI in general, is a watershed moment in the history of technology and is likened to the dawn of the Internet and the smartphone. Generative AI has shown limitless potential in its ability to hold intelligent conversations, pass exams, generate complex programs/code, and create eye-catching images and video. While GPUs run most Gen AI models in the cloud – both for training and inference – this is not a long-term scalable solution, especially for inference, owing to factors that include cost, power, latency, privacy, and security. This article addresses each of these factors along with motivating examples to move Gen AI compute workloads to the edge.
Most applications run on high-performance processors – either on device (e.g., smartphones, desktops, laptops) or in data centers. As the share of applications that utilize AI expands, these processors with only CPUs are inadequate. Furthermore, the rapid expansion in Generative AI workloads is driving an exponential demand for AI-enabled servers with expensive, power-hungry GPUs that in turn, is driving up infrastructure costs. These AI-enabled servers can cost upwards of 7X the price of a regular server and GPUs account for 80% of this added cost.
Additionally, a cloud-based server consumes 500W to 2000W, whereas an AI-enabled server consumes between 2000W and 8000W – 4x more! To support these servers, data centers need additional cooling modules and infrastructure upgrades – which can be even higher than the compute investment. Data centers already consume 300 TWH per year, almost 1% of the total worldwide power consumption. If the trends of AI adoption continue, then as much as 5% of worldwide power could be used by data centers by 2030. Additionally, there is an unprecedented investment into Generative AI data centers. It is estimated that data centers will consume up to $500 billion for capital expenditures by 2027, mainly fueled by AI infrastructure requirements.
powerconsumption.png

The electricity consumption of Data centers, already 300 TwH, will go up significantly with the adoption of generative AI.
AI compute cost as well as energy consumption will impede mass adoption of Generative AI. Scaling challenges can be overcome by moving AI compute to the edge and using processing solutions optimized for AI workloads. With this approach, other benefits also accrue to the customer, including latency, privacy, reliability, as well as increased capability.

Compute follows data to the Edge

Ever since a decade ago, when AI emerged from the academic world, training and inference of AI models has occurred in the cloud/data center. With much of the data being generated and consumed at the edge – especially video – it only made sense to move the inference of the data to the edge thereby improving the total cost of ownership (TCO) for enterprises due to reduced network and compute costs. While the AI inference costs on the cloud are recurring, the cost of inference at the edge is a one-time, hardware expense. Essentially, augmenting the system with an Edge AI processor lowers the overall operational costs. Like the migration of conventional AI workloads to the Edge (e.g., appliance, device), Generative AI workloads will follow suit. This will bring significant savings to enterprises and consumers.
The move to the edge coupled with an efficient AI accelerator to perform inference functions delivers other benefits as well. Foremost among them is latency. For example, in gaming applications, non-player characters (NPCs) can be controlled and augmented using generative AI. Using LLM models running on edge AI accelerators in a gaming console or PC, gamers can give these characters specific goals, so that they can meaningfully participate in the story. The low latency from local edge inference will allow NPC speech and motions to respond to players' commands and actions in real-time. This will deliver a highly immersive gaming experience in a cost effective and power efficient manner.
In applications such as healthcare, privacy and reliability are extremely important (e.g., patient evaluation, drug recommendations). Data and the associated Gen AI models must be on-premise to protect patient data (privacy) and any network outages that will block access to AI models in the cloud can be catastrophic. An Edge AI appliance running a Gen AI model purpose built for each enterprise customer – in this case a healthcare provider – can seamlessly solve the issues of privacy and reliability while delivering on lower latency and cost.
Picture2.png

Generative AI on edge devices will ensure low latency in gaming and preserve patient data and improve reliability for healthcare.
Many Gen AI models running on the cloud can be close to a trillion parameters – these models can effectively address general purpose queries. However, enterprise specific applications require the models to deliver results that are pertinent to the use case. Take the example of a Gen AI based assistant built to take orders at a fast-food restaurant – for this system to have a seamless customer interaction, the underlying Gen AI model must be trained on the restaurant’s menu items, also knowing the allergens and ingredients. The model size can be optimized by using a superset Large Language Model (LLM) to train a relatively small, 10-30 billion parameter LLM and then use additional fine tuning with the customer specific data. Such a model can deliver results with increased accuracy and capability. And given the model’s smaller size, it can be effectively deployed on an AI accelerator at the Edge.

Gen AI will win at the Edge

There will always be a need for Gen AI running in the cloud, especially for general-purpose applications like ChatGPT and Claude. But when it comes to enterprise specific applications, such as Adobe Photoshop’s generative fill or Github copilot, Generative AI at Edge is not only the future, it’s also the present. Purpose-built AI accelerators are the key to making this possible.



The Author

Ravi Annavajjhala
As a Silicon Valley veteran, and CEO of Kinara Inc, Ravi Annavajjhala brings more than 20 years of experience spanning business development, marketing, and engineering, building leading-edge technology products and
bringing them to market. In his current role as chief executive officer of Deep Vision, Ravi serves on
its board of directors and has raised $50M taking the company’s Ara-1 processor from pre-silicon to
full-scale production and to ramp the 2nd generation processor, Ara-2, in volume. Prior to joining
Deep Vision, Ravi held executive leadership positions at Intel and SanDisk where he played key roles
in driving revenue growth, evolving strategic partnerships, and developing product roadmaps that
led the industry with cutting-edge features and capabilities.
 
  • Like
  • Fire
  • Love
Reactions: 37 users
Not sure if posted before.

Was on AKD1000.

Degree project

Degree Project in Computer Science and Engineering, First Cycle 15 credits
Date: June 8, 2023
Supervisor: Jörg Conradt
Examiner: Pawel Herman
Swedish title: Medicinsk bildanalys med neuromorfisk hårdvara: On-Edge-träning med Akida
Brainchip
School of Electrical Engineering and Computer Science


Degree Project in Computer Science and Engineering
First Cycle, 15 credits
Neuromorphic Medical Image
Analysis at the Edge
On-Edge training with the Akida Brainchip

Future work lists some additional considerations on external inputs to refine results.

Paper HERE
 
  • Like
  • Love
  • Fire
Reactions: 20 users

IloveLamp

Top 20
Very interesting like from Nikunj.......... 🤔😃

Then again............ there's lots to like!!!


LINK TO ARTICLE -
The really astonishing thing is it can apparently outperform GPUs and CPUs specifically designed to tackle AI inference. For example, Numenta took a workload for which Nvidia reported performance figures with its A100 GPU, and ran it on an augmented 48-core 4th-Gen Sapphire Rapids CPU. In all scenarios, it was faster than Nvidia’s chip based on total throughput. In fact, it was 64 times faster than a 3rd-Gen Intel Xeon processor and ten times faster than the A100 GPU.


1000006912.png

1000006916.png

1000006914.png
 
Last edited:
  • Like
  • Fire
  • Thinking
Reactions: 27 users

jtardif999

Regular
  • Haha
  • Fire
  • Like
Reactions: 7 users

Diogenese

Top 20
Very interesting like from Nikunj.......... 🤔😃

Then again............ there's lots to like!!!


LINK TO ARTICLE -
The really astonishing thing is it can apparently outperform GPUs and CPUs specifically designed to tackle AI inference. For example, Numenta took a workload for which Nvidia reported performance figures with its A100 GPU, and ran it on an augmented 48-core 4th-Gen Sapphire Rapids CPU. In all scenarios, it was faster than Nvidia’s chip based on total throughput. In fact, it was 64 times faster than a 3rd-Gen Intel Xeon processor and ten times faster than the A100 GPU.


View attachment 47693
View attachment 47696
View attachment 47697
Numenta does not use spikes. They are addicted to AI fast food - MACs:

US2022108157A1 HARDWARE ARCHITECTURE FOR INTRODUCING ACTIVATION SPARSITY IN NEURAL NETWORK

1697931296942.png



[0072] A multiply circuit 330 may take various forms. In one embodiment, a multiply circuit 330 is a multiply-accumulate circuit (MAC) that includes multiply units and accumulators. The multiply units may be used to perform multiplications and additions. A multiply unit is a circuit with a known structure and may be used for binary multiplication or floating-point multiplication. An accumulator is a memory circuit that receives and stores values from the multiply units. The values may be stored individually or added together in the accumulator.
 
  • Like
  • Haha
  • Love
Reactions: 19 users

IloveLamp

Top 20
Numenta does not use spikes. They are addicted to AI fast food - MACs:

US2022108157A1 HARDWARE ARCHITECTURE FOR INTRODUCING ACTIVATION SPARSITY IN NEURAL NETWORK

View attachment 47698


[0072] A multiply circuit 330 may take various forms. In one embodiment, a multiply circuit 330 is a multiply-accumulate circuit (MAC) that includes multiply units and accumulators. The multiply units may be used to perform multiplications and additions. A multiply unit is a circuit with a known structure and may be used for binary multiplication or floating-point multiplication. An accumulator is a memory circuit that receives and stores values from the multiply units. The values may be stored individually or added together in the accumulator.
For me -

Brainchip employee liking Senior designer @microsoft post is positive

Brainchip employee Liking post about intels advances is also positive

I checked out Numentas LinkedIn page they advertise themselves as a SOFTWARE company.

If you read the article posted, they attribute the gains to the numenta software imo

Perhaps Intel is using their software to integrate akida ip and claiming the advantages are from Numenta when actually the advantages will come from our ip?

1000006918.png


The performance and efficiency advances they're claiming make me highly suspicious

Not the longest of bows to draw imo

Waiting fot the intel / SiFive ip agreement like...

1000006917.gif
 
Last edited:
  • Like
  • Haha
  • Fire
Reactions: 14 users

Earlyrelease

Regular
Numenta does not use spikes. They are addicted to AI fast food - MACs:

Diogenese. On behalf of the silent majority that lurk on these pages can I just call out what a magnificent contribution you bring to these pages. Your input saves us non engineering mortals, hours of research and angst. Thanks for the supreme effort. Drinks are on me when we have out $3 party late next year.
 
  • Like
  • Fire
  • Love
Reactions: 13 users

Diogenese

Top 20
For me -

Brainchip employee liking Senior designer @microsoft post = half win

Brainchip employee Liking post about intels advances = half win

I checked out Numentas LinkedIn page they advertise themselves as a SOFTWARE company.

If you read the article posted, they attribute the gains to the numenta software imo

Perhaps Intel is using their software to integrate akida ip and claiming the advantages are from Numenta when actually the advantages will come from our ip?

View attachment 47700

The performance and efficiency advances they're claiming make me highly suspicious

Not the longest of bows to draw imo

Waiting fot the intel ip agreement like...

View attachment 47699
Hi Ill,

Yes. Numenta make a big thing about sparsity. They use a winner-take-all approach which ignores the lower probability possibilities from other neurons.

If Numenta's patents are indicative of their software architecture, their system will use a lot more power than Akida. In fact, Akida could greatly improve both the power efficiency and speed of their system. Their patent even allows for the possibility of different AI accelerators. Unfortunately, Numenta think ML is software, so if Akida were to be used, Numenta would be largely redundant.

US2022108156A1 HARDWARE ARCHITECTURE FOR PROCESSING DATA IN SPARSE NEURAL NETWORK 20201005

1697938198834.png


[0042] AI accelerator 104 may be a processor that is efficient at performing certain machine learning operations such as tensor multiplications, convolutions, tensor dot products, etc. In various embodiments, AI accelerator 104 may have different hardware architectures. For example, in one embodiment, AI accelerator 104 may take the form of field-programmable gate arrays (FPGAs). In another embodiment, AI accelerator 104 may take the form of application-specific integrated circuits (ASICs), which may include circuits along or circuits in combination with firmware.
...
[0048] Machine learning models 140 may include different types of algorithms for making inferences based on the training of the models. Examples of machine learning models 140 include regression models, random forest models, support vector machines (SVMs) such as kernel SVMs, and artificial neural networks (ANNs) such as convolutional network networks (CNNs), recurrent network networks (RNNs), autoencoders, long short term memory (LSTM), reinforcement learning (RL) models. Some of the machine learning models may include a sparse network structure whose detail will be further discussed with reference to FIG. 2B through 2D. A machine learning model 140 may be an independent model that is run by a processor. A machine learning model 140 may also be part of a software application 130 . Machine learning models 140 may perform various tasks.

Numenta have their foot in Intel's door:

Home | Numenta

1697938698669.png


... but that may not prevent Akida from coming down the chimney.
 
Last edited:
  • Like
  • Love
  • Fire
Reactions: 31 users

IloveLamp

Top 20
Hi Ill,

Yes. Numenta make a big thing about sparsity. They use a winner-take-all approach which ignores the lower probability possibilities from other neurons.

If Numenta's patents are indicative of their software architecture, their system will use a lot more power than Akida. In fact, Akida could greatly improve both the power efficiency and speed of their system. Their patent even allows for the possibility of different AI accelerators. Unfortunately, Numenta think ML is software, so if Akida were to be used, Numenta would be redundant.

US2022108156A1 HARDWARE ARCHITECTURE FOR PROCESSING DATA IN SPARSE NEURAL NETWORK 20201005

View attachment 47706

[0042] AI accelerator 104 may be a processor that is efficient at performing certain machine learning operations such as tensor multiplications, convolutions, tensor dot products, etc. In various embodiments, AI accelerator 104 may have different hardware architectures. For example, in one embodiment, AI accelerator 104 may take the form of field-programmable gate arrays (FPGAs). In another embodiment, AI accelerator 104 may take the form of application-specific integrated circuits (ASICs), which may include circuits along or circuits in combination with firmware.
...
[0048] Machine learning models 140 may include different types of algorithms for making inferences based on the training of the models. Examples of machine learning models 140 include regression models, random forest models, support vector machines (SVMs) such as kernel SVMs, and artificial neural networks (ANNs) such as convolutional network networks (CNNs), recurrent network networks (RNNs), autoencoders, long short term memory (LSTM), reinforcement learning (RL) models. Some of the machine learning models may include a sparse network structure whose detail will be further discussed with reference to FIG. 2B through 2D. A machine learning model 140 may be an independent model that is run by a processor. A machine learning model 140 may also be part of a software application 130 . Machine learning models 140 may perform various tasks.

Numenta have their foot in Intel's door:

Home | Numenta

View attachment 47707

... but that may not prevent Akida from coming down the chimney.
Thanks for explaining in a way even i can understand 😚

1000006921.gif
 
  • Haha
  • Like
  • Fire
Reactions: 12 users

Neuromorphia

fact collector
Last edited:
  • Like
  • Haha
  • Love
Reactions: 16 users

IloveLamp

Top 20
Cheif data scientist at the Indian Institute of Technology gives us a nice mention here along side all the big players......😌


1000006940.png
1000006942.png
 
  • Like
  • Fire
  • Love
Reactions: 60 users

IloveLamp

Top 20
  • Like
  • Fire
  • Thinking
Reactions: 31 users

TopCat

Regular
Infineon using snn for radar...

Akida?

View attachment 47721
I’d like it to be Akida, but from other papers I’ve come across I think Infineon research with spinnaker as the photo suggests. But maybe they could use Akida as it’s commercially available 🤞🤞
 
  • Like
  • Thinking
  • Wow
Reactions: 13 users

TECH

Regular
Hang on to your hats, our 4C will more than likely be released this coming week.

I have already placed my bet, but would be over the moon to be 100% wrong !....the journey continues.

Another thing to remember is that it's taken us this long to achieve what's already been achieved through hard work, self discipline
and founders who are at the top of their game, they have both been very flexible, employed first class staff and let staff go as the need
arose, so if we are, say 3 years ahead (conservative maybe) of other companies attempting to achieve excellence at the far edge, well they
to will be faced with a similar timeline in my view, dispite being large corporations, time will prove me right or wrong I guess.

See you during the week.....Tech (NZ)
 
  • Like
  • Fire
  • Thinking
Reactions: 24 users
I’d like it to be Akida, but from other papers I’ve come across I think Infineon research with spinnaker as the photo suggests. But maybe they could use Akida as it’s commercially available 🤞🤞
Yeah true but remember we have just been in the hackathon with Infineon and Sony for the pedestrian detection.

Would be great if Infineon had some kind of revelation with us.

I know Nikunj speaks about how sensors from Sony or Infineon can be used with Akida in his TinyML video on the hackathon.




 
  • Like
  • Fire
  • Love
Reactions: 48 users

IloveLamp

Top 20
Yeah true but remember we have just been in the hackathon with Infineon and Sony for the pedestrian detection.

Would be great if Infineon had some kind of revelation with us.

I know Nikunj speaks about how sensors from Sony or Infineon can be used with Akida in his TinyML video on the hackathon.





Ooh i forgot about that! Nice one fm


1000006581.gif
 
  • Like
  • Haha
  • Fire
Reactions: 14 users

TechGirl

Founding Member

WHAT IS BRAINCHIP USED FOR?​

What is Brainchip used for?

Understanding the Applications of Brainchip: Revolutionizing Artificial Intelligence​

In recent years, the field of artificial intelligence (AI) has witnessed remarkable advancements, pushing the boundaries of what was once thought possible. One such groundbreaking technology that has emerged is Brainchip, a neuromorphic computing platform that mimics the functioning of the human brain. With its unique architecture and capabilities, Brainchip is being utilized in a multitude of applications, revolutionizing various industries.
At its core, Brainchip is designed to process vast amounts of data in real-time, enabling rapid decision-making and pattern recognition. This innovative technology utilizes spiking neural networks (SNNs), which are inspired by the way neurons communicate in the human brain. By leveraging SNNs, Brainchip can efficiently process and analyze complex data sets, making it ideal for applications that require high-speed and low-power processing.

One of the primary applications of Brainchip lies in the field of surveillance and security. Traditional video surveillance systems often struggle to analyze and interpret large volumes of video footage in real-time. However, Brainchip’s advanced capabilities allow it to process video streams in real-time, detecting and identifying objects, faces, and even abnormal behavior. This technology has the potential to revolutionize the way security systems operate, enhancing public safety and reducing response times.
Another significant application of Brainchip is in the field of autonomous vehicles. Self-driving cars rely heavily on AI algorithms to navigate and make split-second decisions. Brainchip’s ability to process data in real-time and recognize patterns makes it an ideal solution for autonomous vehicles. By leveraging Brainchip’s capabilities, self-driving cars can analyze sensor data, detect obstacles, and make informed decisions, ultimately improving safety and efficiency on the roads.
Furthermore, Brainchip is also being utilized in the healthcare industry. Medical professionals are constantly faced with vast amounts of patient data that need to be analyzed accurately and efficiently. Brainchip’s high-speed processing and pattern recognition capabilities enable healthcare providers to quickly analyze medical images, detect anomalies, and diagnose diseases. This technology has the potential to revolutionize medical diagnostics, leading to faster and more accurate diagnoses, ultimately saving lives.
It is important to note that Brainchip is not limited to these applications alone. Its versatility and adaptability make it suitable for a wide range of industries, including robotics, industrial automation, and even gaming. As the technology continues to evolve, we can expect to see Brainchip being integrated into various sectors, transforming the way we live and work.

In conclusion, Brainchip is a revolutionary technology that is transforming the field of artificial intelligence. With its ability to process vast amounts of data in real-time and recognize patterns, Brainchip is being utilized in various applications, including surveillance and security, autonomous vehicles, healthcare, and more. As this technology continues to advance, it holds the potential to reshape industries and improve our daily lives in ways we never thought possible.
Sources:
– “Brainchip: The World’s First Neuromorphic Computing Platform” – Brainchip Holdings Ltd.
– “Brainchip Technology: A New Era of AI” – Analytics Insight
– “How Brainchip Works: The Future of AI” – TechRadar
 
  • Like
  • Love
  • Fire
Reactions: 90 users
IBM came out of stealth mode today with NorthPole, an extension of TrueNorth…



19 Oct 2023
News
6 minute read

A new chip architecture points to faster, more energy-efficient AI​

A new chip prototype from IBM Research’s lab in California, long in the making, has the potential to upend how and where AI is used efficiently.

image

A new chip prototype from IBM Research’s lab in California, long in the making, has the potential to upend how and where AI is used efficiently.

We’re in the midst of a Cambrian explosion in AI. Over the last decade, AI has gone from theory and small tests to enterprise-scale use cases. But the hardware used to run AI systems, although increasingly powerful, was not designed with today’s AI in mind. As AI systems scale, the costs skyrocket. And Moore’s Law, the theory that the density of circuits in processors would double each year, has slowed.

But new research out of IBM Research’s lab in Almaden, California, nearly two decades in the making, has the potential to drastically shift how we can efficiently scale up powerful AI hardware systems.

Since the birth of the semiconductor industry, computer chips have primarily followed the same basic structure, where the processing units and the memory storing the information to be processed are stored discretely. While this structure has allowed for simpler designs that have been able to scale well over the decades, it’s created what’s called the von Neumann bottleneck, where it takes time and energy to continually shuffle data back and forth between memory, processing, and any other devices within a chip. The work by IBM Research’s Dharmendra Modha and his colleagues aims to change this, taking inspiration from how the brain computes. “It forges a completely different path from the von Neumann architecture,” according to Modha.

Over the last eight years, Modha has been working on a new type of digital AI chip for neural inference, which he calls NorthPole. It’s an extension of TrueNorth, the last brain-inspired chip that Modha worked on prior to 2014. In tests on the popular ResNet-50 image recognition and YOLOv4 object detection models, the new prototype device has demonstrated higher energy efficiency, higher space efficiency, and lower latency than any other chip currently on the market, and is roughly 4,000 times faster than TrueNorth.

The first promising set of results from NorthPole chips were published today in Science. NorthPole is a breakthrough in chip architecture that delivers massive improvements in energy, space, and time efficiencies, according to Modha.
Using the ResNet-50 model as a benchmark, NorthPole is considerably more efficient than common 12-nm GPUs and 14-nm CPUs. (NorthPole itself is built on 12 nm node processing technology.) In both cases, NorthPole is 25 times more energy efficient, when it comes to the number of frames interpreted per joule of power required. NorthPole also outperformed in latency, as well as space required to compute, in terms of frames interpreted per second per billion transistors required. According to Modha, on ResNet-50, NorthPole outperforms all major prevalent architectures — even those that use more advanced technology processes, such as a GPU implemented using a 4 nm process.

How does it manage to compute with so much efficiency than existing chips? One of the biggest differences with NorthPole is that all of the memory for the device is on the chip itself, rather than connected separately. Without that von Neumann bottleneck, the chip can carry out AI inferencing considerably faster than other chips already on the market. NorthPole was fabricated with a 12-nm node process, and contains 22 billion transistors in 800 square millimeters. It has 256 cores and can perform 2,048 operations per core per cycle at 8-bit precision, with potential to double and quadruple the number of operations with 4-bit and 2-bit precision, respectively. “It’s an entire network on a chip,” Modha said.

IBM_NP_PCIe-PCB-Rear.png
The NorthPole chip on a PCIe card.

Architecturally, NorthPole blurs the boundary between compute and memory,” Modha said. "At the level of individual cores, NorthPole appears as memory-near-compute and from outside the chip, at the level of input-output, it appears as an active memory.” This makes NorthPole easy to integrate in systems and significantly reduces load on the host machine.

But the biggest advantage of NorthPole is also a constraint: it can only easily pull from the memory it has onboard. All of the speedups that are possible on the chip would be undercut if it had to access information from another place.
Via an approach called scale-out, NorthPole can actually support larger neural networks by breaking them down into smaller sub-networks that fit within NorthPole’s model memory, and connecting these sub-networks together on multiple NorthPole chips. So while there is ample memory on a NorthPole (or collectively on a set of NorthPoles) for many of the models that would be useful for specific applications, this chip is not meant to be a jack of all trades. “We can’t run GPT-4 on this, but we could serve many of the models enterprises need,” Modha said . “And, of course, NorthPole is only for inferencing.”
This efficacy means that the device also doesn’t need bulky liquid-cooling systems to run — fans and heat sinks are more than enough — meaning that it could be deployed in some rather small spaces.


Potential applications for NorthPole​

While research into the NorthPole chip is still ongoing, its structure lends itself to emerging AI use cases, as well as more well-established ones.

In testing, NorthPole team focused primarily on computer vision-related uses, in part because funding for the project came from the U.S. Department of Defense. Some of the primary applications in consideration were detection, image segmentation, and video classification. But it was also tested in other arenas, such as natural language processing (on the encoder-only BERT model) and speech recognition (on the DeepSpeech2 model). The team is currently exploring mapping decoder-only large language models to NorthPole scale-out systems.

When you think of these AI tasks, all sorts of fantastical use cases spring to mind, from autonomous vehicles, to robotics, digital assistants, or spatial computing. Many sorts of edge applications that require massive amounts of data processing in real time could be well-suited for NorthPole. For example, it could potentially be the sort of device that’s needed to move autonomous vehicles from machines that require set maps and routes to operate on a small scale, to ones that can think and react to the rare edge-case situations that make navigating in the real world so challenging even for proficient human drivers. These sorts of edge-cases are the exact sweet spot for future NorthPole applications. NorthPole could enable satellites that monitor agriculture and manage wildlife populations, monitor vehicle and freight for safer and less congested roads, operate robots safely, and detect cyber threats for safer businesses.

What’s next

This is just the start of the work for Modha on NorthPole. The current state of the art for CPUs is 3 nm — and IBM itself is already years into research on 2 nm nodes. That means there’s a handful of generations of chip processing technologies NorthPole could be implemented on, in addition to fundamental architectural innovations, to keep finding efficiency and performance gains.

BIC-Group-Photo_2023-08-10_no-caption.png
Modha, center, with most of the team working on NorthPole.

But for Modha, this is just one important milestone along a continuum that has dominated the last 19 years of his professional career. He’s been working on digital brain-inspired chips throughout that time, knowing that the brain is the most energy-efficient processor we know, and searching for ways to replicate that digitally. TrueNorth was fully inspired by the structures of neurons in the brain — and had as many digital “synapses” in it as the brain of a bee. But sitting on a park bench in 2015 in San Francisco, Modha said he was thinking through his work to date. He had the belief that there was something in marrying the best of traditional processing devices with the structure of processing in the brain, where memory and processing are interspersed throughout the brain. The answer was “brain-inspired computing, with silicon speed,” according to Modha.

Over the next eight years, Modha and his colleagues were single-minded and hermetic in their goal of turning this vision into a reality. Toiling inconspicuously in Almaden, the team didn’t give any lectures or publish any papers on their work, until this year. Each person brought different skills and perspective yet everyone collaborated so that as a whole the team’s contribution was much greater than the sum of the parts. Now, the plan is to show what NorthPole could do, while exploring how to translate the designs into smaller chip production processes and further exploring the architectural possibilities.

This work stemmed from simple ideas — how can we make computers that work like the brain — and after years of fundamental research, has come up with an answer. Something that is really only possible today at a place like IBM Research, where there is the time and space to explore the big questions in computing, and where they can take us. “NorthPole is a faint representation of the brain in the mirror of a silicon wafer,” Modha said.


Here is a 61 page PDF file for the techies…





View attachment 47481

Biblical level praise from Italy

IMG_6779.jpeg


IMG_6778.jpeg

IMG_6777.jpeg




 
  • Like
  • Fire
  • Love
Reactions: 52 users

Mt09

Regular
For me -

Brainchip employee liking Senior designer @microsoft post is positive

Brainchip employee Liking post about intels advances is also positive

I checked out Numentas LinkedIn page they advertise themselves as a SOFTWARE company.

If you read the article posted, they attribute the gains to the numenta software imo

Perhaps Intel is using their software to integrate akida ip and claiming the advantages are from Numenta when actually the advantages will come from our ip?

View attachment 47700

The performance and efficiency advances they're claiming make me highly suspicious

Not the longest of bows to draw imo

Waiting fot the intel / SiFive ip agreement like...

View attachment 47699
 

Attachments

  • IMG_5117.jpeg
    IMG_5117.jpeg
    74.1 KB · Views: 58
  • IMG_5118.jpeg
    IMG_5118.jpeg
    47 KB · Views: 47
  • Haha
  • Like
Reactions: 3 users

IloveLamp

Top 20
Show me how it's done oh great one, or should i take note from all your high quality offerings? Rhetorical question.

How about you contribute something before critisising others that try?

In fact let's revisit my aim in 18 months, my arrows are still in the air for the most part.
 
Last edited:
  • Like
  • Love
  • Fire
Reactions: 23 users
Top Bottom