BRN Discussion Ongoing

Diogenese

Top 20
Evening BrainShit ,

1 , if one could find a more salubrious forum name would be great.
But yes , truely sense & feel ones torment.

2, Not 100% certain , but pretty sure this is a rehashed announcement from some time ago..... with the pressent day date attached , purely to confuse.

Regards,
Esq.
I wonder how long ISL had been playing with Akida before they got the Airforce radar SBIR?

This patent was filed in mid 2021, so 5 months before the announcement.

US11256988B1 Process and method for real-time sensor neuromorphic processing 20210719

1713527356114.png

1713527429102.png


[0003] The invention relates to the general field of deep learning neural networks, which has enjoyed a great deal of success in recent years. This is attributed to more advanced neural architectures that more closely resemble the human brain. Neural networks work with functionalities similar to the human brain. The invention includes both a training cycle and a live (online) operation. The training cycle includes five elements and comprises the build portion of the deep learning process. The training cycle requirements ensure adequate convergence and performance. The live (online) operation includes the live operation of a Spiking Neural Network (SNN) designed by the five steps of the training cycle. The invention is part of a new generation of neuromorphic computing architectures, including Integrated Circuits (IC). This new generation of neuromorphic computing architectures includes IC, deep learning and machine learning.



1. A method of providing real-time sensor neuromorphic processing, the method comprising the steps of:
providing a training cycle and a live operation cycle;
wherein the training cycle includes:
(1) the establishment of a build portion or training cycle of a deep learning process with AI;
wherein the build portion or training cycle begins with the process taking performance requirements in the form of generation of a scenario as inputs;
(2) selecting a sensor model application, with associated performance specifications set forth in the generation of a scenario;
(3) providing a Hi-Fi Radio-Frequency (RF) sensor model which is used to augment any real data for training;
(4) providing a computer model surrogate which is used instead of, or in addition to, a non-surrogate computer model;
(5) the sensor model application being one or more of radar, sonar or LIDAR;
(6) specifying an operating environment details wherein the Hi-Fi sensor model generates requisite training data and/or training environment;
(7) the Hi-Fi sensor model generates training data in a quantity to ensure convergence of DNN neuron weights, wherein as an input enters the node, the input gets multiplied by a weight value and the resulting output is either observed, or passed to the next layer in the neural network;
(8) raw sensor training data is preprocessed into a format suitable for presentation to a DNN from a DNN interface, the sensor model application training data is forwarded to the DNN;
(9) the training environment information is then output from the DNN to a DNN-to-SNN operation through a DNN-to-SNN conversion; and
(10) thereafter, the DNN is converted to an SNN and the SNN outputs the SNN information to an neuromorphic integrated circuit (IC), creating a neuromorphic sensor application;
(11) providing and utilizing a statistical method which ensures reliable performance of the neuromorphic integrated circuit, wherein reliability is 99%
.
 
Last edited:
  • Like
  • Love
  • Fire
Reactions: 14 users

Diogenese

Top 20
Assuming Apple make the decision to pursue their own way and manage to cobble something together which is workable (which is highly likely, given their resources).

What are the chances of them offering this tech, to their competitors, to also use?...

The problems Apple had, with their flagship iPhone 15 overheating last year, goes to show they are maybe not as "hot shit" as they think they are..

Maybe they are literally...

Happy to change my opinion, if they come to their senses 😛
If their potential customers have done their DD, they will be aware of Akida.

Using MACs makes the Apple system much less efficient and slower than Akida.

That would make it a different equation for the potential customers in not having the sunk costs of developing a second-rate in-house system.

As you point out, using the iPhone as a handwarmer did use up the battery quickly.
 
  • Like
  • Fire
  • Haha
Reactions: 10 users

manny100

Regular
If their potential customers have done their DD, they will be aware of Akida.

Using MACs makes the Apple system much less efficient and slower than Akida.

That would make it a different equation for the potential customers in not having the sunk costs of developing a second-rate in-house system.

As you point out, using the iPhone as a handwarmer did use up the battery quickly.
Diogenes, just a query concerning tge original patent that expires in 2028.
It's been improved over and over with several more patents added over the years until we got GEN 2 with GEN 3 next year with possibly another patent application.
How do you see the expiry of the original patent effecting our business?
My view is that its old tech now with ongoing further patents improving the original, but it will still enable others to add their own improvements.
I guess it depends mainly on how much market traction we can gain over the next few years with Edge AI starting to take off.
 
  • Like
  • Thinking
Reactions: 3 users
I wonder how long ISL had been playing with Akida before they got the Airforce radar SBIR?

This patent was filed in mid 2021, so 5 months before the announcement.

US11256988B1 Process and method for real-time sensor neuromorphic processing 20210719

View attachment 61118
View attachment 61119

[0003] The invention relates to the general field of deep learning neural networks, which has enjoyed a great deal of success in recent years. This is attributed to more advanced neural architectures that more closely resemble the human brain. Neural networks work with functionalities similar to the human brain. The invention includes both a training cycle and a live (online) operation. The training cycle includes five elements and comprises the build portion of the deep learning process. The training cycle requirements ensure adequate convergence and performance. The live (online) operation includes the live operation of a Spiking Neural Network (SNN) designed by the five steps of the training cycle. The invention is part of a new generation of neuromorphic computing architectures, including Integrated Circuits (IC). This new generation of neuromorphic computing architectures includes IC, deep learning and machine learning.



1. A method of providing real-time sensor neuromorphic processing, the method comprising the steps of:
providing a training cycle and a live operation cycle;
wherein the training cycle includes:
(1) the establishment of a build portion or training cycle of a deep learning process with AI;
wherein the build portion or training cycle begins with the process taking performance requirements in the form of generation of a scenario as inputs;
(2) selecting a sensor model application, with associated performance specifications set forth in the generation of a scenario;
(3) providing a Hi-Fi Radio-Frequency (RF) sensor model which is used to augment any real data for training;
(4) providing a computer model surrogate which is used instead of, or in addition to, a non-surrogate computer model;
(5) the sensor model application being one or more of radar, sonar or LIDAR;
(6) specifying an operating environment details wherein the Hi-Fi sensor model generates requisite training data and/or training environment;
(7) the Hi-Fi sensor model generates training data in a quantity to ensure convergence of DNN neuron weights, wherein as an input enters the node, the input gets multiplied by a weight value and the resulting output is either observed, or passed to the next layer in the neural network;
(8) raw sensor training data is preprocessed into a format suitable for presentation to a DNN from a DNN interface, the sensor model application training data is forwarded to the DNN;
(9) the training environment information is then output from the DNN to a DNN-to-SNN operation through a DNN-to-SNN conversion; and
(10) thereafter, the DNN is converted to an SNN and the SNN outputs the SNN information to an neuromorphic integrated circuit (IC), creating a neuromorphic sensor application;
(11) providing and utilizing a statistical method which ensures reliable performance of the neuromorphic integrated circuit, wherein reliability is 99%
.
Maybe they were one of the reasons AKD1000 was delayed because early adopters wanted changes such as CNN2SNN converters. Makes sense to me.

SC
 
  • Thinking
  • Like
Reactions: 2 users

Diogenese

Top 20
Diogenes, just a query concerning tge original patent that expires in 2028.
It's been improved over and over with several more patents added over the years until we got GEN 2 with GEN 3 next year with possibly another patent application.
How do you see the expiry of the original patent effecting our business?
My view is that its old tech now with ongoing further patents improving the original, but it will still enable others to add their own improvements.
I guess it depends mainly on how much market traction we can gain over the next few years with Edge AI starting to take off.
Hi manny,

The 2008 patent is of great value in that the main claim is very broad in scope and is directed to ML:

US2010076916A1 Autonomous Learning Dynamic Artificial Neural Computing Device and Brain Inspired System 20080921

An information processing system intended for use in artificial intelligence, consisting of a plurality of artificial neuron circuits connected in an array, comprising:

a first plurality of dynamic synapse circuits, comprising a means of learning and responding to input signals by producing a compounding strength value simulating a biological Post Synaptic Potential,

a temporal integrator circuit that integrates and combines individually simulated Post Synaptic Potential values over time, and thus constitutes an artificial membrane potential value,

a second plurality of dynamic soma circuits each capable of producing one or more pulses when the integrated membrane potential value has reached or exceeded a stored variable threshold value
,

That said, there are a number of subsequent developments which greatly improve performance, chief of these (until TeNNs) is N-of-M coding which greatly increased sparsity and reduced power usage and latency.

Another landmark patent is ...

US11468299B2 Spiking neural network 20181101

1713536435888.png


A neuromorphic integrated circuit, comprising:
a spike converter circuit configured to generate spikes from input data;
a reconfigurable neuron fabric comprising a neural processor comprising a plurality of spiking neuron circuits configured to perform a task based on the spikes and a neural network configuration; and
a memory comprising the neural network configuration, wherein the neural network configuration comprises a potential array and a plurality of synapses, and the neural network configuration defines connections between the plurality of spiking neuron circuits and the plurality of synapses, the potential array comprising membrane potential values for the plurality of spiking neuron circuits, and the plurality of synapses having corresponding synaptic weights,
wherein the neural processor is configured to:
select a spiking neuron circuit in the plurality of spiking neuron circuits based on the selected spiking neuron circuit having a membrane potential value that is a highest value among the membrane potential values for the plurality of spiking neuron circuits;
determine that the membrane potential value of the selected spiking neuron circuit reached a learning threshold value associated with the selected spiking neuron circuit; and
perform a Spike Time Dependent Plasticity (STDP) learning function based on the determination that the membrane potential value of the selected spiking neuron circuit reached the learning threshold value associated with the selected spiking neuron circuit
.


... which showcases my all-time favourite drawing. As you can see, this has a priority of 2018, so it will last well into the 2030s.


CNN2SNN is certainly a vital commercial development, but, for me, not as exciting as N-of-M or TeNNS.

This patent introduced a number of "minor" improvements:

WO2020092691A1 AN IMPROVED SPIKING NEURAL NETWORK 20181101

[0038] But conventional SNNs can suffer from several technological problems. First, conventional SNNs are unable to switch between convolution and fully connected operation. For example, a conventional SNN may be configured at design time to use a fully-connected feedforward architecture to learn features and classify data. Embodiments herein (e.g., the neuromorphic integrated circuit) solve this technological problem by combining the features of a CNN and a SNN into a spiking convolutional neural network (SCNN) that can be configured to switch between a convolution operation or a fully- connected neural network function. The SCNN may also reduce the number of synapse weights for each neuron. This can also allow the SCNN to be deeper (e.g., have more layers) than a conventional SNN with fewer synapse weights for each neuron.
Embodiments herein further improve the convolution operation by using a winner-take-all (WTA) approach for each neuron acting as a filter at particular position of the input space. This can improve the selectivity and invariance of the network. In other words, this can improve the accuracy of an inference operation.

[0039] Second, conventional SNNs are not reconfigurable. Embodiments herein solve this technological problem by allowing the connections between neurons and synapses of a SNN to be reprogrammed based on a user defined configuration. For example, the connections between layers and neural processors can be reprogrammed using a user defined configuration file.

[0040]
Third, conventional SNNs do not provide buffering between different layers of the SNN. But buffering can allow for a time delay for passing output spikes to a next layer. Embodiments herein solve this technological problem by adding input spike buffers and output spike buffers between layers of a SCNN.

[0041] Fourth, conventional SNNs do not support synapse weight sharing. Embodiments herein solve this technological problem by allowing kernels of a SCNN to share synapse weights when performing convolution. This can reduce memory requirements of the SCNN.

[0042] Fifth, conventional SNNs often use l-bit synapse weights. But the use of l-bit synapse weights does not provide a way to inhibit connections. Embodiments herein solve this technological problem by using ternary synapse weights. For example, embodiments herein can use two-bit synapse weights. These ternary synapse weights can have positive, zero, or negative values. The use of negative weights can provide a way to inhibit connections which can improve selectivity. In other words, this can improve the accuracy of an inference operation.

[0043] Sixth, conventional SNNs do not perform pooling. This results in increased memory requirements for conventional SNNs. Embodiments herein solve this technological problem by performing pooling on previous layer outputs. For example, embodiments herein can perform pooling on a potential array outputted by a previous layer. This pooling operation reduces the dimensionality of the potential array while retaining the most important information.

[0044] Seventh, conventional SNN often store spikes in a bit array. Embodiments herein provide an improved way to represent and process spikes. For example, embodiments herein can use a connection list instead of bit array. This connection list is optimized such that each input layer neuron has a set of offset indexes that it must update. This enables embodiments herein to only have to consider a single connection list to update all the membrane potential values of connected neurons in the current layer.

[0045]
Eighth, conventional SNNs often process spike by spike. In contrast, embodiments herein can process packets of spikes. This can cause the potential array to be updated as soon as a spike is processed. This can allow for greater hardware parallelization.

[0046] Finally, conventional SNNs do not provide a way to import learning (e.g., synapse weights) from an external source. For example, SNNs do not provide a way to import learning performed offline using backpropagation. Embodiments herein solve this technological problem by allowing a user to import learning performed offline into the neuromorphic integrated circuit
.


The TeNNs patents:


WO2023250092A1 METHOD AND SYSTEM FOR PROCESSING EVENT-BASED DATA IN EVENT-BASED SPATIOTEMPORAL NEURAL NETWORKS 20220622



WO2023250093A1 METHOD AND SYSTEM FOR IMPLEMENTING TEMPORAL CONVOLUTION IN SPATIOTEMPORAL NEURAL NETWORS 20220622

1713538720562.png



1713538738493.png



Disclosed is a neural network system generally relates to the field of neural networks (NNs). In particular, the present disclosure relates to event-based convolutional neural networks (NNs) that are trained to process spatial and temporal data using kernels represented by polynomial expansion. The event-based convolutional neural networks (NNs) are spatiotemporal neural networks. According to an embodiment, an explicit temporal convolution capability is added through Temporal Event-based Neural Networks (TENN) models, or TENNs in the spatiotemporal neural networks. The TENNs includes a plurality of temporal and spatial convolution layers that combine spatial and temporal features of data for low-level and high-level features. The TENNs as disclosed herein are configured to perform in a buffer mode and recurrent mode that effectively learns both spatial and temporal correlations from the input data.


The FIFO (first-in - first-out) buffer acts like a data conveyor belt which facilitates comparison of incoming data with previously temporarily stored data.
 
  • Love
  • Like
  • Fire
Reactions: 7 users
Top Bottom