@DingoBorat
Your crystal ball working well.
Suspect we get an Ann Mon morn as you say with a new Patent published 26 June.
Just came up and doesn't have all the paperwork on the website yet, just the below.
US2025209313A1
METHOD AND SYSTEM FOR IMPLEMENTING ENCODER PROJECTION IN NEURAL NETWORKS
Bibliographic data
Global Dossier
Applicants
BRAINCHIP INC [US]
Inventors
COENEN PHD OLIVIER JEAN-MARIE DOMINIQUE [US]; PEI YAN RU [US]
Classifications
IPC
G06F17/16; G06N3/048;
CPC
G06F17/16 (US); G06N3/048 (US);
Priorities
US202363614220P·2023-12-22; US202418991246A·2024-12-20
Application
US202418991246A·2024-12-20
Publication
US2025209313A1·2025-06-26
Published as
US2025209313A1
en
METHOD AND SYSTEM FOR IMPLEMENTING ENCODER PROJECTION IN NEURAL NETWORKS
Abstract
Disclosed is a neural network system that includes a memory and a processor. The memory is configured to store a plurality of storage buffers corresponding to a current neural network layer, and implement a neural network that includes a plurality of neurons for the current neural network layer and a corresponding group among a plurality of groups of basis function values. The processor is configured to receive an input data sequence into the first plurality of storage buffers over a first time sequence and project the input data sequence on a corresponding basis function values by performing, for each connection of a corresponding neuron, a dot product of the first input data sequence within a corresponding storage buffer with the corresponding basis function values and thereby determine a corresponding potential value for the corresponding neurons. Thus, utilizing the corresponding potential values, the processor generates a plurality of encoded output responses.
Well, kinda sucks we didn't get the Ann this morning but you'd like to think once they're awake and "working" in the US we might get it late today or tomoz morn?
Anyway, additional details and drawings have now been added to the initial publication I spotted over the weekend.
@Diogenese given the references to Pleiades and Gen AI would it be fair to conclude this patent is around the aforementioned and LLM inroads for the Edge? I see Rudy also one of the inventors so appears likely to me.
[0004] Recently different methods have been explored to generate AI-based content using the aforementioned NNs. It is critical for some applications for a neural network to be able to generate content. The neural networks that are capable of generating the AI-based content are called generative models or generative networks. One of the commonly known generative networks is RNNs and the other commonly known generative networks are transformers. The RNNs were traditionally used to generate the content before the emergence of the transformers.
Transformers and variants have been at the basis of the advances in generative models, particularly in the domain of large language models (LLMs).
[0005] For each of the above-discussed NN models including ANN, CNN, and RNN, the computation process is very often performed in the cloud for generating the content. However, in order to have a better user experience and privacy, and for various commercial reasons, an implementation of the computation process has started moving from the cloud to edge devices. In order to generate AI-based content, there are mainly two solutions available in the state of the art i.e., RNNs and transformers. However, RNNs are difficult to train because of the recurrence they take more time to train. Transformers generate content without having to make use of recurrence, which permits parallelized training. The transformers are capable of being trained efficiently in the cloud by leveraging Graphics Processing Unit (GPU) or Tensor Processing Unit (TPU) for parallel computation.
[0006] Further, with the increasing complexity of the NN models, there is a corresponding increase in the computational requirements required to execute highly complex NN models, for example, the transformer based models. Thus, a huge computational processing and a large memory are required for executing highly complex transformer based models.
[0007]
Thus, there lies a need for a method and system to reduce the computational requirements of the above-discussed NN models while still meeting desired accuracy expectations, in order to facilitate more efficient content generation, particularly for the edge devices.
SUMMARY
[0008] This summary is provided to introduce a selection of concepts in a simplified form that is further described below in the Detailed Description section. This summary is not intended to identify or exclude key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
[0009]
For generating content, the neurons of NN models, such as polynomial expansion in adaptive distributed event-based systems (PLEAIDES) models, perform a temporal convolution operation on input signals with the temporal kernels. In some aspects, the temporal kernels are represented as an expansion over a basis function with kernel coefficients Gi . In some aspects, the kernel coefficients are trainable parameters of the neural network during learning. In some aspects, the kernel coefficients remain constant while generating content using convolutions during inference. Even though the recurrent mode of PLEIADES decomposes the convolution with a kernel onto a set of basis functions, the contribution from each may not be used individually, but summed together to provide a scalar value of the convolution. Such a scalar value has more limited power in generating signals than if a contribution, coefficient, from each basis could be used