BRN Discussion Ongoing

7für7

Top 20
How is the situation actually regarding licenses in the tech sector?
Specifically for MegaChips – how long do they have to use the IP before the license expires? If they have something going on with Akida, wouldn’t they have to move seamlessly into an extension? Or could a new agreement possibly be treated as a kind of grace period during the transition phase while the details are being negotiated?
 
  • Thinking
  • Like
Reactions: 2 users

Rach2512

Regular
 
  • Love
  • Like
  • Fire
Reactions: 8 users
Manny100

The Sony chip does not support on chip learning. That is the key and why its 2nd or 3rd rate compared to AKIDA.
The guts of the difference between on chip learning and a Sony pre trained chip example - AKIDA can recognise new faces (one shot learning) where as Sony can only recognise faces already pretrained on the chip.
So its really only a matter of time before Sony need AKIDA just to keep up with what consumers want.
If they move into robotics, wearables, or adaptive consumer devices, on‑chip learning its AKIDA they need.
I cannot see a world where any of the majors would not be looking at AKIDA right now. Will any let the competition steal the market?
Some simple examples that only AKIDA offers - cameras only - plenty of other examples:
Smart security cameras
: Learn to recognize new faces, pets, or vehicles locally without cloud retraining.
Professional cameras: Adapt autofocus and exposure to a photographer’s unique shooting style.
Consumer Cameras: Personalize scene recognition (e.g., learning your preferred portrait style).
 
  • Like
  • Fire
  • Love
Reactions: 8 users
Appears another GitHub user playing with Akida. I believe it was probs discovered sometime ago maybe as the first repositories are from last year but maybe not..haven't searched to check.


Jalil32/skin-cancer-detection

Skin Cancer Classifier using convolutional neural network deployed to a neuromorphic processor (Akida from BrainChip). Built using Tensorflow.

  • Jupyter Notebook
  • ·
  • 0
  • ·
  • Updated
    6 hours ago
Wondering if one of the two contributors could be this person who works for



Jalil Inayat-Hussain​

Visagio The University of Western Australia​

 
  • Like
  • Love
  • Fire
Reactions: 14 users
How is the situation actually regarding licenses in the tech sector?
Specifically for MegaChips – how long do they have to use the IP before the license expires? If they have something going on with Akida, wouldn’t they have to move seamlessly into an extension? Or could a new agreement possibly be treated as a kind of grace period during the transition phase while the details are being negotiated?
Have you ever thought could be buying direct from the chip manufacturer now, hence no notification required?
 
  • Like
Reactions: 2 users
Apologies on my grammar, I am not the brightest in this field.
Please I own that title


1764319546805.gif
 
  • Haha
  • Love
  • Like
Reactions: 8 users

7für7

Top 20
Have you ever thought could be buying direct from the chip manufacturer now, hence no notification required?
That doesn’t really answer my question…

My question was about what happens in the case of an extension or a new licensing agreement,
not whether there should be an announcement about them even considering an extension in the first place.

But thanks anyway.
 
  • Like
  • Thinking
Reactions: 2 users
That doesn’t really answer my question…

My question was about what happens in the case of an extension or a new licensing agreement,
not whether there should be an announcement about them even considering an extension in the first place.

Well ain’t it obvious we all become very rich

1764320589569.gif
 
  • Haha
Reactions: 1 users
  • Like
  • Love
  • Fire
Reactions: 15 users

Xray1

Regular
That doesn’t really answer my question…

My question was about what happens in the case of an extension or a new licensing agreement,
not whether there should be an announcement about them even considering an extension in the first place.

But thanks anyway.
Note: This is an old Co announcement concerning the MegaChips Licensing Program Agreement and that renewal of this agreement " IF " it were to be imo renewed was or should have taken place on or about the 25th November 2025 .

The Co announcement states:
Additional information on MegaChips agreement Laguna Hills, Calif. – 25 November, 2021 – BrainChip Holdings Ltd (ASX: BRN), (OTCQX: BCHPY) As announced on Monday, Brainchip has entered into a Licensing Program Agreement with MegaChips, a pioneer in the Application Specific Integrated Circuit (ASIC) industry. The agreement has an initial term of four years under which Brainchip grants MegaChips a non-exclusive, worldwide intellectual property license for use in designing and manufacturing BrainChip’s Akida technology into external customers’ system on chip designs. MegaChips will pay a license fee for the grant of the license, which will be paid in tranches over the next two years, with the first payment received on signing the agreement. Brainchip also agrees to provide proof of concept engineering services, Akida sales and development support and software support to MegaChips’ customers under the agreement and has agreed to enter into a Distribution Agreement with MegaChips for the distribution of certain Akida products in Japan. In addition to the upfront license fee, Brainchip has the ability to generate additional revenue under the agreement (which is not yet quantifiable) from: • royalties on the sale of products to MegaChips’ customers (following the design and manufacture of the Akida technology into the customer’s products) calculated as a percentage of the net sales price of such products (with the percentage dependent on the volume of products sold); • license fees for application specific product developments; • project fees for proof of concept development projects with MegaChip’s customers for specific custom networks; and • fees for support services and licensing of software associated with the Akida IP. The remaining terms of the agreement are subject to strict confidentiality provisions as between Brainchip and MegaChips. Brainchip anticipates that it will recognize aggregate revenue of approximately US$2 million under the agreement over the current financial year and the financial year ending 31 December 2022 and expects to recognize additional revenue from the license fee and additional revenue opportunities noted above in subsequent financial years. Page | 2 Brainchip considers that this agreement is highly significant to its growth strategy as it not only provides a licensing fee for its Akida technology but also provides broader opportunities to generate revenue from MegaChip’s substantial global customer base which it would be unlikely to be able to access directly.
 
  • Like
  • Fire
Reactions: 13 users
Not particularly fussed to be honest and enjoying adding to my holdings at these low prices. The shareprice will rise in good time and we could be seeing an announcement that will see a rapid jump at any point. Its getting close to a very exciting period for brainchip imo. U.S defense/government/space use comes before mass adoption and that has been going on for some time already.

Just relax, its so easy to do!
Watch us now and expect some exciting news in December/ January, Everybody is watching Mr Heir
 
  • Like
Reactions: 4 users
BRN doing some Git updates not long ago. I'll put some snips on what ONNX is below.

Creating a ONNX2Akida on Akida.


Brainchip-Inc/MetaONNX

Brainchip Akida 2.0 tools, runtime, examples and documentation for bringing ONNX models to Akida hardware.
  • 0
  • ·
  • Updated
    2 hours ago

Overview​

The onnx2akida package provides tools to evaluate the compatibility of an ONNX model with Akida hardware. The tool quantizes ONNX models, attempts conversion to Akida format, and reports which nodes are hardware-compatible. It can also generate hybrid models that combine both Akida-compatible submodels and standard ONNX operators, enabling inference through the Akida runtime together with ONNXRuntime

From BRN:


Overview


The MetaONNX Framework

The MetaONNX Framework is a dedicated toolchain designed specifically for deploying ONNX models on the Akida 2nd Generation Neuromorphic Processor. MetaONNX enables the execution of any ONNX model regardless of operator support on Akida hardware by combining neuromorphic acceleration with CPU fallback execution for operators that are not supported by Akida. Built on the industry-standard ONNX format, MetaONNX provides tools for analyzing model compatibility, automatically partitioning models into Akida-accelerated and CPU-executed subgraphs, and generating heterogeneous inference models that maximize hardware utilization while ensuring complete model execution.

MetaONNX is based on the onnx2akida Python package installed from the PyPI repository via the pip command. The framework provides:

  • ONNX model ingestion - accepts any valid ONNX model as input, regardless of operator support on Akida hardware,
  • automatic graph partitioning - analyzes the model graph and intelligently partitions it into subgraphs for optimal execution on Akida hardware and CPU,
  • AkidaOp library - a core component that identifies and assigns supported neural network operators to the Akida 2nd Generation hardware,
  • CPU fallback execution - ensures operators not supported by Akida can be executed on CPU,
  • device estimation tools - determines the minimum Akida hardware configuration required for a given model,
  • developer-friendly APIs - Python API and CLI tools (onnx2akida, onnx2akida-device) integrated with ONNX Runtime for seamless deployment workflows.
The onnx2akida toolkit works with models from any framework that supports ONNX export, including TensorFlow (via tf2onnx), PyTorch (via torch.onnx), and Hugging Face models (via Optimum). This broad compatibility enables developers to bring their existing models to the Akida platform without being constrained by hardware-specific operator limitations.

MetaONNX execution flow
MetaONNX execution flow

The MetaONNX examples

The examples section includes tutorials demonstrating the onnx2akida workflow on various model architectures. These examples illustrate:

  • How to analyze ONNX model compatibility with Akida hardware
  • Converting models from popular frameworks (TensorFlow, PyTorch, Hugging Face)
  • Generating and using hybrid inference models
  • Estimating hardware requirements for deployment





What is ONNX (AI summary off my browser)?

ONNX, or Open Neural Network Exchange, is an open-source, standardized format for representing machine learning models. It allows models trained in one framework, such as PyTorch or TensorFlow, to be exported and used in other frameworks or environments without major changes. This interoperability makes models more portable and easier to deploy across various platforms and hardware.

Key functions and benefits
    • Interoperability:
      ONNX acts as a common language that bridges different AI frameworks, enabling seamless model exchange and preventing vendor lock-in.
    • Portability:
      Models can be trained in one environment and then deployed in another, such as taking a model from a research framework to a production environment on a mobile device or cloud service.
    • Framework-agnostic:
      It provides a common set of operators and a file format that is compatible with a wide range of machine learning tools and frameworks.
    • Performance optimization:
      Once a model is in the ONNX format, it can be run using an optimized accelerator like ONNX Runtime, which is designed for high-performance inference across multiple platforms and hardware.
    • Streamlined development:
      It simplifies the process of moving models from training to deployment, which helps speed up innovation in the AI community.
 
  • Like
  • Fire
  • Love
Reactions: 22 users

IloveLamp

Top 20
1000014261.jpg
1000014264.jpg
 
  • Like
  • Love
  • Fire
Reactions: 12 users

Tothemoon24

Top 20


Neuromorphic Accelerator Performance Bottlenecks Modeled, Revealing 3.38x, 3.86x Gains through Optimized Sparsity​

Rohail T.

November 28, 2025BY ROHAIL T.
Neuromorphic Accelerator Performance Bottlenecks Modeled, Revealing 3.38x, 3.86x Gains through Optimized Sparsity
Neuromorphic accelerators represent a potentially transformative approach to machine learning, promising significant gains in speed and energy efficiency through their brain-inspired architectures, but realising this potential requires a deep understanding of their performance limitations. Jason Yik, Walter Gallego Gomez, and Andrew Cheng, alongside Benedetto Leto, Alessandro Pierro, Noah Pacik-Nelson, and colleagues, address this challenge by presenting the first comprehensive analysis of performance bottlenecks in these novel systems. Their work reveals that traditional methods for optimising accelerator performance, which rely on broad measures of sparsity, often fail to capture the nuances of neuromorphic architectures, and the team establishes three distinct bottleneck states, memory-bound, -bound, and traffic-bound, that dictate performance. By combining theoretical modelling with extensive testing on real neuromorphic hardware, including the Brainchip AKD1000, Synsense Speck, and Intel Loihi 2, the researchers develop a ‘floorline’ performance model that accurately predicts workload performance and guides optimisation strategies, ultimately achieving substantial improvements in both speed and energy efficiency, up to 3. 86times faster and 3. 38times more energy efficient, compared to existing methods.

Spiking Neural Networks and Neuromorphic Hardware​

This overview summarizes research into neuromorphic computing, a field aiming to build computer systems that mimic the brain’s efficiency. The research spans hardware design, software development, and application exploration, with a central focus on spiking neural networks (SNNs) and event-based processing, promising significant reductions in power consumption and improvements in efficiency compared to traditional computing architectures. Alongside hardware development, researchers are creating the software tools and algorithms needed to program and utilize neuromorphic hardware. The Neurobench framework provides a standardized platform for benchmarking neuromorphic algorithms and systems, crucial for comparison and progress.
Dataflow synthesis techniques simplify programming by automatically generating SNNs from high-level descriptions, and techniques to reduce computational cost and memory footprint, such as sparse layers and regularization, are also being investigated. These efforts aim to make neuromorphic computing accessible to a wider range of users. While much of the research focuses on the underlying technology, some studies explore potential applications in areas like pattern recognition, machine learning, robotics, and sensor processing, where event-based vision and other sensor applications are particularly well-suited. A recurring theme throughout the research is optimizing energy efficiency and minimizing communication between cores and chips, achieved through efficient network partitioning and core placement, and utilizing sparsity in network weights and activations.
Research also explores specific technologies and approaches, including memristors, approximate computing, and event-based vision. A clear trend emerges towards building and testing actual neuromorphic hardware, rather than solely relying on simulations, with sparsity consistently emerging as a key optimization technique. This work mirrors established methodologies used for conventional architectures, but adapted to the unique characteristics of neuromorphic systems. Researchers developed a simplified analytical model to understand how memory, computation, and communication scale with sparsity and parallelization, informing experiments on the physical hardware and validating theoretical predictions. The team meticulously profiled the three accelerators, measuring performance across various workloads and configurations.
This involved detailed characterization of memory access patterns, computational throughput, and inter-core communication traffic, identifying three distinct accelerator bottleneck states: memory-bound, compute-bound, and traffic-bound. Crucially, the work revealed that conventional performance metrics are insufficient for accurately predicting neuromorphic accelerator performance due to load imbalance at the neurocore level, necessitating neurocore-aware metrics for effective optimization. Building on these insights, the researchers synthesized the “floorline model,” a visual tool analogous to the widely-used roofline model for conventional architectures. This model visually indicates performance bounds for a given neural network architecture and informs how to optimize network instantiation.
The team then developed a two-stage optimization methodology, combining sparsity-aware training with floorline-informed partitioning, achieving substantial performance improvements, demonstrating up to a 4. 29x runtime improvement and a 4. Scientists moved beyond conventional performance metrics to gain a deeper understanding of these novel architectures, revealing that simple measures like network-wide sparsity are often poor indicators of actual performance gains. Experiments demonstrate that increasing weight sparsity alone does not substantially improve runtime for convolutional neural networks (CNNs) on the AKD1000 and Loihi 2, although a slight energy benefit was observed. However, for linearly connected networks, weight sparsity provided performance gains comparable to increasing activation sparsity, with the benefits of weight sparsity dependent on the hardware implementation.
Further investigation into activation sparsity revealed a linear correlation between sparsity and performance when applied uniformly across network layers on the AKD1000 and Loihi 2. However, non-uniform sparsity schedules disrupted this correlation, highlighting the importance of balanced workload distribution. On the Speck accelerator, analysis of varying sparsity schedules indicated that the final network layer often represents the performance bottleneck. The research establishes three distinct accelerator bottleneck states, memory-bound, compute-bound, and traffic-bound, and identifies workload configurations likely to exhibit each state, providing a crucial foundation for optimizing neuromorphic workloads.

Neuromorphic Accelerator Bottlenecks And Performance Limits​

This work presents the first systematic study of performance limitations and bottlenecks in real neuromorphic accelerators, establishing a foundational understanding of their capabilities. Researchers developed analytical models and conducted extensive empirical characterization using three distinct accelerator platforms, identifying three key bottleneck states: memory-bound, compute-bound, and traffic-bound. The study reveals how workload configurations and sparsity levels influence which bottleneck dominates performance, offering insights into optimizing applications for these novel architectures. These insights were synthesized into a floorline performance model, analogous to the widely-used roofline model for conventional architectures, providing a visual representation of performance bounds and informing optimization strategies. The research establishes a crucial foundation for maximizing the efficiency of emerging computing platforms.
 
  • Like
  • Love
  • Fire
Reactions: 14 users

Iseki

Regular
It's concerning that after 4 years neither MegaChips nor Renesas could get Akida to work.
You'd have thought that BRN would have given them a free extension to use Akida2 or pico to keep them in the ecosystem.

Or is it true that Akida2 isn't ready for silicon?

Who knows? Who really cares any more?
 
  • Like
Reactions: 1 users

IloveLamp

Top 20
Such a well placed / timed ad......🤣🤷


1000014268.jpg
 
  • Haha
  • Like
Reactions: 5 users

Tothemoon24

Top 20
Some very big statements from Sean .

13 minute mark & onwards



IMG_1821.jpeg
 
Last edited:
  • Like
  • Love
  • Fire
Reactions: 16 users

Rough Timeline: When 'Mass Market' Might Happen​

Here’s a rough projection for when edge-AI chips go “mass market”:

  • 2025–2026: Rapid ramp-up phase. Many new devices (smartphones, wearables, IoT gadgets, entry-level “AI-enabled” devices) increasingly include on-device AI acceleration. Enterprises begin embedding edge-AI systematically.
  • 2027–2030: Broad adoption period. Edge-AI capabilities in a large portion of new consumer electronics, vehicles (ADAS / smart-vehicle features), smart home devices, industrial IoT. Market size begins reflecting a “mainstream mass-market” equilibrium.
  • By 2030 (and beyond): Edge-AI chips could be ubiquitous — in millions to billions of devices globally — across phones, PCs, cars, home electronics, sensors, industrial machines.
So around 2027–2030 is a plausible window for full mass-market adoption (from the vantage point of late 2025) — though parts of the market (smartphones, premium devices) are already there.
 
  • Like
Reactions: 6 users
Top Bottom