BRN Discussion Ongoing

Time is ticking and very fast and me like probably many other share holders are waiting for this one decent contract in the next 2-3 months for Sean to save his job. If not he should be

View attachment 76948
The recent upgrade to the BRN website to include a detailed list of were Pico will be going and into what products to me means they are about to sign a deal Very soon
Go brainchip
 
  • Like
  • Fire
  • Wow
Reactions: 13 users

Tothemoon24

Top 20
The good Doctor is an absolute ⭐


IMG_0583.jpeg



DeepSeek R1's breakthrough results are a boon for BrainChip and the extreme-edge AI industry. Here's how it will impact inference at the extreme edge:

DeepSeek has shattered the "No one ever got fired for going with IBM" mentality, liberating customers from the behemoths' models. This paves the way for a Cambrian explosion of specialized LLMs delivered by smaller companies. It's akin to breaking the 4-minute mile – now that it's been done, we're poised to witness an explosion of innovation. This is excellent news for smaller companies like BrainChip.

Now imagine AI models running on whispers of energy, delivering powerful performance without massive data centers or cloud connectivity. This is the promise of extreme-edge AI, which is BrainChip's sweet spot.

While DeepSeek is an impressive transformer based model, TENNs technology, based on State Space Models (SSMs), currently offers a superior solution for extreme-edge applications:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

2. Competitive Performance:preliminary experiments pitting the DeepSeek 1.5B model against TENNs 1.2B model in various tasks, such as trip planning and simple programming, showed comparable or slightly better results for TENNs.

3. Extremely Low Training Costs: BrainChip's focus on small specialized models means training costs are less than a premium economy flight from Los Angeles to Sydney.

DeepSeek's success highlights areas where TENNs can be further enhanced. We can leverage many of the tricks learned from DeepSeek to improve their TENNs LLM even more.

The future of extreme-edge AI is bright, with DeepSeek demonstrating that small companies can compete effectively in this space.
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 119 users

IloveLamp

Top 20
View attachment 76949

DeepSeek R1's breakthrough results are a boon for BrainChip and the extreme-edge AI industry. Here's how it will impact inference at the extreme edge:

DeepSeek has shattered the "No one ever got fired for going with IBM" mentality, liberating customers from the behemoths' models. This paves the way for a Cambrian explosion of specialized LLMs delivered by smaller companies. It's akin to breaking the 4-minute mile – now that it's been done, we're poised to witness an explosion of innovation. This is excellent news for smaller companies like BrainChip.

Now imagine AI models running on whispers of energy, delivering powerful performance without massive data centers or cloud connectivity. This is the promise of extreme-edge AI, which is BrainChip's sweet spot.

While DeepSeek is an impressive transformer based model, TENNs technology, based on State Space Models (SSMs), currently offers a superior solution for extreme-edge applications:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

2. Competitive Performance:preliminary experiments pitting the DeepSeek 1.5B model against TENNs 1.2B model in various tasks, such as trip planning and simple programming, showed comparable or slightly better results for TENNs.

3. Extremely Low Training Costs: BrainChip's focus on small specialized models means training costs are less than a premium economy flight from Los Angeles to Sydney.

DeepSeek's success highlights areas where TENNs can be further enhanced. We can leverage many of the tricks learned from DeepSeek to improve their TENNs LLM even more.

The future of extreme-edge AI is bright, with DeepSeek demonstrating that small companies can compete effectively in this space.


That's what i thought i knew but wanted to hear. You're a keeper Tony, buy that man a beer.

Bring on 2025, the instos are coming for your shares MAKE THEM PAY THROUGH THE NOSE.

imo dyor
 
  • Like
  • Fire
  • Love
Reactions: 58 users
  • Like
  • Fire
  • Love
Reactions: 71 users

Diogenese

Top 20
The good Doctor is an absolute ⭐


View attachment 76950


DeepSeek R1's breakthrough results are a boon for BrainChip and the extreme-edge AI industry. Here's how it will impact inference at the extreme edge:

DeepSeek has shattered the "No one ever got fired for going with IBM" mentality, liberating customers from the behemoths' models. This paves the way for a Cambrian explosion of specialized LLMs delivered by smaller companies. It's akin to breaking the 4-minute mile – now that it's been done, we're poised to witness an explosion of innovation. This is excellent news for smaller companies like BrainChip.

Now imagine AI models running on whispers of energy, delivering powerful performance without massive data centers or cloud connectivity. This is the promise of extreme-edge AI, which is BrainChip's sweet spot.

While DeepSeek is an impressive transformer based model, TENNs technology, based on State Space Models (SSMs), currently offers a superior solution for extreme-edge applications:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

2. Competitive Performance:preliminary experiments pitting the DeepSeek 1.5B model against TENNs 1.2B model in various tasks, such as trip planning and simple programming, showed comparable or slightly better results for TENNs.

3. Extremely Low Training Costs: BrainChip's focus on small specialized models means training costs are less than a premium economy flight from Los Angeles to Sydney.

DeepSeek's success highlights areas where TENNs can be further enhanced. We can leverage many of the tricks learned from DeepSeek to improve their TENNs LLM even more.

The future of extreme-edge AI is bright, with DeepSeek demonstrating that small companies can compete effectively in this space.
"equivalent to breaking the 4 minute mile" = Nvidia sliding down the Bannister, while Akida takes the express lift (= elevator in USese) to the top Landying.
 
Last edited:
  • Like
  • Fire
  • Haha
Reactions: 38 users

Diogenese

Top 20
The good Doctor is an absolute ⭐


View attachment 76950


DeepSeek R1's breakthrough results are a boon for BrainChip and the extreme-edge AI industry. Here's how it will impact inference at the extreme edge:

DeepSeek has shattered the "No one ever got fired for going with IBM" mentality, liberating customers from the behemoths' models. This paves the way for a Cambrian explosion of specialized LLMs delivered by smaller companies. It's akin to breaking the 4-minute mile – now that it's been done, we're poised to witness an explosion of innovation. This is excellent news for smaller companies like BrainChip.

Now imagine AI models running on whispers of energy, delivering powerful performance without massive data centers or cloud connectivity. This is the promise of extreme-edge AI, which is BrainChip's sweet spot.

While DeepSeek is an impressive transformer based model, TENNs technology, based on State Space Models (SSMs), currently offers a superior solution for extreme-edge applications:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

2. Competitive Performance:preliminary experiments pitting the DeepSeek 1.5B model against TENNs 1.2B model in various tasks, such as trip planning and simple programming, showed comparable or slightly better results for TENNs.

3. Extremely Low Training Costs: BrainChip's focus on small specialized models means training costs are less than a premium economy flight from Los Angeles to Sydney.

DeepSeek's success highlights areas where TENNs can be further enhanced. We can leverage many of the tricks learned from DeepSeek to improve their TENNs LLM even more.

The future of extreme-edge AI is bright, with DeepSeek demonstrating that small companies can compete effectively in this space.
... and I've still got 2 wishes left!
 
  • Like
  • Haha
  • Love
Reactions: 10 users
IMG_0208.jpeg


First port of call for Utae is to conduct internal audits on the whereabouts of Brainchip revenue landing date ……!
 
  • Like
  • Fire
  • Haha
Reactions: 9 users

Diogenese

Top 20


The A.I. Model wars are intensifying, with 2 new Chinese ones introduced, in the last several hours.

One by Alibaba no less..

Thanks DB,

I mistook the Alibaba patent for DeepSeeks because it had a "Wenfeng" as inventor.

CN118798303A Large language model training method, question and answer method, equipment, medium and productPatent Translate 20240913

ALIYUN FEITIAN HANGZHOU CLOUD COMPUTING TECH CO LTD

Inventors FENG WENFENG; ZHANG YUEWEI; ZENG ZHENYU

The invention provides a large language model training method, a question and answer method, equipment, a medium and a product, and relates to the technical field of artificial intelligence, the training method comprises the following steps: obtaining long text training data, the sequence length of the long text training data being greater than the maximum length of an input text sequence of a pre-trained large language model; increasing a rotation angle base number of a rotation position code of the pre-trained large language model to obtain a modified pre-trained large language model; and training the modified pre-trained large language model by using the long text training data to obtain a trained large language model. In the embodiment, the pre-trained large language model is trained by acquiring the long text training data and increasing the base number of the rotation angles of the rotation position codes, so that the length of the input text sequence is amplified, and the trained large language model can process the long text sequence; and the answer integrity and accuracy of the large language model on questions dependent on long texts and multi-document comparison are improved.
 
  • Like
  • Fire
Reactions: 6 users

Diogenese

Top 20
I agree Dio, but I think Sean reiterated again in that recent podcast interview he did that we are (still) an IP company. Someone correct me if my memory is not serving me well (I couldn’t be bothered listening to it again 😬)
Yes - he did say that, but there are several Akida 1 SoC products.

PCIe, M2, Raspberry Pi, Edge Box, Edge Server, Bascom Hunter 3U VPX SNAP card, ...

I wonder where the chips are coming from?
 
  • Like
  • Thinking
  • Fire
Reactions: 12 users

rgupta

Regular
Yes - he did say that, but there are several Akida 1 SoC products.

PCIe, M2, Raspberry Pi, Edge Box, Edge Server, Bascom Hunter 3U VPX SNAP card, ...

I wonder where the chips are coming from?
I wonder the where the products are selling, there is no sales revenue either.
Jokes apart I assume akida 1000 was fabricated based on demand related orders and there is possibility company can order more chips from tsmc.
 
  • Like
Reactions: 3 users

AusEire

Founding Member.
  • Haha
Reactions: 2 users

AusEire

Founding Member.
That's what i thought i knew but wanted to hear. You're a keeper Tony, buy that man a beer.

Bring on 2025, the instos are coming for your shares MAKE THEM PAY THROUGH THE NOSE.

imo dyor
Can we ensure it's not VB?
 
  • Haha
Reactions: 4 users

JDelekto

Regular
The good Doctor is an absolute ⭐


View attachment 76954


DeepSeek R1's breakthrough results are a boon for BrainChip and the extreme-edge AI industry. Here's how it will impact inference at the extreme edge:

DeepSeek has shattered the "No one ever got fired for going with IBM" mentality, liberating customers from the behemoths' models. This paves the way for a Cambrian explosion of specialized LLMs delivered by smaller companies. It's akin to breaking the 4-minute mile – now that it's been done, we're poised to witness an explosion of innovation. This is excellent news for smaller companies like BrainChip.

Now imagine AI models running on whispers of energy, delivering powerful performance without massive data centers or cloud connectivity. This is the promise of extreme-edge AI, which is BrainChip's sweet spot.

While DeepSeek is an impressive transformer based model, TENNs technology, based on State Space Models (SSMs), currently offers a superior solution for extreme-edge applications:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

2. Competitive Performance:preliminary experiments pitting the DeepSeek 1.5B model against TENNs 1.2B model in various tasks, such as trip planning and simple programming, showed comparable or slightly better results for TENNs.

3. Extremely Low Training Costs: BrainChip's focus on small specialized models means training costs are less than a premium economy flight from Los Angeles to Sydney.

DeepSeek's success highlights areas where TENNs can be further enhanced. We can leverage many of the tricks learned from DeepSeek to improve their TENNs LLM even more.

The future of extreme-edge AI is bright, with DeepSeek demonstrating that small companies can compete effectively in this space.
This has me excited:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

If I understand this correctly it seems to have a fixed memory requirement regardless of the context length. I notice that when using local inferencing with models that support up to a 128k context window, the smaller the context window size I choose, the less memory it consumes (in addition to the memory used by the model data itself).

Removing the context window altogether is like the holy grail for LLMs. You could feed it volumes of information to give it more and more context for a more accurate response to your query.

To put it into perspective, there are about 0.75 tokens per word, with around 96,000 words in a 128k context window. Average novels run around 60,000 to 100,000 words, with "Harry Potter and the Sorcerer's Stone" with about 77,000 words.

Probably an oversimplification, but imagine stuffing a model up front with the Harry Potter novel and being able to ask it to summarize, look for specific parts of the text, create character biographies, write a new imaginative piece using the same writing style, isolate heroes vs. villains, etc.

Now an alternate route, like filling up context for a programming language, examples, the documentation for some 3rd party APIs, and then asking the model to write code to solve a specific problem using all those tools.

If they can do extreme RAG (Retrieval Augment Generation, where current data is retrieved and used to provide the model context) in memory-constrained Edge cases, that is a boon for TENNS. An additional advantage when performing accurate and up-to-date inferencing in conjunction with Akida's ability to update its model as it learns from new sensor input.
 
  • Like
  • Fire
  • Love
Reactions: 66 users

Schwale

Regular
This has me excited:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

If I understand this correctly it seems to have a fixed memory requirement regardless of the context length. I notice that when using local inferencing with models that support up to a 128k context window, the smaller the context window size I choose, the less memory it consumes (in addition to the memory used by the model data itself).

Removing the context window altogether is like the holy grail for LLMs. You could feed it volumes of information to give it more and more context for a more accurate response to your query.

To put it into perspective, there are about 0.75 tokens per word, with around 96,000 words in a 128k context window. Average novels run around 60,000 to 100,000 words, with "Harry Potter and the Sorcerer's Stone" with about 77,000 words.

Probably an oversimplification, but imagine stuffing a model up front with the Harry Potter novel and being able to ask it to summarize, look for specific parts of the text, create character biographies, write a new imaginative piece using the same writing style, isolate heroes vs. villains, etc.

Now an alternate route, like filling up context for a programming language, examples, the documentation for some 3rd party APIs, and then asking the model to write code to solve a specific problem using all those tools.

If they can do extreme RAG (Retrieval Augment Generation, where current data is retrieved and used to provide the model context) in memory-constrained Edge cases, that is a boon for TENNS. An additional advantage when performing accurate and up-to-date inferencing in conjunction with Akida's ability to update its model as it learns from new sensor input.
Thank you. This is an extremely informative post. It summarises how the process works and how components feed into one another.
 
  • Like
  • Love
Reactions: 13 users
This has me excited:

1. No explosive KV cache: Unlike traditional transformer models that rely on rapidly expanding key-value (KV) caches, TENNs sidesteps this issue with fixed memory requirements, regardless of input length. This fundamental property enables LLMs at the extreme edge.

If I understand this correctly it seems to have a fixed memory requirement regardless of the context length. I notice that when using local inferencing with models that support up to a 128k context window, the smaller the context window size I choose, the less memory it consumes (in addition to the memory used by the model data itself).

Removing the context window altogether is like the holy grail for LLMs. You could feed it volumes of information to give it more and more context for a more accurate response to your query.

To put it into perspective, there are about 0.75 tokens per word, with around 96,000 words in a 128k context window. Average novels run around 60,000 to 100,000 words, with "Harry Potter and the Sorcerer's Stone" with about 77,000 words.

Probably an oversimplification, but imagine stuffing a model up front with the Harry Potter novel and being able to ask it to summarize, look for specific parts of the text, create character biographies, write a new imaginative piece using the same writing style, isolate heroes vs. villains, etc.

Now an alternate route, like filling up context for a programming language, examples, the documentation for some 3rd party APIs, and then asking the model to write code to solve a specific problem using all those tools.

If they can do extreme RAG (Retrieval Augment Generation, where current data is retrieved and used to provide the model context) in memory-constrained Edge cases, that is a boon for TENNS. An additional advantage when performing accurate and up-to-date inferencing in conjunction with Akida's ability to update its model as it learns from new sensor input.
The only thing that would get me excited now is Brainchip actually selling their product and making some explosive revenue. NVIDIA would never have thought something like this could've happened, but it did. Imagine waking up one morning and finding out that the PRC have managed to develop and sell some type of device with on chip learning, independant of the cloud. Don't think because its Chinese people won't buy it and its at the right price..... also no secutity concern because it doesn't need to be connected to the cloud....
Tony needs to give Sean and the rest of the selling machine a good kick up the arse.
 
  • Like
  • Love
  • Fire
Reactions: 16 users

charles2

Regular

View attachment 76951
This rather eloquent viewpoint or a version thereof, should be a news release from Brainchip touting not only our product but our leadership.

Time to stop playing nice
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 13 users

Rach2512

Regular
  • Like
  • Fire
  • Love
Reactions: 8 users

Hrdwk

Regular
This might sound stupid, But I wonder if they would consider doing what DeepSeek did to ChatGPT to DeepSeek with Akida?
 
  • Like
Reactions: 2 users

uiux

Regular
  • Like
  • Fire
  • Love
Reactions: 43 users
Haven't had much time this week & just skim reading posts where I can.

Just ran an ASX Broker data on BRN to go with a couple I saved the other week to see who's doing what.

Not gone through properly myself yet but thought I'd post them here fwiw as about to shutdown lappy.

One I hadn't seen before was in the first date range, Instinet Aust (part of the Nomura Group it appears) who bought $440k worth for someone(s).

Newest date range to oldest covering ~ 3 days ago (T+3 data) back to 1/12.

BRN 15.1.25 - 30.1.25.jpg

BRN 1.1.25 - 17.1.25.jpg


BRN 1.12.24 - 15.1.25.jpg
 
  • Like
  • Love
  • Fire
Reactions: 31 users
Top Bottom