BRN Discussion Ongoing

AARONASX

Holding onto what I've got
IMO, AI in the cloud is getting more expensive everyday...$40,000+ for a chip wow!.... it seems to me that Akida is looking day-after-day a more cost effective option for customers if they can integrate this in the cloud and the edge this will upset those that spend all this money for nothing in 1-2 years time. hence why NDAs and very little information might be a good thing for ALL of us, we don't want to be road blocked until it is too late to stop Brainchips team...while i wish there was more information, sometime I (we) need step back and look at the bigger picture.

 
  • Like
  • Fire
  • Love
Reactions: 32 users

IloveLamp

Top 20



1000014581.jpg
1000014583.jpg
 
  • Like
  • Fire
  • Thinking
Reactions: 18 users

TheDrooben

Pretty Pretty Pretty Pretty Good
  • Like
  • Fire
Reactions: 18 users

Euks

Regular
Nice find. Do you think this is us?

No mentioned in the product description but lots of familiar phrases… “Fanless” “supports neural networks”

IMG_4670.png



 
  • Like
  • Fire
Reactions: 21 users

IloveLamp

Top 20
Can I ask you please about the Arm Cortex M55 and if we know if BRN has or can be integrated into this yet ?.
The company has said it "can be"
 
  • Like
Reactions: 6 users

7für7

Top 20
  • Thinking
  • Like
Reactions: 2 users

toasty

Regular
Being a CEO should be about increasing the value of a company and I believe BRNs CEO is doing that in spades! Being the CEO should not be about trying to control the day to day share price fluctuations. If you think the day to day SP machinations have anything to do with company value in a pre-revenue company then this post is wasted on you.
You will note that I didn't mention the day to day share price. I referred to the value of the company. Whether we like it or not, that is measured in the wider market by commercial success and a share price that attaches to that.

As I have said many times, I am a true believer in the technology but am concerned about how management is parlaying the technology into commercial success.

Pre-revenue valuations for publicly listed companies are based on something that may or may not happen in the future. Patents have an intrinsic value but only if they attach to a product that the market believes will be able to generate revenue. Likewise partnerships. In my view management has not been transparent enough to convince the market that commercial success is inevitable and consequently the value of the company is depressed. The current state of the enterprise may speak value to you but clearly it does not do so to the market. And to be quite frank its what the market perceives that dictates enterprise value, not what you or I would like to believe.
 
  • Like
  • Thinking
Reactions: 17 users

7für7

Top 20
And this kind of system I spotted recently in japan when I was for a business trip in a bakery… interesting name

 
  • Fire
  • Like
Reactions: 4 users

IMG_1539.jpeg


IMG_1538.jpeg
 
  • Like
  • Fire
  • Thinking
Reactions: 7 users

buena suerte :-)

BOB Bank of Brainchip
Is this the first time we have publicly stated that the company intends to "enable radical innovation required to bring LLM's to the edge.
Although I must admit my disappointment whenever I see another CR I was then very pleased to read what it was for.
Overall happy days
I agree Boab..I think this is going to be our much needed big step up moving forward !!!!

“The company will also bolster the CTO function, enabling radical innovation required to bring large language models, multi-modal operation and other state of art AI to the edge and ensure we remain the industry leaders in hyper-efficient Edge AI” SH
 
  • Like
  • Fire
  • Love
Reactions: 12 users

AARONASX

Holding onto what I've got
Nice find. Do you think this is us?

No mentioned in the product description but lots of familiar phrases… “Fanless” “supports neural networks”

View attachment 59974


Mentions that the CPU -Intel® Atom® x6414RE Processor SoC

Intel® Gaussian & Neural Accelerator 1.0


possibly not, as been around back in 2022, unless works together.

1711581130348.png
 
  • Like
  • Fire
Reactions: 5 users

Diogenese

Top 20
I need help from an expert here, but isn’t Vision Transformers at the Edge an Akida 2 speciality?

View attachment 59958
ViT is a new feature of Akida 2, but, according to Prof. Wiki, the idea of transformers has been around for several years, although the concept was applied to vision a couple of years ago.
 
  • Like
Reactions: 7 users

Bravo

If ARM was an arm, BRN would be its biceps💪!

Sony Semiconductor Brings Inference Close To The Edge​

Steve McDowell
Contributor
Chief Analyst & CEO, NAND Research.


https://www.forbes.com/sites/stevem...close-to-the-edge/?sh=660e80bb34f9#open-web-0
Mar 27, 2024,04:12pm EDT
Sony Semiconductor Solutions Group

Sony Semiconductor Solutions Group
NURPHOTO VIA GETTY IMAGES

AI only matters to businesses if the technology enables competitive differentiation or drives increased efficiencies. The past year has seen technology companies focus on training models that promise to change enterprises across industries. While training has been the focus, the recent NVIDIA GTC event showcased a rapid transition towards inference, where actual business value lay.


AI at the Retail Edge​

Retail is one of the industries that promises to benefit most from AI. Generative AI and large language models aside, retail organizations are already deploying image recognition systems for diverse tasks, from inventory control and loss prevention to customer service.

Earlier this year, Nvidia published its 2024 State of AI in Retail and CPG report that takes a survey-based approach to understanding the use of AI in the retail sector. Nvidia found that 42% percent of retailers already use AI, with an additional 34% assessing or piloting AI programs. Narrowing the aperture, among large retailers with revenues of more than $500 million, the adoption of AI stretches to 64%. That’s a massive market.


The challenge for retailers and the array of ecosystem partners catering to them is that AI can be complex. Large language models and generative AI require infrastructure that scales beyond the capabilities of many retail locations. Using the cloud to solve those problems isn't always practical, either, as applications like vision processing need to be done at the edge, where the data lives.


Sony’s Platform Approach to On-Device Inference​

Sony Semiconductor Solutions Corporation took on the challenge of simplifying vision processing and inference, resulting in the introduction of its AITRIOS edge AI sensing platform. AITRIOS addresses six significant cloud challenges based IoT systems, including handling large data volumes, enhancing data privacy, reducing latency, conserving energy, ensuring service continuity, and securing data.


AITRIOS accelerates the deployment of edge AI-powered sensing solutions across industries, enabling a comprehensive ecosystem for creating solutions that blend edge computing and cloud technologies.


 
  • Like
  • Thinking
  • Love
Reactions: 12 users

TheDrooben

Pretty Pretty Pretty Pretty Good
Nice find. Do you think this is us?

No mentioned in the product description but lots of familiar phrases… “Fanless” “supports neural networks”

View attachment 59974


I saw that it was updated Cupcake "2" and fanless so Larry has his fingers crossed......although no specific mention. We are in a new configuration of Cupcake edge AI servers
 
  • Like
Reactions: 10 users

DJM263

LTH - 2015
Can I ask you please about the Arm Cortex M55 and if we know if BRN has or can be integrated into this yet ?.
Hi @Smoothsailing

I actually sent the company an email back when the Arm Cortex M55 was announced (23/11/23) asking if Akida had been used in the same way as in the Cortex M85!

The following is the only response I got. Understanding PVDM is very busy so i thought I would just sit back and let things play out...

23/11/2023

Thank you for your email.

I have forwarded your email to Peter van der Made, and asked him to provide a response in regard to your question about Cortex M-52.

When Peter sends his response, I will forward it to you.

Regards


Tony Dawe
Director, Global Investor Relations

+61 (0)405 989 743
 
  • Like
  • Fire
Reactions: 11 users

stuart888

Regular
The Nvidia Youtube Channel is unreal. I also like Broadcom Youtube to see the Open System they are building to compete with Nvidia.

Note: Nvidia is Broadcom's largest customer!

Meta and Google buy Neural custom chips from Broadcom, and they got a third customer announced, Amazon, Apple, or Bytedance? Broadcom builds their custom chips start to finish in 18 months or less. They have a process perfected over 10 years. Very interesting.



NVIDIA Holodeck is a virtual reality (VR) innovation platform that brings designers, peers, and stakeholders together from anywhere in the world to build and explore creations in a highly realistic, collaborative, and physically simulated VR environment.

See how NVIDIA Holodeck can bring teams together, speed up productivity, and improve your entire creative process.
 
  • Fire
  • Like
Reactions: 6 users
Hi @Smoothsailing

I actually sent the company an email back when the Arm Cortex M55 was announced (23/11/23) asking if Akida had been used in the same way as in the Cortex M85!

The following is the only response I got. Understanding PVDM is very busy so i thought I would just sit back and let things play out...

23/11/2023

Thank you for your email.

I have forwarded your email to Peter van der Made, and asked him to provide a response in regard to your question about Cortex M-52.

When Peter sends his response, I will forward it to you.

Regards



Tony Dawe
Director, Global Investor Relations

+61 (0)405 989 743
Thanks my reason for this question is in relationship to this latest announcement regarding the Ambiq using the M55 and if we could be involved . 🫤
 
  • Like
Reactions: 3 users

stuart888

Regular

View attachment 59970
What did you say Akida SNN, power saver? Low Power AI MD!
"Artificial Intelligence may never be able to fully replace a doctor 👨‍⚕️, but it can certainly help supplement their work, speed up diagnostic processes, or offer data-driven 🧠 analyses to assist with decision making."
Go AI!
 
  • Like
  • Fire
Reactions: 8 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
Here's an article about Sony's LEV-2 lunar robot with it's low-power IoT board computer SPRESENSE.

BTW, a few months ago Rob Telson "liked" a post about SPRESENSE in terms of it's intelligent HVAC capabilities.

Let's hope Sony is keeping abreast of our impressive incursions in the space vertical! We'd make great partners!

Screenshot_20230817_091848_LinkedIn.jpg








Sony's Technology Leading to Space:​

Behind the scenes of LEV-2 lunar exploration robot development​

March 7, 2024

On January 25, 2024, a photograph of Japan's first successful pinpoint lunar landing by the Smart Lander for Investigating Moon (SLIM), was released. The historic photo was captured by the Transformable Lunar Robot LEV-2 (nicknamed "SORA-Q"), jointly developed by Sony Group Corporation and Japan Aerospace Exploration Agency (JAXA), TOMY Company, Ltd. (TOMY), and Doshisha University. We spoke with Masaharu Nagata (Exploratory Deployment Group, Technology Platform, Sony Group Corporation), who led the Sony team, about what became the world's first fully autonomous robot lunar exploration project, and Sony's technologies that made it possible.

img01.jpg

World's first fully autonomous robot​

that captured SLIM landing on the moon​

LEV-2, jointly developed by the four parties, is the world's smallest and lightest fully autonomous lunar exploration robot at approx. 80mm in diameter (before transformation) and weighing around 250g. TOMY's knowledge and ideas from toy development have been put to use in the endearing spherical form and the frame design which allows it to travel across the lunar surface in a motion inspired by sea turtles.

LEV-2 was launched into space on board the SLIM lander and was released onto the lunar surface together with LEV-1, an ultra-compact Lunar Excursion Vehicle, on January 20, 2024. It then autonomously found the SLIM, taking pictures as it moved along, and transmitted high-quality image data of the lunar lander and its surroundings to the LEV-1, thereby delivering photos of the lunar surface to earth.

©JAXA/TOMY/Sony Group Corporation/Doshisha University

IoT board computer "SPRESENSE" and Sony's technology combine​

to make a super compact and lightweight robot​

The development of the lunar exploration robot began in 2016 by JAXA's Space Exploration Innovation Hub and TOMY. Sony joined in 2019, and Doshisha University followed in 2021. Sony's engineers were assigned the mission to realize a fully autonomous robot from the concept created by JAXA and TOMY for "a sphere-shaped robot that transforms to move around on uneven terrain." In addition, Sony led the selection of components as well as the technological development for the robot's motion control and imaging.

During the early stage of development, the team singled out a low-power IoT board computer "SPRESENSE™" developed and commercialized by Sony Semiconductor Solutions, which they determined to be the most suitable technology for what would be the world's smallest and lightest lunar exploration robot. Despite SPRESENSE's small size and low power consumption, it boasts a high performance programmable to work with camera and sensing devices enabling the robot to drive autonomously, capture images and perform other operations. After repeated environmental tests to verify its resistance to radiation and external shocks, SPRESENSE was selected as the core processor to control all the robot's movements. The development environment and source code are open to the public, which was also an advantage in working together with partners.

img03.jpg
Operation board installed on LEV-2
SPRESENSE mainboard mounted with components such as "ISX012" image sensors
However, minimizing the system load would be necessary to allow all tasks, including autonomous mobility, image capturing, and wireless communication, to be operated by SPRESENSE. To this end, engineers who had honed their skills in device design and development worked steadily on configuring the circuits and boards.

A further challenge was image processing, which is a vital element in enabling the robot to travel to the proper position as well as capture images. It was not possible to perfectly replicate the space environment and the conditions of the lunar landing for the tests on earth. "Without knowing the correct answer, we considered how to reliably detect and photograph the SLIM, and finally devised a system by which the LEV-2 could recognize the SLIM from the gold color of its exterior insulation. We then repeatedly experimented with every conceivable situation, including various light and shadow contrasts and reflections, to improve the accuracy of the technology. "

img04.jpg
Image data on the lunar captured and transmitted by LEV-2
(Test image received through the data transmission by test radio between LEV-2 to the radio station via LEV-1)

Expertise in manufacturing​

enabling a new approach in space exploration​

Nagata is also involved in Sony's "Earth MIMAMORI Platform" project, which aims to contribute to the sustainability of the Earth and society through Sony's technologies, and has been working on new communication technologies that leverage outer space. Through such experience, he is aware how challenging outer space development is, where you face risks and harsh environment that are completely different from those on earth, including impact of launch, temperature changes and cosmic rays.
"Every single component needs to be specialized for space, which requires a great deal of time and effort to develop. With that in mind, I feel that the development of the robot by these four parties proceeded quite quickly," he says.

In fact, an automated robot prototype was completed within about a year after Sony's participation in 2019. This allowed ample time for field testing and environmental testing by JAXA. This speedy development was made possible by a new approach taken by TOMY and Sony, both with roots in manufacturing, though in different fields.

"LEV-2 utilizes consumer devices for its components, and its frame is simple enough to be assembled by hand. Moreover, both companies excel at mass production of consistent and high-quality products. We were able to efficiently improve the overall quality of LEV-2 during its development by building multiple prototypes equipped with the same technology and testing them in parallel to verify the technologies each party was responsible for."

This was a novel approach from a new perspective, where in conventional development experiments were repeated with the single actual vehicle to be launched into space.

img05.jpg
Prototype of LEV-2 with the same core technology
as the actual robot that completed lunar exploration
Recently, more companies are showing an interest in space and entering the field by developing new technologies such as satellites or launching new businesses. Looking back on the joint research, Nagata says, "We were able to prove with our own hands that the technologies and skills we have honed at Sony can contribute to space, which is very gratifying. I also realized firsthand that fundamental principles of manufacturing remains the same, whether on the ground or in space. I would like to continue exploring the potential of Sony's technologies and development expertise in new fields such as outer space."

img06.jpg


Sony's R&D team who worked on the joint research project
 
Last edited:
  • Like
  • Wow
  • Thinking
Reactions: 17 users
Top Bottom