BRN Discussion Ongoing

✅
 

Attachments

  • DE83C084-6063-43D5-9CBE-55C4BC57708D.jpeg
    DE83C084-6063-43D5-9CBE-55C4BC57708D.jpeg
    34.5 KB · Views: 57
  • E783FB43-A871-4D60-97E0-0F2ED02EDA7F.jpeg
    E783FB43-A871-4D60-97E0-0F2ED02EDA7F.jpeg
    126.3 KB · Views: 62
  • F0C1C56C-DF3F-4799-8C2A-E1F023EBB911.jpeg
    F0C1C56C-DF3F-4799-8C2A-E1F023EBB911.jpeg
    90.1 KB · Views: 64
  • 473F7769-74E2-4934-A99E-4720A59DC6AC.jpeg
    473F7769-74E2-4934-A99E-4720A59DC6AC.jpeg
    139.3 KB · Views: 60
  • 2FDACDCA-E532-44C3-B3D4-8D2F7A3C0B8B.jpeg
    2FDACDCA-E532-44C3-B3D4-8D2F7A3C0B8B.jpeg
    84.6 KB · Views: 59
  • F26C4D72-DDDC-466F-9EAD-EED9C97852E5.jpeg
    F26C4D72-DDDC-466F-9EAD-EED9C97852E5.jpeg
    188 KB · Views: 56
  • 389CF46D-EDCE-420F-95B3-7DDEAFFB1A7E.jpeg
    389CF46D-EDCE-420F-95B3-7DDEAFFB1A7E.jpeg
    207.8 KB · Views: 52
  • 39D0A7A0-1D8B-4388-833C-4C9890B443D8.jpeg
    39D0A7A0-1D8B-4388-833C-4C9890B443D8.jpeg
    135.3 KB · Views: 52
  • 54EECA1C-56D3-4362-BB0E-58B49E59F9A4.jpeg
    54EECA1C-56D3-4362-BB0E-58B49E59F9A4.jpeg
    146.2 KB · Views: 51
  • A0C85BD7-6A4D-4FA6-9F6D-CF999D71F805.jpeg
    A0C85BD7-6A4D-4FA6-9F6D-CF999D71F805.jpeg
    172.4 KB · Views: 57
  • Like
  • Fire
Reactions: 26 users

Andy38

The hope of potential generational wealth is real

Neuromorphic Sensing: Coming Soon to Consumer Products​

By Sally Ward-Foxton 07.20.2022 0
Share Post





What does “neuromorphic” mean today?
“You will get 10 different answers from 10 different people,” laughed Luca Verre, CEO of Prophesee. “As companies take the step from ‘this is what we believe’ to ‘how can we make this a reality,’ what neuromorphic means will change.”

Most companies doing neuromorphic sensing and computing have a similar vision in mind, he said, but implementations and strategies will be different based on varying product, market, and investment constraints.
ADVERTISEMENT
SCROLL DOWN TO CONTINUE READING
“The reason why… all these companies are working [on neuromorphic technologies] is because there is a fundamental belief that the biological model has superior characteristics compared to the conventional,” he said. “People make different assumptions on product, on system integration, on business opportunities, and they make different implementations… But fundamentally, the belief is the same.”
Prophesee CEO Luca Verre Luca Verre (Source: Prophesee)
Verre’s vision is that neuromorphic technologies can bring technology closer to human beings, which ultimately makes for a more immersive experience and allows technologies such as autonomous driving and augmented reality to be adopted faster.

“When people understand the technology behind it is closer to the way we work, and fundamentally natural, this is an incredible source of reassurance,” he said.

WHICH MARKETS FIRST?

Prophesee is already several years into its mission to commercialize the event–based camera using its proprietary dynamic vision sensor technology. The company has collaborated with camera leader Sony to make a compact, high–resolution event–based camera module, the IMX 636. In this module, the photodiode layer is stacked directly on top of the CMOS layer using Sony’s 3D die stacking process.
According to Verre, the sector closest to commercial adoption of this technology is industrial machine vision.
“Industrial is a leading segment today because historically we pushed our third–generation camera into this segment, which was a bigger sensor and more tuned for this type of application,” he said. “Industrial has historically been a very active machine vision segment, in fact, it is probably one of the segments that adopted the CCD and CMOS technologies at the very beginning… definitely a key market.”
Prophesee Sony IMX 636 event-based sensor Prophesee’s and Sony’s jointly developed IMX 6363 event-based sensor (Source: Prophesee)
The second key market for the IMX 636 is consumer technologies, driven by the shrink in size enabled by Sony’s die–stacking process. Consumer applications include IoT cameras, surveillance cameras, action cameras, drones, consumer robots, and even smartphones. In many cases, the event–based camera is used alongside a full–frame camera, detecting motion so that image processing can be applied to capture better quality images, even when the subject is moving.
“The reason is very simple: event–based cameras are great to understand motion,” he said. “This is what they are meant for. Frame–based cameras are more suited to understanding static information. The combination of the dynamic information from an event–based camera and static information from a frame–based camera is complementary if you want to capture a picture or video in a scene where there’s something moving.”
Event data can be combined with full–frame images to correct any blur on the frame, especially for action cameras and surveillance cameras.
“We clearly see some traction in this in this area, which of course is very promising because the volume typically associated with this application can be quite substantial compared to industrial vision,” he said.
Prophesee is also working with a customer on automotive driver monitoring solutions, where Verre said event–based cameras bring advantages in terms of low light performance, sensitivity, and fast detection. Applications here include eye blinking detection, tracking or face tracking, and micro–expression detection.

APPROACH TO COMMERCIALIZATION

Prophesee's EV4 evaluation kit's EV4 evaluation kit Prophesee’s EV4 evaluation kit (Source: Prophesee)
Prophesee has been working hard on driving commercialization of event–based cameras. The company recently released a new evaluation kit (EVK4) for the IMX 636. This kit is designed for industrial vision with a rugged housing but will work for all applications (Verre said several hundred of these kits have been sold). The company’s Metavision SDK for event–based vision has also recently been open–sourced in order to reduce friction in the adoption of event–based technology. The Metavision community has around 5,000 registered members today.
“The EDK is a great tool to further push and evangelize the technology, and it comes in a very typical form factor,” Verre said. “The SDK hides the perception of complexity that every engineer or researcher may have when testing or exploring a new technology… Think about engineers that have been working for a couple of decades on processing images that now see events… they don’t want to be stretched too much out of their comfort zone.”
New to the Metavision SDK is a simulator to convert full frames into events to help designers transition between the way they work today and the event domain. Noting a reluctance of some designers to move away from full frames, Verre said the simulator is intended to show them there’s nothing magic about events.
“[Events are] just a way of capturing information from the scene that contains much more temporal precision compared to images, and is actually much more relevant, because typically you get only what is changing,” he said.
The simulator can also reconstruct image full frames from event data, which he says people find reassuring.
“The majority of customers don’t pose this challenge any longer because they understand that they need to see from a different perspective, similar to when they use technology like time of flight or ultrasound,” he said. “The challenge is when their perception is that this is another image sensor… for this category of customer, we made this tool that can show them the way to transition stepwise to this new sensing modality… it is a mindset shift that may take some time, but it will come.”
Applications realized in the Prophesee developer communityinclude restoring some sight for the blind, detecting and classifying contaminants in medical samples, particle tracking in research, robotic touch sensors, and tracking space debris.

HARDWARE ROADMAP

In terms of roadmap, Prophesee plans to continue development of both hardware and software, alongside new evaluation kits, development kits, and reference designs. This may include system reference designs which combine Prohpesee sensors with specially developed processors. For example, Prohpesee partner iCatch has developed an AI vision processor SoC that interfaces natively with the IMX 636 and features an on–chip event decoder. Japanese AI core provider DMP is also working with Prophesee on an FPGA–based system, and there are more partnerships in the works, said Verre.
Prophesee Sony IMX 636 event-based camera Prophesee and Sony IMX 636 is a fourth-generation product. Prophesee said future generations will reduce pixel pitch and ease integration with conventional computing platforms (Source: Prophesee)
“We see that there is growing interest from ecosystem partners at the SoC level, but also the software level, that are interested in building new solutions based on Prophesee technology,” he said. “This type of asset is important for the community, because it is another step towards the full solution — they can get the sensor, camera, computing platform, and software to develop an entire solution.”
Where does event–based sensor hardware go from here? Verre cited two key directions the technology will move in. The first is further reduction of pixel size (pixel pitch) and overall reduction of the sensor to make it suitable for compact consumer applications such as wearables. The second is facilitating the integration of event–based sensing with conventional SoC platforms.
Working with computing companies will be critically important to ensure next–generation sensors natively embed the capability to interface with the computing platform, which simplifies the task at the system level. The result will be smarter sensors, with added intelligence at the sensor level.
“We think events make sense, so let’s do more pre-processing inside the sensor itself, because it’s where you can make the least compromise,” Verre said. “The closer you get to the acquisition of the information, the better off you are in terms of efficiency and low latency. You also avoid the need to encode and transmit the data. So this is something that we are pursuing.”
As foundries continue to make progress in the 3D stacking process, stacking in two or even three layers using the most advanced CMOS processes can help bring more intelligence down to the pixel level.
How much intelligence in the pixel is the right amount?
Verre said it’s a compromise between increasing the cost of silicon and having sufficient intelligence to make sure the interface with conventional computing platforms is good enough.
“Sensors don’t typically use advanced process nodes, 28nm or 22nm at most,” he said. “Mainstream SoCs use 12nm, 7nm, 5nm, and below, so they’re on technology nodes that can compress the digital component extremely well. The size versus cost equation means at a certain point it’s more efficient, more economical [to put the intelligence] in the SoC.”
There is also a certain synergy to combining event–based sensors with neuromorphic computing architectures.
“The ultimate goal of neuromorphic technology is to have both the sensing and processing neuromorphic or event–based, but we are not yet there in terms of maturity of this type of solution,” he said. “We are very active in this area to prepare for the future — we are working with Intel, SynSense, and other partners in this area — but in the short term, the mainstream market is occupied by conventional SoC platforms.
Prophesee’s approach here is pragmatic. Verre said the company’s aim is to try to minimize any compromises to deliver benefits that are superior to conventional solutions.
“Ultimately we believe that events should naturally stream asynchronously to a compute architecture that is also asynchronous in order to benefit fully in terms of latency and power,” he said. “But we need to be pragmatic and stage this evolution, and really capitalize on the existing platforms out there and work with key partners in this space that are willing to invest in software–hardware developments and to optimize certain solution for certain markets.”
Love this post @chapman89 ! Basically saying everyone and everything needs neuromorphic tech and Akida can make event based processing on the device a REALITY for anyone, even those old codgers currently using CNN’s. It’s a beautiful thing…Pantene 💪
 
  • Like
  • Love
  • Fire
Reactions: 37 users

chapman89

Founding Member
Love this post @chapman89 ! Basically saying everyone and everything needs neuromorphic tech and Akida can make event based processing on the device a REALITY for anyone, even those old codgers currently using CNN’s. It’s a beautiful thing…Pantene 💪
Yes, and let’s not forget that Prophesee looked to Brainchip to make their event based cameras happen with Sony.

 
  • Like
  • Love
  • Fire
Reactions: 52 users
D

Deleted member 118

Guest
Finally got myself banned from HC. Was worth the lols though. Been here watching TSE from the start. Great job everybody.

 
  • Like
  • Fire
Reactions: 5 users

MDhere

Regular
From what Nviso have previously said we’re not in this generation of Nicobo. Fingers crossed for the next gen.
Hasn't there already been 1 trial generation of Nicobo and he means it was that generation that we were not in ?
I want to get a Nicobo for my granddaughter and one for my mum, ok maybe one for me to BUT only when AKIDA is in it.
Thanks all for the slides, will have a good look at them today.
Happy Thursday fellow brners. My $2 orange kite coming up.
 
  • Like
  • Fire
  • Love
Reactions: 23 users
D

Deleted member 118

Guest
Nasdaq smashing it again, bring on the

 
Last edited by a moderator:
  • Like
  • Haha
  • Fire
Reactions: 24 users

chapman89

Founding Member
Samsung Director liked my ARM & Brainchip post 😮‍💨


6EC6BE22-DDF3-433F-98B5-F76894AACB05.jpeg
 
  • Like
  • Fire
  • Love
Reactions: 106 users
F

Filobeddo

Guest
Last edited by a moderator:
  • Like
  • Fire
  • Love
Reactions: 10 users

AlpineLife

Member
AFR, typo or deliberate? I errr to the later, as they in bed with the shorters?
14.14% our BRN firmly at #2.
😳🤬
 

Attachments

  • 31E73899-F1DC-460F-B451-33C4E38806EB.jpeg
    31E73899-F1DC-460F-B451-33C4E38806EB.jpeg
    66.2 KB · Views: 188
  • Like
  • Wow
  • Sad
Reactions: 20 users
D

Deleted member 118

Guest
Hi all, if I'm welcome here I will end all activities on the sh$t show that has become of HC. If I've offended some in the past and the island votes NO I understand. I'm the first to admit that many strong opinions rubbed me the wrong way 18 months ago but over time I've witnessed the good being done by the group and I enjoy the humour here. As mum used to say listen more n talk less is my lesson.

Great to see you here eventually lol
 
  • Like
Reactions: 7 users
Samsung Director liked my ARM & Brainchip post 😮‍💨


View attachment 12015

More and more big players seem to know (or publicly show they know and like!)

Just waiting for my tax return to clear so I can load the bus more. I was hoping to get to 10k shares before this ole bus takes flight…

Either way I keep topping up, this price is still peanuts!

Also every time I see that BRN 10X Microsoft that someone here has, I stop, pause, Google, calculate and sit back with a “holy crap surely not…” I mean even with my modest holdings. That would see our family out with way waaaay more than I ever could hope for.

Full disclosure. I like to dream about such an incredible SP. But realistically, if this enables me to clear mortgage, then that is good enough. All else is just a bonus 😁.

Sorry got sidetracked, delirious from sitting in my virtual hot tub for all of yesterday 🤣😂
 
  • Like
  • Love
  • Fire
Reactions: 44 users
F

Filobeddo

Guest
Last edited by a moderator:
  • Like
  • Love
  • Fire
Reactions: 25 users
Fire Oops GIF

Beaker prepare the 🔥


 
  • Like
  • Haha
  • Fire
Reactions: 13 users
D

Deleted member 118

Guest
  • Like
Reactions: 3 users

stuart888

Regular
Question: "What is the Brainchip Akida?"

Answer: From the Head of Machine Learning at Edge Impulse, Daniel Situanyake.

Hint: He loves it! This Daniel dominates this entire video. Class act.


 
  • Like
  • Love
  • Fire
Reactions: 47 users
  • Haha
  • Like
  • Fire
Reactions: 16 users
Sorry guys my Dad came over & distracted me & I missed it.

I'm sure it will be available soon as Tim is a big promoter.

let's go help GIF by kate spade new york's go help GIF by kate spade new york
Hope you enjoyed the time with your father. Family is everything all else is just white noise♥️
 
  • Like
  • Love
Reactions: 15 users
F

Filobeddo

Guest
Lol. So did a Crime Scene Invesiigator. Not sure on that one. Hope it means we're murdering the opposition.

SC

Careful, Old mate CSI is one of us 😉😂
 
  • Haha
  • Like
Reactions: 6 users
  • Haha
  • Like
Reactions: 3 users
D

Deleted member 118

Guest
Question: "What is the Brainchip Akida?"

Answer: From the Head of Machine Learning at Edge Impulse, Daniel Situanyake.

Hint: He loves it! This Daniel dominates this entire video. Class act.



Akida discussed from about 19 minute mark
 
  • Like
  • Fire
  • Love
Reactions: 14 users
Top Bottom