BRN Discussion Ongoing

Sirod69

bavarian girl ;-)
I always want to find any conection with Qualcomm

Qualcomm

Qualcomm1.067.540 Follower:innen17 Std. • vor 17 Stunden

After 3 years in the making, we’re proud to announce the #Snapdragon W5+ Gen 1 wearable platform: our most advanced leap yet. With breakthrough performance, new thresholds in low power consumption, and a smaller-than-ever chip size enabling sleeker devices, this platform unlocks premium, interactive, ambient and always-on experiences. Learn all you need to know about our latest wearables powerhouse: https://bit.ly/3uStCVp
Bewerten Sie diese Übersetzung


Nach 3 Jahren in der Entwicklung sind wir stolz darauf, die #Snapdragon W5+ Gen 1 Wearable-Plattform ankündigen zu können: unser bisher fortschrittlichster Sprung. Mit bahnbrechender Leistung, neuen Schwellenwerten für einen niedrigen Stromverbrauch und einer geringeren Chipgröße als je zuvor, die schlankere Geräte ermöglicht, ermöglicht diese Plattform erstklassige, interaktive, Umgebungs- und Always-On-Erlebnisse. Erfahren Sie alles, was Sie über unser neuestes Wearables-Kraftpaket wissen müssen: https://bit.ly/3uStCVp
 
  • Like
  • Love
  • Thinking
Reactions: 22 users

Cgc516

Regular
I LIKE IT ALL AND HAVE THAT FEELING IN MY BONES BOYZ!!!!!!!!!

Hey quietly we would not be still up near our daily top IF this was not a major move upwards! Shorters would have KILLED this many times today if its supposed to drop.

Wish I had more funds .................same ole story. Bugger. :(

Yak52 :cool:
Yes quietly, there won’t be any speedy tickets!
 
  • Like
Reactions: 3 users

Sirod69

bavarian girl ;-)
I also wish everyone a wonderful day and I'm going for a swim now, in this heat
1658318929238.png


Edge Impulse

Edge Impulse23.236 Follower:innen8 Std. • vor 8 Stunden

"By putting machine learning on a sensor, engineers can build devices that don’t need an internet connection for basic tasks."

Here's a good read from Stacey Higginbotham featuring the recent work of Pete Warden along with several other researchers — and what it means for embedding ML intelligence into sensor 2.0 devices.

cc: Matthew Stewart, Brian Plancher, Colby Banbury, Shvetank Prakash, Emma Belliveau Zain Asgar, Sachin Katti, Vijay Janapa Reddi

 
  • Like
  • Love
Reactions: 27 users

VictorG

Member
  • Haha
  • Like
  • Love
Reactions: 15 users
F

Filobeddo

Guest
Haha 😁 yeah, what’s that one ‘wanna earn 14 bucks the hard way?.’ 😂😂

Legend



Classic
 
  • Haha
  • Fire
Reactions: 4 users
F

Filobeddo

Guest
That's not Rodney Dangerfield 😛



Tell you what, the shorters are getting no respect at the moment 🤣


Yeah , well try this one



'figured out I'm bisexual..yeah I have sex twice a year' 😁
 
  • Haha
  • Like
  • Fire
Reactions: 9 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
I think this perfectly encapsulates how I feel atm, about our little community of BrainChippers. ❤️Together we will help ground each other and together we will help each-other to FLY!!!!!!!!!!!!!🐦



PS: I raided the medicine cabinet a tiny bit too.🤭
PPS: TRUSTING OTHERS IS HARDER THAN IT LOOKS
 
Last edited:
  • Like
  • Love
  • Fire
Reactions: 32 users

cassip

Regular
Tim was asked if he new about the next phase of AKIDA.
He ummned and aarghed before saying something like we have NDA's with clients. I dont recall correctly but because of his hesitation I felt he new something. My opinion only.
He was also asked if he was happy with the relationship with Brainchip which he was. He also mentioned something about a joint presentation to a US company??
Anyone else have a better recollection?
Hi Boab,
I noticed the same as you and then I got a call. I am not sure about the "sort of" hesitation. Also because of this it would be really interesting to know if he said something about a joint presention to a US company. (?)
Regards, cassip
 
  • Like
Reactions: 5 users

cassip

Regular
... by the way: what was his answer to the question why they asked for someone´s BRN-shareholdership?
 
  • Like
Reactions: 2 users

Boab

I wish I could paint like Vincent
... by the way: what was his answer to the question why they asked for someone´s BRN-shareholdership?
He said you need to take it up with the company who is promoting the possible listing. I can't remember their name.
Will be god when we get the replay and we can pause and rewind..
 
  • Like
Reactions: 3 users

cassip

Regular
He said you need to take it up with the company who is promoting the possible listing. I can't remember their name.
Will be god when we get the replay and we can pause and rewind..
Thank you for your reply
 
  • Like
Reactions: 1 users
F

Filobeddo

Guest
I think this perfectly encapsulates how I feel atm, about our little community of BrainChippers. ❤️Together we will help ground each other and together we will help each-other to FLY!!!!!!!!!!!!!🐦



PS: I raided the medicine cabinet a tiny bit too.🤭
PPS: TRUSTING OTHERS IS HARDER THAN IT LOOKS


That’s ok @Bravo as long as it’s just ‘medicinal’ and they’re just ‘tiny’ sips 😉
 
Last edited by a moderator:
  • Haha
  • Like
Reactions: 5 users

cosors

👀
Finally got myself banned from HC. Was worth the lols though. Been here watching TSE from the start. Great job everybody.
In my case, they finally did it after four months and three letters of cancellation. That chapter is closed. May their stock dry up.
I'm rather tame compared to this revoluzer.
Screenshot_2022-05-11-19-41-46-52_7d6541707e0ad471ad1a839839bd7d1b.jpg
 
Last edited:
  • Haha
  • Like
  • Love
Reactions: 26 users
Pressure intensifying


For any Rage Against The Machine fans -

“Burn, burn, yes you’re gonna burn”

7D421477-13D8-4585-AE2C-1F41C9A83334.jpeg


 
  • Like
  • Fire
  • Love
Reactions: 19 users
 
  • Like
  • Fire
  • Love
Reactions: 13 users

Sirod69

bavarian girl ;-)

Neuromorphic Sensing: Coming Soon to Consumer Products​

There is also a certain synergy to combining event–based sensors with neuromorphic computing architectures.

“The ultimate goal of neuromorphic technology is to have both the sensing and processing neuromorphic or event–based, but we are not yet there in terms of maturity of this type of solution,” he said. “We are very active in this area to prepare for the future — we are working with Intel, SynSense, and other partners in this area — but in the short term, the mainstream market is occupied by conventional SoC platforms.
(Robin Telson likes)

 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 28 users

chapman89

Founding Member

Neuromorphic Sensing: Coming Soon to Consumer Products​

There is also a certain synergy to combining event–based sensors with neuromorphic computing architectures.

“The ultimate goal of neuromorphic technology is to have both the sensing and processing neuromorphic or event–based, but we are not yet there in terms of maturity of this type of solution,” he said. “We are very active in this area to prepare for the future — we are working with Intel, SynSense, and other partners in this area — but in the short term, the mainstream market is occupied by conventional SoC platforms.
(Robin Telson likes)

8EEEB7D1-4A21-438A-9234-06D66ECBCDE0.jpeg
 
  • Like
  • Fire
  • Love
Reactions: 55 users

chapman89

Founding Member

Neuromorphic Sensing: Coming Soon to Consumer Products​

By Sally Ward-Foxton 07.20.2022 0
Share Post





What does “neuromorphic” mean today?
“You will get 10 different answers from 10 different people,” laughed Luca Verre, CEO of Prophesee. “As companies take the step from ‘this is what we believe’ to ‘how can we make this a reality,’ what neuromorphic means will change.”

Most companies doing neuromorphic sensing and computing have a similar vision in mind, he said, but implementations and strategies will be different based on varying product, market, and investment constraints.
ADVERTISEMENT
SCROLL DOWN TO CONTINUE READING
“The reason why… all these companies are working [on neuromorphic technologies] is because there is a fundamental belief that the biological model has superior characteristics compared to the conventional,” he said. “People make different assumptions on product, on system integration, on business opportunities, and they make different implementations… But fundamentally, the belief is the same.”
Prophesee CEO Luca Verre Luca Verre (Source: Prophesee)
Verre’s vision is that neuromorphic technologies can bring technology closer to human beings, which ultimately makes for a more immersive experience and allows technologies such as autonomous driving and augmented reality to be adopted faster.

“When people understand the technology behind it is closer to the way we work, and fundamentally natural, this is an incredible source of reassurance,” he said.

WHICH MARKETS FIRST?

Prophesee is already several years into its mission to commercialize the event–based camera using its proprietary dynamic vision sensor technology. The company has collaborated with camera leader Sony to make a compact, high–resolution event–based camera module, the IMX 636. In this module, the photodiode layer is stacked directly on top of the CMOS layer using Sony’s 3D die stacking process.
According to Verre, the sector closest to commercial adoption of this technology is industrial machine vision.
“Industrial is a leading segment today because historically we pushed our third–generation camera into this segment, which was a bigger sensor and more tuned for this type of application,” he said. “Industrial has historically been a very active machine vision segment, in fact, it is probably one of the segments that adopted the CCD and CMOS technologies at the very beginning… definitely a key market.”
Prophesee Sony IMX 636 event-based sensor Prophesee’s and Sony’s jointly developed IMX 6363 event-based sensor (Source: Prophesee)
The second key market for the IMX 636 is consumer technologies, driven by the shrink in size enabled by Sony’s die–stacking process. Consumer applications include IoT cameras, surveillance cameras, action cameras, drones, consumer robots, and even smartphones. In many cases, the event–based camera is used alongside a full–frame camera, detecting motion so that image processing can be applied to capture better quality images, even when the subject is moving.
“The reason is very simple: event–based cameras are great to understand motion,” he said. “This is what they are meant for. Frame–based cameras are more suited to understanding static information. The combination of the dynamic information from an event–based camera and static information from a frame–based camera is complementary if you want to capture a picture or video in a scene where there’s something moving.”
Event data can be combined with full–frame images to correct any blur on the frame, especially for action cameras and surveillance cameras.
“We clearly see some traction in this in this area, which of course is very promising because the volume typically associated with this application can be quite substantial compared to industrial vision,” he said.
Prophesee is also working with a customer on automotive driver monitoring solutions, where Verre said event–based cameras bring advantages in terms of low light performance, sensitivity, and fast detection. Applications here include eye blinking detection, tracking or face tracking, and micro–expression detection.

APPROACH TO COMMERCIALIZATION

Prophesee's EV4 evaluation kit Prophesee’s EV4 evaluation kit (Source: Prophesee)
Prophesee has been working hard on driving commercialization of event–based cameras. The company recently released a new evaluation kit (EVK4) for the IMX 636. This kit is designed for industrial vision with a rugged housing but will work for all applications (Verre said several hundred of these kits have been sold). The company’s Metavision SDK for event–based vision has also recently been open–sourced in order to reduce friction in the adoption of event–based technology. The Metavision community has around 5,000 registered members today.
“The EDK is a great tool to further push and evangelize the technology, and it comes in a very typical form factor,” Verre said. “The SDK hides the perception of complexity that every engineer or researcher may have when testing or exploring a new technology… Think about engineers that have been working for a couple of decades on processing images that now see events… they don’t want to be stretched too much out of their comfort zone.”
New to the Metavision SDK is a simulator to convert full frames into events to help designers transition between the way they work today and the event domain. Noting a reluctance of some designers to move away from full frames, Verre said the simulator is intended to show them there’s nothing magic about events.
“[Events are] just a way of capturing information from the scene that contains much more temporal precision compared to images, and is actually much more relevant, because typically you get only what is changing,” he said.
The simulator can also reconstruct image full frames from event data, which he says people find reassuring.
“The majority of customers don’t pose this challenge any longer because they understand that they need to see from a different perspective, similar to when they use technology like time of flight or ultrasound,” he said. “The challenge is when their perception is that this is another image sensor… for this category of customer, we made this tool that can show them the way to transition stepwise to this new sensing modality… it is a mindset shift that may take some time, but it will come.”
Applications realized in the Prophesee developer communityinclude restoring some sight for the blind, detecting and classifying contaminants in medical samples, particle tracking in research, robotic touch sensors, and tracking space debris.

HARDWARE ROADMAP

In terms of roadmap, Prophesee plans to continue development of both hardware and software, alongside new evaluation kits, development kits, and reference designs. This may include system reference designs which combine Prohpesee sensors with specially developed processors. For example, Prohpesee partner iCatch has developed an AI vision processor SoC that interfaces natively with the IMX 636 and features an on–chip event decoder. Japanese AI core provider DMP is also working with Prophesee on an FPGA–based system, and there are more partnerships in the works, said Verre.
Prophesee Sony IMX 636 event-based camera Prophesee and Sony IMX 636 is a fourth-generation product. Prophesee said future generations will reduce pixel pitch and ease integration with conventional computing platforms (Source: Prophesee)
“We see that there is growing interest from ecosystem partners at the SoC level, but also the software level, that are interested in building new solutions based on Prophesee technology,” he said. “This type of asset is important for the community, because it is another step towards the full solution — they can get the sensor, camera, computing platform, and software to develop an entire solution.”
Where does event–based sensor hardware go from here? Verre cited two key directions the technology will move in. The first is further reduction of pixel size (pixel pitch) and overall reduction of the sensor to make it suitable for compact consumer applications such as wearables. The second is facilitating the integration of event–based sensing with conventional SoC platforms.
Working with computing companies will be critically important to ensure next–generation sensors natively embed the capability to interface with the computing platform, which simplifies the task at the system level. The result will be smarter sensors, with added intelligence at the sensor level.
“We think events make sense, so let’s do more pre-processing inside the sensor itself, because it’s where you can make the least compromise,” Verre said. “The closer you get to the acquisition of the information, the better off you are in terms of efficiency and low latency. You also avoid the need to encode and transmit the data. So this is something that we are pursuing.”
As foundries continue to make progress in the 3D stacking process, stacking in two or even three layers using the most advanced CMOS processes can help bring more intelligence down to the pixel level.
How much intelligence in the pixel is the right amount?
Verre said it’s a compromise between increasing the cost of silicon and having sufficient intelligence to make sure the interface with conventional computing platforms is good enough.
“Sensors don’t typically use advanced process nodes, 28nm or 22nm at most,” he said. “Mainstream SoCs use 12nm, 7nm, 5nm, and below, so they’re on technology nodes that can compress the digital component extremely well. The size versus cost equation means at a certain point it’s more efficient, more economical [to put the intelligence] in the SoC.”
There is also a certain synergy to combining event–based sensors with neuromorphic computing architectures.
“The ultimate goal of neuromorphic technology is to have both the sensing and processing neuromorphic or event–based, but we are not yet there in terms of maturity of this type of solution,” he said. “We are very active in this area to prepare for the future — we are working with Intel, SynSense, and other partners in this area — but in the short term, the mainstream market is occupied by conventional SoC platforms.
Prophesee’s approach here is pragmatic. Verre said the company’s aim is to try to minimize any compromises to deliver benefits that are superior to conventional solutions.
“Ultimately we believe that events should naturally stream asynchronously to a compute architecture that is also asynchronous in order to benefit fully in terms of latency and power,” he said. “But we need to be pragmatic and stage this evolution, and really capitalize on the existing platforms out there and work with key partners in this space that are willing to invest in software–hardware developments and to optimize certain solution for certain markets.”
 
  • Like
  • Fire
  • Love
Reactions: 58 users
✅
 

Attachments

  • DE83C084-6063-43D5-9CBE-55C4BC57708D.jpeg
    DE83C084-6063-43D5-9CBE-55C4BC57708D.jpeg
    34.5 KB · Views: 70
  • E783FB43-A871-4D60-97E0-0F2ED02EDA7F.jpeg
    E783FB43-A871-4D60-97E0-0F2ED02EDA7F.jpeg
    126.3 KB · Views: 72
  • F0C1C56C-DF3F-4799-8C2A-E1F023EBB911.jpeg
    F0C1C56C-DF3F-4799-8C2A-E1F023EBB911.jpeg
    90.1 KB · Views: 74
  • 473F7769-74E2-4934-A99E-4720A59DC6AC.jpeg
    473F7769-74E2-4934-A99E-4720A59DC6AC.jpeg
    139.3 KB · Views: 71
  • 2FDACDCA-E532-44C3-B3D4-8D2F7A3C0B8B.jpeg
    2FDACDCA-E532-44C3-B3D4-8D2F7A3C0B8B.jpeg
    84.6 KB · Views: 71
  • 389CF46D-EDCE-420F-95B3-7DDEAFFB1A7E.jpeg
    389CF46D-EDCE-420F-95B3-7DDEAFFB1A7E.jpeg
    207.8 KB · Views: 62
  • F26C4D72-DDDC-466F-9EAD-EED9C97852E5.jpeg
    F26C4D72-DDDC-466F-9EAD-EED9C97852E5.jpeg
    188 KB · Views: 67
  • 39D0A7A0-1D8B-4388-833C-4C9890B443D8.jpeg
    39D0A7A0-1D8B-4388-833C-4C9890B443D8.jpeg
    135.3 KB · Views: 64
  • 54EECA1C-56D3-4362-BB0E-58B49E59F9A4.jpeg
    54EECA1C-56D3-4362-BB0E-58B49E59F9A4.jpeg
    146.2 KB · Views: 62
  • A0C85BD7-6A4D-4FA6-9F6D-CF999D71F805.jpeg
    A0C85BD7-6A4D-4FA6-9F6D-CF999D71F805.jpeg
    172.4 KB · Views: 68
  • Like
  • Fire
Reactions: 26 users
Top Bottom