BRN Discussion Ongoing

Tothemoon24

Top 20
IMG_8690.jpeg





Nvidia’s H100 AI #GPUs are taking the tech world by storm, but their reign comes at the price of a hefty energy bill.

According to a report from #CBInsights and #Stocklytics.com, these power-hungry processors are projected to consume a staggering 13,797 GWh in 2024, exceeding the annual energy consumption of nations like #Georgia and #CostaRica.

Imagine this, a legacy data center consumes 10 kW/rack where #CyrusOne, #KKR owned leading global data center operator and developer specialising in #AI applications, consumes 300 kW/rack!

But why do #GPUs consume so much power?

Data center #GPUs consume a substantial amount of power primarily due to their high computational requirements and the complex algorithms they handle. These #GPUs optimize parallel processing tasks like #machinelearning and #dataanalytics, involving simultaneous processing of vast amounts of data.

While #parallel processing speeds up data processing, one demerit is that, at a time most parts of a chip are active. This constant computation, coupled with the execution of complex algorithms, demands significant computational power, thereby increasing energy consumption.

The large-scale deployment of #GPUs in data centers, where racks and clusters utilize hundreds or thousands of #GPUs further amplify their collective power consumption. This combination of factors underscores the considerable energy consumption associated with data center GPUs.

Successfully navigating these challenges and fostering innovation will shape the future landscape of #AI computing.

So, what options do we have?

● #Amazon, frenemy to Nvidia, recently unveiled Arm based Graviton4 and Trainium2 chips holds promise for efficiency gains.

● In the near to medium term, #Neuromorphic computing is being researched aggressively as an alternative to synchronous parallel computing architectures. Neuromorphic computing is an asynchronous computing paradigm which runs on event based ‘spikes’ rather than a clock signal. And drastically lowers the power consumption.

● Big money is going into enabling tech like liquid cooling - #KKR acquired CoolIT Systems for $270 mn and Bosch acquired Jetcool through its venture arm

While CooIT Systems becomes the supplier for Cyrus One, #KKR makes money on both!
 
  • Like
  • Fire
  • Thinking
Reactions: 11 users

Reuben

Founding Member
  • Like
  • Fire
Reactions: 16 users

jtardif999

Regular
  • Like
Reactions: 1 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
what a company, they have a few of our other execs also... i'm sure there is a logical explanation.

View attachment 59874


I'm pretty sure the Financial Controller, Michelle Afflick is Ben Afflick's sister.
 
  • Haha
  • Like
Reactions: 12 users

itsol4605

Regular
  • Haha
Reactions: 6 users

Shadow59

Regular
  • Haha
  • Like
Reactions: 10 users
Presuming already posted but too lazy to check search this time but if not, nice for ARM Design Engineer to acknowledge us in the mix...difference is out of that group, we can integrate both DVS & DAS from multiple sensors :)



Dynamic Vision Sensor (DVS) and the Dynamic Audio Sensor (DAS)​


Kailash Prasad

Kailash Prasad​

Design Engineer @ Arm | PMRF | IIRF | nanoDC Lab…

Published Jan 16, 2024
+ Follow
Have you ever wondered how the human eye👁️ and ear👂 can process complex and dynamic scenes with such high speed and accuracy? Imagine if we could design artificial sensors that mimic the biological mechanisms of vision and hearing, and produce data that is more efficient and meaningful than conventional sensors.

In this post, I will introduce you to two types of neuromorphic sensors: the Dynamic Vision Sensor (DVS) and the Dynamic Audio Sensor (DAS).

These sensors are inspired by the structure and function of the retina and the cochlea, respectively, and use a novel paradigm of event-based sensing. Unlike conventional sensors that capture frames or samples at a fixed rate, event-based sensors only output data when there is a change in the input signal, such as brightness or sound intensity. This results in a stream of asynchronous events that encode the temporal and spatial information of the scene, with high temporal resolution, low latency, and high dynamic range.

📖 - "In simpler terms, these special sensors work like our eyes and ears. They're designed based on the way our eyes' retinas and ears' cochleae function. But what sets them apart is their unique approach called event-based sensing. Unlike regular sensors that take pictures or recordings at a set speed, these event-based sensors only provide information when there's a change. Whether it's a shift in light or a change in sound, they only capture those moments. Instead of a constant flow of data, you get quick updates that show when and where things change. This gives you highly detailed and fast information about what's happening, with minimal delay and a wide range of details. It's like having sensors that focus on the important stuff, making them efficient and responsive."

The DVS is an imaging sensor that responds to local changes in brightness, and outputs events that indicate the pixel address, the polarity (increase or decrease) of the brightness change, and the timestamp. The DVS can achieve a temporal resolution of microseconds⏱️, a dynamic range of 120 dB🔊, and a low power consumption of 30 mW💡. The DVS can also avoid motion blur and under/overexposure that plague conventional cameras. The DVS can be used for applications such as optical flow estimation, object tracking, gesture recognition, and robotics.

The DAS is an auditory sensor that mimics the cochlea, the auditory inner ear. The DAS takes stereo audio inputs and outputs events that represent the activity in different frequency ranges. The DAS can capture sound signals with a frequency range of 20 Hz to 20 kHz🎵, a dynamic range of 60 dB🔊, and a temporal resolution of microseconds⏱️. The DAS can also extract auditory features such as interaural time difference, harmonicity, and speaker identification.

Both the DVS and the DAS are compatible with neuromorphic computing architectures, such as spiking neural networks, that can process the event data in a parallel and distributed manner. This enables low-power and real-time computation of complex tasks such as scene understanding, speech recognition, and sound localization.

Some examples of recent products that use the DVS and the DAS are:

- The Prophesee Metavision Camera, which is a high-resolution DVS camera that can capture fast and complex motions with minimal data and power consumption.

- The Samsung ISOCELL Slim 3P9, which is a smartphone camera sensor that incorporates a DVS mode to enable fast autofocus and video stabilization.

- The iniVation Dynamic Audio Sensor USB board, which is a binaural DAS board that can be interfaced with standard PCs for sound analysis and processing.

- The Brainchip Akida Neuromorphic System-on-Chip, which is a low-power and scalable chip that can integrate multiple DVS and DAS sensors and perform event-based learning and inference.

I hope this post has given you some insights into the exciting field of neuromorphic sensing, and how it can revolutionize the way we perceive and interact with the world. Thank you for reading!

P.S.: This post is based on one of the tutorials on Neuromorphic Computing at the VLSI Design Conference that happened last week.
 
  • Like
  • Fire
  • Love
Reactions: 48 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
Last edited:
  • Haha
Reactions: 4 users

Fox151

Regular
Well said Hop.
Yes, the timeframe is painful, and patience and trust can really test some. As you say, it just depends on peoples circumstances. When that hockey stick appears in revenue, thats when I will start planning a holiday, and Oh, I need a holiday.
Just my own simple opinion, I think we have been told all we need, if you can apply some patience and trust.
View attachment 59821
"WATCH THE FINANCIALS"
Which financials though?!?!
 
  • Like
Reactions: 1 users

Reuben

Founding Member
Interesting what this Wavious actually does if Rob is working for them now.
For anyone speculative... answers in screenshot
 

Attachments

  • Screenshot_20240326_173621_LinkedIn.jpg
    Screenshot_20240326_173621_LinkedIn.jpg
    451.4 KB · Views: 355
  • Like
  • Love
  • Haha
Reactions: 34 users

Quercuskid

Regular
  • Haha
Reactions: 5 users
What will you do if there is no new sales by the next 1/4 Ly
 

Jumpchooks

Regular
As part of my OCD recovery, I have a salad every now and again, and again, and ...

Onion rings - for goodness sake, who thought of that? ... you have to cut up 10 onions to get 10 rings the same size.
I would suggest that you try and get a ring from both top and bottom half from each onion. Therefore, 5 Onions can produce 10 rings of matching diameter. 😎
 
  • Haha
  • Like
  • Thinking
Reactions: 8 users

7für7

Top 20
For anyone speculative... answers in screenshot
To bring speculations again to life….Thats obviously an AI generated conversation. Too smooth in my opinion! Don’t believe everything posted on internet! (Again sarcasm off.. I have to write it because some people take everything serious!
 
  • Like
  • Haha
Reactions: 2 users

IloveLamp

Top 20
  • Like
  • Love
Reactions: 4 users

miaeffect

Oat latte lover
  • Like
  • Haha
  • Love
Reactions: 23 users

TheDon

Regular
  • Like
  • Fire
  • Love
Reactions: 9 users

IloveLamp

Top 20
  • Like
  • Fire
  • Love
Reactions: 10 users

Wags

Regular
Which financials though?!?!
Gidday @Fox151
Im the optimist, and forced to be a realist.
The time frame is way longer than I anticipated and I have accepted that, and my naivety.
The world has seen the headwinds from Covid and semiconductor shortages to wars.
That said, I have seen a continual increase in the market activity and discussion regarding neuromorphic computing. The things we talked about years ago, data centre overloads, power consumption blowouts, clever battery operated sensors etc, seem to be coming to the fore.
I am very far from a technical expert, but I trust in my early research, gut and amazement of the tech. The SP is not what we want to see, but the evolution of the tech seems to be playing out, albiet slower than first thought.
Call me blind, or a fool, or whatever, Im not phased.
Until i read something confirmed to the contrary, all I read makes me feel good about this investment. (except for the timeframe).
Only a few days left in this qtr, so Yes, Im hoping for signs of growing revenue in this qtr. How bout you?
 
  • Like
  • Love
  • Fire
Reactions: 41 users
Top Bottom