BRN Discussion Ongoing

Hi All
I have just made two minor edits to substitute the Table 6. which I could not size at reference 3. thanks to Bravo and by adding reference 9. thanks to Diogenese.
Regards
FF
but remember DYOR
Great post before collating a lot of relevant info.

Was reading this blog post last night and would appear you and this writer are on a similar page.

Too big to post all of it but the research definitely worth a read on the analysis of NVIDIA and pondering where to next....I just copied the concluding paragraphs.



The Engine for AI and The Catalyst Of The Future​

A pure strategic masterclass from Nvidia's trillion-dollar accelerated computing empire​


JARYD HERMANN
7 SEPT 2023


👋 Welcome to How They Grow, my newsletter’s main series. Bringing you in-depth analyses on the growth of world-class companies, including their early strategies, current tactics, and actionable product-building lessons we can learn from them.


Final thoughts in the blog.

For the curious…

What could be next for Nvidia? Neuromorphic Computing? 🧠




Okay, so to understand neuromorphic computing (NC), all you need to know is this formula:




Juust kidding.

In short, the answer is much simpler: The future of chips is making physical computers think more like human brains, where chips use the same physics of computation as our own nervous system.

The term "neuromorphic" comes from the Greek words "neuron" (meaning nerve cell) and "morphe" (meaning form). In the context of computing, it refers to the use of electronic circuits and devices inspired by biological neurons' structure and function.

While still nascent, this is another massive technological feat. You might be thinking—but we’ve done something similar already with AI neural networks, what’s the difference?

Simply, we’ve made great progress on the software side of things in terms of mimicking the logic of how a human brain thinks. But solving those challenges on a physical chip is a different beast.

That’s what NC is solving though, and you can imagine how much more advanced our AI and computing will be when we have the chips and software both working in unison like a brain. 🧠

And just to illustrate the monumental difference on the chips’ side:

  • Your computer operates in binary. That’s 0s and 1s; Yes and Nos. It’s rigid, so, the code we use and the questions we ask these kinds of machines must be structured in rigid way.
  • With NC though, we go from rigid to flexible, as these chips will bring computers to the ability to have a gradient of understanding. I’m no engineer, but from what I’ve read, that’s huge.
Here are a few of the breakthrough benefits NC has pundits excited about:

In more succinct words: Neuromorphic Computing is the key to huge leaps in AI advancements.

Just imagine how fast things would be changing in the AI landscape if these were powering Nvidia’s data centers and AI platforms.


It sounds somewhat far-fetched, but that future is already here and in the works.

As was published in this research paper in Nature:

With the end of Moore’s law approaching and Dennard scaling ending, the computing community is increasingly looking at new technologies to enable continued performance improvements. Neuromorphic computers are one such new computing technology. The term neuromorphic was coined by Carver Mead in the late 1980s1,2, and at that time primarily referred to mixed analogue–digital implementations of brain-inspired computing; however, as the field has continued to evolve and with the advent of large-scale funding opportunities for brain-inspired computing systems such as the DARPA Synapse project and the European Union’s Human Brain Project, the term neuromorphic has come to encompass a wider variety of hardware implementations.
And builders are going after it. Intel is already working on these chips, as are various other startups.

In the long term, NC poses a technological obsolescence risk to traditional GPUs and DPUs. If these types of chips become successful, it could threaten Nvidia's business.

However, because NC has the potential to be a game-changer in many different areas of society, and its consequences could be far-reaching and complex, I have zero doubt in my mind that Jensen and his crew are sitting in a Denny’s somewhere, dice in hand, and mapping out Nvidia’s strategic future over some burnt coffee. ☕


And that, folks, brings us to the end of our Nvidia analysis.

Given how large and complex Nvidia is as a company, ironically, I deliberately kept this shorter than other deep dives so we didn’t risk getting lost in the weeds.

I hope that strategy paid off and that you found this post insightful and enjoyable. 🙏

If you did—and you made it all the way down here to the deep dark footer section—I’d be incredibly grateful if you gave this post a like, share, or just spread the word about HTG to some friends.
 
  • Like
  • Love
  • Fire
Reactions: 57 users

Esq.111

Fascinatingly Intuitive.
Can Nvidia Survive the 4th Industrial Revolution?
by Fact Finder


Though Nvidia is riding high at the moment all indicators are that it has positioned itself on the wrong side of technology history.

While Nvidia has been compressing models to stave off the end of Moore’s Law it’s continued preoccupation with its Von Neumann market dominance has seen it embrace the false dawn offered to it by Large Language Model’s and the cult of CHATGpt.

The fragile nature of Nvidia’s technology future has been exposed in the last week by a small Australian technology company that has been stealthily developing an entirely new, some have said science fiction solution, to the energy resource issue exposed by the power and cost involved in training and running Large Language Models represented by CHATGpt.

The World has been fantasising about what is called Edge Computing for over a decade. The principle underpinning Edge Computing is actually very simple and can best be understood by what might be considered a strange example.

I am sure you have heard of terms like Food Miles, Buy Local, Eat Local, Grow Your Own as ways to decarbonise and save the planet. The simple indisputable proposition being if you reduce the distance between you and your food sources point of production the reduction in transport will reduce the energy consumed in putting the food on your plate.

Putting a bunch of flowers cut from your own garden on the sideboard is infinitely more fuel efficient than trucking, flying, trucking, driving fresh picked flowers from Europe around the globe to you in Australia.

Cutting asparagus in your kitchen garden is infinitely more efficient than buying asparagus from your Local Supermarket that has been cut in Peru and transported to you in Australia.

Now in the above examples I have chosen two products that require refrigeration to keep them fresh after picking to ensure they arrive at your home still useable and which as a result requires transport by jet airliners.

Suffice to say it is immediately obvious that if you are trying to reduce carbon and cost processing your flowers and asparagus at home wins hands down as zero carbon versus tonnes is a no brainer.

Now I know there are practical limitations making this solution difficult for those of us with black thumbs or living in home units to embrace. But that is an argument for another day.

The point is that this is what Edge Computing is all about. It is about reducing compute miles and in so doing cutting dramatically the cost of doing compute and carbon emissions.

For example take a Smart Doorbell. Currently a Smart Doorbell needs to be constantly connected via your homes wireless network to carry out its function of identifying and alerting you to the presence of someone at the door.

In lay terms 24/7 it sits there constantly processing camera frames showing the brick wall next to your front door and sending to the cloud over and over and over an image of the brick wall and receiving back message after message that no one is at the front door. Now you can reduce the power by slowing down the number of photos/frames it takes every second to monitor for someone coming to your front door but if the gap becomes too great between frames someone can come and go in that gap and avoid detection. So this method has built in limitations even when working to design but not so well when bandwidth is congested or connection breaks down.

Enter the neuromorphic Edge Computing revolution.

Edge Computing is about as I said reducing the compute miles. By placing the compute close as possible to the Smart Doorbell if not right up against it you immediately reduce the distance between the camera/sensor and the compute/intelligence. This has the advantage of reducing power consumption, reducing latency (time it takes to send the message back and forth from the camera/sensor to the data centre) and preventing congesting the bandwidth with photos of a blank brick wall affecting your ability to stream Netflix or Sky Sport.

Everywhere Edge Computing is being spoken about and Nvidia as the dominant player in the computing space is calling out about its Edge Computing solutions.

There is probably not one person on the planet with any interest in computing who has not heard of Nvidia or its Nvidia Jetson range of edge computing solutions.

Indeed, the Nvidia Jeson range is a leader in this space across the globe. Its Jetson solutions are to be found everywhere but for how much longer can Jetson dominate for Nvidia when it is hamstrung by old thinking in a World that is transitioning towards the Fourth Industrial Revolution.

So let's take a quick look at what Nvidia publishes about the Nvidia Jetson AGX Orin series, the Jetson Orin NX series and the Jetson Orin Nano series by reference to the advertised performance figures.

Power ranges from 5 watts to 60 Watts

“Power 15W - 60W 15W - 75W 15W - 40W 10W - 25W 10W - 20W 7W - 15W 7W - 10W”

TOPs ranges from 20 TOPs to 275 TOPs

AI Performance 275 TOPS 248 TOPS 200 TOPS 100 TOPS 70 TOPS 40 TOPS 20 TOPS


These numbers would certainly seem impressive to those who were feeding punch cards into the first IBM main frame computer even if you added on the power required to externally cool Jetson running some form of external cooling such as fans.

As impressive as these numbers are they clearly do not offer a power budget that can be reasonably embraced by those looking for an Edge Computing solution. Nvidia have to their credit recognised this and in consequence introduced the Jetson Nano TX2 Series and boasts that the Jetson TX2i, the Jetson TX2, the Jetson TX2 4GB and the Jetson TX2 NX come with AI Performance or 1.26 TFLOPS to 1.33 TFLOPS however the power budget ranges from 7.5 Watts to 20 Watts and on top of these numbers you need to allow for external cooling. You might have noticed that while Nvidia has reduced the form factor the power required remains in the multi watt region. (1)

To address these failings a new form of computing referred to as spiking neural network compute is being championed by Intel and IBM and they have over the last decade reached the point of proving out in research chips the huge benefits to be had by embracing this new style of compute not the least of which is a massively reduced power budget. The research at Intel and IBM goes on apace.

Enter stage left this little-known Australian company that has been listed on the Australian Stock Exchange since 2015. This tiny company with less than 100 employees has quietly some would say stealthily gone about its business and has beaten Intel and IBM to the punch launching its first commercial spiking neural network engineering chip in 2020 and is shortly to release its second generation technology, which it reports on its website as being capable of the following performance figures across three models or iterations of this technology advancement:

ITERATION ONE:

Max
Efficiency


Ideal for always-on, energy-sipping Sensor Applications:

  • Vibration Detection
  • Anomaly Detection
  • Keyword Spotting
  • Sensor Fusion
  • Low-Res Presence Detection
  • Gesture Detection
Extremely Efficient
@Sensor Inference


Either Standalone or with Min-spec MCU.

Configurable to ideal fit:

  • 1 – 2 nodes (4 NPE/node)
  • Anomaly Detection
  • Keyword Spotting
Expected implementations:

  • 50 MHz – 200
  • MHz Up to 100 GOPs
Additional Benefits

Eliminates need for CPU intervention

Fully accelerates most feed-forward networks

  • Optional skip connection and TENNs support for more complex networks
  • Completely customizable to fit very constrained power, thermal, and silicon area budgets
  • Enables energy-harvesting and multi-year battery life applications, sub milli-watt sensors

INTERATION 2:

Sensor
Balanced


Accelerates in hardware most Neural Network Functions:

  • Advanced Keyword Spotting
  • Sensor Fusion
  • Low-Res Presence Detection
  • Gesture Detection & Recognition
  • Object Classification
  • Biometric Recognition
  • Advanced Speech Recognition
  • Object Detection & Semantic Segmentation
Optimal for Sensor Fusion
and Application SoCs


With Min-Spec or Mid-Spec MCU.

Configurable to ideal fit:

  • 3 – 8 nodes (4 NPE/node) 25 KB
  • 100 KB per NPE
  • Process, physical IP and other optimizations
Expected implementations:

  • 100 – 500 MHz
  • Up to 1 TOP
Additional Benefits

  • CPU is free for most non-NN compute
  • CPU runs application with minimal NN-management
  • Completely customizable to fit very constrained power, thermal and silicon area budgets
  • Enables intelligent, learning-enabled MCUs and SoCs consuming tens to hundreds of milliwatts or less

ITERATION 3:

Max
Performance


Detection, Classification, Segmentation, Tracking, and ViT:

  • Gesture Detection
  • Object Classification
  • Advanced Speech Recognition
  • Object Detection & Semantic Segmentation
  • Advanced Sequence Prediction
  • Video Object Detection & Tracking
  • Vision Transformer Networks
Advanced Network-Edge Performance
in a Sensor-Edge Power Envelope


With Mid-Spec MCU or Mid-Spec MPU.

Configurable to ideal fit:

  • 8 – 256 nodes (4 NPE/node) + optional Vision Transformer
  • 100 KB per NPE
  • Process, physical IP and other optimizations
Expected implementations:
  • 800 MHz – 2 GHz
  • Up to 131 TOPs
Additional Benefits
  • CPU is free for most non-NN compute
At this stage I cannot comment on the power budget of these iterations however we do know that the first released chip the AKD1000 which was able to retail for about $US25.00 had a power budget that ran in the micro to milliwatts and was claimed by Edge Impulse (5), Quantum Ventura (3 & 4) and Tata Consulting Services (1) to outperform a GPU from Nvidia by some considerable margin across all performance measurements and this new version is an advancement grown out of the underlying neural fabric supporting AKD1000.

Perhaps the most worrying recent doomsday prediction for Nvidia at the Edge came from Tata Elxsi’s Mr. Sunil Nair Vice President EMEA and Design Digital who posted on his LinkedIn page right beside a post about Tata Elxsi partnering with Nvidia in the Cloud the following:

“Cloud computing is commodity. Edge is where the action is.

Thrilled to see Tata Elxsi and Brainchip partner to enable and integrate ultra-low power neuromorphic processors for use cases that would bring huge savings and transform citizen experience. (especially the ones over spending on Nvidia.)”


Mr. Nair has been with Tata Elxsi since 1997. While the partnership with Tata Elxsi has only recently been announced Brainchip has been working with Tata Consulting Services (1) TATA Groups research arm since at least 2019 when they jointly presented AKD1000 performing a live gesture recognition demonstration. Since that time Tata Consulting Services has released a number of peer reviewed papers covering the use of AKD1000 and it can be said that Mr. Nair would be very well informed when it comes to the benefits that Brainchip’s AKIDA technology solutions can bring to the Edge.

The full release of the next generation referred to as AKIDA 2.0 up till today has been restricted to a number of select customers however the company has recently advised in a CEO investor presentation that the full public release is imminent. This prediction seems to be holding true as in the past week Brainchip’s website has been updated with substantial information regarding AKIDA 2.0 signalling it is getting close to the launch date.

The interesting aspect of Brainchip Inc is that while it has remained largely unknown to the general public and the investment world in its quiet way it has been accumulating a very long and impressive list of corporate and academic engagements including the following publicly acknowledged group and according to Mr. Rob Telson Vice President of Ecosystems & Partnerships, they have hundreds of companies testing AKIDA technology boards:

1. FORD, 2. VALEO, 3. RENESAS, 4. NASA, 5. TATA Consulting Services, 6. MEGACHIPS, 7. MOSCHIP, 8. SOCIONEXT, 9. PROPHESEE, 10. VVDN, 11. TEKSUN, 12. Ai LABS, 13. NVISO, 14. EMOTION 3D, 15. ARM, 16. EDGE IMPULSE, 17. INTEL Foundries, 18. GLOBAL FOUNDRIES, 19. BLUE RIDGE ENVISIONEERING, 20. MERCEDES BENZ, 21. ANT 61, 22. QUANTUM VENTURA, 23. INFORMATION SYSTEM LABORATORIES, 24. INTELLISENSE SYSTEMS, 25. CVEDIA, 26. LORSER INDUSTRIES, 27. SiFIVE, 28. IPRO Silicon IP, 29. SALESLINK, 30. NUMEN, 31. VORAGO, 32. NANOSE, 33. BIOTOME, 34. OCULI, 35. Magik Eye, 36. GMAC, 37. TATA Elxsi, 38. University of Oklahoma, 41. Arizona State University, 42. Carnegie Mellon University, 43. Rochester Institute of Technology, 44. Drexel University, 45. University of Virginia.

It should be noted that Brainchip has been at pains in its literature and presentations to explain that the AKIDA technology is processor and sensor agnostic and being fully digital, scalable, and portable across all foundries. The AKD1000 was produced successfully first time and every time in 28nm at TSMC and only recently the AKD1500 was received back from Global Foundries successfully fabricated in 22nm FDSOI first time and every time.

The individual who gave the underlying AKIDA technology life is Peter van der Made who is also one of the founders of Brainchip and his full vision is to create a beneficial form of Artificial General Intelligence by about 2030. This vision plays out in a series of steps and AKIDA 3.0 is presently in development and according to the company’s CEO Sean Hehir in a very recent investor presentation each step is targeted to take 12 to 18 months. Historically Peter van der Made and his cofounder Anil Mankar have impressed with their capacity to deliver on their technology development time lines and by always having a little extra what they like to call secret sauce with each new technology reveal.

This little extra secret sauce with AKIDA 2.0 was the release of the TeNNs (6 & 9) and ViT (7) which provide an unprecedented leap into the future from what even the most optimistic expected to be possible at the far Edge using energy harvesting to power these devices. It is impossible to do justice to what they bring to the Edge Compute revolution in this article but fortunately even though patents are pending Brainchip has published a White Paper (6) and videos providing easy to follow plain English explanations. (7)

By the way anyone up for a bit of regression analysis (8) using AKIDA technology Brainchip also has that covered. When others were opining that spiking neural networks could not do regression analysis Brainchip Inc was demonstrating it running on AKD1000 for monitoring vibration in rail infrastructure.

There is so much to delight those who love to read about and explore science fiction becoming reality when peeling back the petals of the rose that is the Brainchip AKIDA technology revolution.

In concluding in Australia the ignorant and poor of intellect have treated at times the vision of Peter van der Made with a savagery of doubt usually reserved for those who claim to have been abducted by aliens and as is usually the case these critics have been members of the so called sophisticated investor class and even though they have little credibility in their areas of claimed expertise they drown in their own ignorance when it comes to the science of neuromorphic computing. If tempted to listen to such individuals about the science of neuromorphic computing one is well served to remember the life of Robert Goddard - https://www.msn.com/en-au/news/aust...1&cvid=297d1358fb8c45ea9b1ae445a4985d75&ei=51

REFERENCES:

1
.Low Power & Low Latency Cloud Cover Detection in Small Satellites Using On-board Neuromorphic Processors

Chetan Kadway, Sounak Dey, Arijit Mukherjee, Arpan Pal, Gilles Bézard

2023 International Joint Conference on Neural Networks (IJCNN), 1-8, 2023

Emergence of small satellites for earth observation missions has opened up new horizons for space research but at the same time posed newer challenges of limited power and compute resource arising out of the size & weight constraints imposed by these satellites. The currently evolving neuromorphic computing paradigm shows promise in terms of energy efficiency and may possibly be exploited here. In this paper, we try to prove the applicability of neuromorphic computing for on-board data processing in satellites by creating a 2-stage hierarchical cloud cover detection application for multi-spectral earth observation images. We design and train a CNN and convert it into SNN using the CNN2SNN conversion toolkit of Brainchip Akida neuromorphic platform. We achieve 95.46% accuracy while power consumption and latency are at least 35x and 3.4x more efficient respectively in stage-1 (and 230x & 7x in stage-2) compared to the equivalent CNN running on Jetson TX2.

https://ieeexplore.ieee.org/abstract/document/10191569/

2.An energy-efficient AkidaNet for morphologically similar weeds and crops recognition at the Edge

Vi Nguyen Thanh Le, Kevin Tsiknos, Kristofor D Carlson, Selam Ahderom

2022 International Conference on Digital Image Computing: Techniques and Applications (DICTA), 1-8, 2022

Wild radish weeds have always been a persistent problem in agriculture due to their quick and undesirable spread. Therefore, the accurate identification and effective control of wild radish in canola crops at early growth stage play an indispensable role in reducing herbicide rates and enhancing agricultural productivity. In this paper, an energy efficient and lightweight AkidaNet model is developed to accurately identify broad-leaf weeds and crops with similar morphology at four different growth stages. Experiments performed on a published bccr-segset dataset show that our proposed method achieves competitive performance, a classification accuracy of 99.73%, compared to several well-known CNNs architectures. Next, we quantized and converted the model into a Spiking Neural Network for implementation on a spike-based neuromorphic hardware device. The converted model is not only superior in low-latency and low-power consumption but also retains a similar accuracy to the original model. We also employ Grad-CAM to validate whether our model focuses on important features in plant images to identify wild radish weeds in crops

https://www.researchgate.net/profil...r-weeds-and-crops-recognition-at-the-Edge.pdf


3.Table 6.

View attachment 44776

https://www.sciencedirect.com/scien...c65de04&pid=1-s2.0-S1877050922017860-main.pdf (Page 494)

4. “In this federally funded phase 2 program, Quantum Ventura is creating state-of-the-art cybersecurity applications for the U.S. Department of Energy under the Small Business Innovation Research (SBIR) Program. The program is focused on “Cyber threat-detection using neuromorphic computing,” which aims to develop an advanced approach to detect and prevent cyberattacks on computer networks and critical infrastructure using brain-inspired artificial intelligence.

“Neuromorphic computing is an ideal technology for threat detection because of its small size and power, accuracy, and in particular, its ability to learn and adapt, since attackers are constantly changing their tactics,” said Srini Vasan, President and CEO of Quantum Ventura Inc. “We believe that our solution incorporating Brainchip’s Akida will be a breakthrough for defending against cyber threats and address additional applications as well.””

https://brainchip.com/brainchip-and-quantum-ventura-partner-to-develop-cyber-threat-detection/

5. Running the out-of-the-box demos on the Akida Raspberry Pi development kit I have was very impressive achieving, according to the statistics, approximately 100 FPS for a 9-mW power dissipation.

All told I am very impressed with the BrainChip Akida neuromorphic processor. The performance of the networks implemented is very good, while the power used is also exceptionally low. These two parameters are critical parameters for embedded solutions deployed at the edge.

Project links

  1. Visal wake word: https://studio.edgeimpulse.com/studio/224143
  2. Anomaly detection: https://studio.edgeimpulse.com/studio/261242
  3. CIFAR10: https://studio.edgeimpulse.com/studio/257103
  4. Keyword spotting: https://studio.edgeimpulse.com/studio/257193
Adiuvo is a consultancy that provides embedded systems design, training, and marketing services. Taylor is Founder and Principal Consultant of the company, teaches about embedded systems at the University of Lincoln, and is host of the podcast “The Embedded Hour.”

https://www.edgeimpulse.com/blog/brainchip-akida-and-edge-impulse

6.https://brainchip.com/temporal-event-based-neural-networks-a-new-approach-to-temporal-processing/

7.

8.https://brainchip.com/brainchip-demonstrates-regression-analysis-with-vibration-sensors/

9.
https://brainchip.com/wp-content/uploads/2023/06/TENNs_Whitepaper_Final.pdf

Pg 8:
Unlike standard CNN networks that only operate on the spatial dimensions, TENNs contain both temporal and spatial convolution layers. They may combine spatial and temporal features of the data at all levels from shallow to deep layers. In addition, TENNs efficiently learn both spatial and temporal correlations from data in contrast with state-space models that mainly treat time series data with no spatial components. Given the hierarchical and causal nature of TENNs, relationships between elements that are both distant in space and time may be constructed for efficient continuous data processing (such as video, raw speech, and medical data).


My opinion only so DYOR

Good Afternoon Fact Finder ,

Very nice 👌.

One cannot really add much to such a peice ...

Bugger it , some loose numbers, which I have posted before.

* Bloody hard work simultaneously taking a photo with one hand , pointing with the other ( FAIR PRICE RANGE ) indicater, whilst under the influence.
Rest assured , I shall not be operating any heavey machinery today.


Regards,
Esq.
 

Attachments

  • 20230917_130055.jpg
    20230917_130055.jpg
    4.9 MB · Views: 249
  • Like
  • Love
  • Fire
Reactions: 46 users
I find it really strange that the big head has been responsible for at least 5 decent contributing posters to leave this forum.🤔
Whatever reason @zeeb0t has for not sending him to the naughty corner if we all ignore hiim it has the same effect! I put them on ignore a long time ago.
 
  • Like
  • Fire
  • Love
Reactions: 10 users

IloveLamp

Top 20
  • Like
  • Fire
Reactions: 3 users

Tothemoon24

Top 20
Below link is a recently released Neuromorphic medical paper .
24 pages , very insightful !
Wearables, hearing aids ,vision , ect .
Maybe an enjoyable Sunday evening read for those interested

Click link Botton left corner

IMG_7543.jpeg



IMG_7545.jpeg




 
  • Like
  • Love
  • Fire
Reactions: 30 users

Diogenese

Top 20
Whatever reason @zeeb0t has for not sending him to the naughty corner if we all ignore hiim it has the same effect! I put them on ignore a long time ago.
So are we talking about a one-trick-pony suffering from relevance deprivation?
 
  • Haha
  • Like
Reactions: 9 users
  • Like
  • Love
  • Fire
Reactions: 45 users

Gies

Regular
IMG_8780.jpeg
 
  • Like
  • Fire
  • Haha
Reactions: 7 users

Gies

Regular
Interesting Thomas Hülsing replied on my remark !
 
  • Like
  • Love
  • Fire
Reactions: 25 users

MDhere

Regular
Interesting Thomas Hülsing replied on my remark !
well Adam Osseriran from Brainchip liked it so its a yes from me :)
 
  • Like
  • Fire
  • Love
Reactions: 25 users

Gies

Regular
This fantastic technology!
 

Attachments

  • IMG_8781.png
    IMG_8781.png
    397 KB · Views: 286
  • Like
  • Fire
  • Love
Reactions: 18 users

Dallas

Regular
  • Like
  • Fire
Reactions: 9 users

Dallas

Regular
Warum erwähnt ein führender Mercedes-Entwickler diese Türklingel und warum erwähnt BrainChip immer wieder Türklingeln ❗🤠👋🤔
 
  • Like
  • Thinking
  • Fire
Reactions: 11 users

Cartagena

Regular
Interesting Thomas Hülsing replied on my remark !
Hi Gies, great work, as an airline captain in 777 and airbus I understand you would have a lot of knowledge in high tech avionics systems and Brainchip's predictive maintenance I guess is the main potential application for the avionics area. Awesome work and very exciting that Thomas Hulsing replied to your comment 👍
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 21 users

Tothemoon24

Top 20
IMG_7547.jpeg


Artificial Intelligence Computing Challenge

CPU/GPU platforms vs. Neuromorphic Versions
– Price
– Power Consumption
– Size

...and who do think is the winner in all categories...?

Here you can find detail information:

pg. 494
We roughly estimate the power draw of one chip to be between 100 mW and 1 W based on BrainChip’s implementation of a network to process the CIFAR dataset.
In addition, BrainChip’s processor may fit into USB keys, making them amenable to deployment in handheld/portable units which DHS seeks.
On the other hand, the 3D CNN run on a GPU would require much more SWaP-C.
In Table 6, we detail the SWaP-C differences between CPU/GPU platforms and neuromorphic versions.

#edgecomputing #neuromorphic #computing BrainChip

IMG_7548.jpeg
 
  • Like
  • Fire
  • Love
Reactions: 52 users

Cartagena

Regular
Hi Gies, great work, as an airline captain in 777 and airbus I understand you would have a lot of knowledge in high tech avionics systems and Brainchip's predictive maintenance I guess is the main potential application for the avionics area. Awesome work and very exciting that Thomas Hulsing replied to your comment 👍

View profile for Thomas Hülsing
Thomas Hülsing 3rd+
System-of-Systems Engineering
32m
Follow

Artificial Intelligence Computing Challenge CPU/GPU platforms vs. Neuromorphic Versions – Price – Power Consumption – Size ...and who do think is the winner in all categories...? Here you can find detail information: https://lnkd.in/e7gz7ZHU pg. 494 We roughly estimate the power draw of one chip to be between 100 mW and 1 W based on BrainChip’s implementation of a network to process the CIFAR dataset. In addition, BrainChip’s processor may fit into USB keys, making them amenable to deployment in handheld/portable units which DHS seeks. On the other hand, the 3D CNN run on a GPU would require much more SWaP-C. In Table 6, we detail the SWaP-C differences between CPU/GPU platforms and neuromorphic versions. #edgecomputing #neuromorphic #computing BrainChip
  • No alternative text description for this image
  • Image



11 Comment
Like
CommentShare
 

Attachments

  • Screenshot_20230917-220604.png
    Screenshot_20230917-220604.png
    513.7 KB · Views: 79
Last edited:
  • Like
  • Love
  • Fire
Reactions: 27 users

IloveLamp

Top 20
  • Like
  • Fire
  • Thinking
Reactions: 22 users

cosors

👀
Please don't misunderstand, I'm laughing because your markers aren't still as gallant as Bravo's and it looks more like a screen break on my current device. Thanks for your contribution!

____
I appreciate that very much! So my laughter is not serious when it comes to the matter at hand.
 
Last edited:
  • Haha
  • Like
Reactions: 5 users

Gies

Regular
Please don't misunderstand, I'm laughing because your markers aren't still as gallant as Bravo's and it looks more like a screen break on my current device. Thanks for your contribution!

____
I appreciate that very much! So my laughter is not serious when it comes to the matter at hand.
No worries mate.
I’m not very handy with those things.
It’s a good change that AKIDA is openly recommended and recognised
 
  • Like
  • Love
Reactions: 9 users

Tothemoon24

Top 20
IMG_7549.jpeg



𝐔𝐧𝐥𝐨𝐜𝐤𝐢𝐧𝐠 𝐭𝐡𝐞 𝐅𝐮𝐭𝐮𝐫𝐞 𝐨𝐟 𝐇𝐞𝐚𝐥𝐭𝐡𝐜𝐚𝐫𝐞 𝐰𝐢𝐭𝐡 𝐍𝐞𝐮𝐫𝐨𝐦𝐨𝐫𝐩𝐡𝐢𝐜 𝐂𝐨𝐦𝐩𝐮𝐭𝐢𝐧𝐠

Neuromorphic computing is an exciting field aiming to replicate the intricate workings of the human brain in electronic devices. Unlike traditional digital chips, neuromorphic chips use artificial neurons and synapses to process information with natural analog signals. These chips have the potential to transform healthcare, especially in creating point-of-care devices for disease detection.

However, a significant challenge in neuromorphic computing is training the neural network on the chip, which can be time-consuming and energy-inefficient. Researchers from Eindhoven University of Technology and Northwestern University have introduced a groundbreaking neuromorphic biosensor capable of on-chip learning by processing real-time patient data. This innovation eliminates the need for external training, making the biosensor more adaptable and responsive.

To showcase its effectiveness, the researchers used the biosensor to diagnose cystic fibrosis, an inherited condition affecting the lungs and digestive system. The biosensor detects cystic fibrosis by measuring chloride anion levels in sweat samples. It uses an organic electrochemical transistor to sense these anions and a neuromorphic chip to process the signal and classify it as normal or abnormal, adjusting its decision threshold with each sample.

The results show that this biosensor achieves high accuracy and sensitivity in cystic fibrosis detection while consuming minimal power and providing rapid response times. Its flexibility and biocompatibility also make it suitable for wearable applications. This promising study paves the way for intelligent biosensors and personalized healthcare devices grounded in neuromorphic computing.

For more details, you can access the published study in Nature Electronics through this link
 
  • Like
  • Love
  • Fire
Reactions: 26 users
Top Bottom