BRN Discussion Ongoing

rgupta

Regular
They get shares because we don’t have money to pay their high salaries.
That is no true. They designed the things and we investors are following. They know if they plan bigger salaries then holders will ask for more transparency. This way they can hide why they have to pay that much. In reality it is better for holders to pay salary because that will help us tax offset. e.g Sean get half a million worth of shares he pay 200,000 worth of tax by selling his shares. If we pay him 500,000 worth of money he will pay the same tax but we can account that in expenses and claim the same from our profits in future ( if there will be)
But the plans were made by management to suit them.
Dyor
 

yogi

Regular
  • Like
Reactions: 1 users

manny100

Regular
You cannot compare salaries of MDs and CEOs of US based companies to those of Aus based companies.
Sean as part of his package gets circa $US500K cash and the rest as shares and options (81%).
$US500K cash for a CEO is peanuts in the US.
If BRN goes belly up and his shares worthless he has worked for Jack Shiete.
His package is effectively designed to reward performance.
If his 5 year plan plays out he will be very, very wealthy via his package with 81% equities.
If you look at the recent Ann concerning Tony Viana share sales you will see 250k of restricted shares have vested and therefore attract income tax..
Note restricted shares for Tony after the transaction has reduced by the 250k vested shares.
Note that he sold 85k of those 250k shares at 23.5 cents to pay income tax on those 250k vested/he now owns shares.
The net result is that he owns out right an extra 165k shares (250k less 85k sold). The increase in his holding is reflected in the ann.
So in effect Tony has via his salary package purchased an extra 165k shares.
Contrary to what the downrampers on the crapper say these shares are not free as they are paid for via pay package.
Shares as part of a salary package are a great incentive strategy.
The irony is if you believe the company is a dud then senior staff are actually being paid dud money.
If you believe the company will eventually succeed then senior staff will be very, very well off financially via their holdings.
The bottom line is that the BOD is accumulating.
 
  • Like
  • Fire
  • Love
Reactions: 52 users

rgupta

Regular
You cannot compare salaries of MDs and CEOs of US based companies to those of Aus based companies.
Sean as part of his package gets circa $US500K cash and the rest as shares and options (81%).
$US500K cash for a CEO is peanuts in the US.
If BRN goes belly up and his shares worthless he has worked for Jack Shiete.
His package is effectively designed to reward performance.
If his 5 year plan plays out he will be very, very wealthy via his package with 81% equities.
If you look at the recent Ann concerning Tony Viana share sales you will see 250k of restricted shares have vested and therefore attract income tax..
Note restricted shares for Tony after the transaction has reduced by the 250k vested shares.
Note that he sold 85k of those 250k shares at 23.5 cents to pay income tax on those 250k vested/he now owns shares.
The net result is that he owns out right an extra 165k shares (250k less 85k sold). The increase in his holding is reflected in the ann.
So in effect Tony has via his salary package purchased an extra 165k shares.
Contrary to what the downrampers on the crapper say these shares are not free as they are paid for via pay package.
Shares as part of a salary package are a great incentive strategy.
The irony is if you believe the company is a dud then senior staff are actually being paid dud money.
If you believe the company will eventually succeed then senior staff will be very, very well off financially via their holdings.
The bottom line is that the BOD is accumulating.
You are right BOD are accumulating but for the market they are selling on market. If company pay them money in leu of their shares then they may be net buyers. So they sell shares on behalf of company for their tax purposes, if company sell those shares and pay them money, that will mean BOD may but on market and that will be good outcome for share holders.
But the only draw back will be the losses which are projected @approx 20 million a year may rise by another 10-15 million a year. But for accounting purpose and better clarity that would be in interest of company and holders.
Dyor
 
  • Like
Reactions: 1 users
Anyone know any jobs going needing DoD clearance....anyone?....would be nice :)



ITL ventures into neuromorphic computing​

By Megan Saxton
U.S. Army Engineer Research and Development Center
Published Sept. 23, 2024
Updated: Sept. 23, 2024


Neuromorphic Computing

PHOTO DETAILS / DOWNLOAD HI-RES 1 of 1

The U.S. Army Engineer Research and Development Center (ERDC) Information Technology Laboratory (ITL) Edge Computing Lab has long been on the cutting-edge of this field and is now exploring something new: neuromorphic computing.

In recent years, edge computing has revolutionized the technology landscape for users situated in remote areas or away from primary devices. By bringing computation and data storage closer to the location where it is needed, response times, reliability and performance are greatly improved, latency and bandwidth costs are reduced and privacy and security are enhanced. The U.S. Army Engineer Research and Development Center (ERDC) Information Technology Laboratory (ITL) Edge Computing Lab has long been on the cutting-edge of this field and is now exploring something new: neuromorphic computing.

“Neuromorphic computing is a process in which computers are designed and engineered to mirror the structure and function of the human brain,” said Dr. Raju Namburu, ITL chief technology officer and a senior scientific technical manager. “Using artificial neurons and synapses, neuromorphic computers simulate the way our brains process information, allowing them to solve problems, recognize patterns and make decisions more quickly and efficiently than the traditional high-performance computing systems we use today.”

The driving force behind ITL’s research into this emerging technology is the U.S. military’s need to know more, sooner, to allow rapid, decisive action on the multi-domain battlefield. The battlespace has become characterized by highly distributed processing, heterogeneous and mobile assets with limited battery life, communications- dominated but restricted network capacity and operating with time-critical needs in a rapidly changing hostile environment. Distributed and low power edge processing is one of the essential technologies for maintaining overmatch in various emerging operational and contested environments, as is the need to take advantage of machine learning (ML) and generative artificial intelligence (AI).

“Overall, neuromorphic chips offer the DoD community a number of potential benefits including improved performance, resilience, cost-efficiency, security, privacy, power-efficiency, signal processing, ML capabilities and more,” said Dr. Ruth Cheng, a computer scientist in ITL’s Supercomputing Research Center. “By keeping an eye on developments in this technology, the DoD community can ensure it remains at the forefront of military and defense innovation.”

“Computations performed at the molecular, atomic, and neuro scales mimicking the human brain are showing tremendous viability,” added Namburu. “We just started this work on next generation advanced computing, which is significantly different from traditional computing systems historically used at ERDC. Neuromorphic computing represents a paradigm shift in computing, promising significant advancements in ML, generative AI, scientific applications and sensor processing compared to traditional computing. Moreover, neuromorphic chips emulate the brain's plasticity, enabling learning and adaptation over time, unlike traditional systems.”

Ongoing efforts edge computing efforts include agnostic graphics processing unit (GPU) ray tracing development, benchmarking deep neural networks, sensor-data management, ML for underwater invasive plants, railcar inspection, photogrammetry, reservoir frameworks, decentralized edge computing, bi-directional digital twins and algorithms for anomaly detection. ITL is also exploring emerging AI chips for edge computing including novel algorithms and sustainable software.

“Overall, edge computing is helping to enable new use cases and provide better experiences to the users by making applications faster, more reliable and more secure,” said Cheng. “Neuromorphic chips are well-suited for edge computing, which is becoming increasingly important in military and defense applications, and ITL is already aiding in this process that will touch everything from lowering the cost of deployments by eliminating the need for expensive, high-powered servers and data centers to support of mobile and autonomous systems. This is the future.”
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 34 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
Altmann is not the messiah ...

When he bought the $51M worth of chips off the plan, Rain were touting pie-in-the-sky technology - a spaghetti bowl of self-organizing nanowires which miraculously formed an analog neural network, substituting chaos for complexity.

Rain have since completely changed their design and now offer a hybrid (Frankenstein) digital/analog MAC NN. Analog is very efficient in performing 1-bit multiplications, but, due to manufacturing variability, the output amplitude varies between different neuron multiplier.

To overcome this, Rain interposes ADC and DAC circuits within or between the NN layers.

US2024249190A1 GRADIENT COMPUTATION IN HYBRID DIGITALLY TIED ANALOG BLOCKS WITH ARBITRARY CONNECTIVITY BY EQUILIBRIUM PROPAGATION 20230119

View attachment 73528

Having started as a pure analog comapny, Rain are late to the hybrid digital/analog party.

In Rain they use MACs.

US2024143541A1 COMPUTE IN-MEMORY ARCHITECTURE FOR CONTINUOUS ON-CHIP LEARNING 20221028

View attachment 73547


A system capable of providing on-chip learning comprising a processor and a plurality of compute engines coupled with the processor. Each of the compute engines including a compute-in-memory (CIM) hardware module and a local update module. The CIM hardware module stores a plurality of weights corresponding to a matrix and is configured to perform a vector-matrix multiplication for the matrix. The local update module is coupled with the CIM hardware module and configured to update at least a portion of the weights.


Akida and Rain are competitors with incompatible technologies.

Thanks Diogenese.

RainAI has attracted some top quality people such as Jean-Didier Allegrucci, who worked on Apple's s SoCs for over 17 years. And also former Meta architecture leader Amin Firoozshahian. And both of these appointments have been within the last 12 months.

I'll admit, I find it a bit of a pity that our technologies seem to be incompatible as they seem like they're shaping up to be a formidable competitor in our midst if I'm not mistaken.

Or maybe not, but the first thing that struck me quite a while ago was that Altman went with Rain AI when they were spruiking "NEUROMORPHIC" capabilities. And now they aren't or can't spruik it any longer because they didn't find the right path there.

The question is, Altman (OpenAI) and Masayoshi Son (SoftBank) need to find the company with the best technology to make their goals come to fruition but is that company going to be Rain AI?

I mean, it's probably silly to ask that I suppose. But we are still in the "neuromorphic" business and Rain AI aren't, so that must be worth something...

Random thought bubbles...
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 15 users

DK6161

Regular
Insiders still hold circa 15.5% to 16.5% of SOI - that shows confidence.
LDN (Sean's predecessor) is still in the Top 50 holders - that shows confidence.
Sean takes 81% of his salary in shares and options - that shows confidence.
All the BOD take levels of shares and options as part of their pay- that shows confidence.
The above gives me confidence.
Thanks for this. Every now and then you just need posts like this to keep the hope alive
 
  • Like
Reactions: 2 users

Diogenese

Top 20
Thanks Diogenese.

RainAI has attracted some top quality people such as Jean-Didier Allegrucci, who worked on Apple's s SoCs for over 17 years. And also former Meta architecture leader Amin Firoozshahian. And both of these appointments have been within the last 12 months.

I'll admit, I find it a bit of a pity that our technologies seem to be incompatible as they seem like they're shaping up to be a formidable competitor in our midst if I'm not mistaken.
Yes. Those two would understand the new Rain tech. I doubt that either would have joined Rain if they were pushing the original nanowire NN which Altmann signed up for - think Theranos.

That's a very astute observation about neuromorphic. The RISC-V collaboration is not neuromorphic.

At one stage, rain were touting themselves as digital AI using RISC-V ( https://rain.ai/approach ), . Their recent patents are for hybrid analog/digital so they may have a couple of strings to their bow. In any event, they have been spending a lot on early stage R&D, ie, more on the R than the D.

Their new design will probably work. I haven't seen any performance figures.

There are several companies pushing hybrid analog/digital including Qualcomm.
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 15 users
Always positive my friend… if you’re negativ on something, stop it, change it, move on, or let it be.


IP announcement, next week for sure folks!

I don't use "bold" lightly.
 
  • Haha
  • Like
  • Thinking
Reactions: 14 users
Verrückte Deutsche.
 
You cannot compare salaries of MDs and CEOs of US based companies to those of Aus based companies.
Sean as part of his package gets circa $US500K cash and the rest as shares and options (81%).
$US500K cash for a CEO is peanuts in the US.
If BRN goes belly up and his shares worthless he has worked for Jack Shiete.
His package is effectively designed to reward performance.
If his 5 year plan plays out he will be very, very wealthy via his package with 81% equities.
If you look at the recent Ann concerning Tony Viana share sales you will see 250k of restricted shares have vested and therefore attract income tax..
Note restricted shares for Tony after the transaction has reduced by the 250k vested shares.
Note that he sold 85k of those 250k shares at 23.5 cents to pay income tax on those 250k vested/he now owns shares.
The net result is that he owns out right an extra 165k shares (250k less 85k sold). The increase in his holding is reflected in the ann.
So in effect Tony has via his salary package purchased an extra 165k shares.
Contrary to what the downrampers on the crapper say these shares are not free as they are paid for via pay package.
Shares as part of a salary package are a great incentive strategy.
The irony is if you believe the company is a dud then senior staff are actually being paid dud money.
If you believe the company will eventually succeed then senior staff will be very, very well off financially via their holdings.
The bottom line is that the BOD is accumulating.
Well said. Also opportunities for them to buy shares is limited especially if they are in possession of knowledge that would be seen as insider knowledge. I'm hoping they are in possession of a shit load of that knowledge. 😀

SC
 
  • Like
  • Fire
  • Haha
Reactions: 12 users

manny100

Regular
Well said. Also opportunities for them to buy shares is limited especially if they are in possession of knowledge that would be seen as insider knowledge. I'm hoping they are in possession of a shit load of that knowledge. 😀

SC
Sean is all in up to his neck financially with BRN.
When Sean talks BRN up he is backing it up with his hard earned.
You are right, Sean taking 81% of his salary as equity avoids any 'inside trading' allegations/problems.
This applies to the BOD who are net accumulators of equity.
 
  • Like
  • Love
Reactions: 20 users

Frangipani

Top 20
Intriguing interview with Lavinia Andreea Danielescu, Director of the Future Technologies R&D group at Accenture Labs and one of the co-inventors of the Accenture patents “Neuromorphic smooth control of robotic arms” (filed on 24 May 2022, granted on 27 August 2024), “Self-learning neuromorphic acoustic model for speech recognition” (filed on 16 September 2022, granted on 12 November 2024) and “Self-learning neuromorphic gesture recognition models” (filed on 22 November 2022, application first published on 23 May 2024), all three of which mention both Akida and Loihi as examples of neuromorphic processors:




The Innovator
Latest articles

Interview Of The Week: Andreea Danielescu, Future Technologies Expert​

9 hours ago
by Jennifer L. Schenker
7 min read
Andreea.png

12dQrIJourU0CgiToONiVcQ-1-150x150.png

Written by Jennifer L. Schenker

Andreea Danielescu is the Director of the Future Technologies R&D group at Accenture Labs. Her group focuses on new emerging technologies that blend the physical and digital, including biotechnology, smart materials, energy harvesting and storage and neuromorphic computing. Her specific areas of expertise also include conversational and gestural interfaces, wearable technologies, and AI and tech ethics. Prior to Accenture she worked as an engineer and researcher on conversational interfaces at both Facebook and Intel. Danielescu received her bachelor’s in computer science and mathematics from University of Arizona and her PhD from Arizona State University in computer science with an arts, media, and engineering concentration. She holds over 10 patents, has over 50 peer reviewed publications and is a Senior Member of the ACM, a digital library that serves as a research, discovery and networking platform for computing educators, researchers, and professionals. Danielescu, a speaker at the XPANSE conference in Abu Dhabi Nov. 20-22, recently spoke to The Innovator about emerging technology trends and how to prepare for the future.

Q: Tell us about Accenture’s Future Technologies R&D group.

AD: The group is currently focusing on neuromorphic computing [an approach to computing that uses physical artificial neurons to do computations, mimicking the human brain], smart materials and biomaterials. We work on Edge-based solutions that are low power, energy harnessing solutions and smart materials. It is part of a larger focus on sustainability. A lot of our clients are looking for alternatives to plastics and new packaging materials. We think about new designs, such as finding an alternative to glucose monitors or diabetic pumps that need to be returned for disassembly. We don’t have our own materials lab. We work a lot with university partners and run about 2 dozen projects at any given time.

Q: Can you give some examples of your projects?


AD: One example is dissolvable, degradable electronics. Advances in manufacturing methods for electronics most often aim to produce highly integrated and reliable devices for long-term use. While these features have brought benefits to customers, they also have many side effects. One of these is e-waste, which is the fastest-growing waste stream in the world. Significant effort has been put toward e-waste recycling, but the composition of electronic devices makes the recycling process far more challenging than that of other materials like metals and cardboard. This is exacerbated by the increasing rate at which we produce smart devices, leading to much more e-waste than today’s recycling methods can handle.

In our work, we investigated how materials, fabrication tools and methods, and different types of destruction (melting, dissolving, etc.) can be combined to make devices with sustainability, transiency, and interactivity at the core. We worked on three practical approaches for building such devices including: laser-transferring edible gold foil onto 3D printed chocolates, inkjet printing conductive traces on water-soluble PVA sheets and fabricating electronic devices using natural beeswax.

An example I discussed during a panel at EXPANSE, is a seed carrier that can be dropped from drones and drill seeds into the soil to help with reforestation in areas where it is difficult or dangerous for humans to reach.

Another project involved leveraging the inherent material properties of natural materials, such as paper, leaf skeletons, and chitosan, along with silver nanowires, to create a new decomposable portable heater capable of being electrically controlled. This leaf powered heater can reach temperatures of >70°C and is flexible, lightweight, low-cost, and reusable. Use cases include heating snacks or beverages on the go or simple heating of wax strips for hair removal.

We are currently working on ways to create sensors out of textiles. Everything from sensing and actuation to the power and communications can be textile-based. The next step will be textile-based intelligence. Applications include biometrics for health and wellness embedded in your clothing. The more you know about your body the better off you are, MXenes, which were discovered in 2011 by Yury Gogotsi at Drexel University are finally approaching enough maturity to provide practical solutions to e-textiles that weren’t possible with only conductive copper or silver yarns. By creating fully textile-based systems you also eliminate the filaure points of hard (traditional electronics) to soft (textile) connections that have made many e-textiles impractical in the past.

Q: What changes do you see coming?

AD: In our conversations with clients, academia and the industry about what is top of mind we start to see trends and understand what people are worried about now. How will this change the market? Supply chain management and product authenticity for high-end goods and how we think about counterfeits need to be rethought.

This is not just a technology question. Ensuring ethical practices throughout the supply chain is dependent on humans. Unless multiple parties that don’t know each other have a vested interest in a particular outcome you will never have a neutral outcome. Tracking supply chains relies on a reporting structure that is reliant on people with vested interests, so it is a social problem just as much as a technological one.

More and more consumers are demanding ethically and sustainably sourced goods. How do you provide that information to your clients? How do you develop interactive elements into a product that can give detailed information? This all goes hand-in-hand with tracking things on the supply chain. We will see more and more companies adding interactive elements into products. An example of this would be a smart label on a wine bottle that uses near-field communication (NFC) on mobile phones to access a website about that wine, giving consumers much more information without adding package. An edible RFID sensor can be added to chocolate that allows consumers to scan it with their phone to enter a sweepstakes and win a ticket to an event.

NFC is built into everything so we can do this with what we have out there already. Printing gold leaf onto a chocolate bar can be done using very cheap materials that are readily available. Ubiquitous computing is already here. We just need to leverage what is already out there and add a little bit more.

Privacy and personalization will also become more important. Neuomorphic computing uses low latency and allows privacy preserving computing at the edge. If you don’t need to send information to the Cloud you can safeguard privacy and this will increase opportunities for personalized services. The ability to create a digital signature on textiles by varying pressure and tempo can provide solutions that are more robust to counterfeiting by providing more biomarkers for each person’s signature than is available today through touch screens.

As we reduce power usage requirements we can move more computational systems into the environment, improving efficiency and convenience while also protecting privacy. For example, low-power sensors can be used to structural health monitoring, providing early warnings of bridge failures. Low cost, degradable sensors can be used to help with crop management by providing real-time updates on soil and plant health and disease, allowing farmers to respond more quickly to changing conditions and increase the likelihood of good crop yields.

Personalization can also be applied to communication. The voices of digital assistants or on text-to-speech devices will be customized so that people – whether they are male or female or non-binary -can find voices that sound like them.

Q: Based on your experience, what is the best way for corporates to anticipate and prepare for the future?

AD: To prepare for the future you need dedicated resources. It can be challenging to justify but you need to properly resource a future-facing division and hire the right people. An interdisciplinary team is critical to this. R&D requires much longer cycles, on the order of three-to-five years minimum to explore a new area and start to develop applications and methods to scale the technology. Technology forecasting services can be helpful. Conversations with a wide range of researchers, clients, and subject matter experts in the industries of interest are also critical to identifying potential problems that could benefit from a novel solution and emerging technologies that are worth exploring. . R &D requires much longer cycles.




About the author​

12dQrIJourU0CgiToONiVcQ-1-150x150.png

Jennifer L. Schenker​

Jennifer L. Schenker, an award-winning journalist, has been covering the global tech industry from Europe since 1985, working full-time, at various points in her career for the Wall Street Journal Europe, Time Magazine, International Herald Tribune, Red Herring and BusinessWeek. She is currently the editor-in-chief of The Innovator, an English-language global publication about the digital transformation of business. Jennifer was voted one of the 50 most inspiring women in technology in Europe in 2015 and 2016 and was named by Forbes Magazine in 2018 as one of the 30 women leaders disrupting tech in France. She has been a World Economic Forum Tech Pioneers judge for 20 years. She lives in Paris and has dual U.S. and French citizenship.
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 20 users

Bravo

If ARM was an arm, BRN would be its biceps💪!

SNAPSHOT​

A team of NASA personnel and contractors has prototyped a new set of algorithms that will enable instruments in space to process data more efficiently. Using these algorithms, space-based remote sensors will be able to provide the most important data to scientists on the ground more quickly and may also be able to autonomously determine which Earth phenomena are the most important to observe.

Photo of the International Space Station in orbit with the Earth in the background

The International Space Station, where Steve Chien and his team prototyped a new set of AI algorithms that will reduce data latency and improve dynamic targeting capabilities for satellites. (Credit: NASA/ISS)
Earth-observing instruments can gather a world’s worth of information each day. But transforming that raw data into actionable knowledge is a challenging task, especially when instruments have to decide for themselves which data points are most important.

“There are volcanic eruptions, wildfires, flooding, harmful algal blooms, dramatic snowfalls, and if we could automatically react to them, we could observe them better and help make the world safer for humans,” said Steve Chien, a JPL Fellow and Head of Artificial Intelligence at NASA’s Jet Propulsion Laboratory.

Engineers and researchers from JPL and the companies Qualcomm and Ubotica are developing a set of AI algorithms that could help future space missions process raw data more efficiently. These AI algorithms allow instruments to identify, process, and downlink prioritized information automatically, reducing the amount of time it would take to get information about events like a volcanic eruption from space-based instruments to scientists on the ground.

These AI algorithms could help space-based remote sensors make independent decisions about which Earth phenomena are most important to observe, such as wildfires.

“It’s very difficult to direct a spacecraft when we’re not in contact with it, which is the vast majority of the time. We want these instruments to respond to interesting features automatically,” said Chien

Chien prototyped the algorithms using commercially available advanced computers onboard the International Space Station (ISS). During several different experiments, Chien and his team investigated how well the algorithms ran on Hewlett Packard Enterprise’s Spaceborne Computer-2 (SBC-2), a traditional rack server computer, as well as on embedded computers.

These embedded computers include the Snapdragon 855 processor, previously used in cell phones and cars, and the Myriad X processor, which has been used in terrestrial drones and low Earth orbit satellites.

Including ground tests using PPC-750 and Sabertooth processors – which are traditional spacecraft processors – these experiments validated more than 50 image processing, image analysis, and response scheduling AI software modules.

The experiments showed these embedded commercial processors are very suitable for space-based remote sensing, which will make it much easier for other scientists and engineers to integrate the processors and AI algorithms into new missions.

The full results of these experiments were published in a series of three papers at the 2022 IEEE Geoscience and Remote Sensing Symposium, which can be accessed through the links below.

Chien explains that while it is easiest to deploy AI algorithms from ground computers to larger, rack-mounted servers like the SBC-2, satellites and rovers have less space and power, which means they would need to use smaller, low-power, embedded processors similar to the Snapdragon or Myriad units.

By processing the data onboard, these AI algorithms prevent important or urgent information from being buried within larger data transmissions. A researcher wouldn’t have to downlink and process an entire transmission to see that a hurricane is intensifying or a harmful algal bloom has formed.

“A large image could have gigabytes of data, so it might take a day to get it to the ground and process it. But you don’t need to process all that data to identify a wildfire. These algorithms pre-process data onboard so that researchers get the most important information first,” said Chien.

These algorithms could be useful not only for Earth-observing instruments, but also for instruments observing other planets as well. The proposed Europa Lander mission, for example, could use Chien’s algorithms to help search for life on the Jovian moon.

“There are several missions that are in concept development right now that could use this technology. They’re still in the early phases of development, but these are missions that need the kind of onboard analysis, understanding, and response these algorithms enable,” said Chien.

The team is also testing neural network models to interpret Mars satellite imagery. “Someday such a neural net could enable a satellite to detect a new impact ejecta, evidence of a meteorite impact, and alert other spacecraft or take follow-up images,” said JPL Data Scientist Emily Dunkel. “Rovers could also use these processors with neural networks to determine where it is safe for the rover to drive,” Dr. Dunkel added.

“We used the CogniSat framework to deploy models to the Myriad X, reducing the effort to develop deep learning models for onboard use. This experience helps prove that this advanced hardware and software system is ready now for space missions,” said Léonie Buckley, Senior Engineer at Ubotica.

As climate change continues to alter the world we live in, information systems like Chien’s allow scientific instruments to be as dynamic as the Earth systems they observe.

“We don’t often think about the fact that we’re walking around with more computing power in our cell phones than supercomputers had forty years ago. It’s an amazing world we live in, and we’re trying to incorporate those advancements into NASA missions,” said Chien. View attachment 35765 View attachment 35765

Following on from @Tothemoon24 's post, which described how NASA’s JPL (Jet Propulsion Laboratory) was working with Qualcomm and Ubotica on developing a set of AI algorithms that could help future space missions process raw data more efficiently to detect events like a volcanic eruptions, wildfires, flooding, harmful algal blooms, dramatic snowfalls, etc.

Well, I was just reading this article below dated 14 June 2024, and it talks about Ubotica having been awarded Horizon Europe funding through the "Meseo project", which aims to revolutionise Earth observation systems. The project is a collaboration between various companies including GMV and Airbus!

In trying to get to the bottom of whether we are linked to the work with NASA JPL, Qualcom and Ubotica, I came upon this research paper dated October 2023, which has probably been posted in TSEx previously. The paper titled "NimbleAI: 3D-Integrated Neuromorphic Vision Sensor-Processor" was written by Xabier Iturbe , Gianluca Furano (ESA) and Didier Keymeulen (NASA JPL), and it talks about using SNN's to help monitor the earth's surface for phenomena such as explosive eruptions (think volcanoes) and the like.

It's interesting because back in March 2024, one of the authors of this paper, Xabier Iturbe, liked a post about BrainChip's Edge Box on LinkedIn (see below).

Anyway, if you take a look at the MESEO Projects website, you can see Airbus is listed as a partner! And it makes me wonder whether the agreement we signed with Airbus could also have something to do with the MESEO project.

It's a bit unfortunate that I only just discovered this website today because it says they held a Webinar on the 28th of November, which would have been great to tune into to check for any other potential signs of our involvement.


Screenshot 2024-11-30 at 10.15.05 am.png





Screenshot 2024-11-30 at 10.29.15 am.png







Screenshot 2024-11-30 at 10.35.03 am.png




Screenshot 2024-11-30 at 10.41.36 am.png





Screenshot 2024-11-30 at 12.13.59 pm.png




Dublin’s Ubotica joins EU project to boost Earth observation​

by Leigh Mc Gowran
14 Jun 2024
Save article
A satellite in space above the Earth.

Image: © Dabarti/Stock.adobe.com
The Meseo project aims to create a system that can support the vast amount of data that satellites transfer, to improve various Earth observation services.
Irish space-tech company Ubotica has been awarded Horizon Europe funding through the Meseo project, which aims to revolutionise Earth observation systems.
Meseo is a collaborative research project that aims to make Europe’s space sector more competitive. The Horizon Europe-backed project aims to develop a scalable multi-mission Earth observation system that can support large-scale data processing.
This system will be designed to manage and process the huge amount of data that numerous satellites must transfer. The aims to reduce communication bandwidth requirements and utilise advanced technologies to optimise power consumption, processing capabilities and system usability.
The heart of this project is an Earth observation coordination centre, where specific software called processing functions will manage satellite data and products, to guarantee their ownership and quality. The goal is to create an accessible ecosystem that is open to any Earth observation product and service.
The project is a collaboration between various companies including GMV and Airbus. Ubotica is responsible for the development and deployment of AI-driven image triaging and in-line processing.
“We are excited to contribute to Meseo’s ambitious objectives and we look forward to the innovative advancements this partnership will bring to the space industry,” said Dr Aubrey Dunne, Ubotica co-founder and CTO. “Our AI-driven solutions are poised to revolutionise how data is processed on board, enhancing overall efficiency and effectiveness, and the Meseo system will significantly benefit from the integration of these solutions.”
Founded in 2017 and based at Dublin City University, Ubotica has had exciting developments in recent years, including striking a partnership with IBM and forming a corporate entity in the US to expand its presence in the country.

Earlier this year, Ubotica successfully launched its CogniSAT-6 satellite, as part of its mission to improve Earth observation services. This satellite is designed to provide real-time data to support various activities, such as monitoring crop health or tracking illegal fishing.




 

Attachments

  • Screenshot 2024-11-30 at 11.11.19 am.png
    Screenshot 2024-11-30 at 11.11.19 am.png
    92.6 KB · Views: 30
Last edited:
  • Like
  • Love
  • Fire
Reactions: 37 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
And do you know what else is pretty cool?!!!

It's that Xabier Iturbe (Nimble AI's project co-ordinator), the same guy that liked the Linkedin post about BrainChip's Edge Box in March 2024 (see above), just posted this on LinkedIn a mere 9 hours ago.

I love some of the quotes included here!

"We expect up to 57% penetration of neuromorphic chips in most major applications by 2034" - YOLE Intelligence

"Neuromorphic computing will have a substantial impact on existing products and markets, taking three to six years to cross over from early-adopter status to early majority adoption" - Gartner.


I checked where this Gartner quote first came from and it came from an article dated 19 January 2023, so it will now be roughly "two to five years from early adoption to majority adoption".




Screenshot 2024-11-30 at 12.43.59 pm.png



January 19, 2023

Contributor: Tuong Nguyen

The 2023 Gartner Emerging Technologies and Trends Impact Radar shows product leaders where to capitalize on market opportunities.

Gartner research reveals four emerging technologies and trends to which tech vendors and product leaders will need to respond, calibrating their tech strategies, investments and tools to stay ahead:

  1. Smart world expands with increased fusion of physical-digital experiences.
  2. Productivity revolution accelerates with advances in artificial intelligence (AI) tools and tech.
  3. Transparency and privacy get more scrutiny amid exponential growth in corporate and personal data collection.
  4. New critical technology enablers create new business and monetization opportunities.

What’s on the 2023 Gartner Emerging Technologies and Trends Impact Radar?​

These trends surfaced in our 2023 Gartner Emerging Technologies and Trends Impact Radar, which highlights 26 emerging trends and technologies to which vendors must respond, whether they are a new or established player in that space.

The Impact Radar portrays the maturity, market momentum and influence of technologies, making it a handy tool for product leaders to identify and track the technologies and trends that will help them improve and differentiate their products, remain competitive and capitalize on market opportunities.

Download now: Your Detailed Guide to Gartner Emerging Tech Impact Radar 2023

2023 Gartner Emerging Technologies and Trends Impact Radar

Four Emerging Technologies Disrupting the Next Three to Eight Years​

Most of this year’s emerging technologies and trends are three to eight years away from reaching widespread adoption but represent significant innovation in the years ahead.

Let’s look at four we think will prove especially interesting.

No. 1: Neuromorphic computing​

  • A critical enabler, neuromorphic computing provides a mechanism to more accurately model the operation of a biological brain using digital or analog processing techniques.
  • It will take three to six years to cross over from early-adopter status to early majority adoption.
  • Neuromorphic computing will have a substantial impact on existing products and markets.
Neuromorphic computing systems simplify product development, enabling product leaders to develop AI systems that can better respond to the unpredictability of the real world. Their autonomous capabilities quickly react to real-time events and information, and will form the basis of a wide range of future AI-based products. Early use cases include event detection, pattern recognition and small dataset training.

We expect breakthrough neuromorphic devices by the end of 2023, but it will likely take five years for these devices to reach early majority adoption.

The impact is likely to be significant, though, as neuromorphic computing is expected to disrupt many of the current AI technology developments, delivering power savings and performance benefits not achievable with current generations of AI chips.

No. 2: Self-supervised learning​

  • Self-supervised learning accelerates productivity by using an automated approach to annotating and labeling data.
  • It will take six to eight years to cross over from early-adopter status to early majority adoption.
  • Self-supervised learning will have a significant impact on existing products and markets.
Self-supervised models learn how information relates to other information; for example, which situations typically precede or follow another, and which words often go together.

Self-supervised learning has only recently emerged from academia and is currently practiced by a limited number of AI companies. A few companies focused on computer vision and NLP products have recently added self-supervised learning to their product roadmaps, however.

The potential impact and benefits of self-supervised learning are extensive, as it will extend the applicability of machine learning to organizations with limited access to large datasets. Its relevance is most prominent in AI applications that typically rely on labeled data, primarily computer vision and NLP.

No. 3: Metaverse​

  • The metaverse fuels the smart world by providing an immersive digital environment.
  • It will take eight-plus years to cross over from early-adopter status to early majority adoption.
  • The metaverse will have a very substantial impact on existing products and markets.
The metaverse enables persistent, decentralized, collaborative, interoperable digital content that intersects with the physical world’s real-time, spatially organized and indexed content.

It is an example of a combinatorial trend in which a number of individually important, discrete and independently evolving trends and technologies interact with one another to give rise to another trend. The emerging, supporting technologies and trends include (but are not limited to) spatial computing and the spatial web; digital persistence; multientity environments; decentralization tech; high-speed, low-latency networking; sensing technologies; and AI applications.

The features and functionality these ETT bring to the metaverse will need to reach an early majority in order for the metaverse to cross the chasm. We consider all current examples to be precursors or premetaverse offerings because they are potentially capable and compatible but do not yet meet the definition of the metaverse.

While the benefits and opportunities from the metaverse are not immediately viable, emerging metaverse solutions give an indicator of potential use cases. We expect the transition toward the metaverse to be as significant as the one from analog to digital.

No. 4: Human-centered AI​

  • Human-centered AI (HCAI) is a common AI design principle calling for AI to benefit people and society, which could improve transparency and privacy.
  • It will take three to six years to reach early majority adoption.
  • HCAI will have a substantial impact on existing products and markets.
HCAI assumes a partnership model of people and AI working together to enhance cognitive performance, including learning, decision making and new experiences. HCAI is sometimes referred to as “augmented intelligence,” “centaur intelligence” or “human in the loop,” but in a wider sense, even a fully automated system must have human benefits as a goal.

HCAI enables vendors to manage AI risks, and to be ethical, responsible and more efficient with automation, while complementing AI with a human touch and with common sense. Many AI vendors have already shifted their positions to the more impactful and responsible HCAI approach. The technology-centric approach of developing AI products has led to numerous negative impacts, urging vendors to rethink their AI product strategies.

The potential impact of HCAI is high because it leverages human abilities to make humans more productive and remove avoidable limitations, biases and blind spots.

In short:

  • The Gartner Emerging Tech Impact Radar highlights the technologies and trends that have the most potential to disrupt a broad cross section of markets.
  • The trends are organized around four key themes, which are critical for product leaders to evaluate as part of their competitive strategy.
  • Product leaders must explore these technologies now to capitalize on market opportunities.
Tuong Nguyen is a Director Analyst within the Emerging Technologies and Trends team in Gartner Research. He undertakes analysis on immersive technologies, metaverse, computer vision, SLAM and human-machine interfaces. He advises tech provider product leaders how to factor emerging tech and trends into creating and evolving highly successful product offerings.

Gartner Tech Growth & Innovation Conference​

Join the world’s leading IT and business leaders to get an update on accelerating tech growth in a new era of transformation and technology trends.


 
  • Like
  • Fire
  • Love
Reactions: 39 users
IP announcement, next week for sure folks!
Come on spill the beans and let us know what asx company I should be putting some money on Monday?
 
  • Haha
  • Like
  • Thinking
Reactions: 8 users

Diogenese

Top 20
delayed reaction:

A coupe of years ago there was talk of potential patent litigation between iniVation and Prophesee.

Does the merger of iniVation and SynSense preclude further cooperation between Prophesee and SynSense?

https://www.synsense.ai/synsense-an...orm-leading-neuromorphic-technology-provider/

SynSense and iniVation join forces to form leading neuromorphic technology provider​

2024-02-01

Who would Prophesee turn to for NN if this is the case?

I'd thought that Prophesee had deveoped its low-fi 320*320 pixel array so as not to overstretch SynSense capabilities.

https://brainchip.com/brainchip-introduces-second-generation-akida-platform/

BrainChip Introduces Second-Generation Akida Platform

Introduces Vision Transformers and Spatial-Temporal Convolution for radically fast, hyper-efficient and secure Edge AIoT products, untethered from the cloud
Laguna Hills, Calif. – March 6, 2023
...
“At Prophesee, we are driven by the pursuit of groundbreaking innovation addressing event-based vision solutions. Combining our highly efficient neuromorphic-enabled Metavision sensing approach with Brainchip’s Akida neuromorphic processor holds great potential for developers of high-performance, low-power Edge AI applications. We value our partnership with BrainChip and look forward to getting started with their 2nd generation Akida platform, supporting vision transformers and TENNs,” said Luca Verre, Co-Founder and CEO at Prophesee.


Luca Verre, Co-Founder and CEO,Prophesee

Interestingly, on their ML page, Prophesee have a number of linked pages, including a password protected page from July 2023.
https://www.prophesee.ai/?s=machine+learning

Protected: Protected: DRAFT Metavision® Intelligence – Event-Based Vision Software

Jul 18, 2023

Password Protected​



This is 4 months after Luca's comment about TENNs was published. NDA?
 
  • Like
  • Thinking
  • Love
Reactions: 31 users

Tothemoon24

Top 20
IMG_0060.jpeg





Faster neuron-style processing to be capable of lightning-quick processing while using 90 per cent power
Mercedes-Benz says it’s being forced to explore new ways of crunching huge amounts of data as it ramps up autonomous driving technology because its current processers are too slow and consume too much energy.
Existing Level 2 adaptive cruise control uses between 70-100W of energy, according to the German car-maker, but the more sophisticated Level 3 cruise, introduced on its latest S-Class, consumes around 400W.
The next-generation Level 4 driverless aids sucks-up as much as 3000W in operation while the final Level 5 being developed to see cars, vans and trucks operate without any human involvement at all will use as much as 20kW of power – and that’s not sustainable on its future electric vehicles.

This has forced engineers to look elsewhere for its solution to more efficient computing.
“The most efficient computer processing unit we know is the brain. As always, biology provides the answer, for a human to carry out the equivalent of L5 autonomy the brain consumes just 20W of power – a fraction of what we were achieving in the lab,” a Mercedes scientist told carsales.
Mercedes will eventually use new-generation neuromorphic computers – currently being developed – that are not only better and react faster are 10 times more efficient than current systems.
mercedes benz neuromorphic computing 3

mercedes benz neuromorphic computing 5

mercedes benz neuromorphic computing 4

Sadly, to fully understand the tech needs a doctorate but the most basic explanation is that it uses silicon neurons to form a neural network that mimics the brain.
While existing tech wastes time processing everything a camera ‘sees’, the silicon neurons only react to spikes of information and the resulting event-triggered computation is both quicker and more efficient by up to 90 per cent in its current use, compared to existing tech.
 
  • Like
  • Fire
  • Love
Reactions: 28 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
delayed reaction:

A coupe of years ago there was talk of potential patent litigation between iniVation and Prophesee.

Does the merger of iniVation and SynSense preclude further cooperation between Prophesee and SynSense?

https://www.synsense.ai/synsense-an...orm-leading-neuromorphic-technology-provider/

SynSense and iniVation join forces to form leading neuromorphic technology provider​

2024-02-01

Who would Prophesee turn to for NN if this is the case?

I'd thought that Prophesee had deveoped its low-fi 320*320 pixel array so as not to overstretch SynSense capabilities.

https://brainchip.com/brainchip-introduces-second-generation-akida-platform/

BrainChip Introduces Second-Generation Akida Platform

Introduces Vision Transformers and Spatial-Temporal Convolution for radically fast, hyper-efficient and secure Edge AIoT products, untethered from the cloud
Laguna Hills, Calif. – March 6, 2023
...
“At Prophesee, we are driven by the pursuit of groundbreaking innovation addressing event-based vision solutions. Combining our highly efficient neuromorphic-enabled Metavision sensing approach with Brainchip’s Akida neuromorphic processor holds great potential for developers of high-performance, low-power Edge AI applications. We value our partnership with BrainChip and look forward to getting started with their 2nd generation Akida platform, supporting vision transformers and TENNs,” said Luca Verre, Co-Founder and CEO at Prophesee.


Luca Verre, Co-Founder and CEO,Prophesee

Interestingly, on their ML page, Prophesee have a number of linked pages, including a password protected page from July 2023.
https://www.prophesee.ai/?s=machine+learning

Protected: Protected: DRAFT Metavision® Intelligence – Event-Based Vision Software

Jul 18, 2023

Password Protected​



This is 4 months after Luca's comment about TENNs was published. NDA?


Hi Diogense,

Wow, great point!

The other thing that's really good for us is the time-line of events that were involved. Just to recap, SynSense and Prophesee announced their partnership on October 15, 2021. But then on the podcast in March 2023 with Rob Telson and Luca Verre, Luca spoke in glowing terms about BrainChip.

If you listen around the 26 minute mark, Luca talks about how BrainChip and Prophesee are natural partners because they both have very complementary technologies. He then recalls just how excited Christoph was, saying that beforehand they only had half the story and now they can tell their customers the full story. Luca also said because of this they can now push offerings to their customers to unprecedented levels in their industry.

I can't imagine Synsense being very happy hearing Luca waxing lyrical about BrainChip in this manner.

 
  • Like
  • Love
  • Fire
Reactions: 33 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
That's a great point @manny100.

It might be a simplistic way of looking at things, but for me it boils down to two particular areas:
  • performance and efficiency calculated by TOPS/watt, and
  • the unique processing capabilities that BrainChip's technology can bring to the table
If BrainChip's technology can deliver on both of these areas in a meaningful way, then there should be no reason for Qualcomm not to want to integrate it into their products. As has been mentioned numerous times, we are trying to position ourselves as a partner rather than a competitor to behemoths like Qualcomm and NVIDIA.

In terms of determining the TOPS/watt capabilities of Qualcomm's chips, it's a bit tricky because there doesn't seem to be any specified values from the manufacturer on power consumption measurements. In this instance, the authors of one article (published 20 June 2024, linked below) estimated via their own power consumption measurements that "the most efficient power range of the Snapdragon X Elite chips seems to be 20-30 Watts".

AKIDA's Pico on the other hand operates in the microwatt (μW) to milliwatt (mW) range.

When it comes to the process that BrainChip's technology allows for, we can look to Max Maxfiled's latest article "Taking the Size and Power of Extreme Edge AI/ML to the Extreme Minimum" dated 21 Nov 2024. The obvious benefit is that you can "feed the event-based data from the camera directly to the event-based Akida processor, thereby cutting latency and power consumption to a minimum", as compared to other available techniques.

The big question is whether Qualcomm would see any value in adopting this type technology into their own products and I think that Judd Heape, VP for product management of camera, computer vision and video at Qualcomm Technologies, might have actually answered that question in an EE Times article dated 22 March 2023 when he stated the following.


EXTRACT - EE Times article dated 22 March 2023 (Interview with Judd Heape, VP for product management of camera, computer vision and video at Qualcomm Technologies).

View attachment 73247





EXTRACT - Notebook Check 20 June 2024
View attachment 73241


EXTRACT - Notebook Check 20 June 2024
View attachment 73244


EXTRACT - EE Journal"Taking the Size and Power of Extreme Edge AI/ML to the Extreme Minimum" 21 Nov 2024.
View attachment 73246


Links


What's also interesting is that in March 2023 at pretty much the same time that BrainChip and Prophesee recorded this podcast, Judd Heape (VP for product management of camera, computer vision and video at Qualcomm Technologies), was interviewed by EE times on Prophesee and Qualcomm's partnership.

In the article titled "Experts Weigh Impact of Prophesee-Qualcomm Deal" dated 22 March 2023, Judd talked about how Qualcomm are interested in low power use cases like gesture recognition to control the car while your driving (see above post). Can you imagine how much extra power that would consume in a vehicle? So if gesture recognition is to become a reality, then both Prophesee and Qualcomm would be looking for something that isn't just low power, but ULTRA LOW POWER to manage all of that additional processing work I suppose.

 
  • Like
  • Love
  • Fire
Reactions: 26 users
Top Bottom