BRN Discussion Ongoing

Traders can’t predict the market, but maybe their faces can..

In the AGE just now this morning. Note the University and the Technology involved. Also the timelines.

"The saying goes that our eyes are the window to the soul. Perhaps over time they’ll serve a less romantic purpose, as windows to making money.

Researchers at Carnegie Mellon University in Pittsburgh, one of the leading institutions for artificial-intelligence research, have embarked on a study using facial-recognition algorithms to track the expressions of traders. Their goal: finding correlations between mood swings and market swings. If the traders look enthusiastic, it might be time to buy. Are there more furrowed brows than usual? Could be time to sell. The provisional US patent application was filed on September 13, 2022"

See link for full article and more details.
Traders can't predict the market ..

Maybe Akida will come to our rescue as shareholders creating value in ways we don't image
 
Last edited:
  • Like
  • Thinking
  • Fire
Reactions: 25 users

Lex555

Regular
Musk Ox is planning on using software NNs.

Musk Shares Details on FSD Beta v11: Neural Nets to Be Used for Vehicle Control

January 15, 2023
By Nuno Cristovao


https://www.notateslaapp.com/softwa...11-neural-nets-to-be-used-for-vehicle-control

...

Neural Nets for Vehicle Behavior​

A week ago Musk said this upgrade will include 'many major improvements.' Last night Musk revealed some additional details. He said there will be "many small things," one of which will be that Tesla will begin to use neural nets for vehicle navigation and control, instead of just vision.

Today Tesla uses neural networks to determine the vehicle's surroundings, where objects are, what they are, and their distances from the vehicle to create a 3D environment known as 'vector space.' With this information, the vehicle can then plan a path and navigate around these objects toward its destination.

However, based on Musk's comment, it sounds like Tesla is currently only using neural nets to determine its environment and not for controlling the vehicle. This means that how the vehicle behaves, how it finds a path, and how it moves is still a process that is coded traditionally.

In the same way that Tesla uses millions of images to determine what a stop sign or traffic cone is, it sounds like Tesla will now use a large number of examples to determine how to best control the vehicle in various situations
.

Surely he can't be doing mission critical functions on the internet.

Sounds like Akida could improve Tesla mileage by 100 km or more.
Interesting indeed dio. If Akida improved efficiency by 100km a manufacturer such as Tesla could reduce pack size by ~20% for same range. As of 2022 cost of battery pack was approximately $138USD per kWh.

Meaning for a Model S, 20kWh could be reduced which would save $2760. That’s big money when manufacturing millions of cars a year.
 
  • Like
  • Love
  • Fire
Reactions: 24 users

Bravo

If ARM was an arm, BRN would be its biceps💪!
Greetings Groovy People,

Check out this article published a few hours ago. This mentions the partnership between Prophesee and Datalogic and also discusses other companies using event-based vision systems such as Sony and Nikon. Because such systems can significantly improve efficiency and increase the amount of data collected it is going to be indispensable in automating a range of manufacturing processes, including counting, quality inspection, and predictive maintenance. It says here " As investment in this space increases, the market is expected to drive growth in other industries at an exponential rate through at least 2030".

Sweet!🍯





Event-Based Vision: Where Tech Meets Biology​

January 17, 2023
By Brad MarleyContributing Editor
Vision-System.jpg

Machine vision systems are helpful in automating a range of manufacturing processes, including counting, quality inspection, and predictive maintenance. However, most vision systems in use today rely on frame-based image capture technology that has been around for more than a hundred years.
The next iteration in vision systems relies on what changes in a particular scene, or a specific “event” that happens. The technology takes cues from human biology, namely how efficiently eyes work to process massive amounts of visual data. Event-based vision is based on neuromorphic computing in which machines process information much like how the brain processes information. This can significantly improve efficiency and increase the amount of usable data collected.

A Smart Catalyst for Change​

“The introduction of 3D- and AI- (artificial intelligence) based vision systems have really changed the game when it comes to event-based vision systems in the manufacturing space,” said Dan Simmons, senior sales engineer, Datalogic, which has its U.S. headquarters in Eugene, Ore. “Before you had to know how to program a system to do what you wanted it to do. The introduction of AI helps create a vision system that helps you learn what is ‘good’ and what is ‘bad.’”
Simmons explained that AI learns the deviations from good and bad, and from there it makes better determinations without having to wait for a human operator to step in and make changes as it goes. But an AI system is only as good as the images you give it, he added.
On the factory floor, there are three application categories for event-based vision systems. The first lies in optical character recognition. In this scenario, the camera is being used for traceability. The AI can view characters that normally cannot be viewed with a standard vision system that is capable of optical character recognition. This could mean anything from reading characters on a non-flat surface to characters that have a very low print quality.
The second category focuses on error proofing. When it comes to quality, mistakes aren’t tolerated if a company wants to ensure success and produce the products its customers have come to expect—even if we’re just talking about junk food.
“I remember hearing about a use case where a food manufacturer wanted to ensure the right cookie was being placed in the right bag during its trip down the factory line,” Simmons said. “The company taught the vision system to simply recognize that there was writing on the package.”
He explained that it came down to the contrast between black and white, with the system able to decipher black writing on the white bag to let it proceed. If no writing was detected, the process was stopped.
The third application falls within cameras that require calibration, such as vision-guided robotics or applications that require measurement, such as an outer diameter measurement.
For example, machines can be taught to view the gap between spark plugs to ensure width accuracy. Then it becomes a simple pass or fail report if the gap isn’t accurate. The manufacturer can then archive that data to use when teaching a next-generation system what it needs to know for a similar job.
Understanding that manufacturers don’t always have the time or capability to facilitate these teachable moments, Datalogic rolled out its IMPACT Robot Guidance system that helps customers take advantage of smart robots by quickly and easily interfacing between any smart camera or vision processor.
“Our IMPACT software is proven to let users solve not only guidance, but many other machine vision applications with an intuitive drag and drop interface,” said Simmons. “With more than 100 vision tools, our customers won’t have to fret about not finding a guidance system that fits their needs.”
automotive-quality-control.jpg
The APDIS Laser is a fast, fully automated, non-contact inspection replacement to traditional CMMs for automotive quality control on the shop floor. (Provided by Nikon Metrology)

Making an Impact on Metrology​

Proper measurement is vital when it comes to quality assurance and part calibration to help mitigate risk and ensure parts are built to proper specifications.
One company thriving in the metrology space is Nikon, a name most would recognize as a producer of high-end cameras, camera lenses, and microscopes. Whereas once you saw cameras hanging around the necks of tourists and amateur photographers, the phones in our pockets have taken over, leaving Nikon with a gap to fill in its offerings.
Nikon brings more than one hundred years of experience in lenses and scopes. The company revolutionized quality control and metrology across a wide range of clients, using innovative techniques such as laser-radar systems to help automotive companies, for instance, measure gaps between door frames and window holes in automobile frames.
“The benefits of an event-based vision system are very similar to what we offer our manufacturing customers,” said Pete Morken, senior application engineer, Nikon Metrology, which has a U.S. office in Brighton, Mich. “Our systems help to measure whether or not a part is good or bad simply by scanning a car body on the assembly line.”
With Nikon’s laser-radar stations, manufacturers can measure the geometry of parts—car doors, whole car chassis, etc.—as alternatives to the slow, lumbering horizontal-arm coordinate measurement machine systems (CMMS).
In a typical CMMS, information is gathered slowly offline by the software and stored in a database where it can be accessed later when decisions about non-conformances need to be made. But where it lacks the ability of a laser-radar system is the speed that a company like Nikon can offer to make sure that information is used more efficiently.
“With our laser-radar system, the measurement that our customers obtain can be collected, analyzed, and reported more quickly, using more data, to see improved process quality,” Morken said. “The use of pre-defined positions eliminates the requirement for further programming after installation, so the measurement program can happen immediately and continuously.”
The camera company is well-positioned to improve measurement possibilities for customers.
“Nikon has always lived at the cutting edge of technology, even as far back as its photography advances that re-shaped how we take pictures,” added Morken. “Bringing in an event-based vision system could do for metrology what the company once did for budding photographers.”
As technology advances, companies are starting to see how combining artificial intelligence with vision systems represents that next iteration of this process, and how it can re-shape how manufacturers are able to view products and parts.
At its core, a vision system enables machines to “see” necessary objects, whether it’s a part in a bin or package of cookies. In the past, companies would have to teach the machine the parts or products it needed to scan, and the machine was then limited by what it had learned. If there was a flaw in the product, the machine might not know it was an imperfection because it wasn’t taught to recognize it.
As previously noted, artificial intelligence and machine learning can be used to teach manufacturing systems the difference between good and bad parts. As a result, algorithms become less important while the AI does most of the work.
Clean_MV430.jpg
Nikon Metrology’s APDIS Laser Radar.

Seeing is Collecting​

Audio-video giant Sony is working to take vision systems to the next level. Similar to Nikon’s transformation, Sony aims to carve out its spot in the event-based vision system industry by creating sensors that act like retinas in the human eye.
The tiny sensors are becoming ever smaller, which allows more of them to be fitted on a device to boost data collection volumes. The use of these sensors goes far beyond the manufacturing floor. As the technology improves, Sony sees deployment within collision avoidance systems, drones, and event-based 3D cameras.
Sony recently introduced what it touts as the world’s first intelligent vision sensors equipped with AI processing functionality. One highlight: The new chip will be to identify people and objects.
This would allow cameras with the chip to identify stock levels on a store shelf or use heat maps to track and analyze customer behavior. It could even count and forecast the number of customers in a given location, providing valuable data to calculate when foot traffic is highest.
Where the technology stands to shine the most in manufacturing is around data management. Advanced sensors can identify objects and send a description of what they see without having to include an accompanying image that takes up space in the database. This could reduce storage requirements by up to 10,000 times, leaving companies with more space to gather critical data that they previously haven’t been able to access, while giving AI a looser leash to capture relevant information.

Working Together​

As technology evolves, partnerships between companies in the event-based vision system space and those that want to deploy across other industries will become commonplace.
Datalogic is joining forces with Paris-based Prophesee, a company that invented advanced neuromorphic vision systems and is working to build the next generation of industrial products.
“We are conducting a very fruitful partnership with Prophesee,” said Michele Benedetti, chief technology officer at Datalogic. “Neuromorphic vision is a fascinating technology inspired by the behavior of the human biological system, exactly like neural networks. We believe that the combination of these technologies will provide innovative solutions to our customers.”
As investment in this space increases, the market is expected to drive growth in other industries at an exponential rate through at least 2030, according to a Grand View Research report on the U.S. machine vision market. The increasing demand for quality inspection, as well as the need for vision-guided robotic systems, is expected to fuel that growth.
While long-term forecasts for emerging technologies are far from an exact science, the future for event-based vision systems looks promising—giving manufacturers cause to be fitted with a pair of 20/20 rose-colored specs.

 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 54 users

Learning

Learning to the Top 🕵‍♂️
More Free Advertising!


There is a short podcast at the above link also

BrainChip Holdings Ltd. (ASX:BRN) has published a paper titled "Benchmarking AI Inference at the Edge: Measuring Performance and Efficiency for Real-World Deployments," which "evaluates the current state of edge AI benchmarks and the need to continually improve metrics that measure performance and efficiency of real-world, power-conscious edge AI deployments." Anil Mankar, the company's Chief Development Officer, explained"

"While there's been a good start, current methods of benchmarking for edge AI don't accurately account for the factors that affect devices in industries such as automotive, smart homes and Industry 4.0. We believe that as a community, we should evolve benchmarks to continuously incorporate factors such as on-chip, in-memory computation and model sizes to complement the latency and power metrics that are measured today."
--------
A report published by Brand Essence Research finds that the global market for Conversational AI is projected to grow from $8.24 billion USD in 2022 to $32.51 billion by 2028, registering a compound annual growth rate (CAGR) of 21.6 percent in the forecast period. The following excerpt from the report's summary outlines the role of COVID-19 in influencing the market's growth:
--------
Learning 🏖
 
  • Like
  • Fire
  • Love
Reactions: 48 users

31D28A18-6EE1-4A4C-9936-F2CC50952793.jpeg
 
  • Like
  • Fire
  • Love
Reactions: 39 users

Learning

Learning to the Top 🕵‍♂️
Not directly about BrainChip. But we know Brainchip's partner Prophesee is working with Sony;


Counterpoint: Sony's smartphone camera sensor business is on the rise thanks to iPhone upgrades
PETER, 17 JANUARY 2023

The smartphone market shrunk in 2022, which impacted the suppliers of image sensors. Some fared better than others though – Sony was the only supplier to see its revenue grow on a yearly basis.

And it was mostly thanks to Apple upgrading the cameras on the iPhone 14 series. The two Pro models brought new 48MP sensors in the main cameras and larger 12MP sensors in the ultra wide cameras. The selfie cam on all four models was upgraded with autofocus too. Apple exclusively uses Sony sensors, you can see the breakdown by camera type below:.
Screenshot_20230118_061510_Chrome.jpg

The Sony sensors inside the last two generations of Apple iPhonesThe Sony sensors inside the last two generations of iPhones (source: Counterpoint BoM analysis service)
Adding it all together, Sony made an extra $6 per unit for a total of around $300 million in the second half of 2022. The end result is that Sony took in 54% of the total revenue for the year, up 5 percentage points compared to 2021.

Samsung LSI did well for itself, even though its revenue share contracted by 1 percentage point to 29%. The company raked in the benefits of high resolution, small pixel size sensors (sub-0.7µm pixels).

The affordable 50MP sensors proved quite popular and Samsung shipped an estimated 200 million of them in 2022. These are used in the main cameras of lower end phones and in the selfie cameras of more premium devices. The company still dominates the 100+ megapixel sensor market and shipped an estimated 150 million units since it launched the first one.
Screenshot_20230118_061602_Chrome.jpg

Last year the smartphone image sensor market contracted by 6% compared to 2021, but the total revenue remained above $13 billion. Sony and Samsung took in the lion’s share of that, 83% in total.

Source

Learning 🏖
 
  • Like
  • Fire
  • Love
Reactions: 30 users

Moonshot

Regular
2023 year of the Akida Spikformer?

Conclusion

In this work we explored the feasibility of implementing the self-attention mechanism and Transformer in Spiking Neuron Networks and propose Spikformer based on a new Spiking Self-Attention (SSA). Unlike the vanilla self-attention mechanism in ANNs, SSA is specifically designed for SNNs and spike data. We drop the complex operation of softmax in SSA, and instead perform matrix dot- product directly on spike-form Query, Key, and Value, which is efficient and avoids multiplications. In addition, this simple self-attention mechanism makes Spikformer work surprisingly well on both static and neuromorphic datasets. With directly training from scratch, Spiking Transformer outperforms the state-of-the-art SNNs models. We hope our investigations pave the way for further research on transformer-based SNNs models.

https://arxiv.org/pdf/2209.15425.pdf
 
  • Like
  • Fire
  • Love
Reactions: 20 users

BaconLover

Founding Member
Researchers at Carnegie Mellon University in Pittsburgh, one of the leading institutions for artificial-intelligence research, have embarked on a study using facial-recognition algorithms to track the expressions of traders. Their goal: finding correlations between mood swings and market swings. If the traders look enthusiastic, it might be time to buy. Are there more furrowed brows than usual? Could be time to sell. The provisional US patent application was filed on September 13, 2022"

I like Mellon.


“We have incorporated experimentation with BrainChip’s Akida development boards in our new graduate-level course, “Neuromorphic Computer Architecture and Processor Design” at Carnegie Mellon University during the Spring 2022 semester,” said John Paul Shen, Professor, Electrical and Computer Engineering Department at Carnegie Mellon. “Our students had a great experience in using the Akida development environment and analyzing results from the Akida hardware. We look forward to running and expanding this program in 2023.”

 
  • Like
  • Love
  • Fire
Reactions: 60 users

gex

Regular
On Mellon, was doing a search and this comes up. can't get any further. Could be nothing mind you

1673999719466.png
 
  • Like
  • Fire
  • Love
Reactions: 29 users

somme

Member
I like Mellon.


“We have incorporated experimentation with BrainChip’s Akida development boards in our new graduate-level course, “Neuromorphic Computer Architecture and Processor Design” at Carnegie Mellon University during the Spring 2022 semester,” said John Paul Shen, Professor, Electrical and Computer Engineering Department at Carnegie Mellon. “Our students had a great experience in using the Akida development environment and analyzing results from the Akida hardware. We look forward to running and expanding this program in 2023.”

Hi Guys,

Just gone through the promo for the new Apple Macbook Pro that arrived in my Email as a Apple user..
5 times faster than Intel, better battery life, uses less power, greater capacity etc etc.
Has anybody else seen this? Could this be what we are looking for?
 
  • Like
  • Thinking
  • Love
Reactions: 23 users

Shadow59

Regular
Hi Guys,

Just gone through the promo for the new Apple Macbook Pro that arrived in my Email as a Apple user..
5 times faster than Intel, better battery life, uses less power, greater capacity etc etc.
Has anybody else seen this? Could this be what we are looking for?
...or could this be why Intel has suddenly decided to include Akida in their offerings.
 
  • Like
  • Thinking
  • Fire
Reactions: 23 users
Shorters seem to be stuck in there positions, there isn’t enough liquidity for them to get out. I beat they are trying to hold us down until we move out of the ASX index which should create liquidity for them to close their positions. Fact is they are racing against time, any moment we could have an Ann that blows them out of the water.
 

Evermont

Stealth Mode
Markus Schafer’s neuromorphic article is finally up


For those not aware, Markus Schafer is the CTO of Development & Procurement and is also a Member of the Board of Management of Mercedes-Benz. Markus is highly regarded across the industry and has almost 55,000 followers alone on LinkedIn. His post on neuromorphic computing highlighting BrainChip has only been live 14 hours but has already been re-posted 10 times extending the tentacles across multiple audiences.

A few may consider the content underwhelming, however for me, and IMO this coverage does a number of things.

1. Takes opportunity to identify BrainChip as one of two leading developers (alongside Intel)
2. Provides irrefutable proof of neuromorphic computings strategic importance to Mercedes future fleet development
3. Indicates that Mercedes are researching wide use cases that will have a direct impact on future AI functionality across their vehicles
4. Validates the ongoing partnership with BrainChip
5. Confirms that experts are working closely to examine new applications

6. Provides clue that even small node application is transferring to measurable advancement across multiple applications
7. Reinforces that a single use case for Akida (Hey Mercedes) has demonstrated efficiency improvements of 5-10 times
8. Clearly flags that efficiency improvements are paramount given the increasing use of AI technology within vehicles

Cheers.

NB - Perhaps also a timely reminder on the LinkedIn comments, let's keep it professional please.

1674001036857.png

1674001065519.png
 
Last edited:
  • Like
  • Love
  • Fire
Reactions: 71 users

Learning

Learning to the Top 🕵‍♂️
  • Like
  • Fire
  • Love
Reactions: 38 users

Dhm

Regular
  • Like
  • Fire
Reactions: 17 users
Just (ICYMI) BrainChip sharing Markus Schafer’s Post, Great advertisement! Thanks Markus.

It's had been mentioned, but fantastic for a CTO of Mercedes referencing Neuromorphic Computing to only 'BrainChip and Intel' 😎🎉🥳

View attachment 27350
Just (ICYMI) BrainChip sharing Markus Schafer’s Post, Great advertisement! Thanks Markus.

It's had been mentioned, but fantastic for a CTO of Mercedes referencing Neuromorphic Computing to only 'BrainChip and Intel' 😎🎉🥳

View attachment 27350

Learning 🏖
Thanks to the 1000 eyes who voted and ensured neuromorphic computing was the clear winner!!
 
  • Like
  • Love
  • Fire
Reactions: 40 users

Diogenese

Top 20
Have we discussed CEVA before in any detail? The reason I ask is that they seem to be describing what we have.

https://www.chipestimate.com/techtalk.php?d=2023-01-17

View attachment 27366
View attachment 27368
View attachment 27369
Hi Dhm,

AS Ella continually reminds us:
"Taint what you do,
It's the way that you do it."

CEVA Figure 1 "Mixed precision neural Engine - 4k MAC"

1674005154254.png


Using MACs and 16 bits is incompatible with Akida SNN. These are von Neumann hangovers.

While they refer to sparsity, there in no indication that CEVA are using N-of-M coding.

This CEVA patent application relates to the cited article:

EP3709225A1 SYSTEM AND METHOD FOR EFFICIENT UTILIZATION OF MULTIPLIERS IN NEURAL-NETWORK COMPUTATIONS

1674005320675.png


A system and method for performing neural network calculations may include selecting a size in bits for representing a plurality of weight elements of the neural network based on a value of the weight elements. In each computational cycle: if the size in bits of a weight element of the plurality of weight elements is N, configuring an N∗K multiply accumulator to perform one multiply-accumulate operation of a K-bit data element and the N-bit weight element; and if the size in bits of at least two N/M-bit weight elements of the plurality of weight elements is N/M, configuring the N∗K multiply accumulator to perform up to N/M multiply-accumulate operations, each of a K-bit data element and an N/M-bit weight element, where N, K and M are integers bigger than one, N is a power of 2, M is even and N≥M.

...
Typically, the neurons and links within a NN are represented by mathematical constructs, such as activation functions and matrices of data elements and weights. A processor, e.g. CPUs or graphics processing units (GPUs), or a dedicated hardware device may perform the relevant calculations.

[0004] NN calculations require performing a huge amount of multiplications, e.g., of the data elements and weights. Typical hardware implementations of NN usually support 16-bit fixed-point precision arithmetic processing. However, the power consumption of such devices becomes a problem in many NN applications.

[0005] Attempts to reduce the power consumption have been made, for example, by reducing the bit precision to 8, 4 or even 1 bit. While reducing the bit precision may indeed reduce the power consumption, it may at the same time reduce the accuracy of the neural network.

SUMMARY OF THE INVENTION
[0006] According to embodiments of the present invention, there is provided a system and method for efficient utilization of multipliers in neural network computations by an execution unit. The method may include for example determining a size in bits of weight elements; configuring an N∗ K multiply accumulator to perform at least two multiply operations in parallel, if the size in bits of at least two weight elements is not bigger than N /M, where K is an integer bigger than one, each of N and M is a power of 2 and N≥M
.
 

Attachments

  • 1674005280153.png
    1674005280153.png
    49.6 KB · Views: 48
  • Like
  • Fire
  • Love
Reactions: 28 users
Good morning everyone.

Can we stop sharing Motley Fool articles here?

BRN holders share it everywhere and then we complain that picklebro writes non stop about Brainchip.

They get maximum exposure with Brainchip when holders write and discuss their articles constantly. They have every reason to write it because we create the buzz, make a platform and share it far and wide.

Stop posting their bs articles here and then the buzz goes away and they'll find another company.

This is my "soft" opinion only, you can continue to share it if you wish, thought I'd make the suggestion 😉 .
You're 100% right BaconLover 👍

They bait the hook perfectly, with their "views" and we take the bait each time and run with it.

It's a deliberate parasitic strategy and it's working for them.

They obviously need the help, to engage in such an unethical practice, but why should we give it to them?

Have a belly laugh at their bullshit, if you want, but don't spread it around unless your intention is to help their cause.

There are better causes out there..
 
  • Like
  • Love
  • Fire
Reactions: 33 users
So who’s going to send Picklebro a copy of Markus Schaefer’s blog? No need to rub it in too much. Remember that humility is a valuable and admirable attribute. Maybe just very subtly highlight that Markus states that neuromorphic computing is a “highly significant field of computing“ and that we’re described as being “leading developers“ and, without tooting our own horn too much, just quickly mention that we were mentioned ahead of Intel and then maybe also highlight this statement “Together with intense parallel execution on neuromorphic chips, the new processing principles require us to go beyond the application of existing #AI frameworks to neuromorphic chips.”

And then sign off saying “Nah-nah-née-nah-nah”. 😝

Just joking - about the sign-off part that is.🤭
I would rather no one give him a heads up on anything and let him get trampled in the ensuring stampede!!

SC
 
  • Like
  • Haha
  • Fire
Reactions: 19 users
Top Bottom