BRN Discussion Ongoing

HopalongPetrovski

I'm Spartacus!
Yes that could have been me and yes we will do our best to make sure dubious posters from HC or elsewhere don't start 'infiltrating' this place, for want if a better word.
Another thing to remember when comparing new and old holders is that most of the old holders held their shares well below 10c for a very long time. To be fair.
Hi Guys and Gals.
I have been in BRN since 2015 and been “underwater” for the vast majority of that time.
So believe me when I tell you that I understand some of the emotion and internal self talk that arises.

I held and averaged down for years at a paper loss without publicly moaning and bitching about it, gradually building my position, because I believe in the Company and what they were trying to build, and even though all the way through I hoped for fast gratification on the share price front, I understood that this is largely out of my hands.

For much of that time there was none to little information available regarding our Company nor the area of neuromorphic computing.
Indeed the thread over at the other place was the most prolific source I found.
Over time it became obvious that everyone posting was running their own agenda, some more consciously than others.
A few genuine helpful souls seeking betterment for all, some benign, some hostile, and many who were merely avaricious.
The worst part about it was the unjust and seemingly biased agenda driven moderation favouring a particular element inflicted upon us by the people running the site.
I think Fact Finder described it best when he termed it as a “conflict model” designed to engage likes and keep people coming back, much like the “if it bleeds, it leads” model adopted by much of the general media.

And so when Zeeb0t set this alternative up many of us abandoned the old crapper and settled in here instead and for most of the time it has been a breath of fresh air.

All the genuine holders here want our Company to succeed and our share price to reflect that to justify our investment and faith in the Company.
But, much of what happens in a changing world is beyond the control of the people running our Company.
Also, sometimes errors are made, and the Company needs to regroup and move in another direction.
Also, growth is hard, and people looking with hindsight at either share price’s or other lost opportunities may be indulging in woulda, shoulda, coulda fantasies. Sure, its good and wise to learn from revealed errors and discern apparent patterns, but in the end we are constantly changing with the world and no single action eternally defines us.

So, in an effort to stop the decline here and avoid creating another cesspool like the crapper, lets perhaps cut out the “gotcha” posts that just propagate more of the same.
If you don’t like or see any particular relevance with any particular comment, thread or personality just skip on over till you find something that resonates positively.
If you find something that is absolutely egregious then bring it to Zeeb0ts attention.
He is very effective at eliminating obvious threats.

And if you are here just to promote discord or to try to instil fear and doubt in our community, then fuck off!
AKIDA BALLISTA
AKIDA EVERYWHERE
GLTAH
 
  • Like
  • Love
  • Fire
Reactions: 112 users

LexLuther77

Regular
Yet the price is being suppressed… why aren’t we moving with other tech on this delightful green day 🤔😤
 
  • Like
  • Thinking
Reactions: 3 users

wilzy123

Founding Member
Yet the price is being suppressed… why aren’t we moving with other tech on this delightful green day 🤔😤

Tech is up a meagre 2%.... hold onto your rocket gifs

1673828509240.png
 
  • Like
  • Fire
Reactions: 10 users

Mccabe84

Regular
Yet the price is being suppressed… why aren’t we moving with other tech on this delightful green day 🤔😤
My personal opinion is shorts might be starting to close their positions.. I’m waiting for the days to come out put to see though. But hey I’m probably wrong
 
  • Like
Reactions: 6 users

misslou

Founding Member
Hi Guys and Gals.
I have been in BRN since 2015 and been “underwater” for the vast majority of that time.
So believe me when I tell you that I understand some of the emotion and internal self talk that arises.

I held and averaged down for years at a paper loss without publicly moaning and bitching about it, gradually building my position, because I believe in the Company and what they were trying to build, and even though all the way through I hoped for fast gratification on the share price front, I understood that this is largely out of my hands.

For much of that time there was none to little information available regarding our Company nor the area of neuromorphic computing.
Indeed the thread over at the other place was the most prolific source I found.
Over time it became obvious that everyone posting was running their own agenda, some more consciously than others.
A few genuine helpful souls seeking betterment for all, some benign, some hostile, and many who were merely avaricious.
The worst part about it was the unjust and seemingly biased agenda driven moderation favouring a particular element inflicted upon us by the people running the site.
I think Fact Finder described it best when he termed it as a “conflict model” designed to engage likes and keep people coming back, much like the “if it bleeds, it leads” model adopted by much of the general media.

And so when Zeeb0t set this alternative up many of us abandoned the old crapper and settled in here instead and for most of the time it has been a breath of fresh air.

All the genuine holders here want our Company to succeed and our share price to reflect that to justify our investment and faith in the Company.
But, much of what happens in a changing world is beyond the control of the people running our Company.
Also, sometimes errors are made, and the Company needs to regroup and move in another direction.
Also, growth is hard, and people looking with hindsight at either share price’s or other lost opportunities may be indulging in woulda, shoulda, coulda fantasies. Sure, its good and wise to learn from revealed errors and discern apparent patterns, but in the end we are constantly changing with the world and no single action eternally defines us.

So, in an effort to stop the decline here and avoid creating another cesspool like the crapper, lets perhaps cut out the “gotcha” posts that just propagate more of the same.
If you don’t like or see any particular relevance with any particular comment, thread or personality just skip on over till you find something that resonates positively.
If you find something that is absolutely egregious then bring it to Zeeb0ts attention.
He is very effective at eliminating obvious threats.

And if you are here just to promote discord or to try to instil fear and doubt in our community, then fuck off!
AKIDA BALLISTA
AKIDA EVERYWHERE
GLTAH
I love this post so much. Especially the last sentence 😂 Thanks for your wise thoughts Hoppy ❤🔥🙏
 
  • Like
  • Haha
  • Love
Reactions: 26 users
Nice to see a paper presented to give some measurable comparison to others.

Obviously doesn't cover all competitors and I'm sure I read once before when formal MLPerf testing done it's not cheap...someone can correct that if not the case.

Anyway, I see certain players in the industry have been talking about lowering / creating a newer benchmark standard.

Recent article below with some relevant statements imo.

Maybe they need to lower it already as per the comment in the conclusion...though presume not many out there could meet it yet haha



Will Floating Point 8 Solve AI/ML Overhead?​


Less precision equals lower power, but standards are required to make this work.
JANUARY 12TH, 2023 - BY: KAREN HEYMAN


While the media buzzes about the Turing Test-busting results of ChatGPT, engineers are focused on the hardware challenges of running large language models and other deep learning networks. High on the ML punch list is how to run models more efficiently using less power, especially in critical applications like self-driving vehicles where latency becomes a matter of life or death.

AI already has led to a rethinking of computer architectures, in which the conventional von Neumann structure is replaced by near-compute and at-memory floorplans. But novel layouts aren’t enough to achieve the power reductions and speed increases required for deep learning networks. The industry also is updating the standards for floating-point (FP) arithmetic.

“There is a great deal of research and study on new data types in AI, as it is an area of rapid innovation,” said David Bell, product marketing director, Tensilica IP at Cadence. “Eight-bit floating-point (FP8) data types are being explored as a means to minimize hardware — both compute resources and memory — while preserving accuracy for network models as their complexities grow.”

As part of that effort, researchers at Arm, Intel, and Nvidia published a white paper proposing “FP8 Formats for Deep Learning.” [1]

“Bit precision has been a very active topic of debate in machine learning for several years,” said Steve Roddy, chief marketing officer at Quadric. “Six or eight years ago when models began to explode in size (parameter count), the sheer volume of shuffling weight data into and out of training compute (either CPU or GPU) became the performance limiting bottleneck in large training runs. Faced with a choice of ever more expensive memory interfaces, such as HBM, or cutting bit precision in training, a number of companies experimented successfully with lower-precision floats. Now that networks have continued to grow exponentially in size, the exploration of FP8 is the next logical step in reducing training bandwidth demands.”

How we got here

Floating-point arithmetic is a kind of scientific notation, which condenses the number of digits needed to represent a number. This trick is pulled off by an arithmetic expression first codified by IEEE working group 754 in 1986, when floating-point operations generally were performed on a co-processor.

IEEE 754 describes how the radix point (more commonly known in English as the “decimal” point) doesn’t have a fixed position, but rather “floats” where needed in the expression. It allows numbers with extremely long streams of digits (whether originally to the left or right of a fixed point) to fit into the limited bit-space of computers. It works in either base 10 or base 2, and it’s essential for computing, given that binary numbers extend to many more digits than decimal numbers (100 = 1100100).

Fig. 1: 12.345 as a base-10 floating-point number. Source: Wikipedia

Fig. 1: 12.345 as a base-10 floating-point number. Source: Wikipedia

While this is both an elegant solution and the bane of computer science students worldwide, its terms are key to understanding how precision is achieved in AI. The statement has three parts:
  1. A sign bit, which determines whether the number is positive (0) or negative (1);
  2. An exponent, which determines the position of the radix point, and
  3. A mantissa, or significand, which represents the most significant digits of the number.
Fig. 2: IEEE 754 Floating Point scheme. Source: WikiHow

Fig. 2: IEEE 754 floating-point scheme. Source: WikiHow
As shown in figure 2, while the exponent gains 3 bits in a 64-bit representation, the mantissa jumps from 32 bits to 52 bits. Its length is key to precision.

IEEE 754, which defines FP32 bit and FP64, was designed for scientific computing, in which precision was the ultimate consideration. Currently, IEEE working group P3109 is developing a new standard for machine learning, aligned with the current (2019) version of 754. P3109 aims to create a floating-point 8 standard.

Precision tradeoffs

Machine learning often needs less precision than a 32-bit scheme. The white paper proposes two different flavors of FP8: E4M3 (4-bit exponent and 3-bit mantissa) and E5M2 (5-bit exponent and 2-bit mantissa).

“Neural networks are a bit strange in that they are actually remarkably tolerant to relatively low precision,” said Richard Grisenthwaite, executive vice president and chief architect at Arm. “In our paper, we showed you don’t need 32 bits of mantissa for precision. You can use only two or three bits, and four or five bits of exponent will give you sufficient dynamic range. You really don’t need the massive precision that was defined in 754, which was designed for finite element analysis and other highly precise arithmetic tasks.”
Consider a real-world example: A weather forecast needs the extreme ranges of 754, but a self-driving car doesn’t need the fine-grained recognition of image search. The salient point is not whether it’s a boy or girl in the middle of the road. It’s just that the vehicle must immediately stop, with no time to waste on calculating additional details. So it’s fine to use a floating point with a smaller exponent and much smaller mantissa, especially for edge devices, which need to optimize energy usage.

“Energy is a fundamental quantity and no one’s going to make it go away as an issue,” said Martin Snelgrove, CTO of Untether AI. “And it’s also not a narrow one. Worrying about energy means you can’t afford to be sloppy in your software or your arithmetic. If doing a 32-bit floating point makes everything easier, but massively more power consuming, you just can’t do it. Throwing an extra 1,000 layers at something makes it slightly more accurate, but the value for power isn’t there. There’s an overall discipline about energy — the physics says you’re going to pay attention to this, whether you like it or not.”

In fact, to save energy and performance overhead, many deep learning networks had already shifted to an IEEE-approved 16-bit floating point and other formats, including mantissa-less integers. [2]

“Because compute energy and storage is at a premium in devices, nearly all high-performance device/edge deployments of ML always have been in INT8,” Quadric’s Roddy said. “Nearly all NPUs and accelerators are INT-8 optimized. An FP32 multiply-accumulate calculation takes nearly 10X the energy of an INT8 MAC, so the rationale is obvious.”

Why FP8 is necessary

The problem starts with the basic design of a deep learning network. In the early days of AI, there were simple, one-layer models that only operated in a feedforward manner. In 1986, David Rumelart, Geoffrey Hinton, and Ronald Williams published a breakthrough paper on back-propagation [3] that kicked off the modern era of AI. As their abstract describes, “The procedure repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector. As a result of the weight adjustments, internal ‘hidden’ units, which are not part of the input or output, come to represent important features of the task domain, and the regularities in the task are captured by the interactions of these units.”
In other words, they created a system in which better results could be achieved by adding more and more layers into a model, which would be improved by incorporating “learned” adjustments. Decades later, their ideas so vastly improved machine translation and transcription that college professors remain unsure whether undergraduates’ essays have been written by bots.
But additional layers require additional processing power. “Larger networks with more and more layers were found to be progressively more successful at neural networks tasks, but in certain applications this success came with an ultimately unmanageable increase in memory footprint, power consumption, and compute resources. It became imperative to reduce the size of the data elements (activations, weights, gradients) from 32 bits, and so the industry started using 16-bit formats, such as Bfloat16 and IEEE FP16,” according to the paper jointly written by Arm/Intel/Nvidia.

“The tradeoff fundamentally is with an 8-bit floating-point number compared to a 32-bit one,” said Grisenthwaite. “I can have four times the number of weights and activations in the same amount of memory, and I can get far more computational throughput as well. All of that means I can get much higher performance. I can make the models more involved. I can have more weights and activations at each of the layers. And that’s proved to be more useful than each of the individual points being hyper-accurate.”
Behind these issues are the two basic functions in machine learning, training and inference. Training is the first step in which, for example, the AI learns to classify features in an image by reviewing a dataset. With inference, the AI is given novel images outside of the training set and asked to classify them. If all goes as it should, the AI should distinguish that tails and wings are not human features, and at finer levels, that airplanes do not have feathers and a tube with a tail and wings is not a bird.

“If you’re doing training or inference, the math is identical,” said Ron Lowman, strategic marketing manager for IoT at Synopsys. “The difference is you do training over a known data set thousands of times, maybe even millions of times, to train what the results will be. Once that’s done, then you take an unknown picture and it will tell you what it should be. From a math perspective, a hardware perspective, that’s the big difference. So when you do training, you want to do that in parallel, rather than doing it in a single hardware implementation, because the time it takes to do training is very costly. It could take weeks or months, or even years in some cases, and that just costs too much.”
In industry, training and inference have become separate specialties, each with its own dedicated teams.

“Most companies that are deploying AI have a team of data scientists that create neural network architectures and train the networks using their datasets,” said Bob Beachler, vice president of product at Untether AI. “Most of the autonomous vehicle companies have their own data sets, and they use that as a differentiating factor. They train using their data sets on these novel network architectures that they come up with, which they feel gives them better accuracy. Then that gets taken to a different team, which does the actual implementation in the car. That is the inference portion of it.”

Training requires a wide dynamic range for the continual adjustment of coefficients that is the hallmark of backpropagation. The inference phase is computing on the inputs, rather than learning, so it needs much less dynamic range. “Once you’ve trained the network, you’re not tweaking the coefficients, and the dynamic range required is dramatically reduced,” explained Beachler.
For inference, continuing operations in FP32 or FP16 is just unnecessary overhead, so there’s a quantization step to shift the network down to FP8 or Integer 8 (Int8), which has become something of a de facto standard for inference, driven largely by TensorFlow.

“The idea of quantization is you’re taking all the floating point 32 bits of your model and you’re essentially cramming it into an eight-bit format,” said Gordon Cooper, product manager for Synopsys’ Vision and AI Processor IP. “We’ve done accuracy tests and for almost every neural network-based object detection. We can go from 32-bit floating point to Integer 8 with less than 1% accuracy loss.”
For quality/assurance, there’s often post-quantization retraining to see how converting the floating-point value has affected the network, which could iterate through several passes.

This is why training and inference can be performed using different hardware. “For example, a common pattern we’ve seen is accelerators using NVIDIA GPUs, which then end up running the inference on general purpose CPUs,” said Grisenthwaite.

The other approach is chips purpose-built for inference.

“We’re an inference accelerator. We don’t do training at all,” says Untether AI’s Beachler. “We place the entire neural network on our chip, every layer and every node, feed data at high bandwidth into our chip, resulting in each and every layer of the network computed inside our chip. It’s massively parallelized multiprocessing. Our chip has 511 processors, each of them with single instruction multiple data (SIMD) processing. The processing elements are essentially multiply/accumulate functions, directly attached to memory. We call this the Energy Centric AI computing architecture. This Energy Centric AI Computing architecture results in a very short distance for the coefficients of a matrix vector to travel, and the activations come in through each processing element in a row-based approach. So the activation comes in, we load the coefficients, do the matrix mathematics, do the multiply/accumulate, store the value, move the activation to the next row, and move on. Short distances of data movement equates to low power consumption.”

In broad outline, AI development started with CPUs, often with FP co-processors, then moved to GPUs, and now is splitting into a two-step process of GPUs (although some still use CPUs) for training and CPUs or dedicated chips for inference.

The creators of general-purpose CPU architectures and dedicated inference solutions may disagree on which approach will dominate. But they all agree that the key to a successful handoff between training and inference is a floating-point standard that minimizes the performance overhead and risk of errors during quantization and transferring operations between chips. Several companies, including NVIDIA, Intel, and Untether, have brought out FP8-based chips.

“It’s an interesting paper,” said Cooper. “8-bit floating point, or FP8, is more important on the training side. But the benefits they’re talking about with FP8 on the inference side is that you possibly can skip the quantization. And you get to match the format of what you’ve done between training and inference.”

Nevertheless, as always, there are still many challenges still to consider.

“The cost is one of model conversion — FP32 trained model converted to INT8. And that conversion cost is significant and labor intensive,” said Roddy. “But if FP8 becomes real, and if the popular training tools begin to develop ML models with FP8 as the native format, it could be a huge boon to embedded inference deployments. Eight-bit weights take the same storage space, whether they are INT8 or FP8. The energy cost of moving 8 bits (DDR to NPU, etc.) is the same, regardless of format. And a Float8 multiply-accumulate is not significantly more power consumptive than an INT8 MAC. FP8 would rapidly be adopted across the silicon landscape. But the key is not whether processor licensors would rapidly adopt FP8. It’s whether the mathematicians building training tools can and will make the switch.”

Conclusion

As the quest for lower power continues, there’s debate about whether there might even be a FP4 standard, in which only 4 bits carry a sign, an exponent, and mantissa.
People who follow a strict neuromorphic interpretation have even discussed binary neural networks, in which the input functions like an axon spike, just 0 or 1.

“Our sparsity level is going to go up,” said Untether’s Snelgrove. “There are hundreds of papers a day on new neural net techniques. Any one of them could completely revolutionize the field. If you talk to me in a year, all of these words could mean different things.”
At least at the moment, it’s hard to imagine that lower FPs or integer schemes could contain enough information for practical purposes. Right now, various flavors of FP8 are undergoing the slow grind towards standardization. For example, Graphcore, AMD, and Qualcomm have also brought a detailed FP8 proposal to the IEEE. [4]

“The advent of 8-bit floating point offers tremendous performance and efficiency benefits for AI compute,” said Simon Knowles, CTO and co-founder of Graphcore. “It is also an opportunity for the industry to settle on a single, open standard, rather than ushering in a confusing mix of competing formats.”

Indeed, everyone is optimistic there will be a standard — eventually. “We’re involved in IEEE P3109, as are many, many companies in this industry,” said Arm’s Grisenthwaite. “The committee has looked at all sorts of different formats. There are some really interesting ones out there. Some of them will stand the test of time, and some of them will fall by the wayside. We all want to make sure we’ve got complete compatibility and don’t just say, ‘Well, we’ve got six different competing formats and it’s all a mess, but we’ll call it a standard.”
 
  • Like
  • Love
  • Fire
Reactions: 26 users

stuart888

Regular
Hi Stuart,

I'm not sure I follow your drift.

"I was thinking about the input spike data. For more simple pattern detection, my assumption it is typically an array of numbers. I could be way off with the voice or text, but video pixels would be numbers. Sound would be numbers, and all packed into an array, I believe. A big array, lots of numbers."

There's quite a lot I don't understand about how Akida works, but here is a summary of my understanding.

Sensor data is analog. For instance, a pixel puts out a voltage signal whose amplitude is proportional to the strength of light hitting the pixel.

A digital camera will then have an Analog to Digital Converter (ADC) to convert the analog pixel signal to digital bits, say 16 bits. As you say, these digital numbers are arranged in an array corresponding to the pixel array.

The matrix of these digital bits is what CNN operates on to classify the object being photographed. When the CNN compares small segments (sliding windows) of the matrix with the images stored in the 16-bit image model library, it performs a Multiply Accumulate computation (MAC) by multiplying the two 16-bit numbers together for each pixel. The processing is sequential as the sampling window "slides" across the larger matrix.

Akida uses N-of-M coding and the JAST rules directly on the analog pixel signal.

The strength of the light impinging on a pixel determines both the amplitude of the analog signal and the time when the signal is generated, the stronger the light input, the faster the analog signal is generated. N-of-M coding utilizes this timing to select the strongest incoming signals (N signals) and to ignore the later arriving (M-N) signals because most of the image information is contained in the stronger signals. The pixels are processed in parallel so the timing of the input analog signals can be compared, hence the millions of neurons.

In addition, where the content of adjacent pixels is the same (within a predetermined threshold value), only the first pixel is processed. This provides sparsity in processing so only edges of uniform image segments are processed.

Thus
CNN requires the ADC process for every pixel - Akida does not need ADC.
CNN processes every pixel - Akida only processes N-of-M pixels
CNN processes every pixel - Akida only processes changes in pixel strength.
CNN uses MAC for every pixel - Akida does not use MAC.
CNN runs entirely in software on a CPU. Akida runs in a SoC almost without CPU involvement*
CNN requires the fetching of instructions and image data from memory (von Neumann bottleneck) - Akida stores the data with the NPUs.

*Akida 2-bit and 4-bit operation now uses of the order of 3% CPU operation and 97% silicon for recurrent NNs.
https://brainchip.com/4-bits-are-enough/
You @Diogenese are Mel Fisher Gold. Much thanks.

1673831138344.png


 
  • Like
  • Wow
Reactions: 10 users

Mccabe84

Regular
Brainchip just shared it, seems to be more happening this year than last year

 
  • Like
  • Fire
Reactions: 36 users
Found this blog site from Taiwan / China by looks, which I thought had some nice info on various Tiny ML MCU's in one spot.

Curious as to which ones we may have compatibility with :unsure:

Need to use translator obviously.



1673832488577.png
 
  • Like
Reactions: 4 users

jk6199

Regular
Another positive article and the price still held down lol.

It’s pretty obvious the pressure builds more and the wick is getting shorter, boom!
 
  • Like
  • Fire
Reactions: 18 users

stuart888

Regular

Hi @Diogenese

Have you come across this before. Unfortunately need to pay to get full access.

Advanced Materials
Research Article

A MoS2 Hafnium Oxide Based Ferroelectric Encoder for Temporal-Efficient Spiking Neural Network​


Yu-Chieh Chien, Heng Xiang, Yufei Shi, Ngoc Thanh Duong, Sifan Li, Kah-Wee Ang
First published: 11 November 2022

https://doi.org/10.1002/adma.202204949
Read the full text

PDF PDF
TOOLS

SHARE

Abstract​

Spiking neural network (SNN), where the information is evaluated recurrently through spikes, has manifested significant promises to minimize the energy expenditure in data-intensive machine learning and artificial intelligence. Among these applications, the artificial neural encoders are essential to convert the external stimuli to a spiking format that can be subsequently fed to the neural network. Here, a molybdenum disulfide (MoS2) hafnium oxide-based ferroelectric encoder is demonstrated for temporal-efficient information processing in SNN. The fast domain switching attribute associated with the polycrystalline nature of hafnium oxide-based ferroelectric material is exploited for spike encoding, rendering it suitable for realizing biomimetic encoders. Accordingly, a high-performance ferroelectric encoder is achieved, featuring a superior switching efficiency, negligible charge trapping effect, and robust ferroelectric response, which successfully enable a broad dynamic range. Furthermore, an SNN is simulated to verify the precision of the encoded information, in which an average inference accuracy of 95.14% can be achieved, using the Modified National Insitute of Standards and Technology (MNIST) dataset for digit classification. Moreover, this ferroelectric encoder manifests prominent resilience against noise injection with an overall prediction accuracy of 94.73% under various Gaussian noise levels, showing practical promises to reduce the computational load for the neural network.
So true. Brainchip's SNN solution reduces "energy expenditure", beneficial AI. CES was all about low power, green, smarts. We win there.🍾🍾🍾

What a broad use-case covering a myriad of everything, low power smarts. Every CTO and CEO says the same thing. Brainchip's solutions are energy saving, perfect time, year, it is all coming together like the perfect play book.

I think the eco-green movement is going to be louder and louder, and products are going to want to sell green, low power smarts. Cloud was beloved, now frowned upon slightly. Cloud has its place in AI inference for sure, for the gas hog!🛻🚙🚕

Mr Spike SNN has manifested significant promises to minimize the energy expenditure in data-intensive machine learning and artificial intelligence.
 
  • Like
  • Fire
  • Love
Reactions: 28 users

Mercfan

Member
Weren't you the one that had to change your Avatar because it was a teen girl in her underpants who turned out to be a porn star who you said was your daughter?

Street cred now = 0
I only changed it because i didn't need stupid comments from people.
 
When is the next quarterly being released? Anyone got the date?
No but Morningstar Corporate Cenlar has the following for February.


We22 nd
Books Close:
YTMAZ2
Div Pay Date:
CGFPB
Report (Annual):
BRN*
, CRN*, NIC*, NZM*, SCA*, SCW*, VNT*
Report (Interim):
1AD*, ACQ*, AEF*, AFL*, AGI*, AIM*, ALC*, ALT*, AMH*, ANP*, APA*, AVG*, AVN*, AYU*, BCK*, BEL*, BIT*, BNO*, BUB*, CBL*, CBR*, CDO*, CGO*, COS*, CQR*, CTE*, CUE*, CUV*, CXL*, DMP*, DUG*, ECP*, ECS*, EML*, EVS*, EXP*, EZL*, FCT*, FGL*, FRI*, GFL*, HDN*, HGV*, HLS*, HTG*, HVM*, HZN*, ICR*, IDT*, IDX*, IMM*, INV*, JAY*, JCS*, KLS*, KYP*, KZA*, LBL*, LGL*, LYL*, MCP*, MEZ*, MGX*, MLG*, MMS*, NMR*, NXD*, NXT*, OAK*, OCC*, ODA*, OEQ*, OPH*, OPN*, PFP*, PFT*, PIC*, PLS*, PRU*, PSI*, PTB*, PTM*, PXA*, QRI*, QUE*, REG*, RFG*, RMS*, RMY*, RNO*, S66*, SBM*, SFC*, SGP*, SPK*, SRV*, SYM*, TGF*, TIP*, UBN*, UNI*, WAT*, WGB*, WLE*, WMA*, WOR*, WOW*, WQG*, WTC*, X2M*, YOW*, Z2U*, ZLD*
Report (Prelim):
29M*, BFL*, BRN*, CRN*, CYC*, EXP*, MFD*, NIC*, NXS*, NZM*, RCT*, RIO*, SCA*, SCG*, SCW*, SYD*, VLS*, VNT*

SC
 
  • Like
Reactions: 6 users

stuart888

Regular
Just posted, edge impulse are putting out a lot of tweets lately


Awesome @Mccabe84. Just think about the exact verbiage, hunting for clues!

Brainchip Akida version one: "easy, quick, and advanced model development and deployment".

Spiking SNN Akida is now easy and quick to develop and implement. Massive help by Edge Impulse to Brainchip. Maybe equal to the other announcements.

Edge Impulse is also boosting mega confidence in the Brainchip Technical Team to deliver SNN solutions that are best in class. What a great educational partnership. 👨‍🎓
 
  • Like
  • Fire
Reactions: 29 users
Another positive article and the price still held down lol.

It’s pretty obvious the pressure builds more and the wick is getting shorter, boom!
Hopefully my money has cleared into my supers investement account before boom... then i add my 20% from super, sit back and wait for BOOOM :)
 
  • Like
  • Fire
  • Love
Reactions: 21 users

wilzy123

Founding Member
  • Haha
  • Like
Reactions: 14 users

mkm109

Regular
Yeah, there's a lot of "Artificial" intelligence in parliament house!
Hummm may be remove the word intelligence..
;-)
 
  • Haha
  • Like
  • Fire
Reactions: 11 users

miaeffect

Oat latte lover
  • Haha
  • Like
Reactions: 12 users

Diogenese

Top 20
Nice to see a paper presented to give some measurable comparison to others.

Obviously doesn't cover all competitors and I'm sure I read once before when formal MLPerf testing done it's not cheap...someone can correct that if not the case.

Anyway, I see certain players in the industry have been talking about lowering / creating a newer benchmark standard.

Recent article below with some relevant statements imo.

Maybe they need to lower it already as per the comment in the conclusion...though presume not many out there could meet it yet haha



Will Floating Point 8 Solve AI/ML Overhead?​


Less precision equals lower power, but standards are required to make this work.
JANUARY 12TH, 2023 - BY: KAREN HEYMAN


While the media buzzes about the Turing Test-busting results of ChatGPT, engineers are focused on the hardware challenges of running large language models and other deep learning networks. High on the ML punch list is how to run models more efficiently using less power, especially in critical applications like self-driving vehicles where latency becomes a matter of life or death.

AI already has led to a rethinking of computer architectures, in which the conventional von Neumann structure is replaced by near-compute and at-memory floorplans. But novel layouts aren’t enough to achieve the power reductions and speed increases required for deep learning networks. The industry also is updating the standards for floating-point (FP) arithmetic.

“There is a great deal of research and study on new data types in AI, as it is an area of rapid innovation,” said David Bell, product marketing director, Tensilica IP at Cadence. “Eight-bit floating-point (FP8) data types are being explored as a means to minimize hardware — both compute resources and memory — while preserving accuracy for network models as their complexities grow.”

As part of that effort, researchers at Arm, Intel, and Nvidia published a white paper proposing “FP8 Formats for Deep Learning.” [1]

“Bit precision has been a very active topic of debate in machine learning for several years,” said Steve Roddy, chief marketing officer at Quadric. “Six or eight years ago when models began to explode in size (parameter count), the sheer volume of shuffling weight data into and out of training compute (either CPU or GPU) became the performance limiting bottleneck in large training runs. Faced with a choice of ever more expensive memory interfaces, such as HBM, or cutting bit precision in training, a number of companies experimented successfully with lower-precision floats. Now that networks have continued to grow exponentially in size, the exploration of FP8 is the next logical step in reducing training bandwidth demands.”

How we got here

Floating-point arithmetic is a kind of scientific notation, which condenses the number of digits needed to represent a number. This trick is pulled off by an arithmetic expression first codified by IEEE working group 754 in 1986, when floating-point operations generally were performed on a co-processor.

IEEE 754 describes how the radix point (more commonly known in English as the “decimal” point) doesn’t have a fixed position, but rather “floats” where needed in the expression. It allows numbers with extremely long streams of digits (whether originally to the left or right of a fixed point) to fit into the limited bit-space of computers. It works in either base 10 or base 2, and it’s essential for computing, given that binary numbers extend to many more digits than decimal numbers (100 = 1100100).

Fig. 1: 12.345 as a base-10 floating-point number. Source: Wikipedia

Fig. 1: 12.345 as a base-10 floating-point number. Source: Wikipedia

While this is both an elegant solution and the bane of computer science students worldwide, its terms are key to understanding how precision is achieved in AI. The statement has three parts:
  1. A sign bit, which determines whether the number is positive (0) or negative (1);
  2. An exponent, which determines the position of the radix point, and
  3. A mantissa, or significand, which represents the most significant digits of the number.
Fig. 2: IEEE 754 Floating Point scheme. Source: WikiHow

Fig. 2: IEEE 754 floating-point scheme. Source: WikiHow
As shown in figure 2, while the exponent gains 3 bits in a 64-bit representation, the mantissa jumps from 32 bits to 52 bits. Its length is key to precision.

IEEE 754, which defines FP32 bit and FP64, was designed for scientific computing, in which precision was the ultimate consideration. Currently, IEEE working group P3109 is developing a new standard for machine learning, aligned with the current (2019) version of 754. P3109 aims to create a floating-point 8 standard.

Precision tradeoffs

Machine learning often needs less precision than a 32-bit scheme. The white paper proposes two different flavors of FP8: E4M3 (4-bit exponent and 3-bit mantissa) and E5M2 (5-bit exponent and 2-bit mantissa).

“Neural networks are a bit strange in that they are actually remarkably tolerant to relatively low precision,” said Richard Grisenthwaite, executive vice president and chief architect at Arm. “In our paper, we showed you don’t need 32 bits of mantissa for precision. You can use only two or three bits, and four or five bits of exponent will give you sufficient dynamic range. You really don’t need the massive precision that was defined in 754, which was designed for finite element analysis and other highly precise arithmetic tasks.”
Consider a real-world example: A weather forecast needs the extreme ranges of 754, but a self-driving car doesn’t need the fine-grained recognition of image search. The salient point is not whether it’s a boy or girl in the middle of the road. It’s just that the vehicle must immediately stop, with no time to waste on calculating additional details. So it’s fine to use a floating point with a smaller exponent and much smaller mantissa, especially for edge devices, which need to optimize energy usage.

“Energy is a fundamental quantity and no one’s going to make it go away as an issue,” said Martin Snelgrove, CTO of Untether AI. “And it’s also not a narrow one. Worrying about energy means you can’t afford to be sloppy in your software or your arithmetic. If doing a 32-bit floating point makes everything easier, but massively more power consuming, you just can’t do it. Throwing an extra 1,000 layers at something makes it slightly more accurate, but the value for power isn’t there. There’s an overall discipline about energy — the physics says you’re going to pay attention to this, whether you like it or not.”

In fact, to save energy and performance overhead, many deep learning networks had already shifted to an IEEE-approved 16-bit floating point and other formats, including mantissa-less integers. [2]

“Because compute energy and storage is at a premium in devices, nearly all high-performance device/edge deployments of ML always have been in INT8,” Quadric’s Roddy said. “Nearly all NPUs and accelerators are INT-8 optimized. An FP32 multiply-accumulate calculation takes nearly 10X the energy of an INT8 MAC, so the rationale is obvious.”

Why FP8 is necessary

The problem starts with the basic design of a deep learning network. In the early days of AI, there were simple, one-layer models that only operated in a feedforward manner. In 1986, David Rumelart, Geoffrey Hinton, and Ronald Williams published a breakthrough paper on back-propagation [3] that kicked off the modern era of AI. As their abstract describes, “The procedure repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector. As a result of the weight adjustments, internal ‘hidden’ units, which are not part of the input or output, come to represent important features of the task domain, and the regularities in the task are captured by the interactions of these units.”
In other words, they created a system in which better results could be achieved by adding more and more layers into a model, which would be improved by incorporating “learned” adjustments. Decades later, their ideas so vastly improved machine translation and transcription that college professors remain unsure whether undergraduates’ essays have been written by bots.
But additional layers require additional processing power. “Larger networks with more and more layers were found to be progressively more successful at neural networks tasks, but in certain applications this success came with an ultimately unmanageable increase in memory footprint, power consumption, and compute resources. It became imperative to reduce the size of the data elements (activations, weights, gradients) from 32 bits, and so the industry started using 16-bit formats, such as Bfloat16 and IEEE FP16,” according to the paper jointly written by Arm/Intel/Nvidia.

“The tradeoff fundamentally is with an 8-bit floating-point number compared to a 32-bit one,” said Grisenthwaite. “I can have four times the number of weights and activations in the same amount of memory, and I can get far more computational throughput as well. All of that means I can get much higher performance. I can make the models more involved. I can have more weights and activations at each of the layers. And that’s proved to be more useful than each of the individual points being hyper-accurate.”
Behind these issues are the two basic functions in machine learning, training and inference. Training is the first step in which, for example, the AI learns to classify features in an image by reviewing a dataset. With inference, the AI is given novel images outside of the training set and asked to classify them. If all goes as it should, the AI should distinguish that tails and wings are not human features, and at finer levels, that airplanes do not have feathers and a tube with a tail and wings is not a bird.

“If you’re doing training or inference, the math is identical,” said Ron Lowman, strategic marketing manager for IoT at Synopsys. “The difference is you do training over a known data set thousands of times, maybe even millions of times, to train what the results will be. Once that’s done, then you take an unknown picture and it will tell you what it should be. From a math perspective, a hardware perspective, that’s the big difference. So when you do training, you want to do that in parallel, rather than doing it in a single hardware implementation, because the time it takes to do training is very costly. It could take weeks or months, or even years in some cases, and that just costs too much.”
In industry, training and inference have become separate specialties, each with its own dedicated teams.

“Most companies that are deploying AI have a team of data scientists that create neural network architectures and train the networks using their datasets,” said Bob Beachler, vice president of product at Untether AI. “Most of the autonomous vehicle companies have their own data sets, and they use that as a differentiating factor. They train using their data sets on these novel network architectures that they come up with, which they feel gives them better accuracy. Then that gets taken to a different team, which does the actual implementation in the car. That is the inference portion of it.”

Training requires a wide dynamic range for the continual adjustment of coefficients that is the hallmark of backpropagation. The inference phase is computing on the inputs, rather than learning, so it needs much less dynamic range. “Once you’ve trained the network, you’re not tweaking the coefficients, and the dynamic range required is dramatically reduced,” explained Beachler.
For inference, continuing operations in FP32 or FP16 is just unnecessary overhead, so there’s a quantization step to shift the network down to FP8 or Integer 8 (Int8), which has become something of a de facto standard for inference, driven largely by TensorFlow.

“The idea of quantization is you’re taking all the floating point 32 bits of your model and you’re essentially cramming it into an eight-bit format,” said Gordon Cooper, product manager for Synopsys’ Vision and AI Processor IP. “We’ve done accuracy tests and for almost every neural network-based object detection. We can go from 32-bit floating point to Integer 8 with less than 1% accuracy loss.”
For quality/assurance, there’s often post-quantization retraining to see how converting the floating-point value has affected the network, which could iterate through several passes.

This is why training and inference can be performed using different hardware. “For example, a common pattern we’ve seen is accelerators using NVIDIA GPUs, which then end up running the inference on general purpose CPUs,” said Grisenthwaite.

The other approach is chips purpose-built for inference.

“We’re an inference accelerator. We don’t do training at all,” says Untether AI’s Beachler. “We place the entire neural network on our chip, every layer and every node, feed data at high bandwidth into our chip, resulting in each and every layer of the network computed inside our chip. It’s massively parallelized multiprocessing. Our chip has 511 processors, each of them with single instruction multiple data (SIMD) processing. The processing elements are essentially multiply/accumulate functions, directly attached to memory. We call this the Energy Centric AI computing architecture. This Energy Centric AI Computing architecture results in a very short distance for the coefficients of a matrix vector to travel, and the activations come in through each processing element in a row-based approach. So the activation comes in, we load the coefficients, do the matrix mathematics, do the multiply/accumulate, store the value, move the activation to the next row, and move on. Short distances of data movement equates to low power consumption.”

In broad outline, AI development started with CPUs, often with FP co-processors, then moved to GPUs, and now is splitting into a two-step process of GPUs (although some still use CPUs) for training and CPUs or dedicated chips for inference.

The creators of general-purpose CPU architectures and dedicated inference solutions may disagree on which approach will dominate. But they all agree that the key to a successful handoff between training and inference is a floating-point standard that minimizes the performance overhead and risk of errors during quantization and transferring operations between chips. Several companies, including NVIDIA, Intel, and Untether, have brought out FP8-based chips.

“It’s an interesting paper,” said Cooper. “8-bit floating point, or FP8, is more important on the training side. But the benefits they’re talking about with FP8 on the inference side is that you possibly can skip the quantization. And you get to match the format of what you’ve done between training and inference.”

Nevertheless, as always, there are still many challenges still to consider.

“The cost is one of model conversion — FP32 trained model converted to INT8. And that conversion cost is significant and labor intensive,” said Roddy. “But if FP8 becomes real, and if the popular training tools begin to develop ML models with FP8 as the native format, it could be a huge boon to embedded inference deployments. Eight-bit weights take the same storage space, whether they are INT8 or FP8. The energy cost of moving 8 bits (DDR to NPU, etc.) is the same, regardless of format. And a Float8 multiply-accumulate is not significantly more power consumptive than an INT8 MAC. FP8 would rapidly be adopted across the silicon landscape. But the key is not whether processor licensors would rapidly adopt FP8. It’s whether the mathematicians building training tools can and will make the switch.”

Conclusion

As the quest for lower power continues, there’s debate about whether there might even be a FP4 standard, in which only 4 bits carry a sign, an exponent, and mantissa.
People who follow a strict neuromorphic interpretation have even discussed binary neural networks, in which the input functions like an axon spike, just 0 or 1.

“Our sparsity level is going to go up,” said Untether’s Snelgrove. “There are hundreds of papers a day on new neural net techniques. Any one of them could completely revolutionize the field. If you talk to me in a year, all of these words could mean different things.”
At least at the moment, it’s hard to imagine that lower FPs or integer schemes could contain enough information for practical purposes. Right now, various flavors of FP8 are undergoing the slow grind towards standardization. For example, Graphcore, AMD, and Qualcomm have also brought a detailed FP8 proposal to the IEEE. [4]

“The advent of 8-bit floating point offers tremendous performance and efficiency benefits for AI compute,” said Simon Knowles, CTO and co-founder of Graphcore. “It is also an opportunity for the industry to settle on a single, open standard, rather than ushering in a confusing mix of competing formats.”

Indeed, everyone is optimistic there will be a standard — eventually. “We’re involved in IEEE P3109, as are many, many companies in this industry,” said Arm’s Grisenthwaite. “The committee has looked at all sorts of different formats. There are some really interesting ones out there. Some of them will stand the test of time, and some of them will fall by the wayside. We all want to make sure we’ve got complete compatibility and don’t just say, ‘Well, we’ve got six different competing formats and it’s all a mess, but we’ll call it a standard.”
Hi Fmf,

As someone once said:

"4 bits are enough".

...

and what an outlandish idea - binary NNs!

As PvdM has pointed out, the IBM simulation demonstrated that 4-bit deep-learning models in vision, speech, and language, lose little in comparison with 16-bit deep learning.

https://brainchip.com/4-bits-are-enough/

"To dive a little bit deeper into the value of 4-bit, in its 2020 NeurIPS paper IBM described the various pieces that are already present and how they come together. They prove the readiness and the benefit through several experiments simulating 4-bit training for a variety of deep-learning models in computer vision, speech, and natural language processing. The results show a minimal loss of accuracy in the models’ overall performance compared with 16-bit deep learning. The results are also more than seven times faster and seven times more energy efficient."

... and Akida does it with not a MAC in sight, just massively parallel neurons and skeletal sparsity.


As long ago as September 2022, ARM, Intel and Nvidia came to understand that:

Neural networks are a bit strange in that they are actually remarkably tolerant to relatively low precision,” said Richard Grisenthwaite, executive vice president and chief architect at Arm. “In our paper*, we showed you don’t need 32 bits of mantissa for precision. You can use only two or three bits, and four or five bits of exponent will give you sufficient dynamic range. You really don’t need the massive precision that was defined in 754 [ IEEE 754 floating-point scheme ], which was designed for finite element analysis and other highly precise arithmetic tasks.”


*

FP8 Formats for Deep Learning​

Paulius Micikevicius, Dusan Stosic, Neil Burgess, Marius Cornea, Pradeep Dubey, Richard Grisenthwaite, Sangwon Ha, Alexander Heinecke, Patrick Judd, John Kamalu, Naveen Mellempudi, Stuart Oberman, Mohammad Shoeybi, Michael Siu, Hao Wu
FP8 is a natural progression for accelerating deep learning training inference beyond the 16-bit formats common in modern processors. In this paper we propose an 8-bit floating point (FP8) binary interchange format consisting of two encodings - E4M3 (4-bit exponent and 3-bit mantissa) and E5M2 (5-bit exponent and 2-bit mantissa). While E5M2 follows IEEE 754 conventions for representatio of special values, E4M3's dynamic range is extended by not representing infinities and having only one mantissa bit-pattern for NaNs. We demonstrate the efficacy of the FP8 format on a variety of image and language tasks, effectively matching the result quality achieved by 16-bit training sessions. Our study covers the main modern neural network architectures - CNNs, RNNs, and Transformer-based models, leaving all the hyperparameters unchanged from the 16-bit baseline training sessions. Our training experiments include large, up to 175B parameter, language models. We also examine FP8 post-training-quantization of language models trained using 16-bit formats that resisted fixed point int8 quantization.
Subjects:Machine Learning (cs.LG)
Cite as:arXiv:2209.05433 [cs.LG]
(or arXiv:2209.05433v2 [cs.LG] for this version)
https://doi.org/10.48550/arXiv.2209.05433
Focus to learn more

Submission history​

From: Paulius Micikevicius [view email]
[v1] Mon, 12 Sep 2022 17:39:55 UTC (117 KB)
[v2] Thu, 29 Sep 2022 20:47:07 UTC (117 KB)


Thinking of mantissas and exponents locks the thought process and the mathematical implementation into MACs (Multiply Accumulate calculations).

It's like making a traffic code for dinosaurs.
 
Last edited:
  • Like
  • Fire
  • Haha
Reactions: 38 users

chapman89

Founding Member
Here is a research paper from 2021 I came across which contemplates the use of SNN and DVS for predator robots funded by DARPA.

They conclude by suggesting Loihi 1 but as we know AKIDA can do everything better than Loihi 1 and 2 combined again.

https://cpb-us-w2.wpmucdn.com/sites...ent-Cameras_to_Central_Pattern_Generation.pdf

“This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TCDS.2021.3097675, IEEE Transactions on Cognitive and Developmental Systems
An End-to-end Spiking Neural Network Platform for Edge Robotics: From Event-Cameras to Central Pattern Generation
Ashwin Lele Yan Fang
Abstract—Learning to adapt one’s gait with environmental changes plays an essential role in locomotion of legged robots which remains challenging for constrained computing resources and energy budget, as in the case of edge-robots. Recent ad- vances in bio-inspired vision with dynamic vision sensors (DVS) and associated neuromorphic processing can provide promising solutions for end-to-end sensing, cognition and control tasks. However, such bio-mimetic closed-loop robotic systems based on event-based visual sensing and actuation in the form of spiking neural networks (SNN) have not been well explored. In this work, we program the weights of a bio-mimetic multi- gait central pattern generator (CPG) and couple it with DVS- based visual data processing to show a spike-only closed-loop robotic system for a prey-tracking scenario. We first propose a supervised learning rule based on stochastic weight updates to produce a multi-gait producing Spiking-CPG (SCPG) for hexa- pod robot locomotion. We then actuate the SCPG to seamlessly transition between the gaits for a nearest prey tracking task by incorporating SNN based visual processing for input event-data generated by the DVS. This for the first time, demonstrates the natural coupling of event data flow from event-camera through SNN and neuromorphic locomotion. Thus, we exploit bio-mimetic dynamics and energy advantages of spike-based processing for autonomous edge-robotics.
I. KEYWORDS:
Edge Intelligence, Spiking Neural Networks, Central Pattern Generation, Hexapods, Dynamic Vision Sensor (DVS) Cameras

VII. CONCLUSION
We develop an end-to-end neuromorphic system that takes event-based visual data from a dynamic visual sensor and generates adaptive gait patterns for the locomotion of a hexa- pod robot in a predator-prey tracking scenario. The proposed method is fully bio-inspired carrying out the sensing to ac- tuation in event-based processing. For learning various gaits, we propose a supervision-based weight adaptation algorithm to learn multiple gaits in a single CPG. Furthermore, the programmed SCPG is coupled with DVS to form a closed- loop control system that processes binary events to mimic predatory behaviour when the hexapod is approaching prey. Benefiting from the event-driven sensing and data processing, the proposed method shows high energy efficiency if this is implemented on well-known neuromorphic hardware (Intel’s Loihi platform). We demonstrate the feasibility of end-to-end neuromorphic systems for resource-constrained edge robotics.
ACKNOWLEDGEMENT
This work was supported by CBRIC, one of six centers in JUMP, a Semiconductor Research Corporation (SRC) program sponsored by DARPA”

Bit science fiction predator robots stalking the enemy but it could also be used for finding victims in disaster scenes. If we add in Prophesee to the AKIDA equation the turning of this research into reality is not far off.

Many more use cases will begin to emerge, thanks to the advancement of akida and Peter and Anil.
 
  • Like
  • Fire
  • Love
Reactions: 44 users
Top Bottom