BRN Discussion Ongoing

MDhere

Top 20
Good Morning Chippers,

Just cut open my peice of morning fruit & look what I found.

😃 .

Impatient shareholder , YES.

* Artists impression , any conclusions drawn from imagery ( šŸŽ ) are purely those of viewers vivid imagination and are NOT those expressed or implied even in slightest by BrainChip management or those in their engage .

Bring on 2nd Generation AKIDA & some revinue.

Regards,
Esq.
pretty sure i read that they are going straighr it IP LICENSING with 2nd gen no need for chip dev. if i need to dig that info up but I pretty sure i read that. in any case nice chip. will give u a beer next tine if u can chisel one of those for me 🤣
 
  • Like
  • Fire
  • Haha
Reactions: 7 users

buena suerte :-)

BOB Bank of Brainchip
Always interesting bits sometimes when doing an Edgar search for a few mins.

These guys report filed end June show they were flicking out stock to short.

PACE SELECT ADVISORS TRUST

Only small amount but they all add up and I haven't bothered going through all lodgements for all US listed funds etc.

View attachment 39337


On the flip side and again a small amount....a back of couch lunch money purchase, is IBM grabbed close to 600,000 shares at some point...probs through State Street or someone.
Lets hope they don't stop there and sift out all the other illegal crap that's going on in the ASX!!! 😔😔

 
  • Like
  • Love
Reactions: 14 users

cosors

šŸ‘€

This made me think of Brainchips demo from way back in 2017/18 ish. Of a car learning to drive a course trough trial and error and adapting to its surroundings compared to traditional machine learning where you have to retrain the model again and again.

It made me wonder how much this robot would improve with the help of Akida. Man I would like to see a demo like this with Akida!

I can’t find that original demo so if anyone who knows where it is could post a link that would be great!:)

Thanks for the very amusing and very interesting video! I would also like to see the Brainchip team get in touch with AI Warehouse.

"In every ā€œAI learns to walkā€ video I’ve seen, the AI either learns to walk in a weird, non-human way, or they use motion capture of a real person walking and simply train the AI to imitate that. I thought it was weird that nobody tried to train it to walk properly from scratch (without any external data), so I wanted to give it a shot! That’s what I said 4 months ago. It’s been really difficult, but I’ve finally managed to do it, so please watch the whole video! The final result ended up being awesome :)

NOTE: From the last video, you guys made it clear you didn’t like that Albert had his brain reset, so from now his brain is here to stay (hopefully)! The next video I make with Albert will start with the brain we trained in this video, so with every video Albert will become more and more capable until he eventually learns to break out of my computer and take over the world. You also can only see one Albert, but there are actually 200 copies of Albert and the room he’s in training behind the camera to speed up the training.

If you want to learn more about how Albert actually works, you can read the rest of this very long comment I wrote explaining exactly how I trained him! (and please let the video play in the background while reading so YouTube will show Albert to more people) I created everything using Unity and ML-Agents. Albert is controlled entirely by an artificial brain (neural network) which has 5 layers, the first layer consists of the inputs (the information Albert is given before taking action, like his limb positions and velocities), the last layer tells him what actions to take and the middle 3 layers, called hidden layers, are where the calculations are performed to convert the inputs into actions. His brain was trained using the standard algorithm in reinforcement learning; proximal policy optimization (PPO).

For each of Albert’s limbs I’ve given him (as an input) the position, velocity, angular velocity, contacts (if it’s touching the ground, wall or obstacle) and the strength applied to it. I’ve also given him the distance from each foot to the ground, direction of the closest target, the direction his body’s moving, his body’s velocity, the distance from his chest to his feet and the amount of time one foot has been in front of the other. As for his actions, we allow Albert to control each body part’s rotation and strength (with some limitations so his arm can’t bend backwards, for example).

Just like the last videos, Albert was trained using reinforcement learning. For each of Albert's attempts, we calculate a score for how 'good' it was and make small, calculated adjustments to his brain to try to encourage the behaviors that led to a higher score and avoid those that led to a lower score. You can think of increasing Albert’s score as rewarding him and decreasing his score as punishing him, or you can think about it like natural selection where the best performing Alberts are most likely to reproduce. For this video there are 13 different types of rewards (ways to calculate Albert's score), we start off with only a couple and with each new room add more, always in an attempt to get him to walk.

Room 1: We start off very simple in the first room, we reward him based on how much he moved to the target and we punish him for moving in the wrong direction. This led to Albert doing the worm towards the target, since he figured out that was the easiest way for him to move the quickest/get the highest score. It would have been possible to get Albert to walk in a janky way by just rewarding him for moving towards the target and also punish him for falling as a team at Google (DeepMind) showed in 2017, but I thought it would make for a more enjoyable video if he starts off with the worm and over time learns to use his legs, rather than immediately being able to partially walk.

Room 2: In the second room we start checking if his limbs hit the ground. If the limb that hits the ground is a foot we reward him (but only if it's in front of his other foot, more on that later), if it isn’t, we punish him. I also made it so Albert wasn’t rewarded at all unless his chest was high enough to force it to at least be partially standing. As seen in the video, this encourages him to not fall over and encourages him to use his feet to do it. We also introduced a new reward designed to encourage smoother movement; if he approaches the maximum strength allowed on a limb he's punished, and he's rewarded if he uses a strength of almost 0. This encourages him to opt for the more human-like movement of using a bit of strength from many limbs as opposed to a lot of strength from one limb.

Room 3: This is where we start to polish Albert’s gait that developed in room 2 and teach him to turn. From here on we start using the chest height calculation as another direct reward where the higher his chest is the more he’s rewarded in an attempt to get him to stand up as straight as possible. These rewards so far give Albert a decent gait, however he’s still not using both of his feet (which was by far the hardest part of this project), so room 4 is designed to do exactly that.

Room 4: We get Albert to take more steps from a few additional rewards. To start, we introduce a 2 second timer that resets when one foot goes in front of the other. We reward Albert whenever this timer is above 0 (the front foot has been in front for < 2 seconds), and we punish him whenever the timer goes below 0 (the front foot has been in front > 2 seconds). We add another reward proportional to the distance of his steps to encourage him to take larger steps. To smooth out the movement, we also add a punishment every frame proportional to the difference in his body’s velocity from the previous frame to the current frame, so if he’s moving at a perfectly consistent velocity he isn’t punished at all, and if he makes very quick erratic movements he’s punished a lot.

Room 5: For the final room the only change I made to the reward function was to go back to an earlier version of a reward. Throughout the other rooms I had been tinkering with how I should reward Albert’s feet being grounded, my initial thought was to only reward the front foot for being grounded to try to get him to put more weight on his front foot when taking steps, but somewhere along the way I changed it to just rewarding Albert for any foot being grounded, and that was the version Albert trained with in rooms 3 and 4. For this final room I switched back to the old front foot grounded reward which resulted in a much nicer looking walk. Also, the video makes it seem like I never reset Albert’s brain, that isn't entirely true, I had to occasionally reset it because of something called decaying plasticity.

Decaying plasticity was a big issue. Basically, Albert’s brain specializes a lot from training in one room, then training in the next room on top of that brain is difficult because he first needs to unlearn that specialization from the first room. The best way to solve the issue is by resetting a random neuron every once in a while so over time he ā€œforgetsā€ the specialization of the network without it ever being noticeable, the problem is I don’t know how to do that through ML-Agents. My solution was to keep training on top of the same brain, but if Albert’s movement doesn’t converge as needed I record another attempt trained from scratch, then stitch the videos together when their movements are similar. If you know how to reset a single neuron in ML-Agents please let me know! The outcome from both methods is exactly the same, but it would be a smoother experience having the neurons reset over time instead of all at once.

For rooms 1 to 4 I only allowed Albert to make a decision every 5 game ticks, but for the final room I removed that constraint and let him make decisions every frame. I found if Albert makes a decision every game tick it’s too difficult for him to commit to any proper movements, he ends up just making very small movements like slightly pushing his front foot forward when he should be taking a full step. The 5 game tick decision time forces him to commit to his decision for at least 5 game ticks so he ends up being more careful when moving a limb. When I recorded him beating the final room I removed this limitation because he’s already learned to commit to his actions so allowing him to make a decision every tick just results in a smoother motion.

If you’re still reading this thank you for being so interested in the project! I’d like to upload much more often than once every few months, and to do that I need some help. I have 2 part time positions open, one for a Unity AI Developer and one for a Unity Scene Designer. It would start off as part time (paid per project) but I’d love to get someone full time provided they’re skilled enough:) If you think you’d be able to help, please apply here for the AI Developer position: forms.gle/rExRJCKcxNmxnBRu5 and here for the Scene Designer position: forms.gle/VafZTMZ8QMruSBiRA I’ve hidden these job postings in this long pinned comment to make sure anybody who applies is interested enough in the videos to actually read the whole comment, so thank you for reading all the way through!:D Also if you have any ideas for how to improve the AI (or solve the issue of decaying plasticity with ML-Agents), include the text "Technical idea" in your comment so I can find it easier! Thank you so much for watching, this video took me 4 months to make, so please, if you enjoyed it or learned something from it, share it with someone you think will also enjoy it! :)"
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 17 users

Jannemann

Member
Hey @TopCat, I reckon I’ve got exciting news to share! Keep your paws crossed that our sleuthing skills will soon be validated by an official announcement - I can already hear you happily purring away!

I only just found the time to watch the recorded CVPR 2023 Workshop presentations by Kynan Eng (CEO of iniVation) and Nandan Nayampally - they both have such wonderfully soothing voices, by the way.

The mosaic of conjecture we’ve both been creating over the past couple of weeks is almost complete; here are two more tiles to lay, one of them a shiny golden one.

While the videos’ tech content is way above my pay grade, both presentations were quite illuminating, nevertheless! Let me share with you what piqued my interest:

I first watched Kynan Eng’s talk on iniVation’s new Aeveon sensor:



The wallaby featured on the iniVation website must have hopped away and is now presumably roaming freely somewhere around Lake Zurich, as it didn’t make an appearance during the presentation. Instead, it had been substituted by a beautiful, iridescent hummingbird, legendary for its rapid wing beat, which is notoriously difficult to capture on regular camera without blurring.

Watch from 8 min onwards, where Kynan Eng starts to talk about the new, soon-to-be released Aeveon sensor:
ā€žWhatā€˜s important to note here is, this is mainly digital. Weā€˜ve moved quite a distance away from the analog circuit in current DVS pixelsā€¦ā€œ

And from 10:45 min onwards: ā€žSo the chip doesn’t exist, yet. We are getting very close to the first tape-out of it now, but because it is in digital, we are able to create an emulator for this, for the chip.ā€


Well, I immediately took a liking to the sound of ā€œdigitalā€, especially since Synsense, iniVation’s sister start-up and partner for their Speck SoC, is into analog SNNs, but it got even better, when I then listened to Nandan Nayampally’s virtual presentation (which explains why he was nowhere to be seen in the CVPR 2023 Workshop group photo - I had actually hoped to spot him next to the likes of Kynan Eng, Tobi Delbrück or AndrĆ© van Schaik, which could have hinted at the fact that he was in conversation with one of them just before the group photo was taken. Clever idea of mine, eh?!)





Just over a minute into the presentation, I couldn’t believe my ears: ā€œLet me start with the key technology changes that we are making to support our partners like Prophesee, iniVation and other folks that are building not only event-based solutionsā€¦ā€

WHAT???!!! My real-time auditory sensor processing DID get that right, didn’t it?! Listen for yourself and correct me if I’m wrong…

Did he really just let slip that iniVation is indeed a Brainchip partner?!

They obviously didn’t get listed under ā€œpartnersā€ on the ā€œBrainchip at a glanceā€ presentation slide, as there hasn’t been any official announcement so far, but IMO this (unintentionally) revealing statement strongly supports our hypothesis.

Admittedly, it could have been a lapsus linguae, a disclosure of our Brainchip CMO’s secretly harboured wish of teaming up with iniVation, but honestly, how probable is that?! Much more likely it was an inadvertent divulgence in front of a small audience of computer vision researchers, plus the Aeveon sensor screams Akida, doesn’t it?
I suppose we’ll find out soon, as ā€œwe are getting very close to the first tape-outā€ of the new wonder chip.

My opinion only, DYOR.


Hi,

i wasnt online for a while now..so i dont know if you got this information.

Mitre > Inivation

The company Mitre (USA)
https://www.mitre.org/
has some interesting connections to the world of neuromorphic computing. They already tested neuromorphic stuff since 2015 from IBM, Intel and so on. So they know this space. Mitre is a big company in the US for Cybersecurity, health, ISS Spacestadion, they are working close with the government and much more.

I attached a file where you can read about a vision camera with Inivation which "WAS" probably working with Synsense at this time. For Mitre...

"Extreme Machine Vision
High-perfomance neuromorphic vision systems for demanding real-time applications"


With that new informations it could be now BRN > Inivation > Mitre..

This is on Inivations Homepage and they admit to work with Brainchip.


it's an old video, but remember the news from edgeimpuls.

https://docs.edgeimpulse.com/expert...g-projects/brainchip-akida-traffic-monitoring

Same usecase with traffic.

So at the bottomline Mitre knows BRN, maybe already more.... Could be interesting to have an eye on Mitre.

I hope this is a new information.
Have a nice day.
 

Attachments

  • Doucette.pdf
    1.5 MB · Views: 256
  • Like
  • Fire
  • Love
Reactions: 42 users
Question: Does anyone know why Brainchip has 2 Linkedin profiles? I once thought that it could be region related, or the one with less followers was an older profile, but neither of those 2 scenarios make sense when you look at the employees on each of the 2 sites.

The positive is that there are now 92 employees across both profiles.

(I hope the answer is that one profile wont be big enough for us in the near future :) )
 
  • Like
  • Haha
  • Fire
Reactions: 16 users
Could someone tell me are we still partners with socionext?

BrainChip and Socionext Provide a New Low-Power Artificial Intelligence Platform for AI Edge Applications.​

It's just that Socionext today on the Tokyo stock exchange was down nearly 23%,
and I was wondering why?

View attachment 39360
Here is an article why:


Haven't looked into why 3 of the top shareholders have sold but looks like it was planned, and doesn't seem to be a real issue for the long term.
 
  • Like
  • Thinking
  • Sad
Reactions: 8 users
Here is an article why:


Haven't looked into why 3 of the top shareholders have sold but looks like it was planned, and doesn't seem to be a real issue for the long term.
Looks like Socionext has also had an impact on other major chip-related firms, such as Intel, Renensas and Texas Instruments who've also had a decline:

 
  • Like
  • Fire
Reactions: 10 users

Tothemoon24

Top 20
3. Development of Neuromorphic Chips
Researchers from the Royal Melbourne Institute of Technology (RMIT) University in Australia have achieved significant breakthroughs in the development of neuromorphic chips.





Written by BIS Research | Jul 4, 2023 8:23:45 AM
The global semiconductor industry plays a pivotal role in powering modern technology, ranging from smartphones and computers to automobiles and advanced medical devices.
However, recent disruptions and challenges in the semiconductor supply chain have highlighted vulnerabilities in the global electronics ecosystem. In response, global economies are intensifying their efforts to innovate and strengthen semiconductor manufacturing technologies, aiming to counter supply chain challenges and ensure a reliable and resilient electronics industry.
This article explores the recent global initiatives boosting semiconductor manufacturing and supply chain resilience.

1. Successful Implementation of IoT in Chip Manufacturing
Semiconductor manufacturers are actively addressing the unique requirements of Internet of Things (IoT) devices, such as smaller sizes, diverse connectivity options, and lower power consumption. They are focusing on the development of sensors and integrated circuits to meet these demands. Flexible multifunctional chipsets are being developed that incorporate more circuits. These chipsets combine microcontrollers and analytics, enhancing the resilience of IoT devices and bringing computing closer to the source. The implementation of IoT in chip manufacturing brings financial benefits through continuous process and asset monitoring and also improves visibility into production operations.
For instance, Taiwanese startup IMOSTAR offers multi-band IoT chips that integrate multiple low-power IoT radios into a single chip, resulting in space and cost savings. These chips feature compact and versatile monolithic antennas, expanding the application range of IoT devices and simplifying their manufacturing process. Similarly, Chinese startup Nano-Core Chip specializes in artificial intelligence of things (AIoT) chips. Their chips leverage event-driven architecture, dynamic charge domain signal chains, closed-loop circuit topology, and memory-computing fusion simulation. These features enable high energy efficiency and a small chip area, supporting AI computations with low latency and high storage density.

2. Integration of AI in Manufacturing Workflows
Semiconductor companies are integrating AI into manufacturing workflows to optimize operations and enhance product quality.
semiconductor1

For instance, South Korean startup Rebellions specializes in domain-specific AI processors that bridge silicon architectures and deep learning algorithms. By modifying processor architecture using silicon kernels, they accelerate machine learning computations, improve performance, and reduce deployment costs. Meanwhile, US-based startup Gauss Labs offers AI-based solutions for semiconductor manufacturing. Its solutions utilize machine sensor measurements and metrology data to predict factory anomalies and provide guidance to engineers, enabling AI-driven precision manufacturing and minimizing disruptions in the process.

3. Development of Neuromorphic Chips
Researchers from the Royal Melbourne Institute of Technology (RMIT) University in Australia have achieved significant breakthroughs in the development of neuromorphic chips. They developed a single-chip device using doped indium oxide. This device mimics human vision and memory, capturing, processing, and storing visual information akin to the human eye, optic nerve, and memory system.
The neuromorphic chip enables ultra-fast decision-making, eliminates the need for energy-intensive computation, and facilitates real-time processing. Through longer memory retention without frequent electrical signals, this advancement reduces energy consumption while enhancing performance. This chip can find applications in bionic vision, autonomous operations, food shelf-life assessment, and advanced forensics.

4. International Collaborations and Partnerships Countering Supply Chain Challenges
The complexity and global nature of the industry require close cooperation among nations, companies, and research institutions. Through strategic alliances and information sharing, countries can pool resources, expertise, and technology to address issues such as raw material shortages, production bottlenecks, and logistics disruptions.
For instance, in March 2023, India and the U.S. joined forces through the India-U.S. initiative on Critical and Emerging Technologies (iCET) to reshape global semiconductor supply chains. The iCET focuses on collaboration in areas such as AI, quantum computing, semiconductors, telecommunications, defense, and space, aiming to address regulatory and supply chain barriers as well as export control issues.

Conclusion

The future of the semiconductor industry holds great promise as advancements in technology address current vulnerabilities, paving the way for transformative breakthroughs and driving economic growth.
The advent of artificial intelligence, 5G, the Internet of Things, and autonomous vehicles will drive increased demand for semiconductors. Emerging areas such as quantum computing and neuromorphic engineering promise to revolutionize the industry further, shaping a world of limitless possibilities.
Interested to know more about the growing technologies in your industry vertical? Get the latest market studies and insights from BIS Research. Connect with us at
 
  • Like
  • Fire
Reactions: 8 users

Jchandel

Regular
A new video from Edge AI and Vision Alliance

Related to the video posted earlier but now posted by Edge Impulse on their LinkedIn page:
1688647212416.jpeg
 

Attachments

  • IMG_7302.jpeg
    IMG_7302.jpeg
    1.1 MB · Views: 115
  • Like
  • Fire
  • Love
Reactions: 55 users
Another item maybe or maybe not posted but I haven't seen it myself.

There was a conference Nov last year in Syd and I see a team from BRN did a presso as below.

Haven't tried to search the paper yet.

Appears based on Akidanet models for agri crop / weed ID at the edge.



View attachment 27769

View attachment 27770
Just googling for dots and this presso / paper I posted about back in Jan popped up.

Couldn't find it back then but it's HERE if anyone wanted a read.

@Diogenese thoughts whenever if you have time or anything of interest in it?

TIA
 
  • Like
  • Love
Reactions: 9 users

Diogenese

Top 20
Just googling for dots and this presso / paper I posted about back in Jan popped up.

Couldn't find it back then but it's HERE if anyone wanted a read.

@Diogenese thoughts whenever if you have time or anything of interest in it?

TIA
Hi Fmf,

The paper discusses a model (AkidaNet) used for distinguishing weed from crop and describes how it was adapted to run on Akida 1000 using CNN2SNN.
The model has sveral layers, the first using 8-bits to deal with the model, the remaining layers running on 4 bit weights and activations. This is interesting in that the recently publishes BRN patent application added ALUs to the NPU, presumably to handle 8-bits more efficiently, so these results would not reflect the performance of Akida 2, or whatever version has the ALUs.


. AkidaNet

In this paper, we propose a new lightweight and energy-efficient model and convert it to a Spiking Neural Network(SNN) for implementation on a NSoC hardware platform for weed identification. AkidaNet is built with reference to thewell-known MobileNetV1 and VGG-11 CNN architectures. The full architecture of AkidaNet has been developed to enhance power efficiency and reduce computational latency as shown in Fig.3. Specifically, the model begins with four3Ɨ3regular convolutional layers (Conv2D), followed by eleven depthwise separable convolutional layers (Separable-Conv2D), and ends with a softmax activation with 4 outputs for crops/weeds classification. As can be observed in Fig.4,the differences among the standard convolution block, theMobileNetV1 DSC block, and the AkidaNet DSC block are illustrated in detail. The AkidaNet structure allows the NSoC platform to process the AkidaNet DSC blocks more efficiently with less memory. Furthermore, it is important to note that instead of using max pooling operations in this model, a stride of 2 is used for the specific layers including Conv2D 0, Conv2D 2, SeparableConv2D 4, Separable-Conv2D 6 and SeparableConv2D 12 to reduce the resolution of the outputs of all standard convolution and AkidaNet DSC blocks. In addition, all standard convolution and depthwise separable convolution operations have a kernel size of 3Ɨ3and zero padding. Finally, the Classification top layer is kept as a separable convolutional layer instead of a dense layer to reduce memory usage and allow AkidaNet’s architecture to utilize the NSoC hardware more efficiently. With regard to the AKD1000 NSoC hardware version, it currently supports most layers found in feed-forward network architectures such as Dense, Standard Convolutional, Depth-wise Separable Convolutional, Batch Normalization, MaxPooling Layers. AkikaNet’s optimized architecture came from a limited architecture search that measured accuracy, power, and latency of each model variation deployed on the NSoC. Additionally, we also investigated the impacts of max pooling and stride-2 convolution on reducing spatial dimensions. As a result, the stride-2 option gave better performance on power and accuracy. The AkidaNet model can be adjusted to further reduce computational cost by scaling the number of output channels in each layer with a hyper-parameter α, or the width multiplier, inspired by the original MobileNet paper [43]. Specifically, the range of αis from 0 to 1, with typical settings of 0.25, 0.5,0.75 and 1. For example, if α= 1, it will be the baseline AkidaNet. If α= 0.5, it will generate a model with only half the output channels used in each layer. Decreasing the output channels results in the decreasing of the number of model parameters, which reduces the final size of the model. Therefore, changing the width multiplier αin AkidaNet allows one to meet the resource constraints of the NSoC platform for real-time weed recognition in practice with a trade off , accuracy, and model size.



D. AkidaNet to Spiking Neural Network Conversion

The goal of the AkidaNet design is allow target applications to achieve low latency and power consumption when implemented on the AKD1000 NSoC hardware platform. To optimize the performance of this model on edge devices, a potential solution mentioned in this paper is the use of Spiking Neural Networks (SNNs). According to [50, 51, 52], before converting the architecture of AkidaNet to SNN for execution on the AKD1000 NSoC, the model needs to be quantized to use 4-bit activations and 4-bit parameters. The quantization process plays an indispensable role in significantly reducing model size and power consumption when compared with CPU or GPU implementations. After that, the quantized AkidaNet model is converted to an SNN in a process called CNN2SNNas shown in Fig.5. Particularly, all the communication between neurons in the SNN takes the form of ā€œspikesā€ or ā€œevents ā€œcorresponding to binary impulses that are generated when a neuron crosses a threshold level of activation. If all neurons on an Neural Processing Unit (NPU) do not cross the threshold ,it will generate no output. Hence, this feature is the key to the efficiency of SNNs and contributes to further minimizing computational cost.



E. Evaluation of AkidaNet and MobileNetV1 models onImageNet-1K dataset

We trained both MobileNetV1 and AkidaNet models from scratch on the ImageNet-1K classification dataset [54] using Tensorflow/Keras for 90 epochs on a single NVIDIA GeForceRTX 2080 Ti GPU with a batch size of 128 images. Weused Stochastic Gradient Descent (SGD) with momentumoptimizer and the learning rate was set to 0.1 for the first10 epochs and then decreased to 0.0001 with an exponentialdecay. We also used the L2 weight regularizer in Conv2Dkernels and in pointwise kernels of the SeparableConv2Dlayers. In addition, we applied basic data augmentation (i.e.,random resized cropping and horizontal flipping). We trainedat the resolution of 160x160 and evaluated at the resolutionof 224x224 based on [55]. Finally, the last step is Akidaconversion and evaluation. The model was quantized to 4bit weights and activations except for the first convolutionallayer’s weights, which were quantized to 8 bits. After that, themodel was tuned for an additional 10 epochs with an initiallearning rate of 0.0001 that was kept constant for 2 epochsand then decreased to 1e-8.



F. Evaluation of the AkidaNet model on the bccr-segsetcrop/weed image dataset

We performed extensive experiments to evaluate the performance of our proposed method and some baseline approaches. All experiments were executed on an Ubuntu machine with an Intel Core 3.7GHz i7 CPU, 32Gb RAM and a GeforceGTX 1080Ti GPU. Our network was implemented in Python using TensorFlow and the MetaTF ML framework [50]. We compared our model against the following models: ResNet-50,VGG-16, and InceptionV3. For all experiments, we used 80% (24,000 images) of the dataset for training the model and 20% (6,000 images) for testing. In order to have a fair comparison, all models were trained with an Adam optimizer, 10 epochs, a batch size 32, and a learning rate 0.0001. The input size of VGG-16, ResNet-50, AkidaNet α= 0.25, and AkidaNetα= 0.5was 224 Ɨ224, while the input size of InceptionV3was 299 Ɨ299. Moreover, All CNN models, including the AkidaNet variants, were pre-trained with the ImageNet-1kdataset and then trained on our bccr-segset dataset. In the next steps, we quantized the Keras AkidaNet model and then converted the resulting quantized model to an SNN for execution on the AKD1000 NSoC. Specifically, AkidaNet models with α= 0.25 and α= 0.5were quantized by using the cnn2snn.quantize function in MetaTF framework [50]. The first layer of weights were quantized to 8-bits while the remaining model weights and activations were quantized with 4-bits in order to meet the requirements of the NSoC hardware. After having obtained a quantized model with satisfactory performance, the Keras quantized model was converted into the SNN model by using the cnn2snn.convert function [50],which returns a model in an Akida-compatible format. This model file format can be run efficiently on the NSoC hardware device in inference mode.
 
  • Like
  • Love
  • Wow
Reactions: 36 users

cosors

šŸ‘€
Totally off topic.
Who opens the 3000th page with her/his post?


"Hello Fellow Chippers,

Thanks to ZeeBot for creating this amazing new forum so we can all have civilized thoughtful BrainChip discussions.

I'll get us started, 2022 has already exceeded my expectations & we are only in the first few days of February.

Can't wait to see what the year brings.

Love all you guys & Girls xxxx"
 
Last edited:
  • Like
  • Love
  • Fire
Reactions: 16 users

cosors

šŸ‘€
Is this finally heading in the direction of MosChip?

ASIC Verification Engineer​

BrainChip Hyderabad, Telangana, India

Responsibilities

- Test plan, Test bench development, execution, and debugging
- Block and Chip level- IP/ASIC/SOC/CPU/AMS Verification
- SystemVerilog with Testbench methodologies UVM
- Verilog, VHDL, C++, Vera, e, System C
- Protocols: PCIe / USB / DDR / Ethernet / ARM šŸ’Ŗ/CNN/RNN
- Scripting: Perl, Tcl, Unix sc
ripting

https://www.linkedin.com/jobs/view/...at-brainchip-3650966374/?originalSubdomain=in
 
Last edited:
  • Like
  • Fire
  • Love
Reactions: 18 users

Esq.111

Fascinatingly Intuitive.
  • Like
  • Thinking
  • Fire
Reactions: 14 users

cosors

šŸ‘€
@TechGirl where are you it's your turn! ;)
 
  • Love
  • Like
  • Fire
Reactions: 5 users

AARONASX

Holding onto what I've got
 
  • Like
  • Love
  • Fire
Reactions: 32 users

Tothemoon24

Top 20
Click on the above for the full read & a insightful interview

The Air Force Research Laboratory (AFRL) posted the new video covering aspects of the Autonomous Aircraft Experimentation (AAx) initiative on the Defense Visual Information Distribution Service(DVIDS) website earlier today. AAx's main focus is on testing and refining artificial intelligence and machine learning-driven autonomous capabilities for use on future advanced uncrewed aircraft and helping to move those technologies out of the laboratory and onto actual operational platforms.

We are trying to figure out how to integrate artificially trained neural networks, trained in a simulation... into the real world," Bill "Evil" Gray, the chief test pilot at the Air Force's Test Pilot School, explains in the newly released video. "In this case [through AAx], integrate them into controlling an airplane."
"We need to recognize that AI [artificial intelligence] is here. It's here to stay. It's a powerful tool," Air Force Col. Tucker "Cinco" Hamilton, the service's chief of AI Test and Operations, says at another point in the footage. "Collaborative Combat Aircraft and that type of autonomy is revolutionary. And will be the future battle space."
 
  • Like
  • Fire
Reactions: 16 users

manny100

Top 20
  • Like
  • Love
Reactions: 13 users

Draed

Regular
3000? Just taking a pot shot at it....🤣
 
  • Haha
  • Like
Reactions: 3 users
Top Bottom