Hopefull final episode.



E. Data Preprocessing
Fig. 5. Annotation images of dataset.
In this method, we start by deleting duplicate images from
datasets and filtering out datasets that don’t fit our criteria. By
removing irrelevant information, we focus on the photographs
that are most important to our project. The next key step is
annotating these images, which is essential for getting the
model ready for training. Each image is given the proper
tags or categories during annotation, which helps to properly
organize the data. Additionally, the machine learning system
can recognize patterns in the photos thanks to this annotation
process, which enables it to produce precise predictions. Next,
the model set the parameter of images to 96 X 96 size. We may
go on to the following step, which entails training the model
after the photographs have been annotated. In this stage, the
model is taught how to recognize and understand numerous
patterns and characteristics inside the photos using the anno-
tated data. The model continually improves its comprehension
and grows better at generating precise predictions through
a series of iterations and modifications. Generally speaking,
the process starts with data filtering to get rid of unwanted
datasets and duplicates. After that, each image is given the
proper labels via the annotation process, which facilitates data
organization and makes it possible for the machine learning
system to recognize patterns. In the end, the model is trained
using these annotated images, honing its forecasting skills and
deepening its comprehension of the visual information.
F. Implementation
We delivered the dataset to the three modules for training
after Data Preprocessing. They are further discussed below:
1) Creating Impulse: We will start the essential procedures
at this point in order to train our model successfully. To begin
with, we performed a few configuration changes, including
translating the entire dataset into dimensions of 96X96. We
used a resize setting that allowed for the shortest axis of the
photos to ensure compatibility. Then, we used the BrainChip
Akida model to do feature extraction from the dataset’s
images. We were able to draw out important details and useful
information from each image using this procedure. It is crucial
to remember that we will preserve and use each and every
one of these extracted characteristics to train our model in the
future. By following these steps, we are setting up an efficient
training procedure that will allow our model to develop and
produce precise predictions based on the supplied information.
2) Image: The color depth parameters are converted into
the Red-Green-Blue (RGB) format in the next stage, enabling
additional analysis and modification. We then started the
feature extraction procedure, successfully identifying all the
crucial traits that were present in the dataset. We created
a graph that successfully demonstrates the properties of the
dataset to display this data visually. This graph is shown in
Fig. 6, which offers a clear visual depiction of the retrieved
characteristics.
Fig. 6. Feature explorer.
3) Object Detection: In this phase, we set up the parameters
that will be used to train our model. We kept the validation
set size at 20%, the learning rate at 0.001, and the number
of epochs at 100. In addition, we trained our dataset using
the Akida Fomo model. We started the model training process
after making these adjustments, and as a consequence, we got
a remarkable training accuracy of 95.0%. Fig. 7 shows the
precise measurements and results of this training procedure.
In addition, we created a quantized version of the model in
addition to the outcomes. This quantization ensures that the
model performs in real-time without any hiccups even on
devices with modest Random-Access Memory (RAM) and
storage capabilities. The quantized model maintains its ability
to carry out the required activities while greatly reducing
the memory and storage needs. This development makes it
possible to install and use the model effectively in contexts
with limited resources.
4) Classification: In this part, we evaluated the model using
photos from the dataset and were successful in obtaining a
testing accuracy of 90%, which we believe can improve in the
future with a larger training dataset set so that the model may
be fine-tuned and have additional features for training. Fig. 8
depicts the testing accuracy. Overall, we are satisfied with the
performance of the model and optimistic about its potential for
further improvement. It is clear that the team has put in a lot
of effort into developing and evaluating the model. The testing
accuracy of 90% is impressive, and the team’s optimism about
Fig. 7. Model Training Accuracy.
the model’s potential for further improvement is encouraging.
It will be interesting to see how the model performs with a
larger training dataset and additional features for training.
Fig. 8. Model Testing Accuracy.
Fig. 9. Model output.
IV. CONCLUSION
In conclusion, our study made use of a sizable dataset
made up of a variety of occurrences, such as chain stealing
and wallet stealing. The videos in the collection were first
broken down into individual frames, and then redundant and
duplicate pictures were eliminated. After that, we worked on
training our model and annotating the pictures. Our model,
known as Akida FOMO, showed excellent accuracy of 95.0%
throughout the training phase and testing accuracy of 90%
after the training phase was complete. Therefore, we can state
with confidence that the use of the Edge Impulse platform and
the BrainChip Akida FOMO model contributed significantly to
the production of insightful findings for our research. The high
levels of accuracy reached by our model demonstrate that the
incorporation of Akida FOMO into our study was a resounding
success. Although setting up the dataset and annotating the
photos took a lot of time and work, the outcomes were
unquestionably trustworthy and the effort was justified. We are
adamant that future research into computer vision and object
identification has a great deal of potential when the cutting-
edge technology of BrainChip is paired with the Edge Impulse
platform. The positive results of our research demonstrate the
strong combination’s prospective trajectory and demonstrate
its potential to open up new directions for future developments
in this area.
V. FUTURE DIRECTION
There are a number of fascinating future areas that may
be investigated in the field of computer vision and object
identification in order to build upon the accomplishments and
beneficial results of our study using the Edge Impulse platform
and the Akida FOMO model.
Improved Model Performance: Although our existing model
achieved outstanding accuracy rates, there is still potential
for improvement. The model parameters may be adjusted in
future study, along with enhanced training methods, more
datasets, and a wider variety of training examples. The model
can push the limits of accuracy even further by continu-
ously improving the model’s performance. Real-time Object
Recognition: The Edge Impulse platform opens up oppor-
tunities for real-time object recognition applications when
combined with Akida FOMO’s capabilities. The development
of a system that can recognize and detect chain-snatching
or wallet-snatching instances in real-time may result from
broadening our study, which might have a substantial impact
on public safety and crime prevention initiatives. Generali-
sation and Transfer Learning: Exploring the possibilities of
transfer learning strategies might be an attractive area for
future study. We may be able to attain greater accuracy rates
and speed up the training process by utilizing pre-trained
models on related object identification tasks and fine-tuning
them with our particular dataset. The model’s usefulness
and applicability can also be increased by investigating its
capacity to generalize across various scenarios and contexts.
Scalability and Deployment: As our study develops, the Akida
FOMO model’s scalability and deployment need to be taken
into account. The model’s deployment on resource-constrained
edge devices, such as security cameras or smartphones, may be
facilitated by optimizing the model’s architecture and training
procedure to decrease computing needs and memory footprint.
This would make it possible to use our research’s results in the
real world, broadening its influence beyond the boundaries of a
sterile laboratory. Expansion to Other Applications: Although
the majority of our study was devoted to chain-snatching and
wallet-snatching occurrences, the approaches and techniques
created may be applied to a number of other fields and appli-
cations. Exploring the model’s capability to identify and detect
various objects or events, such as pedestrian identification or
traffic sign recognition, can advance computer vision in more
general ways and improve safety protocols in a variety of
scenarios.
Finally, the effective integration of Akida FOMO into the
Edge Impulse platform opens the door for interesting new
research trajectories. We may improve the area of computer
vision and object identification by continually improving
the model, investigating real-time applications, using transfer
learning, assuring scalability, and extending to new domains,
eventually helping society with increased safety and efficiency.