D
Deleted member 3781
Guest
Not even 20
I see the big news coming from tomorrow any time now
| Aspect | Akida with OnâChip Learning | Chips without OnâChip Learning |
|---|---|---|
| Adaptability | Can learn new patterns locally (e.g., new anomalies, new commands) and immediately use them in communication. | Limited to preâtrained models; cannot adapt on their own. Must be retrained in the cloud or offline and redeployed. |
| Communication Content | Sends newly learned events or adapted patterns to other devices (e.g., ânew anomaly detectedâ). | Sends only preâdefined outputs from its fixed model (e.g., âclass = 2â). |
| Collaboration | Devices can share discoveries: one node learns something new and alerts others, enabling systemâwide adaptation. | Devices can only share fixed detections; if one encounters a new pattern, others wonât recognize it until all are updated externally. |
| Autonomy | High â devices can selfâimprove and coordinate without cloud or human intervention. | Lower â devices depend on external retraining and updates to handle new situations. |
| Resilience | Ideal for remote or disconnected environments (spacecraft, autonomous vehicles, industrial plants) where cloud updates arenât possible. | Vulnerable in disconnected environments â performance degrades if new patterns appear that werenât in the original training. |
| Efficiency of Communication | Communicates only meaningful, newly learned events, reducing bandwidth. | Communicates fixed outputs, which may include irrelevant or incomplete information if the model is outdated. |
| System Intelligence | Creates a distributed, adaptive network â like a ânervous systemâ that learns and shares in real time. | Creates a static network â devices can coordinate, but only within the limits of their preâtrained knowledge. |
thinklucid.com
âSony invested in Brainchipâ
Here might be why.
"
Sony IMX501: Vision and Intelligence in One Package
Parallel DSP Cores for Efficient AI Inference
The main engine for AI processing is the DSP subsystem core. This core is comprised of a High-Computation-Intensity (CI) DSP core, Standard-CI DSP core, and Tensor DMAs (TDMAs). The High-CI and Standard-CI DSPs work in parallel, executing neural network operations and moving the data directly into TDMA and then into L2 memory. Image and inference data are then transferred via the MIPI interface to the cameraâs FPGA."
ie Sony cameras also comes with an FPGA chip, which can run akida2.
While akida may not yet be in the sensor chip ( a la Prophesee) we might be next to it, performing deeper inferencing on the entire image
Sony IMX501 AI Sensor and Brain Builder for Edge AI Processing
AI at the edge: LUCIDâs Triton Smart camera uses Sonyâs IMX501 for efficient on-camera inference for smart manufacturing, logistics, and embedded vision systemsthinklucid.com
Sure. There will be inferencing on two levels. On the pixels (can you make the image sharper) and on the image on the FPGA chip ( is it a duck?) and akida could come in the second category is my point.Sony's IMX501 intelligent vision sensor contains its own proprietary parallel DSP cores for efficient AI inference and on-sensor processing, designed in-house, and uses a unique architecture to allow AI tasks to be run at the edge without external processors or cloud dependency.
Source: https://www.provideocoalition.com/sony-reveals-the-worlds-first-intelligent-vision-sensor-with-ai/#:~:text=Sonyâs original DSP(Digital Signal Processor) dedicated to AI signal processing
Sure. There will be inferencing on two levels. On the pixels (can you make the image sharper) and on the image on the FPGA chip ( is it a duck?) and akida could come in the second category is my point.
I only say this because the sony adverts show a lot of image composition boxes..
" "Developing advanced technologies that can be easily adapted, implemented, and integrated into both existing and future systems is key to delivering reliable and resilient communications capabilities to the warfighter."
McLaina Mazzone, Science & Technology Assistant Program Manager, PEO C4I, PMW/A 170"
The above from the Navy Transition success story.
The point is that Bascom Hunter can integrate its card with AKIDA once testing is finalised straight into existing equipment.
Parsons should be able to do the same so our wait once the chips have been delivered may not be that long if they initially for drones.
Manned craft will be longer of course.
Who amongst us doesn't see us succeeding?
That's the question.
This sector isn't all about us and our technology, I understand that, for us to even think of dictating how the edge ai market will play out over the next 3 years plus would be a totally arrogant view to hold, one which our company, I'd suggest, wouldn't even contemplate.
Unless we establish ourselves in this market on our on two feet with real growing revenue streams, we are definitely a takeover target by a small number of contenders, one's who we are all aware of..keep an open mind, this share market bullshit will come to an abrupt end soon enough, let's go Sean, it's time to giddy up!![]()
investor.brainchip.com
That is impressive performance. No wonder the potential and current applications are varied.![]()
BrainChip Unveils Breakthrough AKD1500 Edge AI Co-Processor at Embedded World North America - BrainChip Investor Portal
BrainChip Unveils Breakthrough AKD1500 Edge AI Co-Processor at Embedded World North America Laguna Hills, Calif. â November 4th, 2025 â BrainChip Holdings Ltd (ASX: BRN, OTCQX: BRCHF, ADR: BCHPY), a global leader in ultra-low power, fully digital, event-based neuromorphic AI, today announced...investor.brainchip.com
"...800 giga operations per second (GOPS) while operating under 300 milliwattsâsetting a new benchmark for edge AI efficiency."
That sounds incredible massiv power to me.....
View attachment 92755