Is this real or somebody's BSI am not fluent in English but….. could it be that there are several grammatical errors on that flyer
„mims nimos“, „ayschonnotus“ ?? “mimics … asynchronous“.??
„instany / localy locally / rel-tiime“ … „instantly / locally / real time“.??
„devices that smart“ … „devices that are smart“.
„do’t“ … „don’t“??
View attachment 92668
| Aspect | Akida with On‑Chip Learning | Chips without On‑Chip Learning | 
|---|---|---|
| Adaptability | Can learn new patterns locally (e.g., new anomalies, new commands) and immediately use them in communication. | Limited to pre‑trained models; cannot adapt on their own. Must be retrained in the cloud or offline and redeployed. | 
| Communication Content | Sends newly learned events or adapted patterns to other devices (e.g., “new anomaly detected”). | Sends only pre‑defined outputs from its fixed model (e.g., “class = 2”). | 
| Collaboration | Devices can share discoveries: one node learns something new and alerts others, enabling system‑wide adaptation. | Devices can only share fixed detections; if one encounters a new pattern, others won’t recognize it until all are updated externally. | 
| Autonomy | High — devices can self‑improve and coordinate without cloud or human intervention. | Lower — devices depend on external retraining and updates to handle new situations. | 
| Resilience | Ideal for remote or disconnected environments (spacecraft, autonomous vehicles, industrial plants) where cloud updates aren’t possible. | Vulnerable in disconnected environments — performance degrades if new patterns appear that weren’t in the original training. | 
| Efficiency of Communication | Communicates only meaningful, newly learned events, reducing bandwidth. | Communicates fixed outputs, which may include irrelevant or incomplete information if the model is outdated. | 
| System Intelligence | Creates a distributed, adaptive network — like a “nervous system” that learns and shares in real time. | Creates a static network — devices can coordinate, but only within the limits of their pre‑trained knowledge. | 
						
					
					thinklucid.com
				“Sony invested in Brainchip”
Here might be why.
"
Sony IMX501: Vision and Intelligence in One Package
Parallel DSP Cores for Efficient AI Inference
The main engine for AI processing is the DSP subsystem core. This core is comprised of a High-Computation-Intensity (CI) DSP core, Standard-CI DSP core, and Tensor DMAs (TDMAs). The High-CI and Standard-CI DSPs work in parallel, executing neural network operations and moving the data directly into TDMA and then into L2 memory. Image and inference data are then transferred via the MIPI interface to the camera’s FPGA."
ie Sony cameras also comes with an FPGA chip, which can run akida2.
While akida may not yet be in the sensor chip ( a la Prophesee) we might be next to it, performing deeper inferencing on the entire image
Sony IMX501 AI Sensor and Brain Builder for Edge AI Processing
AI at the edge: LUCID’s Triton Smart camera uses Sony’s IMX501 for efficient on-camera inference for smart manufacturing, logistics, and embedded vision systemsthinklucid.com
Sure. There will be inferencing on two levels. On the pixels (can you make the image sharper) and on the image on the FPGA chip ( is it a duck?) and akida could come in the second category is my point.Sony's IMX501 intelligent vision sensor contains its own proprietary parallel DSP cores for efficient AI inference and on-sensor processing, designed in-house, and uses a unique architecture to allow AI tasks to be run at the edge without external processors or cloud dependency.
Source: https://www.provideocoalition.com/sony-reveals-the-worlds-first-intelligent-vision-sensor-with-ai/#:~:text=Sony’s original DSP(Digital Signal Processor) dedicated to AI signal processing
Sure. There will be inferencing on two levels. On the pixels (can you make the image sharper) and on the image on the FPGA chip ( is it a duck?) and akida could come in the second category is my point.
I only say this because the sony adverts show a lot of image composition boxes..
" "Developing advanced technologies that can be easily adapted, implemented, and integrated into both existing and future systems is key to delivering reliable and resilient communications capabilities to the warfighter."
McLaina Mazzone, Science & Technology Assistant Program Manager, PEO C4I, PMW/A 170"
The above from the Navy Transition success story.
The point is that Bascom Hunter can integrate its card with AKIDA once testing is finalised straight into existing equipment.
Parsons should be able to do the same so our wait once the chips have been delivered may not be that long if they initially for drones.
Manned craft will be longer of course.
Who amongst us doesn't see us succeeding?
That's the question.
This sector isn't all about us and our technology, I understand that, for us to even think of dictating how the edge ai market will play out over the next 3 years plus would be a totally arrogant view to hold, one which our company, I'd suggest, wouldn't even contemplate.
Unless we establish ourselves in this market on our on two feet with real growing revenue streams, we are definitely a takeover target by a small number of contenders, one's who we are all aware of..keep an open mind, this share market bullshit will come to an abrupt end soon enough, let's go Sean, it's time to giddy up!![]()