uiux
Regular
That's like search for "coca cola" and being amazed coca cola is the first result LOL
MetaTF is a brand
Ok I have had a word ... we are now on the way up there!Can someone rally us to $1.20 today please.
Thanks in advance
Oh dear. I'm learning everyday.That's like search for "coca cola" and being amazed coca cola is the first result LOL
MetaTF is a brand
And the brand awareness is getting out thereThat's like search for "coca cola" and being amazed coca cola is the first result LOL
MetaTF is a brand
Big buy just now wiped out $1.16Approximately 0.1% of shares on issue have been traded/churned this morning on ComSec trading platform.
Tightly held is an understatement. Imagine when the big funds really want to buy in.
FONLAB..... Fear of no lubricant at barrelGood Morning Chippers,
Gross shorts for yesterday, 4th August 2022,
726,880 shares.
Getting that feeling again , possibly a decent rise this afternoon.
New shorting terminology..
FOBBOBL :- fear of being bent over by lender.
Or
FOROOC :- fear of running out of collateral.
.
Regards,
Esq.
Fingers crossed someone swipes out the lines from $1.165 to $1.80
That the easiest million $ I’ve ever spentDONE !!!!![]()
My $35k, 2022 Kia has a good level of automated driving capabilities. I absolutely love the lane-following feature that requires no steering input from the driver, combined with radar to automatically stay a safe distance away from the vehicle in front, and adaptive cruise control (all the way down to a standstill if necessary) . I do however dislike the timed warning for me to keep my hands on the wheel. Under expressway conditions it could easily do this all the steering by itself. The warning for me to keep my eyes open, now that’s one that I do agree with. Just joking! I never close my eyes while driving.Leveling up
With Drive Pilot, Mercedes has placed itself in an unusual position. The SAE autonomy scale goes from Level 0, signifying complete manual operation, to Level 5, signifying cars that can drive themselves in all conditions. Level 3 is the halfway point between those two extremes but, crucially, it’s also the only level that blends human and machine control.
According to the SAE, when a Level 3 system is operating, the car is driving. However, it can still ask the driver to take over control. That stipulation is unique to Level 3. At Levels 4 and 5, the car shouldn’t require human intervention, per the SAE standards, while at Levels 0-2 the driver is considered to be in charge at all times. That includes systems like General Motors’ Super Cruise and Ford’s BlueCruise, where drivers may be merely supervising without touching the steering wheel or pedals.
Another luxury?
The need to add more sensors and connectivity also contradicts a popular misconception: that Level 3 tech inevitably leads to fully autonomous driving. Just as new hardware was needed to get from Level 2 to Level 3, even more will likely be needed to reach Levels 4 and 5. That will also require mapping more roads, and taking Drive Pilot beyond divided highways. That adds development costs and pushes the limits of the technology, which may not work as well beyond the relatively controlled environment of highways.
![]()
We tested the self-driving Mercedes tech so advanced, it’s not allowed in the U.S.
Mercedes-Benz Drive Pilot reaches a new level of sophistication for driver-assist systems. We hit the track to see what it can do.www.digitaltrends.com
Hi Sirod,I've been asked to post the following on our forum, I've put it all through the translator:
Where could the Akida chip technology be installed soon????
![]()
BrainChip Partners with Prophesee
BrainChip Partners with Prophesee Optimizing Computer Vision AI Performance and Efficiencywww.prophesee.ai
BrainChip partners with Prophesee to optimize the performance and efficiency of Computer Vision AI
Laguna Hills, California - June 14, 2022 - BrainChip Holdings Ltd (ASX: BRN, OTCQX: BRCHF, ADR: BCHPY), the world's first commercial neuromorphic AI IP manufacturer, and Prophesee , the inventor of the world's most advanced neuromorphic vision systems, today announced a technology partnership that will provide next-generation platforms for OEMs looking to integrate event-based vision systems with high AI performance coupled with ultra-low-power technologies.
Inspired by human vision, Prophesee's technology uses a patented sensor design and AI algorithms that mimic the eye and brain to reveal what was previously invisible using standard frame-based technology. Prophesee's computer vision systems unlock new potential in areas such as autonomous vehicles, industrial automation, IoT, security and surveillance, and AR/VR.
Akida, BrainChip's first-to-market neuromorphic processor, mimics the human brain to analyze only the essential sensory inputs at the point of acquisition and process data with unprecedented efficiency, precision and power savings. Keeping AI/ML local to the chip, independent of the cloud, also drastically reduces latency.
"We successfully ported data from Prophesee's neuromorphic camera sensor to process inference on Akida with impressive performance," said Anil Mankar, co-founder and CDO of BrainChip. "This combination of intelligent vision sensors with Akida's ability to process data with unparalleled efficiency, precision and energy savings at the point of capture truly advances cutting-edge AI enablement, providing manufacturers with an out-of-the-box solution."
CAUTION!! Absolutely the most informative section of the report:
“By combining our Metavision solution with Akida-based IP, we are better able to provide a complete high-performance, ultra-low-power solution to OEMs looking to leverage edge-based visual technologies as part of their product offering,” said Luca Verre, CEO and co-founder of Prophesee.”
For more information on the BrainChip and Prophete partnership, contact sales@brainchip.com.
"We successfully ported data from Prophesee's neuromorphic camera sensor to process inference on Akida with impressive performance," said Anil Mankar, BrainChip co-founder and CDO. "This combination of intelligent vision sensors with Akida's ability to process data with unparalleled efficiency, precision and energy savings at the point of capture truly advances cutting-edge AI enablement, providing manufacturers with an out-of-the-box solution."
“By combining our Metavision solution with Akida-based IP, we are better able to provide a complete high-performance, ultra-low-power solution to OEMs looking to leverage edge-based visual technologies as part of their product offering,” said Luca Verre, CEO and co-founder of Prophesee."
Remember the keyword Metavision!!
![]()
Xperi Develops World-first Neuromorphic In-cabin Monitoring Technologies
Xperi develops world-first neuromorphic driver monitoring solution (DMS), powered by Prophesee Metavision® - Event-Based Vision sensor.www.prophesee.ai
ABOUT PROPHESEE
Prophesee is the inventor of the world's most advanced neuromorphic vision systems.
The company has pioneered an event-based vision approach to machine vision. This new category of vision enables a significant reduction in power, latency and data processing requirements to reveal what was previously invisible to traditional frame-based sensors.
Prophesee's patented Metavision® sensors and algorithms mimic how the human eye and brain work to dramatically improve efficiency in areas such as autonomous vehicles, industrial automation, IoT, security and surveillance, and AR/VR.
![]()
Xperi Develops World-first Neuromorphic In-cabin Monitoring Technologies
Xperi develops world-first neuromorphic driver monitoring solution (DMS), powered by Prophesee Metavision® - Event-Based Vision sensor.www.prophesee.ai
New sensor technologies complement the capabilities of standard vision systems, improve occupant safety, enable greater personalization and expand today's cabin experience into tomorrow's third room experience
CALABASAS, California – April 21, 2021 – DTS® , a global leader in next-generation audio, imaging and sensing technology and a wholly owned subsidiary of Xperi Holding Corporation (NASDAQ: XPER) (“Xperi”), today announced a global first neuromorphic driver monitoring solution (DMS) powered by Prophesee Metavision® – Event-Based Vision Sensor .
Using the raw data from the Metavision® – Event-Based Vision sensor, the DTS AutoSense* team developed key driver monitoring functions such as gaze tracking, head posture, identification and eyelid opening.
With more than 20 years of world-leading experience in neural solutions and computer vision, and with billions of products powered by their solutions, the Xperi team maps and adapts interior surveillance (ICM) technologies to existing and future sensor-centric trends .
“Intelligent sensing is required for the next-generation cabin experience. The key way to get there is an advanced research strategy focused on sensor fusion,” said Petronel Bigioi, Chief Technology Officer, Xperi. "DTS AutoSense's advanced research team is focused on cutting-edge sensing solutions, and neuromorphic sensing is one of the technologies that can shape the future of the industry."
The neuromorphic sensors capture information at an equivalent frame rate of 10,000 frames per second without the need for active lighting. This enables better low-light performance for driver monitoring functions, as well as unprecedented features such as saccadic eye movements or micro-expressions, critical next steps for next-generation passenger experiences.
Xperi's implementation has 12 patents pending and relies on an end-to-end data generation and training system designed to take into account the specifics of the sensor. The training dataset was generated based on Xperi's extensive computer vision infrastructure, reusing ground truths from the visible and near-infrared spectrum, synthesizing a completely novel approach in neuromorphic recognition.
"Being able to instantly recognize the most subtle, almost imperceptible facial and eye movements can be life-saving," Petronel Bigioi said. "The powerful low-light imaging capabilities, fast response times, and low-power operation that enable this groundbreaking neuromorphism-based technology application make it ideal for the safety-critical applications of driver monitoring systems, while expanding the possibilities for further enhancement and personalization of the vehicle cabin."
Xperi uses Prohesee's Metavision products.
Take a look at the customer list at Xperi, all the well-known car manufacturers in the world are there!
![]()
Connected Car - Xperi
Xperi is transforming the automotive experience by bringing high-quality multimedia and personalization to the connected car, just like at home.xperi.com
What do you think?
I live outside a major metropolis, Ubers and taxis are bloody difficult to find at the best of times. Will this be better or worse with fully autonomous Uber? I expect they’ll be restricted to well-defined roads in the larger cities and don’t expect them to reach most parts of Australia. Not even outer suburbia.People will still want to own one instead of waiting for 1 to turn up or go on a camping trip etc
It's good to get a user perspective.My $35k, 2022 Kia has a good level of automated driving capabilities. I absolutely love the lane-following feature that requires no steering input from the driver, combined with radar to automatically stay a safe distance away from the vehicle in front, and adaptive cruise control (all the way down to a standstill if necessary) . I do however dislike the timed warning for me to keep my hands on the wheel. Under expressway conditions it could easily do this all the steering by itself. The warning for me to keep my eyes open, now that’s one that I do agree with. Just joking! I never close my eyes while driving.
I am not one who wants the car to always drive itself under all conditions, just on longer trips on well-marked roads would suit me. So I’d love the car manufacturers to simplify it to staying within the lanes on a well marked highway. All the other stuff seems an over-complication to me. In fact all the other stuff may cause problems when people neglect to take over manual control when they need to, or worse still, lose crucial driving abilities and can’t take over manual control.
Quick chart to try consolidate as much info in simple visual.
Have highlighted other possible areas of supp / resist related to previous SP action.
1. Summary table of various oscillators, pivots & MA's
2. Red / Green auto Fib based on a qtrly function. Lines change colour as SP moves below or above to become supp / resist.
3. Tries to visualise supply / demand / sentiment. Can see tightened recently as bull / bears battle.
4. RSI bias paints the SP bars. Want the area to break upwards (green) above the centre channel. If sits in centre channel can indicate poss ranging zone.
Not seeing a lot of vol support for the upward momentum though. Not a lot of supply coming in so buyers just nibbling away at what's available at the moment. Could change quickly imo if we get to a tipping point where buyers make a rush and manage to swamp / swallow any supply that does show up in the resist area heading into now.
View attachment 13396