“Entering the Era of Multimodal Perception,” a Presentation from Connected Vision Advisors

embeddedvision 32 views 17 slides Sep 16, 2024
Slide 1
Slide 1 of 17
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17

About This Presentation

For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/09/entering-the-era-of-multimodal-perception-a-presentation-from-connected-vision-advisors/

Simon Morris, Serial Tech Entrepreneur and Start-Up Advisor at Connected Vision Advisors, presents the “Entering t...


Slide Content

Entering the Era of
Multimodal Perception
Simon C. Morris
Serial Tech Entrepreneur and Start-Up Advisor
Connected Vision Advisors

Generative AI –All the Buzz
2© 2024 Connected Vision Advisors
Forbes 2Oct 2023: The 10 Biggest Generative AI Trends For 2024
Generate content, like:
1.Writing articles
2.Composing music
3.Designing graphics
4.Generating code
5.Etc…

There’s Another Opportunity
3© 2024 Connected Vision Advisors
Reasoning is perhaps the
ultimate goal of AI, but
with human evolution it
was both multisensory
integration and
reasoning that lead to
life saving decisions.
Obvious examples:
•Sight + sound
•Smell + taste

Multisensory –Better Decisions –Safer
4© 2024 Connected Vision Advisors
This Photoby Unknown Author is licensed under CC BY-SA

•Sensor fusion -the process of combining sensor data or data derived
from disparate sources
•The result -less uncertainty or a better decision compared to when the
sensor sources were used individually.
•Edge perception today -combination of camera, lidar, time of flight
(ToF), radar, ultrasonic, GPS sensor data
Multisensory Integration –Fusion on the Edge
5© 2024 Connected Vision Advisors

In the last three years
alone, there have been
over 1.7 million patents
filed and granted in radar-
camera fusion
Proof of the Importance of Fusion Opportunity
6© 2024 Connected Vision Advisors
Source: GlobalData’sreport on Artificial Intelligence in Automotive: Radar Camera Fusion
0
2
4
6
8
10
Range x FOV
Range Resolution
Elevation Resolution
Angular Resolution
Velocity
Discrimination
Processing
Overhead
Device Cost
Low Light
Performance
Dust/Fog/Snow
Robustness to
interference
Colour Detection
Object Classification
Object
Segmentation
Object Tracking
VisionRadar Lidar

Fusion Patent Volume 2020-2022 –Automotive
7© 2024 Connected Vision Advisors
Bubble size = patent
volumes 2020-2022
Source: GlobalDataPatent Analytics

Fusion for Edge Perception
8© 2024 Connected Vision Advisors
Camera Radar Fusion -YouTube
Ref: JunhyukHyun, Yonsei University
Korea, Computational Intelligence Lab
Camera Ready Fusion

Fusion for Edge Perception
9© 2024 Connected Vision Advisors
High-level fusion (decision fusion) -incumbent
• GPS + inertial navigation systems, or
• Fusing objects and trajectories of objects

Low-level fusion (early fusion of the raw data)
• Fuse lidar point cloud data + RGB camera data train AI model
• Fuse radar RAD data + RGB camera data train AI model
Fusion for Edge Perception –Where’s It Going?
10© 2024 Connected Vision Advisors
IEEE Transactions on Intelligent Vehicles ( Volume: 9, Issue: 1, January 2024)
Mid-level fusion (fusing features)
• Radar or lidar point cloud + vision object detection
Latest research shows that AI models trained on early fusion data will provide better detection

1.Huge raw data bandwidth (raw radar data 30 Gbps +) to process –needs
very high AI performance, but low cost, power effective for edge
•Incumbents exceeding 20 TOPs: e.g. QCOM, TI, Ambarella, Renesas
•Challenging start-ups exceeding 20 TOPs: e.g. Hailo, Blaize, Brainchip, EdgeCortix,
untether, DeepX,
2.Lack of early-fusion training data to train AI models and non-public
Challenges to Adoption of Mid & Low-Level Fusion
11© 2024 Connected Vision Advisors

3.Need for a perception processing framework for calibrating, synchronizing
different data sources
4.An easier (lower cost, time) means of collecting calibrated data sources in
the field
5.An easier (lower cost, time) means to annotate fused data, auditing
annotation and train an edge AI model based on this data
Challenges to Adoption to Mid & Low-Level Fusion
12© 2024 Connected Vision Advisors

Anywhere where multi-mode perception sensing is required such as:
•Automotive (ADAS to AD, driver monitoring)
•Off-road (agriculture, construction, mining, material movement,
defense)
•Factory, warehouse mobile robotics
•Drones (autonomy without pilot or GNSS)
•Security/surveillance
•Etc…
The Untapped Opportunity
13© 2024 Connected Vision Advisors

Semiconductor
•Incumbents exceeding 20 TOPs: e.g.
•Challenger start-ups exceeding 20 TOPs: e.g. Tens Torrents, SiMa.ai, Rebellions,
Picks & Shovels: Positioned for Edge Fusion Perception
14© 2024 Connected Vision Advisors

Edge AI Fusion Perception Frameworks -very few
Picks & Shovels: Positioned for Edge Fusion Perception
15© 2024 Connected Vision Advisors
AI Models Optimized for the Edge -focus primarily on vision only, not fusion
Data Annotation Tools -focus primarily on image and video annotation

1.Perception based on data fusion yields superior perception
2.Challenges of “early fusion” are being overcome
3.Fusion will move this decade from incumbent “late fusion” designs to
“early fusion” implementations
4.Opportunities exist to differentiate throughout the value-chain to
address the demands of early fusion
Take Aways
16© 2024 Connected Vision Advisors

References/Resources
17© 2024 Connected Vision Advisors
GlobalDataPatent Analytics: Key players in radar camera fusion
JunhyukHyun, Yonsei University Korea, Computational Intelligence Lab
GlobalData’sreport onArtificial Intelligence in Automotive: Radar Camera Fusion
Forbes 2Oct 2023: The 10 Biggest Generative AI Trends For 2024
IEEE Transactions on Intelligent Vehicles ( Volume: 9, Issue: 1, January 2024) : Radar-Camera Fusion for
Object Detection & Semantic Segmentation in Autonomous Driving: A Comprehensive Review