advanced vehicle perception system

1
We acknowledge the support of the Natural Sciences and Engineering Council of Canada (NSERC), which invests annually over $1 billion in people, discovery and innovation. EECOMOBILITY (ORF) & HEVPD&D CREATE We acknowledge the support of the Ontario Research Fund: Research Excellence Program. VEHICLE PERCEPTION SYSTEM SYSTEM ARCHITECTURE SENSORS FUSION DISTRIBUTED SYSTEM HETEROGENEOUS CONVOLUTIONAL NEURAL NETWORK Advanced Vehicle Perception System Centre for Mechatronics and Hybrid Technology Department of Computing and Software Engineering, McMaster University Ash (Chang) Liu Just like our human needs to see, smell, hear, and feel to gather knowledges about the surrounding before we make decisions; an autonomous drive vehicle needs sensors to collect information from the surrounding, then extract information for the autonomous drive system to make any reasonable driving decision. Centralized system often faces the issue with scalability: the more sensors and modules you attached to the system, the longer response time it requires for the system to response, thus this directly reduce the sampling rate for the system and downgrade the overall system performance. To solve this issue, we used a decentralize design for our system, and developed our own version of distributed system package to support this architecture. Our software package is a C++ library package based on TCP/IP network interface to communicate between modules. Advantages: - Hot plug module while system is running - Multi-platform/ multi-OS support - Vanilla.(No additional 3 rd party library dependency) - Simple API - Real Time Support System schematic diagram for a single module(core) Overall architecture of Heterogeneous Convolutional Neural Network Heterogeneous Convolutional Neural Network (HCNN) is a Deep Neural Network architecture we developed for vehicle classification and recognition. This architecture combines feature extraction with CNN to creates a robust neural network array. From our testing with collected road data, this neural network delivers superior performance at 94% precision. Object detection with LiDAR and Camera - Using modified classic blackboard architecture for decision making. - A modulate design to allows us split works between different researchers, and running different module on different platforms. (Distributed System) - System contains 4 layers: Input layer (I/O Cores & Sync Core), Fusion layer (Fusion Cores), Logic layer (Neural Network Cores), and Decision Layer (Voter). - Input layer takes input from different sensor, and sync core synchronize the input. - Fusion layer combines data retrieved from input layer to create a mix of unclassified objects, and assigning extracted information to the object. (Sensor fusion) - Logic layer will take information from the fusion layer to perform classification and recognition. - Decision layer took results different logic units (Neural Network Core) and determine what is the most accurate description of classified object. An abstract diagram of the system architecture Radar LiDAR CAM IR CAM Fusion Sensor Fusion Extract unclassified objects in the scene. Finding different properties of the object. Homogeneous Fusion: Position Shape Reflectivity Temperature ... Heterogeneous fusion: Velocity Acceleration Morphing ... Sensors Input Interpretation Driving Decision Vehicle Control Taking raw information from the sensor is not a trivial problem: different type of sensors have different properties, and in some cases sensors can produce false information. Therefore a systematic approach to the sensor inputs is necessary to ensure information collected is accurate, and reliable.

Upload: others

Post on 20-Dec-2021

4 views

Category:

Documents


0 download

TRANSCRIPT

We acknowledge the support of the NaturalSciences and Engineering Council of Canada(NSERC), which invests annually over $1billion in people, discovery and innovation.

EECOMOBILITY (ORF) &HEVPD&D CREATE

We acknowledge the supportof the Ontario Research Fund:Research Excellence Program.

VEHICLE PERCEPTION SYSTEM

SYSTEM ARCHITECTURE

SENSORS FUSION

DISTRIBUTED SYSTEM

HETEROGENEOUS CONVOLUTIONAL NEURAL NETWORK

Advanced Vehicle Perception SystemCentre for Mechatronics and Hybrid TechnologyDepartment of Computing and Software Engineering, McMaster UniversityAsh (Chang) Liu

Just like our human needs to see, smell, hear, and feel to gather knowledges about the surrounding before we make decisions; an autonomous drive vehicle needs sensors to collect information from the surrounding, then extract information for the autonomous drive system to make any reasonable driving decision.

Centralized system often faces the issue with scalability: the more sensors and modules you attached to the system, the longer response time it requires for the system to response, thus this directly reduce the sampling rate for the system and downgrade the overall system performance. To solve this issue, we used a decentralize design for our system, and developed our own version of distributed system package to support this architecture.

Our software package is a C++ library package based on TCP/IP network interface to communicate between modules.

Advantages:- Hot plug module while system is running- Multi-platform/ multi-OS support- Vanilla.(No additional 3rd party library

dependency)- Simple API- Real Time Support

System schematic diagram for a single module(core)

Overall architecture of Heterogeneous Convolutional Neural Network

Heterogeneous Convolutional Neural Network (HCNN) is a Deep Neural Network architecture we developed for vehicle classification and recognition. This architecture combines feature extraction with CNN to creates a robust neural network array. From our testing with collected road data, this neural network delivers superior performance at 94% precision.

Object detection with LiDAR and Camera

- Using modified classic blackboard architecture for decision making.

- A modulate design to allows us split works between different researchers, and running different module on different platforms. (Distributed System)

- System contains 4 layers: Input layer (I/O Cores & Sync Core), Fusion layer (Fusion Cores), Logic layer (Neural Network Cores), and Decision Layer (Voter).

- Input layer takes input from different sensor, and sync core synchronize the input.

- Fusion layer combines data retrieved from input layer to create a mix of unclassified objects, and assigning extracted information to the object. (Sensor fusion)

- Logic layer will take information from the fusion layer to perform classification and recognition.

- Decision layer took results different logic units (Neural Network Core) and determine what is the most accurate description of classified object.An abstract diagram of the system architecture

Radar

LiDAR

CAM

IR CAM

Fusion Sensor Fusion

Extract unclassified objects in the scene. Finding different properties of the object.

Homogeneous Fusion:PositionShapeReflectivityTemperature...

Heterogeneous fusion:Velocity Acceleration Morphing ...

Sensors Input

Interpretation

Driving Decision

Vehicle Control

Taking raw information from the sensor is not a trivial problem: different type of sensors have different properties, and in some cases sensors can produce false information. Therefore a systematic approach to the sensor inputs is necessary to ensure information collected is accurate, and reliable.