Mathworks (Sensor Fusion for Autonomous Systems)
The goal of this project is to develop a sensor fusion algorithm for vehicle pose estimation using AI-based techniques. Vehicle pose refers to a vehicle's position and direction. Vehicles typically employ a variety of on-board sensors such as cameras and LiDAR sensors to gather environmental information and estimate vehicle pose. Traditionally, pose estimation was done using classic filtering techniques such as Kalman Filtering. However, we are now exploring Machine Learning (ML) and Artificial Intelligence (AI) based methods to help us with estimation. In this project, we will design a software application that can fuse vehicle sensor data from multiple sensors and use ML to analyze the data and estimate a vehicle's pose. General components of the work to be done include:
- Identifying a limited set of sensor types and data that we consider most relevant to vehicle pose estimation.
- Identifying an appropriate ML method through survey of recent literature and industry practices.
- Training and testing of the ML algorithm.
- Revision of software application with other appropriate sensor data. The idea behind this step is to determine if additional sensor data adds value to our vehicle pose estimation method.
The description of this project from Mathworks can be found here.
The group are divided in two teams:
Group 1: Computer Vision Team |
Angel Tinajero |
Gerardo Ibarra |
Hagop Arabian |
Jonathan Santos |
Deepanker Seth |
Group 2: Object Detection Team |
Daniel Gallegos |
Roberto Garcia |
David Neilson |
Xiao Hang Wang |
Patrick Emmanuel Sangalang |
Team Leader: David Neilsen
Team Communication Leader: Hagop Arabian
- Hagop Arabian
- Daniel Gallegos
- Roberto Garcia
- Gerardo Ibarra
- David Neilsen
- Patrick Sangalang
- Jonathan Santos
- Deepanker Seth
- Angel Tinajero
- Xiao Wang