Research Proposal: Does RG-chromaticity and depth differ from RGB in distinguishing anomalous objects in stereo image datasets for self-driving vehicles?
Self-driving vehicles are being tested live, on public streets, but when we attempt to validate the underlying image models’ performance, there remains some unexplained underperformance, which prevents reliable validation processes.
Vehicles are dangerous, massive objects that operate at speed near exposed members of the public. Self-driving technologies seek to reduce the probability of collisions, by both removing human inattention and unreliable analysis. However, partially automated systems consistently result in greatly increased human inattention, and the computational analysis it is replaced with is challenging to validate as reliable.
In 2018, the first pedestrian fatality by a self-driving vehicle was observed.
Further dependability research has been advocated. This project is part of asking: What is causing these apparently capable models to be assessed as operating incorrectly?
Please sign in
If you are a registered user on Laidlaw Scholars Network, please sign in