Trying to emulate human eyesight in self driving cars is hilariously misguided.
1. We have a computer behind our eyes so advanced that we may never be able to come close to replicating it. It is capable of identifying, tracking and predicting the behaviour of multiple objects in real-time even in reduced visibility and can infer new objects without training e.g. a green firetruck or a RV with a satellite dish.
2. Our eyes are connected to a very adaptive and movable object i.e. our head. In order to perceive depth and identity objects e.g. an actual person versus a photo of a person we continuously move our head around in multiple dimensions. A car can't do this.
I would refer everyone to the countless examples of Tesla's Autopilot recognising humans in bus signs, sides of trucks etc and attempting to do auto-avoidance. That is an unsolvable problem with only optical cameras.
Metal fatigue (and other wear and tear) might be a reason; our muscles are self healing and regenerating. It would likely be cheaper and less error prone to simply have more sensors, not moving at all or with much narrower ranges of motion to simulate depth perception. Training a computer to understand a simulacrum is an entirely different challenge, I think.
Of course cars can move their dozen or so cameras hundreds of times a second. But good luck getting a DNN to process all of that as well as the system handling tasks like identifying traffic lights, pedestrians.
Head of Tesla AI has already stated that even with the new Nvidia hardware they struggle to meet the computational requirements.
Those are two largely unrelated problems. Reconstructing geometry from multiple views is a separate (very well understood) problem from analysis to understand what the shapes mean.
1. We have a computer behind our eyes so advanced that we may never be able to come close to replicating it. It is capable of identifying, tracking and predicting the behaviour of multiple objects in real-time even in reduced visibility and can infer new objects without training e.g. a green firetruck or a RV with a satellite dish.
2. Our eyes are connected to a very adaptive and movable object i.e. our head. In order to perceive depth and identity objects e.g. an actual person versus a photo of a person we continuously move our head around in multiple dimensions. A car can't do this.
I would refer everyone to the countless examples of Tesla's Autopilot recognising humans in bus signs, sides of trucks etc and attempting to do auto-avoidance. That is an unsolvable problem with only optical cameras.