- cross-posted to:
- technology@lemmit.online
- cross-posted to:
- technology@lemmit.online
Driverless cars worse at detecting children and darker-skinned pedestrians say scientists::Researchers call for tighter regulations following major age and race-based discrepancies in AI autonomous systems.
This has been the case with pretty much every single piece of computer-vision software to ever exist…
Darker individuals blend into dark backgrounds better than lighter skinned individuals. Dark backgrounds are more common that light ones, ie; the absence of sufficient light is more common than 24/7 well-lit environments.
Obviously computer vision will struggle more with darker individuals.
No it’s because they train AI with pictures of white adults.
It literally wouldn’t matter for lidar, but Tesla uses visual cameras to save money and that weighs down everyone else’s metrics.
Lumping lidar cars with Tesla makes no sense
Visible light is racist.
deleted by creator
If the computer vision model can’t detect edges around a human-shaped object, that’s usually a dataset issue or a sensor (data collection) issue… And it sure as hell isn’t a sensor issue because humans do the task just fine.
Do they? People driving at night quite often have a hard time seeing pedestrians wearing dark colors.
Sounds like you have never reviewed dash camera video or low light photography.
Which cars are equipped with human eyes for sensors?