There might be something fundamentally wrong with the data used to train the AI systems we are increasingly relying on, scientists have warned.
Much of the data that is used to train those systems could rely on images that were taken without consent, are not diverse and do not protect privacy, researchers have warned. As well as being potentially unethical in itself, that could also have consequences for those people who rely on the AI systems trained on that data.
The warning comes after researchers released the Fair Human-Centric Image Benchmark, which is developed by Sony AI and aims to offer an ethically-sourced set of training images. Researchers can then use that data to evaluate whether systems have been trained using similarly ethically collected sets of data.
The research f

The Independent Technology

AlterNet
Raw Story