Though AI technology is perceived to be the next big thing, it also seems to have implications that go beyond its generic notion of making life easier for humans.
Artificial
Intelligence (AI) is one of the emerging concepts shaping the
world of technology today. From analysing big data to powering
autonomous vehicles, this technology has created new avenues for
humans to experiment and explore. Though the AI is perceived to be
the next big thing in technology, it also seems to have implications
that go beyond the generic notion of it making life easier for
humans.
A
research by Stanford and Google
has revealed that a machine learning agent that was meant to form
street maps from aerial images was found to be hiding information in
a clandestine manner to use it later stealthily. This incident came
to light when researchers were working on improving the process of
turning satellite imagery into Google Maps’ Street View feature,
according to a news report in technology and start-up news portal the
Tech Crunch.
This
is not the first time that an AI-based program has been found to
learn new tricks to improvise its work beyond the purpose it was
designed for. Last year, Facebook AI Research Lab (FAIR) also had to
shut down chatbots powered by the social media giant’s AI-engine
after they appeared to have been conversing in a language that only
they understood.
The
two examples above show that AI has, along with other things, also
imbibed a dark side from humans. Though it is designed to work on
data sourced from humans, its basic nature to learn, adapt and
improvise makes it change its course from what it is designed for to
what it is capable of — that can be horrendous and scary.
The
AI is also seen to have inherited a sense of bias from humans. In
2016, an investigation by ProPublica,
a US-based non-profit organisation, found that an AI-based software
(COMPAS) used by judges in some US states to calculate risk score of
a person re-committing a crime again, was biased against people of
colour.
Business Standar
No comments:
Post a Comment