In a new study, researchers detailed a way to train AI algorithms to reach human levels of performance in a popular 3D multiplayer game - a modified version of Quake III Arena in Capture the Flag mode.
Business
Standard : Since the earliest days of virtual chess and
solitaire, video games have been a playing field for developing
artificial
intelligence (AI). Each victory of machine against human has
helped make algorithms smarter and more efficient. But in order to
tackle real world problems – such as automating complex tasks
including driving and negotiation – these algorithms must navigate
more complex environments than board games, and learn teamwork.
Teaching AI how to work and interact with other players to succeed
had been an insurmountable task – until now.
In
a new study, researchers detailed a way to train AI algorithms to
reach human levels of performance in a popular 3D multiplayer game –
a modified version of Quake III Arena in Capture the Flag mode.
Even
though the task of this game is straightforward – two opposing
teams compete to capture each other’s flags by navigating a map –
winning demands complex decision-making and an ability to predict and
respond to the actions of other players.
This
is the first time an AI has attained human-like skills in a
first-person video
game. So how did the researchers do it?
The
robot learning curve
In
2019, several milestones in AI research have been reached in other
multiplayer strategy games. Five “bots” – players controlled by
an AI – defeated a professional e-sports team in a game of DOTA 2.
Professional human players were also beaten by an AI in a game of
StarCraft II. In all cases, a form of reinforcement learning was
applied, whereby the algorithm learns by trial and error and by
interacting with its environment.
The
five bots that beat humans at DOTA 2 didn’t learn from humans
playing – they were trained exclusively by playing matches against
clones of themselves. The improvement that allowed them to defeat
professional players came from scaling existing algorithms. Due to
the computer’s speed, the AI could play in a few seconds a game
that takes minutes or even hours for humans to play. This allowed the
researchers to train their AI with 45,000 years of gameplay within
ten months of real-time.
The
Capture the Flag bot from the recent study also began learning from
scratch. But instead of playing against its identical clone, a cohort
of 30 bots was created and trained in parallel with their own
internal reward signal.
No comments:
Post a Comment