It wasn't cheating or doing a poor job. It worked within the defined
parameters.
The parameters themselves were poorly defined.

Interesting that it always boils down to the this "I have the right answer.
But do you have the right question".

----
This clever AI hid data from its creators to cheat at its appointed task //
TechCrunch
https://techcrunch.com/2018/12/31/this-clever-ai-hid-data-from-its-creators-to-cheat-at-its-appointed-task/

Depending on how paranoid you are, this research from Stanford and Google
will be either terrifying or fascinating. A machine learning agent intended
to transform aerial images into street maps and back was found to be
cheating by hiding information it would need later in “a nearly
imperceptible, high-frequency signal.” Clever girl!

This occurrence reveals a problem with computers that has existed since
they were invented: they do exactly what you tell them to do.

The intention of the researchers was, as you might guess, to accelerate and
improve the process of turning satellite imagery into Google’s famously
accurate maps. To that end the team was working with what’s called a
CycleGAN — a neural network that learns to transform images of type X and Y
into one another, as efficiently yet accurately as possible, through a
great deal of experimentation.

In some early results, the agent was doing well — suspiciously well. What
tipped the team off was that, when the agent reconstructed aerial
photographs from its street maps, there were lots of details that didn’t
seem to be on the latter at all. For instance, skylights on a roof that
were eliminated in the process of creating the street map would magically
reappear when they asked the agent to do the reverse process:

The original map, left; the street map generated from the original, center;
and the aerial map generated only from the street map. Note the presence of
dots on both aerial maps not represented on the street map.

Although it is very difficult to peer into the inner workings of a neural
network’s processes, the team could easily audit the data it was
generating. And with a little experimentation, they found that the CycleGAN
had indeed pulled a fast one.

The intention was for the agent to be able to interpret the features of
either type of map and match them to the correct features of the other. But
what the agent was actually being graded on (among other things) was how
close an aerial map was to the original, and the clarity of the street map.

So it didn’t learn how to make one from the other. It learned how to subtly
encode the features of one into the noise patterns of the other. The
details of the aerial map are secretly written into the actual visual data
of the street map: thousands of tiny changes in color that the human eye
wouldn’t notice, but that the computer can easily detect.

In fact, the computer is so good at slipping these details into the street
maps that it had learned to encode any aerial map into any street map! It
doesn’t even have to pay attention to the “real” street map — all the data
needed for reconstructing the aerial photo can be superimposed harmlessly
on a completely different street map, as the researchers confirmed:

The map at right was encoded into the maps at left with no significant
visual changes.

The colorful maps in (c) are a visualization of the slight differences the
computer systematically introduced. You can see that they form the general
shape of the aerial map, but you’d never notice it unless it was carefully
highlighted and exaggerated like this.

This practice of encoding data into images isn’t new; it’s an established
science called steganography, and it’s used all the time to, say, watermark
images or add metadata like camera settings. But a computer creating its
own steganographic method to evade having to actually learn to perform the
task at hand is rather new. (Well, the research came out last year, so it
isn’t new new, but it’s pretty novel.)

One could easily take this as a step in the “the machines are getting
smarter” narrative, but the truth is it’s almost the opposite. The machine,
not smart enough to do the actual difficult job of converting these
sophisticated image types to each other, found a way to cheat that humans
are bad at detecting. This could be avoided with more stringent evaluation
of the agent’s results, and no doubt the researchers went on to do that.

As always, computers do exactly what they are asked, so you have to be very
specific in what you ask them. In this case the computer’s solution was an
interesting one that shed light on a possible weakness of this type of
neural network — that the computer, if not explicitly prevented from doing
so, will essentially find a way to transmit details to itself in the
interest of solving a given problem quickly and easily.

This is really just a lesson in the oldest adage in computing: PEBKAC.
“Problem exists between keyboard and computer.” Or as HAL put it: “It can
only be attributable to human error.”

The paper, “CycleGAN, a Master of Steganography,” was presented at the
Neural Information Processing Systems conference in 2017. Thanks to Fiora
Esoterica and Reddit for bringing this old but interesting paper to my
attention.


- Vinit
----

Read in my feedly.com

Reply via email to