
S1 Settings
We make use of three datasets, namely CIFAR-10, BOSS1.0.1, and ImageNet for the evaluation. CIFAR-10 and ImageNet are
widely used for various vision-related tasks and BOSS1.0.1 is the dataset commonly covered by the stenography fields. For the
overall experiments, for CIFAR-10, we utilize its training dataset for training and the testing dataset for testing. In the case of
BOSS1.0.1, it contains 10,000 images in total. We split the dataset into two: 9,000 for training and 1,000 for testing. In the case
of ImageNet, for all experiments including PixelCNN++ training, we utilize its training dataset for training and its validation
dataset for testing. We use 32 × 32 × 3 for CIFAR-10, 128 × 128 × 1 for Boss-1.0.1 dataset, and 64 × 64 × 3 for ImageNet. For
experimental settings, we spend approximately 1 week to train a PixelCNN++ model for each dataset using four NVIDIA P100
GPUs. For training Deep Steganography and ISGAN, we need less than 3h for each case on one NVIDIA GTX 1080 GPU. In
the case of Deep Steganography and ISGAN, we train each model using the hyperparameter proposed in each paper.
S2 Where is a Secret Image Encoded?
We further examine how and where the secret image is encoded in case of the deep learning based steganographic algorithms
as mentioned in Sec. 1. We could utilize the newly acquired discoveries to find a more appropriate way to destroy as much of
the secret image as possible while maintaining the quality of the cover image.
As illustrated in Fig. S1, after we increase the pixel value of the R channel of a random single pixel in the stego image by 1,
we evaluate its effect on the decoded secret image by measuring the residual between the secret image decoded by unmodified
stego image and randomly modified stego image, respectively. An increase by 1 in any position in any color channel of the
stego image has an impact across all color channels in the decoded secret image. We see that the largest impact commonly
occurs on the same channel.
We also witness that the secret image is encoded in a distributive but location-limited way. In other words, the pixel in-
formation of the left top corner of the secret image is only encoded on the left top corner of the cover image and likewise
for other areas. Based on this finding, we confirm that active steganalysis algorithms adjusting the stego image generated by
the deep learning-based encoding algorithms should apply the alternation of all the pixels of the stego image for the definitive
steganography removal.
Stego Image
Decoded
Secret Image
R
G
B
R
G
B
Figure S1: Impact of increasing a pixel value of the stego image by 1 on the decoded secret image. The left side of the figure is the residual
on each RGB channel between the stego image and modified stego image. The right side of the figure is the residual on each RGB channel
between the secret images decoded by the stego image and modified stego image, respectively.