Structure-Aware Motion Deblurring Using
Structure-Aware Motion Deblurring Using
30, 2021
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
WEN et al.: STRUCTURE-AWARE MOTION DEBLURRING USING MULTI-ADVERSARIAL OPTIMIZED CycleGAN 6143
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
6144 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
through a variety of technologies, such as bilateral filtering realize the deblurring task by unsupervised use of unpaired
and gradient amplitude. In [29], Xu et al. propose an training data, [16], [41] only target at the specific image
L0-regularized gradient prior based on the sharp edge infor- domain deblurring problem, while [42] will encode other
mation for blind image deblurring. In [12], Pan et al. develop factors (color, texture, etc., instead of blurred information) into
an optimization method based on L 0 -regularized intensity and the generated deblurred image. Different from these previous
gradient prior to generate reliable intermediate results for blur methods, our unsupervised method can solve the demand
kernel estimation. In [30], Sun et al. use dictionary learning of paired training data problems for the image deblurring.
to predict the sharp edges with the sharp edge patches of Meanwhile, we utilize the multi-adversarial architecture and
clear images for deblurring. In [31], Pan et al. describe a structure-aware mechanism to further remove the unpleasant
blind image deblurring method with the dark channel prior. artifacts and maintain structure information effectively.
In [32], Kim et al. propose to estimate the motion flow and the
latent sharp image simultaneously based on the total variation III. P ROPOSED M ETHOD
(TV)-L1 model. In [33], Bai et al. propose a multi-scale latent Our overall flowchart is shown in Fig. 2. In Fig. 2, G B
structure prior and gradually restore the sharp images from the and G S are two generator sub-networks which transform from
coarse-to-fine scales on a blurry image. Recently, thanks to the the sharp image to the blurred image and from the blurred
powerful semantic analysis and deep mining ability of CNNs, image to the sharp image, respectively. D B and D S are the
more works tend to use large-scale samples to solve the blind discriminators to distinguish the real images and generated
image deblurring problems. images, and give feedback to the generators. Different from the
2) CNNs Based Kernel Estimation Method for Deblurring:
traditional CycleGAN [1], we use the form of multi-adversarial
In recent years, CNNs have played an unparalleled advan-
in different resolution constraints to gradually improve the
tage in solving computer vision problems including image
quality of the generated images and use skip connections to
deblurring and achieved many promising results [7]–[9]. Some
make the low-level information better guide the high-level
methods use CNNs to estimate the blur kernel to achieve the
generation structure. Meanwhile, we design a structure-aware
deblurring task. For instance, Sun et al. mainly estimate the
mechanism by introducing the multi-scale edge constraints
probabilistic distribution of the unknown motion blur kernel
in the multi-adversarial architecture to make the adversarial
based on CNN for deblurring [34]. However, these methods
network generate persuasive structural information at different
have strict requirements for paired training data and cannot
resolutions, and edge map is also used as part of the input
directly realize the transformation from the blurred image to
to facilitate the network’s retention of structural information.
the sharp image, and still cannot avoid errors in the process of
Besides, we add a variety of loss functions (structural loss
blur kernel estimation based on CNNs [35], [36]. In contrast,
MS-SSIM and perceptual loss obtained by VGG16) to fur-
our approach can avoid these errors, since our method is based
ther strengthen the constraints to reduce the generated false
on the unsupervised image-to-image translation with unpaired
information. Compared with other methods, our method can
training data and can directly realize the transformation from
not only solve the demand of paired data problem, but also
blurred images to sharp images without kernel estimation
can maintain more structural information and achieve a better
process. In this paper, we show a comparison with [12], [31],
deblurring effect.
[34] to verify our advantages in Session IV-E.
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
WEN et al.: STRUCTURE-AWARE MOTION DEBLURRING USING MULTI-ADVERSARIAL OPTIMIZED CycleGAN 6145
Fig. 2. The flowchart of our structure-aware multi-adversarial optimized CycleGAN. Our architecture relies on the unsupervised image-to-image translation to
learn the mapping between blurred images and deblurred (sharp) images with unpaired training data. G S and G B are two generator sub-networks for translating
blurred image to sharp image and translating sharp image to blurred image, respectively. D S64 , D S128 and D S256 are the discriminators to determine whether
the image generated by G S is real or fake at three resolutions. D B64 , D B128 and D B256 are the discriminators to determine whether the image generated by
G B is real or fake at three resolutions. We restore sharp images by this multi-adversarial manner to iteratively generate high-resolution from low-resolution
images. In addition, we introduce the structure-aware mechanism by adding edge input to guide the generation procedure and multi-scale edge losses to
maintain more structure details at different resolutions. Besides, we utilize cycle-consistency loss, perceptual loss and MS-SSIM loss to enforce constraints
on the structure generation.
sometimes lost [16], [43]. In order to solve these problems, images with three resolution levels. Then, three independent
we expect to improve the generation effect step by step with discriminators will judge the authenticity of the generated
multi-adversarial architecture and structure-ware mechanism. images on different resolutions and feed information to the
generators. The hidden layers with different resolutions in the
B. Multi-Adversarial Generative Network network are constrained and the feature maps are iteratively
optimized to generate higher quality results. Additionally,
As discussed in Section II-B, the classical GAN-based
structure often introduces artifacts when generating realistic the generated edge maps at three different resolutions are used
for multi-scale edge constraints to improve the structure reten-
images, especially with the increase of resolution. To solve this
tion performance of the network. We also use skip connections
problem, a multi-scale way is preferred to improve the quality
of the generated images [10]. Ideally, a mature multi-scale to take full advantage of the low-level information to guide the
deconvolution process.
approach not only can significantly improve the network per-
For a blurred image b, generator G S generates synthesized
formance but also need to minimize parameters to reduce time
consumption and hardware burden. However, the parameters sharp image sb1 , sb2 , sb3 as outputs. The sb3 , which presents
the output of the last deconvolution layer, is sent as the input of
in some multi-scale approaches at each scale are still inde-
pendent of each other in some multi-scale methods [10], [20]. G B to generate three reconstructions b1 ,
b2 and b3 . Similarly,
for a deblurred (sharp) image s as input, G B will output
Given this, we introduce the multi-adversarial architecture in
synthesized blurred images bs1 , bs2 and bs3 . And with bs3 as
our unsupervised deblurring model to make full use of the
input information and avoid the problem of false information the input, the generator G S will produce three reconstructions
s1 ,
s2 and s3 . We then supervise these different outputs to
increasing with the increase of resolution.
force them closer to the target at different resolutions. D S64,
Inspired by the traditional encoder-decoder network struc-
ture [44], the generator G S in our proposed multi-adversarial D S128 and D S256 are defined for G S . D B64 , D B128 and D B256
are defined for G B . Three resolutions of 64 × 64, 128 × 128
network is shown in Fig. 3. The input of the generator
and 256 × 256 are applied on the corresponding deconvolution
sub-network G S is the blurred image and the corresponding
layers, respectively. The adversarial losses can be written as
edge map obtained by Sobel operator. The edge map used as
Eq. (1) and Eq. (2):
part of the input can provide additional structural information
to the network. G S contains a series of convolution layers,
L adv (G S , D Si ) = E b∼ p(b) log(1 − D Si (G S (b)i ))
deconvolution layers and upper sampling layers. Feature maps
+ E si ∼ p(si ) log(D Si (si )) (1)
are generated from each deconvolution layer through a 3 × 3
convolution forward layer with output images at different L adv (G B , D Bi ) = E s∼ p(s) log(1 − D Bi (G B (s)i ))
resolutions. From Fig. 3, generator G S can produce the output + E bi ∼ p(bi ) log(D Bi (bi )) (2)
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
6146 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
Fig. 3. Network structure of the proposed multi-adversarial generator. G S is the generator sub-network for the translation from the blurred image to the
deblurred (sharp) image. The input of the generator sub-network G S is the blurred image and the corresponding edge map obtained by Sobel operator. By the
multi-adversarial manner, G S can produce three different resolution outputs (64 × 64, 128 × 128 and 256 × 256). Multi-adversarial supervision is achieved
through multiple discriminators in the hidden layers. Discriminators D S64 , D S128 , D S256 are defined for G S at three different resolutions, respectively.
In addition, the generated edge maps at three different resolutions are used for multi-scale edge constraints to improve the structure retention performance of
the network. The specific parameters of the generator sub-network are shown in the figure so that we can train our multi-adversarial model with a specific
size and test the image of any size.
where G S (b)i = sbi , G B (s)i = bsi and i = 1, 2, 3 corresponds image quality assessment and image restoration tasks. In [16],
to the three different resolutions. bi and si are the blurred an unsupervised network for deblurring with a reblurring cost
image and sharp image at i t h resolution, respectively. D Bi and and a scale-space gradient cost is proposed. In [11], Vasu et al.
D Si are the discriminators corresponding to G B and G S at i t h first investigate the relationship between the edge profiles and
scale, respectively. the camera motion, and then incorporate the edge profiles into
As for the cycle-consistency loss in the traditional an existing blind deblurring framework. In [45], a two-stage
CycleGAN, it can be improved to multiple resolutions: edge-aware network is proposed to improve image deblurring
according to the feature that human eyes pay more attention to
L cycbi = bi − bi 1 = G B (G S (b)3 )i − bi 1 (3) edge sharpening. Although several structure-aware strategies
L cycsi = si − si 1 = G S (G B (s)3 )i − si 1 (4) have been successively applied to deblurring problems, it is
still difficult to maintain structure information and reduce
where G S (b)3 = sb3 and G B (s)3 = bs3 . The final
inherent ambiguity in unsupervised deblurring tasks.
multi-adversarial objective function is defined as:
In order to preserve the structural information of the
L Mult iG AN (G S , G B , D S , D B ) deblurred image to the maximum extent, we introduce the
3 structure-aware mechanism by taking the corresponding edge
= (L adv (G S , D Si ) + L adv (G B , D Bi ) map as part input and adding multi-scale edge constraint func-
i=1 tions in the multi-adversarial architecture. Different from the
+ μi (L cycbi + L cycsi )) (5) structure-aware mechanism in other image processing tasks,
the structure-aware mechanism in our unsupervised deblurring
Simplified as: model not only includes the input edge clues for structural
3 information assistance but also includes multi-scale edge con-
L Mult iG AN = (L adv i + μi L cyci ) (6) straints for generating the deblurring with different resolutions.
i=1 Besides, the multi-scale edge constraints can be organically
combined with the multi-adversarial strategy to promote the
where μi is the weight parameter at i t h resolution to balance
generation of structural information in unsupervised networks.
the different components. L cyci = L cycsi + L cycbi , and L adv i = We have verified that both of them can effectively promote
L adv (G S , D Si ) + L adv (G B , D Bi ) the structure retention ability of the network and generate
a more satisfactory deblurring effect through the ablation
C. Structure-Aware Mechanism for Deblurring experiments.
The high-frequency details of the image are weakened The proposed structure-aware mechanism can emphasize
to some extent due to the blurring process, how to restore the protection of image geometry to alleviate the important
the structure and details as much as possible in the image ambiguity problem of the original CycleGAN. In this paper,
deblurring task is very important. Previous studies [11], [16], the proposed structure-aware mechanism network is shown
[45] prove that image edge is of great significance in subjective in Fig. 3. Due to the input edge guidance, the Eq. (1) and
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
WEN et al.: STRUCTURE-AWARE MOTION DEBLURRING USING MULTI-ADVERSARIAL OPTIMIZED CycleGAN 6147
Fig. 4. Comparative experiment of structure maintenance effect. (a) The original blurred image. (b) Deblurring result using CycleGAN [1]. (c) Deblurring
result with edge map as input. (d) Deblurring result with edge loss. (e) Deblurring result with both edge map as input and edge loss. It shows our method
is more satisfying, especially in the yellow rectangles.
Eq. (2) can be revised as Eq. (7) and Eq. (8): combination can better improve the motion deblurring perfor-
mance as shown in Fig. 4(e).
L adv (G S , D Si ) = E b∼ p(b) log(1 − D Si (G S (b, be )i ))
+ E si ∼ p(si ) log(D Si (si , sie )) (7) D. The Network Structure
1) Generator: The generator in our architecture is shown
L adv (G B , D Bi ) = E s∼ p(s) log(1 − D Bi (G B (s, s e )i ))
in Fig. 3. It contains a series of convolution layers
+ E bi ∼ p(bi ) log(D Bi (bi , bie )) (8) and residual blocks. Specific as follows: C7S1 − 64,
where be and s e are the edge maps of the image b and C3 − 128, C3 − 256, R B256 × 9, T C64, T C32, C7S1 − 3,
image s obtained by Sobel operator, respectively. bie and sie where, C7S1 − k represents a 7 × 7 ConvBNReLU
are the responding edge maps at i t h resolution. By this edge (Convolution+BatchNorm+ReLU) block with stride 1 and
guidance manner, we can take the advantage of the additional k filters, C3 − k represents a 3 × 3 ConvBNReLU block
edge information to make the generated images in the target with stride 2 and k filters. R Bk × n denotes k filters and
domain contain similar edge structure information of the n residual blocks which contain two 3 × 3 convolution lay-
source domain and better guide the discriminator to distinguish ers, T Ck represents a 3 × 3 TConvBNReLU (Transposed
the generated images from the real images. However, even the Convolution+BatchNorm+ReLU) block with stride 1/2 and
edge guidance can improve the accuracy of discrimination, k filters. In addition, we introduce the structure-aware archi-
we find that the generated deblurred image still exits the tecture (including edge input guidance and multi-scale edge
problems of ringing and oversharp. constrains) in G S and G B during training process.
In order to handle the existing problems and force the struc- 2) Discriminator: The discriminator is also shown in Fig. 3.
ture of the generated deblurred image to match its correspond- Classic PatchGANs [47] is used as a discriminator to classify
ing sharp image, we introduce the multi-scale edge losses in overlapping image blocks and determine whether they are real
the multi-adversarial structure. Since our unsupervised method or false. All the discriminator networks at three resolutions
has no access to the corresponding reference image and it mainly include: C64−C128−C256−C512, here Ck presents
is difficult to generate an accurate corresponding edge map, a 4 × 4 ConvBNLeakyReLU (Convolution + BatchNorm +
we follow the heuristic from [16], [46] and utilize the fact LeakyReLU) block with stride 2 and k filters. The parameter of
that the resized image bη which is obtained by shrinking a LeakyReLU is set to 0.2 in our experiment. According to the
blurred image b with a factor of η is sharper than the image specific parameters of the generator and discriminator, we can
b itself. Thus, we introduce the multi-scale edge losses to train our multi-adversarial model with a specific size and test
enforce the edge of the generated deblurred image to match the images of any size.
its corresponding sharp image. The factor of η in our model
E. Loss Functions
is set to 0, 1/2 and 1/4 for three different scales respectively.
1) Multi-Scale SSIM Loss: The perceptually motivated met-
Then, the introduced multi-scale edge losses are defined as:
ric Structural SIMilarity index (SSIM) [48] has often been
L Gradbi = sbi − bi 1 = (G S (b)i ) − bi 1 (9) used to measure the similarity of two images. To preserve the
information of contrast, luminance, structure in the generated
L Grad = bs − si = (G B (s)i ) − si 1 (10)
si i 1 images and alleviate the ambiguity problem of CycleGAN,
where is the Sobel operator to calculate the gradient map we use the Multi-scale SSIM loss (MS-SSIM) based on SSIM
of an image, and L Gradi = L Gradbi + L Gradsi . between bi and bi in our model. The MS-SSIM we used is
Fig. 4 shows the effect of just using the edge loss and adding defined as:
edge as an input to the generator. From Fig. 4, most structure α
M
β
information can be migrated to the target domain with edge L M S S I Mbi = 1− l M (bi ,
bi ) M [c j (bi ,
bi )] j [m j (bi ,
bi )]γ j
input in Fig. 4(c), and most artificial noise can be effectively j =1
eliminated through multi-scale edge losses in Fig. 4(d). The (11)
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
6148 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
2μbi μ
b +C 1 2σbi σ
b +C 2
where l(bi ,
bi ) = μ2b +μ
i
2 +C , c(bi ,
bi ) = i
σb2 +σ2 +C 2
and IV. E XPERIMENTAL R ESULTS
bi 1 bi
i i
σb +C A. Implementation Details
m(bi ,
3
bi ) = i bi
σbi σ . (bi , bi ) denotes the image pair of
b +C 3
i We conduct our training and testing experiments on a
input image and the reconstructed image, respectively. μbi , workstation with Intel Xeon E5 CPU and NVIDIA 2080ti
μbi , σbi , σ
bi , σbi
bi indicate the means, standard deviations and GPU. The model we used is implemented with Pytorch
cross-covariance of the image pair (bi , bi ), respectively. C1 , platform [50]. For fairness, all the experiments are set in the
C2 and C3 are the constants determined according to refer- same data set and environment except for special instructions.
ence [48]. l(bi , bi ), c(bi ,
bi ) and m(bi ,
bi ) denote the compar- Throughout our experiments, we use ADAM [51] solver for
ison components of luminance, contrast and structure between model training with parameters β1 = 0.9 and β2 = 0.999.
bi and bi , respectively. α, β and γ are the hyper-parameters Limited by the memory, the batch-size is set to 2 for all the
set according to [48], which are used to control the relative methods. The initial learning rate is fixed to 0.0002 for the
weight of the three comparison components. first 30 epoches and then decay to one-tenth every 30 epoches.
Similarly, the MS-SSIM loss function L M S S I Msi between Totally, 200 epoches already satisfy the convergence condition.
si and si is defined as the same way, and the total MS-SSIM
loss at i t h resolution is L M S S I Mi = L M S S I Mbi + L M S S I Msi . B. Datasets and Metrics
2) Perceptual Loss: Previous work [38] shows that cyclic
For the blurred text images, we use the dataset
perceptual-consistency losses have the ability to preserve
BMVC_TEXT [52] which totally contains 66K text images
original image structure by investigating the combination of
with the size 300 × 300. This dataset contains both defocus
high-level and low-level features extracted from the second
blur generated by anti-aliased disc and motion blur generated
and fifth pooling layers of VGG16 [49] architecture. Accord-
by a random walk. The blurred images in BMVC_TEXT are
ing to [38], the formulation of cyclic perceptual-consistency
divided into two parts: the training set and the test set (50% of
loss is given below, where (bi , bi ) refers to the blurred and
the total, and no crossover is ensured), and the corresponding
ground truth image set, φ is a VGG16 [38], [49] feature
sharp images are divided in the same way. During the training
extractor from the second and fifth pooling layers:
process, we crop the image into 128 × 128 image blocks
2 in both the blur set and the sharp set. The parameter ω1
L Percept ualbi = φ(
bi ) − φ(bi )2 (12)
is set to 5, parameters ω2 and ω3 are set to 0.5, ω4 is set
Similarly, L Percept ualsi between si and si is defined as the to 10 and ω5 is set to 0 in Eq. (15) because we find that
the perceptual loss L Percept ual has little impact on overall
same way, and the total perceptual loss at i t h resolution is
performance. To compare with other classical deblurring meth-
L Percept uali = L Percept ualsi + L Percept ualbi .
ods, we choose the algorithms given by Pan et al. [12], [31],
3) Identity Preserving Loss: In addition, we use an identity
Xu et al. [29], Sun et al. [34], MS-CNN [10], Deblur-
preserving loss to reinforce the identity information of the
GAN [14]. We also choose other unsupervised methods Cycle-
input image during the unpaired image-to-image translation.
GAN [1], Madam et al. [16] and UID-GAN [43] that trained
Thus, information such as the color of the input and output
on the same text training dataset with our unpaired data.
images can be mapped as accurately as possible. The identity
For the blurred face images, the CelebA dataset [53] which
preserving loss between the source domain and target domain
mainly includes more than 200K face images with size
can be defined as:
178 × 218 are used. We first select 200K data from the data
set, where 100K is the sharp images and the other 100K is
L I d bi = G B (b)i − bi 1 (13)
the blurred images. In addition, we select 2000 images from
L I d si = G S (s)i − si 1 (14) the remaining images for testing. We scale all the images to
128 × 128 and ensure that there is no paired data during the
The total identity preserving loss at i t h resolution is L I di = unsupervised algorithm training. The method of generating
L I d bi + L I d si . From the above loss functions described in blurred images by sharp images is consistent with the method
Eq. (1) ∼ Eq. (14), the total loss for our deblurring model proposed in UID-GAN [43]. The parameters ω1 ∼ ω4 are
is: set in the same way as BMVC_TEXT [52] dataset, and the
parameter ω5 is set to 5.
3
L = (L adv i + ω1 L cyclei + ω2 L Gradi + ω3 L M S S I Mi For the motion blurred images, the same as [10], we firstly
i=1 use the GoPro dataset proposed in [10] to train our model.
+ ω4 L I di + ω5 L Percept uali ) (15) Since our model is based on the unsupervised image-to-image
translation, during the training process, we firstly segregate the
where, ω1 , ω2 , ω3 , ω4 and ω5 are non-negative constants GoPro dataset into two parts. We just use the blurred images
to adjust different influence on overall deblurring effects. i from one part and the clean (sharp) image from the second
denotes the component at i t h resolution. Similar to other part so that there are no corresponding pairs while the training
previous methods [1], [10], parameters ω1 , ω2 , ω3 , ω4 and process. 2103 blurred/clear unpaired images in GoPro dataset
ω5 in Eq. (15) are set according to the data characteristics for are used for training and the remaining 1111 images are used
different cases and we weight each loss empirically to balance for evaluation. We ensure no overlap in the training pairs and
the importance of each component. randomly crop the image into 256 × 256 image blocks in both
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
WEN et al.: STRUCTURE-AWARE MOTION DEBLURRING USING MULTI-ADVERSARIAL OPTIMIZED CycleGAN 6149
TABLE I
A BLATION S TUDY ON THE E FFECTIVENESS OF D IFFERENT C OMPONENTS
IN O UR M ODEL . A LL THE R ESULTS A RE T ESTED ON THE G O P RO
D ATASET [10]. G S M EANS THE T RANSLATION F ROM THE B LUR
D OMAIN TO THE S HARP D OMAIN , AND G B M EANS THE
T RANSLATION F ROM THE S HARP D OMAIN
TO THE B LUR D OMAIN
the blur set and the sharp set. The parameter ω1 is set to 5,
parameters ω2 and ω3 are set to 0.5, ω4 is set to 10 and
ω5 is set to 1 in Eq. (15). We use PSNR and SSIM two
metrics to show quantitative comparisons with other deblurring
algorithms.
C. Ablation Study
To analyze the effectiveness of each important component
or loss (perceptual etc.), we perform an ablation study in
this section. Both quantitative and qualitative results on the
GoPro dataset are presented for the following six variants of
our method by adding each component gradually: 1) origi- Fig. 5. Stability analysis for our proposed model. (a) The overall loss
nal CycleGAN method [1]; 2) adding the multi-adversarial variation. (b) The perceptual loss variation. (c) The multi-scale edge losses
structure; 3) adding edge map input component; 4) adding variation of our method at resolution 256×256. (d), (e) and (f) are the identity
loss variation at resolution 64 × 64, 128 × 128 and 256 × 256, respectively.
multi-scale edge constraints; 5) adding multi-scale SSIM loss; (a), (b), (c) and (d) show that different losses of our model can steadily
6) adding all the above components. decrease with the increase of iteration times during the training process.
We present the PSNR and SSIM for each variant in Table I. (d), (e) and (f) indicate the identity preserving loss of our model decrease
steadily with the increase of iteration times at different resolutions.
G S (blur-sharp) means the translation from the blurred domain
to the sharp domain, and G B (sharp-blur) means the trans-
lation from the sharp domain to the blurred domain. From To illustrate the stability of the proposed model, Fig. 5
shows the different loss change curves of our proposed meth-
Table I, we can see that the multi-adversarial structure signif-
ods. Fig. 5(a) is the overall loss variation curve. Fig. 5(b) is
icantly improves the deblurring performance because of the
multi-resolution constraints. Meanwhile, the structure-aware the perceptual loss variation curve. Fig. 5(c) is the multi-scale
edge losses variation of our method at resolution 256 × 256.
mechanism (with the edge as input and multi-scale edge
Fig. 5(d), Fig. 5(e) and Fig. 5(f) indicate that the identity
constraints) can also preserve the structure and details because
preserving loss of our model can decrease steadily with the
of the additional edge information and edge constraints. Even
increase of iteration times at different resolutions (64 × 64,
the original CycleGAN basically implements the unsupervised
translation from blurred to sharp and from sharp to blurred, 128 × 128 and 256 × 256, respectively). As seen from the
change curve of all losses, different types of losses and losses
it introduces the unpleasant noise information (colors, textures,
with different resolutions can steadily decline with the increase
etc.). In contrast, with adding the multi-adversarial structure,
discriminators are able to determine whether the resulting of iteration times during the training process, which fully
indicates that our model is relatively stable.
clear image is true or false from multiple resolutions and then
feedback to the generators. With the edge map as part of the
input, more structure-guided information can be transferred to D. Parameter Sensitivity
the target domain. With the multi-scale edge constraints to As we mentioned in Section III-E, the weight ω1 for
guide the deblurring process, some unwanted ringing artifacts cycle-consistency loss L cycle , ω4 for identity preserving loss
at the boundary of the generated images can be removed L I d , ω5 for perceptual loss L Percept ual need to be tuned
effectively. With the multi-scale SSIM loss, the generated so that the deblurred image neither stays too close to the
image can preserve the luminance, contrast and structure original blurred image, nor contains many artifacts. The
information effectively. The overall deblurring performance quantitative performance is shown in Fig. 6. From Fig. 6,
in Table I also shows that there is a close relationship we can see that parameter ω4 setting for L I d is greatly
between our multi-adversarial learning and the structure-aware different from the traditional CycleGAN based task (such as
mechanism. for Photo-Sketch). As our method is based on multi-resolution
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
6150 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
WEN et al.: STRUCTURE-AWARE MOTION DEBLURRING USING MULTI-ADVERSARIAL OPTIMIZED CycleGAN 6151
TABLE II
P EAK S IGNAL - TO -N OISE R ATIO AND S TRUCTURAL S IMILARITY M EASURE , M EAN ON THE BMVC_TEXT [52] AND FACE D ATASETS [53]
TABLE III
P EAK S IGNAL - TO -N OISE R ATIO AND S TRUCTURAL S IMILARITY M EASURE , M EAN ON THE G O P RO D ATASET [10]
Fig. 8. Comparison of deblurred images by our method and other popular approaches on some images from BMVC_TEXT dataset [52]. (a) Blurred images.
(b) Deblurring results using Pan et al. [12]. (c) Deblurring results using Pan et al. [31]. (d) Deblurring results using Xu et al. [29]. (e) Deblurring results
using Sun et al. [34]. (f) Deblurring results using MS-CNN [10]. (g) Deblurring results using CycleGAN [1]. (h) Our results. It shows the characters in our
results are much clearer.
Fig. 9. Comparison of deblurred images by our method and other popular approaches on one sample from GoPro Dataset [10]. (a) Blurred image. (b) Deblurring
results using Pan et al. [12]. (c) Deblurring results using Xu et al. [29]. (d) Deblurring results using Sun et al. [34]. (e) Deblurring results using MS-CNN [10].
(f) Deblurring results using CycleGAN [1]. (g) Deblurring result using DiscoGAN [17]. (h) Our results. It shows our results are more satisfying, especially
in the pink and yellow rectangles.
methods will introduce new artifacts, while our method can on Köhler dataset in Fig. 10, which also verifies our better
better maintain the structure in the areas such as the girl’s performance compared with both supervised and unsupervised
head flower or arm. We also provide the visual contrast effect methods.
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
6152 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
Fig. 10. Comparison of deblurred images by our method and other popular approaches on one sample taken from Köhler Dataset [55]. (a) Blurred image.
(b) Deblurring result using Pan et al. [12]. (c) Deblurring result using Xu et al. [29]. (d) Deblurring result using Sun et al. [34]. (e) Deblurring result using
MS-CNN [10]. (f) Deblurring result using CycleGAN [1]. (g) Deblurring result using DiscoGAN [17]. (h) Our results. It shows our results are more satisfying,
especially in the pink and yellow rectangles.
Fig. 11. Comparison of deblurred images by our method and other popular approaches on one real image taken from Lai Dataset [54]. (a) Blurred image.
(b) Deblurring result using [31]. (c) Deblurring result using [29]. (d) Deblurring result using [12]. (e) Deblurring result using [34]. (f) Deblurring result using
[16]. (g) Deblurring result using CycleGAN [1]. (h) Deblurring result using [17]. (i) Deblurring result using [47]. (j) Deblurring result by our method.
3) Real Dataset: In order to compare the effects of different involved) to form the winning matrix. The quantitative results
deblurring algorithms on real blurred images, we use the in Table IV show that the methods based on CNNs usually
model trained on GoPro data set to test the real blurred images have better effect than the convolutional methods, and our
in the real set of Lai dataset [54]. Since the real blurred method can achieve a more satisfied deblurring effect in real
images do not provide the corresponding sharp images, it is blurred images compared with most existing methods. From
impossible to evaluate the deblurring effect with the full ref- Fig. 11, our method shows superior performance compared
erence image quality evaluation methods (Such as SSIM and with other methods, especially in the girl’s eyes and mouth.
PSNR). Therefore, we compare the deblurring performance According to the above experiments, we can conclude that
of different algorithms in the real blurred images with the our method has obvious advantages in solving the deblur-
help of subjective user analysis. Inspired by [56], we use ring task on all the test datasets when comparing with the
the Bradley-Terry model to estimate the subjective score. most existing unsupervised deblurring methods [1], [16],
Each blurred image is processed with the deblurring methods [43]. We can also infer that our unsupervised deblurring
Pan et al. [12], Xu et al. [29], Whyte et al. [6], Sun et al. method can achieve competitive results with the supervised
[30], MS-CNN [10], CycleGAN [1] and DeblurGAN [14]. deblurring algorithm [10], [12], [14], [29] in most datasets
We test all these methods with corresponding models trained except for the GoPro dataset. We believe this is mainly due
on GoPro. Together with the original blurred images, all these to CycleGAN’s lack of ability to generate high-resolution
results are sent for pairwise comparison (22 human raters are images and the difficulty for unpaired data learning compared
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
WEN et al.: STRUCTURE-AWARE MOTION DEBLURRING USING MULTI-ADVERSARIAL OPTIMIZED CycleGAN 6153
TABLE IV
AVERAGE S UBJECTIVE E VALUATION S CORES OF D EBLURRING P ERFORMANCE ON THE R EAL D ATASET [54]
TABLE V R EFERENCES
T HE AVERAGE RUNNING T IME C OMPARISONS OF O UR M ETHOD W ITH
O THER S EVERAL C LASSICAL M ETHODS ON [1] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image
BMVC_TEXT D ATASET [52] translation using cycle-consistent adversarial networks,” in Proc. IEEE
Int. Conf. Comput. Vis. (ICCV), Oct. 2017, pp. 2242–2251.
[2] V. Papyan and M. Elad, “Multi-scale patch-based image restora-
tion,” IEEE Trans. Image Process., vol. 25, no. 1, pp. 249–261,
Jan. 2016.
[3] M. Temerinac-Ott, O. Ronneberger, P. Ochs, W. Driever, T. Brox, and
H. Burkhardt, “Multiview deblurring for 3-D images from light-sheet-
based fluorescence microscopy,” IEEE Trans. Image Process., vol. 21,
no. 4, pp. 1863–1873, Apr. 2012.
with paired data. Since our deblurring method is based on [4] A. Danielyan, V. Katkovnik, and K. Egiazarian, “BM3D frames and
variational image deblurring,” IEEE Trans. Image Process., vol. 21,
unsupervised learning and can be trained with finite unpaired no. 4, pp. 1715–1728, Apr. 2012.
training data. Compared with other supervised-based methods, [5] W. Dong, L. Zhang, G. Shi, and X. Wu, “Image deblurring and super-
our unsupervised deblurring method has a wider application resolution by adaptive sparse domain selection and adaptive regular-
ization,” IEEE Trans. Image Process., vol. 20, no. 7, pp. 1838–1857,
value. Jul. 2011.
[6] O. Whyte, J. Sivic, A. Zisserman, and J. Ponce, “Non-uniform deblurring
F. Evaluation of the Running Time for shaken images,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis.
Pattern Recognit., Jun. 2010, pp. 491–498.
Table V shows the average running time per image compar- [7] T. M. Nimisha, A. K. Singh, and A. N. Rajagopalan, “Blur-invariant
isons of several classical deblurring methods with 512 × 512 deep learning for blind-deblurring,” in Proc. IEEE Int. Conf. Comput.
on the test dataset of BMVC_TEXT dataset [52]. According Vis. (ICCV), Oct. 2017, pp. 4762–4770.
[8] C. J. Schuler, M. Hirsch, S. Harmeling, and B. Scholkopf, “Learning
to Table V, we can see that the proposed unsupervised to deblur,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 38, no. 7,
method achieves the state-of-the-art deblurring quality, while pp. 1439–1451, Jul. 2016.
maintains relatively high and competitive speed in comparison [9] X. Xu, J. Pan, Y.-J. Zhang, and M.-H. Yang, “Motion blur kernel
estimation via deep learning,” IEEE Trans. Image Process., vol. 27,
to most existing supervised and unsupervised methods on no. 1, pp. 194–205, Jan. 2018.
BMVC_TEXT dataset [52]. Even though the time used is [10] S. Nah, T. H. Kim, and K. M. Lee, “Deep multi-scale convo-
slightly longer than CycleGAN [1] and MS-CNN [10] due to lutional neural network for dynamic scene deblurring,” in Proc.
IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017,
the multi-adversarial and multiple constraints structure, we get pp. 257–265.
a better deblurring effect. In future work, we are committed to [11] S. Vasu and A. N. Rajagopalan, “From local to global: Edge profiles to
further streamlining the network and improving its operational camera motion in blurred images,” in Proc. IEEE Conf. Comput. Vis.
Pattern Recognit. (CVPR), Jul. 2017, pp. 558–567.
efficiency. [12] J. Pan, Z. Hu, Z. Su, and M.-H. Yang, “Deblurring text images via L0-
regularized intensity and gradient prior,” in Proc. IEEE Conf. Comput.
V. C ONCLUSION AND F UTURE W ORK Vis. Pattern Recognit., Jun. 2014, pp. 2901–2908.
In this paper, we propose a structure-aware motion [13] S. Ramakrishnan, S. Pachori, A. Gangopadhyay, and S. Raman, “Deep
generative filter for motion deblurring,” in Proc. IEEE Int. Conf. Comput.
deblurring method based on a multi-adversarial optimized Vis. Workshops (ICCVW), Oct. 2017, pp. 2993–3000.
CycleGAN model. Unlike previous work, our CycleGAN [14] O. Kupyn, V. Budzan, M. Mykhailych, D. Mishkin, and J. Matas,
based method can avoid the error of the kernel estimation and “DeblurGAN: Blind motion deblurring using conditional adversarial
networks,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit.,
does not need the paired training data to make the training Jun. 2018, pp. 8183–8192.
more flexible. In addition, the multi-adversarial constraints in [15] X. Tao, H. Gao, X. Shen, J. Wang, and J. Jia, “Scale-recurrent network
the generator of CycleGAN we used are different from the for deep image deblurring,” in Proc. IEEE/CVF Conf. Comput. Vis.
Pattern Recognit., Jun. 2018, pp. 8174–8182.
traditional multi-scale manner to ensure that the results closest [16] N. T. Madam, S. Kumar, and A. N. Rajagopalan, “Unsupervised
to sharpening images are generated at different resolutions. class-specific deblurring,” in Proc. Eur. Conf. Comput. Vis. Cham,
Besides, we introduce a structure-aware method based on Switzerland: Springer, 2018, pp. 353–369.
[17] T. Kim, M. Cha, H. Kim, J. K. Lee, and J. Kim, “Learning to discover
edge clues so that the generated deblurred image can keep cross-domain relations with generative adversarial networks,” in Proc.
more structural information as much as possible. Extensive Int. Conf. Mach. Learn., 2017, pp. 1857–1865.
experiments on the different benchmark datasets demonstrate [18] J. Johnson, A. Alahi, and F.-F. Li, “Perceptual losses for real-time style
the effectiveness of the method we proposed. In the future, transfer and super-resolution,” in Proc. Eur. Conf. Comput. Vis., 2016,
pp. 694–711.
we are committed to solving the problem of significant target [19] Z. Wang, E. P. Simoncelli, and A. C. Bovik, “Multiscale structural
deblurring and further reducing the complexity of the network. similarity for image quality assessment,” in Proc. 37th Asilomar Conf.
Besides, we will further explore an unsupervised motion Signals, Syst. Comput., vol. 2, 2003, pp. 1398–1402.
[20] Y. Gan, X. Xu, W. Sun, and L. Lin, “Monocular depth estimation with
blur method with better performance and apply the proposed affinity, vertical pooling, and label enhancement,” in Proc. Eur. Conf.
network model to the video deblurring problem. Comput. Vis., 2018, pp. 232–247.
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
6154 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 30, 2021
[21] C. J. Schuler, H. C. Burger, S. Harmeling, and B. Schölkopf, “A machine [47] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation
learning approach for non-blind image deconvolution,” in Proc. IEEE with conditional adversarial networks,” in Proc. IEEE Conf. Comput. Vis.
Conf. Comput. Vis. Pattern Recognit., Jun. 2013, pp. 1067–1074. Pattern Recognit. (CVPR), Jul. 2017, pp. 5967–5976.
[22] S. Oh and G. Kim, “Robust estimation of motion blur kernel using a [48] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image
piecewise-linear model,” IEEE Trans. Image Process., vol. 23, no. 3, quality assessment: From error visibility to structural similarity,” IEEE
pp. 1394–1407, Mar. 2014. Trans. Image Process., vol. 13, no. 4, pp. 600–612, Apr. 2004.
[23] P. Chandramouli, M. Jin, D. Perrone, and P. Favaro, “Plenoptic image [49] K. Simonyan and A. Zisserman, “Very deep convolutional networks for
motion deblurring,” IEEE Trans. Image Process., vol. 27, no. 4, large-scale image recognition,” in Proc. Int. Conf. Learn. Represent.,
pp. 1723–1734, Apr. 2018. 2015, pp. 1–14.
[24] Y. Wen, B. Sheng, P. Li, W. Lin, and D. D. Feng, “Deep color guided [50] A. Paszke et al., “Automatic differentiation in PyTorch,” in Proc. Neural
coarse-to-fine convolutional network cascade for depth image super- Inf. Process. Syst. Workshop, 2017, pp. 1–4.
resolution,” IEEE Trans. Image Process., vol. 28, no. 2, pp. 994–1006, [51] D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimiza-
Feb. 2019. tion,” in Proc. Int. Conf. Learn. Represent., 2015, pp. 1–15.
[25] R. Fergus, B. Singh, A. Hertzmann, S. T. Roweis, and W. T. Freeman, [52] M. Hradiš, J. Kotera, P. Zemčík, and F. Šroubek, “Convolutional neural
“Removing camera shake from a single photograph,” ACM Trans. networks for direct text deblurring,” in Proc. Brit. Mach. Vis. Conf.,
Graph., vol. 25, no. 3, pp. 787–794, Jul. 2006. 2015, pp. 6:1–6:13.
[26] Q. Shan, J. Jia, and A. Agarwala, “High-quality motion deblurring from [53] Z. Liu, P. Luo, X. Wang, and X. Tang, “Deep learning face attributes
a single image,” ACM Trans. Graph., vol. 27, no. 3, pp. 73:1–73:10, in the wild,” in Proc. IEEE Int. Conf. Comput. Vis., Dec. 2015,
2008. pp. 3730–3738.
[27] L. Xu and J. Jia, “Two-phase kernel estimation for robust motion [54] W.-S. Lai, J.-B. Huang, Z. Hu, N. Ahuja, and M.-H. Yang, “A com-
deblurring,” in Proc. Eur. Conf. Comput. Vis., 2010, pp. 157–170. parative study for single image blind deblurring,” in Proc. IEEE Conf.
[28] D. Krishnan, T. Tay, and R. Fergus, “Blind deconvolution using a Comput. Vis. Pattern Recognit., Jun. 2016, pp. 1701–1709.
normalized sparsity measure,” in Proc. CVPR, Jun. 2011, pp. 233–240. [55] R. Köhler, M. Hirsch, B. Mohler, B. Schölkopf, and S. Harmeling,
[29] L. Xu, S. Zheng, and J. Jia, “Unnatural L0 sparse representation for “Recording and playback of camera shake: Benchmarking blind decon-
natural image deblurring,” in Proc. IEEE Conf. Comput. Vis. Pattern volution with a real-world database,” in Proc. Eur. Conf. Comput. Vis.,
Recognit., Jun. 2013, pp. 1107–1114. 2012, pp. 27–40.
[30] L. Sun, S. Cho, J. Wang, and J. Hays, “Edge-based blur kernel estimation [56] O. Kupyn, T. Martyniuk, J. Wu, and Z. Wang, “DeblurGAN-v2: Deblur-
using patch priors,” in Proc. IEEE Int. Conf. Comput. Photography ring (orders-of-magnitude) faster and better,” in Proc. IEEE Int. Conf.
(ICCP), Apr. 2013, pp. 1–8. Comput. Vis., Oct. 2019, pp. 8877–8886.
[31] J. Pan, D. Sun, H. Pfister, and M.-H. Yang, “Deblurring images via dark
channel prior,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 40, no. 10,
pp. 2315–2328, Oct. 2018.
[32] T. H. Kim and K. M. Lee, “Segmentation-free dynamic scene deblur-
ring,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2014, Yang Wen received the M.Eng. degree in com-
pp. 2766–2773. puter science from Xidian University, Xi’an, China,
[33] Y. Bai, H. Jia, M. Jiang, X. Liu, X. Xie, and W. Gao, “Single-image in 2015. She is currently pursuing the Ph.D. degree
blind deblurring using multi-scale latent structure prior,” IEEE Trans. in computer science with the Department of Com-
Circuits Syst. Video Technol., vol. 30, no. 7, pp. 2033–2045, Jul. 2020. puter Science and Engineering, Shanghai Jiao Tong
[34] J. Sun, W. Cao, Z. Xu, and J. Ponce, “Learning a convolutional neural University, Shanghai, China. Her current research
network for non-uniform motion blur removal,” in Proc. IEEE Conf. interests include motion deblurring, convolutional
Comput. Vis. Pattern Recognit. (CVPR), Jun. 2015, pp. 769–777. neural networks, image/video processing, and com-
[35] D. Ren, W. Zuo, D. Zhang, J. Xu, and L. Zhang, “Partial deconvolution puter graphics.
with inaccurate blur kernel,” IEEE Trans. Image Process., vol. 27, no. 1,
pp. 511–524, Jan. 2018.
[36] D. Gong et al., “From motion blur to motion flow: A deep learning
solution for removing heterogeneous motion blur,” in Proc. IEEE Conf.
Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 3806–3815.
[37] C. Ledig et al., “Photo-realistic single image super-resolution using
a generative adversarial network,” in Proc. IEEE Conf. Comput. Vis. Jie Chen received the B.Eng. degree in com-
Pattern Recognit. (CVPR), Jul. 2017, pp. 105–114. puter science from Nanjing University, Nanjing,
[38] D. Engin, A. Genç, and H. K. Ekenel, “Cycle-Dehaze: Enhanced Cycle- China. She is currently a Senior Chief Engineer and
GAN for single image Dehazing,” in Proc. IEEE/CVF Conf. Comput. Senior Architect with Samsung Electronics (China)
Vis. Pattern Recognit. Workshops (CVPRW), Jun. 2018, pp. 938–946. Research and Development Centre, Nanjing. She is
[39] L. Xu, J. S. J. Ren, C. Liu, and J. Jia, “Deep convolutional neural also the Head of the AI Department. Her current
network for image deconvolution,” in Proc. Neural Inf. Process. Syst., research interests include computer vision and big
2014, pp. 1790–1798. data.
[40] L. Li, J. Pan, W.-S. Lai, C. Gao, N. Sang, and M.-H. Yang, “Dynamic
scene deblurring by depth guided model,” IEEE Trans. Image Process.,
vol. 29, pp. 5273–5288, 2020.
[41] B. Lu, J.-C. Chen, and R. Chellappa, “Unsupervised domain-specific
deblurring via disentangled representations,” in Proc. IEEE Conf. Com-
put. Vis. Pattern Recognit., Jun. 2019, pp. 10217–10226.
[42] Q. Yuan, J. Li, L. Zhang, Z. Wu, and G. Liu, “Blind motion deblurring Bin Sheng (Member, IEEE) received the B.A.
with cycle generative adversarial networks,” Vis. Comput., vol. 36, no. 8, degree in English and the B.Eng. degree in computer
pp. 1591–1601, Aug. 2020. science from the Huazhong University of Science
[43] B. Lu, J.-C. Chen, and R. Chellappa, “UID-GAN: Unsupervised image and Technology, Wuhan, China, in 2004, the M.Sc.
deblurring via disentangled representations,” IEEE Trans. Biometrics, degree in software engineering from the University
Behav., Identity Sci., vol. 2, no. 1, pp. 26–39, Jan. 2020. of Macau, Taipa, Macau, in 2007, and the Ph.D.
[44] L. Wang, V. Sindagi, and V. Patel, “High-quality facial photo-sketch degree in computer science and engineering from
synthesis using multi-adversarial networks,” in Proc. 13th IEEE Int. The Chinese University of Hong Kong, Sha Tin,
Conf. Autom. Face Gesture Recognit. (FG ), May 2018, pp. 83–90. Hong Kong, in 2011. He is currently a Full Professor
[45] Z. Fu, Y. Zheng, H. Ye, Y. Kong, J. Yang, and L. He, “Edge-aware deep with the Department of Computer Science and Engi-
image deblurring,” CoRR, vol. abs/1907.02282, pp. 1–9, Jul. 2019. neering, Shanghai Jiao Tong University, Shanghai,
[46] Y. Bahat and M. Irani, “Blind Dehazing using internal patch recurrence,” China. His current research interests include virtual reality and computer
in Proc. IEEE Int. Conf. Comput. Photography (ICCP), vol. 8691, graphics. He is an Associate Editor of IEEE T RANSACTIONS ON C IRCUITS
May 2016, pp. 783–798. AND S YSTEMS FOR V IDEO T ECHNOLOGY .
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.
WEN et al.: STRUCTURE-AWARE MOTION DEBLURRING USING MULTI-ADVERSARIAL OPTIMIZED CycleGAN 6155
Zhihua Chen received the Ph.D. degree in com- Ping Tan (Senior Member, IEEE) received the Ph.D.
puter science from Shanghai Jiao Tong University, degree in computer science and engineering from
Shanghai, China, in 2006. He is currently a Full Pro- The Hong Kong University of Science and Technol-
fessor with the Department of Computer Science and ogy, Clear Water Bay, Hong Kong, in 2007. He is
Engineering, East China University of Science and currently an Associate Professor with the School
Technology, Shanghai. His current research inter- of Computing Science, Simon Fraser University,
ests include image/video processing and computer Burnaby, BC, Canada. His current research interests
vision. include computer vision and computer graphics.
He has served as an Area Chair for IEEE CVPR,
ACM SIGGRAPH, and ACM SIGGRAPH Asia.
He has served as an Editorial Board Member of
IEEE T RANSACTIONS ON PATTERN A NALYSIS AND M ACHINE I NTELLI -
GENCE and the International Journal of Computer Vision.
Authorized licensed use limited to: National Cheng Kung Univ.. Downloaded on July 20,2021 at 04:18:24 UTC from IEEE Xplore. Restrictions apply.