PDF文库 - 千万精品文档,你想要的都能搜到,下载即用。

◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf

What is love11 页 8.583 MB下载文档
◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf
当前文档共11页 2.88
下载后继续阅读

◆ 朱振峰, 孟莹莹, 孔德强, 张幸幸, Yandong, 赵耀, “To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene,” IEEE Transactions on Circuits and Systems for Video Yechnology , vol. 31, no. 2, pp. 492-502, Feb. 2021..pdf

492 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 31, NO. 2, FEBRUARY 2021 To See in the Dark: N2DGAN for Background Modeling in Nighttime Scene Zhenfeng Zhu , Yingying Meng, Deqiang Kong, Xingxing Zhang , Yandong Guo, and Yao Zhao , Senior Member, IEEE Abstract— Due to the deteriorated conditions of illumination lack and uneven lighting, the performance of traditional background modeling methods is greatly limited for the surveillance of nighttime video. To make background modeling under nighttime scene performs as well as in daytime condition, we put forward a promising generation-based background modeling framework for foreground surveillance. With a pre-specified daytime reference image as background frame, the GAN based generation model, called N2DGAN, is trained to transfer each frame of nighttime video to a virtual daytime image with the same scene to the reference image except for the foreground part. Specifically, to balance the preservation of background scene and the foreground object(s) in generating the virtual daytime image, we presented a two-pathway generation model, in which the global and local sub-networks were well combined with spatial and temporal consistency constraints. For the sequence of generated virtual daytime images, a multi-scale Bayes model was further proposed to characterize pertinently the temporal variation of background. We manually labeled ground truth on the collected nightime video datasets for performance evaluation. The impressive results illustrated in both the main paper and supplementary show the effectiveness of our proposed approach. Index Terms— GAN, background model, foreground detection, Bayes theory. I. I NTRODUCTION ACKGROUND modeling originates in numerous applications, especially in visual surveillance [1]–[7]. In the last decades, state-of-the-art approaches for background modeling have been proposed for visual surveillance under daytime scenes. On a whole, they are popularly dominated by a family B Manuscript received November 9, 2019; revised March 5, 2020; accepted March 26, 2020. Date of publication April 15, 2020; date of current version February 4, 2021. This work was supported in part by the Science and Technology Innovation 2030—New Generation Artificial Intelligence Major Project under Grant 2018AAA0102101, in part by the National Natural Science Foundation of China under Grant 61976018 and Grant 61532005, and in part by the Fundamental Research Funds for the Central Universities under Grant 2018JBZ001. This article was recommended by Associate Editor W.-H. Peng. (Corresponding author: Zhenfeng Zhu.) Zhenfeng Zhu, Yingying Meng, Xingxing Zhang, and Yao Zhao are with the Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China, and also with the Beijing Key Laboratory of Advanced Information Science and Network Technology, Beijing 100044, China (e-mail: zhfzhu@bjtu.edu.cn; mengyingying@bjtu.edu.cn; zhangxing@ bjtu.edu.cn; yzhao@bjtu.edu.cn). Deqiang Kong is with Microsoft Multimedia, Beijing 100080, China (e-mail: kodeqian@microsoft.com). Yandong Guo is with the OPPO Research Institute, Beijing 100026, China (e-mail: guoyandong@oppo.com). Color versions of one or more of the figures in this article are available online at https://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TCSVT.2020.2987874 Fig. 1. Flowchart of three kinds of background modeling methods for foreground object detection in nighttime surveillance video. (a) Conventional. (b) Enhancement-based. (c) Generation-based. of statistical based methods, like GMM [4] and KDE [2]. Besides, Codebook [3] and ViBe [1] are also two representative methods that achieve good performance for modeling the background. Most recently, several deep learning based works [5]–[7] for background modeling were also proposed. Despite the achievements by these approaches, all of them face quite a challenge in the case of illumination lack and uneven lighting at night, especially in the presence of dynamic background, change of light, and some extreme weather conditions such as rain, snow and fog. As we can see from Fig.1 (a), the conventional background modeling methods, like GMM, etc., fail to distinguish the foreground object from background due to the deteriorated condition of illumination lack. To deal with such a case, an intuitive way as shown in Fig.1 (b) is to perform image enhancement through E(·) first, and then build background model H E (·) on the bases of the enhanced frames E tn ’s, just like background modeling under daytime scene. However, since these enhancement methods [8]–[10] are not task-driven, they usually lose sight of pixel-wise consistency of inter-frame, and yet it is of great significance for background modeling. Although the captured images can be brightened by equipping a light with the camera, there are still deficiencies in the use of these devices. First, it will inevitably increase the cost of surveillance. Second, compared with the imaging under the natural light in the daytime, the visual quality by this way is still unsatisfactory. For example, some problems including serious noise, blurring, and unbalanced illumination will further bring difficulties for some vision processing tasks including object detection. In addition, it also won’t work well for monitoring distant object. 1051-8215 © 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See https://www.ieee.org/publications/rights/index.html for more information. Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. ZHU et al.: TO SEE IN THE DARK: N2DGAN FOR BACKGROUND MODELING IN NIGHTTIME SCENE To address this issue, we make a novel contribution in integrating generative model into background modeling. Fig.1 (c) shows the proposed generation-based background modeling framework. With a pre-specified daytime reference image I d as ground-truth background frame, the generation model G(·) is trained for transferring each frame Itn of nighttime video to a virtual daytime image G nt with the same scene to the reference image except for the foreground region. Furthermore, the background model HG (·) can be built to obtain FtG with the detected foreground object. In fact, the unique reference image plays a significant role for enforcing the pixel-wise temporal consistency of inter-frames in the generation of virtual daytime images. To the best of our knowledge, this paper is one of the first attempts to introduce GANs based deep learning network for background modeling. In summary, the following points highlight several contributions of the paper: • This paper proposes a reasonable and innovative solution, i.e., N2DGAN, to the longstanding problem of foreground object detection under nighttime scene. As a promising generation-based framework, it makes background modeling work as well as in daytime condition. • To simultaneously preserve background scene and the foreground object(s) in generating the virtual daytime image, we present a two-pathway generation model, in which the global and local sub-networks are seamlessly combined with spatial and temporal consistency constraints. • For the sequence of generated virtual daytime images, a multi-scale Bayes model is proposed to characterize pertinently the temporal variation of background. Thus, while suppressing effectively noise coming from virtual daytime image generation, we can ensure the favorable detection of foreground objects. • We collect a benchmark dataset including indoor and outdoor scenes with manually labeled ground truth, which can serve as a good benchmark for the research community. II. R ELATED W ORK A. Nighttime Image Enhancement Here we simply divide image enhancement methods into two categories: reference based and non-reference based methods. Non-reference based methods mainly focus on how to improve low contrast images. As a naive method, Histogram Equalization (HE) [9] spreads out the most frequent intensity values, thus gaining a higher contrast for the areas of lower contrast. The purpose of Retinex based image enhancement [MSR] [8] is to estimate the illumination from original image, thereby decomposing reflectance image and eliminating the influence of uneven illumination. In recent years, some deep learning based low light image enhancement approaches were also proposed, such as LIME [44], LLNet [45], and Struct [46]. Although they generally can achieve better perceptual quality than HE and MSR, they highly depend on the amount of training data. 493 Reference based methods [11]–[13] usually combine images of a scene at different time intervals by image fusion. These methods usually produce unnatural effects in the enhanced images. Besides, it would increase signal to noise ratio, which is adverse for further video analysis and applications such as foreground detection. B. Generative Adversarial Nets As a novel way to train generative models, GANs [14] proposed by Goodfellow et al. has received extensive applications in various of visual tasks [15]–[20]. In [15], GANs was applied for image completion with globally and locally consistent adversarial training. Reference [16] used back-propagation on a pretrained image generative network for image inpainting. To transfer the original image into a cartoon style, domain transfer network (DTN) [17] was proposed. In [18], [19], GANs has been employed for image super-resolution and image deblurring. Recently, a general-purpose solution to image-to-image translation based on conditional adversarial networks, also known as pixel2pixel network, was proposed in [20], and shows good performances on a variety of tasks like photo generation and semantic segmentation. In our previous work [21], a generative adversarial networks (GANs) based framework for nighttime image enhancement was proposed. C. Background Modeling Algorithms Broadly speaking, background modeling methods can be divided into two categories: pixel-based methods and block-based methods. One of the most popular pixel-based methods is Gaussian mixture models(GMM) [4], [22], [23]. It models the distribution at each pixel observed over time using a summation of weighted Gaussian distribution. Such methods generally perform well with the multi-modal nature of many practical situations. However, if high or low frequency changes appear in the background, the model can’t be adaptively tuned in time and even may miss some information about fast moving objects. Consequently, Elgammal et al. [24] have developed a non-parametric background model, which estimates the probability of observing pixel intensity values based on a sample of intensity values for each pixel. Different from GMM and KDE, some deep learning based works [5]–[7] for background modeling have also been proposed in recent years. But these models are all supervised and require many manually labeled data for model training. Thus, they obviously lack of scalability to a scene unknown beforehand, which also means their performances greatly depend on the collected training datasets. Block-based methods [25], [26] divide each frame into multiple overlapped or non-overlapped small blocks, and then model the background using the features of each block. Compared with pixel-based methods, the image blocks can capture more spatial distribution information, which makes block-based methods insensitive to the local shift in the background. However, the detection performance will largely depend on the block-dividing technique, especially for small moving targets. Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. 494 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 31, NO. 2, FEBRUARY 2021 Fig. 2. The network architecture of N2DGAN. I n is the input nighttime image, and we divide it into M blocks (Bin , i = 1, 2, . . . , M) as the input of each local generator. Then, the output of global generator and local generators are concatenated together. Finally, after two convolutional layers, the output is the daytime image G θG (I n ), where I d is the reference daytime image. More details about our model architecture are provided in the appendix. III. N IGHTTIME TO DAYTIME G ENERATIVE A DVERSARIAL N ETWORKS (N2DGAN) To maintain spatial and temporal consistency in the generation process, our goal is to train a generation model G(·) as in Fig.1 to transfer each frame Itn of nighttime video to a virtual daytime image with the same scene to the unique reference image I d except for the foreground region. Specifically, this generation problem can be formulated as: 1  L(G θG (Itn ), I d ) N N θ̂G = argmi n θG (1) t =1 where N is the number of training pairs, L(·, ·) denotes a weighted combination of several loss components. For the intent of learning the generation function G(·), the GANs is applied [14] due to its powerful generating ability. In particular, we propose a two path-way network N2DGAN with a generator network G θG and a discriminator network Dθ D parameterized by θG and θ D , respectively. For the discriminator network Dθ D , we will have the following maximization problem given G θG : θ̂ D = argmax E I d ∼Pd [Dθ D (I d )] − E Itn ∼Pn [Dθ D (G θG (Itn ))] θ D ∈D (2) where Pd is the real daytime image distribution, and Pn is the nighttime image distribution. Here we adopt the same formulation for Eq.(2) as in WGAN [27], D is the set of 1-Lipschitz function, and weight clipping is utilized to enforce the Lipschitz constraint. In essence, Eq.(1) and Eq.(2) are the alternative optimizing of a min-max optimization problem jointly parameterized by θG and θ D . A. Architecture The overview on the network architecture of the proposed N2DGAN is shown in Fig.2. To leverage the preserving of background scene and foreground object(s) in generating virtual daytime image, a two-pathway generator is proposed with global sub-network for maintaining background scene and M local sub-networks attending to capture local foreground information. As illustrated in Fig.2, both the global and local sub-networks are designed in an Encoder-Decoder manner as in most cases with modules of norm Convolution-BatchNorm-Relu, and each layer is followed by three residual blocks [28]. Following the architecture of “U-Net” adopted in [20], a fusion subnet is also designed to connect both the “Encoder” and the “Decoder” since symmetric layers can share some common information. This will be helpful for facilitating the information flow of foreground object between the input and output in the network chain. The details of the model architecture are provided in the Appendix. B. Loss Function The loss function L(·, ·) in Eq.(1) plays a significant role in training GANs model. For an input nighttime image Itn , several kinds of loss functions are exploited to make the generated virtual daytime image G θG (Itn ) retain most of the image information such as structure, objects, and texture as in the pre-specified reference image Id . 1) Adversarial Loss: To encourage the generated images move towards the real daytime image manifold and generate images with more details, the adversarial loss is first considered for distinguishing the generated image G θG (I n ) from the daytime image I d .    1  −Dθ D G θG Itn N N L adv = (3) t =1 where N is the number of nighttime images in the training dataset. Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. ZHU et al.: TO SEE IN THE DARK: N2DGAN FOR BACKGROUND MODELING IN NIGHTTIME SCENE 495 Fig. 3. Multi-scale Bayes inference framework for foreground detection. In the training phase, phase, for input frame Itn , multi-scale daytime images [G θG (Itn )]si , i = 1, 2, . . . Q, are generated, then we model the temporal distribution of each pixel of the generated image at each scale. D (si ) is discriminator θD network for scale si . In the testing phase, that is, foreground detection phase, for the input frame Ttn , the pre-trained N2DGAN network outputs its corresponding multi-scale daytime images [G θG (Ttn )]si , i = 1, 2, . . . Q. Then the multiple probabilities of each pixel belonging to the background at each scale are integrated elegantly based on Bayes inference for detecting the foreground object. 2) Perceptual Loss: In order to minimize the high-level  perceptual and semantic differences between G θG Itn and I d while preventing unexpected overfitting coming from I d , we follow the idea that minimizes the difference in convolutional layer of a pre-trained network [29] between two images. The motivation behind it lies in that the neural network pre-trained by image classification task has already learnt effective representation, which can be transferred into other tasks such as our enhancement processing. Specifically, we define φi as the activation of the i t h convolutional layer of the pre-trained network, and the perceptual loss is defined as: i  i  2 1  1 1  [φi (G θG (Itn ))]x,y − [φi (I d )]x,y Lp= C Wi Hi C W i=1 x=1 y=1 H (4) where C is the number of convolutional layers, Wi and Hi describe the dimensions of the respective feature maps within the VGG network. 3) Pixel-Wise Loss: To facilitate further background modeling task with spatial consistency of intra-frame and pixel-wise temporal consistency of inter-frame, the most widely used pixel-wise MSE loss (Eq.(5)) and total variation loss (Eq.(6)) are also adopted. W H   2 1 1  d (5) Ii, j − [G θG Itn ]i, j W H i=1 j =1  W,H  2  2  Ltv = ( Iˆtd )i+1, j − ( Iˆtd )i, j + ( Iˆtd )i, j+1 −( Iˆtd )i, j L mse = i=1, j =1 (6) where Iˆtd denotes the generated virtual daytime image G θG (Itn ). Since each of the loss functions mentioned above is provided with an unique view on characterizing the visual quality of the generated virtual image, an intuitive way is to make a combination of them. Thus, we have the final overall loss function as: L = λadv L adv + λt v L t v + λmse L mse + λ p L p (7) where λadv , λt v , λmse , and λ p are weights of the corresponding terms, respectively. IV. M ULTI -S CALE BAYES I NFERENCE FOR F OREGROUND D ETECTION N2DGAN ensures that there is a detectable difference between foreground object and background. All these characters match the major premise of GMM, that the background is more frequently visible than the foreground and that its variance is significantly slight. However, as we know, the neural network has the properties of both randomness and uncertainty. Thus, there exists inevitably pixel-wise difference between Ii,d j and [G θG (I n )]i, j in generating G θG (I n ), and it essentially can be regarded as some kind of random noise arising from both spatial and temporal domains. In other words, given the total error value, this difference at pixel (i, j ) may also occur at any other pixels with equal probability. This case will be doomed to bring some unexpected negative influence on pixel-level background modeling. To mitigate this issue, inspired by some works on multi-scale multiplication for edge detection [37], [38] that tend to yield significant localized detection, we extend N2DGAN to a multi-scale generative model as shown in Fig.3 to facilitate the background modeling to be noise-free. Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. 496 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 31, NO. 2, FEBRUARY 2021 TABLE I A. Multi-Scale Generation T HE D ETAILS OF OUR F OUR D ATASETS As illustrated in Fig.3, we reformulate the generation problem for background modeling as follows: to train a generator network G θG parametrized by θG , which learns a mapping function from the source domain n of nighttime to the target domain d of daytime. For every input nighttime frame Itn ∈ n , to generate a multi-scale set of images [G θG (Itn )]si , i = 1, . . . . . . , Q, in daytime domain d will be equivalent to: θˆG =  1 argmi n L([G θG (Itn )]si , [I d ]si ) N θG N Q (8) t =1 i=1 where N is the number of training pairs as before, and L(·, ·) denotes the loss function as mentioned above. In addition, we introduce Q adversarial discriminators to distinguish the reference daytime image [I d ]si from the generated virtual daytime image [G θG (Itn )]si under scale si . Particularly, similar to Eq.(2), in each generation task of different scales, the discriminator network will be reformulated as: (si ) = argmax E [I d ]si ∼Pd [D (si ) ([I d ]si )] θ̂ D θD (s ) θ D i ∈D −E I n ∼Pn [D (si ) ([G θG (I n )]si )] (9) θD It should be noted that both the generative network architecture and the discriminator architecture in Fig.3 are same as those in Fig.2. But different from N2DGAN, more convolutional layers are employed to generate multi-scale daytime images. B. Multi-Scale Bayes Inference Based on Scale Multiplication N2DGAN enforces the sequence of generated virtual daytime images G θG (Itn ), t = 1, . . . , N, to be as approximated closely as possible to the pre-specified unique reference frame I d . The inescapable fact, however, is that there is certain difference between them, one is noise accompanied by the neural network, the other part is the foreground region. To suppress effectively noise coming from virtual daytime generation while strengthening the discriminant of foreground object, a multiscale Bayes model is proposed to characterize pertinently the temporal variation of background. For each pixel Ii,n j , we use P(B | Ii,n j ) to serve as the background model, denoting the probability of pixel Ii,n j to be background. Given the multi-scale representations s [G θG (I n )]i,kj , k = 1, . . . , Q, for pixel Ii,n j , the background model P(B | Ii,n j ) can be given with Bayes criterion by Eq.(10). Here, · represents rounding down to the nearest whole number. On the assumption that the generation of virtual daytime images with different scales is independent of each other, thus the background model given by Eq.(10), as shown at the bottom of this page, will further reduce to the following Eq.(11), as shown at the bottom of this page. As we can see from Eq.(11), the background model P(B | Ii,n j ) is equivalent to the multi-scale multiplication of multiple background models at different scales. In addition, s for each background model P(B | [G θG (I n )]ki/2k−1 , j/2k−1  ), k = 1, . . . , Q, a single gaussian model instead of GMM can be simply applied as shown in Fig.3. V. E XPERIMENTAL R ESULTS We evaluate the proposed background modeling approach visually and quantitatively, by comparing with state-of-the-arts and providing extensive ablation studies A. Datasets and Experiment Settings 1) Datasets and Evaluation Metrics: Our work in this paper mainly focuses on background modeling under nighttime scene with low illumination. However, to the best of our knowledge, there are no public open datasets to evaluate such a task. For this reason, we collect several benchmark datasets by a Canon IXY 210F video camera including indoor and outdoor scenes with manually labeled ground truth. The details about our four datasets, including Lab, Tree, Lake1, and Lake2, are shown in Tab.I.1 For each dataset, the corresponding pre-specified daytime images that serve as ground truth background frames are also provided. It is worth of noting that both the ‘Lake’ and ‘Tree’ datasets were captured outdoor on windy 1 The datasets and code of our https://github.com/anqier0468/N2DGAN. s method will be released at s P(B | Ii,n j ) = P(B | [G θG (I n )]si,1j , . . . , [G θG (I n )]Qi/2 Q−1 , j/2 Q−1  )P([G θG (I n )]si,1j , . . . , [G θG (I n )]Qi/2 Q−1 , j/2 Q−1  | Ii,n j ) (10) P(B | Ii,n j ) = ∝ Q  k=1 Q  k=1 s s P(B | [G θG (I n )]ki/2k−1 , j/2k−1  )P([G θG (I n )]ki/2k−1 , j/2k−1  | Ii,n j ) P(B | [G θG (I n )]ski/2k−1 , j/2k−1  ) Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. (11) ZHU et al.: TO SEE IN THE DARK: N2DGAN FOR BACKGROUND MODELING IN NIGHTTIME SCENE 497 TABLE II Q UANTITATIVE C OMPARISON OF F OREGROUND D ETECTION A CCURACY BY D IFFERENT M ETHODS nights, and the ’Lab’ dataset is taken indoors where we control the intensity of lighting by pulling curtains and switching incandescent lights on purpose. Actually, these datasets are much challenging for background modeling task since they feature the undulation of lake, reflection of lights in the water, leaves shaking, and illumination variation. In order to make a quantitative evaluation, the foreground object(s) in the datasets are also manually labeled. Following the previous works on foreground detection, IoU is employed as our evaluation metric [39]. 2) Implementation Details: The training of the proposed N2DGAN model is implemented on 2 NVIDIA TITAN Xp GPUs. The first 300 frames of each nighttime video in Tab.I paired with the corresponding daytime reference image are used to train the model, and the remaining are for testing. All of the images are downscaled to resolution of 256 × 256. Specially, we split each image into multiple image blocks with size 32 × 32, and then each block is used as the input of each local generator subnet. Considering the computational efficiency, only two scales are adopted, i.e., Q = 2, to eliminate the influence of noise and spatial shift of background pixels. Based on RMSProp, the mini-batch gradient descent method is used with a batch size of 4 and a learning rate of 10−4 . Since WGAN [27] is used as the backbone of our generation model, the weight need to be clampped to a fixed box (−0.01, 0.01) after each gradient update to avoid gradient vanishing and mode collapse problems during the learning process. In all our experiments, we uniformly set the total training epoch number to 30, and empirically set λadv = 10−1 , λ p = 10−5 , λmse = 10−1 , and λt v = 10−3 in Eq.(7) to maintain the same order of magnitude. Besides, the first 3 convolutional layers of VGG network is used to calculate perceptual loss. B. Performance Evaluation 1) Comparisons With State-of-the-Arts: We compare the proposed method with state-of-the-arts on foreground detection, including eight typical background modeling methods and four enhancement-based methods. Implementations of all these methods are based on the BGSLibrary [40] with default parameters. The quantitative comparison results on 3 sequences are shown in Tab.II. Obviously, our method can always achieve the best performance. Specifically, for sequences ’Tree’ and ’Lake2’, our method even outperforms the state-of-the-art method SUBSENSE [33] by 75% and 44%, respectively. Fig.4 presents the qualitative comparison results, which illustratively show that the proposed N2DGAN performs better. This mainly lies in two facts: 1) compared with the directly background modeling methods, generating daytime images makes the flatten pixel distribution sharper and easier to detect foreground; 2) compared with the enhancement-based methods, the unique daytime reference frame in generative process ensures the inter-frame consistency of the generated daytime images. 2) Stability Comparison: Consistency stability of successive frames is of great importance for further background modeling. For the t-th frame It  R W ×H , we use the following metric to measure the stability between It and its adjacent frame It +1 . st = Si m(It , It +1 ) (12) where Si m(·, ·) denotes the distance between It and It +1 . Here, we adopt Kullback-Leibler Divergence, which represents stronger stability if its value is close to 0. Fig.5 shows the stability comparison of several representative methods on a randomly selected sequence consisting of 200 consecutive frames from the test set of three datasets. As we can see, the result of our method (yellow line) is quite close to real nighttime images (blue line), while both HE (red line) and MSR (green line) show distinct difference from real sequence. Particularly, the large fluctuation by MSR also indicates that the pixel values between two adjacent frames differ greatly, which is unfavorable for background modeling. To sum up, with the joint constrain of spatial and temporal Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. 498 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 31, NO. 2, FEBRUARY 2021 Fig. 4. Qualitative comparison of different foreground detection methods. (a) nighttime image, (b) groundtruth, (c) our method N2G-GAN, (d) GMG [35], (e) ASOM [32], (f) FASOM [31], (g) LOBSTER [30], (h) GMM [4], (i) MCueBGS [41], (j) SUBSENSE [33], (k) MRF-UV [36], (l) HE [9]+GMM [4], (m) MSR [8]+GMM [4], (n) HE [9]+SUBSENSE [33], (o) MSR [8]+SUBSENSE [33]. Fig. 5. Sequence stability comparison with HE, MSR and N2DGAN on Lab (left), Tree (middle), and Lake1 (right) datasets. randomly selected frames Lake1-13th and Lake1-158th, Fig. 7 and Fig. 8 present visually the comparison results with conventional background modeling methods and enhancement based methods, which clearly shows that the proposed N2DGAN achieves the best performance. For the experiments on evaluating the robustness to illumination variation, Fig. 9 illustrates the comparison results on the indoor nighttime dataset Lab with illumination variation. As we can see from Fig.9, the N2DGAN model is much more insensitive to the instantaneous changes in light compared with other state-of-the-art background modeling methods. Here, the enhancement result based on HE (Fig. 9(b)) is only utilized to clarify the foreground object since it is not easy to find the ground truth in dark. Two factors must be credited for our high resilience to noise and illumination change. The first originates from our model design, which allows noisy pixel to be outfitting to the reference background image. The second lies in our background model on successive multi-scale images, which is more robust to noise by hierarchical Bayes modeling. D. Ablation Study Fig. 6. Foreground detection accuracy on Lake1 dataset with different levels of noises. consistency, our generation based method performs better than enhancement-based method. C. Robustness to Noise and Illumination Variation Some extreme weather conditions such as rain, snow, and fog usually bring great challenges to background modeling. To demonstrate N2DGAN’s scalability under such environment, additive Gaussian noise is added to nighttime video sequence to simulate extreme weather. We randomize the noise standard deviation σ  {[0, 5] , [5, 10] , [10, 15] , [15, 20]} separately for each testing example. As illustrated in Fig.6, the behaviors are quantitatively different in all three datasets. This demonstrates that our method is the only technique that manages to perform well with different levels of noises. On two 1) Global and Local Consistency Evaluation: To verify the effectiveness of combining both local and global consistency together in our model, we first perform foreground detection when using global subnetwork alone, called N2DGAN(global). As observed from Fig.10, small targets in nighttime images are lost in this case. Meanwhile, when using local subnet alone, called N2DGAN(local), the detected foreground objects are incomplete on the edge of patches, since there exists blocking-artifact problem caused by patch enhancement. Additionally, quantitative comparison results shown in Tab.II (bottom) demonstrate that our baseline improves detection accuracy by more than 10% and 5% compared with N2DGAN(global) and N2DGAN(local), respectively. Evaluation on Multi-scale Bayes modeling To further demonstrate the effectiveness of our background modeling on successive multi-scale images of daytime domain, we attempt to perform on a single scale generated images [G θG (In )]s1 . As illustrated in Fig.11, due to the fact that multi-scale bayes model can suppress the noise caused by network, then Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. ZHU et al.: TO SEE IN THE DARK: N2DGAN FOR BACKGROUND MODELING IN NIGHTTIME SCENE 499 Fig. 7. Performance evaluation of robustness to noise on Lake1-13th. (a) The input nighttime image. (b) Groundtruth. (c) N2DGAN. (d) GMG [35], (e) IMBGS [34]. (f)ASOM [32], (g)FASOM [31], (h)LOBSTER [30], (i)GMM [4], (j)SUBSENSE [33], (k)MRF-UV [36], (l) HE [9]+GMM [4], (m) MSR [8]+GMM [4], (n) HE [9]+SUBSENSE [33], (o) MSR [8]+SUBSENSE [33]. Fig. 8. Performance evaluation of robustness to noise on Lake1-158th. (a) The input nighttime image. (b) Groundtruth. (c) N2DGAN. (d) GMG [35], (e) IMBGS [34]. (f)ASOM [32], (g)FASOM [31], (h)LOBSTER [30], (i)GMM [4], (j)SUBSENSE [33], (k)MRF-UV [36], (l) HE [9]+GMM [4], (m) MSR [8]+GMM [4], (n) HE [9]+SUBSENSE [33], (o) MSR [8]+SUBSENSE [33]. Fig. 9. Performance evaluation of robustness to illumination variation. (a) nighttime image, (b) HE, (c) N2DGAN, (d) GMG [35], (e) ASOM [32], (f) FASOM [31], (g) LOBSTER [30], (h) GMM [4], (i) MCueBGS [41], (j) SUBSENSE [33], (k) MRF-UV [36], (l) HE [9]+GMM [4], (m) MSR [8]+GMM [4], (n) HE [9]+SUBSENSE [33], (o) MSR [8]+SUBSENSE [33]. our baseline N2DGAN makes the foreground region more remarkable. E. Time Complexity Analysis For a background model, the computational complexity is one of the key issues worthy of attention. For our N2DGAN model, its computational complexity mainly consists of two parts, i.e., virtual daytime image generation and foreground detection. In the foreground detection stage, since our model holds only a single gaussian model which can be off-line available and without need for online model updating, thus the time complexity in this stage is much lower than the traditional GMM and can be negligible compared with the one in generation stage. Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. 500 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 31, NO. 2, FEBRUARY 2021 TABLE IV A RCHITECTURE OF THE L OCAL G ENERATOR S UB -N ETWORK Fig. 10. Comparison of foreground detection results using global subnet and local subnet independently. (a) Input nighttime image. (b) Ground truth. (c) N2DGAN. (d) N2DGAN(global). (e) N2DGAN(local). TABLE V A RCHITECTURE OF THE G LOBAL G ENERATOR S UB -N ETWORK Fig. 11. Comparison of foreground detection results using multi scales and single sale. (a) nighttime image. (b) Groundtruth. (c) Generated image [G θG (In )]s1 . (d) N2DGAN. (e) The detection result using single scale s1 . TABLE III T IME C OMPLEXITY A NALYSIS OF G ENERATION P ROCESS By comparison, to generate a virtual daytime image will occupy most of the time with a frame rate of 8 fps without any code optimisation. As shown in Table III, the frame rate of N2DGAN(local) to generate 8 × 8 blocks of local subimages2 is around 10 fps, which is much more slowly than N2DGAN(global) with a frame rate of 56 fps. However, considering that we can generate each block of local virtual sub-image in parallel, the frame rate of N2DGAN(local) will dramatically increased, approximating around 640 fps in an ideal situation. It means that the global generation process with 56 fps will dominate the overall computational complexity of 2 For an input nighttime image, it is divided into 8 × 8 blocks and then the corresponding local generation sub-network is trained on each block to generate a local sub-image. For details please refer to Section III. the virtual daytime image generation. By this way, the need for online real-time foreground object detection can be met. VI. C ONCLUSION For the challenge of background modeling under daytime scene, an innovative N2DGAN model is proposed, which paves a new way completely different from the existing methods. To the best of our knowledge, this is the first time to introduce GANs based deep learning for this practical problem. As an unsupervised model, N2DGAN is provided with good scalability and practical significance. As for the time complexity of N2DGAN, it takes about 0.125 seconds (8 fps) for each frame. Considering each local generation model can be implemented in parallel, the proposed N2DGAN could be highly parameterizable. Besides, some model compression works [42], [43] can also be feasible Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. ZHU et al.: TO SEE IN THE DARK: N2DGAN FOR BACKGROUND MODELING IN NIGHTTIME SCENE solutions for network acceleration. It is also worth noting that we have assumed that the images in the training set is free of foreground objects. To overcome this limitation, we will try to extend this method to deal with the situation that the images in the training set may contain some foreground objects. A PPENDIX The detailed structures of the global sub-network and local sub-network are provided in Table IV and Table V, respectively. Each convolution layer is followed by 3 residual block [28]. Then we simply concatenate the output from each local generator and the global generator to produce a fused feature tensor and then feed it to the successive convolution layers to generate the final output. R EFERENCES [1] O. Barnich and M. Van Droogenbroeck, “ViBe: A universal background subtraction algorithm for video sequences,” IEEE Trans. Image Process., vol. 20, no. 6, pp. 1709–1724, Jun. 2011. [2] A. Elgammal, R. Duraiswami, D. Harwood, and L. S. Davis, “Background and foreground modeling using nonparametric kernel density estimation for visual surveillance,” Proc. IEEE, vol. 90, no. 7, pp. 1151–1163, Jul. 2002. [3] K. Kim, T. H. Chalidabhongse, D. Harwood, and L. Davis, “Real-time foreground-background segmentation using codebook model,” RealTime Imag., vol. 11, no. 3, pp. 172–185, 2005. [4] Z. Zivkovic, “Improved adaptive Gaussian mixture model for background subtraction,” in Proc. 17th Int. Conf. Pattern Recognit. (ICPR), 2004, pp. 28–31. [5] L. A. Lim and H. Yalim Keles, “Foreground segmentation using convolutional neural networks for multiscale feature encoding,” Pattern Recognit. Lett., vol. 112, pp. 256–262, Sep. 2018. [6] L. Yang, J. Li, Y. Luo, Y. Zhao, H. Cheng, and J. Li, “Deep background modeling using fully convolutional network,” IEEE Trans. Intell. Transp. Syst., vol. 19, no. 1, pp. 254–262, Jan. 2018. [7] D. Zeng and M. Zhu, “Background subtraction using multiscale fully convolutional network,” IEEE Access, vol. 6, pp. 16010–16021, 2018. [8] D. J. Jobson, Z. Rahman, and G. A. Woodell, “A multiscale retinex for bridging the gap between color images and the human observation of scenes,” IEEE Trans. Image Process., vol. 6, no. 7, pp. 965–976, Jul. 1997. [9] D. J. Ketcham, “Real-time image enhancement techniques,” OSA Image Process., vol. 74, no. 2, pp. 120–125, 1976. [10] Y. Gong, Y. Lee, and T. Q. Nguyen, “Nighttime image enhancement applying dark channel prior to raw data from camera,” in Proc. Int. SoC Design Conf. (ISOCC), Oct. 2016, pp. 173–174. [11] Y. Cai, K. Huang, T. Tan, and Y. Wang, “Context enhancement of nighttime surveillance by image fusion,” in Proc. 18th Int. Conf. Pattern Recognit. (ICPR), 2006, pp. 980–983. [12] Y. Rao, “Image-based fusion for video enhancement of night-time surveillance,” Opt. Eng., vol. 49, no. 12, Dec. 2010, Art. no. 120501. [13] R. Raskar, A. Ilie, and J. Yu, “Image fusion for context enhancement and video surrealism,” in Proc. Int. Symp. Non-Photorealistic Animation Rendering, 2004, p. 4. [14] I. Goodfellow et al., “Generative adversarial nets,” in Proc. Adv. Neural Inf. Process. Syst., 2014, pp. 2672–2680. [15] S. Iizuka, E. Simo-Serra, and H. Ishikawa, “Globally and locally consistent image completion,” ACM Trans. Graph., vol. 36, no. 4, pp. 1–14, Jul. 2017. [16] R. Yeh et al., “Semantic image inpainting with perceptual and contextual losses,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 1–10. [17] Y. Taigman, A. Polyak, and L. Wolf, “Unsupervised cross-domain image generation,” in Proc. Int. Conf. Learn. Represent., 2016, pp. 1–8. [18] C. Ledig et al., “Photo-realistic single image super-resolution using a generative adversarial network,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 105–114. [19] S. Zheng, Z. Zhu, J. Cheng, Y. Guo, and Y. Zhao, “Edge heuristic GAN for non-uniform blind deblurring,” IEEE Signal Process. Lett., vol. 26, no. 10, pp. 1546–1550, Oct. 2019. 501 [20] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-Image translation with conditional adversarial networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 1125–1134. [21] Y. Meng, D. Kong, Z. Zhu, and Y. Zhao, “From night to day: GANs based low quality image enhancement,” Neural Process. Lett., vol. 50, no. 1, pp. 799–814, Aug. 2019. [22] P. Kaewtrakulpong and R. Bowden, “An improved adaptive background mixture model for real-time tracking with shadow detection,” in Video-Based Surveillance Systems. Berlin, Germany: Springer, 2002, pp. 135–144. [23] C. Stauffer and W. E. L. Grimson, “Adaptive background mixture models for real-time tracking,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., Jun. 1999, pp. 246–252. [24] A. M. Elgammal, D. Harwood, and L. S. Davis, “Non-parametric model for background subtraction,” in Proc. Eur. Conf. Comput. Vis., 2000, pp. 751–767. [25] M. Heikkila and M. Pietikainen, “A texture-based method for modeling the background and detecting moving objects,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no. 4, pp. 657–662, Apr. 2006. [26] S. Liao, G. Zhao, V. Kellokumpu, M. Pietikainen, and S. Z. Li, “Modeling pixel process with scale invariant local patterns for background subtraction in complex scenes,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., Jun. 2010, pp. 1301–1306. [27] M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein generative adversarial networks,” in Proc. Int. Conf. Mach. Learn., 2017, pp. 1–11. [28] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Dec. 2016, pp. 770–778. [29] L. A. Gatys, A. S. Ecker, and M. Bethge, “Image style transfer using convolutional neural networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Aug. 2016, pp. 2414–2423. [30] P.-L. St-Charles and G.-A. Bilodeau, “Improving background subtraction using local binary similarity patterns,” in Proc. IEEE Winter Conf. Appl. Comput. Vis., Mar. 2014, pp. 509–515. [31] L. Maddalena and A. Petrosino, “A fuzzy spatial coherence-based approach to background/foreground separation for moving object detection,” Neural Comput. Appl., vol. 19, no. 2, pp. 179–186, Mar. 2010. [32] L. Maddalena and A. Petrosino, “A self-organizing approach to background subtraction for visual surveillance applications,” IEEE Trans. Image Process., vol. 17, no. 7, pp. 1168–1177, Jul. 2008. [33] P.-L. St-Charles, G.-A. Bilodeau, and R. Bergevin, “Flexible background subtraction with self-balanced local sensitivity,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. Workshops, Jun. 2014, pp. 408–413. [34] D. Bloisi and L. Iocchi, “Independent multimodal background subtraction.,” in Proc. Comput. Modeling Objects Represented Images Fundam. Methods Appl., 2012, pp. 39–44. [35] A. B. Godbehere, A. Matsukawa, and K. Goldberg, “Visual tracking of human visitors under variable-lighting conditions for a responsive audio art installation,” in Proc. Amer. Control Conf., 2012, pp. 4305–4312. [36] Z. Zhao, T. Bouwmans, X. Zhang, and Y. Fang, “A fuzzy background modeling approach for motion detection in dynamic backgrounds,” in Proc. Int. Conf. Multimedia Signal Process., 2012, pp. 177–185. [37] Z. Zhu, H. Lu, and Y. Zhao, “Scale multiplication in odd Gabor transform domain for edge detection,” J. Vis. Commun. Image Represent., vol. 18, no. 1, pp. 68–80, Feb. 2007. [38] L. Zhang and P. Bao, “Edge detection by scale multiplication in wavelet domain,” Pattern Recognit. Lett., vol. 23, no. 14, pp. 1771–1784, Dec. 2002. [39] Z. Liu, K. Huang, and T. Tan, “Foreground object detection using top-down information based on EM framework,” IEEE Trans. Image Process., vol. 21, no. 9, pp. 4204–4217, Sep. 2012. [40] A. Sobral, “BGSLibrary: An OpenCV C++ background subtraction Library,” in Proc. IX Workshop Vis. Comput. (WVC), Rio de Janeiro, Brazil, 2013, pp. 1–8. [41] S. J. Noh and M. Jeon, “A new framework for background subtraction using multiple cues,” in Proc. Asian Conf. Comput. Vis., 2012, pp. 493–506. [42] K. Jia, D. Tao, S. Gao, and X. Xu, “Improving training of deep neural networks via singular value bounding,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 4344–4352. [43] X. Yu, T. Liu, X. Wang, and D. Tao, “On compressing deep models by low rank and sparse decomposition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jul. 2017, pp. 7370–7379. [44] X. Guo, Y. Li, and H. Ling, “LIME: Low-light image enhancement via illumination map estimation,” IEEE Trans. Image Process., vol. 26, no. 2, pp. 982–993, Feb. 2017. Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply. 502 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 31, NO. 2, FEBRUARY 2021 [45] K. G. Lore, A. Akintayo, and S. Sarkar, “LLNet: A deep autoencoder approach to natural low-light image enhancement,” Pattern Recognit., vol. 61, pp. 650–662, Jan. 2017. [46] M. Li, J. Liu, W. Yang, X. Sun, and Z. Guo, “Structure-revealing lowlight image enhancement via robust retinex model,” IEEE Trans. Image Process., vol. 27, no. 6, pp. 2828–2841, Jun. 2018. Zhenfeng Zhu received the M.E. degree from the Harbin Institute of Technology, Harbin, China, in 2001, and the Ph.D. degree from the National Laboratory of Pattern Recognition (NLPR), Institute of Automation, CAS, Beijing, China, in 2005, respectively. He was a Visiting Scholar with the Department of Computer Science and Engineering, Arizona State University, Tempe, AZ, USA, in 2010. He is currently a Professor with the School of Computer and Information Technology, Beijing Jiaotong University, Beijing. He has authored or coauthored more than 100 articles in journals and conferences, including the IEEE T RANSACTIONS ON K NOWLEDGE AND D ATA E NGINEERING (T-KDE), the IEEE T RANSACTIONS ON N EURAL N ETWORKS AND L EARNING S YSTEMS (T-NNLS), the IEEE T RANSACTIONS ON C IRCUITS AND S YSTEMS FOR V IDEO T ECHNOLOGY (T-CSVT), the IEEE T RANSACTIONS ON M ULTIMEDIA (T-MM), the IEEE T RANSACTIONS ON C YBERNETICS (T-CYB), the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), AAAI, IJCAI, and ACM Multimedia. His group has won the Honorable Award (Rank First in stage two) in KDD Cup2016 competition and Rank3 in CIKM Cup2016 competition, respectively. His current research interests include image and video understanding, computer vision, machine learning, and recommendation systems. Yingying Meng received the B.S. degree in computer science and technology from the Shandong University of Science and Technology, Qingdao, China, in 2016, and the M.E. degree in signal and information processing from the Institute of Information Science, Beijing Jiaotong University, Beijing, China, in 2019. She is currently a Software Engineer with the China CITIC Bank. Deqiang Kong received the M.E. degree in signal and information processing from the Institute of Information Science, Beijing Jiaotong University, Beijing, China, in 2018. He is currently an Applied Scientist with Microsoft, Beijing. His research interests include video understanding, natural language processing, and recommendation systems. Xingxing Zhang received the B.S. degree in communication engineering from Henan Normal University, Xinxiang, China, in 2015. She is currently pursuing the Ph.D. degree with the Institute of Information Science, Beijing Jiaotong University, Beijing, China. Her research interests include data analysis, image and video understanding, computer vision, and machine learning. Yandong Guo received the B.S. and M.S. degrees in ECE from the Beijing University of Posts and Telecommunications, China, in 2005 and 2008, receptively, and the Ph.D. degree in ECE from Purdue University at West Lafayette in 2013, under the supervision of Prof. Bouman and Prof. Allebach. He is currently the Chief Scientist of intelligent perception with the OPPO Research Institute. He also holds an Adjunct Professor position at the Beijing University of Posts and Telecommunications, and the University of Electronic Science and Technology of China. Before he joined OPPO in 2020, he was the Chief Scientist with XPeng Motors, China, and a Researcher with the Microsoft Research, Redmond, WA, USA. His professional interests lie in the broad area of computer vision, imaging systems, human behavior understanding and biometric, and autonomous driving. Yao Zhao (Senior Member, IEEE) received the B.S. degree from Fuzhou University, Fuzhou, China, in 1989, the M.E. degree from Southeast University, Nanjing, China, in 1992, and the Ph.D. degree from the Institute of Information Science, Beijing Jiaotong University (BJTU), Beijing, China, in 1996. From 2001 to 2002, he was a Senior Research Fellow with the Information and Communication Theory Group, Delft University of Technology, Delft, The Netherlands. Since 2001, he has been a Professor with BJTU, where he is currently the Director of the Institute of Information Science. His current research interests include image/video coding, digital watermarking and forensics, and video analysis and understanding. Dr. Zhao serves on the editorial boards of several international journals, including as an Associate Editor of the IEEE T RANSACTIONS ON C YBER NETICS , the IEEE S IGNAL P ROCESSING L ETTERS , and Circuits, Systems, and Signal Processing, and an Area Editor of Signal Processing: Image Communication. He was named as a Distinguished Young Scholar by the National Science Foundation of China in 2010 and was elected as a Chang Jiang Scholar of the Ministry of Education of China in 2013. Authorized licensed use limited to: Beijing Jiaotong University. Downloaded on December 28,2021 at 01:51:08 UTC from IEEE Xplore. Restrictions apply.

相关文章