0.8% Nyquist computational ghost imaging via non-experimental deep learning
0.8% 奈奎斯特计算鬼成像的非实验性深度学习实现
We present a framework for computational ghost imaging based on deep learning and customized pink noise speckle patterns. The deep neural network in this work, which can learn the sensing model and enhance image reconstruction quality, is trained merely by simulation. To demonstrate the subNyquist level in our work, the conventional computational ghost imaging results, reconstructed imaging results using white noise and pink noise via deep learning are compared under multiple sampling rates at different noise conditions. We show that the proposed scheme can provide highquality images with a sampling rate of 0.8 even when the object is outside the training dataset, and it is robust to noisy environments. This method is excellent for various applications, particularly those that require a low sampling rate, fast reconstruction efficiency, or experience strong noise interference.
我们提出了一种基于深度学习(deep learning)和定制粉红噪声散斑图案的计算鬼成像框架。本工作中的深度神经网络仅通过模拟训练就能学习传感模型并提升图像重建质量。为验证本工作的亚奈奎斯特采样水平,我们在不同噪声条件下的多种采样率场景中,对比了传统计算鬼成像结果、使用白噪声和粉红噪声通过深度学习重建的成像结果。实验表明,即使目标物体不在训练数据集中,该方案仍能在0.8%采样率下获得高质量图像,并对噪声环境具有强鲁棒性。该方法特别适用于需要低采样率、快速重建效率或面临强噪声干扰的各类应用场景。
I. INTRODUCTION
I. 引言
Ghost imaging (GI) [1–4] is an innovative method for measuring the spatial correlations between light beams. With GI, the signal light field interacts with the object and is collected by a single-pixel detector, and the reference light field, which does not interact with the object, falls onto the imaging detector. Therefore, the image information is not present in either beam alone but only revealed in their correlations. Computational ghost imaging (CGI) [5, 6] was proposed to further ameliorate and simplify this framework. In CGI, The reference arm that records the speckles is replaced by loading pre-generated patterns directly onto the spatial light modulator or the digital micro mirror device (DMD). The unconventional image is then revealed by correlating the sequentially recorded intensities at the single-pixel detector with the corresponding patterns. CGI finds a lot of applications such as wide spectrum imaging [7–9], remote sensing [10], and quantum-secured imaging [11].
鬼成像 (GI) [1–4] 是一种测量光束间空间相关性的创新方法。在鬼成像中,信号光场与物体相互作用后被单像素探测器收集,而未与物体相互作用的参考光场则投射到成像探测器上。因此,图像信息并不单独存在于任一光束中,而是仅体现在二者的相关性中。计算鬼成像 (CGI) [5, 6] 的提出进一步优化并简化了这一框架。在计算鬼成像中,记录散斑的参考臂被直接加载预生成图案的空间光调制器或数字微镜器件 (DMD) 所取代。随后,通过将单像素探测器依次记录的强度与对应图案相关联,即可重构出非常规图像。该技术在宽光谱成像 [7–9]、遥感 [10] 和量子安全成像 [11] 等领域具有广泛应用。
However, CGI generally requires a large number of samplings to reconstruct a high-quality image, or the signal would have been submerged under correlation fluctuations and environmental noise. To suppress the environmental noise and correlation fluctuations, the required minimum number of sampling is proportional to the total pixel number of the pattern applied on DMD, i.e., the Nyquist sampling limit [12, 13]. The image could have a meager quality with a limited sampling number. This demanding requirement hindered CGI from fully replacing conventional photography. A large number of schemes have been proposed to improve CGI’s speed and decrease the sampling rate (sub-Nyquist). For instance, compressive sensing imaging can reconstruct images with a relatively low sampling rate by exploiting the sparsity of the objects [14–17]. It nevertheless largely depends on the sparsity of objects and is sensitive to noise [18]. Ortho normalized noise patterns can be used to suppress the noise and improve the image’s quality under a limited sampling number [19, 20]. In particular, the ortho normalized colored noise patterns can break the Nyquist limit down to ∼5 [20]. Fourier and sequence-ordered Walsh-Hadamard patterns, which are orthogonal to each other in time or spatial domain, were also applied to the sub-Nyquist imaging [21–23]. The Russian doll [24] and cake-cutting [25] ordering of Walsh-Hadamard patterns can minimize the sampling ratio to 5%-10% Nyquist limit.
然而,计算鬼成像 (CGI) 通常需要大量采样才能重建高质量图像,否则信号会被相关波动和环境噪声淹没。为抑制环境噪声和相关波动,所需的最小采样数与数字微镜器件 (DMD) 上加载图案的总像素数成正比,即奈奎斯特采样极限 [12, 13]。在有限采样数下,图像质量可能极差。这一严苛要求阻碍了CGI完全取代传统摄影。目前已有大量方案被提出以提高CGI速度并降低采样率(亚奈奎斯特)。例如,压缩感知成像通过利用物体的稀疏性,能以较低采样率重建图像 [14-17],但其效果很大程度上依赖于物体的稀疏性且对噪声敏感 [18]。正交归一化噪声图案可用于抑制噪声,并在有限采样数下提升图像质量 [19, 20]。特别是正交归一化彩色噪声图案能将奈奎斯特极限降低至 ∼5 [20]。在时域或空域相互正交的傅里叶排序和沃尔什-哈达玛序列排序图案也被应用于亚奈奎斯特成像 [21-23]。俄罗斯套娃 [24] 和蛋糕切割 [25] 排序的沃尔什-哈达玛图案能将采样率降至奈奎斯特极限的5%-10%。
Recently, the deep learning (DL) technique is employed to identify images [26, 27] and improve the quality of images with the deep neural network (DNN) [28–36]. Specifically, computational ghost imaging via deep learning (CGIDL) has shown a minimum ratio of Nyquist limit down to ∼5 [29, 33]. However, such work’s DNNs are trained by experimental CGI results. Only when the training environment is highly identical to the environment used for image reconstruction can the DNN be effective. This limits its universal applications and restricts it to achieve quick reconstructions. Usually at least thousands of inputs have to be generated for the training, which would be very time-consuming if conducting experimental training each time. Some studies have been performed to test the effectiveness of non-experimental CGI training DNN, the minimum ratios of the Nyquist limit were up to a few percent [30, 31, 35]. However, the sampling ratio is much higher for objects outside of training dataset than those in the training dataset [33]. Therefore, despite the proliferation of numerous algorithms, retrieving high-quality images outside of the training group with a meager Nyquist limit ratio by non-experimental training remains a challenge for the CGIDL system.
近年来,深度学习(DL)技术被用于图像识别[26, 27],并通过深度神经网络(DNN)提升图像质量[28–36]。其中,基于深度学习的计算鬼成像(CGIDL)已实现最低至∼5奈奎斯特极限采样率[29, 33]。但这类研究的DNN需通过实验CGI结果进行训练,只有当训练环境与图像重建环境高度一致时,DNN才能有效工作。这限制了其普适性应用,并阻碍了快速重建的实现。通常需要生成至少数千组训练输入,若每次进行实验训练将极其耗时。已有研究测试非实验CGI训练DNN的效果,其奈奎斯特极限最低采样率可达百分之几[30, 31, 35],但对于训练集外的物体,其采样率远高于训练集内物体[33]。因此,尽管算法层出不穷,如何通过非实验训练在极低奈奎斯特极限采样率下获取训练集外的高质量图像,仍是CGIDL系统面临的挑战。
This letter aims to minimize the necessary sampling number further and improve the imaging quality with the combination of DL and colored noise CGI. Recently, it has been shown that the synthesized colored noise patterns possess unique non-zero correlations between neighborhood pixels via amplitude modulation in the spatial frequency domain [37, 38]. In particular, The pink noise CGI owns positive cross-correlations in the second-order correlation [37]. It gives a good image quality under a boisterous environment or pattern distortion when the traditional CGI method fails. Combining DL with pink noise CGI shows that the imaging can be retrieved under an extremely low sampling rate \mathrm{'\sim0.8%} ). We also show that we can get training patterns from the simulation without introducing the environmental noises, i.e., there is no need to get DNN training with a large number of experimental training inputs. The object used in the experiment can be independent of the training dataset, which can largely benefit CGIDL in the real application.
本信旨在进一步减少必要采样数量,并通过深度学习(DL)与彩色噪声关联成像(CGI)的结合提升成像质量。最新研究表明,通过空间频域的振幅调制,合成彩色噪声模式在相邻像素间具有独特的非零相关性[37,38]。特别是粉红噪声CGI在二阶关联中呈现正交叉相关性[37],当传统CGI方法失效时,它能在嘈杂环境或模式畸变下保持良好成像质量。实验表明,深度学习与粉红噪声CGI结合可在极低采样率( \mathrm{'\sim0.8%} )下重建图像。我们还证明可以从仿真中获取训练模式而无需引入环境噪声,即无需通过大量实验训练输入来训练深度神经网络(DNN)。实验中使用的物体可独立于训练数据集,这将极大促进CGIDL在实际应用中的发展。
II. DEEP LEARNING
II. 深度学习
FIG. 1: Architecture of DNN. It consists of four convolution layers, one image input layer, one fully connected layer (yellow), the rectified linear unit, and the batch normalization layers (red). In the upper line are CGI results (training inputs) and handwriting ground truths (training labels); In the bottom line are CGI results from the experiment (test inputs) and CGIDL results (test outputs) with block style.
图 1: DNN架构。该架构包含四个卷积层、一个图像输入层、一个全连接层(黄色)、修正线性单元以及批量归一化层(红色)。上方为CGI生成结果(训练输入)和手写真实值(训练标签);下方为实验获得的CGI结果(测试输入)和采用块状风格的CGIDL输出结果(测试输出)。
Our DNN model, as shown in Fig. 1, uses four convolution layers, one image input layer, and one fully connected layer. Small 3×3 receptive fields are applied throughout the whole convolution layers for better performance [39]. Batch normalization layers (BNL), rectified Linear Unit (ReLU) layers and zero padding are added between each convolution layer. The BNL is functioned to avoid internal covariate shift during the training process and speed up the training of DNN [40]. The ReLU layer applies a threshold operation to each element of the inputs [41]. The zero padding part was designed to maintain the characteristic of input images’ boundaries. To customize the size of training pictures, both the input and output layers are set to be 54×98 . The solver for training is employed by the Stochastic Gradient Descent with Momentum Optimizer (SGDMO) to reduce the oscillation via using momentum. The parameter vector can be updated via equation Eq. (1), which demonstrates the updating process during the iteration.
我们的 DNN (Deep Neural Network) 模型如图 1 所示,使用了四个卷积层、一个图像输入层和一个全连接层。整个卷积层都采用小的 3×3 感受野以获得更好的性能 [39]。每个卷积层之间都添加了批量归一化层 (BNL)、修正线性单元 (ReLU) 层和零填充。BNL 的作用是避免训练过程中的内部协变量偏移,并加速 DNN 的训练 [40]。ReLU 层对输入的每个元素应用阈值操作 [41]。零填充部分旨在保持输入图像边界的特征。为了自定义训练图片的大小,输入层和输出层都设置为 54×98。训练求解器采用带动量的随机梯度下降优化器 (SGDMO),通过使用动量来减少振荡。参数向量可以通过方程 Eq. (1) 进行更新,该方程展示了迭代过程中的更新过程。
where ℓ is the iteration number, α is the learning rate, θ is the parameter vector, and E(θ) is the loss function, mean square error (MSE). The MSE is defined as
其中 ℓ 为迭代次数,α 为学习率,θ 为参数向量,E(θ) 为损失函数,即均方误差 (MSE)。MSE定义为
Here, G represents the pixel value of the resulted imaging. G(o) represents pixels that the light ought to be transmitted, i.e., the object area, while G(b) represents pixels that the light ought to be blocked, i.e., the background area. X is the ground truth calculated by
此处,G 表示成像结果的像素值。G(o) 表示光线应当透过的像素(即目标区域),而 G(b) 表示光线应当被阻挡的像素(即背景区域)。X 是通过计算得到的地面真实值。
The third part on the right hand side of the equation is the feature of SGDMO, analog to the momentum where γ determines the contribution of the previous gradient step to the current iteration [42]. Two strategies are applied to avoid over-fitting of training images. At the end of DNN, a dropout layer is applied with probability of dropping out input elements being 0.2, which is aimed to reduce the connection between convolution layers and the fully connected layer [43]. Meanwhile, we adopted a step decay schedule for the learning rate. The learning rate dropped from 10−3 to 10−4 after 75 epochs, which constrain the fitting parameters within a reasonable region. Lower the learning rate could avoid over fitting significantly with constant maximum epochs.
方程右侧的第三部分是SGDMO的特征,类似于动量,其中γ决定了前一步梯度对当前迭代的贡献[42]。为避免训练图像过拟合,采用了两种策略。在DNN末端应用了丢弃概率为0.2的dropout层,旨在减少卷积层与全连接层之间的关联[43]。同时,我们采用了学习率逐步衰减策略,学习率在75个周期后从10−3降至10−4,从而将拟合参数限制在合理范围内。在保持最大周期数不变的情况下,降低学习率能显著避免过拟合。
B. Network training
B. 网络训练
The proposed CGIDL scheme requires a training process based on pre-prepared dataset. After training in simulation, it owns ability to reconstruct the images. We use a set of 10000 digits from the MNIST handwritten digit database [44] as training images. All images are resized and normalized to 54×98 to test the smaller sampling ratio. These training images are reconstructed by the CGI algorithm. The training images and reconstruction training images then feed the DNN model as inputs and outputs, respectively, as shown in Fig. 2(a). The white noise and pink noise speckle patterns are used separately for the training process, using exactly the same protocol. The maximum epochs are set as 100, and the training iteration is 31200. The program is implemented via MATLAB R2019a Update 5 (9.6.0.1174912, 64-bit), and the DNN is implemented through DL Toolbox. The GPU-chip NVIDIA GTX1050 is used to accelerate the speed of the computation.
提出的CGIDL方案需要基于预先准备的数据集进行训练。在仿真环境中完成训练后,该方案将具备图像重建能力。我们使用MNIST手写数字数据库[44]中的10000个数字作为训练图像,所有图像均被调整尺寸并归一化为54×98以测试更小的采样率。这些训练图像通过CGI算法进行重建,随后将原始训练图像与重建训练图像分别作为DNN模型的输入和输出,如图2(a)所示。训练过程中分别采用白噪声和粉红噪声散斑图案,且保持完全相同的协议参数。最大训练周期设为100次,训练迭代次数为31200次。程序通过MATLAB R2019a Update 5(9.6.0.1174912, 64位)实现,DNN部分采用深度学习工具箱(DL Toolbox)构建,并利用NVIDIA GTX1050显卡加速计算。
The trained DNN is then tested by simulation and used for retrieving CGI results in the experiments. In the testing part, the CGI algorithm generates reconstructed images from testing images with both the MNIST handwritten digits and block style digits, where the later set is different from images in the training group. As shown in Fig. 2(b), the trained DNN, fed with reconstruction testing images, generates CGIDL results. Comparing the difference between CGIDL and testing images, we could measure the quality of the trained DNN. Well-performed DNN can be used for retrieving CGI in the experiment.
训练好的深度神经网络 (DNN) 随后通过仿真进行测试,并用于在实验中检索计算鬼成像 (CGI) 结果。在测试环节,CGI 算法分别对 MNIST 手写数字和方块风格数字的测试图像进行重建,其中后者与训练组的图像不同。如图 2(b) 所示,输入重建测试图像后,训练好的 DNN 会生成 CGIDL 结果。通过对比 CGIDL 与测试图像的差异,我们可以评估训练好的 DNN 的质量。性能良好的 DNN 可用于实验中检索 CGI。
The schematic of the experiment is shown in Fig. 2(c). A CW laser is used to illuminate the DMD, on which the noise patterns are loaded. The pattern generated by the DMD is then projected onto the object. In our experiment, the size of the noise patterns is 216×392 DMD pixels (54 × 98 independent pixels), in which the independent changeable mirrors count for 4×4 pixels. Each DMD pixel is 16μm×16μm in size.
实验示意图如图 2(c)所示。实验中采用连续波激光器照射加载了噪声图案的 DMD (数字微镜器件),随后由 DMD 生成的图案被投射到目标物体上。实验所用噪声图案尺寸为 216×392 个 DMD 像素(54 × 98 个独立像素),其中每个可独立控制的微镜对应 4×4 个像素单元。每个 DMD 像素的物理尺寸为 16μm×16μm。
In the CGI process, the quality of the images is proportional to the sampling rate, which is the ratio between the number of illumination patterns Npattern and Npixel [45, 46]:
在CGI过程中,图像质量与采样率成正比,即照明模式数量Npattern与像素数量Npixel之比 [45, 46]:
FIG. 2: The flow chart of CGIDL consists of three parts: (a) training, (b) test, and (c) experiment. The DNN model is trained with CGI results from database via simulation. The simulation testing process and experimental measuments use both the handwriting digits and block style digits. The experimental part for CGI uses pink noise and white noise speckle patterns, and their CGI results are ameliorated by trained DNN model.
图 2: CGIDL流程图包含三部分: (a) 训练, (b) 测试, (c) 实验。DNN模型通过仿真使用数据库中的CGI结果进行训练。仿真测试过程和实验测量同时使用手写数字和方块风格数字。CGI实验部分采用粉红噪声和白噪声散斑图案, 其CGI结果通过训练后的DNN模型进行优化。
In the following, We compared the trained network using white noise speckle patterns (DL white) and pink noise speckle patterns (DL pink), as well as the conventional CGI (CGI white) in terms of reconstruction performance with respect to the sampling ratio β .
我们比较了使用白噪声散斑图案 (DL white) 和粉红噪声散斑图案 (DL pink) 训练的网络,以及传统 CGI (CGI white) 在不同采样率 β 下的重建性能。
III. SIMULATION
III. 仿真
To test the robustness of our method to different datasets, noise, and its performs at different sampling rates, we performed a set of simulations. Two sets of testing images are used in the simulation. One of which is the handwriting digits 1-9 from the training set, the other is the block style digits 1-9, which are completely independent of training images. These images have 28×28 pixels and are resized into 54×98 by widening and amplification. We started our simulation from the comparison of the CGI white, DL white and DL pink without noise at β=5 , as shown in Fig. 3. The upper part is with the handwriting digits 1-9, the lower part is with the block style digits 1-9. Apparently, at this low sampling rate, the traditional CGI method fails to retrieve the images in both cases. On the other hand, both DL methods work much better than the traditional CGI. For digits from the training dataset, both methods work almost equally well. For digits from outside the training dataset, DL pink works already better than DL white. For example, the DL white barely can distinguish digits ’3’ and ’8’, but DL pink can retrieve all the digits images.
为验证本方法对不同数据集、噪声及采样率的鲁棒性,我们进行了一系列仿真实验。仿真采用两组测试图像:一组来自训练集的手写数字1-9,另一组是与训练图像完全独立的方块风格数字1-9。这些图像原始尺寸为28×28像素,通过加宽放大调整为54×98像素。我们首先在β=5无噪声条件下比较CGI白光、DL白光与DL粉光的性能,如图3所示。上半部分为手写数字1-9的重建结果,下半部分为方块数字1-9的重建结果。显然,在此低采样率下,传统CGI方法在两类情况下均无法重建图像;而两种DL方法均显著优于传统CGI。对于训练集内的数字,两种DL方法表现相当;对于训练集外数字,DL粉光已优于DL白光——例如DL白光难以区分"3"和"8",而DL粉光可完整重建所有数字图像。
In real application, there always exist noise in the measurement. It is therefore worthwhile checking the performances of different methods under the influence of noise. We then performed another set of simulations with added grayscale random noise. The signal-to-noise ratio (SNR) in logarithmic decibel scale is defined as
在实际应用中,测量中总是存在噪声。因此,有必要检查不同方法在噪声影响下的性能。我们随后进行了一组添加灰度随机噪声的模拟实验。对数分贝尺度的信噪比 (SNR) 定义为
SNR=10logPsPb,
信噪比 (SNR) 计算公式:
SNR=10logPsPb,
where Ps is the average signa