一种基于图像融合和卷积神经网络的相位恢复波前传感方法
基金项目: 国家级-国家自然科学基金 (61471039)
摘要: 相位恢复法利用光波传输中某一(或某些)截面上的光强分布来传感系统波前,其结构简单,不易受震动及环境干扰,被广泛应用于光学遥感,像差检测等领域。传统相位恢复法采用迭代计算,很难满足实时性要求,且在一定程度上依赖于迭代转换或迭代优化初值。为克服上述问题,本文提出了一种基于卷积神经网络的相位恢复波前传感方法,该方法采用基于小波变换的图像融合技术对焦面和离焦面图像进行融合处理,可在不损失图像信息的同时简化卷积神经网络的输入。网络模型训练完成后可依据输入的融合图像直接输出表征波前相位的4-9阶Zernike系数,且波前传感精度RMS可达0.015λ,λ=632.8nm。通过引入噪声及离焦量误差,验证了该方法对噪声具有一定鲁棒性,且相对离焦量误差在7.5%内时,波前传感精度RMS仍可达0.05λ。此外,分析了当系统实际像差阶数与网络训练阶数不同时,本方法所能实现的传感精度,并给出了进一步改进措施。
A phase retrieval wavefront sensing based on image fusion and convolutional neural network
- Received Date:
18 August 2020
Abstract: The conventional phase retrieval wavefront sensing approaches mainly refer to a series of iterative algorithms,such as G-S algorithms ,Y-G algorithms and error reduction algorithms ,these methods use intensity information to calculate the wavefront phase. However, most of the traditional phase retrieval algorithms are difficult to meet the real-time requirements and depend on the iteration initial value used in iterative transformation or iterative optimization to some extent, their practicality is limited. To overcome these problems, in this paper ,a phase-diversity phase retrieval wavefront sensing method based on wavelet transform image fusion and convolutional neural network is proposed. Specifically, the image fusion method based on wavelet transform is used to fuse the point spread functions at the in-focus and defocus image planes, so as to simplify the network inputs without losing the image information. CNN can directly extract image features and fit the required nonlinear mapping. In this paper, the CNN is utilized to establish the nonlinear mapping between the fusion images and wavefront distortions (represented by Zernike polynomials), that is, the fusion images are taken as the input data and the corresponding Zernike coefficients as the output data. The network structure of the training in this paper is 22 layers, including 1 input layer, 13 convolution layers, 6 pooling layers, 1 flatten layer and 1 full connection layer, that is, the output layer. The size of the convolution kernel is 3×3 and the step size is 1. The pooling method selects the maximum pooling and the size of the pooling kernel is 2×2. The activation function is ReLU, the optimization function is Adam, the loss function is the MSE, and the learning rate is 0.0001. The number of training data is 10000, which is separated into three parts:training set,validation set and test set, accounting for 80%, 15% and 5% respectively. Trained CNN can directly output the Zernike coefficients of order 4-9 to a high precision with these fusion images serving as the input, which is more in line with the real-time requirements. Abundant simulation experiments prove that the wavefront sensing precision is RMS 0.015λ, when the dynamic ranges of Zernike coefficients of order 4-9 are [-0.5λ,0.5λ]. In practical application, according to the system aberration characteristics, the number of network output layer units can be changed and the network structure can be adjusted based on the method presented in this paper so as to train the new network suitable for higher order aberration to realize high-precision wavefront sensing. It is also proved that the proposed method has certain robustness to noise, and when the relative defocus error is within 7.5%, the wavefront sensor accuracy is acceptable.