Abstract
The deep learning wavefront sensor (DLWFS) allows the direct estimate of Zernike coefficients of aberrated wavefronts from intensity images. The main drawback of this approach is related to the use of massive convolutional neural networks (CNNs) that are lengthy to train or estimate. In this paper, we explore several options to reduce both the training and estimation time. First, we develop a CNN that can be rapidly trained without compromising accuracy. Second, we explore the effects given smaller input image sizes and different amounts of Zernike modes to be estimated. Our simulation results demonstrate that the proposed network using images of either $8 \times 8$, $16 \times 16$, or $32 \times 32$ will dramatically reduce training time and even boost the estimation accuracy of Zernike coefficients. From our experimental results, we can confirm that a $16 \times 16$ DLWFS can be quickly trained and is able to estimate the first 12 Zernike coefficients–skipping piston, tip, and tilt–without sacrificing accuracy and significantly speeding up the prediction time to facilitate low-cost, real-time adaptive optics systems.
© 2021 Optical Society of America
Full Article | PDF ArticleMore Like This
Yohei Nishizaki, Matias Valdivia, Ryoichi Horisaki, Katsuhisa Kitaguchi, Mamoru Saito, Jun Tanida, and Esteban Vera
Opt. Express 27(1) 240-251 (2019)
Yushuang Li, Dan Yue, and Yihao He
Appl. Opt. 61(14) 4168-4176 (2022)
Yulong He, Zhiwei Liu, Yu Ning, Jun Li, Xiaojun Xu, and Zongfu Jiang
Opt. Express 29(11) 17669-17682 (2021)