single-shot and lensless complex-amplitude imaging with … · another approach for phase imaging...

5
Vol.:(0123456789) 1 3 Optical Review (2018) 25:593–597 https://doi.org/10.1007/s10043-018-0452-1 REGULAR PAPER Single‑shot and lensless complex‑amplitude imaging with incoherent light based on machine learning Ryoichi Horisaki 1,2  · Kazuki Fujii 1  · Jun Tanida 1 Received: 8 April 2018 / Accepted: 25 July 2018 / Published online: 31 July 2018 © The Author(s) 2018 Abstract We present a single-shot incoherent light imaging method for simultaneously observing both amplitude and phase without any imaging optics, based on machine learning. In the proposed method, an object with a complex-amplitude field is illu- minated with incoherent light and is captured by an image sensor with or without a coded aperture. The complex-amplitude field of the object is reconstructed from a single captured image using a state-of-the-art deep convolutional neural network, which is trained with a large number of input and output pairs. In experimental demonstrations, the proposed method was verified with a handwritten character database, and the effect of a coded aperture printed on an overhead projector film in the reconstruction was examined. Our method has advantages over conventional wavefront sensing techniques using incoherent light, namely simplification of the optical hardware and improved measurement speed. This study shows the importance and practical impact of machine learning techniques in various fields of optical sensing. Keywords Phase retrieval · Digital holography · Incoherent holography · Coded aperture · Deep learning · Convolutional neural network 1 Introduction An optical field is expressed as a complex-amplitude, which describes both the amplitude and phase of a light wave, in wave optics [1]. Conventional cameras just observe ampli- tude information only. Phase information is important for visualizing transparent objects, e.g., biological cells in biology, and compensating for aberrations/scattering with adaptive optics, e.g., deep tissue imaging in biomedicine and telescopic imaging through atmospheric turbulence in astronomy. An established approach for complex-amplitude imaging, in which the transmittance/reflectance and phase delay of an object are observed, is digital holography (DH) [25]. DH basically assumes coherent illumination, such as a laser, but several incoherent DH methods have been proposed [610]. One issue with these incoherent DH methods is a tradeoff between the number of shots and the space–bandwidth prod- uct due to the use of a spatial or temporal reference carrier. Another approach for phase imaging with incoherent light is to use a Shack–Hartmann wavefront sensor [11]. Single- shot complex-amplitude/phase imaging methods based on the Shack–Hartmann wavefront sensor have utilized a lens array or a holographic optical element to observe a stereo or plenoptic image, and this results in a tradeoff between the spatial and angular resolutions [1217]. Thus, these methods involve a compromise between the space–bandwidth product and the simplicity of the optical setup. In this paper, to solve the above issues, we propose a method for single-shot complex-amplitude imaging with incoherent light and no imaging optics. We verify the effec- tiveness of a deep convolutional neural network as the recon- struction algorithm and the use of a coded aperture (CA) to improve the reconstruction fidelity. This approach extends the range of applications of complex-amplitude imaging and shows the importance of machine learning techniques in optical sensing. * Ryoichi Horisaki [email protected] 1 Department of Information and Physical Sciences, Graduate School of Information Science and Technology, Osaka University, 1-5 Yamadaoka, Suita, Osaka 565-0871, Japan 2 JST, PRESTO, 4-1-8 Honcho, Kawaguchi-shi, Saitama 332-0012, Japan

Upload: others

Post on 26-Jul-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Single-shot and lensless complex-amplitude imaging with … · Another approach for phase imaging with incoherent light is to use a Shack–Hartmann wavefront sensor [11]. Single-shot

Vol.:(0123456789)1 3

Optical Review (2018) 25:593–597 https://doi.org/10.1007/s10043-018-0452-1

REGULAR PAPER

Single‑shot and lensless complex‑amplitude imaging with incoherent light based on machine learning

Ryoichi Horisaki1,2 · Kazuki Fujii1 · Jun Tanida1

Received: 8 April 2018 / Accepted: 25 July 2018 / Published online: 31 July 2018 © The Author(s) 2018

AbstractWe present a single-shot incoherent light imaging method for simultaneously observing both amplitude and phase without any imaging optics, based on machine learning. In the proposed method, an object with a complex-amplitude field is illu-minated with incoherent light and is captured by an image sensor with or without a coded aperture. The complex-amplitude field of the object is reconstructed from a single captured image using a state-of-the-art deep convolutional neural network, which is trained with a large number of input and output pairs. In experimental demonstrations, the proposed method was verified with a handwritten character database, and the effect of a coded aperture printed on an overhead projector film in the reconstruction was examined. Our method has advantages over conventional wavefront sensing techniques using incoherent light, namely simplification of the optical hardware and improved measurement speed. This study shows the importance and practical impact of machine learning techniques in various fields of optical sensing.

Keywords Phase retrieval · Digital holography · Incoherent holography · Coded aperture · Deep learning · Convolutional neural network

1 Introduction

An optical field is expressed as a complex-amplitude, which describes both the amplitude and phase of a light wave, in wave optics [1]. Conventional cameras just observe ampli-tude information only. Phase information is important for visualizing transparent objects, e.g., biological cells in biology, and compensating for aberrations/scattering with adaptive optics, e.g., deep tissue imaging in biomedicine and telescopic imaging through atmospheric turbulence in astronomy.

An established approach for complex-amplitude imaging, in which the transmittance/reflectance and phase delay of an object are observed, is digital holography (DH) [2–5]. DH basically assumes coherent illumination, such as a laser, but several incoherent DH methods have been proposed [6–10].

One issue with these incoherent DH methods is a tradeoff between the number of shots and the space–bandwidth prod-uct due to the use of a spatial or temporal reference carrier.

Another approach for phase imaging with incoherent light is to use a Shack–Hartmann wavefront sensor [11]. Single-shot complex-amplitude/phase imaging methods based on the Shack–Hartmann wavefront sensor have utilized a lens array or a holographic optical element to observe a stereo or plenoptic image, and this results in a tradeoff between the spatial and angular resolutions [12–17]. Thus, these methods involve a compromise between the space–bandwidth product and the simplicity of the optical setup.

In this paper, to solve the above issues, we propose a method for single-shot complex-amplitude imaging with incoherent light and no imaging optics. We verify the effec-tiveness of a deep convolutional neural network as the recon-struction algorithm and the use of a coded aperture (CA) to improve the reconstruction fidelity. This approach extends the range of applications of complex-amplitude imaging and shows the importance of machine learning techniques in optical sensing.

* Ryoichi Horisaki [email protected]

1 Department of Information and Physical Sciences, Graduate School of Information Science and Technology, Osaka University, 1-5 Yamadaoka, Suita, Osaka 565-0871, Japan

2 JST, PRESTO, 4-1-8 Honcho, Kawaguchi-shi, Saitama 332-0012, Japan

Page 2: Single-shot and lensless complex-amplitude imaging with … · Another approach for phase imaging with incoherent light is to use a Shack–Hartmann wavefront sensor [11]. Single-shot

594 Optical Review (2018) 25:593–597

1 3

2 Method

In the proposed method, a complex-amplitude object field illuminated with incoherent light is captured by an image sensor with or without a coded aperture (CA), as shown in Fig. 1. A CA has been used to improve the reconstruc-tion conditions because its Fourier spectrum is significantly extended compared with that of a conventional non-coded aperture. Using CAs, single-shot amplitude imaging with incoherent light and single-shot reference-free DH with coherent light have been demonstrated [18–23].

The imaging process of the proposed method is written in general form as

where g ∈ ℝN2×1 is the captured intensity image and

[∙] is the forward operator. f ∈ ℂN2×1 , f r ∈ ℝ

N2×1 , and f� ∈ ℝ

N2×1 are the complex-amplitude, amplitude, and phase of the object field, respectively, and f = f r exp(if�) . Here, N is the number of pixels along one spatial direction, ℝ

P×Q is a P × Q matrix of real numbers, and ℂP×Q is a P × Q matrix of complex numbers.

The inversions for reconstructing the amplitude and the phase are separable, as indicated by Eq.  (2). In this paper, for simplicity, we solve the following two inversions independently:

where −1r[∙] and −1

�[∙] are the inverse operators for the

amplitude and the phase, respectively.We use a regression algorithm for the inversions of

Eqs. (3) and (4) in the reconstruction process. Regressions, including deep learning, are typical techniques in machine learning, and have applied regressions to optical sensing

(1)g = [f ]

(2)= [f r exp(if�)],

(3)f̂ r = −1r[g],

(4)f̂𝜙 = −1𝜙[g],

and control [24–29]. In addition, several studies for phase retrieval, ghost imaging, and superresolution imaging with deep learning have been reported [30–32]. In this paper, we modify a deep convolutional residual network (ResNet [33]) for each of the inversions. Residual learning in ResNet pre-vents stagnation during the learning process and optimizes deep layers efficiently. ResNet has been used for phase retrieval and computer-generated holography and showed favorable results [29, 30].

The network architecture for the reconstruction is shown in Fig. 2. The network is composed of multi-scale ResNets, as shown in Fig. 2a. K is the number of filters at each of the convolutional layers. “D” is a downsampling block, as shown in Fig. 2b. “U” is an upsampling block, as shown in Fig. 2c. “R” is a residual convolutional block, as shown in Fig. 2d. “S” is a skip convolutional block, as shown in Fig. 2e. The definitions of the layers are as follows [34]: “Conv(S, L)” is a two-dimensional convolutional layer with a filter size S and a stride L. “TConv(S, L)” is a two-dimen-sional transposed convolutional layer with a filter size S and a stride L. “BatchNorm” is a batch normalization layer. “ReLU” is a rectified linear unit layer.

The network architectures for reconstructing the ampli-tude and phase are the same, as shown in Fig. 2a, but they are trained with different datasets. The amplitude datasets are composed of pairs of amplitude images f r and their captured intensity images g . On the other hand, the phase datasets are composed of pairs of phase images f� and their captured intensity images g.

3 Experimental demonstration

The proposed method was demonstrated experimentally. The complex-amplitude object was implemented with two spatial light modulators (SLMs: LC 2012 manufactured by Holoeye, pixel pitch: 36 µm, pixel count: 768 × 1024 ); one of them was operated in the phase mode, and the other one was operated in the amplitude mode. The phase and ampli-tude SLMs were located at positions 5 and 7 cm from an image sensor (PL-B953 manufactured by PixeLink, pixel pitch: 4.65 µm, pixel count: 768 × 1024 ), respectively. Those two SLMs were illuminated with an incoherent green light-emitting diode (M565L3 manufactured by Thorlabs, nominal wavelength: 565 nm, full width at half maximum of spectrum: 103 nm) without any spatial or spectral filter. Our method does not suppose amplitude and phase layers at different distances. A single complex-amplitude layer is readily trainable in the same manner by implementing such an object. In the case with the CA, the CA was printed on an overhead projector film (OHP film: VF-1421N manufac-tured by Kokuyo) by a laser printer (ApeosPort-V C2276 manufactured by Fuji Xerox, resolution: 1200 dpi) and it

Incoherentlight

Complex- amplitude

object

Image sensor

(lensless)

Coded aperture(optional)

Fig. 1 Schematic diagram of single-shot, lensless complex-amplitude imaging with/without a CA

Page 3: Single-shot and lensless complex-amplitude imaging with … · Another approach for phase imaging with incoherent light is to use a Shack–Hartmann wavefront sensor [11]. Single-shot

595Optical Review (2018) 25:593–597

1 3

was located at a position 1 cm from the image sensor. The CA was a random binary pattern with the maximum print-able resolution.

The amplitude and phase images simultaneously dis-played on the two SLM were different handwritten digits randomly selected from the EMNIST database, where the

pixel count of the images was 28 × 28 [35]. The captured intensity images were reduced to 28 × 28 pixels, and then they were expanded to 32 × 32(= N2) with zero padding for adjusting the captured image size to the input one of the network, where the input and output image sizes are a power of two. These resized captured intensity images and the original amplitude and phase images, which were also expanded to 32 × 32(= N2) with zero padding, were provided to the network for reconstructing the amplitude and phase images with zero padding. The output images from the network were cropped to remove the zero padding area, and the final amplitude and phase images with a pixel count of 28 × 28 were produced. The number of convolu-tional filters was 64(= K) . An algorithm called “Adam” was used to optimize the network with an initial learning rate of 0.0001 for the step width in the optimization, a mini-batch size of 50, and a maximum number of epochs of 50 [36]. The number of training pairs was 100,000 in both the amplitude and phase reconstructions.

The experimental results without the CA are shown in Figs. 3 and 4. The trained amplitude and phase networks were respectively tested with 1000 amplitude images and 1000 phase images that had not been used for the training process. Ten examples of the amplitude and phase test images are shown in Fig. 3a and b, respectively. Their captured intensity images from the amplitude and phase images in Fig. 3 are shown in Fig. 4a. The amplitude and phase images reconstructed from Fig. 4a are shown in Fig. 4b and c. The peak signal-to-noise ratios (PSNRs) of the 1000 reconstructed amplitude images and 1000 recon-structed phase images were 20.9 and 14.3 dB, respec-tively. These results demonstrated the effectiveness of

S

N2

(N/2)2×K

N2×K

(N/2)2×K

(N/4)2×K

(N/4)2×K

22×K

22×K

12×K

N2

D

Captured intensity image

Reconstructed amplitude/phase

D

D

R

D

U

U

U

U

U

D

(a)

Conv (3, 2)Conv (3, 2)

BatchNorm

ReLU

Conv (3, 1)

+

BatchNorm

ReLU

(b)

TConv (2, 2)

BatchNorm

ReLU

Conv (3, 1)

+

BatchNorm

ReLU

TConv (2, 2)

(c)

Conv (3, 1)

BatchNorm

ReLU

Conv (3, 1)

+

BatchNorm

ReLU

kcolb

bus-R

R-subblock

(d)

ReLU

Conv (3, 1)

BatchNorm

ReLU

Conv (3, 1)

BatchNorm

(e)

Fig. 2 Network design for the reconstruction. Structures of the a whole network, b D-block, c U-block, d R-block, and e S-block

Fig. 3 Ten examples of a the untrained original amplitude images and b the untrained original phase images, where phases are normalized in the interval [0,�]

Page 4: Single-shot and lensless complex-amplitude imaging with … · Another approach for phase imaging with incoherent light is to use a Shack–Hartmann wavefront sensor [11]. Single-shot

596 Optical Review (2018) 25:593–597

1 3

the deep network for complex-amplitude imaging with incoherent light.

The experimental results with the CA are shown in Fig. 5. The network was trained and tested with the same amplitude and phase images as the previous experiment but with different captured intensity images through the CA. The captured intensity images with the CA from the amplitude and phase images in Fig. 3 are shown in Fig. 5a. The amplitude and phase images were also recon-structed from the captured intensity images, as shown in Fig. 5b and c. The PSNRs of the amplitude and phase images were 23.5 and 15.3 dB, respectively. These results showed that the amplitude and phase reconstructions were improved in the case with the CA.

In the experiments, the calculation time for recon-structing the test images was 0.5 s when using a computer with an Intel Dual Xeon processor with a clock rate of 2.1 GHz and a memory size of 128 GB, and NVIDIA Quadro P4000 graphics processing unit with a memory size of 8 GB. The training time was 18 h.

4 Conclusion

We presented a method for single-shot, lensless complex-amplitude imaging with incoherent light and no imaging optics. A deep convolutional neural network was used for reconstructing both the amplitude and phase from a single intensity image. We demonstrated the proposed method experimentally with a CA implemented on an OHP film using handwritten datasets. The results showed the effec-tiveness of the deep network and the CA for complex-amplitude imaging with incoherent light.

Our method reduces the hardware size/cost and the measurement time in incoherent complex-amplitude imaging, while maintaining the image quality. It is read-ily extendable to a reflective setup and multi-dimensional complex-amplitude imaging, such as three-dimensional spatial imaging and spectral imaging. The CA is not lim-ited to only amplitude modulation but can also be applied to phase or complex-amplitude modulation  [22]. Our method is applicable to non-visible spectral regions, such

Fig. 4 Experimental results without the CA. a The captured intensity images with the amplitude and phase images in Fig.  3. The recon-structed b amplitude images and c phase images, where phases are normalized in the interval [0,�] , reconstructed from the captured images in a

Fig. 5 Experimental results with the CA. a The captured intensity images with the amplitude and phase images in Fig.  3. The recon-structed b amplitude images and c phase images, where phases are normalized in the interval [0,�] , reconstructed from the captured images in a

Page 5: Single-shot and lensless complex-amplitude imaging with … · Another approach for phase imaging with incoherent light is to use a Shack–Hartmann wavefront sensor [11]. Single-shot

597Optical Review (2018) 25:593–597

1 3

as X-rays, infrared light, and terahertz radiation. There-fore, the method is promising in various fields, including, biology, security, and astronomy.

Funding This work was supported by JSPS KAKENHI Grant Num-bers JP17H02799 and JP17K00233, and JST PRESTO Grant Number JPMJPR17PB.

Open Access This article is distributed under the terms of the Crea-tive Commons Attribution 4.0 International License (http://creat iveco mmons .org/licen ses/by/4.0/), which permits unrestricted use, distribu-tion, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

References

1. Goodman, J.W.: Introduction to Fourier Optics. McGraw-Hill, New York (1996)

2. Nehmetallah, G., Banerjee, P.P.: Applications of digital and analog holography in three-dimensional imaging. Adv. Opt. Photonics 4, 472–553 (2012)

3. Osten, W., Faridian, A., Gao, P., Körner, K., Naik, D., Pedrini, G., Singh, A.K., Takeda, M., Wilke, M.: Recent advances in digital holography. Appl. Opt. 53, G44–G63 (2014)

4. Bhaduri, B., Edwards, C., Pham, H., Zhou, R., Nguyen, T.H., Goddard, L.L., Popescu, G.: Diffraction phase microscopy: prin-ciples and applications in materials and life sciences. Adv. Opt. Photonics 6, 57–119 (2014)

5. Memmolo, P., Miccio, L., Paturzo, M., Caprio, G.D., Coppola, G., Netti, P.A., Ferraro, P.: Recent advances in holographic 3D particle tracking. Adv. Opt. Photonics 7, 713–755 (2015)

6. Rosen, J., Brooker, G.: Non-scanning motionless fluorescence three-dimensional holographic microscopy. Nat. Photonics 2, 190–195 (2008)

7. Rosen, J., Brooker, G.: Fresnel incoherent correlation holography (FINCH): a review of research. Adv. Opt. Technol. 1, 151–169 (2012)

8. Vijayakmar, A., Kashter, Y., Kelner, R., Rosen, J.: Coded aperture correlation holographya new type of incoherent digital holograms. Opt. Express 24, 12430–12441 (2016)

9. Quan, X., Matoba, O., Awatsuji, Y.: Single-shot incoherent digital holography using a dual-focusing lens with diffraction gratings. Opt. Lett. 42, 383–386 (2017)

10. Liu, J.-P., Tahara, T., Hayasaki, Y., Poon, T.-C.: Incoherent digital holography: a review. Appl. Sci. 8, 143 (2018)

11. Platt, B.C., Shack, R.: History and principles of Shack–Hartmann wavefront sensing. J. Refract. Surg. 17, S573–7 (1995)

12. Parthasarathy, A.B., Chu, K.K., Ford, T.N., Mertz, J.: Quantitative phase imaging using a partitioned detection aperture. Opt. Lett. 37, 4062–4064 (2012)

13. Saita, Y., Shinto, H., Nomura, T.: Holographic Shack–Hartmann wavefront sensor based on the correlation peak displacement detection method for wavefront sensing with large dynamic range. Optica 2, 411–415 (2015)

14. Gong, H., Agbana, T.E., Pozzi, P., Soloviev, O., Verhaegen, M., Vdovin, G.: Optical path difference microscopy with a Shack–Hartmann wavefront sensor. Opt. Lett. 42, 2122–2125 (2017)

15. Chai, J.-X., Tong, X., Chan, S.-C., Shum, H.-Y.: Plenoptic sam-pling. In: Proc. 27th Annu. Conf. Comput. Graph. Interact. Tech., SIGGRAPH ’00, pp. 307–318. ACM Press/Addison-Wesley Pub-lishing Co., New York (2000)

16. Nakamura, T., Horisaki, R., Tanida, J.: Computational phase modulation in light field imaging. Opt. Express 21, 29523–29543 (2013)

17. Shimano, T., Nakamura, Y., Tajima, K., Sao, M., Hoshizawa, T.: Lensless light-field imaging with Fresnel zone aperture: quasi-coherent coding. Appl. Opt. 57, 2841–2850 (2018)

18. Fenimore, E.M., Cannon, T.M.: Coded aperture imaging with uniformly redundant arrays. Appl. Opt. 17, 337–347 (1978)

19. Levin, A., Fergus, R., Durand, F., Freeman, W.T.: Image and depth from a conventional camera with a coded aperture. ACM Trans. Graph. 26, 70:1–70:9 (2007)

20. Horisaki, R., Ogura, Y., Aino, M., Tanida, J.: Single-shot phase imaging with a coded aperture. Opt. Lett. 39, 6466–6469 (2014)

21. Horisaki, R., Egami, R., Tanida, J.: Experimental demonstration of single-shot phase imaging with a coded aperture. Opt. Express 23, 28691–28697 (2015)

22. Egami, R., Horisaki, R., Tian, L., Tanida, J.: Relaxation of mask design for single-shot phase imaging with a coded aperture. Appl. Opt. 55, 1830–1837 (2016)

23. Horisaki, R., Kojima, T., Matsushima, K., Tanida, J.: Subpixel reconstruction for single-shot phase imaging with coded diffrac-tion. Appl. Opt. 56, 7642–7647 (2017)

24. Ando, T., Horisaki, R., Tanida, J.: Speckle-learning-based object recognition through scattering media. Opt. Express 23, 33902–33910 (2015)

25. Horisaki, R., Takagi, R., Tanida, J.: Learning-based imaging through scattering media. Opt. Express 24, 13738–13743 (2016)

26. Takagi, R., Horisaki, R., Tanida, J.: Object recognition through a multi-mode fiber. Opt. Rev. 24, 117–120 (2017)

27. Horisaki, R., Takagi, R., Tanida, J.: Learning-based single-shot superresolution in diffractive imaging. Appl. Opt. 56, 8896–8901 (2017)

28. Horisaki, R., Takagi, R., Tanida, J.: Learning-based focusing through scattering media. Appl. Opt. 56, 4358–4362 (2017)

29. Horisaki, R., Takagi, R., Tanida, J.: Deep-learning-generated holography. Appl. Opt. 57, 3859–3863 (2018)

30. Sinha, A., Lee, J., Li, S., Barbastathis, G.: Lensless computational imaging through deep learning. Optica 4, 1117–1125 (2017)

31. Lyu, M., Wang, W., Wang, H., Wang, H., Li, G., Chen, N., Situ, G.: Deep-learning-based ghost imaging. Sci. Rep. 7, 17865 (2017)

32. Rivenson, Y., Göröcs, Z., Günaydin, H., Zhang, Y., Wang, H., Ozcan, A.: Deep learning microscopy. Optica 4, 1437–1443 (2017)

33. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (IEEE), pp. 770–778 (2016)

34. Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge (2016)

35. Cohen, G., Afshar, S., Tapson, J., van Schaik, A.: EMNIST: an extension of MNIST to handwritten letters (2017). arXiv :1702.05373 (preprint)

36. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization.In: International Conference on Learning Representations (ICLR) (2015)