Skip to content
BY 4.0 license Open Access Published online by De Gruyter April 23, 2024

Compact biologically inspired camera with computational compound eye

  • Shu-Bin Liu , Xu-Ning Liu , Wei-Jie Fan , Meng-Xuan Zhang and Lei Li ORCID logo EMAIL logo
From the journal Nanophotonics

Abstract

The growing interests have been witnessed in the evolution and improvement of artificial compound eyes (CE) inspired by arthropods. However, the existing CE cameras are suffering from a defocusing problem due to the incompatibility with commercial CMOS cameras. Inspired by the CEs of South American Shrimps, we report a compact biologically inspired camera that enables wide-field-of-view (FOV), high-resolution imaging and sensitive 3D moving trajectory reconstruction. To overcome the defocusing problem, a deep learning architecture with distance regulation is proposed to achieve wide-range-clear imaging, without any hardware or complex front-end design, which greatly reduces system complexity and size. The architecture is composed of a variant of Unet and Pyramid-multi-scale attention, with designed short, middle and long distance regulation. Compared to the current competitive well-known models, our method is at least 2 dB ahead. Here we describe the high-resolution computational-CE camera with 271 ommatidia, with a weight of 5.4 g an area of 3 × 3 cm2 and 5-mm thickness, which achieves compatibility and integration of CE with commercial CMOS. The experimental result illustrates this computational-CE camera has competitive advantages in enhanced resolution and sensitive 3D live moving trajectory reconstruction. The compact camera has promising applications in nano-optics fields such as medical endoscopy, panoramic imaging and vision robotics.

1 Introduction

Over one billion years of evolution, arthropods have developed sophisticated compound eyes (CEs) with extraordinary vision, and the increasing interest in CEs has been observed in their evolution and improvement. Natural insect CEs are natural imaging systems with complex-exceptional capabilities, which include numerous closely distributed ommatidia, which enable to inspire the development of artificial CE. CEs have unique advantages in integration, wide FOV, distortion-free imaging, and sensitive motion tracking ability [1], [2], [3], which promotes the applications in medical endoscopy, panorama, robot vision and micro navigation [4], [5], [6], [7], [8], [9], [10]. In the past decade, great efforts have been devoted to the development of artificial CEs, including macroscopic-array systems and integrated CE. As some typical examples in macroscopic-array-systems field, AWARE-2 [1] and AWARE 40 [11] cameras use array cameras to capture large-scale photography. Moreover, RUSH [12] with 35 CMOSs is proposed to achieve cm-scale FOV and μm-level resolution. DLBP array camera [13] is proposed to achieve panorama and computational zoom imaging. The existing array cameras are suffering from large size and complex design, however, the integrated CEs pave the way for camera integration. Planar CEs were first implemented by combining a microlens array (MLA) with commercial CMOS, whose FOV is relatively small [14]. To enable wide FOV, curved CEs have been successfully fabricated using advanced special surface design and micro-nano fabrication technology [2], [15], [16], [17], [18], [19], [20], [21], [22], [23]. Typically, a CE camera is prepared with an enhanced FOV in the whole directions [17]. A CE camera is proposed with a 14.72-mm size and wide FOV [18]. A curved CE camera is fabricated using specific-complex design and fabrication process [23]. Nevertheless, further integration and enhanced-resolution of the whole CE system becomes significant due to the incompatibility of complex CE and CMOS. To the best of our knowledge, such a compact CE camera without special front-end design, has never been achieved.

In this work, to overcome the defocusing problem, a novel deep learning architecture with distance regulation is proposed to achieve wide-range-clear imaging, without any hardware or specialized front-end design, which greatly reduces system complexity and size. Based on the proposed architecture, a compact computational-CE camera is reported, which enables wide-FOV, high-resolution imaging and sensitive 3D moving trajectory reconstruction, with only a weight of 5.4 g, a 3 × 3-cm2 area and 5-mm thickness.

2 Results

2.1 Principle and concept

Natural Penaeus under a complex environment, with a well-developed visual system densely covered with ommatidia, inspired this work. As illustrated in Figure 1(a), the CE of South American Shrimp is composed of a series of ommatidia under a scanning electron microscope (SEM), which is regarded as large FOV. Artificial CEs perfectly inherit the advantages of natural CEs, however, the direct integration of traditional CEs with planar CMOS will result in defocusing blurring and inability to achieve perfect imaging due to the curved focusing, as illustrated in Figure 1(b). The concept of computational-CE camera is inspired by above inspiration. In Figure 1(c), after passing through a CE, incident light will be focused onto different planes, which results in blurred imaging with an inhomogeneous resolution. To address the challenges of incompatibility between traditional CEs and commercial CMOS, here this work cleverly combines a proposed multi-branch model with CE camera to resolve complete-focused imaging with uniform resolution from degraded ones, without any additional hardware and complex front-end design.

Figure 1: 
Principle and concept of proposed computational-CE camera. (a) CE of the South American Shrimp. (b) Defocusing principles of curved CE with planar CMOS. (c) Principle of our fabricated computational-CE camera.
Figure 1:

Principle and concept of proposed computational-CE camera. (a) CE of the South American Shrimp. (b) Defocusing principles of curved CE with planar CMOS. (c) Principle of our fabricated computational-CE camera.

3 Methods

3.1 PDMS CE fabrication methods

The fabrication of polydimethylsiloxane (PDMS) CE involves six steps, as shown in Figure 2(a)–(f). Firstly, the substrate is fabricated by a nanoArch® S140 3D printing system with 25-um precision (BMF Precision Tech Inc. Shenzhen, China), and the photosensitive resin UTL is used as the printing material. Secondly, the microholes of the substrate are filled with photoresist (SU-8). Then, it is degassed for 1 h in a vacuum machine (FUJIWARA-550D) for degassing and the bubbles floating on the photoresist surface are removed with a tool. Then it was placed in KW-4A coater at a speed of 3000 r/s for 120 s for homogenizing of the glue, thereby forming a preliminary microlens array (MLA) mold. The mold is then placed on a heating platform (CHEMAT-MODELKW-4AH) for initial curing for 15 s, and then exposed to ultraviolet light to final cure, completing the mold fabrication. Then, the PDMS liquid is fabricated by mixing DowCorning’s Sylgard 184 in a ratio of 10:1. The mixture is degassed using a vacuum machine and poured into the mold. Subsequently, it is heat-cured at 75–80 °C for 1.5 h, resulting in the formation of a planar PDMS film with the MLA. Finally, the planar PDMS film with the MLA is pushed to a heated shell slowly from the surface without the MLA. After a few minutes when it dropped to room temperature, the curved shell was torn off with the MLA.

Figure 2: 
Fabrication of PDMS CE. (a) 3D printing substrate using nanoArch® S140. (b) Microholes filled with SU-8. (c) Uncured mold. (d) Fabricated cured mold. (e) PDMS fabrication. (f) Cured PDMS fabrication. (g) Physical images of the substrate. (h) Physical images of the mold. (i) Physical images of the curved PDMS.
Figure 2:

Fabrication of PDMS CE. (a) 3D printing substrate using nanoArch® S140. (b) Microholes filled with SU-8. (c) Uncured mold. (d) Fabricated cured mold. (e) PDMS fabrication. (f) Cured PDMS fabrication. (g) Physical images of the substrate. (h) Physical images of the mold. (i) Physical images of the curved PDMS.

The physical images of the substrate, mold, and curved PDMS filmed with the MLA are shown in Figure 2(g)–(i). In Figure 2(g), the total number of holes in the substrate is 271, and the diameter of each hole is 400 μm, and the diameter of the whole substrate is 9 mm. In Figure 2(i), the diameter of the entire PDMS compound eye is 8 mm, weighs 0.046 g, and the MLA has 271 ommatidia, each of which has a diameter of 400 μm and a refractive index of 1.4.

3.2 Imaging pipeline and designed network architecture

The traditional imaging pipelines are typically limited to complex systems and methods, previously proposed CE cannot usually account for single-frame imaging. As illustrated in Figure 3(a), the captured raw-defocused images in various distances using complex-designed CE, are composited as an image with full focus using complex-image fusion method, such as Laplace pyramid fusion algorithm [24]. In general, the calculated process can be described as:

(1) I x , y = G i = 1 N F ( I i ( x , y ) )

where I x , y and I i (x, y) represent the composited full-focus image and captured raw-defocused image, respectively. F(·) represents the operation for captured raw-defocused image, and G(·) represents the complex-image fusion method. The fidelity of the obtained information can be reduced due to defocusing blurring caused by different distances. Furthermore, it can be foreseen that the blurring caused by different distances varies. To address these challenges, here a novel multi-branch imaging strategy with distance regulation for CE camera is proposed, as illustrated in Figure 3(b). The captured raw-defocused images in various distances are recovered using the multi-branch model with distance regulation, including short-distance, mid-distance and long-distance branch. Using the end-to-end model instead of asynchronous captures only, the final-composited image I x , y for uniform-resolution image can be described as:

(2) I ( x , y ) = F s , m , l ( I i ( x , y ) ; θ )

where I i (x, y) represents the captured raw-defocused image, Fs,m,l represents the deconvolution operation for the end-to-end model in short, middle and long distance, and θ represents the parameters in the model. This integrated imaging strategy avoids complex design and processing, which considerably mitigates blur due to the imperfect imaging and diffraction of the imaging system, in tandem with possible sensor noise.

Figure 3: 
Comparisons of traditional imaging strategy with ours. (a) Traditional imaging pipeline, where blue letters represent focusing state, while gray letters represent defocusing state. (b) Our proposed imaging strategy with distance regulation, where green represents defocusing imaging.
Figure 3:

Comparisons of traditional imaging strategy with ours. (a) Traditional imaging pipeline, where blue letters represent focusing state, while gray letters represent defocusing state. (b) Our proposed imaging strategy with distance regulation, where green represents defocusing imaging.

The detailed architecture for the multi-branch model in Figure 3(b) is illustrated in Figure 4. Figure 4(a) illustrates the multi-branch model is mainly consisting of a designed generator and discriminator. Considering the ambiguity caused by different distances in the real world, a multi-branch strategy based on distance regulation is proposed. For the captured dataset, a slide is specially arranged to ensure that the low resolution (LR) raw images {I i (x, y)} captured by the CE camera are based on different distances. In order to avoid external factors such as jitter as much as possible, we choose a commercial electric zoom lens to capture high-resolution Ground Truth (GT) images {IrealGT}. In theory, according to the arrangement of our PDMS ommatidium array, it can be roughly divided into 10 rings, and the ommatidia on the same ring have the same focal length, object distance, and other characteristics. However, considering that the depth of focus has a certain adjustment range, the imaging range of each ring can take into account a certain area, and it is reasonable to choose the three distances of long, middle, and short distances for adjustment. Here the raw image clusters subject to distance are trained, including short-distance clusters, mid-distance clusters and long-distance clusters. For example, for short-distance clusters, given a set of LR raw images {I i (x, y)} and corresponding GT images {IrealGT}, the same applies to the other two. The designed generator is a variant of Unet structure, which consists of an encoder, multi-scale attention and decoder. Due to regional variations in CE imaging, a piecewise loss is considered, to strengthen the recovery of texture details while reducing the impact of smooth areas. In this work, 700 raw image pairs captured using CE camera and high-performance camera, are trained using our multi-branch model and piecewise loss function. In each branch of the reconstruction model, we attempt to adopt the joint training of adversarial loss [25], perceptual loss [26] and L1 [27] to recover degraded images robustly at different distances.

Figure 4: 
Designed overall architecture with distance regulation for fabricated camera. (a) Overall architecture for designed network. (b) Designed generator. (c) Designed discriminator.
Figure 4:

Designed overall architecture with distance regulation for fabricated camera. (a) Overall architecture for designed network. (b) Designed generator. (c) Designed discriminator.

3.3 Adversarial loss

Adversarial Loss function is used to guide the generator to generate realistic images, here we apply Wasserstein GAN to minimize the Wasserstein distance between predictions and GT for a robust training process. The adversarial loss function can be expressed as:

(3) Advloss = E x Pg D ( x ) E x P r [ D ( x ) ] + λ E x P x x D ( x ) 2 1 2

where Pr represents the GT data {IrealGT}, and Pg represents the generated data { I x , y }.

3.4 Perceptual loss

Perceptual loss is used to improve the visual quality of generated images I x , y , whose goal is to minimize the distance between the generated image and GT image in the feature space. The expression can be described as:

(4) Perceploss = 1 N i = 1 N F s , m , l I ( x , y ) F s , m , l I real GT 2

3.5 L1 loss

L1 loss function is used to reconstruct pixel level similarity, which is equal to calculate the differences between GT images IrealGT and generated images I x , y . The expression can be described as:

(5) L 1 loss = 1 N i = 1 N I real GT I ( x , y )

Hence, the total loss can be calculated using the combination of formulas (3)(5):

(6) Totloss = Advloss + Perceploss + L 1 loss

Using the piecewise loss, all parameters are optimized using the optimization functions.

Figure 4(b) illustrates the overall architecture designed generator, including variant for a combination of Unet and Pyramid-multi-scale attention [28], which enhances the reconstruction for multi-scale texture information. Also, the details of discriminator are illustrated in Figure 4(c), which covers 2D convolution layer, BatchNom layer, LReLU function, fully-connected layer and Sigmoid function. Here the generator is to generate realistic images to deceive the discriminator, and the discriminator is to distinguish the authenticity of an image. Our work is performed on a PC platform (Intel Core i9-10850K CPU @3.6 GHz + GTX3060Ti) equipped with Windows 10 operating system.

3.6 System

The computational-CE camera is a highly integrated system, which has a competitive advantage in scale, weight, and cost. As illustrated in Figure 5(a), the light source and the computational-CE camera are attached to a graduated slide. The computational-CE camera is integrated on a 3 × 3 cm2 circuit board, only with a 5-mm thickness and a weight of 5.4 g, which significantly simplifies system size and complexity. The CE is with a 4-mm radius, as illustrated in Figure 5(b). As illustrated in Figure 5(c), the comparison result demonstrates that the size of CE has an integrated advantage comparing with 1-cent and 25-cent coins. The light source adopts LED illustrated in Figure 5(d) to irradiate the sample, and the distance can be adjusted according to the scale using the slide in Figure 5(e). Here we describe the computational-CE camera with 271 ommatidia, only with a weight of 5.4 g, an area of 3 × 3 cm2 and 5-mm thickness.

Figure 5: 
Experimental setup with compact computational-CE camera. (a) System setup. (b) Computational-CE camera. (c) Comparisons of PDMS CE with coins. (d) Light source. (e) Slide.
Figure 5:

Experimental setup with compact computational-CE camera. (a) System setup. (b) Computational-CE camera. (c) Comparisons of PDMS CE with coins. (d) Light source. (e) Slide.

3.7 Enhanced-contrast reconstruction

Traditional CE cameras perform poorly in the face of natural scene in visible light, which inevitably results in defocusing blur due to the imperfect imaging. However, the experimental result in Figure 6 shows the competitive advantage comparing with the traditional CE camera. The sample I for an Arabic numeral “8” located in the upper right corner is with a defocusing blur due to imperfect imaging in Figure 6(a). However, is recovered effectively in Figure 6(b), and the labeled-yellow region represents the reconstruction detail. It is worth noting that the red and blue line chart are the normalized intensities that vary with the coordinates for the corresponding image, and the positions are the middle lines of the corresponding image (the same follow-up). The sample II for a painted flashlight is reconstructed incompletely in Figure 6(c), such as how many stripes are exactly located on the flashlight. The labeled-yellow in Figure 6(d) illustrates 4 stripes are located on it, which is unreadable from Figure 6(c). The sample III for a painted penguin face is with a blurred reconstruction, as illustrated in Figure 6(e), which cannot be distinguished from the background considerably. As a comparison in Figure 6(f), the reconstructed face using our fabricated camera with the deep learning model is successfully distinguished. For the reconstructed analysis, the normalized intensity statistics varied with coordinate showing greater contrast ratio. Using the computational-CE camera strategy, the experimental result demonstrates significant enhancements in contrast and sharpness.

Figure 6: 
Imaging result comparisons and analysis for our proposed compact computational-CE camera. (a, c, e) Left: the traditional CE camera. (b, d, f) Right: our fabricated CE camera.
Figure 6:

Imaging result comparisons and analysis for our proposed compact computational-CE camera. (a, c, e) Left: the traditional CE camera. (b, d, f) Right: our fabricated CE camera.

In real-world super-resolution reconstruction tasks, efficiency and quality are the golden rules for measuring models. There are classical models that are robust in terms of reconstruction quality and detail, which are often served as the comparison model, such as single image super-resolution (EDSR) [29], flexible style image super-resolution (FxSR) [30], fast super-resolution convolutional neural network (FSRCNN) [31] and efficient sub-pixel convolutional neural network (ESPCN) [32]. As illustrated in Figure 7, we compared the reconstruction results of our model with the world-famous SR models, from which it can be seen that our reconstruction effect is ahead of other models. From the example, the advantage is to resist blurred phenomenon, and the ability to describe texture details is more prominent due to the addition of multiscale attention mechanism. And a quantitative analysis was compared as shown in Table 1. The first column of images represents the input images of different examples, while the 2–6 columns of images represent the results of EDSR, FxSR, FSRCNN, ESPCN and our reconstruction model processing for different examples, respectively. The results of the comparison can be concluded that our reconstruction method is ahead of the reconstruction models in recent years in terms of imaging quality, human eye vision and efficiency. Our reconstruction method has a competitive advantage in quantitative scoring comparing with other SR reconstructed methods.

Figure 7: 
Comparison results with conventional SR reconstruction methods. (a) Input. (b) EDSR. (c) FxSR. (d) FSRCNN. (e) ESPCN. (f) Ours.
Figure 7:

Comparison results with conventional SR reconstruction methods. (a) Input. (b) EDSR. (c) FxSR. (d) FSRCNN. (e) ESPCN. (f) Ours.

Table 1:

Quantitative analysis of comparative results.

Test dataset Upscaling factor Ours EDSR FxSR
PSNR SSIM Time PSNR SSIM Time PSNR SSIM Time
Raw-cap ×4 31.21 0.863 0.11 29.11 0.814 8.29 30.76 0.820 0.35
City-100 30.76 0.890 0.11 29.25 0.828 8.78 30.33 0.826 0.33
Test dataset Upscaling factor FSRCNN ESPCN
PSNR SSIM Time PSNR SSIM Time
Raw-cap ×4 28.47 0.732 0.15 29.48 0.828 0.21 PSNR: dB
City-100 28.16 0.787 0.18 29.66 0.795 0.19 Time: second

3.8 Moving trajectory reconstruction

The fabricated computational-CE camera has competitive advantage in the application in vivo moving trajectory reconstruction. Because insect CE vision is very sensitive to moving objects and has good perception of objects with depth, it is able to provide immediate and effective feedback. Inspired by the insect CE feature, our computational compound eye camera can perceive the position of moving objects in the entire three-dimensional space. To assess the ability in this application, a three-dimensional curved track covered with LED light strips is established, as illustrated in Figure 8(a). A live scarab crawls on a paved track, whose moving trajectory can be recorded using this computational-CE camera. Figure 8(b)–(d) recorded the moving trajectory at different moments using our proposed CE camera for a scarab. The comparison results at 1/10/48 s for the moving live scarab are illustrated in Figure 8(b)–(d), here lower left region represents the raw image without multi-branch reconstruction, and the upper right region represents the reconstructed image with a multi-branch model. The labeled-yellow regions demonstrate enlarged details. At the 1st and 10th seconds, the reconstructed image recovered feet from the degraded image, such as how many (2) feet are exactly displayed next to the torso. From the 1st second to the 48th second, to our findings, we also find that the computational-CE camera first captures the head, then the torso, and finally the buttocks of the scarab if we focus on the upper right region.

Figure 8: 
Recorded moving trajectory at different moments using proposed CE camera for a scarab. (a) A three-dimensional curved track covered with LED light strips. (b–d) Recorded moving trajectory for scarab at different moments. (e) Three-dimensional moving trajectory reconstruction.
Figure 8:

Recorded moving trajectory at different moments using proposed CE camera for a scarab. (a) A three-dimensional curved track covered with LED light strips. (b–d) Recorded moving trajectory for scarab at different moments. (e) Three-dimensional moving trajectory reconstruction.

Unlike a traditional camera that can only determine the object’s distance with known true size of the object, the proposed CE camera enables three-dimensional moving trajectory reconstruction based on the principle of multi-eye vision. When we observe the target objects using the proposed CE camera, ommatidia with different orientations can image the same target from different angles. The reconstruction steps of the scarab’s movement trajectory are as follows: firstly, the proposed compound eye camera is placed in a fixed position and the camera parameters are calibrated. Secondly, the speeded up robust features (SURF) [33] algorithm is used to extract the feature points at different viewing angles of adjacent images. Furthermore, the extracted feature points are effectively matched, and finally, the three-dimensional spatial position of the scarab is measured by triangulation method according to the camera parameters and matching relationship. Based on the recorded information, the moving trajectory reconstruction is reconstructed in Figure 8(e).

3.9 Computational-CE imaging with distance regulation

The comparison results with distance regulation or without it are illustrated in Figure 9(a)–(h). Due to the defocusing blur caused by imperfect imaging, the image details in the left images (a, c, e and g) are blurred, which is unrecognizable. As a comparison, the image details in the right images (b, d, f and h) are significantly recorded. Such as how many fuzzes (3) are exactly located on the foot at 0.5 cm (Figure 9(b)), and such as how many feet (3) are exactly bare on the torso at 2.5 cm (Figure 9(f)), and so on. The experimental result illustrates the multi-branch-regulation reconstruction makes image quality restoration possible at different distances.

Figure 9: 
Comparison results based on distance regulation. (a, c, e, g) Left: results without multi-branch reconstruction. (b, d, f, i) Right: multi-branch reconstruction based on the distance regulation.
Figure 9:

Comparison results based on distance regulation. (a, c, e, g) Left: results without multi-branch reconstruction. (b, d, f, i) Right: multi-branch reconstruction based on the distance regulation.

3.10 Measurement of FOV

In this study, the experimental setup illustrated in Figure 8(a) is used to measure the FOV of our proposed CE camera. Here, the scale is labeled (1 cm). Notably, we can measure the FOV of this computational-CE camera when the entire target object fills the entire acquired image of the camera at a suitable location. As an example, the scarab illustrated in Figure 9(d) fills exactly the entire image, which proves that the FOV can be measured at this time. Based on the known size of the target object d is 2.3 cm and the object distance L is 1.8 cm, therefore, so the FOV θ value can be calculated to be 102° according to the following Equation (7):

(7) θ = 2 arctan d 2 L

4 Conclusions

The topic of integrated compound eyes is a hot topic that has attracted much attention. To overcome the defocusing problem, a deep learning architecture with distance regulation is first proposed to achieve wide-range-clear imaging, without any hardware or specialized front-end design, which greatly reduces system complexity and size. Here we describe the high-resolution computational-CE camera with 271 ommatidia, with a weight of 5.4 g, an area of 3 × 3 cm2 and 5-mm thickness, which achieves compatibility and integration of CE with commercial CMOS. However, the performance of the system is expected to be further improved, the potential ways to further improve the performance include designing multi-layer compound eyes, more robust reconstruction algorithms and so on. In addition, sensor selection is also one of the methods. The compact camera has promising applications in integrated fields such as medical endoscopy, panoramic imaging and vision robotics.


Corresponding author: Lei Li, School of Electronics and Information Engineering, Sichuan University, Chengdu 610065, China, E-mail:

Award Identifier / Grant number: 61927809

Award Identifier / Grant number: 61975139

Acknowledgments

The authors thank the BMF Precision Tech Inc. for fabricating the CE substrate for this work.

  1. Research funding: National Natural Science Foundation of China under Grant No. 61927809 and 61975139.

  2. Author contributions: S.B.L. and X.N.L. contributed equally to this work. S.B.L. and L.L. conceived the project; S.B.L. and X.N.L. designed the computational CE camera; S.B.L. designed the algorithms and analyzed the data; W.J.F. and M.X.Z. conducted the experiments; L.L. revised the original manuscript; all authors discussed the results and commented on the paper.

  3. Conflict of interest: Authors state no conflict of interest.

  4. Data availability: The original code generated and analysed during the current study is available at https://github.com/lsb17844619800/Computational-CE.

References

[1] D. J. Brady, et al.., “Multiscale gigapixel photography,” Nature, vol. 486, no. 7403, pp. 386–389, 2012. https://doi.org/10.1038/nature11150.Search in Google Scholar PubMed

[2] Y. M. Song, et al.., “Digital cameras with designs inspired by the arthropod eye,” Nature, vol. 497, no. 7447, pp. 95–99, 2013. https://doi.org/10.1038/nature12083.Search in Google Scholar PubMed

[3] Z. C. Ma, et al.., “Smart compound eyes enable tunable imaging,” Adv. Funct. Mater., vol. 29, no. 38, p. 1903340, 2019. https://doi.org/10.1002/adfm.201903340.Search in Google Scholar

[4] G. J. Lee, C. Choi, D. H. Kim, and Y. M. Song, “Bioinspired artificial eyes: optic components, digital cameras, and visual prostheses,” Adv. Funct. Mater., vol. 28, no. 24, p. 1705202, 2018. https://doi.org/10.1002/adfm.201705202.Search in Google Scholar

[5] V. Iyer, A. Najafi, J. James, S. Fuller, and S. Gollakota, “Wireless steerable vision for live insects and insect-scale robots,” Sci. Robot., vol. 5, no. 44, p. eabb0839, 2020. https://doi.org/10.1126/scirobotics.abb0839.Search in Google Scholar PubMed

[6] J. Li, et al.., “Ultrathin monolithic 3D printed optical coherence tomography endoscopy for preclinical and clinical use,” Light Sci. Appl., vol. 9, no. 1, pp. 1–10, 2020. https://doi.org/10.1038/s41377-020-00365-w.Search in Google Scholar PubMed PubMed Central

[7] Q. Fan, et al.., “Trilobite-inspired neural nanophotonic light-field camera with extreme depth-of-field,” Nat. Commun., vol. 13, no. 1, p. 2130, 2022. https://doi.org/10.1038/s41467-022-29568-y.Search in Google Scholar PubMed PubMed Central

[8] H. Pahlevaninezhad, et al.., “Nano-optic endoscope for high resolution optical coherence tomography in vivo,” Nat. Photonics, vol. 12, no. 9, pp. 540–547, 2018. https://doi.org/10.1038/s41566-018-0224-2.Search in Google Scholar PubMed PubMed Central

[9] Z. Y. Hu, et al.., “Miniature optoelectronic compound eye camera,” Nat. Commun., vol. 13, no. 1, p. 5634, 2022. https://doi.org/10.1038/s41467-022-33072-8.Search in Google Scholar PubMed PubMed Central

[10] Y. Luo, et al.., “Varifocal metalens for optical sectioning fluorescence microscopy,” Nano Lett., vol. 21, no. 12, pp. 5133–5142, 2021. https://doi.org/10.1021/acs.nanolett.1c01114.Search in Google Scholar PubMed

[11] P. Llull, L. Bange, Z. Phillips, K. Davis, D. L. Marks, and D. J. Brady, “Characterization of the AWARE 40 wide-field-of-view visible imager,” Optica, vol. 2, no. 12, pp. 1086–1089, 2015. https://doi.org/10.1364/optica.2.001086.Search in Google Scholar

[12] J. T. Fan, et al.., “Video-rate imaging of biological dynamics at centimetre scale and micrometre resolution,” Nat. Photonics, vol. 13, no. 11, pp. 809–816, 2019. https://doi.org/10.1038/s41566-019-0474-7.Search in Google Scholar

[13] S. B. Liu, et al.., “Deep learning enables parallel camera with enhanced- resolution and computational zoom imaging,” PhotoniX, vol. 4, no. 1, p. 17, 2023. https://doi.org/10.1186/s43074-023-00095-3.Search in Google Scholar

[14] J. Tanida, et al.., “Thin observation module by bound optics (TOMBO): concept and experimental verification,” Appl. Opt., vol. 40, no. 11, pp. 1806–1813, 2001. https://doi.org/10.1364/ao.40.001806.Search in Google Scholar PubMed

[15] X. W. Cao, et al.., “Single-pulse writing of a concave microlens array,” Opt. Lett., vol. 43, no. 4, pp. 831–834, 2018. https://doi.org/10.1364/ol.43.000831.Search in Google Scholar PubMed

[16] J. Tanida, R. Shogenji, Y. Kitamura, K. Yamada, M. Miyamoto, and S. Miyatake, “Color imaging with an integrated compound imaging system,” Opt. Express, vol. 11, no. 18, pp. 2109–2117, 2003. https://doi.org/10.1364/oe.11.002109.Search in Google Scholar PubMed

[17] D. Floreano, et al.., “Miniature curved artificial compound eyes,” Proc. Natl. Acad. Sci. U. S. A., vol. 110, no. 23, pp. 9267–9272, 2013. https://doi.org/10.1073/pnas.1219068110.Search in Google Scholar PubMed PubMed Central

[18] H. C. Ko, et al.., “A hemispherical electronic eye camera based on compressible silicon optoelectronics,” Nature, vol. 454, no. 7205, pp. 748–753, 2008. https://doi.org/10.1038/nature07113.Search in Google Scholar PubMed

[19] D. Wu, et al.., “High numerical aperture microlens arrays of close packing,” Appl. Phys. Lett., vol. 97, no. 3, p. 031109, 2010. https://doi.org/10.1063/1.3464979.Search in Google Scholar

[20] E. P. Chan and A. J. Crosby, “Fabricating microlens arrays by surface wrinkling,” Adv. Mater., vol. 18, no. 24, pp. 3238–3242, 2006. https://doi.org/10.1002/adma.200601595.Search in Google Scholar

[21] Y. Cheng, J. Cao, Y. Zhang, and Q. Hao, “Review of state-of-the-art artificial compound eye imaging systems,” Bioinspir. Biomim., vol. 14, no. 3, p. 031002, 2019. https://doi.org/10.1088/1748-3190/aaffb5.Search in Google Scholar PubMed

[22] S. H. Park, S. H. Lee, D. Y. Yang, H. J. Kong, and K. S. Lee, “Subregional slicing method to increase three-dimensional nano-fabrication efficiency in two-photon polymerization,” Appl. Phys. Lett., vol. 87, no. 15, p. 154108, 2005. https://doi.org/10.1063/1.2103393.Search in Google Scholar

[23] W. L. Liang, J. G. Pan, and G. D. J. Su, “One-lens camera using a biologically based artificial compound eye with multiple focal lengths,” Optica, vol. 6, no. 3, pp. 326–334, 2019. https://doi.org/10.1364/optica.6.000326.Search in Google Scholar

[24] J. Sun, Q. Han, L. Kou, L. Zhang, K. Zhang, and Z. Jin, “Multi-focus image fusion algorithm based on Laplacian pyramids,” J. Opt. Soc. Am. A, vol. 35, no. 3, pp. 480–490, 2018. https://doi.org/10.1364/josaa.35.000480.Search in Google Scholar

[25] C. Ledig, et al.., “Photo-realistic single image super-resolution using a generative adversarial network,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit, 2017, pp. 4681–4690.10.1109/CVPR.2017.19Search in Google Scholar

[26] S. Karen and Z. Andrew, “Very deep convolutional networks for large-scale image recognition,” ICLR, vol. 1409, no. 1556, 2015.Search in Google Scholar

[27] C. Liu, et al.., “Adaptive smooth l1 loss: a better way to regress scene texts with extreme aspect ratios,” in Proceedings of the 2021 IEEE Symposium on Computers and Communications (ISCC), 2021, pp. 1–7.10.1109/ISCC53001.2021.9631466Search in Google Scholar

[28] H. Wang, C. Wu, J. Chi, X. Yu, Q. Hu, and H. Wu, “Image super-resolution using multi-granularity perception and pyramid attention networks,” Neurocomputing, vol. 443, pp. 247–261, 2021. https://doi.org/10.1016/j.neucom.2021.03.010.Search in Google Scholar

[29] B. Lim, S. Son, H. Kim, S. Nah, and K. M. Lee, “Enhanced deep residual networks for single image super-resolution,” in IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 136–144.10.1109/CVPRW.2017.151Search in Google Scholar

[30] S. H. Park, Y. S. Moon, and N. I. Cho, “Flexible style image super-resolution using conditional objective,” IEEE Access, vol. 10, pp. 9774–9792, 2022, https://doi.org/10.1109/access.2022.3144406.Search in Google Scholar

[31] C. Dong, C. C. Loy, and X. Tang, “Accelerating the super-resolution convolutional neural network,” ECCV, vol. 9906, pp. 391–407, 2016, https://doi.org/10.1007/978-3-319-46475-6_25.Search in Google Scholar

[32] W. Shi, J. Caballero, F. Huszár, J. Totz, and Z. Wang, “Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network,” in Proceedings of the ICCV, 2016, pp. 1874–1883.10.1109/CVPR.2016.207Search in Google Scholar

[33] H. Bay, A. Ess, T. Tuytelaars, and G. L. Van, “Speeded-up robust features (SURF),” Comput. Vis. Image Underst., vol. 110, no. 3, pp. 346–359, 2008. https://doi.org/10.1016/j.cviu.2007.09.014.Search in Google Scholar

Received: 2023-11-07
Accepted: 2024-04-07
Published Online: 2024-04-23

© 2024 the author(s), published by De Gruyter, Berlin/Boston

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 9.5.2024 from https://www.degruyter.com/document/doi/10.1515/nanoph-2023-0782/html
Scroll to top button