Skip to main content

ORIGINAL RESEARCH article

Front. Phys., 16 November 2022
Sec. Optics and Photonics
This article is part of the Research Topic Silicon-Based Optoelectronics and Microelectronics Integration Technology View all 7 articles

A digital hologram-based encryption and compression method for 3D models

Yukai SunYukai Sun1Meirong Zhao
Meirong Zhao1*Pingjuan NiuPingjuan Niu2Yelong ZhengYelong Zheng1Chuntian LiuChuntian Liu1Shiyu ZhangShiyu Zhang1Songqi BaiSongqi Bai1Jinwei GuoJinwei Guo1
  • 1State Key Laboratory of Precision Measuring Technology and Instruments, Tianjin University, Tianjin, China
  • 2Tianjin Key Laboratory of Optoelectronic Detection Technology and System, School of Electronics and Information Engineering, Tiangong University, Tianjin, China

This study proposes a novel method to compress and decompress the 3D models for safe transmission and storage. The 3D models are first extracted to become 3D point clouds, which would be classified by the K-means algorithm. Then, these nearby 3D point clouds are converted into a computer-generated hologram (CGH) by calculating the point distribution on the hologram plane using the optical wavefront propagation method. The computer-generated hologram (CGH) contains the spatial coordinate information on point clouds, which can be decompressed using the convolutional neural network (CNN) method. The decompression accuracy of 3D point clouds is quantitatively assessed by normalized correlation coefficients (NCCs), reflecting the correlation between two points and influenced by the hologram resolution, the convolution kernel, and the diffraction distance. Numerical simulations have shown that the novel method can reconstruct a high-quality 3D point cloud with an accuracy of 0.1 mm.

1 Introduction

With the rapid development of 3D data acquisition equipment and technology, the location information on 3D objects and scenes can be collected and transmitted in real time [14]. Since discrete point clouds in 3D space contain depth information and can accurately reproduce real-world scenes, they can be used to record and reconstruct 3D scenes [57]. Those 3D point clouds can be used in visual displays such as city modeling [8,9] and mobile applications of interaction [10,11]. In particular, the point clouds consisting of a large number of 3D points can accurately represent 3D environments and enable more immersive visual experiences. It will give rise to many new multimedia applications, such as virtual reality and augmented reality [12].

However, the accurate reconstruction of point clouds is positively related to the total amount of data in the point clouds, so the reconstructed point cloud data that can meet the visual needs require huge storage space and transmission bandwidth [13]. It is important to develop a 3D point cloud compression (3PCC) algorithm with high secrecy, low complexity, and high decompression accuracy for real-time applications. In the last decade, many attempts have been made to improve the compression efficiency. Jae Kyun et al. [7] proposed an adaptive technique of range image coding for compressing large-scale 3D point clouds, which adaptively predicts the radial distance of each pixel using its previously encoded neighbors. Queiroz et al. [14] investigated a technique to compress point cloud colors based on hierarchical transformations and arithmetic coding. Vicente Morell et al. [15] proposed a 3D lossy compression system based on plane extraction. These proposed methods mainly reduce the number of point clouds or improve the presentation of point clouds. With the current hardware conditions, it is difficult to store and transmit a large amount of point cloud data.

In this article, we propose a highly encrypted 3D point cloud compression and dimensionality reduction algorithm for 3D point cloud transmission and storage. Figure 1 shows an overview of the compression and decompression algorithm.

FIGURE 1
www.frontiersin.org

FIGURE 1. Flow diagram of the proposed compression and decompression algorithm.

2 Methods

In a 3D coordinate system, the outer surface of a 3D object can be represented by point clouds, which are a series of data points defined by X, Y, and Z coordinates. The point cloud acquisition is from 3D scanners or directly from 3D applications. Figure 2 shows the two point clouds using each of the aforementioned two methods. Figure 2A is generated by using a 3D scanner (FARO Focus3D X130), and Figure 2B is created by a 3D application.

FIGURE 2
www.frontiersin.org

FIGURE 2. Point clouds from (A) the 3D scanner and (B) 3D application.

This study mainly focuses on the compression and encryption of point cloud positions and does not consider the color values of the point clouds. We first segmented the point cloud using an unsupervised clustering segmentation method (the K-means algorithm). It can divide the point cloud into k clusters to ensure that each point is located in a cluster. The clustering of point clouds is the grouping of points by specific features. We initially needed to determine the number of clusters (k) and the centers of k clusters randomly. The simple Euclidean function, Minkowski function, and Cityblock function were applied to calculate the distance between each point and the cluster center. The point would be moved to the particular cluster to keep the shortest distance from the cluster center [16]. Figure 3 shows the flow chart and diagram of the K-means algorithm.

FIGURE 3
www.frontiersin.org

FIGURE 3. (A) Flow chart and (B) diagram of the K-means algorithm.

To achieve point cloud data encryption and compression, we calculated the point distribution on the hologram plane by the propagation of the optical wavefront method, which needs to be carried out in two steps: 1) the points in the point cloud can be assumed to be point sources, as shown in Figure 4, whose distribution in the plane in front of the Fourier lens needs to be calculated. 2) The Fourier lens phase data, which are optical transmissions between two parallel planes, also need to be added to the distribution of the final hologram plane [1719]. The phase distribution is crucial in the reconstruction process of the 3D point cloud since the phase data contain the direction information.

FIGURE 4
www.frontiersin.org

FIGURE 4. Schematic diagram of point cloud hologram generation. The hologram includes depth information.

In the propagation of the optical wavefront method, the phase distribution on the final hologram plane precisely records the real-world object. It can calculate and simulate the optical transmission process from the point source to the hologram plane [20,21]. The complex amplitude H(x, y) on the midplane of the point clouds can be calculated and simulated by the optical wavefront.

Hx,y=j=1kAjrjexpikrj,j=1,,k,(1)

where k is the sparse point number of a point cloud after segmentation, Aj is the wave amplitude, and in our work, Aj = 1, and k=2πλ is the wave number in space. The distance rj between the point (xj, yj, zj) in the point cloud and the point (x, y, 0) in the midplane is given by

rj=x2xj2+y2yj2+zj2.(2)

The convolutional neural network (CNN) can be used to deal with the input fluctuations and obtain the invariance characteristics. It is commonly used in image data and is a variation of the MLP that evolved from the idea of biology. There is a complex distribution of cells in the visual cortex that are sensitive to external inputs, called the “receptive field.” These cells are locally sensitive to the input image and are able to mine images for targets, functioning like a filter. Each filter is superimposed over the entire receptive field. The CNN can identify the local information on objects by enhancing the local connectivity patterns (LCPs) of nodes between adjacent layers.

Although the CNN has dominated many tasks in computer vision [2225] and object detection [2628], the application of the CNN in point cloud processing is still lacking. Most of the approaches are still based on the TIN framework, which leads to limited accuracy and processing speed.

In this work, we apply the CNN algorithm to recognize the hologram and reconstruct the point clouds. In our approach, we need the weight of each feature as a code to find the position of the point cloud. The data must first be preprocessed using the following techniques: data regularization and data degradation. It can increase the recognition rate because the non-zero mean of the training sample does not match the non-zero mean of the testing sample. Also, different image features can be obtained by using different convolution kernels. These convolution kernels are similar to different types of filters and are the core of the CNN. A pooling layer follows the convolutional layer to reduce the dimension. In general, the size of the original convolution output matrix is transformed to half of its original size. The pooling layer would increase the robustness of the system and turn the original detailed description into a coarse description (originally, the matrix size was 28 × 28, and now, it is 14 × 14, but the boundary would lose partial information). Two pooling operations (max pooling and mean pooling) would improve the variance of the extracted features. However, the error in feature extraction would come from two ways: 1) the estimate variance increasing caused by the limited size of the neighborhood. 2) estimated mean shifting caused by the convolutional parameter error. Specifically, mean pooling can reduce the first error and preserve the background information about the image. Max pooling can reduce the second error (the derivative does not affect the other points) and preserve texture information.

As shown in Figure 5, the histogram feature vector (HFV) needs to be extracted before rebuilding the point cloud from the hologram. The HFV consists of vertical projection and horizontal projection of the hologram. Then, convolutions and subsampling are performed to get a feature map of 8 × 8 pixels. The feature map is used to reconstruct the feature vector. Also, the computer-generated holograms can be extracted from the feature vectors and numbered to form a look-up table. When a point cloud hologram is decompressed, the CNN method identifies the encoding of holograms. With the encoding, we can find the unsampled hologram in the look-up table and calculate its spatial value.

FIGURE 5
www.frontiersin.org

FIGURE 5. Process of the CNN based on the trained dictionary.

3 Results

Figure 2B is the point cloud given as input to the preprocessing and K-means algorithms. The algorithm calculates the clustering of the point cloud using the Cityblock function. Figure 6 is the output for spatial segmentation by the K-means algorithm with k = 20.

FIGURE 6
www.frontiersin.org

FIGURE 6. Point clouds are segmented by the K-means algorithm with k = 20.

For the given point cloud, the cluster means mk is calculated as follows:

mk=ikndin,k=1,,K.(3)

The distance D(k) between the cluster and each point is calculated as follows:

Dk=arg mindikmk2,k=1,,K.(4)

Iteratively compute the preceding two steps until the mean value convergence is attained.

Four original models are chosen as the compression objects, as shown in Figures 7A1–A4. The simulation of the models is as follows: 1) the models first generate four groups of point clouds by the sampling process, respectively, as shown in Figures 7B1–B4. 2) The K-means method is applied to simplify the data while preserving the data information since the point cloud of the 3D model has a huge amount of data. For example, the point cloud of the first model (a1) has 10,684 points, the second model (a2) has 1,177 points, the third model (a3) has 35,871 points, and the last model (a4) has 40,248 points. As shown in Figures 7C1–C4, the set of points could (b1–b4) be divided into k groups by the principle of the nearest distance. The k value of (c1), (c2), (c3), and (c4) is 174, 84, 171, and 130, respectively. 3) The k point is transformed into a hologram, calculating the point distribution on the hologram plane. The holograms in Figures 8A1–A4 are generated with the following parameters: the diffraction distance is z = 240 mm, the hologram has 512 × 512 pixels, and the wavelength of the laser used in our experiment is 632.8 nm. 4) Using the CNN algorithm to decompress the hologram, get the point cloud code, and look up the table to know the point cloud distribution.

FIGURE 7
www.frontiersin.org

FIGURE 7. 3D models and point clouds. (A1–A4) 3D models. (b1–b4) Point clouds generated from the models by the sampling process. (C1–C4) Point clouds are divided into k groups.

FIGURE 8
www.frontiersin.org

FIGURE 8. Holograms and decompression of the point cloud. (A1–A4) Holograms generated from the four models. (B1–B4) Point clouds decompressed from holograms. (C1–C4) Simulation results for comparison.

The models and point clouds are shown in Figure 7. The point clouds are divided by the K-means method and kc1 = 174, kc2 = 84, kc3 = 171, and kc4 = 130. The k cluster centers are used as point sources of holograms (see Figures 7C1–C4). The hologram contains the location information on point clouds. The compression and decompression results of the models are shown in Figure 8. The holograms are generated from the point sources and by the propagation of the optical wavefront method, as shown in Figures 8A1–A4. Figures 8B1–B4 shows the decompression points by the CNN method, respectively. Simulation results (Figures 8C1–C4) show that the decompression point clouds have high decompression accuracy and low crosstalk with the number of point clouds increasing.

4 Discussion

The discrete point cloud is compressed using virtual light with a wavelength of 632.8 nm. The compression layer resolution is 256 × 256 pixels, and the compression distance is 24 cm. When compressing a single discrete point cloud in different positions (0, 0, 0), (0, 0.1, 0), and (0.1, 0, 0) and analyzing the compressed data, we can find that the encoding is completely different, although the difference is only 0.1 mm (Figure 9). Therefore, the hologram has enough coding expression space to store the point cloud data, and the three-dimensional space information is converted into two-dimensional information. Fresnel diffraction is performed on the discrete point to transform voxels into pixels, which ensures that the original data and information are preserved in the case of dimensionality reduction.

FIGURE 9
www.frontiersin.org

FIGURE 9. Coding of point sources in different positions.

In the proposed method, the hologram converted from the point clouds can record the space position through diffraction. When decompressing the hologram, the accuracy of point clouds is affected by the hologram resolution, the convolution kernel, and the diffraction distance. The accuracy of decompression point clouds can be quantitatively assessed by normalized correlation coefficients (NCCs). It reflects the correlation between two points and can be expressed as Eq. 5, where R0(i, j) expresses the feature vector of the point (0, 0, 0), R1(i, j) indicates the feature vector of the point (0.1, 0, 0), and (I, j) denotes an element in the feature vector. h and l represent the row and column numbers of the feature vector, respectively.

NCC=i=1hi=1lR0i,jR1i,ji=1hi=1lR0i,jR0i,j.(5)

The NCCs of the hologram feature vector under different parameters are shown in Table 1. There are correlations in the holograms with 512 × 512 pixels, 256 × 256 pixels, and 128 × 128 pixels; the diffraction distance with 20 cm, 24 cm, and 28 cm; and the convolution kernel with 3 × 3 pixels, 6 × 6 pixels, and 9 × 9 pixels. The different correlation coefficients are in the range of 0.0272–0.5944, which show that different parameters have different correlation coefficients. The smaller correlation coefficient indicates that the spatial point cloud accuracy is higher. In particular, the hologram (512 × 512 pixels and z = 24 cm) and the convolution kernel (3 × 3) can extract point clouds with a spatial resolution of 0.1 mm.

TABLE 1
www.frontiersin.org

TABLE 1. NCC in different parameters.

5 Conclusion

In this study, a 3D point cloud compression and decompression method based on holograms and the CNN is proposed. The sparse 3D point clouds generated from the 3D model are transformed into a hologram, and the position of the point cloud can be extracted by the CNN method in the decompression process. The hologram can reduce the dimension of spatial data and realize the three-dimensional data recording. The accuracy of the decompression point clouds would be affected by the hologram resolution, the convolution kernel, and the diffraction distance. Under the conditions of the hologram with 512 × 512 pixels and z = 24 cm, the point cloud can be reconstructed with an accuracy of 0.1 mm. Simulation results and quantitative analysis based on the hologram and CNN method show that the proposed method can meet the demand for three-dimensional model transmission and storage.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material; further inquiries can be directed to the corresponding author.

Author contributions

Conceptualization, YS and MZ; methodology, PN and CL; software, SZ; formal analysis, SB; investigation, JG. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (No. 2020YFC2008703).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Alexiadis DS, Zarpalas D, Daras P. Real-time, full 3-d reconstruction of moving foreground objects from multiple consumer depth cameras. IEEE Trans Multimedia (2013) 15:339–58. doi:10.1109/TMM.2012.2229264

CrossRef Full Text | Google Scholar

2. Zhang C, Cai Q, Chou PA, Zhang Z, Martin-Brualla R. Viewport: A distributed, immersive teleconferencing system with infrared dot pattern. IEEE MultiMedia (2013) 20:17–27. doi:10.1109/MMUL.2013.12

CrossRef Full Text | Google Scholar

3. Masuda H, He J. Tin generation and point-cloud compression for vehicle-based mobile mapping systems. Adv Eng Inform (2015) 29:841–50. doi:10.1016/j.aei.2015.05.007

CrossRef Full Text | Google Scholar

4. Li F, Zhao H. 3d real scene data collection of cultural relics and historical sites based on digital image processing. Comput Intelligence Neurosci (2022) 2022:1–12. doi:10.1155/2022/9471720

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Mitchell DJ, Tiddy GJT, Waring L, Bostock T, McDonald MP. Phase behaviour of polyoxyethylene surfactants with water. mesophase structures and partial miscibility (cloud points). J Chem Soc Faraday Trans 1 (1983) 79:975–1000. doi:10.1039/F19837900975

CrossRef Full Text | Google Scholar

6. Rusu RB, Cousins S. 3d is here: Point cloud library (pcl). In: IEEE International Conference on Robotics and Automation, ICRA 2011; 9-13 May 2011; Shanghai, China (2011). p. 1–4. doi:10.1109/ICRA.2011.5980567

CrossRef Full Text | Google Scholar

7. Ahn J-K, Lee K-Y, Sim J-Y, Kim C-S. Large-scale 3d point cloud compression using adaptive radial distance prediction in hybrid coordinate domains. IEEE J Sel Top Signal Process (2015) 9:422–34. doi:10.1109/JSTSP.2014.2370752

CrossRef Full Text | Google Scholar

8. Pylvänäinen T, Berclaz J, Korah T, Hedau V, Aanjaneya M, Grzeszczuk R. 3d city modeling from street-level data for augmented reality applications. In: 3D Imaging, Modeling, Processing, Visualization and Transmission (3DIMPVT), 2012 Second International Conference, October 13-15, 2012, Zurich, Switzerland (2012). p. 238–45. doi:10.1109/3DIMPVT.2012.19

CrossRef Full Text | Google Scholar

9. Wu Y, Zhou Z. Intelligent city 3d modeling model based on multisource data point cloud algorithm. J Funct Spaces (2022) 2022:1–10. doi:10.1155/2022/6135829

CrossRef Full Text | Google Scholar

10. Korah T, Tsai Y-T. Urban canvas: Unfreezing street-view imagery with semantically compressed lidar pointclouds. In: 10th IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2011; October 26-29, 2011; Basel, Switzerland (2011). p. 271–2. doi:10.1109/ISMAR.2011.6143897

CrossRef Full Text | Google Scholar

11. Xu Y, Tong X, Stilla U. Voxel-based representation of 3d point clouds: Methods, applications, and its potential use in the construction industry. Automation in Construction (2021) 126:103675. doi:10.1016/j.autcon.2021.103675

CrossRef Full Text | Google Scholar

12. Waschbüsch M, Würmlin S, Cotting D, Gross M. Point-sampled 3d video of real-world scenes. Signal Processing: Image Commun (2007) 22:203–16. Special issue on three-dimensional video and television. doi:10.1016/j.image.2006.11.009

CrossRef Full Text | Google Scholar

13. Zhang Y, Kwong S, Jiang G, Wang H. Efficient multi-reference frame selection algorithm for hierarchical b pictures in multiview video coding. IEEE Trans Broadcast (2011) 57:15–23. doi:10.1109/TBC.2010.2082670

CrossRef Full Text | Google Scholar

14. de Queiroz RL, Chou PA. Compression of 3d point clouds using a region-adaptive hierarchical transform. IEEE Trans Image Process (2016) 25:3947–56. doi:10.1109/TIP.2016.2575005

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Morell V, Orts S, Cazorla M, Garcia-Rodriguez J. Geometric 3d point cloud compression. Pattern Recognition Lett (2014) 50:55–62. doi:10.1016/j.patrec.2014.05.016

CrossRef Full Text | Google Scholar

16. Selvakumar J, Lakshmi A, Arivoli T. Brain tumor segmentation and its area calculation in brain mr images using k-mean clustering and fuzzy c-mean algorithm. In: IEEE-International Conference On Advances In Engineering, Science And Management (ICAESM -2012); 30-31 March 2012; Nagapattinam, India (2012). p. 186–90.

Google Scholar

17. Zhang H, Zhao Y, Gu H, Tan Q, Cao L, Jin G. Fourier holographic display system of three-dimensional images using phase-only spatial light modulator. Proc SPIE (2012) 8556:03. doi:10.1117/12.999268

CrossRef Full Text | Google Scholar

18. Zhang H, Zhao Y, Wang Z, Cao L, Jin G. Calculation for high-resolution computer-generated hologram with fully-computed holographic stereogram. In: Proceedings of SPIE - The International Society for Optical Engineering 9271, August 17-21, 2014, San Diego, California (2014). p. 92711T. doi:10.1117/12.2071617

CrossRef Full Text | Google Scholar

19. Jiang Q, Jin G, Cao L. When metasurface meets hologram: Principle and advances. Adv Opt Photon (2019) 11:518–76. doi:10.1364/AOP.11.000518

CrossRef Full Text | Google Scholar

20. Zhang H, Zhao Y, Cao L, Jin G. Fully computed holographic stereogram based algorithm for computer-generated holograms with accurate depth cues. Opt Express (2015) 23:3901. doi:10.1364/OE.23.003901

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Schnars U. Direct phase determination in hologram interferometry with use of digitally recorded holograms. J Opt Soc Am A (1994) 11:2011–5. doi:10.1364/JOSAA.11.002011

CrossRef Full Text | Google Scholar

22. Chua L, Roska T. The cnn paradigm. IEEE Trans Circuits Syst (1993) 40:147–56. doi:10.1109/81.222795

CrossRef Full Text | Google Scholar

23. Roska T, Chua L. The cnn universal machine: An analogic array computer. IEEE Trans Circuits Syst (1993) 40:163–73. doi:10.1109/82.222815

CrossRef Full Text | Google Scholar

24. Xie G-S, Zhang X-Y, Yan S, Liu C-L. Hybrid cnn and dictionary-based models for scene recognition and domain adaptation. IEEE Trans Circuits Syst Video Technol (2017) 27:1263–74. doi:10.1109/TCSVT.2015.2511543

CrossRef Full Text | Google Scholar

25. Yu C, Han R, Song M, Liu C, Chang C-I. A simplified 2d-3d cnn architecture for hyperspectral image classification based on spatial–spectral fusion. IEEE J Sel Top Appl Earth Obs Remote Sens (2020) 13:2485–501. doi:10.1109/JSTARS.2020.2983224

CrossRef Full Text | Google Scholar

26. Chua LO. Cnn: A vision of complexity. Int J Bifurcation Chaos (1997) 07:2219–425. doi:10.1142/S0218127497001618

CrossRef Full Text | Google Scholar

27. Razavian A, Azizpour H, Sullivan J, Carlsson S. Cnn features off-the-shelf: An astounding baseline for recognition. In: 2014 IEEE conference on computer vision and pattern recognition workshops, June 24-27, 2014, Columbus, Ohio (2014). p. 1403. doi:10.1109/CVPRW.2014.131

CrossRef Full Text | Google Scholar

28. Ren S, He K, Girshick R, Sun J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell (2017) 39:1137–49. doi:10.1109/TPAMI.2016.2577031

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: 3D point clouds, compression and decompression, CNN, 3D models, hologram

Citation: Sun Y, Zhao M, Niu P, Zheng Y, Liu C, Zhang S, Bai S and Guo J (2022) A digital hologram-based encryption and compression method for 3D models. Front. Phys. 10:1063709. doi: 10.3389/fphy.2022.1063709

Received: 07 October 2022; Accepted: 24 October 2022;
Published: 16 November 2022.

Edited by:

Hongda Chen, Institute of Semiconductors (CAS), China

Reviewed by:

Qun Ma, Jinhang Technical Physics Research Institute, China
Weiming Yang, Tianjin University of Science and Technology, China
Ying Tong, Tianjin Normal University, China

Copyright © 2022 Sun, Zhao, Niu, Zheng, Liu, Zhang, Bai and Guo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Meirong Zhao, bWVpcm9uZ3poYW9fYWNhZEAxNjMuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.