Home  >  Article  >  Technology peripherals  >  New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

王林
王林forward
2023-12-14 20:30:55970browse

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on raysPictures

Paper link:https://arxiv.org/pdf/2310.19629

Code link:https://github.com/vLAR-group/RayDF

Homepage: The content that needs to be rewritten is: https://vlar-group.github.io/RayDF.html

Rewritten content: Implementation method:

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

The overall process and components of RayDF are as follows (see Figure 1)

1. Introduction

In the machine Learning accurate and efficient 3D shape representation is very important in many cutting-edge applications in the fields of vision and robotics. However, existing implicit expressions based on 3D coordinates require expensive computational costs when representing 3D shapes or rendering 2D images; in contrast, ray-based methods can efficiently infer 3D shapes. However, existing ray-based methods do not take into account the geometric consistency under multiple viewing angles, making it difficult to recover accurate geometric shapes under unknown viewing angles.

To address these problems, this paper proposes a new maintenance method. RayDF, a ray-based implicit expression method for multi-view geometric consistency. This method is based on a simple ray-surface distance field, by introducing a new dual-ray visibility classifier and a multi-view consistency optimization module. optimization module), learn to obtain a ray-surface distance that satisfies the geometric consistency of multiple viewing angles. Experimental results show that the modified method achieves superior 3D surface reconstruction performance on three data sets and achieves a rendering speed 1000 times faster than the coordinate-based method (see Table 1).

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

The following are the main contributions:

    Using ray-surface distance field to represent three-dimensional shape, this expression is better than Existing coordinate-based representations are more efficient.

  • Designed a new dual-ray visibility classifier. By learning the spatial relationship of any pair of rays, the learned ray-surface distance field can maintain geometric consistency under multiple viewing angles.

  • The accuracy and efficiency of this method in three-dimensional shape reconstruction have been proven on multiple data sets.

2. Method

2.1 Overview

As shown in Figure 1, RayDF includes two networks and an optimization module. For the main network ray-surface distance network, you only need to input a ray to get the distance value between the starting point of the ray and the geometric surface point hit by the ray. Among them, as shown in Figure 2, RayDF uses a ball surrounding the three-dimensional scene to parameterize the input ray, and uses the parameterized four-dimensional sphere coordinates (incidence point and exit point) as network input. For the auxiliary network dual-ray visibility classifier, as input a pair of rays and a geometric surface point, predict the mutual visibility between the two rays. After this auxiliary network is trained, it will play a key role in the subsequent multi-view consistency optimization module.

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

Figure 2 Ray parameterization and network structure of ray-surface distance field

2.2 Dual-ray Visibility Classifier

In this method The auxiliary network of is a binary classifier that predicts whether two input rays can simultaneously see a surface point. As shown in Figure 3, the features obtained from the two input rays are averaged to ensure that the predicted results are not affected by the order of the two rays. At the same time, the features obtained by separately encoding surface points are spliced ​​after the ray features to enhance the ray features and thereby improve the accuracy of the classifier.

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

The framework structure of the dual-ray visibility classifier is shown in Figure 3

2.3 Multi-view Consistency Optimization

Designed The main network ray-surface distance network and the auxiliary network dual-ray visibility classifier are used as a basis, and the key module of multi-view consistency optimization is introduced to conduct two-stage training on the two networks.

(1) First construct the ray pairs for training for the auxiliary network dual-ray visibility classifier. For a ray in a picture (corresponding to a pixel in the picture), the corresponding space surface point can be known through its ray-surface distance. Project it to the remaining viewing angles in the training set to obtain another ray; and this ray There is a corresponding ray-surface distance. The article sets a threshold of 10 mm to determine whether two rays are visible to each other.

(2) The second stage is to train the main network ray-surface distance network to make its predicted distance field meet multi-view consistency. As shown in Figure 4, for a main ray and its surface points, the surface point is uniformly sampled with the center of the sphere to obtain several multi-view rays. Pair the main ray with these multi-view rays one by one, and their mutual visibility can be obtained through the trained dual-ray visibility classifier. Then predict the ray-surface distance of these rays through the ray-surface distance network; if the main ray and a certain sampling ray are mutually visible, then the surface points calculated by the ray-surface distances of the two rays should be the same point; according to The corresponding loss function is designed and the main network is trained, which ultimately enables the ray-surface distance field to meet multi-view consistency.

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

2.4 Surface Normal Derivation and Outlier Points Removal

Since the depth value at the edge of the scene surface often has mutations (discontinuity), and neural The network is a continuous function. The above-mentioned ray-surface distance field can easily predict inaccurate distance values ​​at the edge of the surface, resulting in noise on the geometric surface at the edge. Fortunately, the designed ray-surface distance field has a good feature, as shown in Figure 5. The normal vector of each estimated three-dimensional surface point can be easily found in closed form through automatic differentiation of the network. Therefore, the normal vector Euclidean distance of the surface point can be calculated during the network inference stage. If the distance value is greater than the threshold, the surface point is regarded as an outlier and eliminated, thereby obtaining a clean three-dimensional reconstructed surface.

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

Figure 5 Surface normal calculation

3. Experiments

In order to verify the effectiveness of the proposed method, we performed experiments on three data sets Experiments were conducted on. The three data sets are the object-level synthetic data set Blender [1], the scene-level synthetic data set DM-SR [2], and the scene-level real data set ScanNet [3]. We selected seven baselines for performance comparison. Among them, OF [4]/DeepSDF [5]/NDF [6]/NeuS [7] are coordinate-based level-set methods, DS-NeRF [8] is a depth-supervised NeRF-based method, and LFN [9] and PRIF [10] are two ray-based baselines

Due to the ease of the RayDF method to directly add a radiance branch to learn textures, it can be compared with baseline models that support predicting radiance fields. Therefore, the comparative experiments of this paper are divided into two groups. The first group (Group 1) only predicts distance (geometry), and the second group (Group 2) predicts both distance and radiance (geometry and texture)

3.1 Evaluation on Blender Dataset

As can be seen from Table 2 and Figure 6, in Group 1 and 2, RayDF achieved better results in surface reconstruction, especially in the most important ADE indicator. Better than coordinate- and ray-based baselines. At the same time, in terms of radiance field rendering, RayDF also achieved performance comparable to DS-NeRF and better than LFN and PRIF.

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

Figure 6 Visual comparison of Blender data set

3.2 Evaluation on DM-SR Dataset

As can be seen from Table 3, RayDF surpasses all baselines in the most critical ADE indicator. At the same time, in the Group 2 experiment, RayDF was able to obtain high-quality new view synthesis while ensuring that the accurate surface shape was restored (see Figure 7).

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

Figure 7 Visual comparison of DM-SR data set

3.3 Evaluation on ScanNet Dataset

Table 4 compares the performance of RayDF and baselines in challenging real-world scenarios. In the first and second groups, RayDF significantly outperforms baselines in almost all evaluation metrics, showing clear advantages in recovering complex real-world 3D scenes

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

The following is a rewrite of the visual comparison of Figure 8 ScanNet dataset: In Figure 8, we show the visual comparison results of the ScanNet dataset

3.4 Ablation Study

We conducted an ablation experiment on the Blender dataset. Table 5 in the paper shows the key The ablation experimental results of the dual-ray visibility classifier

  • are shown in Table 5 (1). Without the help of the dual-ray visibility classifier, the ray-surface distance field will not be able to detect the new angle of view. The rays predict reasonable distance values ​​(see Figure 9).
  • In the input of the classifier, the input surface point coordinates are selected as auxiliary, as shown in Table 5 (2) and (3), if the surface point distance value is selected as auxiliary or not Providing auxiliary information, the classifier will obtain lower accuracy and F1 score, resulting in insufficient visibility information provided for the ray-surface distance network, thereby predicting incorrect distance values.
  • As shown in Table 5 (4), by inputting a pair of rays in an asymmetric manner, the trained classifier has a higher accuracy but a lower F1 score. This shows that this classifier is significantly less robust than a classifier trained with symmetric input rays.

Other resection operations can be viewed in the paper and the paper appendix

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

need to be re- The written content is: Figure 9 shows the visual comparison of using a classifier and not using a classifier

4. Conclusion

When using the ray-based multi-view consistency framework for research, the paper A conclusion is drawn that three-dimensional shape representations can be learned efficiently and accurately through this method. In the paper, a simple ray-surface distance field is used to represent the geometry of three-dimensional shapes, and a novel dual-ray visibility classifier is used to further achieve multi-view geometric consistency. Experiments on multiple data sets have proven that the RayDF method has extremely high rendering efficiency and excellent performance. Further extensions to the RayDF framework are welcome. You can view more visualization results on the homepage

The content that needs to be rewritten is: https://vlar-group.github.io/RayDF.html

New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays

The content that needs to be rewritten is: Original link: https://mp.weixin.qq.com/s/dsrSHKT4NfgdDPYcKOhcOA

The above is the detailed content of New title: Real-time rendering evolved! Innovative method of 3D reconstruction based on rays. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:51cto.com. If there is any infringement, please contact admin@php.cn delete