An Efficient End-to-End 3D Voxel Reconstruction based on Neural Architecture Search
Abstract
Using neural networks to represent 3D objects has become popular. However, many previous works employ neural networks with fixed architecture and size to represent different 3D objects, which lead to excessive network parameters for simple objects and limited reconstruction accuracy for complex objects. For each 3D model, it is desirable to have an end-to-end neural network with as few parameters as possible to achieve high-fidelity reconstruction. In this paper, we propose an efficient voxel reconstruction method utilizing neural architecture search (NAS) and binary classification. Taking the number of layers, the number of nodes in each layer, and the activation function of each layer as the search space, a specific network architecture can be obtained based on reinforcement learning technology. Furthermore, to get rid of the traditional surface reconstruction algorithms (e.g., marching cube) used after network inference, we complete the end-to-end network by classifying binary voxels. Compared to other signed distance field (SDF) prediction or binary classification networks, our method achieves significantly higher reconstruction accuracy using fewer network parameters.
I Introduction
With the increasing development of three-dimensional (3D) deep learning, the task of 3D object representation and reconstruction has become a research hotspot. 3D objects can be represented explicitly or implicitly. Common explicit representations include point clouds, meshes, and voxels, among others. The point clouds representation stores the position of each point, and may also contain color and normal vectors, etc. Mesh establishes the connectivity between points and forms facets for rendering models. Voxel is a dense grid representation that requires a lot of storage. In addition, the octree representation can be used to dynamically adjust the spatial resolution based on the local details of the object. Along with the above explicit representations, neural networks have accomplished various 3D tasks in the areas of computer vision and computer graphics. For example, PointNet [1] first uses MLP layers to obtain the high-dimensional feature of each point, which is then used for object classification and segmentation. Volumetric 3D convolutional neural network (CNN) is introduced in VoxNet [2], and is still used for object classification. ONet [3] extracts latent vectors of one category of objects from their single-view images, point clouds, or voxels, and accomplishes reconstruction by predicting the occupancy of voxels in a reconstructed model.
Among the implicit representations, the signed distance field (SDF) is the most popular since it is a continuous and probability-like representation. For a 3D point, the closer it is to the surface of the object, the smaller the absolute value of its SDF. Based on SDF representation, neural networks have achieved good performance in many tasks, especially in object reconstruction. For example, NI [4] first proposes to use a multi-layer perception (MLP) network to overfit the SDF for each individual object. Therefore, neural compression (through storing network parameters) and reconstruction (through network inference and subsequent surface reconstruction) are achieved. NGLOD [5] adopts the idea of local fitting to significantly improve the reconstruction accuracy for an individual model. The reconstruction error of NI or NGLOD designed for an individual object is obviously lower than the above mentioned ONet designed for one category of objects. However, the compression task is ignored by NGLOD, since the method requires storing a large number of latent vectors of grid nodes, possibly even more than the number of vertices and faces of the model itself. Different from the reconstruction task of 3D models, the recent famous NeRF works [6] [7][8] encode 3D scenes, including shapes, texture and illumination, from a set of calibrated images. Although they also use MLP networks to predict opacity that is similar to SDF, their goal is nearly irrelevant to the neural compression and reconstruction of known 3D models, which is the focus of this article.

Faced with the success of SDF prediction networks, the first problem is why all existing high-precision reconstruction methods predict SDF. Theoretically, it is much harder for a neural network to predict the SDF value of each point in the space than to predict its occupancy. However, to the best of our knowledge, there is no neural network proposed to classify binary voxels for high-fidelity single-object reconstruction. Predicting the occupancy of binary voxels with the same MLP network should yield higher reconstruction accuracy and fewer failures than predicting SDF values.
The second problem is that in MLP networks for model reconstruction, the number of layers, the number of neurons in each layer, and the activation functions are all set to be fixed. However, the complexity of each model is different. 3D models may be convex or concave, with or without holes. In this sense, the number of layers and the number of neurons in each layer used to correctly reconstruct objects should be different. On the other hand, activation functions tend to behave differently in different datasets [11]. For the task of model reconstruction, each object is equivalent to a dataset with different shapes and data distributions. Naturally, for an individual object, the activation function of each layer selected by learning will be more suitable for model reconstruction.
In this paper, we propose an efficient end-to-end 3D voxel reconstruction based on neural architecture search (NAS) [12] [13]. NAS can find a specific network for an individual 3D object in terms of the number of layers, the number of nodes in each layer, and the activation function in each layer. Moreover, directly predicting the occupancy of voxels not only alleviates the fitting difficulty for complex objects, but also avoids the use of surface reconstruction algorithms after obtaining SDF values. The solution of the two problems mentioned above brings about a significant improvement in the reconstruction accuracy. The comparison of six reconstructed voxel models for four methods with their ground truth (GT) is shown in Fig. 1. The voxels in the -st row is the simplest of the six, and all four methods achieve roughly the same reconstruction accuracy. The voxels in the -nd and -th rows become more complex, and NI fails to reconstruct them correctly. While the voxels in the -rd row is not very complicated in geometric shape, only our result is visually acceptable. For the voxels in the -th row, the bow details of the boat can only be correctly reconstructed by our method. The voxels in the -th row is the most complex. For our result, although the legs of the stools are not reconstructed well, there are no noticeable errors in the arresting surface of several stools in all results. Moreover, only the proposed method has the adaptive number of network parameters for different objects, realizing the idea that simpler objects should have fewer parameters of neural representation.
Our contributions are summarized as follows:
1) MLP is used to directly predict the occupancy of each voxel, which significantly improves the reconstruction accuracy. Meanwhile, this end-to-end approach avoids the surface reconstruction required by SDF prediction methods.
2) The network architecture search (NAS) technology is used to find a specific network architecture for each object. The number of the network parameters can vary adaptively with the complexity of the object.
3) The network size is added to the reward, and a post-processing step after NAS is designed. By doing so, the number of network parameters is further reduced, while maintaining almost the same accuracy.
II Related Works
This paper is mainly related to two directions, which are the neural implicit reconstruction of 3D models and the technology of network architecture search. They will be illustrated in the following two subsections, respectively.
II-A Neural Implicit Reconstruction
With the development of 3D deep learning, there is a growing body of work studying implicit neural representation and reconstruction. Here we investigate two sub-directions closely related to our work. The first sub-direction in implicit neural representation and reconstruction is the prediction of SDFs from dense samples using MLP networks [14] [15] [16] [4] [5] [17] [18]. For example, DeepSDF [14] early proposes to learn and reconstruct continuous SDFs for a category of 3D objects using an MLP network. FFN [15] maps Fourier features and learns high-frequency functions in low-dimensional domains to improve the fitting capability of MLP. To overcome the difficulty of fitting SDFs of one category of objects, NI [4] firstly proposes to overfit the global shape of an individual object with a small MLP network. This method actually implements a lossy compression of a 3D model by storing MLP parameters. However, the default network with 7553 parameters may fail in reconstruction, especially for complex objects. NGLOD [5] learns the latent vectors of octree vertices in different levels of details (LOD) to predict local SDFs of an object. Although the reconstruction accuracy can be improved obviously, the storage capacity in this local-fitting method is greatly increased as the latent vectors of a large number of LOD vertices needs to be stored.
The second sub-direction in implicit neural reconstruction is the prediction of occupancy of voxels using encoder-decoder networks [3] [19] [20]. Similar to ONet [3] reviewed in Sec. I, IM-Net [19] learns the generative models of shapes for one category of objects. CONet [20] combines convolutional encoders with implicit occupancy decoders to represent detailed reconstruction of objects and 3D scenes. However, the reconstruction accuracy of these works is obviously lower than the SDF prediction networks. Moreover, the network size of these works is much larger than most of MLP networks.
In theory, predicting occupancy of points is much easier than predicting their SDF values with the same network and training data. Therefore, in this paper, we adopt this idea to directly predict the occupancy of binary voxels.
II-B network architecture search
Neural Architecture Search (NAS) methods essentially aim to provide an automated way to design architectures as an alternative to manual architectures. Our work is closely related to reinforcement learning based NAS work [12] [13] [21] [22] [23]. For example, owing to the weight sharing idea, ENAS [23] can significantly reduce the computational power required to traverse the search space [21]. Although later researchers propose a different framework DARTS [24] [25] [26], this kind of approach is not suitable for MLP architectures search.
In addition, the process of NAS needs to accurately evaluate the performance of each network architecture. A straightforward solution is to train an architecture from scratch and then test it on the validation dataset, which is very time consuming. Instead of accurately evaluating the network architecture on the target task, the researchers proposed the proxy task method. The proxy task means training on subsets of dataset or fine-tuning with fewer epochs [23] [27] [28] [29]. Although these methods improve the speed of NAS, a rough evaluation inevitably treats some promising network architectures as poor networks. A post-processing step after NAS is proposed in this paper. As a result, those potential networks with fewer network parameters can be found.
III The Proposed Method
The purpose of this paper is to utilize NAS technology to search for specific network architectures for different individual objects, while completing end-to-end neural reconstruction through binary classification of voxels. Compared with previous works [3] [4] [5], this adaptive reconstruction method achieves higher reconstruction accuracy using fewer network parameters. The following subsections describe the proposed method in detail.
III-A Binary Voxel Classification
Binary voxels can be directly visualized as one of the explicit representations of 3D objects. In general, voxels inside an object are defined as 1, while voxels outside the object are defined as 0. Training a neural network to classify the binary voxels of a given model in 3D space enables end-to-end neural representation and reconstruction. Therefore, the post-processing steps, such as surface reconstruction using marching cubes [4] [5], can be avoided.
Objects are normalized in a 3D space denoted by . The entire normalized space is divided equally into parts, each of which is regarded as a voxel. The set of voxels inside the object is denoted by . For each voxel in , the neural network outputs the probability of , which is between 0 and 1. To optimize the parameters of the neural network, the cross-entropy classification loss is used as the following,
(1) |
where is the -th voxel in sampled voxels, which will be explained in Sec. III-E.
III-B Search Space of Neural Architecture
Most of previous works employ MLP as their global or local SDF fitting network. Therefore, the number of layers, the number of neurons in each layer, and the activation functions in each layer are chosen as our neural architecture search space.
Specifically, in order to reduce the search time and take into account the reasonable search range, the search space for the number of nodes is {8, 12, 16, 20, 24, 28, 32, 40, 48, 56, 64}. Moreover, we let the controller decide the activation function for each layer, making the network architecture more expressive. The search space of activation functions is {} [30] [31] [11], each of which can be represented by
(2) |
where and are usually set to 1.
In the above selection of activation functions, the traditional activation functions and are not included. This is because we experimentally find that adding these activation functions will reduce the classification performance of the network. As stated in the previous NAS works [32] [33] [21], the choice and design of the search space plays a crucial role in the success of NAS. Ablation experiments shown in Sec. IV-D demonstrate the performance of adding activation functions into the search space.

III-C Process of NAS
In order to search a ‘proper’ neural architecture specifically for a given object, we utilize a mature NAS algorithm ENAS [23]. The searching process is drawn in Fig. 2. The controller samples the MLP layers from the search space consisting of different numbers of nodes and activation functions. After an MLP network is trained, its reward can be obtained on the validation set. The reward is then fed back to the controller RNN for policy-gradient descent.
Unlike ENAS, the search strategy in our method focuses on how to choose a better MLP network with different numbers of neurons and activation functions. However, the raw reward in ENAS only considers the classification score, and the internal controller always generates network architectures that make the classification reward higher. Since one of the expectations of the neural reconstruction is to minimize the number of network parameters [4], the impact of network size should be factored into the reward. Thus, the reward is designed as
(3) |
where denotes the classification accuracy of the evaluated network architecture for all voxels. is an expected accuracy, which is set to 0.98. is equal to the parameter amount 7553 of the default network of NI. denotes the size of the evaluated network architecture. is also a fixed value 21121, which is the number of the largest network architecture parameter in our search space.
It is worth noting that if the controller generates an output layer in the process of generating MLP layers, the number of MLP layers will no longer increase. This explains that the number of MLP layers can also be searched during the NAS process.
III-D Post-processing Step after NAS
After completing the NAS process, we introduce a post-processing step to select smaller neural architectures. There are two reasons for this step. First, to speed up the NAS process, we use a proxy task [21] that is similar to NAS-FPN [34] to shorten the training time of the target task. We train the proxy task epochs instead of epochs used to train the target network. This early termination method speeds up the convergence time of rewards by a factor of . Since the proxy task does not train the network to converge, the ranking of the network accuracy cannot represent the final ranking of the network. Therefore, the network with accuracy slightly lower than the highest accuracy can be considered as candidates.
Second, although we have used the network size reward term in Eq. 3 to influence the network chosen by NAS, a fixed weight is not applicable to different objects. A network with a little lower network size and classification accuracy may be excluded, compared to the network with the least reward in the NAS process. Post-processing selection of multiple candidates would greatly alleviate this problem.
Specifically, we filter out all candidate networks whose accuracy is lower than the highest accuracy in the NAS process up to a threshold ( in experiments). Then the network with the smallest size will be selected. The ablation experiments shown in Sec. IV-C validate the effectiveness of the proposed post-processing step.
III-E Other Details of Data Processing, Sampling, and NAS Configuration
In data processing, 3D models are firstly normalized in a sphere with the radius . Then we utilize PyMesh library to voxelize a model to resolution (N=128 by default).
In voxel sampling, we first sample all surface voxels and their outer layer voxels as the support samples of classification boundary. Then, we down-sample other non-support voxels and copy the support voxels to the same number. Thus, the total sample number is the number of other voxels, which is roughly equivalent to 1M.
In the NAS configuration, the maximum number of total network layers and hidden layers is 8 and 6, respectively. The controller samples 6 MLP layers of different architectures at a time. Due to the weight sharing strategy proposed in ENAS, we terminate the search by only sampling 5 times.
IV Experiments
Various experiments are conducted to verify reconstruction quality of the proposed end-to-end method. Sec. IV-A describes the used datasets and metrics. Sec. IV-B shows the experimental results, compared with other methods. Sec. IV-C gives the first ablation experiment of removing NAS and other improvements. Sec. IV-D gives the second ablation experiment of activation functions. The pre-trained network models for all displayed objects can be reproduced with our source code in https://github.com/cscvlab/VoxelReconstruction-NAS.
IV-A Datasets and Metrics
Datasets we used in this paper include Thingi10K [9], Thingi32 and ShapeNet150. Thingi10K is composed of 3D-printing models, which have been tested in NI [4]. NGLOD [5] mainly verifies two other small datasets: Thingi32 and ShapeNet150. Thingi32 contains 32 simple shapes in Thingi10K. ShapeNet150 contains 150 shapes in the ShapeNet dataset [35], including 50 cars, 50 airplanes, and 50 chairs.
The metrics for evaluation are common 3D intersection over union (IoU) and Chamfer distance (CD). The former metric 3D IoU is defined as the ratio of the intersection and the union voxels of a reconstructed model and its ground truth model. The latter metric CD is defined as the bi-directional minimum distance [36] from the surface voxels of a reconstructed model to the surface voxels of its ground truth model, which is expressed by
(4) |
where and denote the voxel number of and , respectively. and denote each voxel in and , respectively. In all the following results, the value of CD is magnified by a factor of 1000 for convenience of display.
IV-B Comparison with Previous Methods
We compare our approach with one voxel reconstruction method ONet [3], and two SDF prediction methods NI [4] (with the default configuration) and NGLOD [5] (with LOD level 1). First, the comparison is conducted on Thingi32 and ShapeNet150 datasets. The experimental results of the previous three methods are shown in the upper part of Table I, and our results are shown in the last row. Note that the results of NI-Thingi32 and ONet-ShapeNet150 are obtained through the official trained networks. The results of NI-ShapeNet150 and NGLOD on two datasets are obtained by our training using the official codes. Since ONet is trained on one category of objects, it cannot be used for objects in Thingi32.
Method | Thingi32 | ShapeNet150 | ||||||
---|---|---|---|---|---|---|---|---|
Size | IoU | CD | Size | IoU | CD | |||
ONet [3] | – | – | – | 6M | 60.7 | 4.613 | ||
NI [4] | 7553 | 96.8 | 0.117 | 7553 | 82.7 | 4.326 | ||
NGLOD [5] | 8737 | 97.4 | 0.088 | 8737 | 82.5 | 1.163 | ||
Ours w.o. NAS | 7553 | 97.8 | 0.089 | 7553 | 95.5 | 0.084 | ||
|
8837 | 98.0 | 0.076 | 8551 | 96.0 | 0.071 | ||
|
7626 | 97.2 | 0.110 | 7461 | 95.0 | 0.103 | ||
Ours w. NAS | 5452 | 97.4 | 0.100 | 5860 | 95.7 | 0.082 |
Since Thingi32 only contains models of simple shapes, NI, NGLOD and our method perform similarly, and they can all reconstruct these models without noticeable errors. For ShapeNet150, the performance differences of the four methods are easily distinguishable. Since there are some complicated objects in ShapeNet150, ONet and NI cannot reconstruct them correctly. This results in a large increase in the metric CD. NGLOD can handle more objects correctly than ONet and NI, and gain better CD and IoU. The proposed method shows significant improvement in CD and IoU, which indicates that our method does not suffer as much performance degradation as other methods when dealing with complex objects. Fig. 3 depicts three reconstructed voxels (one for each category in ShapeNet150) for all four methods as a visual comparison.



As ShapeNet150 contains only 150 objects in three categories, we further conduct another experiment on Thingi10K dataset. We train the NAS networks for models randomly selected from Thingi10K, and compare the reconstruction results with NI. Fig. 4 depicts the histograms of CD and IoU for the two methods, respectively. The proposed method is obviously superior to NI.
IV-C Ablation Experiments of Removing NAS, Size Reward and Post-processing Step
This ablation experiment aims to observe the influence of removing NAS, the proposed size reward, and the post-processing step. The experimental results are shown in the lower part of Table I (rows to ). The -th row shows our results without NAS. The network here is much the same as NI. The main difference is that instead of predicting SDF values, we directly classify voxels, and over-sample the support voxels. As NI successfully reconstructs the 3D models in Thingi32, the method in the -th row achieves only a slight improvement. However, for the more complicated ShapeNet150 dataset, the improvements of IoU and CD are significant. This means that binary classification is much easier to be fitted by a same network than SDF prediction, which experimentally validates the first theoretical problem described in Sec. I.
The -th row shows our results using NAS, but without adding network size to reward and the post-processing step. Since there is no scheme to control the network scale, the NAS obtains networks with more parameters at average. Owing to larger network parameters, the method in the -th row gains the best IoU and CD performance on two datasets.
The -th row shows our results using NAS and size reward, but without the post-processing step. Since the reward is improved to control network size, the mean number of network parameters is roughly the same to it in the -th row, and is reduced by compared to the -th row. As a result, the performance of the method in the -th row degrades slightly.
The advantage of the post-processing step is reflected in the last row. Compared to the -th row, although the performance is promoted slightly, the mean size of the networks is significantly reduced. This verifies the effectiveness of the proposed method in selecting a suitable network architecture.
IV-D Ablation Experiment of Activation Functions
In order to choose suitable activation functions as the candidates in our search space, we test five activation functions, which are Sigmoid, Tanh, ReLU, ELU, and Swish, on Thingi32 and ShapeNet150 separately. This is also an ablation experiment, as NAS only searches different numbers of layers and nodes at the moment. There is no change to the setting except that only one activation function is used during NAS. The experimental results are shown in Table II.
The activation function ReLU, ELU or Swish, can achieve good accuracy, while the results of Sigmoid are dramatically worse, especially on ShapeNet150. Although the mean network size of Sigmoid appears to be minimal, we experimentally find that the reconstruction for Sigmoid is likely to fail. In order to reduce invalid searches, we remove Sigmoid and Tanh from the final search space of activation functions. Compared to using one fixed activation function, using three activation functions {ReLU, ELU, Swish} as the search space reduces CD and IoU very slightly, but searches obviously smaller networks.
Method | Thingi32 | ShapeNet150 | |||||||
---|---|---|---|---|---|---|---|---|---|
Size | IoU | CD | Size | IoU | CD | ||||
Sigmoid | 4369 | 80.8 | 2.565 | 3217 | 58.3 | 18.420 | |||
Tanh | 5953 | 96.5 | 0.155 | 7098 | 94.9 | 0.125 | |||
ReLU | 6455 | 98.2 | 0.071 | 6040 | 95.4 | 0.132 | |||
ELU | 6149 | 96.7 | 0.147 | 6697 | 94.6 | 0.110 | |||
Swish | 5886 | 97.4 | 0.106 | 6236 | 95.6 | 0.098 | |||
|
5452 | 97.4 | 0.100 | 5860 | 95.7 | 0.082 |
V Conclusion
This paper proposes a neural implicit reconstruction method of 3D objects based on network architecture search (NAS). Without any surface reconstruction algorithm (e.g., marching cube [37]), we employ an end-to-end network by directly classifying binary voxels. Although the basic idea is straightforward to some extent, the proposed approach outperforms the state-of-the-art methods [4] [5] using SDF prediction network and the marching cube algorithm. From various conducted experiments, we can conclude that the combination of different layers, node numbers, activation functions (searched by NAS), and using binary classification together lead to the improvement of reconstruction quality, especially at classification boundaries. Furthermore, the number of network parameters is added to the reward during NAS, which reduces the storage of the neural implicit representation. In other words, the further improvement of the compression ratio enhances the storage advantage of neural implicit representation over traditional explicit representations.
One disadvantage of the proposed method is that its flexibility is temporarily limited, since the learned discrete voxels have a fixed resolution. For continuous SDF prediction networks, voxel models at any resolution can be generated by the marching cube algorithm without re-training the networks. This problem may be addressed by future work incorporating the octree representation of voxel, which can progressively classify more subdivided voxels.
Acknowledgment
This work is supported by Natural Science Foundation of Shanghai (Grant No. 21ZR1401200), Shanghai Sailing Program (21YF1401300), and the Foundation of Key Laboratory of Artificial Intelligence, Ministry of Education, P.R. China (AI2020003).
References
- [1] R. Q. Charles, H. Su, M. Kaichun, and L. J. Guibas, “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 77–85.
- [2] D. Maturana and S. Scherer, “Voxnet: A 3d convolutional neural network for real-time object recognition,” in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2015, pp. 922–928.
- [3] L. Mescheder, M. Oechsle, M. Niemeyer, S. Nowozin, and A. Geiger, “Occupancy networks: Learning 3d reconstruction in function space,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 4455–4465.
- [4] T. Davies, D. Nowrouzezahrai, and A. Jacobson, “On the effectiveness of weight-encoded neural implicit 3d shapes,” arXiv preprint arXiv:2009.09808, 2020.
- [5] T. Takikawa, J. Litalien, K. Yin, K. Kreis, C. Loop, D. Nowrouzezahrai, A. Jacobson, M. McGuire, and S. Fidler, “Neural geometric level of detail: Real-time rendering with implicit 3d shapes,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 11 353–11 362.
- [6] B. Mildenhall, P. P. Srinivasan, M. Tancik, J. T. Barron, R. Ramamoorthi, and R. Ng, “Nerf: Representing scenes as neural radiance fields for view synthesis,” in European conference on computer vision (ECCV), 2020, pp. 405–421.
- [7] M. Niemeyer and A. Geiger, “Giraffe: Representing scenes as compositional generative neural feature fields,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 11 448–11 459.
- [8] S. Fridovich-Keil, A. Yu, M. Tancik, Q. Chen, B. Recht, and A. Kanazawa, “Plenoxels: Radiance fields without neural networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 5501–5510.
- [9] Q. Zhou and A. Jacobson, “Thingi10k: A dataset of 10,000 3d-printing models,” arXiv preprint arXiv:1605.04797, 2016.
- [10] M. Nimier-David, D. Vicini, T. Zeltner, and W. Jakob, “Mitsuba 2: a retargetable forward and inverse renderer,” ACM Transactions on Graphics, vol. 38, pp. 1–17, 2019.
- [11] P. Ramachandran, B. Zoph, and Q. V. Le, “Searching for activation functions,” arXiv preprint arXiv:1710.05941, 2017.
- [12] B. Zoph and Q. V. Le, “Neural architecture search with reinforcement learning,” arXiv preprint arXiv:1611.01578, 2016.
- [13] B. Baker, O. Gupta, N. Naik, and R. Raskar, “Designing neural network architectures using reinforcement learning,” arXiv preprint arXiv:1611.02167, 2016.
- [14] J. J. Park, P. Florence, J. Straub, R. Newcombe, and S. Lovegrove, “Deepsdf: Learning continuous signed distance functions for shape representation,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 165–174.
- [15] M. Tancik, P. Srinivasan, B. Mildenhall, S. Fridovich-Keil, N. Raghavan, U. Singhal, R. Ramamoorthi, J. Barron, and R. Ng, “Fourier features let networks learn high frequency functions in low dimensional domains,” in Proceedings of the International Conference on Neural Information Processing Systems (NeurIPS), 2020, pp. 7537–7547.
- [16] V. Sitzmann, J. N. Martel, A. W. Bergman, D. B. Lindell, and G. Wetzstein, “Implicit neural representations with periodic activation functions,” arXiv preprint arXiv:2006.09661, 2020.
- [17] R. Chabra, J. E. Lenssen, E. Ilg, T. Schmidt, J. Straub, S. Lovegrove, and R. Newcombe, “Deep local shapes: Learning local sdf priors for detailed 3d reconstruction,” in European Conference on Computer Vision (ECCV), 2020, pp. 608–625.
- [18] S. Zhang, H. Cao, Y. Liu, S. Cai, Y. Zhang, Y. Li, and X. Chi, “Sn-graph: A minimalist 3d object representation for classification,” in IEEE International Conference on Multimedia and Expo (ICME), 2021, pp. 1–6.
- [19] Z. Chen and H. Zhang, “Learning implicit fields for generative shape modeling,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 5932–5941.
- [20] S. Peng, M. Niemeyer, L. Mescheder, M. Pollefeys, and A. Geiger, “Convolutional occupancy networks,” in European Conference on Computer Vision (ECCV), 2020, pp. 523–540.
- [21] B. Zoph, V. Vasudevan, J. Shlens, and Q. V. Le, “Learning transferable architectures for scalable image recognition,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 8697–8710.
- [22] E. Real, A. Aggarwal, Y. Huang, and Q. V. Le, “Regularized evolution for image classifier architecture search,” in Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), 2019, pp. 4780–4789.
- [23] H. Pham, M. Y. Guan, B. Zoph, Q. V. Le, and J. Dean, “Efficient neural architecture search via parameter sharing,” arXiv preprint arXiv: Learning, 2018.
- [24] H. Liu, K. Simonyan, and Y. Yang, “Darts: Differentiable architecture search,” arXiv preprint arXiv:1806.09055, 2018.
- [25] H. Liang, S. Zhang, J. Sun, X. He, W. Huang, K. Zhuang, and Z. Li, “Darts+: Improved differentiable architecture search with early stopping,” arXiv preprint arXiv:1909.06035, 2019.
- [26] H. Wang, R. Yang, D. Huang, and Y. Wang, “idarts: Improving darts by node normalization and decorrelation discretization,” IEEE Transactions on Neural Networks and Learning Systems, (online first), 2021.
- [27] T. Elsken, J. H. Metzen, and F. Hutter, “Efficient multi-objective neural architecture search via lamarckian evolution,” arXiv preprint arXiv:1804.09081, 2018.
- [28] A. Klein, S. Falkner, S. Bartels, P. Hennig, and F. Hutter, “Fast bayesian hyperparameter optimization on large datasets,” Electronic Journal of Statistics, vol. 11, no. 2, pp. 4945–4968, 2017.
- [29] B. Wu, X. Dai, P. Zhang, Y. Wang, F. Sun, Y. Wu, Y. Tian, P. Vajda, Y. Jia, and K. Keutzer, “Fbnet: Hardware-aware efficient convnet design via differentiable neural architecture search,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 10 726–10 734.
- [30] R. H. R. Hahnloser, R. Sarpeshkar, M. Mahowald, R. J. Douglas, and H. S. Seung, “Digital selection and analogue amplification coexist in a cortex-inspired silicon circuit,” Nature, vol. 405, pp. 947–951, 2000.
- [31] D.-A. Clevert, T. Unterthiner, and S. Hochreiter, “Fast and accurate deep network learning by exponential linear units (elus),” arXiv preprint arXiv:1511.07289, 2015.
- [32] M. Tan, B. Chen, R. Pang, V. Vasudevan, M. Sandler, A. Howard, and M. Le QV, “Platformaware neural architecture search for mobile. 2019 ieee,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 2815–2823.
- [33] B. Zoph and Q. V. Le, “Neural architecture search with reinforcement learning,” arXiv preprint arXiv:1611.01578, 2016.
- [34] G. Ghiasi, T.-Y. Lin, and Q. V. Le, “Nas-fpn: Learning scalable feature pyramid architecture for object detection,” in IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 7029–7038.
- [35] A. X. Chang, T. Funkhouser, L. Guibas, P. Hanrahan, Q. Huang, Z. Li, S. Savarese, M. Savva, S. Song, H. Su et al., “Shapenet: An information-rich 3d model repository,” arXiv preprint arXiv:1512.03012, 2015.
- [36] N. Ravi, J. Reizenstein, D. Novotny, T. Gordon, W.-Y. Lo, J. Johnson, and G. Gkioxari, “Accelerating 3d deep learning with pytorch3d,” arXiv preprint arXiv:2007.08501, 2020.
- [37] W. E. Lorensen and H. E. Cline, “Marching cubes: A high resolution 3d surface construction algorithm,” in Proceedings of the 14th annual conference on Computer graphics and interactive techniques (SIGGRAPH), 1987, pp. 163–169.