This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\SpecialIssuePaper\BibtexOrBiblatex\electronicVersion\PrintedOrElectronic
\teaser[Uncaptioned image]

Our approach represents a scalar field as a neural network that conditions on a point in the field’s domain and produces a scalar value. We obtain highly compressive volume representations in this manner, here showing two levels of compression (middle, right) for the jet volume (left), where even for extreme compression ratios (right, 1307:11307:1), predominant features of the volume are preserved.

Compressive Neural Representations of Volumetric Scalar Fields

Y. Lu1, K. Jiang2, J. A. Levine2\orcid0000-0002-4302-1704, and M. Berger1
1Department of Electrical Engineering and Computer Science, Vanderbilt University, USA 2Department of Computer Science, University of Arizona, USA
Abstract

We present an approach for compressing volumetric scalar fields using implicit neural representations. Our approach represents a scalar field as a learned function, wherein a neural network maps a point in the domain to an output scalar value. By setting the number of weights of the neural network to be smaller than the input size, we achieve compressed representations of scalar fields, thus framing compression as a type of function approximation. Combined with carefully quantizing network weights, we show that this approach yields highly compact representations that outperform state-of-the-art volume compression approaches. The conceptual simplicity of our approach enables a number of benefits, such as support for time-varying scalar fields, optimizing to preserve spatial gradients, and random-access field evaluation. We study the impact of network design choices on compression performance, highlighting how simple network architectures are effective for a broad range of volumes. {CCSXML} <ccs2012> <concept> <concept_id>10003120.10003145</concept_id> <concept_desc>Human-centered computing Visualization</concept_desc> <concept_significance>500</concept_significance> </concept> <concept> <concept_id>10010147.10010257.10010293.10010294</concept_id> <concept_desc>Computing methodologies Neural networks</concept_desc> <concept_significance>300</concept_significance> </concept> <concept> <concept_id>10010147.10010371.10010395</concept_id> <concept_desc>Computing methodologies Image compression</concept_desc> <concept_significance>300</concept_significance> </concept> </ccs2012>

\ccsdesc

[500]Human-centered computing Visualization \ccsdesc[300]Computing methodologies Neural networks \ccsdesc[300]Computing methodologies Image compression

\printccsdesc
volume: 40issue: 3

1 Introduction

The visualization of large-scale field-based data is a fundamental component to many post hoc analyses. Field data, in particular scalar fields, often arise from numerical simulations of scientific phenomena, which require high (3D) spatial and temporal resolution to accurately resolve domain-specific features of interest. The size of such large-scale data presents a number of challenges for visualization, ranging from bandwidth constraints, disk storage, data accessibility, and interactivity. Compression of volumetric scalar fields remains an important tool that can help mitigate these challenges. For visualization purposes, the data is often so large that lossy compression methods are required to enable analysis. The main purpose of lossy methods is to obtain compact volumetric representations in which features of the data remain visually perceptible, at the cost of sacrificing some data precision.

Within the visualization community, the lossy compression of volumetric scalar fields has largely been based on transforms that admit compressive representations, where transform coefficients can be discarded and/or aggressively quantized with minimal loss in data fidelity. A commonality to these methods is the reliance on a rectilinear grid for which predefined bases may be constructed, e.g. Fourier [YL95] or Wavelet bases [WMB11], or where data-dependent bases may be derived [SMP13, BRP16]. However, the success of these methods heavily depends on whether a provided scalar field is a good match for the assumptions made by the transform, e.g. the field is largely composed of low-frequencies for Fourier bases, or admits a low-rank decomposition for factorization-based transforms [BRLP19]. In practice, these assumptions are satisfied to varying degrees, leading to variability in compression ratios and approximation quality amongst existing methods.

In this work we propose a learning-based method for compressing scalar fields. We leverage neural networks that are designed to map a continuously-defined position from the domain to a scalar value in the range [PFS19, MON19]. The use of neural networks places little assumptions on data characteristics; instead, the network learns what is necessary to approximate the given field. In particular, by limiting the capacity of the network, such that the number of weights of the network is less than the volume resolution, we obtain compressive volume representations that are optimized to best approximate the field at its sampled values. Hence, the neural network is the compressed volume representation – the level of compression, in part, follows from the number of network weights, and the original sampled field can be reconstructed by evaluating the neural network at the given positions.

Our neural network design is conceptually straightforward. We build on the approach of SIREN [SMB20], wherein fully-connected layers and periodic activation functions (using sinusoids) comprise the network architecture. Periodic activation functions, coupled with careful initialization for stable optimization, have been recently shown to outperform more advanced architectures, e.g. frequency-based embeddings with ReLU activations [MST20], for a variety of representation tasks. We illustrate how residual connections that preserve SIREN activation distributions lead to networks that are robust across architecture choices. Further, we observe that the learned, per-layer, weights of SIREN are distributed in such a manner that they may be quantized to a small number of bits, with minimal impact on performance.

By limiting network capacity, and performing weight quantization, we obtain highly compact representations of scalar fields, as demonstrated in Fig. Compressive Neural Representations of Volumetric Scalar Fields. Further, the conceptual simplicity of our approach provides a number of benefits. Due to the use of periodic activation functions, we can target the optimization of higher-order derivatives, as studied in Sitzmann et al. [SMB20]. In particular, when spatial gradients are available, we can use this information to regularize the network. This leads to better-behaved scalar fields whose isosurfaces are less noisy under high compression ratios. We also show how it is straightforward to adapt our approach to compressing time-varying scalar fields simply by adjusting the network inputs. Furthermore, the networks enable field access at arbitrary points in the domain, eschewing the reconstruction of the entire volume at once – we show how this can benefit visualization applications such as volume rendering.

Summarizing, the main contributions of our approach are:

  1. 1.

    A compression technique for volumetric scalar fields based on implicit neural representations.

  2. 2.

    An evaluation of this compression approach, directly comparing it to a recent state-of-the-art technique (tthresh [BRLP19]) as well as exploring implementation design choices; and

  3. 3.

    Experimental results on compressing volumes, including studying gradient preservation and preserving time-varying data.

2 Related Work

2.1 Compression of Volumetric Data

Lossy compression of large scale volumes has been an important topic in recent years, particularly as our ability to simulate and acquire data grows. Many of the early works in this area focused on using discrete cosine transforms [YL95] or wavelet-based compression [GWGS02, IP99, Mur93, WMB11], often specifically focusing on interactive rendering applications [RGG13, SW03]. By using such transforms, these techniques enable compression by allowing the end user to separate features within a transformed space (i.e. removing high frequencies while preserving low frequencies) and ultimately sparsify the data representation. ZFP performs transformations customized at the block level to achieve compression with fast I/O access [Lin14]. More recent techniques, such as TAMRESH [SMP13] and TTHRESH [BRLP19] employ tensor decomposition [BRP16]. TTHRESH is particularly notable as a recent state-of-the-art compression technique that we directly compare our results against in this work.

Other techniques focus on data fitting as the workhorse for compression. For example, ISABELA [LSE11] and SZ [DC16, TDCC17] use curve fitting to approximate the input data. Liang et al. outperform SZ and ZFP using an improved Lorenzo prediction [LDT18]. Data fitting is related to our neural network based approach, although the method to achieve the fit is different. Using data models can be a powerful mechanism for providing a compression technique. Peterka et al. [PNG18] study the use of nonuniform rational B-spline functions (NURBS) for compressive volume representations via adaptive refinement of control points. This work shows the benefit of smooth representations for analytically-defined gradients, however an important distinction in our approach is that we can directly optimize for gradients. Other examples focus on models that preserve specific aspects of the data, e.g. topological features [SPCT18], graph-based models like SQ [IKK12], and dictionaries [DMG20]. COVRA uses an octree decomposition to model data blocks which are compressed at multiple resolutions [GIGM12]. Data fitting broadly considers the tradeoff between how many samples are used and how much precision is devoted to each sample. Recently, Hoang et al. provide a precision-resolution tree that is used for efficient selection of either fewer data samples or data samples encoded with fewer bits [HSB20].

Refer to caption
Figure 1: An overview of our compression method, split into the steps of encoding (left) and decoding (right). We encode a provided scalar field with a coordinate-based neural network, where the network ff is tasked with predicting the field’s value 𝐗[i,j,k]\mathbf{X}[i,j,k] given coordinates (x,y,z)(x,y,z). In the network diagram, red arrows represent fully-connected layers whose weights are quantized, blue arrows are unquantized fully-connected layers, and dashed black arrows represent residual connections. Decoding amounts to function evaluation, where we may reconstruct the field at grid vertices, or arbitrary positions, and thus perform volume rendering with the network in lieu of the sampled field.

2.2 Deep Learning for Volume Representation

Recently, numerous researchers have deployed deep learning and used neural networks to enable a variety of visualization tasks. Berger et al. utilize generative adversarial networks (GANs) to analyzing the space of visual parameters in a volume renderer, providing a new interface for transfer function design [BLL18]. He et al. generalize the idea of learning conditioned on parameter spaces to include ensembles of volumes with InSituNet [HWG19]. Han et al. encode multivariate volume data for variable-to-variable translation [HZX20]. Jakob et al. consider the problem of interpolation using neural networks defined on a large ensemble of vector fields [JGG20]. While the applications are different, all of these works share a common thread in that they construct representations that blend visualization tasks (e.g. rendering, interpolation) with data representations.

Closely related to compression is the task of super-resolution, for which many techniques have been developed in the computer vision communities. More recently, researchers have shown that deep learning can be used to achieve super-resolution on volumetric data. Xie et al. developed tempoGAN for super-resolution of volumetric fluids in computer graphics [XFCT18]. Weiss et al. developed new techniques for isosurface rendering that similarly used deep learning for super-resolution [WCTW19]. Han and Wang developed neural networks for super-resolution in temporal [HW19] and spatial [HW20] super-resolution, and Guo et al. achieved spatial super-resolution in vector fields [GYH20]. Notably, many of these techniques are tested against possible compression techniques, but none of the above were specifically designed for compression. Thus, while they are compressive, their main strengths lie elsewhere.

Recently, the machine learning community has explored using deep learning to construct implicit representations, which can be used to model a variety of data. We utilize a similar approach in our network design, and as seen in recent works this can achieve superior performance. Park et al. developed DeepSDF for representing shape using signed distance fields [PFS19]. Recent extensions to this work include overfit neural networks [DNJ20] and DualSDF [HAESB20] that allow for tradeoffs in the representation precision and flexbility. Similarly, Chen and Zhang use implicit fields for generative shape modeling [CZ19]. Alternatively, instead of training to produce a signed distance field, Mescheder et al. propose to construct occupancy fields [MON19]. Besides just representing shape, other recent applications include texture synthesis [OMN19] and view synthesis [MST20]. As this is an emerging area, we are still discovering new ways to improve these networks. Notable recent works that informed our network design include using periodic activations [SMB20] and Fourier feature mappings [TSM20]. Our approach is distinct from prior works in that we investigate how to make implicit neural representations compressive for arbitrary scalar fields [DNJ20, SMB20, TSM20].

3 Compression Method

The basic idea behind our compression approach is to represent a provided scalar field as a learned function, one that is parameterized by a set of weights whose size is smaller than the field’s size. Fig. 1 shows an overview. Specifically, we assume the input to our method is a scalar field sampled on a regular grid, which we treat as a dd-tensor 𝐗s1×s2××sd\mathbf{X}\in\mathbb{R}^{s_{1}\times s_{2}\times\cdots\times s_{d}}. When d=3d=3 this represents a scalar field whose domain is a 3D Cartesian coordinate system, while d=4d=4 may represent a time-varying scalar field in 4 dimensions. We associate each sample of the field with an indexing tuple of integers 𝐢=(i1,i2,,id)\mathbf{i}=(i_{1},i_{2},\cdots,i_{d}), an integer for each dimension, such that 𝐗[i1,i2,,id]\mathbf{X}[i_{1},i_{2},\cdots,i_{d}] returns the value of the field at this sample, and further, assume this index is associated with a real-valued dd-dimensional point from the field’s domain, 𝐩𝐢d\mathbf{p}_{\mathbf{i}}\in\mathbb{R}^{d}.

Our goal is to learn a function fΘ:df_{\Theta}:\mathbb{R}^{d}\rightarrow\mathbb{R}, such that for a given sample 𝐢=(i1,i2,,id)\mathbf{i}=(i_{1},i_{2},\cdots,i_{d}) and corresponding point 𝐩𝐢\mathbf{p}_{\mathbf{i}}, we would like fΘ(𝐩𝐢)f_{\Theta}(\mathbf{p}_{\mathbf{i}}) to be as close to 𝐗[i1,i2,,id]\mathbf{X}[i_{1},i_{2},\cdots,i_{d}] as possible. The vector Θm\Theta\in\mathbb{R}^{m} denotes the function’s set of parameters. Assuming its size mm is less than the size of the field, defined as C=j=1dsjC=\prod_{j=1}^{d}s_{j}, then we can obtain a representation of the field that (a) serves as a good approximation to 𝐗\mathbf{X}, and (b) requires smaller storage.

What specific form should the function take? In order to handle complex volumetric fields typically produced in numerical simulations, we require functions that place as few assumptions as possible on the characteristics of the fields, e.g. smoothness, spectral properties, or sampling rate. To this end, we define the functions as deep neural networks, building off of recent work in implicit neural representations for shape modeling [PFS19, MON19, SMB20]. Notably, these methods can model fine-grained details, e.g. intricate geometric structures of shapes, only limited by the capacity of the network rather than the resolution of some underlying regular grid, as is commonly used with convolutional networks [HW19]. We argue that such coordinate-based neural networks are a good fit for arbitrary volumetric fields, not just shape-based data. By framing compression as optimization, we allow the network to learn what is necessary for approximating the sampled field under a compression budget. In this paper we study how to effectively construct such functions, considering the trade-offs in approximation quality and level of compression. At a high-level, our approach follows a standard compression setup, comprised of 2 steps: an encode step to find the function, and a decode step that allows us to reconstruct the sampled field.

3.1 Neural Encoder

The purpose of the encode step is to train the neural network to accurately model the field at its provided samples. More specifically, we aim to minimize the following mean-squared error loss:

minΘ𝐢(fΘ(𝐩𝐢)𝐗[i1,i2,,id])2.\min_{\Theta}\sum_{\mathbf{i}}\left(f_{\Theta}(\mathbf{p}_{\mathbf{i}})-\mathbf{X}[i_{1},i_{2},\cdots,i_{d}]\right)^{2}. (1)

In practice, the loss function is optimized by performing stochastic gradient descent, assembling minibatches at each iteration of training by randomly selecting samples from the provided field.

Refer to caption
(a) Asteroid
Refer to caption
(b) Ionization
Figure 2: We show the impact of residual connections (green) vs. non-residual connections (red) as a function of network depth, as well as the impact of network weight quantization (dashed lines), fixing the number of network weights mm to be C50\frac{C}{50}. We find that residual connections are more robust to depth and improve performance, while weight quantization leads to a small drop in performance, but is volume-dependent.

What remains is the design of the network. We opt for simplicity in network architectures, to avoid the brittleness of hyperparameter tuning and support a diverse set of volumetric fields. To this end, we build off of the approach of SIREN [SMB20], where a neural network is defined by a sequence of fully-connected layers that use sinusoidal activation functions – no normalization, e.g. batch normalization [IS15], is necessary for successful training. SIREN has several benefits. First, as demonstrated by Sitzmann et al., training is ensured stable through a principled initialization scheme that controls for the distribution of activations at layers, namely that all layers are arcsine-distributed. Secondly, the use of sinusoidal activations admits functions that are CC^{\infty} differentiable. Our approach takes advantage of both of these properties.

Given the network design of SIREN, to setup the network we need only define the fully-connected layers. To this end, our method accepts two inputs: the number of layers ll, and the total number of network weights mm. We derive an integer kk such that the first weight matrix has size k×dk\times d, the last weight matrix has size d×1d\times 1, and all intermediary weight matrices are of size k×kk\times k, such that the total number of weights of the network is approximately mm. Bias vectors are similarly derived from inputs mm and ll. All weights are of 32-bit floating-point precision. The resulting set of weight matrices and bias vectors, collectively, represents our vector of parameters Θ\Theta, e.g. the compressed representation. Note that it is also possible to vary the sizes of the weight matrices, e.g. to gradually increase the number of hidden units as a function of layer depth. Experimentally, we found this made little difference, thus for simplicity we chose a fixed number of hidden units in intermediary layers.

The number of layers, ll, should be large enough to handle the modeling of arbitrarily complex fields. In practice, however, setting ll too large can be detrimental, posing challenges for training due to vanishing gradients. As an example, we show results for two different volumes in Fig. 2 (red solid lines), varying the number of layers, where the quality of the volume is measured through peak signal-to-noise ratio (PSNR), the mean-squared error normalized by the data range. Note that, at a certain point, as the model depth increases performance decreases. Further, the network depth that leads to highest PSNR varies between the two volumes.

Refer to caption
(a) Layer 3
Refer to caption
(b) Layer 9
Refer to caption
(c) Layer 15
Refer to caption
(d) Quantization v. Depth
Figure 3: In (a-c) we show the distribution of weights for our network trained on the Ionization volume for layers 3, 9, and 15, number of weights m=C50m=\frac{C}{50}. We find that the distribution is approximately zero-mean normal, with slight variations, hence amenable to quantization. In (d) we vary the number of bits used for quantization, adjusting the number of weights mm to ensure an approximately equal compression ratio. We find that 99 bits is a good trade-off between number of network weights, and level of quantization.

In order to ensure our models are robust with respect to the number of layers across different volumes, we enrich SIREN with residual connections [HZRS16]. Specifically, we modify the identity mappings of He et al. [HZRS16] to ensure that the distribution of activations in each layer are within the range [1,1][-1,1], in accordance with the assumptions of SIREN [SMB20]. Each residual block in our network takes the following form, omitting bias vectors for clarity (c.f. Fig. 1 dashed black arrows):

𝐚i+1=12(𝐚i+sin(𝐌i+12sin(𝐌i+11𝐚i))),\mathbf{a}_{i+1}=\frac{1}{2}\left(\mathbf{a}_{i}+\sin\left(\mathbf{M}^{2}_{i+1}\sin(\mathbf{M}^{1}_{i+1}\mathbf{a}_{i})\right)\right), (2)

where 𝐚i\mathbf{a}_{i} represents a vector of activations from a previous block ii, 𝐚i+1\mathbf{a}_{i+1} is a vector of activations at the next block i+1i+1, and 𝐌i+11\mathbf{M}^{1}_{i+1} and 𝐌i+12\mathbf{M}^{2}_{i+1} are a pair of learnable matrices associated with block i+1i+1. In Fig. 2 (green solid lines) we show the effect of residual connections – note we treat each block as containing 2 layers. Note that (a) training is more stable as the network depth increases, and (b) we obtain a boost in performance over non-residual connections. Consequently, unless otherwise stated, all of our networks employ 8 residual blocks.

3.1.1 Network Weight Quantization

Thus far, the compression ratio of our method can be expressed as Cm\frac{C}{m}, assuming the network weights and volume are of the same precision. However, further compression can be gained through quantizing the network weights. Weight quantization has been considered for classification-based convolutional networks [HMD15], but to the best of our knowledge, not yet studied for coordinate-based MLPs. For our SIREN-based network, we make several observations:

  1. 1.

    The first and last layers of the network are parameterized by small weight matrices, as discussed above. Experimentally, we found that they require high precision for good performance and thus we do not quantize the matrices (c.f. Fig. 1 blue arrows).

  2. 2.

    Instead, most of the parameters of our network exist in intermediary layers, parameterized by matrices of k×kk\times k. It is these matrices that we quantize (c.f. Fig. 1 red arrows).

  3. 3.

    Since the distribution of activations in SIREN take a specific form, we find that the distribution of weights in each layer take on, approximately, a normal distribution, with some small variations. Fig. 3(a)-(c) demonstrates this for three different layers, where empirically, we find that deeper layers have lower spread, and a higher concentration of values around zero.

Motivated by these observations, we employ the technique of Han et al. [HMD15] and cluster the weights, individually, for each intermediary layer using k-means. We set the number of clusters to be the precision at which we would like to represent the weights, namely if we wish to represent the weights with a precision of bb bits, then the number of clusters will be set to 2b2^{b}. After clustering, we assign each weight its corresponding bb-bit index, and store these, alongside the corresponding floating point-precision cluster centers, as the quantized weight representation.

What should the number of bits, bb, be set to in practice? In Fig. 2 we show the two network options with their corresponding quantized representations (as dashed lines), setting bb to 8. We find that the drop in performance in quantization is dependent on the specific volume, with the ionization volume most impacted. Thus, we cannot be too aggressive in quantization, yet there exists a size trade-off in the precision at which to represent weights, and the number of weights to assign to the network. Based on this, in Fig. 3(d) we plot the PSNR for a sequence of networks that all have approximately the same size, but consider both the number of weights and weight quantization. We increase the number of bits used for quantization along the x-axis, and compensate by decreasing the number of weights of the network. We observe that at 99 bits, e.g. 512512 clusters, the network takes a minimal drop in PSNR – similar results are observed in different volumes. Consequently, all networks in the paper represent weights of intermediary layers with 9 bits.

3.1.2 Gradient Regularization

As previously mentioned, the use of sinusoidal activation functions in our network implies that we can take arbitrary-order derivatives of the function. Thus we can optimize for the derivatives of the function, in addition to just the provided field values. This allows us to obtain compressed fields whose higher-order information is preserved. Specifically, we modify our loss in Eq. 1 to ensure that the spatial gradient of the function is close to provided gradients of the original scalar field:

minΘ\displaystyle\min_{\Theta} 𝐢(fΘ(𝐩𝐢)𝐗[i1,i2,,id])2+\displaystyle\sum_{\mathbf{i}}\left(f_{\Theta}(\mathbf{p}_{\mathbf{i}})-\mathbf{X}[i_{1},i_{2},\cdots,i_{d}]\right)^{2}+ (3)
λfΘ(𝐩𝐢)𝐗[i1,i2,,id]22,\displaystyle\lambda\lVert\nabla f_{\Theta}(\mathbf{p}_{\mathbf{i}})-\mathbf{X}^{{}^{\prime}}[i_{1},i_{2},\cdots,i_{d}]\rVert^{2}_{2},

where 𝐗\mathbf{X}^{{}^{\prime}} denotes the (provided) scalar field gradient, and we set λ=0.05\lambda=0.05, which we found to give a good balance between the gradient and scalar value fit. In some cases, 𝐗\mathbf{X}^{{}^{\prime}} may be accompanied by the field, e.g. within FEM numerical simulations gradients are often computed from the finite element basis. In other cases, we may numerically approximate gradients using finite differencing schemes, however for certain volumes, e.g. acquired medical images, the estimated gradients may not be reliable for regularization.

Table 1: We compare the impact of gradient regularization (“grad”) without using such regularization (“no grad”) for the jet volume, for a compression ratio of 677:1677:1. We find this regularization yields gradients closer to the target (Grad PSNR) with minimal drop in field approximation (PSNR).
Loss PSNR FD-Grad PSNR Net-Grad PSNR
no grad 51.8 52.0 50.3
grad 51.6 54.7 55.5
Refer to caption
(a) ground truth
Refer to caption
(b) no gradient reg.
Refer to caption
(c) gradient reg.
Figure 4: We show isosurfaces for the reconstructed jet volume, comparing with, and without, gradient regularization (c.f. Table 1) in the loss. Note that gradient regularization helps to suppress erroneous high-frequency details.

To show the benefit of gradient regularization we train 2 networks – one with gradient regularization and one without – to compress the jet volume, under a compression ratio of 677:1677:1 for each. Namely, each network is trained with number of weights m=C200m=\frac{C}{200}, along with weight quantization. We use central differencing to numerically estimate gradients from the provided field. We compare the compression results in terms of the PSNR of the original field, as well as PSNR of the gradient field, following Han et al. [HTZ19]. The numerically estimated gradient from the provided scalar field is taken as the target gradient field, while we compare to 2 forms of gradients from the decompressed volumes: central differencing of the reconstructed volume (denoted FD-Grad), as well as the analytical gradients obtained by differentiating the network (denoted Net-Grad).

Table 1 summarizes the results. We find that incorporating gradient regularization leads to a small decrease in PSNR of the scalar field, but leads to a boost in performance in capturing the target gradient. Interestingly, without gradient regularization, analytical gradients from the network (Net-Grad PSNR) suffer quite a bit relative to numerically-estimated gradients (FD-Grad PSNR). From a visualization perspective, inaccurate gradients can have a large impact on isosurfaces of the scalar field, whose normals point in the same direction as the gradients. We highlight this phenomenon in Fig. 4. Note that without gradient regularization, we obtain surfaces with erroneous high-frequency details, whereas gradient regularization leads to smoother surfaces, reflective of ground truth.

Refer to caption
(a) Direct Neural Rendering
Refer to caption
(b) Gradient-based Shading
Figure 5: Our neural network can be directly used for volume rendering, rather than reconstructing the volume, shown here for asteroid for (a) a basic volume renderer where our network is evaluated on-demand during ray marching, and (b) the computation of network gradients to enable direct illumination.

3.2 Neural Decoder

Once the network has been trained, and the volume encoded, decoding is rather straightforward, as it amounts to evaluating the neural network. To reconstruct the scalar field we simply evaluate our network at all grid vertices within the volume. This evaluation is relatively efficient given the data parallelism of the network, e.g. a sequence of matrix multiplications and sin\sin function evaluations.

Beyond reconstruction, a key advantage of neural scalar field representations is the support for random access function evaluation. We may evaluate our function at any position, not just grid vertices, hence the network serves as an interpolant. Furthermore, we can compute higher-order information from the network directly, eliminating the need to numerically estimate derivatives. These features allow us to use the network directly for visualization purposes, rather than reconstructing the full volume a priori. As an illustration, we have developed a neural volume renderer with our network, where at each step of ray marching, we evaluate our network at all current ray positions to obtain function values, see Fig. 5a for an illustration. Moreover, it is trivial to support direct illumination via the computation of function gradients as we ray march, subsequently normalized for shading purposes. We highlight this feature in Fig. 5b. The quality of the volume-rendered images suggests that our network serves as a good interpolant, not merely overfitting to the sampled field’s values.

4 Experiments

Refer to caption
Figure 6: Compression results for our approach, neurcomp. Left-to-right: asteroid (timestep 0), ionization, isotropic_p, magnetic, and rt. Top row: original volumes. Middle row: medium compression. Bottom row: high levels of compression.
Refer to caption
Figure 7: A comparison of our method with tthresh [BRLP19] for 3D scalar fields, plotting compression ratio against PSNR.

To evaluate our approach we have run our method on a variety of datasets under different levels of compression; Table 2 summarizes them. Specifically, from the Johns Hopkins Turbulence Database (JHTD) [LPW08], mhd_p and mhd_bx are respectively pressure and x-coordinate magnetic fields from a magneto-hydrodynamic isotropic turbulence simulation, and isotropic_p is a pressure field from a forced isotropic turbulence simulation. The ionization volume is a temperature field from an ionization front instability simulation [WN], the jet volume is mixture fraction from a simulation of jet flames [GBG14], the magnetic volume is from a magnetic reconnection simulation [GLDL14], and rt is from a Rayleigh-Taylor instability simulation [CCM04]. Volumes for JHTD and rt are formed by taking centered spatial crops to access the original field values where the central portion of the simulations take place.

Furthermore, we have evaluated our method on two time-varying volumes isabel and asteroid (Section 4.2). The dataset isabel corresponds to hurricane Isabel’s “QVapor” field [WBK04] of its first twelve time steps. Note that we cropped out the ten lowest planes in the zz-coordinate to remove the NaNs in the data that correspond to land. The dataset asteroid uses the first ten timesteps from the field “v02” from a deep water asteroid impact simulation [PG18].

Fig. 6 shows a gallery of compression results on our approach, which we refer to as neurcomp. We also compare our method to the state-of-the-art compression technique tthresh [BRLP19], a method for compressing arbitrary tensor data. We limit our comparisons to only tthresh as Ballester et al. have already demonstrated superior results to a wide variety of existing compression methods. Unless otherwise stated, we use OSPRay [WJA16] for rendering and isosurfacing for visual comparisons.

Implementation Details. We use Adam [KB14] to optimize our networks, where we found the best results by setting the learning rate to be inversely proportional to the number of weights of the network. Specifically, the learning rate is set as a linear function, varying from 21052\cdot 10^{-5} for 5M parameters, to 10410^{-4} for 800K parameters. We ensure that our network makes a prescribed number of passes over the entire volume, where we found 7575 iterations to be more than enough for convergence, decaying the learning rate by a factor of 55 every 2020 iterations. At each iteration we set the batch size to fill up the GPU memory, in practice ranging from 16K to 64K grid points from the volume sampled uniformly at random.

Table 2: Name, resolution, precision in bits, and file size in MB for the datasets in our experiments. For time-varying datasets asteroid and isabel, the 4-th dimension of the resolution is the number of timesteps.
Dataset Resolution Precision Size
mhd_p 255×255×255255\times 255\times 255 float32 63
ionization 248×248×600248\times 248\times 600 float32 141
jet 512×336×768512\times 336\times 768 float32 504
mhd_bx 512×512×512512\times 512\times 512 float32 512
isotropic_p 512×512×512512\times 512\times 512 float32 512
magnetic 512×512×512512\times 512\times 512 float32 512
rt 512×512×512512\times 512\times 512 float32 512
asteroid 300×300×300×10300\times 300\times 300\times 10 float32 1030
isabel 500×500×90×12500\times 500\times 90\times 12 float32 1030

4.1 3D Scalar Fields

We first compare our method with tthresh for 3D scalar fields. We run our method over a sequence of compression levels, and plot the resulting compression ratio against PSNR. For these experiments we do not use gradient regularization, but rather, delegate this to Section 4.4. As tthresh accepts an error/accuracy tolerance (e.g. PSNR) as input parameter, rather than compression ratio, we run tthresh for a sequence of PSNR values such that they give compression ratios that are roughly in the range of what we consider.

Fig. 7 shows the quantitative results. We find that our method is, largely, an improvement over tthresh, with the only exception being the jet volume for low compression ratios. In particular, we find that our method obtains larger gains in performance the higher the compression ratio. This suggests that when our networks are underparameterized (e.g. much fewer parameters than the resolution of the data), they remain robust, and can still obtain good approximations to the given scalar field.

Refer to caption
(a) Magnetic field (xx)
Refer to caption
(b) Pressure field
Figure 8: We show visual comparisons between our method and tthresh for the magneto-hydrodynamic simulation. For high compression ratios, we find that our method produces smoother results (a) and can better retain features in the volume (b).

To further qualitatively assess this robustness in high compression regimes, we visually compare volume renderings of our method with tthresh for the magneto-hydrodynamic simulation, namely its magnetic field (mhd_bx) and pressure field (mhd_p) in Fig. 8. We find that our method is able to reproduce smoother surfaces in the rendering, as shown in the magnetic field plot (a), while in the pressure field (b), we find that components of the volume, corresponding to the chosen transfer function, are better retained through our method. Indeed, we find that our method gracefully degrades for extreme compression ratios, unlike block-based methods [Lin14, DC16] that produce visible artifacts at block boundaries. Global decomposition-based methods [BRP16, BRLP19] also do not suffer from locality issues, but we find that they can produce high-frequency noise in high compression regimes. We demonstrate this in Fig. 9 for the rt dataset, comparing our method with tthresh under approximately the same PSNR values. We first note that our method produces compressed representations approximately half the size of tthresh. Further, we find that our method produces smoother scalar fields, more reflective of the ground truth as shown on top, unlike the high frequencies reproduced by tthresh.

Refer to caption
Figure 9: An evaluation of our method under extreme compression ratios for the rt dataset. We find our method gracefully degrades in performance, whereas tthresh tends to reproduce high frequencies.
Refer to caption
Refer to caption
Figure 10: Comparison against tthresh of our approach’s performance on the time-varying datasets. Top row: asteroid, bottom row: isabel. From left-to-right we vary compression ratios, denoting them as X/YX/Y where XX is the compression ratio of neurcomp and YY is the compression ratio of tthresh.

4.2 Time-varying Scalar Fields

We evaluate our approach’s capability to compress time-varying data by comparing our method with tthresh on isabel and asteroid, assembling the first twelve time steps from isabel and ten consecutive time steps towards the beginning of the asteroid simulation. While we show results per time-step, we note that both methods treat the data as 4D volumes.

Refer to caption
Figure 11: We assess the effect of spatial resolution on performance, evaluating our method on spatial crops of different size for the 5123512^{3} isotropic_p volume.

Fig. 10 show comparisons between our method and tthresh over multiple timesteps with regard to 3 different compression ratios. Note that the compression ratios for our method and tthresh are slightly different, since it is difficult to precisely set the compression ratio for tthresh. As the top of Fig. 10 shows, our approach achieves superior results on asteroid dataset when compared with tthresh, with more stability on both ends. In part, this is due to the fine resolution of sampling in time, where the simulation evolves slowly. For the isabel dataset, our method performs comparable to tthresh at low compression ratio, except for the start and end timesteps where we outperform, as shown in the bottom of Fig.  10. For higher compression levels, however, our method gains larger boosts in performance over tthresh at all timesteps, showing the preference of our method for high-compression ratio scenarios.

4.3 Spatial Resolution

Here we examine the impact of spatial resolution on the quality of our compression method. In particular, we wish to gain insight on the following question: do coordinate-based MLPs benefit from high spatial resolution? To test this, we took the isotropic_p 5123512^{3} volume, and grabbed centered crops of size 2563256^{3} and 1283128^{3}. We expect this volume to contain similar statistics for sufficiently large spatial crops, since it corresponds to isotropic turbulence at relatively small scales. We then ran our method on the volumes, setting the network size-to-volume ratio (Cm)\frac{C}{m}) constant, namely 50, 100, 150, 200, and quantize the network weights to 9 bits. Note that for larger volumes, the storage of the cluster centers leads to overall higher compression ratios, thus for equivalent Cm\frac{C}{m} values lower resolution volumes are given the benefit of the doubt.

Fig. 11 (top) shows the results, plotting performance as PSNR with respect to ground truth. We can see that the higher the spatial resolution, the better the performance, despite (a) the network sizes setup in the same manner, and (b) the higher compression ratios for the larger volumes. We view this as an encouraging property, namely that network capacity is not strictly tied with the spatial resolution. On the bottom of the figure we show volume renderings from low and high-end compression ratios, illustrating that our approach captures the predominant features of the volume even for high levels of compression.

Table 3: A comparison of tthresh to our gradient-regularized network.
Method Compression PSNR Gradient PSNR
neurcomp 545:1 47.9 50.7
tthresh 258:1 47.6 46.6
Refer to caption
(a) ground truth
Refer to caption
(b) neurcomp
Refer to caption
(c) tthresh
Figure 12: We show isosurfaces for the rt dataset with our gradient-regularized network and tthresh. Though both have roughly the same PSNR, our method suppresses high-frequency details.

4.4 Gradient Regularization

Last, we show how gradient regularization can be used to produce better-behaved scalar fields in comparison to prior works. Specifically, for the rt dataset we obtain target gradients through central finite differencing, and train our method using the loss of Eq. 3. We provide our resulting PSNR as a target accuracy for tthresh, yielding comparable PSNR values as shown in Table 3. However, a gap exists between the methods in terms of the Gradient-based PSNR. As discussed in Sec. 3.1.2, the effect of gradient regularization will be prominent when isosurfacing the volume, and we can observe in Fig. 12 that tthresh indeed produces isosurfaces that contain considerable noise – this corroborates findings in Fig. 8(a) and Fig. 9. Indeed, the preservation of higher-order information is nontrivial through existing compression methods, whereas our method can incorporate gradient preservation as part of optimization.

5 Discussion

Although our method can achieve state-of-the-art results for compressing scalar fields, we acknowledge several limitations. The primary drawback is the high computation time for training. In Fig. 13 we plot training times, as a function of compression ratio, for three scalar fields of different spatial resolution – we observe similar times for different volumes with equivalent resolution. These times were computed on a single node using an NVIDIA V100 GPU with 16GB of memory. We foresee our method as being suitable for large simulations running on supercomputers, where as the simulation completes one could run a deep learning job for compression.

We also note that the training times are rather pessimistic – we find that our network converges to comparable PSNR values in, roughly, 23\frac{2}{3} the reported times, and so there are opportunities to speed up training. However, as shown, for large volumes training can be quite slow (requiring a few hours per volume). Interestingly, and perhaps counter-intuitively, as the compression ratio increases it takes less time to compress the data, due to the fact that higher compression ratios utilize smaller network architectures. So, unlike many standard techniques that start from a lossless perspective and work harder to compress, we control this at the architecture level.

Refer to caption
Figure 13: We plot training times for three scalar fields, as a function of compression ratio.

We will address this limitation in future work by investigating multiresolution methods for training [JJHZ20], adapting our network to take spatial locality into account. A major benefit of our approach to compression is that we do not need to store the entire volume in memory at once, since during training we need only access random samples of the field. Addressing limitations in the time required to train will enable us to realize this benefit.

Another limitation related to high computation time is the lack of generalization: the function that is learned is specific to a single volume. We argue that tailoring a neural network to a single volume is precisely what gives us such good performance, as similarly demonstrated in recent work within shape representations [DNJ20]. Nevertheless, we have also shown the efficacy of our method for time-varying scalar fields, and so for future work, we plan to incorporate other factors common to simulation-based data, e.g. data composed of multiple fields and associated with simulation parameters, within our network design. Indeed the notion of generalization is limited for volumetric data within the visualization community, often restricted to fields/time steps produced from a single simulation [HW19, HZX20] or fixing these parameters and instead considering multiple simulation parameters [HWG19]. Our approach is general enough to accommodate these factors.

The use of neural networks for representing volumetric data places some limitations on the types of volumes that we can support. In particular, our method faces limitations in handling noisy volumes, e.g. those produced from medical imaging. The main issue is not in capturing the predominant features of the volume – our method performs well on this matter, please see Fig. 14 for an example on the foot volume. Rather, the use of neural networks makes it challenging to preserve the noise itself. This is also problematic for gradient regularization, where it is challenging to extract useful gradient signal for our network if noise is present. On the other hand, we demonstrated that simple finite difference schemes for simulation data proves helpful for regularization, and we expect higher-order numerical schemes for derivative estimation should also prove useful. Beyond gradients, we can also regularize the learned function with any type of differential expression, and this can lead the way to preserving the physics associated with a given simulation as part of our compression approach.

Refer to caption
(a) ground truth
Refer to caption
(b) 25:1 compression; PSNR: 35
Figure 14: We show our method’s performance for a medical image, here in the case of foot (2563256^{3}, 8-bit precision).

Despite the limitations, we are optimistic about the use of neural networks for scientific data compression. Besides the clear advantages shown over other architectures [SMB20], coordinate-based MLPs offer significant flexibility in designing constraints for data preservation. Future work may show that such networks can better preserve a variety of features in volumetric data.

Acknowledgements

This work is supported in part by the National Science Foundation (NSF) under grant numbers IIS-2007444 and IIS-2006710, and by the U.S. Department of Energy, Office of Science, Office of Advanced Scientific Computing Research, under Award Number(s) DE-SC-0019039.

References

  • [BLL18] Berger M., Li J., Levine J. A.: A generative model for volume rendering. IEEE transactions on visualization and computer graphics 25, 4 (2018), 1636–1650.
  • [BRLP19] Ballester-Ripoll R., Lindstrom P., Pajarola R.: TTHRESH: Tensor compression for multidimensional visual data. IEEE transactions on visualization and computer graphics (2019).
  • [BRP16] Ballester-Ripoll R., Pajarola R.: Lossy volume compression using Tucker truncation and thresholding. The Visual Computer 32, 11 (2016), 1433–1446.
  • [CCM04] Cook A. W., Cabot W., Miller P. L.: The mixing transition in Rayleigh-Taylor instability. Journal of Fluid Mechanics 511 (2004), 333–362. doi:10.1017/S0022112004009681.
  • [CZ19] Chen Z., Zhang H.: Learning implicit fields for generative shape modeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019), pp. 5939–5948.
  • [DC16] Di S., Cappello F.: Fast error-bounded lossy HPC data compression with SZ. In 2016 ieee international parallel and distributed processing symposium (ipdps) (2016), IEEE, pp. 730–739.
  • [DMG20] Díaz J., Marton F., Gobbetti E.: Interactive spatio-temporal exploration of massive time-varying rectilinear scalar volumes based on a variable bit-rate sparse representation over learned dictionaries. Computers & Graphics (2020).
  • [DNJ20] Davies T., Nowrouzezahrai D., Jacobson A.: Overfit neural networks as a compact shape representation. arXiv preprint arXiv:2009.09808 (2020).
  • [GBG14] Gyulassy A., Bremer P.-T., Grout R., Kolla H., Chen J., Pascucci V.: Stability of dissipation elements: A case study in combustion. Computer Graphics Forum 33, 3 (2014), 51–60.
  • [GIGM12] Gobbetti E., Iglesias Guitián J. A., Marton F.: COVRA: A compression-domain output-sensitive volume rendering architecture based on a sparse representation of voxel blocks. Computer Graphics Forum 31, 3pt4 (2012), 1315–1324.
  • [GLDL14] Guo F., Li H., Daughton W., Liu Y.-H.: Formation of hard power laws in the energetic particle spectra resulting from relativistic magnetic reconnection. Phys. Rev. Lett. 113 (Oct. 2014), 155005. doi:10.1103/PhysRevLett.113.155005.
  • [GWGS02] Guthe S., Wand M., Gonser J., Straßer W.: Interactive rendering of large volume data sets. In IEEE Visualization, 2002. VIS 2002. (2002), IEEE, pp. 53–60.
  • [GYH20] Guo L., Ye S., Han J., Zheng H., Gao H., Chen D. Z., Wang J.-X., Wang C.: SSR-VFD: Spatial super-resolution for vector field data analysis and visualization. In 2020 IEEE Pacific Visualization Symposium (PacificVis) (2020), IEEE Computer Society, pp. 71–80.
  • [HAESB20] Hao Z., Averbuch-Elor H., Snavely N., Belongie S.: DualSDF: Semantic shape manipulation using a two-level representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2020), pp. 7631–7641.
  • [HMD15] Han S., Mao H., Dally W. J.: Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015).
  • [HSB20] Hoang D., Summa B., Bhatia H., Lindstrom P., Klacansky P., Usher W., Bremer P.-T., Pascucci V.: Efficient and flexible hierarchical data layouts for a unified encoding of scalar field precision and resolution. IEEE Transactions on Visualization and Computer Graphics (2020).
  • [HTZ19] Han J., Tao J., Zheng H., Guo H., Chen D. Z., Wang C.: Flow field reduction via reconstructing vector data from 3-d streamlines using deep learning. IEEE computer graphics and applications 39, 4 (2019), 54–67.
  • [HW19] Han J., Wang C.: TSR-TVD: Temporal super-resolution for time-varying data analysis and visualization. IEEE Transactions on Visualization and Computer Graphics 26, 1 (2019), 205–215.
  • [HW20] Han J., Wang C.: SSR-TVD: Spatial super-resolution for time-varying data analysis and visualization. IEEE Transactions on Visualization and Computer Graphics (2020).
  • [HWG19] He W., Wang J., Guo H., Wang K.-C., Shen H.-W., Raj M., Nashed Y. S., Peterka T.: InSituNet deep image synthesis for parameter space exploration of ensemble simulations. IEEE transactions on visualization and computer graphics 26, 1 (2019), 23–33.
  • [HZRS16] He K., Zhang X., Ren S., Sun J.: Identity mappings in deep residual networks. In European conference on computer vision (2016), Springer, pp. 630–645.
  • [HZX20] Han J., Zheng H., Xing Y., Chen D. Z., Wang C.: V2V: A deep learning approach to variable-to-variable selection and translation for multivariate time-varying data. IEEE Transactions on Visualization and Computer Graphics (2020).
  • [IKK12] Iverson J., Kamath C., Karypis G.: Fast and effective lossy compression algorithms for scientific datasets. In European Conference on Parallel Processing (2012), Springer, pp. 843–856.
  • [IP99] Ihm I., Park S.: Wavelet-based 3D compression scheme for interactive visualization of very large volume data. Computer graphics forum 18, 1 (1999), 3–15.
  • [IS15] Ioffe S., Szegedy C.: Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International Conference on Machine Learning (2015), pp. 448–456.
  • [JGG20] Jakob J., Gross M., Günther T.: A fluid flow data set for machine learning and its application to neural flow map interpolation. IEEE Transactions on Visualization and Computer Graphics (2020).
  • [JJHZ20] Jiang Y., Ji D., Han Z., Zwicker M.: Sdfdiff: Differentiable rendering of signed distance fields for 3d shape optimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2020), pp. 1251–1261.
  • [KB14] Kingma D. P., Ba J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
  • [LDT18] Liang X., Di S., Tao D., Li S., Li S., Guo H., Chen Z., Cappello F.: Error-controlled lossy compression optimized for high compression ratios of scientific datasets. In 2018 IEEE International Conference on Big Data (Big Data) (2018), IEEE, pp. 438–447.
  • [Lin14] Lindstrom P.: Fixed-rate compressed floating-point arrays. IEEE transactions on visualization and computer graphics 20, 12 (2014), 2674–2683.
  • [LPW08] Li Y., Perlman E., Wan M., Yang Y., Meneveau C., Burns R., Chen S., Szalay A., Eyink G.: A public turbulence database cluster and applications to study lagrangian evolution of velocity increments in turbulence. Journal of Turbulence, 9 (2008), N31.
  • [LSE11] Lakshminarasimhan S., Shah N., Ethier S., Klasky S., Latham R., Ross R., Samatova N. F.: Compressing the incompressible with ISABELA: In-situ reduction of spatio-temporal data. In European Conference on Parallel Processing (2011), Springer, pp. 366–379.
  • [MON19] Mescheder L., Oechsle M., Niemeyer M., Nowozin S., Geiger A.: Occupancy networks: Learning 3d reconstruction in function space. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019), pp. 4460–4470.
  • [MST20] Mildenhall B., Srinivasan P. P., Tancik M., Barron J. T., Ramamoorthi R., Ng R.: NeRF: Representing scenes as neural radiance fields for view synthesis. In ECCV (2020).
  • [Mur93] Muraki S.: Volume data and wavelet transforms. IEEE Computer Graphics and applications 13, 4 (1993), 50–56.
  • [OMN19] Oechsle M., Mescheder L., Niemeyer M., Strauss T., Geiger A.: Texture fields: Learning texture representations in function space. In Proceedings of the IEEE International Conference on Computer Vision (2019), pp. 4531–4540.
  • [PFS19] Park J. J., Florence P., Straub J., Newcombe R., Lovegrove S.: DeepSDF: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019), pp. 165–174.
  • [PG18] Patchett J. M., Gisler G. R.: Deep water impact ensemble data set in 2018 ieee scivis contest, 2018.
  • [PNG18] Peterka T., Nashed Y. S., Grindeanu I., Mahadevan V. S., Yeh R., Tricoche X.: Foundations of multivariate functional approximation for scientific data. In 2018 IEEE 8th Symposium on Large Data Analysis and Visualization (LDAV) (2018), IEEE, pp. 61–71.
  • [RGG13] Rodriguez M. B., Gobbetti E., Guitián J. A. I., Makhinya M., Marton F., Pajarola R., Suter S. K.: A survey of compressed GPU-based direct volume rendering. In Eurographics (STARs) (2013), pp. 117–136.
  • [SMB20] Sitzmann V., Martel J., Bergman A., Lindell D., Wetzstein G.: Implicit neural representations with periodic activation functions. Advances in Neural Information Processing Systems 33 (2020).
  • [SMP13] Suter S. K., Makhynia M., Pajarola R.: TAMRESH - tensor approximation multiresolution hierarchy for interactive volume visualization. Computer Graphics Forum 32, 3pt2 (2013), 151–160.
  • [SPCT18] Soler M., Plainchault M., Conche B., Tierny J.: Topologically controlled lossy compression. In 2018 IEEE Pacific Visualization Symposium (PacificVis) (2018), IEEE, pp. 46–55.
  • [SW03] Schneider J., Westermann R.: Compression domain volume rendering. In IEEE Visualization, 2003. VIS 2003. (2003), IEEE, pp. 293–300.
  • [TDCC17] Tao D., Di S., Chen Z., Cappello F.: Significantly improving lossy compression for scientific data sets based on multidimensional prediction and error-controlled quantization. In 2017 IEEE International Parallel and Distributed Processing Symposium (IPDPS) (2017), IEEE, pp. 1129–1139.
  • [TSM20] Tancik M., Srinivasan P., Mildenhall B., Fridovich-Keil S., Raghavan N., Singhal U., Ramamoorthi R., Barron J., Ng R.: Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems 33 (2020).
  • [WBK04] Wang W., Bruyere C., Kuo B.: Competition data set and description in 2004 ieee visualization design contest, 2004.
  • [WCTW19] Weiss S., Chu M., Thuerey N., Westermann R.: Volumetric isosurface rendering with deep learning-based super-resolution. arXiv preprint arXiv:1906.06520 (2019).
  • [WJA16] Wald I., Johnson G. P., Amstutz J., Brownlee C., Knoll A., Jeffers J., Günther J., Navrátil P.: Ospray-a cpu ray tracing framework for scientific visualization. IEEE transactions on visualization and computer graphics 23, 1 (2016), 931–940.
  • [WMB11] Woodring J., Mniszewski S., Brislawn C., DeMarle D., Ahrens J.: Revisiting wavelet compression for large-scale climate data using JPEG 2000 and ensuring data precision. In 2011 IEEE Symposium on Large Data Analysis and Visualization (2011), IEEE, pp. 31–38.
  • [WN] Whalen D., Norman M. L.: Competition data set and description. in 2008 ieee visualization design contest (2008).
  • [XFCT18] Xie Y., Franz E., Chu M., Thuerey N.: tempoGAN: A temporally coherent, volumetric GAN for super-resolution fluid flow. ACM Transactions on Graphics (TOG) 37, 4 (2018), 1–15.
  • [YL95] Yeo B.-L., Liu B.: Volume rendering of DCT-based compressed 3D scalar data. IEEE Transactions on Visualization and Computer Graphics 1, 1 (1995), 29–43.