ACM Transactions on

Graphics (TOG)

Latest Articles

Poly-Spline Finite-Element Method

We introduce an integrated meshing and finite-element method pipeline enabling solution of partial differential equations in the volume enclosed by a boundary representation. We construct a hybrid hexahedral-dominant mesh, which contains a small number of star-shaped polyhedra, and build a set of high-order bases on its elements, combining... (more)

Video Extrapolation Using Neighboring Frames

With the popularity of immersive display systems that fill the viewer’s field of view (FOV) entirely, demand for wide FOV content has increased. A video extrapolation technique based on reuse of existing videos is one of the most efficient ways to produce wide FOV content. Extrapolating a video poses a great challenge, however, due to the... (more)

Non-line-of-sight Imaging with Partial Occluders and Surface Normals

Imaging objects obscured by occluders is a significant challenge for many applications. A camera that could “see around corners” could... (more)

A Unified Framework for Compression and Compressed Sensing of Light Fields and Light Field Videos

In this article we present a novel dictionary learning framework designed for compression and... (more)


New Submission Requirements

As of October 2018, ACM TOG requires submissions for review to be anonymous. See the Author Guidelines for details.  

About TOG

ACM TOG is the foremost peer-reviewed journal in the area of computer graphics. 

Recent impact factor calculations from Thomson Reuters give ACM TOG an impact factor of 4.096 and an Eigenfactor Score of 0.029, giving it the top ranking among the 104 journals in the Computer Science: Software Engineering category. 

read more
Forthcoming Articles
Dynamic Graph CNN for Learning on Point Clouds

Point clouds provide a flexible and scalable geometric representation suitable for countless applications in computer graphics; they also comprise the raw output of most 3D data acquisition devices. Hence, the design of intelligent computational models that act directly on point clouds is critical, especially when efficiency considerations or noise preclude the possibility of expensive denoising and meshing procedures. While hand-designed features on point clouds have long been proposed in graphics and vision, however, the recent success of convolutional neural networks for image analysis suggests the value of adapting insight from CNN to the point cloud world. We propose a new neural network module EdgeConv suitable for CNN-based high-level tasks on point clouds including classification and segmentation. EdgeConv is differentiable and can be plugged into existing architectures. Compared to existing modules operating in extrinsic space or treating each point independently, EdgeConv has several appealing properties: It incorporates local neighborhood information; it can be stacked to learn global shape properties; and in multi-layer systems affinity in feature space captures semantic characteristics over potentially long distances in the original embedding. Beyond proposing this module, we provide extensive evaluation and analysis revealing that EdgeConv captures and exploits fine-grained geometric properties of point clouds.

Volume Path Guiding Based on Zero-Variance Random Walk Theory

Monte Carlo methods are commonly used for robust rendering of scenes with volumetric participating media. The efficiency of these approaches is directly linked to the manner in which random sampling decisions are made during light path construction. Notably, path construction is influenced by scattering direction and distance sampling, Russian roulette, and splitting strategies. We present a new volumetric path construction technique where all these sampling decisions are guided by a cached estimate of the adjoint light transport solution. Our sampling strategy is based on the theory of zero-variance transport estimators, and it accounts for the spatial and directional variation in volumetric transport. Specifically, we construct paths incrementally by sampling collision distances proportionally to the product of transmittance and the adjoint transport solution (i.e., in-scattered radiance). Scattering directions are likewise sampled according to the product of the phase function and the incident radiance estimate. Combined with an adaptive Russian roulette and splitting strategy tailored to volumes, we demonstrate about an order-of-magnitude variance reduction compared to modern uni-directional methods. Consequently, our approach can render scenes that are otherwise intractable for such methods, while still retaining their simplicity (compared to, e. g., bi-directional methods).

Blockwise Multi-Order Feature Regression for Real-Time Path Tracing Reconstruction

Path tracing produces realistic results including global illumination using a unified simple rendering pipeline. Reducing the amount of noise to imperceptible levels without post-processing requires thousands of samples per pixel (spp), while currently it is only possible to render extremely noisy 1 spp frames in real time with desktop GPUs. However, post-processing can utilize feature buffers, which contain noise-free auxiliary data available in the rendering pipeline. Previously, regression-based noise filtering methods have only been used in offline rendering due to their high computational cost. In this paper we propose a novel regression-based reconstruction pipeline, called Blockwise Multi-Order Feature Regression (BMFR), tailored for path-traced 1 spp inputs that runs in real time. The high speed is achieved with a fast implementation of augmented QR factorization and by using stochastic regularization to address rank-deficient feature data. The proposed algorithm is 1.8× faster than the previous state-of-the-art real-time path tracing reconstruction method while producing better quality frame sequences.

Coding Scheme Optimization for Fast Fluorescence Lifetime Imaging

Neural Rendering and Reenactment of Human Actor Videos

We propose a method for generating (near) video-realistic animations of real humans under user control. In contrast to conventional human character rendering, we do not require the availability of a production-quality photo-realistic 3D model of the human, but instead rely on a video sequence in conjunction with a (medium-quality) controllable 3D template model of the person. With that, our approach significantly reduces production cost compared to conventional rendering approaches based on production-quality 3D models, and can also be used to realistically edit existing videos. Technically, this is achieved by training a neural network that translates simple synthetic images of a human character into realistic imagery. For training our networks, we first track the 3D motion of the person in the video using the template model, and subsequently generate a synthetically rendered version of the video. These images are then used to train a conditional generative adversarial network that translates synthetic images of the 3D model into realistic imagery of the human. We evaluate our method for the reenactment of another person that is tracked in order to obtain the motion data, and show video results generated from artist-designed skeleton motion. Our results outperform the state-of-the-art in learning-based human image synthesis.

Redefining A in RGBA: Towards a Standard for Graphical 3D Printing

Advances in multimaterial 3D printing have the potential to reproduce various visual appearance attributes of an object in addition to its shape. Since many existing 3D file formats encode color and translucency by RGBA textures mapped to 3D shapes, RGBA information is particularly important for practical applications. In contrast to color (encoded by RGB), translucency (encoded by A) is neither linked to any measurable physical nor perceptual quantity and is, therefore, open for interpretation. In this paper, we propose a rigorous definition for A suitable for graphical 3D printing which links both optical material properties and perceptual uniformity for human observers. By deriving our definition from the absorption and scattering coefficients of virtual reference materials with an isotropic phase function, we achieve two important properties. First, a simple adjustment of A is possible, which preserves the translucency appearance if an object is rescaled for printing. Second, determining A for a real material can be achieved by minimizing a distance function between light transport measurements of this material (conducted by commercial spectrophotometers) and simulated measurements of the reference materials. Finally, we derive from visual experiments an embedding of A into a nearly perceptually-uniform scale of translucency for the reference materials.

The Vector Heat Method

This paper describes a method for efficiently computing parallel transport of tangent vectors on curved surfaces, or more generally, any vector-valued data on a curved manifold. More precisely, it extends a vector field defined over any region to the rest of the domain via parallel transport along shortest geodesics. This basic operation enables fast, robust algorithms for extrapolating level set velocities, inverting the exponential map, computing geometric medians and Karcher/Fréchet means of arbitrary distributions, constructing centroidal Voronoi diagrams, and finding consistently ordered landmarks. Rather than evaluate parallel transport by explicitly tracing geodesics, we show that it can be computed via a short-time heat flow involving the connection Laplacian. As a result, transport can be achieved by solving three prefactored linear systems, each akin to a standard Poisson problem. Moreover, to implement the method we need only a discrete connection Laplacian, which we describe for a variety of geometric data structures (point clouds, polygon meshes, etc.). We also study the numerical behavior of our method, showing empirically that it converges under refinement, and augment the construction of intrinsic Delaunay triangulations (iDT) so that they can be used in the context of tangent vector field processing.

All ACM Journals | See Full Journal Index

Search TOG
enter search term and/or author name