Temporal Sample Reuse for Next Event Estimation and Path Guiding for Real-Time Path Tracing

Loading...
Thumbnail Image
Date
2020
Journal Title
Journal ISSN
Volume Title
Publisher
The Eurographics Association
Abstract
Good importance sampling is crucial for real-time path tracing where only low sample budgets are possible. We present two efficient sampling techniques tailored for massively-parallel GPU path tracing which improve next event estimation (NEE) for rendering with many light sources and sampling of indirect illumination. As sampling densities need to vary spatially, we use an octree structure in world space and introduce algorithms to continuously adapt the partitioning and distribution of the sampling budget. Both sampling techniques exploit temporal coherence by reusing samples from the previous frame: For NEE we collect sampled, unoccluded light sources and show how to deduplicate, but also diffuse this information to efficiently sample light sources in the subsequent frame. For sampling indirect illumination, we present a compressed directional quadtree structure which is iteratively adapted towards high-energy directions using samples from the previous frame. The updates and rebuilding of all data structures takes about 1ms in our test scenes, and adds about 6ms at 1080p to the path tracing time compared to using state-of-the-art light hierarchies and BRDF sampling. We show that this additional effort reduces noise in terms of mean squared error by at least one order of magnitude in many situations.
Description

        
@inproceedings{
10.2312:sr.20201135
, booktitle = {
Eurographics Symposium on Rendering - DL-only Track
}, editor = {
Dachsbacher, Carsten and Pharr, Matt
}, title = {{
Temporal Sample Reuse for Next Event Estimation and Path Guiding for Real-Time Path Tracing
}}, author = {
Dittebrandt, Addis
 and
Hanika, Johannes
 and
Dachsbacher, Carsten
}, year = {
2020
}, publisher = {
The Eurographics Association
}, ISSN = {
1727-3463
}, ISBN = {
978-3-03868-117-5
}, DOI = {
10.2312/sr.20201135
} }
Citation