Diego Royo Miguel Crespo Jorge Garcia-Pueyo
mitransient is a library adds support to Mitsuba 3 for doing transient simulations, with amazing support for non-line-of-sight (NLOS) data capture simulations.
- Foundation ready to use: easy interface to convert your algorithms to the transient domain.
- Python-only library for doing transient rendering in both CPU and GPU.
- Several integrators already implemented: transient pathtracing (also adapted for NLOS scenes) and transient volumetric pathtracing.
- Cross-platform: Mitsuba 3 has been tested on Linux (x86_64), macOS (aarch64, x86_64), and Windows (x86_64).
- Temporal domain filtering.
Tip
Check out our online documentation (mitransient.readthedocs.io) and our code examples:
Featuring General Usage, Transient rendering, and Non-line-of-sight (NLOS) transient rendering
This project was created by Miguel Crespo and expanded by Diego Royo and Jorge Garcia-Pueyo. Also see the original Mitsuba 3 license and contributors.
If you use our code in your project, please consider citing us using the following:
@misc{mitransient,
title = {mitransient},
author = {Royo, Diego and Crespo, Miguel and Garcia-Pueyo, Jorge},
year = 2023,
journal = {GitHub repository},
doi = {https://doi.org/10.5281/zenodo.11032518},
publisher = {GitHub},
howpublished = {\url{https://github.com/diegoroyo/mitransient}}
}
Additionally, the NLOS features were re-implemented from our publication Non-line-of-sight transient rendering. Please also consider citing us if you use them:
@article{royo2022non,
title = {Non-line-of-sight transient rendering},
author = {Diego Royo and Jorge García and Adolfo Muñoz and Adrian Jarabo},
year = 2022,
journal = {Computers & Graphics},
doi = {https://doi.org/10.1016/j.cag.2022.07.003},
issn = {0097-8493},
url = {https://www.sciencedirect.com/science/article/pii/S0097849322001200}
}
Conventional rendering is referred to as steady state, where the light propagation speed is assumed to be infinite. In contrast, transient rendering breaks this assumption allowing us to simulate light in motion (see the teaser image for a visual example).
For example, path tracing algorithms integrate over multiple paths that connect a light source with the camera. For a known path, transient path tracing uses the very complex formula of time = distance / speed
(see [Two New Sciences by Galileo]) to compute the time
when each photon arrives at the camera from the path's distance
and light's speed
. This adds a new time
dimension to the captured images (i.e. it's a video now). The simulations now take new parameters as input: when to start recording the video, how long is each time step (framerate), and how many frames to record.
Note: note that the time
values we need to compute are very small (e.g. light takes only ~3.33 * 10^-9 seconds to travel 1 meter), time
is usually measured in optical path distance. See Wikipedia for more information. TL;DR opl = distance * refractive_index
We provide the package via PyPI. To install mitransient
you need to run:
pip install mitransient
Important
In order to run mitransient
you also need to install Mitsuba 3. If you have installed Mitsuba 3 via pip
(pip install mitsuba
) you will only have access to the llvm_ad_rgb
and cuda_ad_rgb
variants. If you want to use other variants (e.g. NLOS simulations can greatly benefit from the llvm_mono
variant which only propagates one wavelength), then we recommend that you compile Mitsuba 3 yourself following this tutorial and enable the following variants: ["scalar_mono", "llvm_mono", "llvm_ad_mono", "cuda_mono", "cuda_ad_mono", "scalar_rgb", "llvm_rgb", "llvm_ad_rgb", "cuda_rgb", "cuda_ad_rgb"]
.
Python >= 3.8
Mitsuba3 >= 3.5.0
- (optional) For computation on the GPU:
Nvidia driver >= 495.89
- (optional) For vectorized / parallel computation on the CPU:
LLVM >= 11.1
At this point, you should be able to import mitsuba
and import mitransient
in your Python code (careful about setting the correct PATH
environment variable if you have compiled Mitsuba 3 yourself, see the section below).
For NLOS data capture simulations, see https://github.com/diegoroyo/tal. tal
is a toolkit that allows you to create and simulate NLOS scenes with an easier shell interface instead of directly from Python.
If you have opted for using a custom (non-default installation through pip
) Mitsuba 3, you have several options for it. The idea here is to be able to control which version of Mitsuba will be loaded on demand.
- One solution is to directly execute
setpath.sh
provided after the compilation of the Mitsuba 3 repo (More info). This shell script will modify thePATH
andPYTHONPATH
variables to load first this version of Mitsuba. - Another solution following the previous one is to directly set yourself the
PYTHONPATH
environment variable as you wish. - Another solution for having a custom version globally available is by using
pip install . --editable
. This will create a symlink copy of the package files inside the correspondingsite-packages
folder and will be listed as a package installed ofpip
and will be available as other packages installed. If you recompile them, you will still have the newest version directly to use. Please follow these instructions:- Go to
<mitsuba-path>/mitsuba3/build/python/drjit
and executepip install . --editable
. - Go to
<mitsuba-path>/mitsuba3/build/python/mitsuba
and executepip install . --editable
.
- Go to
- If you are a user of Jupyter Notebooks, the easiest solution will be to add the following snippet of code to modify the notebook's
PYTHONPATH
:
import sys
sys.path.insert(0, '<mitsuba-path>/mitsuba3/build/python')
import mitsuba as mi
Tip
Check out the examples
folder for practical usage!
You are now prepared to render your first transient scene with mitransient. Running the code below will render the famous Cornell Box scene in transient domain and show a video.
import mitsuba as mi
mi.set_variant('llvm_ad_rgb')
import mitransient as mitr
scene = mi.load_dict(mitr.cornell_box())
transient_integrator = scene.integrator()
transient_integrator.prepare_transient(scene, sensor=0)
img_steady, img_transient = transient_integrator.render(scene, spp=1024)
import numpy as np
img_transient_tonemap = mitr.vis.tonemap_transient(
np.moveaxis(img_transient, 0, 1)
)
mitr.vis.show_video(
img_transient_tonemap,
axis_video=2,
)
mitransient
implements the following plugins which can be used in scene XML files. To view a description of their parameters, click on the name of your desired plugin.
film
:transient_hdr_film
: Transient equivalent of Mitsuba 3'shdrfilm
plugin.
integrators
:transient_path
: Transient path tracing for line-of-sight scenes. If you want to do NLOS simulations, usetransientnlospath
instead.transient_nlos_path
: Transient path tracing with specific sampling routines for NLOS scenes (e.g. laser sampling and hidden geometry sampling of the "Non-Line-of-Sight Transient Rendering" paper).transient_prbvolpath
: Path Replay Backpropagation for volumetric path tracing. Implemented by Miguel Crespo, untested.
sensor
:nlos_capture_meter
: Can be attached to one of the scene's geometries, and measures uniformly-spaced points on such geometry (e.g. relay wall).
Our test suite can be run using pytest
on the root folder of the repo.