# Introduction Lightplane implements a highly memory-efficient differentiable radiance field renderer, and a module for unprojecting features from images to general 3D grids, described in our paper: **Lightplane: Highly-Scalable Components for Neural 3D Fields** *Ang Cao*, *Justin Johnson*, *Andrea Vedaldi*, *David Novotny* [Project Page](https://lightplane.github.io/) | [Arxiv](https://arxiv.org/abs/2404.19760) | [Docs](https://lightplane.github.io/docs/) **Lightplane can differentiably render batches of FullHD images** while consuming **less than a GigaByte** of GPU memory. Rendering / backpropagation speed is comparable to existing methods, which however exceed memory limits already for a single QVGA image (640x480). As such, Lightplane's memory and compute efficiency allows to drastically scale-up the latest neural field methods. ## Documentation Please visit the [documentation website](https://lightplane.github.io/docs/) for a detailed project and API description. ## Overview Lightplane is a pair of highly-scalable components for rendering neural 3D fields. It consists of two components: **Lightplane Renderer** and **Lightplane Splatter**. - **Lightplane Renderer** (3D --> 2D) is a radiance field renderer which, given a camera viewpoint and a scene grid structure (triplane, voxel grid, etc.), renders the corresponding image. - **Lightplane Splatter** (2D --> 3D) is dual to Renderer: it takes image features and lifts/unprojects them to a 3D grid structure. These two components are implemented as a pair of fused auto-optimized [Triton](https://github.com/openai/triton) GPU kernels. Key features include: - **Memory Efficiency**: Lightplane dramatically saves memory usage with competing speed compared to existing methods. When rendering a 256x256 image for backpropgation, it uses 10 MB GPU memory while Pytorch implementation uses 10 GB. - **Highly Scalable**: The memory efficiency of Lightplane significantly boosts the scalability of 3D reconstruction/generation models. It allows lifting/unprojecting of hundred views, rendering high-resolution multiview images, with little memory usage. - **Easy to Extend**: Lightplane is a general design for 3D grid structures. Besides voxel grids and TriPlanes, it can be easily extended to other 3D grids (e.g. hash grid, HexPlane ...). More support is coming. - **Easy to Use**: Lightplane is powerful, configurable and easy to use. ## Installation 1) [Install PyTorch](https://pytorch.org/get-started/locally/#start-locally). 2) Install [COG](https://pypi.org/project/cogapp/), [Triton](https://github.com/openai/triton) and [plotly](https://plotly.com/python/): ```bash pip install cogapp pip install triton pip install plotly ``` 3) Install the LightPlane package: ```bash pip install -e . ``` 4) (Optional for running example code and jupyter notebooks) 1) [Install PyTorch3D](https://github.com/facebookresearch/pytorch3d/blob/main/INSTALL.md) 2) `pip install notebook` 3) `pip install imageio opencv-python lpips matplotlib` 4) `pip install configargparse tqdm` ## Usage Lightplane Renderer and Splatter are packaged into ```torch.nn.Module```, which are easy to use as plugins in neural 3D field pipelines. **Using Lightplane Renderer**: ```python from lightplane import LightplaneRenderer, LightplaneSplatter, Rays # Setup the gpu device the rendering module. device=torch.device("cuda:0") renderer = LightplaneRenderer( num_samples=128, color_chn=3, grid_chn=3, mlp_hidden_chn=32, mlp_n_layers_opacity=2, mlp_n_layers_trunk=2, mlp_n_layers_color=2, ray_embedding_num_harmonics=3, ).to(device) # Create 128 random rendering rays: num_rays = 128 rays = Rays( directions=torch.randn(num_rays, 3) * 0.1, origins=torch.randn(num_rays, 3), grid_idx=torch.zeros(num_rays, dtype=torch.int32), near=torch.full(num_rays, 0.1), far=torch.full(num_rays, 3.0), ) # Create a random triplane. batch_size = 1 channels = 16 H = W = D = 64 # triplane spatial size feature_grid = [ torch.randn(batch_size, 1, H, W, channels, device=device), torch.randn(batch_size, D, 1, W, channels, device=device), torch.randn(batch_size, D, H, 1, channels, device=device), ] # Render the triplane. ( rendered_ray_length, negative_log_transmittances, rendered_features, ) = renderer(rays=rays.to(device), feature_grid=input_grid) # get the rendering alpha mask in range [0, 1] from negative log transmittances alpha = (-negative_log_transmittances).exp() ``` **Using Lightplane Splatter**: The following splats features to a voxel grid using `Rays` from the previous snippet: ```python # Initialize the splatter module splatter = LightplaneSplatter( num_samples=256, grid_chn=channels, ).to(device) # Assign the per-ray image features to splat to the `encoding` field of `rays`. rays.encoding = torch.randn(num_rays, channels, device=device) # Set the size of the output voxel grid which we splat features to: grid_size = [(batch_size, D, H, W, channels)] # Run Splatter. splatted_grid = splatter( rays=rays, grid_size=grid_size, ) ``` ## Examples Please visit [examples](./examples.md) for more detailed example python scripts and jupyter notebooks. ## TODO We plan to implement two major new functionalities: 1) Support for hash-grids 2) Support for temporal grids, e.g. HexPlane ## License Lightplane is released under the [BSD 3.0 License](https://github.com/facebookresearch/lightplane/tree/main/LICENSE). ## Citation ``` @article{cao2024lightplane, author = {Ang Cao and Justin Johnson and Andrea Vedaldi and David Novotny}, title = {Lightplane: Highly-Scalable Components for Neural 3D Fields}, journal = {arXiv}, year = {2024}, } ```