Skip to content

Official implementation of the paper PaletteNeRF: Palette-based Appearance Editing of Neural Radiance Fields

License

Notifications You must be signed in to change notification settings

zfkuang/PaletteNeRF

Repository files navigation

PaletteNeRF: Palette-based Appearance Editing of Neural Radiance Fields

This repository is the official PyTorch implementation of the method described in PaletteNeRF: Palette-based Appearance Editing of Neural Radiance Fields. It is based on the ashawkey's implementation of Instant-NGP (i.e. torch-ngp).

This repository is tested in Ubuntu 20.04 + PyTorch 1.13.1 + RTX 3090.

News: Our paper is accepted by CVPR 2023!

Install

git clone https://github.com/zfkuang/palettenerf.git
cd palettenerf
conda env create -f environment.yml
conda activate palettenerf

Build extension (optional)

We provide two options to build extension, same as torch-ngp: loading the extension at run-time (which is much slower), or building the extension with setup.py:

bash scripts/install_ext.sh

# or
cd raymarching
python setup.py build_ext --inplace 
pip install . 

Run

Data Preparation

We use the same data format as torch-ngp does. Currently, three datasets are supported: NeRF Synthetic Dataset, LLFF Dataset and Mip-NeRF 360 Dataset. For LLFF and Mip-NeRF 360 Dataset, data conversion is required:

python scripts/llff2nerf.py /path/to/the/scene/directory --images images_4 --downscale 4 --datatype <llff/mip360>

for instance:

python scripts/llff2nerf.py ./data/nerf_llff_data/fern --images images_4 --downscale 4 --datatype llff

and everything is ready to go.

Training

For convenience, we encapsulate all training and infering commands in scripts/run_<dataset>.sh. We also manually tuned some parameters of each scene for better reconstruction quality as torch-ngp suggested, and the configurations are under scripts/configs_<dataset>. To train our model, simply run (take lego as an example):

bash scripts/run_blender.sh scripts/configs_blender/lego.sh -m nerf
bash scripts/run_blender.sh scripts/configs_blender/lego.sh -m extract
bash scripts/run_blender.sh scripts/configs_blender/lego.sh -m palette

Running GUI

To edit the appearance with GUI, run the following command after the training is complete:

bash scripts/run_blender.sh scripts/configs_blender/lego.sh -m palette -g

A GUI window will automatically pop-up. Here's an example:

To understand how to operate the GUI, you can check our video demo provided in our project page.

Testing

To quantitatively evaluate our model or render test views, run:

bash scripts/run_blender.sh scripts/configs_blender/lego.sh -m palette -t

(Only available for LLFF and Mip-360 Dataset for now) To render a new video, run:

bash scripts/run_blender.sh scripts/configs_blender/lego.sh -m palette -v

The video trajectories are consistent with the original trajectories from NeRF and Mip-NeRF 360. All results are saved under results_palette.

Semantic-guided editing

Currently our code only supports semantic-guided editing on Mip360 dataset. For convinience, we provide a modified implementation of paper Language-driven Semantic Segmentation (LSeg) at third-party/lang-seg. Follow the installization steps in the repo (make sure you are running it in an independent environment since it might not be compatible with ours), then run:

python extract_lseg_feature.py --datadir <path/to/data/of/mip360>

This script will generate compressed semantic feature maps of all training images. You can then train and test the model using the same commands by replacing -m palette to -m palette_lseg.

Model Zoo

coming soon

Citation

We will highly appreciate it if you would like to cite our work via:

@article{kuang2022palettenerf,
  title={PaletteNeRF: Palette-based Appearance Editing of Neural Radiance Fields},
  author={Kuang, Zhengfei and Luan, Fujun and Bi, Sai and Shu, Zhixin and Wetzstein, Gordon and Sunkavalli, Kalyan},
  journal={arXiv preprint arXiv:2212.10699},
  year={2022}
}

Acknowledgement

  • The implementation of Instant-NGP is adapted from torch-ngp:

    @misc{torch-ngp,
        Author = {Jiaxiang Tang},
        Year = {2022},
        Note = {https://github.com/ashawkey/torch-ngp},
        Title = {Torch-ngp: a PyTorch implementation of instant-ngp}
    }
    
    @article{tang2022compressible,
        title = {Compressible-composable NeRF via Rank-residual Decomposition},
        author = {Tang, Jiaxiang and Chen, Xiaokang and Wang, Jingbo and Zeng, Gang},
        journal = {arXiv preprint arXiv:2205.14870},
        year = {2022}
    }
    
  • The color palette extraction code is adapted from posternerf with the method introduced in Efficient palette-based decomposition and recoloring of images via RGBXY-space geometry:

    @article{tojo2022posternerf,
    title = {Recolorable Posterization of Volumetric Radiance Fields Using Visibility-Weighted Palette Extraction},
    author = {Tojo, Kenji and Umetani, Nobuyuki},
    journal = {Computer Graphics Forum},
    number = {4},
    pages = {149-160},
    volume = {41},
    year = {2022}
    }
    
    @article{Tan:2018:EPD,
    author    = {Tan, Jianchao and Echevarria, Jose and Gingold, Yotam},
    title     = {Efficient palette-based decomposition and recoloring of images via RGBXY-space geometry},
    journal   = {ACM Transactions on Graphics (TOG)},
    volume    = {37},
    number    = {6},
    month     = dec,
    year      = {2018},
    articleno = {262},
    pages     = {262:1--262:10},
    numpages  = {10},
    issn = {0730-0301},
    doi = {10.1145/3272127.3275054},
    publisher = {ACM Press},
    address   = {New York, NY, USA},
    keywords  = {images, layers, painting, palette, generalized barycentric coordinates, convex hull, RGB, color space, recoloring, compositing, mixing}
    }
    
  • The semantic segmentation code is adapted from the official implementation of paper Language-driven Semantic Segmentation (LSeg):

    @inproceedings{
    li2022languagedriven,
    title={Language-driven Semantic Segmentation},
    author={Boyi Li and Kilian Q Weinberger and Serge Belongie and Vladlen Koltun and Rene Ranftl},
    booktitle={International Conference on Learning Representations},
    year={2022},
    url={https://openreview.net/forum?id=RriDjddCLN}
    }
    
  • The GUI is developed with DearPyGui.

About

Official implementation of the paper PaletteNeRF: Palette-based Appearance Editing of Neural Radiance Fields

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published