English | 简体中文
XRNeRF is an open-source PyTorch-based codebase for Neural Radiance Field (NeRF). It is a part of the OpenXRLab project.
xrnerf_demo.mp4
This page provides basic tutorials about the usage of XRNeRF. For installation instructions, please see installation.md.
More details can be found in benchmark.md.
Supported scene-NeRF methods:
(click to collapse)
- NeRF (ECCV'2020)
- Mip-NeRF (ICCV'2021)
- KiloNeRF (ICCV'2021)
- Instant NGP (SIGGRAPH'2022)
Supported human-NeRF methods:
(click to collapse)
- NeuralBody (CVPR'2021)
- AniNeRF (ICCV'2021)
- GNR
Wanna see more methods supported? Post method you want see in XRNeRF on our wishlist.
It is recommended to symlink the dataset root to $PROJECT/data
.
If your folder structure is different, you may need to change the corresponding paths in config files.
xrnerf
├── xrnerf
├── docs
├── configs
├── test
├── extensions
├── data
│ ├── nerf_llff_data
│ ├── nerf_synthetic
│ ├── multiscale
│ ├── ...
For more information on data preparation, please see dataset_preparation.md
We provide detailed installation tutorial for XRNeRF, users can install from scratch or use provided dockerfile.
It is recommended to start by creating a docker image:
docker build -f ./docker/Dockerfile --rm -t xrnerf .
For more information, please follow our installation tutorial.
In XRNeRF, model components are basically categorized as 4 types.
- network: the whole nerf model pipeline, usually contains a embedder, mlp and render.
- embedder: convert point-position and viewdirection data into embedded data, embedder can be function only or with trainable paramters.
- mlp: use the output of embedder as input, and output raw data (the rgb and density value at sampled position) for render, usually contains FC layers.
- render: receive mlp's raw data, output the rgb value at a pixel.
Following some basic pipelines (e.g., NerfNetwork
), the model structure
can be customized through config files with no pains.
To write a new nerf network, you need to inherit from BaseNerfNetwork
,
which defines the following abstract methods.
train_step()
: forward method of the training mode.val_step()
: forward method of the testing mode.
NerfNetwork is a good example which show how to do that.
To be specific, if we want to implement some new components, there are several things to do.
-
create a new file in
xrnerf/models/networks/my_networks.py
.from ..builder import NETWORKS from .nerf import NerfNetwork @NETWORKS.register_module() class MyNerfNetwork(NerfNetwork): def __init__(self, cfg, mlp=None, mlp_fine=None, render=None): super().__init__(cfg, mlp, mlp_fine, render) def forward(self, data): .... def train_step(self, data, optimizer, **kwargs): .... def val_step(self, data, optimizer=None, **kwargs): ....
-
Import the module in
xrnerf/models/networks/__init__.py
from .my_networks import MyNerfNetwork
-
modify the config file from
model = dict( type='NerfNetwork', ....
to
model = dict( type='MyNerfNetwork', ....
To implement some new components for embedder/mlp/render, procedure is similar to above.
- To write a new nerf embedder, you need to inherit from
nn.Module
orBaseEmbedder
, and define theforward
method. BaseEmbedder is a good example. - To write a new nerf mlp, you need to inherit from
nn.Module
orBaseMLP
, and define theforward
method. NerfMLP is a good example. - To write a new nerf render, you need to inherit from
nn.Module
orBaseRender
, and define theforward
method. NerfRender is a good example.
XRNeRF use mmcv.runner.IterBasedRunner
to control training, and mmcv.runner.EpochBasedRunner
to for test mode.
In training mode, the max_iters
in config file decide how many iters.
In test mode, max_iters
is forced to change to 1, which represents only 1 epoch to test.
python run_nerf.py --config configs/nerf/nerf_blender_base01.py --dataname lego
Arguments are:
--config
: config file path.--dataname
: select which data under dataset directory.
We have provided model iter_200000.pth
for test, download from here
python run_nerf.py --config configs/nerf/nerf_blender_base01.py --dataname lego --test_only --load_from iter_200000.pth
Arguments are:
--config
: config file path.--dataname
: select which data under dataset directory.--test_only
: influence on whole testset once.--load_from
: load which checkpoint to test, this will overwrite the originalload_from
in config file to for convenience.
Currently, we provide some tutorials for users to
Except for that,The document also includes the following
If you find this project useful in your research, please consider cite:
@misc{xrnerf,
title={OpenXRLab Neural Radiance Field Toolbox and Benchmark},
author={XRNeRF Contributors},
howpublished = {\url{https://github.com/openxrlab/xrnerf}},
year={2022}
}
The license of our codebase is Apache-2.0. Note that this license only applies to code in our library, the dependencies of which are separate and individually licensed. We would like to pay tribute to open-source implementations to which we rely on. Please be aware that using the content of dependencies may affect the license of our codebase. Some supported methods may carry additional licenses.
We appreciate all contributions to improve XRNeRF. Please refer to CONTRIBUTING.md for the contributing guideline.
XRNeRF is an open source project that is contributed by researchers and engineers from both the academia and the industry. We appreciate all the contributors who implement their methods or add new features, as well as users who give valuable feedbacks. We wish that the framework and benchmark could serve the growing research community by providing a flexible framework to reimplement existing methods and develop their own new models.
- XRPrimer: OpenXRLab foundational library for XR-related algorithms.
- XRSLAM: OpenXRLab Visual-inertial SLAM Toolbox and Benchmark.
- XRSfM: OpenXRLab Structure-from-Motion Toolbox and Benchmark.
- XRLocalization: OpenXRLab Visual Localization Toolbox and Server.
- XRMoCap: OpenXRLab Multi-view Motion Capture Toolbox and Benchmark.
- XRMoGen: OpenXRLab Human Motion Generation Toolbox and Benchmark.
- XRNeRF: OpenXRLab Neural Radiance Field (NeRF) Toolbox and Benchmark.