Revision d6bbefb0b68e6322711b518eac7f9ab4c1cc7b1e authored by Thomas Müller on 08 July 2025, 11:21:56 UTC, committed by Thomas Müller on 08 July 2025, 11:21:56 UTC
1 parent 1edc77e
python_api.cu
/*
* Copyright (c) 2020-2022, NVIDIA CORPORATION. All rights reserved.
*
* NVIDIA CORPORATION and its licensors retain all intellectual property
* and proprietary rights in and to this software, related documentation
* and any modifications thereto. Any use, reproduction, disclosure or
* distribution of this software and related documentation without an express
* license agreement from NVIDIA CORPORATION is strictly prohibited.
*/
/** @file python_api.cpp
* @author Thomas Müller & Alex Evans, NVIDIA
*/
#include <neural-graphics-primitives/common_device.cuh>
#include <neural-graphics-primitives/testbed.h>
#include <neural-graphics-primitives/thread_pool.h>
#include <json/json.hpp>
#include <pybind11/functional.h>
#include <pybind11/numpy.h>
#include <pybind11/pybind11.h>
#include <pybind11/stl.h>
#include <pybind11_json/pybind11_json.hpp>
#include <tiny-cuda-nn/vec_pybind11.h>
#include <filesystem/path.h>
#ifdef NGP_GUI
# include <imgui/imgui.h>
# ifdef _WIN32
# include <GL/gl3w.h>
# else
# include <GL/glew.h>
# endif
# include <GLFW/glfw3.h>
#endif
using namespace nlohmann;
namespace py = pybind11;
namespace ngp {
void Testbed::Nerf::Training::set_image(int frame_idx, pybind11::array_t<float> img, pybind11::array_t<float> depth_img, float depth_scale) {
if (frame_idx < 0 || frame_idx >= dataset.n_images) {
throw std::runtime_error{"Invalid frame index"};
}
py::buffer_info img_buf = img.request();
if (img_buf.ndim != 3) {
throw std::runtime_error{"image should be (H,W,C) where C=4"};
}
if (img_buf.shape[2] != 4) {
throw std::runtime_error{"image should be (H,W,C) where C=4"};
}
py::buffer_info depth_buf = depth_img.request();
dataset.set_training_image(
frame_idx,
{(int)img_buf.shape[1], (int)img_buf.shape[0]},
(const void*)img_buf.ptr,
(const float*)depth_buf.ptr,
depth_scale,
false,
EImageDataType::Float,
EDepthDataType::Float
);
}
void Testbed::override_sdf_training_data(py::array_t<float> points, py::array_t<float> distances) {
py::buffer_info points_buf = points.request();
py::buffer_info distances_buf = distances.request();
if (points_buf.ndim != 2 || distances_buf.ndim != 1 || points_buf.shape[0] != distances_buf.shape[0] || points_buf.shape[1] != 3) {
tlog::error() << "Invalid Points<->Distances data";
return;
}
std::vector<vec3> points_cpu(points_buf.shape[0]);
std::vector<float> distances_cpu(distances_buf.shape[0]);
for (size_t i = 0; i < points_cpu.size(); ++i) {
vec3 pos = *((vec3*)points_buf.ptr + i);
float dist = *((float*)distances_buf.ptr + i);
pos = (pos - m_raw_aabb.min) / m_sdf.mesh_scale + 0.5f * (vec3(1.0f) - (m_raw_aabb.max - m_raw_aabb.min) / m_sdf.mesh_scale);
dist /= m_sdf.mesh_scale;
points_cpu[i] = pos;
distances_cpu[i] = dist;
}
CUDA_CHECK_THROW(cudaMemcpyAsync(
m_sdf.training.positions.data(),
points_cpu.data(),
points_buf.shape[0] * points_buf.shape[1] * sizeof(float),
cudaMemcpyHostToDevice,
m_stream.get()
));
CUDA_CHECK_THROW(cudaMemcpyAsync(
m_sdf.training.distances.data(), distances_cpu.data(), distances_buf.shape[0] * sizeof(float), cudaMemcpyHostToDevice, m_stream.get()
));
CUDA_CHECK_THROW(cudaStreamSynchronize(m_stream.get()));
m_sdf.training.size = points_buf.shape[0];
m_sdf.training.idx = 0;
m_sdf.training.max_size = m_sdf.training.size;
m_sdf.training.generate_sdf_data_online = false;
}
pybind11::dict Testbed::compute_marching_cubes_mesh(ivec3 res3d, BoundingBox aabb, float thresh) {
mat3 render_aabb_to_local = mat3::identity();
if (aabb.is_empty()) {
aabb = m_testbed_mode == ETestbedMode::Nerf ? m_render_aabb : m_aabb;
render_aabb_to_local = m_render_aabb_to_local;
}
marching_cubes(res3d, aabb, render_aabb_to_local, thresh);
py::array_t<float> cpuverts({(int)m_mesh.verts.size(), 3});
py::array_t<float> cpunormals({(int)m_mesh.vert_normals.size(), 3});
py::array_t<float> cpucolors({(int)m_mesh.vert_colors.size(), 3});
py::array_t<int> cpuindices({(int)m_mesh.indices.size() / 3, 3});
CUDA_CHECK_THROW(cudaMemcpy(cpuverts.request().ptr, m_mesh.verts.data(), m_mesh.verts.size() * 3 * sizeof(float), cudaMemcpyDeviceToHost));
CUDA_CHECK_THROW(cudaMemcpy(
cpunormals.request().ptr, m_mesh.vert_normals.data(), m_mesh.vert_normals.size() * 3 * sizeof(float), cudaMemcpyDeviceToHost
));
CUDA_CHECK_THROW(
cudaMemcpy(cpucolors.request().ptr, m_mesh.vert_colors.data(), m_mesh.vert_colors.size() * 3 * sizeof(float), cudaMemcpyDeviceToHost)
);
CUDA_CHECK_THROW(cudaMemcpy(cpuindices.request().ptr, m_mesh.indices.data(), m_mesh.indices.size() * sizeof(int), cudaMemcpyDeviceToHost));
vec3* ns = (vec3*)cpunormals.request().ptr;
for (size_t i = 0; i < m_mesh.vert_normals.size(); ++i) {
ns[i] = normalize(ns[i]);
}
return py::dict(py::arg("V") = cpuverts, py::arg("N") = cpunormals, py::arg("C") = cpucolors, py::arg("F") = cpuindices);
}
// Returns RGBA and depth buffers
std::pair<py::array_t<float>, py::array_t<float>>
Testbed::render_to_cpu(int width, int height, int spp, bool linear, float start_time, float end_time, float fps, float shutter_fraction) {
m_windowless_render_surface.resize({width, height});
m_windowless_render_surface.reset_accumulation();
if (end_time < 0.f) {
end_time = start_time;
}
bool path_animation_enabled = start_time >= 0.f;
if (!path_animation_enabled) { // the old code disabled camera smoothing for non-path renders; so we preserve that behaviour
m_smoothed_camera = m_camera;
}
// this rendering code assumes that the intra-frame camera motion starts from m_smoothed_camera (ie where we left off) to allow for EMA
// camera smoothing. in the case of a camera path animation, at the very start of the animation, we have yet to initialize
// smoothed_camera to something sensible
// - it will just be the default boot position. oops!
// that led to the first frame having a crazy streak from the default camera position to the start of the path.
// so we detect that case and explicitly force the current matrix to the start of the path
if (start_time == 0.f) {
set_camera_from_time(start_time);
m_smoothed_camera = m_camera;
}
auto start_cam_matrix = m_smoothed_camera;
// now set up the end-of-frame camera matrix if we are moving along a path
if (path_animation_enabled) {
set_camera_from_time(end_time);
apply_camera_smoothing(1000.f / fps);
}
auto end_cam_matrix = m_smoothed_camera;
auto prev_camera_matrix = m_smoothed_camera;
for (int i = 0; i < spp; ++i) {
float start_alpha = ((float)i) / (float)spp * shutter_fraction;
float end_alpha = ((float)i + 1.0f) / (float)spp * shutter_fraction;
auto sample_start_cam_matrix = start_cam_matrix;
auto sample_end_cam_matrix = camera_log_lerp(start_cam_matrix, end_cam_matrix, shutter_fraction);
if (i == 0) {
prev_camera_matrix = sample_start_cam_matrix;
}
if (path_animation_enabled) {
set_camera_from_time(start_time + (end_time - start_time) * (start_alpha + end_alpha) / 2.0f);
m_smoothed_camera = m_camera;
}
if (m_autofocus) {
autofocus();
}
render_frame(
m_stream.get(),
sample_start_cam_matrix,
sample_end_cam_matrix,
prev_camera_matrix,
m_screen_center,
m_relative_focal_length,
{0.0f, 0.0f, 0.0f, 1.0f},
{}, // foveation
{}, // prev foveation
{}, // lens
m_visualized_dimension,
m_windowless_render_surface,
!linear
);
prev_camera_matrix = sample_start_cam_matrix;
}
// For cam smoothing when rendering the next frame.
m_smoothed_camera = end_cam_matrix;
py::array_t<float> result_rgba({height, width, 4});
py::buffer_info buf_rgba = result_rgba.request();
py::array_t<float> result_depth({height, width});
py::buffer_info buf_depth = result_depth.request();
CUDA_CHECK_THROW(cudaMemcpy2DFromArray(
buf_rgba.ptr, width * sizeof(float) * 4, m_windowless_render_surface.surface_provider().array(), 0, 0, width * sizeof(float) * 4, height, cudaMemcpyDeviceToHost
));
CUDA_CHECK_THROW(
cudaMemcpy(buf_depth.ptr, m_windowless_render_surface.depth_buffer(), height * width * sizeof(float), cudaMemcpyDeviceToHost)
);
return {result_rgba, result_depth};
}
py::array_t<float> Testbed::render_to_cpu_rgba(
int width, int height, int spp, bool linear, float start_time, float end_time, float fps, float shutter_fraction
) {
return render_to_cpu(width, height, spp, linear, start_time, end_time, fps, shutter_fraction).first;
}
py::array_t<float> Testbed::view(bool linear, size_t view_idx) const {
if (m_views.size() <= view_idx) {
throw std::runtime_error{fmt::format("View #{} does not exist.", view_idx)};
}
auto& view = m_views.at(view_idx);
auto& render_buffer = *view.render_buffer;
auto res = render_buffer.out_resolution();
py::array_t<float> result({res.y, res.x, 4});
py::buffer_info buf = result.request();
float* data = (float*)buf.ptr;
CUDA_CHECK_THROW(cudaMemcpy2DFromArray(
data, res.x * sizeof(float) * 4, render_buffer.surface_provider().array(), 0, 0, res.x * sizeof(float) * 4, res.y, cudaMemcpyDeviceToHost
));
if (linear) {
ThreadPool{}.parallel_for<size_t>(0, res.y, [&](size_t y) {
size_t base = y * res.x;
for (uint32_t x = 0; x < res.x; ++x) {
size_t px = base + x;
data[px * 4 + 0] = srgb_to_linear(data[px * 4 + 0]);
data[px * 4 + 1] = srgb_to_linear(data[px * 4 + 1]);
data[px * 4 + 2] = srgb_to_linear(data[px * 4 + 2]);
}
});
}
return result;
}
#ifdef NGP_GUI
py::array_t<float> Testbed::screenshot(bool linear, bool front_buffer) const {
std::vector<float> tmp(product(m_window_res) * 4);
glReadBuffer(front_buffer ? GL_FRONT : GL_BACK);
glReadPixels(0, 0, m_window_res.x, m_window_res.y, GL_RGBA, GL_FLOAT, tmp.data());
py::array_t<float> result({m_window_res.y, m_window_res.x, 4});
py::buffer_info buf = result.request();
float* data = (float*)buf.ptr;
// Linear, alpha premultiplied, Y flipped
ThreadPool{}.parallel_for<size_t>(0, m_window_res.y, [&](size_t y) {
size_t base = y * m_window_res.x;
size_t base_reverse = (m_window_res.y - y - 1) * m_window_res.x;
for (uint32_t x = 0; x < m_window_res.x; ++x) {
size_t px = base + x;
size_t px_reverse = base_reverse + x;
data[px_reverse * 4 + 0] = linear ? srgb_to_linear(tmp[px * 4 + 0]) : tmp[px * 4 + 0];
data[px_reverse * 4 + 1] = linear ? srgb_to_linear(tmp[px * 4 + 1]) : tmp[px * 4 + 1];
data[px_reverse * 4 + 2] = linear ? srgb_to_linear(tmp[px * 4 + 2]) : tmp[px * 4 + 2];
data[px_reverse * 4 + 3] = tmp[px * 4 + 3];
}
});
return result;
}
#endif
PYBIND11_MODULE(pyngp, m) {
m.doc() = "Instant neural graphics primitives";
m.def("free_temporary_memory", &free_all_gpu_memory_arenas);
py::enum_<ETestbedMode>(m, "TestbedMode")
.value("Nerf", ETestbedMode::Nerf)
.value("Sdf", ETestbedMode::Sdf)
.value("Image", ETestbedMode::Image)
.value("Volume", ETestbedMode::Volume)
.value("None", ETestbedMode::None)
.export_values();
m.def("mode_from_scene", &mode_from_scene);
m.def("mode_from_string", &mode_from_string);
py::enum_<EGroundTruthRenderMode>(m, "GroundTruthRenderMode")
.value("Shade", EGroundTruthRenderMode::Shade)
.value("Depth", EGroundTruthRenderMode::Depth)
.export_values();
py::enum_<ERenderMode>(m, "RenderMode")
.value("AO", ERenderMode::AO)
.value("Shade", ERenderMode::Shade)
.value("Normals", ERenderMode::Normals)
.value("Positions", ERenderMode::Positions)
.value("Depth", ERenderMode::Depth)
.value("Distortion", ERenderMode::Distortion)
.value("Cost", ERenderMode::Cost)
.value("Slice", ERenderMode::Slice)
.export_values();
py::enum_<ERandomMode>(m, "RandomMode")
.value("Random", ERandomMode::Random)
.value("Halton", ERandomMode::Halton)
.value("Sobol", ERandomMode::Sobol)
.value("Stratified", ERandomMode::Stratified)
.export_values();
py::enum_<ELossType>(m, "LossType")
.value("L2", ELossType::L2)
.value("L1", ELossType::L1)
.value("Mape", ELossType::Mape)
.value("Smape", ELossType::Smape)
.value("Huber", ELossType::Huber)
// Legacy: we used to refer to the Huber loss
// (L2 near zero, L1 further away) as "SmoothL1".
.value("SmoothL1", ELossType::Huber)
.value("LogL1", ELossType::LogL1)
.value("RelativeL2", ELossType::RelativeL2)
.export_values();
py::enum_<ESDFGroundTruthMode>(m, "SDFGroundTruthMode")
.value("RaytracedMesh", ESDFGroundTruthMode::RaytracedMesh)
.value("SpheretracedMesh", ESDFGroundTruthMode::SpheretracedMesh)
.export_values();
py::enum_<ENerfActivation>(m, "NerfActivation")
.value("None", ENerfActivation::None)
.value("ReLU", ENerfActivation::ReLU)
.value("Logistic", ENerfActivation::Logistic)
.value("Exponential", ENerfActivation::Exponential)
.export_values();
py::enum_<EMeshSdfMode>(m, "MeshSdfMode")
.value("Watertight", EMeshSdfMode::Watertight)
.value("Raystab", EMeshSdfMode::Raystab)
.value("PathEscape", EMeshSdfMode::PathEscape)
.export_values();
py::enum_<EColorSpace>(m, "ColorSpace").value("Linear", EColorSpace::Linear).value("SRGB", EColorSpace::SRGB).export_values();
py::enum_<ETonemapCurve>(m, "TonemapCurve")
.value("Identity", ETonemapCurve::Identity)
.value("ACES", ETonemapCurve::ACES)
.value("Hable", ETonemapCurve::Hable)
.value("Reinhard", ETonemapCurve::Reinhard)
.export_values();
py::enum_<ELensMode>(m, "LensMode")
.value("Perspective", ELensMode::Perspective)
.value("OpenCV", ELensMode::OpenCV)
.value("FTheta", ELensMode::FTheta)
.value("LatLong", ELensMode::LatLong)
.value("OpenCVFisheye", ELensMode::OpenCVFisheye)
.value("Equirectangular", ELensMode::Equirectangular)
.value("Orthographic", ELensMode::Orthographic)
.export_values();
py::enum_<ECameraPredictionMode>(m, "CameraPredictionMode")
.value("None", ECameraPredictionMode::None)
.value("MatLogLinear", ECameraPredictionMode::MatLogLinear)
.value("MatLogQuadratic", ECameraPredictionMode::MatLogQuadratic)
.value("Se3LogLinear", ECameraPredictionMode::Se3LogLinear)
.value("Se3LogQuadratic", ECameraPredictionMode::Se3LogQuadratic)
.export_values();
py::class_<BoundingBox>(m, "BoundingBox")
.def(py::init<>())
.def(py::init<const vec3&, const vec3&>())
.def("center", &BoundingBox::center)
.def("contains", &BoundingBox::contains)
.def("diag", &BoundingBox::diag)
.def("distance", &BoundingBox::distance)
.def("distance_sq", &BoundingBox::distance_sq)
.def("enlarge", py::overload_cast<const vec3&>(&BoundingBox::enlarge))
.def("enlarge", py::overload_cast<const BoundingBox&>(&BoundingBox::enlarge))
.def("get_vertices", &BoundingBox::get_vertices)
.def("inflate", &BoundingBox::inflate)
.def("intersection", &BoundingBox::intersection)
.def("intersects", py::overload_cast<const BoundingBox&>(&BoundingBox::intersects, py::const_))
.def("ray_intersect", &BoundingBox::ray_intersect)
.def("relative_pos", &BoundingBox::relative_pos)
.def("signed_distance", &BoundingBox::signed_distance)
.def_readwrite("min", &BoundingBox::min)
.def_readwrite("max", &BoundingBox::max);
py::class_<Lens> lens(m, "Lens");
lens.def(py::init<>()).def_readwrite("mode", &Lens::mode).def_property_readonly("params", [](py::object& obj) {
Lens& o = obj.cast<Lens&>();
return py::array{sizeof(o.params) / sizeof(o.params[0]), o.params, obj};
});
py::class_<fs::path>(m, "path").def(py::init<>()).def(py::init<const std::string&>());
py::implicitly_convertible<std::string, fs::path>();
py::class_<Testbed> testbed(m, "Testbed");
testbed.def(py::init<ETestbedMode>(), py::arg("mode") = ETestbedMode::None)
.def(py::init<ETestbedMode, const fs::path&, const fs::path&>())
.def(py::init<ETestbedMode, const fs::path&, const json&>())
.def_readonly("mode", &Testbed::m_testbed_mode)
.def(
"create_empty_nerf_dataset",
&Testbed::create_empty_nerf_dataset,
"Allocate memory for a nerf dataset with a given size",
py::arg("n_images"),
py::arg("aabb_scale") = 1,
py::arg("is_hdr") = false
)
.def("load_training_data", &Testbed::load_training_data, py::call_guard<py::gil_scoped_release>(), "Load training data from a given path.")
.def("clear_training_data", &Testbed::clear_training_data, "Clears training data to free up GPU memory.")
// General control
.def(
"init_window",
&Testbed::init_window,
"Init a GLFW window that shows real-time progress and a GUI. 'second_window' creates a second copy of the output in its own window.",
py::arg("width"),
py::arg("height"),
py::arg("hidden") = false,
py::arg("second_window") = false
)
.def("destroy_window", &Testbed::destroy_window, "Destroy the window again.")
.def(
"init_vr",
&Testbed::init_vr,
"Init rendering to a connected and active VR headset. Requires a window to have been previously created via `init_window`."
)
.def(
"view",
&Testbed::view,
"Outputs the currently displayed image by a given view (0 by default).",
py::arg("linear") = true,
py::arg("view") = 0
)
.def("view_camera", &Testbed::view_camera, "Outputs the current camera matrix of a given view (0 by default).", py::arg("view") = 0)
.def(
"predict_view_camera",
&Testbed::predict_view_camera,
"Outputs the predicted camera matrix of a given view after a given elapsed time (in seconds)."
)
.def_property("camera_prediction_mode", &Testbed::camera_prediction_mode, &Testbed::set_camera_prediction_mode)
#ifdef NGP_GUI
.def_readwrite("keyboard_event_callback", &Testbed::m_keyboard_event_callback)
.def_readwrite("file_drop_callback", &Testbed::m_file_drop_callback)
.def("is_key_pressed", [](py::object& obj, int key) { return ImGui::IsKeyPressed(key); })
.def("is_key_down", [](py::object& obj, int key) { return ImGui::IsKeyDown(key); })
.def("is_alt_down", [](py::object& obj) { return ImGui::GetIO().KeyMods & ImGuiKeyModFlags_Alt; })
.def("is_ctrl_down", [](py::object& obj) { return ImGui::GetIO().KeyMods & ImGuiKeyModFlags_Ctrl; })
.def("is_shift_down", [](py::object& obj) { return ImGui::GetIO().KeyMods & ImGuiKeyModFlags_Shift; })
.def("is_super_down", [](py::object& obj) { return ImGui::GetIO().KeyMods & ImGuiKeyModFlags_Super; })
.def(
"screenshot",
&Testbed::screenshot,
"Takes a screenshot of the current window contents.",
py::arg("linear") = true,
py::arg("front_buffer") = true
)
.def_readwrite("vr_use_hidden_area_mask", &Testbed::m_vr_use_hidden_area_mask)
.def_readwrite("vr_use_depth_reproject", &Testbed::m_vr_use_depth_reproject)
#endif
.def("want_repl", &Testbed::want_repl, "returns true if the user clicked the 'I want a repl' button")
.def(
"frame", &Testbed::frame, py::call_guard<py::gil_scoped_release>(), "Process a single frame. Renders if a window was previously created."
)
.def(
"render",
&Testbed::render_to_cpu_rgba,
"Renders an image at the requested resolution. Does not require a window.",
py::arg("width") = 1920,
py::arg("height") = 1080,
py::arg("spp") = 1,
py::arg("linear") = true,
py::arg("start_t") = -1.f,
py::arg("end_t") = -1.f,
py::arg("fps") = 30.f,
py::arg("shutter_fraction") = 1.0f
)
.def(
"render_with_depth",
&Testbed::render_to_cpu,
"Renders an image at the requested resolution. Does not require a window.",
py::arg("width") = 1920,
py::arg("height") = 1080,
py::arg("spp") = 1,
py::arg("linear") = true,
py::arg("start_t") = -1.f,
py::arg("end_t") = -1.f,
py::arg("fps") = 30.f,
py::arg("shutter_fraction") = 1.0f
)
.def("train", &Testbed::train, py::call_guard<py::gil_scoped_release>(), "Perform a single training step with a specified batch size.")
.def("reset", &Testbed::reset_network, py::arg("reset_density_grid") = true, "Reset training.")
.def("reset_camera", &Testbed::reset_camera, "Reset camera to default state.")
.def(
"reset_accumulation",
&Testbed::reset_accumulation,
"Reset rendering accumulation.",
py::arg("due_to_camera_movement") = false,
py::arg("immediate_redraw") = true
)
.def("reload_network_from_file", &Testbed::reload_network_from_file, py::arg("path") = "", "Reload the network from a config file.")
.def(
"reload_network_from_json",
&Testbed::reload_network_from_json,
"Reload the network from a json object.",
py::arg("json"),
py::arg("config_base_path") = ""
)
.def("override_sdf_training_data", &Testbed::override_sdf_training_data, "Override the training data for learning a signed distance function")
.def(
"calculate_iou",
&Testbed::calculate_iou,
"Calculate the intersection over union error value",
py::arg("n_samples") = 128 * 1024 * 1024,
py::arg("scale_existing_results_factor") = 0.0f,
py::arg("blocking") = true,
py::arg("force_use_octree") = true
)
.def("n_params", &Testbed::n_params, "Number of trainable parameters")
.def("n_encoding_params", &Testbed::n_encoding_params, "Number of trainable parameters in the encoding")
.def(
"save_snapshot",
&Testbed::save_snapshot,
py::arg("path"),
py::arg("include_optimizer_state") = false,
py::arg("compress") = true,
"Save a snapshot of the currently trained model. Optionally compressed (only when saving '.ingp' files)."
)
.def("load_snapshot", py::overload_cast<const fs::path&>(&Testbed::load_snapshot), py::arg("path"), "Load a previously saved snapshot")
.def("load_camera_path", &Testbed::load_camera_path, py::arg("path"), "Load a camera path")
.def(
"load_file",
&Testbed::load_file,
py::arg("path"),
"Load a file and automatically determine how to handle it. Can be a snapshot, dataset, network config, or camera path."
)
.def_property("loop_animation", &Testbed::loop_animation, &Testbed::set_loop_animation)
.def_property("jit_fusion", &Testbed::jit_fusion, &Testbed::set_jit_fusion)
.def(
"compute_and_save_png_slices",
&Testbed::compute_and_save_png_slices,
py::arg("filename"),
py::arg("resolution") = 256,
py::arg("aabb") = BoundingBox{},
py::arg("thresh") = std::numeric_limits<float>::max(),
py::arg("density_range") = 4.f,
py::arg("flip_y_and_z_axes") = false,
"Compute & save a PNG file representing the 3D density or distance field from the current SDF or NeRF model. "
)
.def(
"compute_and_save_marching_cubes_mesh",
&Testbed::compute_and_save_marching_cubes_mesh,
py::arg("filename"),
py::arg("resolution") = ivec3(256),
py::arg("aabb") = BoundingBox{},
py::arg("thresh") = std::numeric_limits<float>::max(),
py::arg("generate_uvs_for_obj_file") = false,
"Compute & save a marching cubes mesh from the current SDF or NeRF model. "
"Supports OBJ and PLY format. Note that UVs are only supported by OBJ files. "
"`thresh` is the density threshold; use 0 for SDF; 2.5 works well for NeRF. "
"If the aabb parameter specifies an inside-out (\"empty\") box (default), the current render_aabb bounding box is used."
)
.def(
"compute_marching_cubes_mesh",
&Testbed::compute_marching_cubes_mesh,
py::arg("resolution") = ivec3(256),
py::arg("aabb") = BoundingBox{},
py::arg("thresh") = std::numeric_limits<float>::max(),
"Compute a marching cubes mesh from the current SDF or NeRF model. "
"Returns a python dict with numpy arrays V (vertices), N (vertex normals), C (vertex colors), and F (triangular faces). "
"`thresh` is the density threshold; use 0 for SDF; 2.5 works well for NeRF. "
"If the aabb parameter specifies an inside-out (\"empty\") box (default), the current render_aabb bounding box is used."
)
// Interesting members.
.def_readwrite("dynamic_res", &Testbed::m_dynamic_res)
.def_readwrite("dynamic_res_target_fps", &Testbed::m_dynamic_res_target_fps)
.def_readwrite("fixed_res_factor", &Testbed::m_fixed_res_factor)
.def_readwrite("background_color", &Testbed::m_background_color)
.def_readwrite("render_transparency_as_checkerboard", &Testbed::m_render_transparency_as_checkerboard)
.def_readwrite("shall_train", &Testbed::m_train)
.def_readwrite("shall_train_encoding", &Testbed::m_train_encoding)
.def_readwrite("shall_train_network", &Testbed::m_train_network)
.def_readwrite("render_groundtruth", &Testbed::m_render_ground_truth)
.def_readwrite("render_ground_truth", &Testbed::m_render_ground_truth)
.def_readwrite("groundtruth_render_mode", &Testbed::m_ground_truth_render_mode)
.def_readwrite("render_mode", &Testbed::m_render_mode)
.def_readwrite("render_near_distance", &Testbed::m_render_near_distance)
.def_readwrite("slice_plane_z", &Testbed::m_slice_plane_z)
.def_readwrite("dof", &Testbed::m_aperture_size)
.def_readwrite("aperture_size", &Testbed::m_aperture_size)
.def_readwrite("autofocus", &Testbed::m_autofocus)
.def_readwrite("autofocus_target", &Testbed::m_autofocus_target)
.def_readwrite("camera_path", &Testbed::m_camera_path)
.def_readwrite("record_camera_path", &Testbed::m_record_camera_path)
.def_readwrite("floor_enable", &Testbed::m_floor_enable)
.def_readwrite("exposure", &Testbed::m_exposure)
.def_property("scale", &Testbed::scale, &Testbed::set_scale)
.def_readonly("bounding_radius", &Testbed::m_bounding_radius)
.def_readwrite("render_aabb", &Testbed::m_render_aabb)
.def_readwrite("render_aabb_to_local", &Testbed::m_render_aabb_to_local)
.def_readwrite("is_rendering", &Testbed::m_render)
.def_readwrite("aabb", &Testbed::m_aabb)
.def_readwrite("raw_aabb", &Testbed::m_raw_aabb)
.def_property("fov", &Testbed::fov, &Testbed::set_fov)
.def_property("fov_xy", &Testbed::fov_xy, &Testbed::set_fov_xy)
.def_readwrite("fov_axis", &Testbed::m_fov_axis)
.def_readwrite("relative_focal_length", &Testbed::m_relative_focal_length)
.def_readwrite("zoom", &Testbed::m_zoom)
.def_readwrite("screen_center", &Testbed::m_screen_center)
.def_readwrite("training_batch_size", &Testbed::m_training_batch_size)
.def("set_nerf_camera_matrix", &Testbed::set_nerf_camera_matrix)
.def("set_camera_to_training_view", &Testbed::set_camera_to_training_view)
.def("first_training_view", &Testbed::first_training_view)
.def("last_training_view", &Testbed::last_training_view)
.def("previous_training_view", &Testbed::previous_training_view)
.def("next_training_view", &Testbed::next_training_view)
.def("compute_image_mse", &Testbed::compute_image_mse, py::arg("quantize") = false)
.def_readwrite("camera_matrix", &Testbed::m_camera)
.def_readwrite("up_dir", &Testbed::m_up_dir)
.def_readwrite("sun_dir", &Testbed::m_sun_dir)
.def_readwrite("default_camera", &Testbed::m_default_camera)
.def_property("look_at", &Testbed::look_at, &Testbed::set_look_at)
.def_property("view_dir", &Testbed::view_dir, &Testbed::set_view_dir)
.def_readwrite("max_level_rand_training", &Testbed::m_max_level_rand_training)
.def_readwrite("visualized_dimension", &Testbed::m_visualized_dimension)
.def_readwrite("visualized_layer", &Testbed::m_visualized_layer)
.def_property_readonly("loss", [](py::object& obj) { return obj.cast<Testbed&>().m_loss_scalar.val(); })
.def_readonly("training_step", &Testbed::m_training_step)
.def_readonly("nerf", &Testbed::m_nerf)
.def_readonly("sdf", &Testbed::m_sdf)
.def_readonly("image", &Testbed::m_image)
.def_readwrite("camera_smoothing", &Testbed::m_camera_smoothing)
.def_readwrite("render_with_lens_distortion", &Testbed::m_render_with_lens_distortion)
.def_readwrite("render_lens", &Testbed::m_render_lens)
.def_property(
"display_gui",
[](py::object& obj) { return obj.cast<Testbed&>().m_imgui.show; },
[](const py::object& obj, bool value) { obj.cast<Testbed&>().m_imgui.show = value; }
)
.def_readwrite("visualize_unit_cube", &Testbed::m_visualize_unit_cube)
.def_readwrite("snap_to_pixel_centers", &Testbed::m_snap_to_pixel_centers)
.def_readwrite("parallax_shift", &Testbed::m_parallax_shift)
.def_readwrite("color_space", &Testbed::m_color_space)
.def_readwrite("tonemap_curve", &Testbed::m_tonemap_curve)
.def_property(
"dlss",
[](py::object& obj) { return obj.cast<Testbed&>().m_dlss; },
[](const py::object& obj, bool value) {
if (value && !obj.cast<Testbed&>().m_dlss_provider) {
if (obj.cast<Testbed&>().m_render_window) {
throw std::runtime_error{"DLSS not supported."};
} else {
throw std::runtime_error{"DLSS requires a Window to be initialized via `init_window`."};
}
}
obj.cast<Testbed&>().m_dlss = value;
}
)
.def_readwrite("dlss_sharpening", &Testbed::m_dlss_sharpening)
.def("crop_box", &Testbed::crop_box, py::arg("nerf_space") = true)
.def("set_crop_box", &Testbed::set_crop_box, py::arg("matrix"), py::arg("nerf_space") = true)
.def("crop_box_corners", &Testbed::crop_box_corners, py::arg("nerf_space") = true)
.def_property(
"root_dir",
[](py::object& obj) { return obj.cast<Testbed&>().root_dir().str(); },
[](const py::object& obj, const std::string& value) { obj.cast<Testbed&>().set_root_dir(value); }
)
;
py::class_<Testbed::Nerf> nerf(testbed, "Nerf");
nerf.def_readonly("training", &Testbed::Nerf::training)
.def_readwrite("rgb_activation", &Testbed::Nerf::rgb_activation)
.def_readwrite("density_activation", &Testbed::Nerf::density_activation)
.def_readwrite("sharpen", &Testbed::Nerf::sharpen)
.def_readwrite("rendering_min_transmittance", &Testbed::Nerf::render_min_transmittance)
.def_readwrite("render_min_transmittance", &Testbed::Nerf::render_min_transmittance)
.def_readwrite("cone_angle_constant", &Testbed::Nerf::cone_angle_constant)
.def_readwrite("visualize_cameras", &Testbed::Nerf::visualize_cameras)
.def_readwrite("render_gbuffer_hard_edges", &Testbed::Nerf::render_gbuffer_hard_edges)
.def_readwrite(
"rendering_extra_dims_from_training_view",
&Testbed::Nerf::rendering_extra_dims_from_training_view,
"If non-negative, indicates the training view from which the extra dims are used. If -1, uses the values previously set by `set_rendering_extra_dims`."
)
.def(
"find_closest_training_view",
&Testbed::Nerf::find_closest_training_view,
"Obtain the training view that is closest to the current camera."
)
.def(
"set_rendering_extra_dims_from_training_view",
&Testbed::Nerf::set_rendering_extra_dims_from_training_view,
"Set the extra dims that are used for rendering to those that were trained for a given training view."
)
.def("set_rendering_extra_dims", &Testbed::Nerf::set_rendering_extra_dims, "Set the extra dims that are used for rendering.")
.def(
"get_rendering_extra_dims", &Testbed::Nerf::get_rendering_extra_dims_cpu, "Get the extra dims that are currently used for rendering."
);
py::class_<BRDFParams> brdfparams(m, "BRDFParams");
brdfparams.def_readwrite("metallic", &BRDFParams::metallic)
.def_readwrite("subsurface", &BRDFParams::subsurface)
.def_readwrite("specular", &BRDFParams::specular)
.def_readwrite("roughness", &BRDFParams::roughness)
.def_readwrite("sheen", &BRDFParams::sheen)
.def_readwrite("clearcoat", &BRDFParams::clearcoat)
.def_readwrite("clearcoat_gloss", &BRDFParams::clearcoat_gloss)
.def_readwrite("basecolor", &BRDFParams::basecolor)
.def_readwrite("ambientcolor", &BRDFParams::ambientcolor);
py::class_<TrainingImageMetadata> metadata(m, "TrainingImageMetadata");
metadata
// Legacy member: lens used to be called "camera_distortion"
.def_readwrite("camera_distortion", &TrainingImageMetadata::lens)
.def_readwrite("lens", &TrainingImageMetadata::lens)
.def_readwrite("resolution", &TrainingImageMetadata::resolution)
.def_readwrite("principal_point", &TrainingImageMetadata::principal_point)
.def_readwrite("focal_length", &TrainingImageMetadata::focal_length)
.def_readwrite("rolling_shutter", &TrainingImageMetadata::rolling_shutter)
.def_readwrite("light_dir", &TrainingImageMetadata::light_dir);
py::class_<NerfDataset> nerfdataset(m, "NerfDataset");
nerfdataset.def_readonly("metadata", &NerfDataset::metadata)
.def_readonly("transforms", &NerfDataset::xforms)
.def_readonly("paths", &NerfDataset::paths)
.def_readonly("render_aabb", &NerfDataset::render_aabb)
.def_readonly("render_aabb_to_local", &NerfDataset::render_aabb_to_local)
.def_readonly("up", &NerfDataset::up)
.def_readonly("offset", &NerfDataset::offset)
.def_readonly("n_images", &NerfDataset::n_images)
.def_readonly("envmap_resolution", &NerfDataset::envmap_resolution)
.def_readonly("scale", &NerfDataset::scale)
.def_readonly("aabb_scale", &NerfDataset::aabb_scale)
.def_readonly("from_mitsuba", &NerfDataset::from_mitsuba)
.def_readonly("is_hdr", &NerfDataset::is_hdr);
py::class_<Testbed::Nerf::Training>(nerf, "Training")
.def_readwrite("random_bg_color", &Testbed::Nerf::Training::random_bg_color)
.def_readwrite("n_images_for_training", &Testbed::Nerf::Training::n_images_for_training)
.def_readwrite("linear_colors", &Testbed::Nerf::Training::linear_colors)
.def_readwrite("loss_type", &Testbed::Nerf::Training::loss_type)
.def_readwrite("depth_loss_type", &Testbed::Nerf::Training::depth_loss_type)
.def_readwrite("snap_to_pixel_centers", &Testbed::Nerf::Training::snap_to_pixel_centers)
.def_readwrite("optimize_extrinsics", &Testbed::Nerf::Training::optimize_extrinsics)
.def_readwrite("optimize_per_image_latents", &Testbed::Nerf::Training::optimize_extra_dims)
.def_readwrite("optimize_extra_dims", &Testbed::Nerf::Training::optimize_extra_dims)
.def_readwrite("optimize_exposure", &Testbed::Nerf::Training::optimize_exposure)
.def_readwrite("optimize_distortion", &Testbed::Nerf::Training::optimize_distortion)
.def_readwrite("optimize_focal_length", &Testbed::Nerf::Training::optimize_focal_length)
.def_readwrite("n_steps_between_cam_updates", &Testbed::Nerf::Training::n_steps_between_cam_updates)
.def_readwrite("sample_focal_plane_proportional_to_error", &Testbed::Nerf::Training::sample_focal_plane_proportional_to_error)
.def_readwrite("sample_image_proportional_to_error", &Testbed::Nerf::Training::sample_image_proportional_to_error)
.def_readwrite("include_sharpness_in_error", &Testbed::Nerf::Training::include_sharpness_in_error)
.def_readonly("transforms", &Testbed::Nerf::Training::transforms)
//.def_readonly("focal_lengths", &Testbed::Nerf::Training::focal_lengths) // use training.dataset.metadata instead
.def_readwrite("near_distance", &Testbed::Nerf::Training::near_distance)
.def_readwrite("density_grid_decay", &Testbed::Nerf::Training::density_grid_decay)
.def_readwrite("extrinsic_l2_reg", &Testbed::Nerf::Training::extrinsic_l2_reg)
.def_readwrite("extrinsic_learning_rate", &Testbed::Nerf::Training::extrinsic_learning_rate)
.def_readwrite("intrinsic_l2_reg", &Testbed::Nerf::Training::intrinsic_l2_reg)
.def_readwrite("exposure_l2_reg", &Testbed::Nerf::Training::exposure_l2_reg)
.def_readwrite("depth_supervision_lambda", &Testbed::Nerf::Training::depth_supervision_lambda)
.def_readonly("dataset", &Testbed::Nerf::Training::dataset)
.def(
"get_extra_dims",
&Testbed::Nerf::Training::get_extra_dims_cpu,
"Get the extra dims (including trained latent code) for a specified training view."
)
.def(
"set_camera_intrinsics",
&Testbed::Nerf::Training::set_camera_intrinsics,
py::arg("frame_idx"),
py::arg("fx") = 0.f,
py::arg("fy") = 0.f,
py::arg("cx") = -0.5f,
py::arg("cy") = -0.5f,
py::arg("k1") = 0.f,
py::arg("k2") = 0.f,
py::arg("p1") = 0.f,
py::arg("p2") = 0.f,
py::arg("k3") = 0.f,
py::arg("k4") = 0.f,
py::arg("is_fisheye") = false,
"Set up the camera intrinsics for the given training image index."
)
.def(
"set_camera_extrinsics",
&Testbed::Nerf::Training::set_camera_extrinsics,
py::arg("frame_idx"),
py::arg("camera_to_world"),
py::arg("convert_to_ngp") = true,
"Set up the camera extrinsics for the given training image index, from the given 3x4 transformation matrix."
)
.def(
"get_camera_extrinsics",
&Testbed::Nerf::Training::get_camera_extrinsics,
py::arg("frame_idx"),
"return the 3x4 transformation matrix of given training frame"
)
.def(
"set_image",
&Testbed::Nerf::Training::set_image,
py::arg("frame_idx"),
py::arg("img"),
py::arg("depth_img"),
py::arg("depth_scale") = 1.0f,
"set one of the training images. must be a floating point numpy array of (H,W,C) with 4 channels; linear color space; W and H must match image size of the rest of the dataset"
);
py::class_<Testbed::Sdf> sdf(testbed, "Sdf");
sdf.def_readonly("training", &Testbed::Sdf::training)
.def_readwrite("mesh_sdf_mode", &Testbed::Sdf::mesh_sdf_mode)
.def_readwrite("mesh_scale", &Testbed::Sdf::mesh_scale)
.def_readwrite("analytic_normals", &Testbed::Sdf::analytic_normals)
.def_readwrite("shadow_sharpness", &Testbed::Sdf::shadow_sharpness)
.def_readwrite("fd_normals_epsilon", &Testbed::Sdf::fd_normals_epsilon)
.def_readwrite("use_triangle_octree", &Testbed::Sdf::use_triangle_octree)
.def_readwrite("zero_offset", &Testbed::Sdf::zero_offset)
.def_readwrite("distance_scale", &Testbed::Sdf::distance_scale)
.def_readwrite("calculate_iou_online", &Testbed::Sdf::calculate_iou_online)
.def_readwrite("groundtruth_mode", &Testbed::Sdf::groundtruth_mode)
.def_readwrite("brdf", &Testbed::Sdf::brdf);
py::class_<Testbed::Sdf::Training>(sdf, "Training")
.def_readwrite("generate_sdf_data_online", &Testbed::Sdf::Training::generate_sdf_data_online)
.def_readwrite("surface_offset_scale", &Testbed::Sdf::Training::surface_offset_scale);
py::class_<Testbed::Image> image(testbed, "Image");
image.def_readonly("training", &Testbed::Image::training).def_readwrite("random_mode", &Testbed::Image::random_mode);
py::class_<Testbed::Image::Training>(image, "Training")
.def_readwrite("snap_to_pixel_centers", &Testbed::Image::Training::snap_to_pixel_centers)
.def_readwrite("linear_colors", &Testbed::Image::Training::linear_colors);
py::class_<CameraKeyframe>(m, "CameraKeyframe")
.def(py::init<>())
.def(
py::init<const quat&, const vec3&, float, float, float, float, float>(),
py::arg("r"),
py::arg("t"),
py::arg("slice"),
py::arg("scale"),
py::arg("fov"),
py::arg("aperture_size"),
py::arg("timestamp")
)
.def(
py::init<const mat4x3&, float, float, float, float, float>(),
py::arg("m"),
py::arg("slice"),
py::arg("scale"),
py::arg("fov"),
py::arg("aperture_size"),
py::arg("timestamp")
)
.def_readwrite("R", &CameraKeyframe::R)
.def_readwrite("T", &CameraKeyframe::T)
.def_readwrite("slice", &CameraKeyframe::slice)
.def_readwrite("scale", &CameraKeyframe::scale)
.def_readwrite("fov", &CameraKeyframe::fov)
.def_readwrite("aperture_size", &CameraKeyframe::aperture_size)
.def_readwrite("timestamp", &CameraKeyframe::timestamp)
.def("m", &CameraKeyframe::m)
.def("from_m", &CameraKeyframe::from_m, py::arg("rv"))
.def("same_pos_as", &CameraKeyframe::same_pos_as, py::arg("rhs"));
py::enum_<EEditingKernel>(m, "EditingKernel")
.value("None", EEditingKernel::None)
.value("Gaussian", EEditingKernel::Gaussian)
.value("Quartic", EEditingKernel::Quartic)
.value("Hat", EEditingKernel::Hat)
.value("Box", EEditingKernel::Box);
py::class_<CameraPath::RenderSettings>(m, "CameraPathRenderSettings")
.def_readwrite("resolution", &CameraPath::RenderSettings::resolution)
.def_readwrite("spp", &CameraPath::RenderSettings::spp)
.def_readwrite("fps", &CameraPath::RenderSettings::fps)
.def_readwrite("duration_seconds", &CameraPath::RenderSettings::duration_seconds)
.def_readwrite("shutter_fraction", &CameraPath::RenderSettings::shutter_fraction)
.def_readwrite("quality", &CameraPath::RenderSettings::quality);
py::class_<CameraPath::Pos>(m, "CameraPathPos").def_readwrite("kfidx", &CameraPath::Pos::kfidx).def_readwrite("t", &CameraPath::Pos::t);
py::class_<CameraPath>(m, "CameraPath")
.def_readwrite("keyframes", &CameraPath::keyframes)
.def_readwrite("update_cam_from_path", &CameraPath::update_cam_from_path)
.def_readwrite("play_time", &CameraPath::play_time)
.def_readwrite("auto_play_speed", &CameraPath::auto_play_speed)
.def_readwrite("loop", &CameraPath::loop)
.def_readwrite("keyframe_subsampling", &CameraPath::keyframe_subsampling)
.def_readwrite("duration_seconds", &CameraPath::duration_seconds)
.def_readwrite("camera_prediction_duration_seconds", &CameraPath::camera_prediction_duration_seconds)
.def_readwrite("camera_prediction_dt", &CameraPath::camera_prediction_dt)
.def_readwrite("editing_kernel_type", &CameraPath::editing_kernel_type)
.def_readwrite("editing_kernel_radius", &CameraPath::editing_kernel_radius)
.def_readwrite("spline_order", &CameraPath::spline_order)
.def_readwrite("render_settings", &CameraPath::render_settings)
.def_readwrite("rendering", &CameraPath::rendering)
.def_readwrite("render_frame_idx", &CameraPath::render_frame_idx)
.def_readwrite("render_start_time", &CameraPath::render_start_time)
.def_readwrite("render_frame_end_camera", &CameraPath::render_frame_end_camera)
.def("clear", &CameraPath::clear)
.def("has_valid_timestamps", &CameraPath::has_valid_timestamps)
.def("make_keyframe_timestamps_equidistant", &CameraPath::make_keyframe_timestamps_equidistant)
.def("sanitize_keyframes", &CameraPath::sanitize_keyframes)
.def("get_pos", &CameraPath::get_pos, py::arg("playtime"))
.def("get_playtime", &CameraPath::get_playtime, py::arg("i"))
.def("get_keyframe", &CameraPath::get_keyframe, py::arg("i"))
.def("eval_camera_path", &CameraPath::eval_camera_path, py::arg("t"))
.def("save", &CameraPath::save, py::arg("path"))
.def("load", &CameraPath::load, py::arg("path"), py::arg("first_xform"))
.def(
"add_camera",
&CameraPath::add_camera,
py::arg("camera"),
py::arg("slice_plane_z"),
py::arg("scale"),
py::arg("fov"),
py::arg("aperture_size"),
py::arg("bounding_radius"),
py::arg("timestamp")
);
}
} // namespace ngp

Computing file changes ...