gaussian-splatting/gaussian_renderer/__init__.py
2024-05-22 00:53:31 +08:00

118 lines
5.3 KiB
Python
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

#
# Copyright (C) 2023, Inria
# GRAPHDECO research group, https://team.inria.fr/graphdeco
# All rights reserved.
#
# This software is free for non-commercial, research and evaluation use
# under the terms of the LICENSE.md file.
#
# For inquiries contact george.drettakis@inria.fr
#
import torch
import math
from diff_gaussian_rasterization import GaussianRasterizationSettings, GaussianRasterizer
from scene.gaussian_model import GaussianModel
from utils.sh_utils import eval_sh
def render(viewpoint_camera, pc : GaussianModel, pipe, bg_color : torch.Tensor, scaling_modifier = 1.0, override_color = None):
"""
渲染场景: 将高斯分布的点投影到2D屏幕上来生成渲染图像
viewpoint_camera: 训练相机集合
pc: 高斯模型
pipe: 管道相关参数
bg_color: Background tensor 必须 on GPU
scaling_modifier:
override_color:
"""
# Create zero tensor. We will use it to make pytorch return gradients of the 2D (screen-space) means
# 创建一个与输入点云(高斯模型)大小相同的 零tensor用于记录屏幕空间中的点的位置。这个张量将用于计算对于屏幕空间坐标的梯度
screenspace_points = torch.zeros_like(pc.get_xyz, dtype=pc.get_xyz.dtype, requires_grad=True, device="cuda") + 0
try:
# 尝试保留张量的梯度。这是为了确保可以在反向传播过程中计算对于屏幕空间坐标的梯度
screenspace_points.retain_grad()
except:
pass
# Set up rasterization configuration
# 计算视场的 tan 值,这将用于设置光栅化配置
tanfovx = math.tan(viewpoint_camera.FoVx * 0.5)
tanfovy = math.tan(viewpoint_camera.FoVy * 0.5)
# 设置光栅化的配置,包括图像的大小、视场的 tan 值、背景颜色、视图矩阵viewmatrix、投影矩阵projmatrix等
raster_settings = GaussianRasterizationSettings(
image_height=int(viewpoint_camera.image_height),
image_width=int(viewpoint_camera.image_width),
tanfovx=tanfovx,
tanfovy=tanfovy,
bg=bg_color,
scale_modifier=scaling_modifier,
viewmatrix=viewpoint_camera.world_view_transform,
projmatrix=viewpoint_camera.full_proj_transform,
sh_degree=pc.active_sh_degree,
campos=viewpoint_camera.camera_center,
prefiltered=False,
debug=pipe.debug
)
# 创建一个高斯光栅化器对象,用于将高斯分布投影到屏幕上
rasterizer = GaussianRasterizer(raster_settings=raster_settings)
# 获取高斯模型的三维坐标、屏幕空间坐标、透明度
means3D = pc.get_xyz
means2D = screenspace_points
opacity = pc.get_opacity
# If precomputed 3d covariance is provided, use it. If not, then it will be computed from scaling / rotation by the rasterizer.
# 如果提供了预先计算的3D协方差矩阵则使用它。否则它将由光栅化器根据尺度和旋转进行计算
scales = None
rotations = None
cov3D_precomp = None
if pipe.compute_cov3D_python:
cov3D_precomp = pc.get_covariance(scaling_modifier)
else:
scales = pc.get_scaling
rotations = pc.get_rotation
# If precomputed colors are provided, use them. Otherwise, if it is desired to precompute colors
# from SHs in Python, do it. If not, then SH -> RGB conversion will be done by rasterizer.
# 如果提供了预先计算的颜色则使用它们。否则如果希望在Python中从球谐函数中预计算颜色请执行此操作。如果没有则颜色将通过光栅化器进行从球谐函数到RGB的转换
shs = None
colors_precomp = None
if override_color is None:
if pipe.convert_SHs_python:
# 将SH特征的形状调整为batch_size * num_points3(max_sh_degree+1)**2
shs_view = pc.get_features.transpose(1, 2).view(-1, 3, (pc.max_sh_degree+1)**2)
# 计算相机中心到每个点的方向向量,并归一化
dir_pp = (pc.get_xyz - viewpoint_camera.camera_center.repeat(pc.get_features.shape[0], 1))
# 计算相机中心到每个点的方向向量,并归一化
dir_pp_normalized = dir_pp/dir_pp.norm(dim=1, keepdim=True)
# 使用SH特征将方向向量转换为RGB颜色
sh2rgb = eval_sh(pc.active_sh_degree, shs_view, dir_pp_normalized)
# 将RGB颜色的范围限制在0到1之间
colors_precomp = torch.clamp_min(sh2rgb + 0.5, 0.0)
else:
shs = pc.get_features
else:
colors_precomp = override_color
# Rasterize visible Gaussians to image, obtain their radii (on screen).
# 调用光栅化器,将高斯分布投影到屏幕上,获得渲染图像和每个高斯分布在屏幕上的半径
rendered_image, radii = rasterizer(
means3D = means3D,
means2D = means2D,
shs = shs,
colors_precomp = colors_precomp,
opacities = opacity,
scales = scales,
rotations = rotations,
cov3D_precomp = cov3D_precomp)
# Those Gaussians that were frustum culled or had a radius of 0 were not visible.
# They will be excluded from value updates used in the splitting criteria.
return {"render": rendered_image,
"viewspace_points": screenspace_points,
"visibility_filter" : radii > 0,
"radii": radii}