About 2,280 results
Open links in new tab
  1. Use Gradient Scaling when True is CPU bound for Nerfacto-huge

    Oct 11, 2023 · Use Gradient Scaling when True is CPU bound for Nerfacto-huge #2511 Open samhodge-aiml opened this issue on Oct 11, 2023 · 5 comments

  2. no gradient propagation in info ["mean2d"] for forward rendering …

    Jun 12, 2025 · I also occurred this, that is a bug in the 3dgut method. May I ask, besides ignoring the gradient of this variable, are there any other good solutions? Or should we just wait for the …

  3. Where is the gradient computation for camera poses …

    May 7, 2025 · Where is the gradient computation for camera poses implemented in GSplat? #668 New issue Closed caikunyang

  4. How to add custom backward gradients · Issue #465 - GitHub

    Oct 24, 2024 · After a lot of trial and error I think the gradient is computing well now, however I still do not know a good procedure to tackle this task: how to modify/add the derivative in the …

  5. Pose optimization for 2dgs? · Issue #546 · nerfstudio-project/gsplat

    Hello, I happened to compare the pose optimization gradient between 3dgs and 2dgs. I found that viewmat has always 0.0 gradient. A minimal example can be found as blew.

  6. Viewmat gradient backward to optimize camera position

    Feb 1, 2024 · In order to re activate camera pose opt in nerfstudio, I cherry pick the gradient backward proposed in branch : https://github.com/nerfstudio-project/gsplat/tree/vickie/camera …

  7. [Bug] 2DGS gradient mismatch with original implementation in

    Apr 12, 2025 · Summary In compute_ray_transforms_aabb_vjp, the gradient of ray_transform for degenerate Gaussians (v_means2d [0] != 0 || v_means2d [1] != 0) seems to be off from the …

  8. Purpose of Gradient Scaling · Issue #100 - GitHub

    Gradient scaling is used in train_mlp_nerf.py, train_ngp_nerf.py and train_mlp_dnerf.py without autocasting. Moreover, gradient unscaling is not performed before optimizer.step (). Hence, …

  9. Weird gradient detach in semantic Nerfacto #1258 - GitHub

    Jan 19, 2023 · I was wondering in the nerfacto semantic head if there would be a gradient issue. Indeed in the following line the gradient is detached preventing from backpropagation.

  10. Scale and normalization of means2d grad #298 - GitHub

    Jul 23, 2024 · I have a question on implementation. In DefaultStrategy, GS growth is done with a comparison of a threshold and a gradient norm. The calculation of the gradient norm is finally …