Welcome to the 1st 3D Restoration and Reconstruction Challenge host at New Trends in Image Restoration and Enhancement (NTIRE) Workshop, in conjunction with CVPR 2026.
This challenge aims to advance robust 3D reconstruction under real-world, in-the-wild degradations. Specifically, the proposed pipeline should account for realistic visual corruptions, restore degraded training views, and reconstruct clean 3D representations for high-quality novel-view synthesis (NVS).
To support this goal, we curate a comprehensive 3D benchmark, RealX3D, and introduce two tracks: (1) 3D Low-Light Enhancement and (2) 3D Smoke Restoration. Each track provides multiple scenes with multi-view images degraded by low-light or smoke, along with clean NVS references for evaluation. Submissions are assessed by comparing rendered NVS against the corresponding ground-truth captures. To simplify the evaluation protocol, both tracks focus on photometric fidelity, and the ground-truth camera poses of the training views are provided.
Participants are encouraged to approach this challenge from two key perspectives.
Top-ranked participants will receive awards and be invited to present their solutions at the NTIRE workshop held in conjunction with CVPR 2026. The challenge results will also be published in the CVPR NTIRE 2026 workshop proceedings.
Low-light imaging is a long-standing challenge in 2D vision, where limited photons lead to severe noise, color shifts, and loss of fine textures, degrading both visual quality and downstream perception. While recent 2D low-light enhancement methods can produce visually pleasing results, they are often optimized for single images and may introduce view-dependent artifacts or hallucinated details. In real applications such as robotics, AR/VR, and autonomous systems, models must operate on multi-view observations and build a coherent 3D representation of the scene. This makes low-light enhancement fundamentally harder in 3D: the method must improve visibility while preserving cross-view consistency so that geometry and appearance remain stable when rendering novel views. The 3D Low-Light Enhancement Challenge targets this gap by encouraging approaches that bridge 2D enhancement and 3D multiview geometry, enabling reliable scene understanding and high-quality restoration and rendering under realistic low-light conditions.
We provide one fully released scene containing both degraded and clean images for debugging and validation. The development set includes 4 scenes, and performance is evaluated based on the submitted restored NVS results. During the testing period, the test set of 3 scenes will be released and evaluated using the same protocol. All low-light images are captured with a shutter speed of 1/400, and the well-lit reference images are captured at 1/10 under same camera settings. Participants may use additional training data or pretrained models, provided that all external resources and training details are clearly documented in the final factsheet.
| Part | Scenes | Train. Images |
Train. Poses |
Test. Images |
Test. Poses |
Leaderboard |
|---|---|---|---|---|---|---|
| Part I | 1 | Not evaluated | ||||
| Part II | 4 | Development (immediate after submission) | ||||
| Part III | 3 | Test (when testing begins) |
For each scene, the dataset is organized as follows:
scene/
train/
0001.JPG
0002.JPG
...
transforms_train.json
transforms_test.json
transforms_train.json and transforms_test.json follow the Blender dataset format and include
camera intrinsics and extrinsics (poses) for training views and NVS views to be submitted. Examples of degraded and clean image pairs are shown below.
Participants are supposed to submit the restored NVS images given the testing view poses of each scene. We evaluate the rendered results against the clean ground-truth references using PSNR, SSIM, and LPIPS. The final ranking is determined by the average per-scene performance across both the Development and Testing sets.
Each frame in the RealX3D dataset provides a pair of degraded and clean images. To acquire these sequences, we used the mechanical rail-dolly system shown in the video, where the camera moved at a constant speed along the rails. The detailed data acquisition protocol can be found below.
RealX3D is a real-capture benchmark for multi-view visual restoration and 3D reconstruction under diverse physical degradations. Corruptions are grouped into four families—illumination, scattering, occlusion, and blurring—and captured at multiple severity levels using a unified acquisition protocol that yields pixel-aligned LQ/GT views. Each scene includes high-resolution capture, RAW images, and dense laser scans, from which we derive world-scale meshes and metric depth.
RealX3D provides 2,407 paired low-quality and reference images, along with the same number of corresponding RAW captures. The dataset is collected across 15 indoor rooms and organized into 55 distinct scenes spanning seven degradation types. The current release includes defocus or camera motion blur with 8 scenes and 271 pairs, where each scene is captured at two blur severity levels; dynamic occlusion with 8 scenes and 271 pairs; reflection with 8 scenes and 271 pairs; extreme low light with 9 scenes and 319 pairs; low-light exposure variation with 9 scenes and 319 pairs; and smoke scattering with 5 scenes and 143 pairs. We further provide laser-scanned point clouds with 5 mm point spacing, along with calibrated camera intrinsics and extrinsics. Each view is paired with a metric depth map stored as a 16-bit PNG in millimeters, and each scene includes a colored mesh reconstructed from the scanned point clouds.
@article{liu2026realx3d,
title={RealX3D: A Physically-Degraded 3D Benchmark for Multi-view Visual Restoration and Reconstruction},
author={Liu, Shuhong and Bao, Chenyu and Cui, Ziteng and Liu, Yun and Chu, Xuangeng and Gu, Lin and Conde, Marcos V and Umagami, Ryo and Hashimoto, Tomohiro and Hu, Zijian and others},
journal={arXiv preprint arXiv:2512.23437},
year={2026}
}