Overview
This challenge workshop aims to accelerate the transition of volumetric video technology from laboratory prototypes to robust, production- and consumer-ready systems, enabling breakthroughs in immersive interactive experiences.
Compression Track
Focuses on balancing storage efficiency and playback performance for practical deployment in real-world applications.
Sparse-View Track
Focuses on accurately reconstructing dynamic 3D humans from minimal camera viewpoints, reducing hardware costs and simplifying capture setups.
Our challenge fosters cross-disciplinary discussion between computer vision, computer graphics, and AR/VR practitioners, uniting expertise in 3D reconstruction, generative AI, and immersive media systems.
News
September 4 (AoE) ⚠️ The compression and sparse-view datasets have been released. For compression track, the FreeTimeGS models and their rendering script are now available. Access these resources at: Google Drive
August 31 (AoE) We are currently completing the final processing of our dataset, and the release is anticipated by September 5. Thank you for your patience.
Important Dates (AoE)
Registration Opens
Get ready to participate in the challenge
Dataset Release
Full dataset available for download
Submission Deadline
Final submission of your results
Results Announcement
Winners will be notified
Workshop & Awards
Join us at SIGGRAPH Asia 2025
Challenge Tracks
The challenge features two tracks focusing on cutting-edge compression and sparse-view reconstruction techniques for volumetric videos.
Compression Track
Optimize file size while maintaining high reconstruction quality metrics. Perfect for teams working on efficient storage and streaming solutions.
Sparse-View Track
Reconstruct dynamic 3D humans from limited camera views. Ideal for teams exploring neural rendering and view synthesis.
Dataset
Our high-fidelity dataset features diverse dynamic human subjects with:
Mixed Focal Lengths
Cinematic-Grade Visual Quality
Challenging Motions
Dataset Structure:
. ├── intri.yml # Camera intrinsics for training views ├── extri.yml # Camera extrinsics for training views ├── test_intri.yml # Camera intrinsics for testing views ├── test_extri.yml # Camera extrinsics for testing views ├── images/ # Multi-view images for training │ ├── 00/ # Camera name │ │ ├── 000000.jpg # Image name using format {frame:06d}.jpg │ │ └── ... │ ├── 01/ │ └── ... ├── masks/ # Multi-view masks for training │ ├── 00/ # Camera name │ │ ├── 000000.jpg │ │ └── ... │ ├── 01/ │ └── ... └── pcds/ # Foreground point clouds ├── 000000.ply └── ...
Note: We provide official Python scripts for data parsing and visualization. All data are provided for research use only.
Evaluation
Hardware: All evaluations conducted on Linux workstation with one NVIDIA RTX 4090 GPU
Evaluatoin: We use PSNR, SSIM, and LPIPS to measure the foreground-only reconstruction quality
Compression Track
Rank = [rank(PSNR) + rank(SSIM) + rank(LPIPS)]/6 + [rank(Size) + rank(Time)]/4
Size: Total on-disk bytes of all content-dependent artifacts required to render. Content-agnostic parts (e.g., shared backbones, shared decoders, ...) are excluded.
Time: Average rendering time of a fixed number of test images with batch size = 1, including preprocessing, decoding, and rendering.
Sparse-View Track
Rank = [rank(PSNR) + rank(SSIM) + rank(LPIPS)]/3
Detailed Requirements
Compression Track
Dataset Split
Validation Set (2 sequences)
- • Full 60-view videos & masks
- • Full 60-view camera parameters
Test Set (5 sequences)
- • 48-view videos & masks
- • Training & testing camera parameters
Baseline
We provide a FreeTimeGS result as baseline. Participants can use it as a starting point (e.g., linear/non-linear quantization and pruning) or develop novel compression methods.
Submission Requirements
Technical Report (PDF, max 4 pages)
SIGGRAPH Asia Technical Communications template recommended
Rendered Results (ZIP)
A zip file containing testing-view images in a specified directory structure:
output/ ├── 00 # Testing camera name defined by test_intri.yml and test_extri.yml │ ├── 000000.jpg │ ├── ... ├── 01 │ ├── 000000.jpg │ ├── ... ...
Model & Scripts (ZIP)
A zip file containing:
- • A .txt file describing all content-dependent files for size computing. One file per line.
- • Your Conda environment .yml file
- • Your compressed models and rendering scripts
- ▪ The rendering code should support the following evaluation commands:
- ▪ The default command should generate a folder named output, with the same structure as Rendered Results (ZIP).
- ▪ Please ensure your code runs non-interactively and reproducibly in a clean environment with the above commands.
conda env create -f [YOUR_ENV_FILE].yml conda activate [YOUR_ENV_NAME] python3 render.py --model [YOUR_MODEL] --intri test_intri.yml --extri test_extri.yml (default) // or python3 render.py --model [YOUR_MODEL] --intri test_intri.yml --extri test_extri.yml --no_image (no image saving for time computing)
Sparse-View Track
Dataset Split
Validation Set (2 sequences)
- • Full 60-view videos & masks
- • Full 60-view camera parameters
Test Set (5 sequences)
- • 8-view videos & masks
- • Training & testing camera parameters
Submission Requirements
Technical Report (PDF, max 4 pages)
SIGGRAPH Asia Technical Communications template recommended
Rendered Results (ZIP)
A zip file containing testing-view images in a specified directory structure:
output/ ├── 00 # Testing camera name defined by test_intri.yml and test_extri.yml │ ├── 000000.jpg │ ├── ... ├── 01 │ ├── 000000.jpg │ ├── ... ...
Submission Guidelines
Submission
- One registration per team
- Team name serves as official identifier
- Maximum 3 submissions per track
File Naming Convention
Report: TeamName.pdf
Results: TeamName.zip
Model: TeamName_Model.zip
Note: All reports are non-archival. Submission link will be provided soon.
Awards
Each track features two prestigious prizes
First Prize
$2,500
Second Prize
$1,500
Winners will present their work via a 5–10 minute pre-recorded video at the workshop.
Workshop Schedule
The workshop will take place during SIGGRAPH Asia 2025 (exact date TBD)
Time (HKT) | Event |
---|---|
13:30 - 14:00 | Welcome Remarks & Challenge Results |
14:00 - 15:00 | Winner Teams Presentation |
15:00 - 15:30 | Coffee Break & Networking |
15:30 - 17:00 | Invited Speaker Presentations |
Keynote Speakers
Distinguished experts in volumetric video and 3D reconstruction
Speaker 1
Details coming soon
Speaker 2
Details coming soon
Speaker 3
Details coming soon
Organizers
Acknowledgements
We thank the following institutions for their support:
We also thank Yuanhong Yu and Yuxuan Lin for their valuable contributions to the development of the website and the dataset preparation.
Contact
For any questions, please contact us at
[email protected]