Loading...

Overview

This challenge workshop aims to accelerate the transition of volumetric video technology from laboratory prototypes to robust, production- and consumer-ready systems, enabling breakthroughs in immersive interactive experiences.

Compression Track

Focuses on balancing storage efficiency and playback performance for practical deployment in real-world applications.

Sparse-View Track

Focuses on accurately reconstructing dynamic 3D humans from minimal camera viewpoints, reducing hardware costs and simplifying capture setups.

Our challenge fosters cross-disciplinary discussion between computer vision, computer graphics, and AR/VR practitioners, uniting expertise in 3D reconstruction, generative AI, and immersive media systems.

News

 September 4 (AoE)    ⚠️ The compression and sparse-view datasets have been released. For compression track, the FreeTimeGS models and their rendering script are now available. Access these resources at: Google Drive

 August 31 (AoE)    We are currently completing the final processing of our dataset, and the release is anticipated by September 5. Thank you for your patience.

Important Dates (AoE)

Registration Opens

Get ready to participate in the challenge

August 15, 2025

Dataset Release

Full dataset available for download

September 1, 2025  By September 5, 2025

Submission Deadline

Final submission of your results

November 23, 2025

Results Announcement

Winners will be notified

November 30, 2025

Workshop & Awards

Join us at SIGGRAPH Asia 2025

December 15-18, 2025

Challenge Tracks

The challenge features two tracks focusing on cutting-edge compression and sparse-view reconstruction techniques for volumetric videos.

Compression Track

Optimize file size while maintaining high reconstruction quality metrics. Perfect for teams working on efficient storage and streaming solutions.

Dataset: 2 validation sequences + 5 test sequences
Download link

Sparse-View Track

Reconstruct dynamic 3D humans from limited camera views. Ideal for teams exploring neural rendering and view synthesis.

Dataset: 2 validation sequences + 5 test sequences
Download link

Dataset

Our high-fidelity dataset features diverse dynamic human subjects with:

Mixed Focal Lengths

Cinematic-Grade Visual Quality

Challenging Motions

Dataset Structure:

.
├── intri.yml               # Camera intrinsics for training views
├── extri.yml               # Camera extrinsics for training views
├── test_intri.yml          # Camera intrinsics for testing views
├── test_extri.yml          # Camera extrinsics for testing views
├── images/                 # Multi-view images for training
│   ├── 00/                 # Camera name
│   │   ├── 000000.jpg      # Image name using format {frame:06d}.jpg
│   │   └── ...
│   ├── 01/
│   └── ...
├── masks/                  # Multi-view masks for training
│   ├── 00/                 # Camera name
│   │   ├── 000000.jpg
│   │   └── ...
│   ├── 01/
│   └── ...
└── pcds/                # Foreground point clouds
    ├── 000000.ply
    └── ...

Note: We provide official Python scripts for data parsing and visualization. All data are provided for research use only.

Evaluation

Hardware: All evaluations conducted on Linux workstation with one NVIDIA RTX 4090 GPU

Evaluatoin: We use PSNR, SSIM, and LPIPS to measure the foreground-only reconstruction quality

Compression Track

Rank = [rank(PSNR) + rank(SSIM) + rank(LPIPS)]/6 + [rank(Size) + rank(Time)]/4

Size: Total on-disk bytes of all content-dependent artifacts required to render. Content-agnostic parts (e.g., shared backbones, shared decoders, ...) are excluded.
Time: Average rendering time of a fixed number of test images with batch size = 1, including preprocessing, decoding, and rendering.

Sparse-View Track

Rank = [rank(PSNR) + rank(SSIM) + rank(LPIPS)]/3

Detailed Requirements

Compression Track

Dataset Split

Validation Set (2 sequences)

  • • Full 60-view videos & masks
  • • Full 60-view camera parameters

Test Set (5 sequences)

  • • 48-view videos & masks
  • • Training & testing camera parameters

Baseline

We provide a FreeTimeGS result as baseline. Participants can use it as a starting point (e.g., linear/non-linear quantization and pruning) or develop novel compression methods.

Submission Requirements

1

Technical Report (PDF, max 4 pages)

SIGGRAPH Asia Technical Communications template recommended

2

Rendered Results (ZIP)

A zip file containing testing-view images in a specified directory structure:

output/
├── 00  # Testing camera name defined by test_intri.yml and test_extri.yml
│   ├── 000000.jpg
│   ├── ...
├── 01
│   ├── 000000.jpg
│   ├── ...
...
3

Model & Scripts (ZIP)

A zip file containing:

  • • A .txt file describing all content-dependent files for size computing. One file per line.
  • • Your Conda environment .yml file
  • • Your compressed models and rendering scripts
  • ▪ The rendering code should support the following evaluation commands:
  • conda env create -f [YOUR_ENV_FILE].yml
    conda activate [YOUR_ENV_NAME]
    python3 render.py --model [YOUR_MODEL] --intri test_intri.yml --extri test_extri.yml (default)
    // or
    python3 render.py --model [YOUR_MODEL] --intri test_intri.yml --extri test_extri.yml --no_image (no image saving for time computing)
    
  • ▪ The default command should generate a folder named output, with the same structure as Rendered Results (ZIP).
  • ▪ Please ensure your code runs non-interactively and reproducibly in a clean environment with the above commands.

Sparse-View Track

Dataset Split

Validation Set (2 sequences)

  • • Full 60-view videos & masks
  • • Full 60-view camera parameters

Test Set (5 sequences)

  • • 8-view videos & masks
  • • Training & testing camera parameters

Submission Requirements

1

Technical Report (PDF, max 4 pages)

SIGGRAPH Asia Technical Communications template recommended

2

Rendered Results (ZIP)

A zip file containing testing-view images in a specified directory structure:

output/
├── 00  # Testing camera name defined by test_intri.yml and test_extri.yml
│   ├── 000000.jpg
│   ├── ...
├── 01
│   ├── 000000.jpg
│   ├── ...
...

Submission Guidelines

Submission

  • One registration per team
  • Team name serves as official identifier
  • Maximum 3 submissions per track

File Naming Convention

Report: TeamName.pdf

Results: TeamName.zip

Model: TeamName_Model.zip

Note: All reports are non-archival. Submission link will be provided soon.

Awards

Each track features two prestigious prizes

🥇

First Prize

$2,500

🥈

Second Prize

$1,500

Winners will present their work via a 5–10 minute pre-recorded video at the workshop.

Workshop Schedule

The workshop will take place during SIGGRAPH Asia 2025 (exact date TBD)

Time (HKT) Event
13:30 - 14:00 Welcome Remarks & Challenge Results
14:00 - 15:00 Winner Teams Presentation
15:00 - 15:30 Coffee Break & Networking
15:30 - 17:00 Invited Speaker Presentations

Keynote Speakers

Distinguished experts in volumetric video and 3D reconstruction

TBD

Speaker 1

Details coming soon

TBD

Speaker 2

Details coming soon

TBD

Speaker 3

Details coming soon

Organizers

Zhiyuan Yu

Zhiyuan Yu

Zhejiang University

Homepage
Jiaming Sun

Jiaming Sun

4DV.ai

Homepage
Siyu Zhang

Siyu Zhang

4DV.ai

Homepage
Sida Peng

Sida Peng

Zhejiang University

Homepage
Ruizhen Hu

Ruizhen Hu

Shenzhen University

Homepage
Xiaowei Zhou

Xiaowei Zhou

Zhejiang University

Homepage

Acknowledgements

We thank the following institutions for their support:

We also thank Yuanhong Yu and Yuxuan Lin for their valuable contributions to the development of the website and the dataset preparation.

Contact

For any questions, please contact us at

[email protected]