Minseop Kwak*, Minkyung Kwon*, Jinhyeok Choi*, Jiho Park,
Seungryong Kim†
KAIST AI
* Equal contribution † Corresponding author
We propose a novel framework for diffusion-based novel view synthesis in which we leverage external representations as conditions, harnessing their geometric and semantic correspondence properties for enhanced geometric consistency in generated novel viewpoints. First, we provide a detailed analysis exploring the correspondence capabilities emergent in the spatial attention of external visual representations. Building from these insights, we propose a representation-guided novel view synthesis through dedicated representation projection modules that inject external representations into the diffusion process, a methodology named ReNoV (Representation-guided Novel View synthesis). Our experiments show that this design yields marked improvements in both reconstruction fidelity and inpainting quality, outperforming prior diffusion‐based novel‐view methods on standard benchmarks and enabling robust synthesis from sparse, unposed image collections.
- Release Code
- Release Pretrained Models
If you find our work useful in your research, please consider citing:
@article{kwak2026renov,
title={Projected Representation Conditioning for High-fidelity Novel View Synthesis},
author={Kwak, Minseop and Kwon, Minkyung and Choi, Jinhyeok and Park, Jiho and Kim, Seungryong},
journal={arXiv preprint},
year={2026}
}