Before joining Peking University, I was a research scientist at
NVIDIA Toronto AI
Lab.
I earned my Ph.D. from the University of Toronto and received both my Master's
and Bachelor's degrees from Shandong University.
I am always actively recruiting Ph.D. students and research interns! Feel free
to drop me a line with your CV and research statement!
My research focuses on Computational Photography and 3D
Vision.
My goal is to integrate 3D Sensing with World
Models to enable intelligent agents to perceive, simulate, and
interact with the physical world.
News
[2025.12] One paper accepted
to
TMLR. We
solve the mode collapse problem by replacing KL with SIM loss.
[2025.12] One paper accepted
to
TIP.
We tackle the challenging blind inverse problems with latent diffusion priors.
[2025.08] One paper accepted
to SIGGRAPH
Asia 2025. We developed a
powerful structured light 3D imaging technique achieving 10x accuracy
improvement over traditional methods.
[2025.03] One paper
accepted to ICCV
2025. Check out GeoSplatting:
Geometry-guided
Gaussian Splatting!
GeoSplatting introduces a novel hybrid representation that grounds 3DGS with
isosurfacing to provide accurate geometry and normals for high-fidelity inverse
rendering.
RainyGS integrates physics simulation with 3DGS to efficiently generate
photorealistic, physically accurate, and controllable dynamic rain effects for
in-the-wild scenes.
We develop a 3D generative model to generate meshes with textures, bridging the
success in the
differentiable surface modeling, differentiable rendering and 2D GANs.
Nvdiffrec reconstructs 3D mesh with materials from multi-view images by combining
diff surface
modeling with diff renderer. The method supports Nvidia Neural Drivesim
DIB-R++ is a high-performant differentiable renderer which combines rasterization
and ray-tracing
together and supports advanced lighitng and material effects. We further embed
it in deep learning
and jointly predict geometry, texture, light and material from a single image.
We explore StyleGAN as a multi-view image generator and
train inverse graphics from StyleGAN images. Once trained,
the invere graphics model further helps disentangle and
manipulate StyleGAN latent code from graphics
knowledge. Our work was featured at NVIDIA GTC 2021 and has become an Omniverse
product.
We predict deformable tetrahedral meshes from images or
point clouds, which support arbitrary topologies. We also
design a differentiable renderer for tetrahedron, allowing
3D reconstrucion from 2D supervison only.
We present optical SGD, a computational imaging technique
that allows an active depth imaging system to
automatically discover optimal illuminations & decoding.
An interpolation-based 3D mesh differentiable renderer
that supports vertex, vertex color, multiple lighting
models, texture mapping and could be easily embedded in
neural networks.
We predict object polygon contours from graph neural
networks, where a novel 2D differentiable rendering loss is
introduced. It renders a polygon countour into a segmentation mask
and back propagates the loss to help optimize the polygon
vertices.
alacarte designs structured light patterns from a maching
learning persepctive, where patterns are automatically
optimized by minimizing the disparity error under any given imaging condition.