Xiaoyu Xiang | 向小雨

I am a staff research scientist at Meta Reality Labs, where I work on 3D Computer Vision and Generative-AI research for AR and VR.

I did my Ph.D. at Purdue University, where I was advised by Jan Allebach. I received my Bachelor's degree from Tsinghua University.

Email  /  CV  /  Bio  /  Google Scholar  /  Twitter  /  Github  /  LinkedIn

profile photo
Research

My current research interests include Generative Model, Computational Photography, Temporal Modeling, and Novel View Synthesis. Below is a partial list of my papers.

prl Garment3DGen: 3D Garment Stylization and Texture Generation
Nikolaos Sarafianos, Tuur Stuyck, Xiaoyu Xiang, Yilei Li, Jovan Popovic, Rakesh Ranjan
arXiv, 2024
[Paper] [Suppl] [Code] [Project Page] [Video] [Data]

Garment3DGen stylizes the geometry and textures of real and fantastical garments that we can fit on top of parametric bodies and simulate.

PlatoNeRF: 3D Reconstruction in Plato's Cave via Single-View Two-Bounce Lidar
Tzofi Klinghoffer, Xiaoyu Xiang*, Siddharth Somasundaram*, Yuchen Fan, Christian Richardt, Ramesh Raskar, Rakesh Ranjan
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2024 (Oral, Best Paper Award Finalist)
[Paper] [Code] [Video] [Dataset]

A method to recover scene geometry from a single view using two-bounce signals captured by a single-photon lidar.

EfficientSAM: Leveraged Masked Image Pretraining for Efficient Segment Anything
Yunyang Xiong, Bala Varadarajan*, Lemeng Wu*, Xiaoyu Xiang, Fanyi Xiao, Chenchen Zhu, Xiaoliang Dai, Dilin Wang, Fei Sun, Forrest Iandola, Raghuraman Krishnamoorthi, Vikas Chandra
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2024 (Highlight)
[Paper] [Code] [Video] [HuggingFace]

A light-weight SAM model that exhibits decent performance with largely reduced complexity.

prl UnSAMFlow: Unsupervised Optical Flow Guided by Segment Anything Model
Shuai Yuan, Lei Luo, Zhuo Hui, Can Pu, Xiaoyu Xiang, Rakesh Ranjan, Denis Demandolx,
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2024
[Paper] [Appendix] [Code] [Poster] [Video] [BibTex]

An unsupervised flow network that leverages object information from the latest foundation model Segment Anything Model (SAM).

prl CAD: Photorealistic 3D Generation via Adversarial Distillation
Ziyu Wan, Despoina Paschalidou, Ian Huang, Hongyu Liu, Bokui Shen, Xiaoyu Xiang, Jing Liao Leonidas Guibas
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2024
[PDF] [Project Page] [Code] [Video]

A new approach for generating high-quality, photoreaslisitc and diverse 3D objects conditioned on a single image and a text prompt.

prl Customizing 360-Degree Panoramas through Text-to-Image Diffusion Models
Hai Wang, Xiaoyu Xiang, Yuchen Fan, Jing-Hao Xue
IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2024
[PDF] [Project Page] [Code]

Synthesize seamless 360-degree panoramas with given text prompts.

SqueezeSAM: User friendly mobile interactive segmentation
Bala Varadarajan, Bilge Soran, Xiaoyu Xiang, Forrest Iandola, Yunyang Xiong, Lemeng Wu, Chenchen Zhu, Naveen Suda, Raghuraman Krishnamoorthi, Vikas Chandra
arXiv, 2023
[Paper] [Code]

SqueezeSAM is 62.5x faster and 31.6x smaller than its predecessor, making it a viable solution for mobile applications.

prl Learning Neural Duplex Radiance Fields for Real-Time View Synthesis
Ziyu Wan, Christian Richardt, Aljaž Božič, Chao Li, Vijay Rengarajan, Seonghyeon Nam, Xiaoyu Xiang, Tuotuo Li, Bo Zhu, Rakesh Ranjan, Jing Liao
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023
[PDF] [WebGL Viewer] [Project Page] [Video]

Represent scenes as neural radiance features encoded on a two-layer duplex mesh, overcoming inaccuracies in 3D surface reconstruction.

Efficient and Explicit Modelling of Image Hierarchies for Image Restoration
Yawei Li, Yuchen Fan, Xiaoyu Xiang, Denis Demandolx, Rakesh Ranjan, Radu Timofte, Luc Van Gool
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2023
[PDF] [Code]

Providing a mechanism to efficiently and explicitly model image hierarchies in the global, regional, and local range for image restoration.

STDAN: deformable attention network for space-time video super-resolution
Hai Wang, Xiaoyu Xiang, Yapeng Tian, Wenming Yang, Qingmin Liao
IEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2023
[PDF] [Code]

A deformable attention network that adaptively captures and aggregates spatial and temporal contexts in dynamic video to enhance reconstruction.

Service

Organizer: Computer Vision for Mixed Reality Workshop, CVPR 2023, 2024

Conference Reviewer: ICLR 2021-2022, CVPR 2021-2024, ICCV 2021-2023, ECCV 2022-2024, NeurIPS 2021-2023, ICML 2022, SIGGRAPH 2024, SIGGRAPH-Asia 2024, WACV 2022-2025

Journal Reviewer: T-PAMI, TNNLS, TMM, NCAA, IEEE Access, Journal of Automatica Sinica, Neurocomputing, Journal of Electronic Imaging

Experience

2021.08~ present       Research Scientist in Meta Reality Labs

2020.08~2021.03       Research Intern in Facebook Reality Labs

2020.06~2020.08       Research Intern in ByteDance

2018.05~2020.05       Research Student in HP Labs

2017.08~2020.05       Graduate Research Assistant in ECE, Purdue University

2015.07~2017.05       Research Engineer in Optical Fiber Research Center, CAEP

2014.07~2014.09       Summer Research Student in DESY

2012.05~2015.07       Undergraduate Research Assistant in Tsinghua University