Xue-qian Li


I am a PhD student at the University of Adelaide. I was a "long-term" research intern at Argo AI, fortunately working with Dr. Simon Lucey and Dr. Jhony Kaesemodel Pontes.

  • Research
  • Publication
  • About Me

Fast Neural Scene Flow
Neural Scene Flow Prior (NSFP) is of significant interest to the vision community due to its inherent robustness to out-of-distribution (OOD) effects and its ability to deal with dense lidar points. The approach utilizes a coordinate neural network to estimate scene flow at runtime, without any training. However, it is up to 100 times slower than current state-of-the-art learning methods. We rediscover the distance transform (DT) as an efficient, correspondence-free loss function that dramatically speeds up the runtime optimization, allowing for the first time real-time performance comparable to learning methods.
Trading Positional Complexity vs. Deepness in Coordinate Networks
We use non-Fourier positional encodings (e.g., shifted Gaussian functions) to show the signal reconstruction is determined by a trade-off between the stable rank of the embedding and the distance preservation between embedded coordinates. Furthermore, employing a complex positional encoding requires only a linear (rather than deep) function to achieve comparable performance, while being orders of magnitude faster than current state-of-the-art.
Neural Prior for Trajectory Estimation
Traditionally, trajectories have been represented by a set of handcrafted bases that have limited expressibility. Here, we propose a neural trajectory prior to capture continuous spatio-temporal information without the need for offline data. We demonstrate how our proposed objective is optimized during runtime to estimate trajectories for two important tasks: Non-Rigid Structure from Motion (NRSfM) and lidar scene flow integration for self-driving scenes.
Neural Scene Flow Prior
We propose a neural scene flow prior that utilizes the architecture of neural networks (MLPs) as an implicit regularizer. Unlike learning-based scene flow methods, optimization occurs at runtime, and our approach needs no offline datasets -- making it ideal for deployment in new environments such as autonomous driving. Also, the implicit and continuous scene flow representation allows us to estimate dense long-term correspondences across a sequence of point clouds.
PointNetLK Revisited
We revisit a recent innovation -- PointNetLK -- and show that the inclusion of an analytical Jacobian can exhibit remarkable generalization properties while reaping the inherent fidelity benefits of a learning framework. Our approach not only outperforms the state-of-the-art in mismatched conditions but also produces results competitive with current learning methods when operating on real-world test data close to the training set.
PCRNet: Point Cloud Registration Network using PointNet Encoding
We provide a novel network to align point clouds that utilizes PointNet encoding. Our PCRNet uses data specific information to make accurate registration estimation, which is robust to noise and large initial misalignment between source point cloud and template point cloud. With single pass, PCRNet performs computationally faster than existing methods, while iterative version of PCRNet can refine the registration result to make highly accurate prediction. Our PCRNet can be applied further to 3D tracking, model replacement.
A Multi-Domain Feature Learning Method for Visual Place Recognition
We propose a multi-domain feature learning-based visual place recognition method with 2 modules: feature extraction module, which uses CapsuleNet to extract features under different circumstances; feature separation module, which enforces a separation in environmental condition-invariant and condition-related domains. Combining 2 modules through sequential matching, we can find potential match.
MRS-VPR: A Multi-Resolution Sampling-based Global Visual Place Recognition Method
In order to perform sequence matching between short-term testing frame and long-term reference frame in visual place recognition, we develop a multi-resolution sampling-based global search method with coarse-to-fine searching and particle filter, which balances the matching accuracy and searching efficiency. By iteratively updating both potential trajectories and frame sequences through low resolution to high resolution, we can find the best match under the highest resolution level.

Conference & arXiv
Xueqian Li, Simon Lucey. Fast Kernel Scene Flow. arXiv preprint.
Jianqiao Zheng, Xueqian Li, Simon Lucey. Convolutional Initialization for Data-Efficient Vision Transformers. arXiv preprint.
Kavisha Vidanapathirana, Shin-Fang Ch'ng, Xueqian Li, Simon Lucey. Multi-Body Neural Scene Flow. 3DV 2024 (Oral, award candidate)
Jianqiao Zheng, Xueqian Li, Sameera Ramasinghe, Simon Lucey. Robust Point Cloud Processing through Positional Embedding. 3DV 2024
Xueqian Li, Jianqiao Zheng, Francesco Ferroni, Jhony Kaesemodel Pointes, Simon Lucey. Fast Neural Scene Flow. ICCV 2023
Jianqiao Zheng, Sameera Ramasinghe, Xueqian Li, Simon Lucey. Trading Positional Complexity vs. Deepness in Coordinate Networks. ECCV 2022
Chaoyang Wang, Xueqian Li, Jhony Kaesemodel Pointes, Simon Lucey. Neural Prior for Trajectory Estimation. CVPR 2022
Xueqian Li, Jhony Kaesemodel Pointes, Simon Lucey. Neural Scene Flow Prior. NeurIPS 2021 (Spotlight)
Xueqian Li, Jhony Kaesemodel Pointes, Simon Lucey. PointNetLK Revisited. CVPR 2021 (Oral)
Vinit Sarode*, Xueqian Li*, Hunter Goforth, Yasuhiro Aoki, Animesh Dhagat, Arun R. Srivatsan, Simon Lucey, Howie Choset. One Framework to Register Them All: PointNet Encoding for Point Cloud Alignment. arXiv preprint. (* equal contribution)
Vinit Sarode*, Xueqian Li*, Hunter Goforth, Yasuhiro Aoki, Arun R. Srivatsan, Simon Lucey, Howie Choset. PCRNet: Point Cloud Registration Network using PointNet Encoding. arXiv preprint. (* equal contribution)
Peng Yin, Lingyun Xu, Xueqian Li, Yin Chen, Yingli Li, R. Arun Srivatsan, Lu Li, Jianmin Ji, and Yuqing He. A Multi-Domain Feature Learning Method for Visual Place Recognition. ICRA 2019
Peng Yin, R. Arun Srivatsan, Yin Chen, Xueqian Li, Hongda Zhang, Lingyun Xu, Lu Li, Zhenzhong Jia, Jiamin Ji and Yuqing He. MRS-VPR: a multi-resolution sampling based global visual place recognition method. ICRA 2019

Education
2021-202?, The University of Adelaide,
Doctor of Philosophy, Mathematics and Computer Sciences.
Working with Dr. Simon Lucey.
2017-2019, Carnegie Mellon University,
Master of Science, Biomedical Engineering.
Working with Dr. Howie Choset, Dr. Arun Srivatsan Rangaprasad, and Dr. Peng Yin in the Biorobotics Lab.
2013-2017, Jilin University,
Bachelor of Science, Biotechnology (major); Bachelor of Engineering, Computer and Application (minor).