Yi Zhou (Joeey, 周易), PhD Candidate

Australian National University (ANU), Australia

Contact: yi.zhou [at] anu.edu.au


ANU Homepage

Recent news
13/04/2018: I finish the PhD defence (thesis presentation).
12/03/2018: I am awarded the NCCR grant for a fellowship
at Prof. Davide Scaramuzza's lab in ETHZ/UZH. NCCR LINK

28/08/2017: I will join the Robotic Perception Group (RPG)
of the University of Zurich since 16 Sep 2017, supervised by
Prof. Davide Scaramuzza.






About me

I'm a PhD candidate at Research School of Engineering, College of Engineering and Computer Science, ANU. My work is under the supervision of Prof. Hongdong Li (ANU) and Dr. Laurent Kneip (ANU).

I'm awarded the CSC Scholarship by the China Scholarship Council (CSC) to support the study at ANU from Sep 2014.

Previously, I worked as a PhD student under the supervision of Prof. Tianmiao Wang and A/Prof. Jianhong Liang at Institute of Robotics, Beijing University of Aeronautics and Astronautics (BUAA). I obtained my bachelor's degree of Aircraft Manufacturing from BUAA in 2012.

Research Interests

  • Computer Vision: Structure from Motion, 3D Vision
  • Robot Vision: Vision-based UAV's GNC, VO/vSLAM, Event Cameras

Publications

Canny-SLAM
Under review.
[Code] [Our Sequence] [Our Result] [ORB_SLAM2 Evaluation on Our Sequence] [Youtube]

Yi Zhou, Laurent Kneip, Hongdong Li
Semi-Dense Visual Odometry for RGB-D Cameras Using Approximate Nearest Neighbour Fields
The 2017 IEEE International Conference on Robotics and Automation (ICRA).
[BibTex] [Arxiv version] [Video]

Yi Zhou, Laurent Kneip, Cristian Rodriguez, Hongdong Li
Divide and Conquer: Effcient Density-Based Tracking of 3D Sensors in Manhattan Worlds
The 13th Asian Conference on Computer Vision (ACCV 2016), Oral presentation
[Abstract] [BibTex] [PDF] [Supplementary Material] [Video1] [Video2] [Sample Code]

3D depth sensors such as LIDARs and RGB-D cameras have become a popular choice for indoor localization and mapping. However, due to the lack of direct frame-to-frame correspondences, the tracking traditionally relies on the iterative closest point technique which does not scale well with the number of points. In this paper, we build on top of more recent and effcient density distribution alignment methods, and notably push the idea towards a highly effcient and reliable solution for full 6DoF motion estimation with only depth information. We propose a divide-and-conquer technique during which the estimation of the rotation and the three degrees of freedom of the translation are all decoupled from one another. The rotation is estimated absolutely and drift-free by exploiting the orthogonal structure in man-made environments. The underlying algorithm is an effcient extension of the mean-shift paradigm to manifold-constrained multiple-mode tracking. Dedicated projections subsequently enable the estimation of the translation through three simple 1D density alignment steps that can be executed in parallel. An extensive evaluation on both simulated and publicly available real datasets comparing several existing methods demonstrates outstanding performance at low computational cost.

Yi Zhou, Laurent Kneip, Hongdong Li
Real Time Rotation Estimation for Dense Depth Senors in Piece-wise Planar Environments
Intelligent Robots and Systems (IROS), 2016 IEEE/RSJ International Conference on
[Abstract] [BibTex] [PDF] [Video]

Low-drift rotation estimation is a crucial part of any accurate odometry system. In this paper, we focus on the problem of 3D rotation estimation with dense depth sensors in environments that consist of piece-wise planar structures, such as corridors and office rooms. An efficient mean-shift paradigm is developed to extract and track planar modes in the surface normal vector distribution on the unit sphere. Robust and piecewise drift-free behavior is achieved by registering the bundle of planar modes from the current frame with respect to a reference frame using a general `1-norm regression scheme. We furthermore add a memory scheme to the regular birth and death of modes, which further compensates accumulated rotational drift when previously discovered modes are revisited. We discuss the robustness issue and evaluate our algorithm on both custom synthetic as well as real publicly available datasets. Our experimental results demonstrate high robustness and effectiveness of the proposed algorithm.

@inproceedings{zhou2016real,
  title={Real-time rotation estimation for dense depth sensors in piece-wise planar environments},
  author={Zhou, Yi and Kneip, Laurent and Li, Hongdong},
  booktitle={Intelligent Robots and Systems (IROS), 2016 IEEE/RSJ International Conference on},
  pages={2271--2278},
  year={2015},
  organization={IEEE}
  }
Yi Zhou, Laurent Kneip, Hongdong Li
A Revisit of Methods for Determining the Fundamental Matrix with Planes
Digital lmage Computing: Techniques and Applications, 2015 International Conference on (DICTA2015)
Note: The proof of the property 1 in this version is a little bit different from the conference version.
[Abstract] [BibTex] [PDF] [Code]
Determining the fundamental matrix from a collection of inter-frame homographies (more than two) is a classical problem. The compatibility relationship between the fundamental matrix and any of the ideally consistent homographies can be used to compute the fundamental matrix. Using the direct linear transformation (DLT), the compatibility equation can be translated into a least squares problem and can be easily solved via SVD decomposition. However, this solution is extremely susceptible to noise and motion inconsistencies, hence rarely used. Inspired by the normalized eight-point algorithm, we show that a relatively simple but non-trivial two-step normalization of the input homographies achieves the desired effect, and the results are at last comparable to the less attractive hallucinated points method. The algorithm is theoretically justified and verified by experiments on both synthetic and real data.

@inproceedings{zhou2015revisit,
  title={A Revisit of Methods for Determining the Fundamental Matrix with Planes},
  author={Zhou, Yi and Kneip, Laurent and Li, Hongdong},
  booktitle={Digital Image Computing: Techniques and Applications (DICTA), 2015 International Conference on},
  pages={1--7},
  year={2015},
  organization={IEEE}
  }
Laurent Kneip,Yi Zhou, Hongdong Li
SDICP: Semi-Dense Tracking based on Iterative Closest Points
The 26TH BRITISH MACHINE VISION CONFERENCE (BMVC2015), Swansea, UK
[Abstract] [BibTex] [PDF] [SDICP(BMVC)] [Video]
This paper introduces a novel strategy for real-time monocular camera tracking over the recently introduced, efficient semi-dense depth maps. We employ a geometric iterative closest point technique instead of a photometric error criterion, which has the conceptual advantage of requiring neither isotropic enlargement of the employed semidense regions, nor pyramidal subsampling. We outline the detailed concepts leading to robustness and efficiency even for large frame-to-frame disparities. We demonstrate successful real-time processing over very large view-point changes and significantly corrupted semi-dense depth-maps, thus underlining the validity of our geometric approach.

@inproceedings{BMVC2015_100,
  title={SDICP: Semi-Dense Tracking based on Iterative Closest Points},
  author={Laurent Kneip and Zhou Yi and Hongdong Li},
  year={2015},
  month={September},
  pages={100.1-100.12},
  articleno={100},
  numpages={12},
  booktitle={Proceedings of the British Machine Vision Conference (BMVC)},
  publisher={BMVA Press},
  editor={Xianghua Xie, Mark W. Jones, and Gary K. L. Tam},
  doi={10.5244/C.29.100},
  isbn={1-901725-53-7},
  url={https://dx.doi.org/10.5244/C.29.100}
}

Previous Work

  • Chao-Lei Wang, Tian-Miao Wang, Jian-Hong Liang, Yi-Cheng Zhang, Yi Zhou. Bearing-only Visual SLAM for Small Unmanned Aerial Vehicles in GPS-denied Environments. International Journal of Automation and Computing (IJAC), pp. 387-396, 2013.
  • Han Gao, Tianmiao Wang, Jianhong Liang, Yi Zhou. Model Adaptive Gait Scheme Based on Evolutionary Algorithm. Industrial Electronics and Applications (ICIEA), 2013 8th IEEE Conference on, pp. 316 - 321, 2013.
  • Chaolei Wang, Tianmiao Wang, Jianhong Liang, Yicheng Zhang, Yi Zhou. Research on monocular visual FastSLAM for a small unmanned helicopter. Chinese High Technology Letters, pp. 1061-1067, 2013.
  • Yi Zhou, Tianmiao Wang, Jianhong Liang, Chaolei Wang, Yicheng Zhang. Structural target recognition algorithm for visual guidance of small unmanned helicopters. Robotics and Biomimetics (ROBIO), 2012 IEEE International Conference on, pp. 908 - 913, 2012.
  • Yicheng Zhang, Tianmiao Wang, Jianhong Liang, Chaolei Wang, Yang Chen, Yi Zhou, Yubao Luan, Han Gao. An implement of RPV control system for small unmanned helicopters. Robotics and Biomimetics (ROBIO), 2012 IEEE International Conference on, pp. 1141 - 1145, 2012.

Honors & Awards

  • NCCR Fellowship Award for the research with ETH/UZH.
    The research was supported by the Swiss National Science Foundation through the National Center of Competence in Research (NCCR) Robotics, 2017~2018. NCCR LINK
  • Chinese Government Scholarship (CSC Scholarship), 2014
  • Outstanding Graduate Student of BUAA (Rank: 9/179), 2012

Academic Services

Nothing at this moment

Teaching & Tutoring

  • Mar 2015- Nov 2015Graduation Project Thesis: Visual Odometry on Android Mobile PhoneYifu Wang (U5434194,ANU)
  • Mar 2015- Nov 2015Graduation Project Thesis: Space Carving based 3D ReconstructionZhirui Wang (U5428281,ANU)
Links
Tools services webmasters counters generators scripts tutorials free