[visionlist] [Jobs] Fully funded PhD position on multimodal perception for UAV localization at ICB UMR CNRS laboratory (France) and JRL UMI CNRS/AIST (Japan) -- DL 24th of May 2024
Carlos Mateo Agullo
carlos-manuel.mateo-agullo at u-bourgogne.fr
Fri Apr 26 09:31:21 -05 2024
PhD position: Dynamic robot localization through event-based vision and
3D point cloud blending
#Location: 18 months in Tsukuba, Japon and 18 months in Dijon, France
#Host institutions: Institut Pascal
#Starting Date: October 1s, 2024
#Funding Duration: 36 months (3 years)
#Supervisors: Dr. HDR. Guillaume Caron, Dr. Carlos Mateo and Prof.
Cédric Demonceaux.
#Application Deadline: May 24th, 2024
#Context:
The VAI team of ICB laboratory at the UB in Dijon, France, is a partner
of the ANR PRCI EVELOC project. EVELOC focuses on providing new 3D
event-based visual localization methods in the field of robotics.
Event-based cameras are a novel bio-inspired type of sensor with low
latency and wide dynamic range. These sensors asynchronously activate
individual pixels to detect changes in light intensity, similar to
natural visual sensors. One of VAI’s most recent works in this area is
dedicated to object detection for autonomous vehicles [1].
The Joint Robotics Laboratory (JRL) at AIST in Tsukuba, Japan, is the
other partner of EVELOC. JRL is focused on 3D point cloud-based object
detection for robots. Especially, JRL is working on object pose tracking
tasks in dynamical conditions by directly aligning dense 3D point clouds
reconstructed from images of intensity variations calculated from events
[2]. Developed methods are aimed at manipulation tasks using humanoid
robots [3].
This thesis will leverage large-scale 3D environment point clouds [4] to
propose new efficient and robust algorithms for directly aligning events
captured by a moving camera on these 3D point clouds. It will take place
over 18 months in Japan, funded by AIST, then 18 months in France,
funded by ANR.
The student will implement and evaluate their research work first on
event-based cameras (Prophesee Gen 3 and 4) mounted on quadruped robots
(Unitree Alien Go) and bipedal robots (HRP-5P) provided by the
supervising team. Obtained results will be published in top-tier venues
(T-RO, IJCV, ICRA, IROS, ICCV, etc).
#Objective:
The goal is to localize a drone equipped with an event camera in an
environment described by a 3D point cloud. This topic is part of the ANR
PRCI EVELOC project, which focuses on the visual localization of robots
equipped with event cameras. It also contributes to strength the
collaboration between the ICB UMR CNRS laboratory (France) and the JRL
UMI CNRS/AIST (Japan).
#Research:
We are looking for one highly motivated PhD student to study multimodal
robot localization on high dynamic scenarios
The PhD student will focus on three open problems:
1. Event-based camera calibration;
2. Align 3D point cloud and event-based images;
3. Localize a drone in high dynamic scenarios
The recruited person will have access to all the necessary equipment to
implement their scientific contributions (computer, computing center,
cameras, etc.). They will start their thesis in Japan and continue after
18 months in France. The thesis is funded by the ANR EVELOC project
(50%) and Japanese funding (50%).
# References
[1] Z. Zhou, Z. Wu, R. Boutteau, F. Yang, C. Demonceaux and D. Ginhac.
RGB-Event Fusion for Moving Object Detection in Autonomous Driving. IEEE
International Conference on Robotics and Automation, 2023, London,
United Kingdom, pp. 7808-7815.
[2] Y. Kang, G. Caron, et al.. Direct 3D model-based object tracking
with event camera by motion interpolation. IEEE Int. Conference on
Robotics and Automation, May 2024, Yokohama, Japan.
[3] K. Chappellet, M. Murooka, G. Caron, F. Kanehiro, A. Kheddar.
Humanoid Loco-Manipulations using Combined Fast Dense 3D Tracking and
SLAM with Wide-Angle Depth-Images. IEEE Transactions on Automation
Science and Engineering, 2023.
[4] I. Ben Salah, S. Kramm, C. Demonceaux, P. Vasseur. Summarizing large
scale3D mesh for urban navigation. Robotics and Autonomous Systems, vol.
152, art. num. 104037, 2022.
#The ideal candidate must have:
1/ a master's degree or equivalent in computer science or another
relevant field,
2/ an excellent academic record,
3/ strong experience in robotics and/or computer vision,
4/ excellent skills in mathematics and coding (C/C++, Matlab, ROS, Python),
5/ excellent written and oral communication skills in English,
6/ enthusiasm for research, teamwork spirit, and the ability to solve
problems independently.
#Application:
Applicants must submit
1/ a one-page cover letter,
2/ curriculum vitae with publications list and contacts of 2 references,
3/ a copy of academic transcripts (bachelor/master grades),
4/ availability (the earliest feasible starting date).
Applicants must be prepared to provide two reference letters upon request.
Once we receive your application and it fits well for the position, you
will be contacted within two weeks.
Applications should be sent, *in a single PDF document*, with the email
subject [PhD application] to:
cedric.demonceaux at u-bourgogne.fr; guillaume.caron at u-picardie.fr;
carlos-manuel.mateo-agullo at u-bourgogne.fr
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://visionscience.com/pipermail/visionlist_visionscience.com/attachments/20240426/bbcae92b/attachment-0001.html>
More information about the visionlist
mailing list