Practical Course: Vision-based Navigation IN2106 (6h SWS / 10 ECTS)
WS 18/19, TU München
Lecturers: Vladyslav Usenko, Nikolaus Demmel
Please direct questions to visnav_ws2018@vision.in.tum.de
TUMOnline course entry: https://campus.tum.de/tumonline/wbLv.wbShowLVDetail?pStpSpNr=950396181&pSpracheNr=1
Date & Location
Lecture & exercises (assignment phase) : Mondays, lectures approx. 2pm to 4pm (starting 2:00 sharp) in 02.05.014, tutoring of exercises approx. 4pm to 6pm in 02.05.014
Tutored lab time (project phase) : Mondays from 2pm to 6pm in lab 02.05.014 (other times for free project work available, tbd)
The course starts on Monday October 22nd, 2pm.
Course Structure
The course will take place in the lab room 02.05.014. In the beginning phase (4-5 weeks), there will be introductory lectures in room 02.05.014. Programming assignment sheets on basic problems will be handed out every week. In a second phase, the students will work in teams of 2-3 students on a practical problem (project). For the rest of the semester, the group meets weekly with their tutors and presents and discusses their progress. At the end of the course, the teams will present their project in a talk and demonstrate their solutions. They will document their project work in a written report. Both the assignments and the project part will be graded, and a final grade will be obtained from that.
For more details see Course Layout below.
Course Registration
Places assigned through TUM matching system.
Requirements:
- Good knowledge of the C/C++ language and basic mathematics such as linear algebra, analysis, and numerics is required
- Participation in at least one of the following lectures of the TUM Computer Vision Group: Variational Methods for Computer Vision, Multiple View Geometry, Autonomous Navigation for Flying Robots. Similar lectures can also be accepted, please contact us.
Number of participants: max. 12
Course Description
Vision-based localization, mapping, and navigation has recently seen tremendous progress in computer vision and robotics research. Such methods already have a strong impact on applications in fields such as robotics and augmented reality.
In this course, students will develop and implement algorithms for visual navigation. For example, vision-based autonomous navigation for platforms such as wheeled robots and quadrocopters, or vision-based localization and mapping with handheld devices will be tackled. This includes, e.g., simultaneous localization and mapping with monocular, stereo, or RGB-D cameras, (semi-)dense 3D reconstruction, obstacle perception and avoidance, or autonomous path planning and execution.
Course Layout
- Lecture & Exercise : 2 hours per week lecture session, Mondays from 2pm to 4pm. 2 hours per week tutored exercises, Mondays from 4pm to 6pm. There are 4-5 lecture & exercise sessions. Each week, the exercise for the following week will be announced and the exercise of the current week will be presented to tutors. The exercises must be done in groups of 2–3 students. The groups should be formed on the first lecture day. Students can use our lab computers in room 02.05.014. Attendance is mandatory.
- Project : Each group will be assigned to a project. Students can work in the lab and consult the tutors on Mondays from 2pm to 6pm. Attendance to meetings with tutors is mandatory. Additional lab time for working freely can be arranged.
- Presentation and demo : Each group will be assigned a time slot on one of the last days of the semester, to present their results and give a live demo, followed by a Q&A session. The presentation shall be 10 minutes long + 5 minutes questions.
- Project Report : Each group writes a report on their project work (10-12 pages, single column, single-spaced lines, 11pt font size; title page, table of content and references will not be accounted for in the page numbers).
Literature
Relevant courses:
- Computer Vision II: Multiple View Geometry, http://vision.in.tum.de/teaching/ss2018/mvg2018
The following book also covers many aspects. You should focus on Part II and III and selected background from Part I as needed:
- Timothy D. Barfoot, "State Estimation for Robotics", July 2017, Cambridge University Press
Free pdf available: http://asrl.utias.utoronto.ca/~tdb/bib/barfoot_ser17.pdf
Less relevant, but still helpful:
- Autonomous Navigation for Flying Robots (EdX course), http://vision.in.tum.de/teaching/ss2015/autonavx
- Computer Vision I: Variational Methods, http://vision.in.tum.de/teaching/ws2017/vmcv2017
Selected publications:
- Edward Rosten et al., "Faster and better: a machine learning approach to corner detection" (https://arxiv.org/pdf/0810.2434.pdf)
- Michael Calonder et al., "BRIEF: Binary Robust Independent Elementary Features" (https://infoscience.epfl.ch/record/149242/files/top_1.pdf
- Ethan Rublee et al., "ORB: an efficient alternative to SIFT or SURF" (http://www.willowgarage.com/sites/default/files/orb_final.pdf)
- Raúl Mur-Artal et al., "ORB-SLAM: A Versatile and Accurate Monocular SLAM System" (http://webdiis.unizar.es/~raulmur/MurMontielTardosTRO15.pdf)
- Ethan Eade, "Lie Groups for 2D and 3D Transformations" (http://ethaneade.com/lie.pdf) –> compare also chapter 7 in the Barfoot book mentioned above
Slides
Additional material can be downloaded from here.