Practical Course: Vision-based Navigation (6h SWS / 10 ECTS)
WS 2016/17, TU München
Please direct questions to firstname.lastname@example.org
Date & Location
Lecture & exercises (assignment phase) : Tuesdays, lectures approx. 2pm to 4pm in seminar room 02.09.023, tutoring of exercises approx. 4pm to 6pm in lab 02.05.014
Tutored lab time (project phase) : Tuesdays from 2pm to 6pm in lab 02.05.014 (other times for free project work available, tbd)
The course will start on 25.10.16 Tuesday at 14:00 in room 02.09.023.
The course will take place in our seminar room 02.09.023 and in our lab. In the beginning phase (4-5 weeks), there will be introductory lectures in our seminar room. Programming assignment sheets on basic problems will be handed out every week. In a second phase, the students will work in teams of 2-3 students on a practical problem (project). For the rest of the semester, the group meets weekly with their tutors and presents and discusses their progress. At the end of the course, the teams will present their project in a talk and demonstrate their solutions. They will document their project work in a written report. Both the assignments and the project part will be graded, and a final grade will be obtained from that.
For more details see Course Layout below.
- Good knowledge of the C/C++ language and basic mathematics such as linear algebra, analysis, and numerics is required
- Prior practical knowledge in CUDA programming, robotics, and computer vision topics is a plus
- Participation in at least one of the following lectures of the TUM Computer Vision Group: Variational Methods for Computer Vision, Multiple View Geometry, Autonomous Navigation for Flying Robots. Similar lectures can also be accepted, please contact us.
Number of participants: max. 12
Vision-based localization, mapping, and navigation has recently seen tremendous progress in computer vision and robotics research. Such methods already have a strong impact on applications in fields such as robotics and augmented reality.
In this course, students will develop and implement algorithms for visual navigation. For example, vision-based autonomous navigation for platforms such as wheeled robots and quadrocopters, or vision-based localization and mapping with handheld devices will be tackled. This includes, e.g., simultaneous localization and mapping with monocular, stereo, or RGB-D cameras, (semi-)dense 3D reconstruction, obstacle perception and avoidance, or autonomous path planning and execution.
- Lecture & Exercise (tba): 2 hours per week lecture session, Tuesdays from 2pm to 4pm. 2 hours per week tutored exercises, Tuesdays from 4pm to 6pm. There are 4-5 lecture & exercise sessions. Each week, the exercise for the following week will be announced and the exercise of the current week will be presented to tutors. The exercises must be done in groups of 2–3 students. The groups should be formed on the first lecture day. Students can use our lab computers in room 02.05.014. Attendance is mandatory.
- Project (tba): Each group will be assigned to a project. Students can work in the lab and consult the tutors on Tuesdays from 2pm to 6pm. Additional lab time for working freely can be arranged.
- Presentation and demo (tba): Each group will be assigned a time slot on one of the last days of the semester, to present their results and give a live demo, followed by a Q&A session.
- Project Report: Each group writes a report on their project work (10-12 pages, single column, single-spaced lines, 11pt font size).
- Multiple View Geometry, http://vision.in.tum.de/teaching/ss2016/mvg2016
- Autonomous Navigation for Flying Robots (EdX course), http://vision.in.tum.de/teaching/ss2014/autonavx
- Variational Methods for Computer Vision, https://vision.in.tum.de/teaching/ws2015/vmcv2015
- LSD-SLAM: Large-Scale Direct Monocular SLAM (J. Engel, T. Schöps, D. Cremers), In European Conference on Computer Vision (ECCV), 2014.
- Semi-Dense Visual Odometry for AR on a Smartphone (T. Schöps, J. Engel, D. Cremers), In International Symposium on Mixed and Augmented Reality, 2014.
- Visual-Inertial Navigation for a Camera-Equipped 25g Nano-Quadrotor (O. Dunkley, J. Engel, J. Sturm, D. Cremers), In IROS2014 Aerial Open Source Robotics Workshop, 2014.
- Collision Avoidance for Quadrotors with a Monocular Camera (H. Alvarez, L.M. Paz, J. Sturm, D. Cremers), In Proc. of The 12th International Symposium on Experimental Robotics (ISER), 2014.
Additional material can be downloaded from here.