From MAE/ECE 148 - Introduction to Autonomous Vehicles
Jump to navigation Jump to search

Team Members

  • Andrew Liang (ECE)
  • Jiansong Wang (MAE)
  • Shane Benetz (ECE)
  • Kevin Kruse (Extension/BIS)

Project Overview/Proposal

The goal of this project is the integration of a Lidar in the existing Framework of the ROS Navigation system.

Robot schematics and pictures

Lap videos

Project Schedule / Gantt Chart

Software Development

Calibrating the existing ROS framework

One especially important thing for our project is a properly calibrated framework.

To achieve this, you can either adjust the controls to make it fit. However, an alternative and more precise solution is to apply a specific color filter. To do this, we first started a live transfer of the current image. Bild1.png Once we had the image, we opened a color tweezer using Microsoft Word and detected the color of the yellow center track. Bild2.png

The program now gives us the color in RGB color space. Using a converter, e.g. https://www.peko-step.com/en/tool/hsvrgb_en.html we can now convert this color code into the HSV color code used by our OpenCV module. Finally we just have to configure this filter correctly and add some tolerances, because especially in twilight the colors vary. Once this process is complete, we get a perfectly calibrated system, as shown in the image below. Bild3.png

Software Subsystems


Links to additional resources (presentations/source code/GitHub)


  • Jack Silberman (Professor)
  • Dominic Nightingale (TA)
  • Haoru Xue (TA)