Skip to content

umigv/cv-nav-integration

Repository files navigation

TODO:

  • Test on Jetson
  • Test with Zed
  • Clean up Code

Current:

  • script runs YOLOP model on saved video + creates a replication with black pixels denoting non-drivable area
  • established bare bones subscriber + publisher nodes

Discuss with Chris:

  • pubish rate (0.5 seconds) -- leave as a parameter
  • how to publish occupancy grid as message (combining publisher method file and drivable lane .py) --- nav occupancy grid message (contains a header -- timestamp etc., data -- as an array)
  • size of pixels

Next Steps:

  • create custom message to publish occupancy grid, robot (x,y) position, and grid size
  • Figure out how to parse ZED input to just recieve left frame
  • how to listen to Zed node (left camera)
  • judge distortion of tape measure in testing images for different angles (perspective transform)

Goals:

  • convert current "replication frame" to occupancy grid via aerial transform, search for zed2i guides
  • listen to ZED nodes + create lane_detection node with publishing

Notes:

Goal: CV output to Nav an occupancy grid of non-drivable area for each frame. Each frame's occupancy grid are ultimately combined to form a "world occupancy grid".

Challenges: computational efficiency + converting local frame to global frame where the relative position of the local origin is unknown.

Resources: https://docs.google.com/document/d/1lL_PE1D-wrfgzJdFESzRF_aXDcSKWAU45HWH284obUQ/edit?usp=sharing

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages