Skip to content

willshw/mfvs

Repository files navigation

Visual Seroving for Object without Prior Model

Package Summary

This package combines object recognition, object tracking, point cloud processing and visual servoing to achieve visual servoing for objects without prior model. This pacakge can be used on robotic arm with RGBD sensor attached to the end effector.

Object Recognition --> Obejct Tracking --> Point Cloud Extraction -> Object Pose Estimation --> Visual Servoing

Object recognition and object tracking work together to enable reliable and uninterrupted tracking of the object of interested. With the bounding box of the tracked object in RGB image, program can extract the point cloud of the object from the structured point cloud from RGBD sensor using the boundng box from RGB image tracking.

At the beginning of a visual servoing session, a template of the object can be captured and cached for object pose estimation which can be estimated by ICP or particle filter from PCL. Position based visual servoing node can set a target pose relative to the estimated pose of the object and do visual servoing.

Package Dependencies

Hardware Used in Project

  1. Realsense D435 RGBD camera
  2. Kinova Mico arm

Topics and Nodes

How to launch package

Calibrate

  1. roslaunch arm_vs calibrate_mico_eye_in_hand.launch
  2. keep arUco tag fixed and move camera frame to capture tag at different orientation
  3. keep camera traslation at minimum and maximize camera rotation
  4. check calibration file under ~/.ros/easy_hand_eye/

Visual Servoing on arUco tag

  1. roslaunch arm_vs pbvs_aruco_tag.launch

Visual Servoing on object

  1. roslaunch arm_vs track.launch
  2. rosservice call /mico_interaction/switch_controllers "controllers_to_be_turned_on: 'velocity' to reset particle filter if pose estimation does not improve

About

model free visual servoing

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published