/Aubo-Deep-Object-Grasping

Performing pick and place operations using deep learning algorithms GPD and YOLO with the Aubo Robotic Manipulator.

Primary LanguageC++

Aubo-Deep-Object-Grasping

A way to perform pick and place operations using deep learning algorithms GPD and YOLO.

Required Installations

The links to where one can install each program are provided.

Hardware

  • Asus Xtion Pro RGB-D Camera
  • Aubo I5 Robot
  • Robotiq 85 Gripper

Running the Programs to Pick Up Object

  1. Run the camera calibrated launch files in seperate terminals
   roslaunch jsk_pcl_ros openni2_local.launch
   roslaunch jsk_pcl_ros openni2_remote.launch
  1. In a new terminal run object detection through darknet_ros by command:
   roslaunch darknet_ros darknet_ros.launch
  1. In a new terminal run the program to filter out all point cloud data except the object that was chosen:
   rosrun my_aubo_i5_robot_config pass_through
  1. To find the transformation of the Aubo Robot base link with respect to the camera frame run:
   roslaunch static_frame.launch
  1. Launch the Aubo Driver in a new terminal:
   roslaunch aubo_i5_moveit_config moveit_planning_execution.launch sim:=false robot_ip:=your-robot-ip 
  1. Grasp the object by running in a new terminal:
   rosrun my_aubo_i5_robot_config pick_place

pick2 place

Inner Workings of the Algorithm

Camera Calibration and Running Camera Drivers

Camera calibration is a very important step. Follow the steps on how to calibrate a Monocular Camera with raw images over ROS(this is called Camera Intrisic Calibration). Once done with the tutorial provided on the ROS wiki, follow the Depth Camera Calibration tutorial.

When the camera calibration is finished we should have a depth error of at 1-2 cm between the the ros topics /camera_remote/depth_registered/points and /camera_remote_uncalibrated/depth_registered/points. The error can be visualized on rviz as such:

image

Open two terminals, in one terminal launch roslaunch jsk_pcl_ros openni2_local.launch, and in the other roslaunch jsk_pcl_ros openni2_remote.launch. This will run the Asus camera driver and load all the calibration files that were found in the camera calibration steps.

Running Aubo Driver

First make sure that the computer can connect to Aubo Robot. To do this go onto the Aubo ipad, go to network settings->tap ifconfig:

image

In the image above insert the inet addr number as shown in the highlighted section into the space bellow Network debuging (this is the robots ip address). Then press ping and it will show 0% packet loss.

Run the next command in the terminal to see if you can connect via the robot ip (in my case the ip is 192.168.1.101; this number will be different for everybody):

   ping 192.168.1.101

If it shows 0% packet loss in the terminal then the robot can be connected to via the robots ip.

To connect to the real robot through moveit in the same terminal run:

   roslaunch aubo_i5_moveit_config moveit_planning_execution.launch sim:=false robot_ip:=your-robot-ip 

Rviz will run with your robot model and the real robot will move just by dragging the endeffector arround and clicking Plan and execute in the Planning tab on Rviz.

Peek 2021-07-13 22-37

Cordinate Frames and Transformations

ROS provides a very useful package to keep track of multple coordinate frames over time called tf. The position of the robot base frame relative to the camera frame will be needed in order to perform any grasping. To do this, Aruco Tags will be used. First you stick the aruco tag onto the end-effector of the robot. Have the tag visible to the camera and run in a terminal:

   roslaunch aruco_ros single.launch 

To view the results of the aruco marker launch file run in a new terminal:

   rosrun image_view image_view image:=/aruco_single/result

image

The aruco marker will create the transformation from the camera frame to the marker frame. This is where the static_transform_publisher will be useful. We can then find the transformation from the end effector frame to the base frame of the robot to get its position relative to the camera frame. To do this edit node with name="robot_base_broadcaster" in the file static_frame.launch and insert the the x y z rz ry rx where the numbers are in args. These values can be found directly from the teach pendant on the Aubo ipad. Then run the command in a new terminal:

   roslaunch static_frame.launch

One can visualize the transformed frames in Rviz:

image

To find the transformation between the camera frame and robot base frame type in the a terminal:

   rosrun tf tf_echo camera_rgb_optical_frame base_link_calculated

Then create a node for static_transform_publisher in the static_frame.launch file using the values provided by the tf_echo command (translation[x, y, z] and Quaternion[x, y, z, w]):

   <node pkg="tf" type="static_transform_publisher" name="cam_broadcaster" args="   x  y  z  x  y  z  w  camera_rgb_optical_frame  base_link_calculated   100" /> 

Object Detection State of the Art (YOLOv4)

YOLOv4 will allow the computer to determine what objects are in the point cloud. Darknet Ros is a convenient way to get bounding boxe pixel coordinates that are published as a rostopic. Subscribing to the topic will allow easy access to continous bounding boxes even if the object is moved. To run YOLOv4 with ros run in a terminal the command:

roslaunch darknet_ros darknet_ros.launch

A new window will pop up with bounding boxes around the objects YOLO has been trained on. A full guide on how to train you own custom weights can be found at this GitHub repository.

image

Cloud Clustering and Centroids

The .cpp file passthrough.cpp filters out the point cloud. It uses a voxel filter, a statistical filter, and a table top filter. Then inorder to only obtain the objects I used a function called Euclidean Cluster Extraction. This will cluster each object, and we can compute the 3D centroids of each object, then convert the centroids into 2D. The 2D Centroids will be compared to the center point of the bounding box from the rostopic(/darknet_ros/bounding_boxes). The object that is selected will have the shortest euclidean distance between the bounding box center point and the 2D centroid.

image

To run the point cloud clustering and object selection; in a new terminal run:

   rosrun my_aubo_i5_robot_config pass_through

This will allow you to enter the object name you want to cluster. To visualize the point cloud in rviz, add the topic /output_cloud_filtered. You should see something similar to the images bellow.

image image

GPD_ROS

We will be inserting the point cloud that is being published from pass_through.cpp on the rostopic names /output_cloud_filtered.

Run in a new terminal:

   roslaunch gpd_ros ur5.launch

This will generate the best grasps on the selected object like in the image bellow.

image

Picking Up Object

With the Aubo Driver running from the command roslaunch my_aubo_i5_robot_config moveit_planning_execution.launch robot_ip:=<your_robot_ip> in a new terminal run:

   rosrun my_aubo_i5_robot_config pick_place

The real robot will then try to go to the grasp pose by finding ik solutions. If it says failed-No IK_solutions found just restart pick_place until it works. If you have tried to restart it more than five times you may just have an unreachable grasp for your robot and will have to find another grasp that is reachable.

An error with the robot controller may occur that says:

   [ERROR] [1415027195.809289492]: Controller is taking too long to execute trajectory (the expected upper bound for the trajectory execution was 4.992567 seconds). Stopping trajectory.

To fix this issue go into the file trajectory_execution.launch. Here, change value="1.2" -> value="3" in the parameter:

   <param name="trajectory_execution/allowed_execution_duration_scaling" value="1.2"/>

Keep increasing the value until the error doesn't occur and the robot motion will be smoother as well.

Installation Possible Problems and their Solutions

Camera Calibration

In order to get the ROS driver for the Asus Xtion Camera run the two commands:

   sudo apt-get install ros-kinetic-openni-launch
   sudo apt-get install ros-kinetic-openni2-launch

So that openni can see the Asus camera; unncomment the line UsbInterface = 2 in the file Globaldefault.ini. This file cannot be editted by simply clicking on the files icon, you will need to run with root privileges:

   sudo nano /etc/openi/Globaldefault.ini

To connect to the camera run command:

   roslaunch openni2_launch openni2.launch

Aubo_Driver

Make sure to install the ros controllers and controller manager:

   sudo apt-get install ros-kinetic-ros-controllers 
   sudo apt-get install ros-kinetic-controller-manager 

One error when trying to launch the Aubo_Driver was:

   error while loading shared libraries: libmoveit_robot_trajectory.so.0.9.15: cannot openshared object file: No such file or directory

The above error means that some of the moveit libraries are not seen by the Aubo driver. To fix this run the following commands in the terminal(make sure to run each command in order):

   cd /opt/ros/kinetic/lib 

   sudo cp -r libmoveit_robot_trajectory.so.0.9.17 .. 

   sudo cp -r libmoveit_robot_model.so.0.9.17 .. 

   sudo cp -r libmoveit_robot_state.so.0.9.17 .. 

   cd .. 

   sudo mv libmoveit_robot_state.so.0.9.17 libmoveit_robot_state.so.0.9.15 

   sudo mv libmoveit_robot_model.so.0.9.17 libmoveit_robot_model.so.0.9.15 

   sudo mv libmoveit_robot_trajectory.so.0.9.17 libmoveit_robot_trajectory.so.0.9.15 

   sudo cp -r libmoveit_robot_state.so.0.9.15 lib/ 

   sudo cp -r libmoveit_robot_model.so.0.9.15 lib/ 

   sudo cp -r libmoveit_robot_trajectory.so.0.9.15 lib/

If you have a different robot from the Aubo Robot used in this repository I suggest looking at the moveit setup assistant. Here you just enter your robots urdf and follow the instructions, and then you can run the exact same steps but with a different robot arm.

Darknet_ros

In order to run git clone with the darknet_ros repository you have to create a ssh key for your GitHub account, otherwise you will get an error saying you do not have permission to download the repository.

I had a little trouble figuring out how to set up the ssh key so here were my steps. Run in a terminal:

   ssh-keygen

Press ENTER to save the key to the default location. Should be at /.ssh/id_rsa.pub. Enter a password or not... it's up to you and then press ENTER again. This should set up the key in the default location. To view the key run:

   cat .ssh/id_rsa.pub   

Copy the generated key from your terminal then go to your GitHub profile. Follow the links: Settings -> SSH and GPG keys. Click "New ssh key", paste the ssh key into the key field. Click "Add key". Now you should be able to clone the darknet ros repository with no problems.

If training a YOLO model on custom objects is necessary follow this guide. I ran into an issue where opencv couldn't find my images. To fix this, I ran the program preprocess.py with the command:

   python preprocess.py   

This will create a test.txt file with 10% of the images and 90% of the images will be put into a train.txt file. This should fix the error with opencv and also save you a lot of time of copying and pasting hundreds of paths in the train.txt and test.txt files.

Training will take a very long time to run if done with CPU only. This is why training on GPU is pretty much a requirement. To do this I downloaded CUDA 10.2 and Cudnn 7.6.5. The downloading is a little tricky and a tutorial can be found here. The only issue that I ran into when installing CUDA was when trying to run make in the directory /cuddn_samples_v7/mnistCUDNN I recieved the error:

   /usr/local/cuda/include/cuda_runtime_api.h:0110:60: error: 'pGraphExec' was not declared in this scope extern __host__ cudaError_t CUDARTAPI cudaGraphInstantiate{cudaGraphExec_t *pGrapheExec, cudaGraph_t graph, cudaGraphNode_t *pErrorNode, char *pLogBuffer...  
   
   /usr/local/cuda/include/cuda_runtime_api.h:0110:60: error: 'CudaGraph_t' was not declared in this scope extern __host__ cudaError_t CUDARTAPI cudaGraphInstantiate{cudaGraphExec_t *pGrapheExec, cudaGraph_t graph, cudaGraphNode_t *pErrorNode, char *pLogBuffer...  
   
   /usr/local/cuda/include/cuda_runtime_api.h:0110:60: error: 'CudaGraphNode_t' was not declared in this scope extern __host__ cudaError_t CUDARTAPI cudaGraphInstantiate{cudaGraphExec_t *pGrapheExec, cudaGraph_t graph, cudaGraphNode_t *pErrorNode, char *pLogBuffer...  

To fix this error open cudnn.h file that should be located in path /usr/include/cudnn.h. To edit run command:

sudo gedit /usr/include/cudnn.h   

Then change the line #include "driver_types.h" to #include <driver_types.h>. Now make should run fine.

You may run into an issue of opencv being to low as python 2.7 is most likely being used. For example it was showing opencv-dev 3.34. I installed opencv from source, then in order for python 2.7 to see the correct version of opencv I had to run the command:

   sudo cp /home/nicholas/opencv-4.5.2/build/lib/cv2.so /opt/ros/kinetic/lib/python2.7/dist-packages/   

GPD and gpd_ros

GPD needs to be installed becuase it is required for gpd_ros as a library. Gpd_ros is just a ros wrapper for GPD. When installing GPD in order to successfully make the project, go to the CMakeList.txt file and comment out the line:

   set(CMAKE_CXX_FLAGS "-O3 -march=native -mtune=intel -msse4.2 -mavx2 -mfma -flto -fopenmp -fPIC -Wno-deprecated -Wenum-compare -Wno-ignored-attributes -std=c++17")

Uncomment the line:

   set(CMAKE_CXX_FLAGS "-fopenmp -fPIC -Wno-deprecated -Wenum-compare -Wno-ignored-attributes -std=c++17")

When making GPD I suggest having pcl version 1.9 installed from source. If done correctly, the CMakeList.txt will find your version of PCL. In the GPD CMakeList.txt I changed the line find_package(PCL 1.8 REQUIRED) -> find_package(PCL 1.9 REQUIRED). A common error I got was that the CMakeList couldn't find the PCLConfig.cmake file. This can be fixed by added the following line above the find_package line that was changed to: set(PCL_DIR "/path/to/PCL/1.9.1/build/PCLConfig.cmake). You may have to do the same thing with the opencv installation as sometimes the CMakeList will not be able to find the OpenCvConfig.cmake file. This can be fixed by using the same function as used for pcl: set(OpenCV_DIR "/path/to/opencv/build/OpenCVConfig.cmake").

GPD_ROS can be ran using the command roslaunch gpd_ros ur5.launch, it will not work right away unfortunately.

In the ur5.launch file change the cloud_type to the point cloud type that will be used. Then change the cloud_topic to the correct rostopic that is needed to be used. The config_file value needs to be changed from:

   <param name="config_file" value="/home/ur5/projects/gpd/cfg/ros_vino_params.cfg" />

To:

   <param name="config_file" value="/path/to/gpd/cfg/ros_eigen_params.cfg" />

The file ros_eigen_params.cfg (path to .cfg file = ~/gpd/cfg/ros_eigen_params.cgf) is where most of the finctionality parameters are, such as setting the gripper dimensions, grasp workspaces, how many grasp samples are considered, etc. The only thing that needs to be changed in order for the launch file to work is the line:

   weights_file = /home/andreas/projects/gpd/lenet/15channels/params/

To:

   weights_file = /path/to/gpd/lenet/15channels/params/

If GPD is running slow on producing grasps for you robot, I suggest lowering the parameter num_samples = 500 to a number lower than 100 and change the parameter workspace_graps = -1 1 -1 1 -1 1 to fit exactly the areas you need to grasp.

Moveit Tutorials

Here are some helpful Moveit tutorials:

ROS Tutorials

Here are some helpful ROS tutorials: