-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
6052038
commit 99cd29e
Showing
1 changed file
with
88 additions
and
1 deletion.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1 +1,88 @@ | ||
# home-manipulation | ||
# home-manipulation | ||
|
||
## Description | ||
This repository contains the ROS workspace and tools needed to run manipulation tasks for the RoBorregos @Home robot. It includes the robot description with URDF and mesh files for the most recent robot including a EAI Dashgo B1 mobile base and a UFactory xArm6. It also contains the MoveIt configuration and packages required for picking, placing and manipulating objects, including 2D and 3D object detector, pointcloud processing and the manipulation pipelines for diverse tasks. | ||
|
||
## Index | ||
|
||
- [Installation](#installation) | ||
- [Docker Installation](#docker-installation) | ||
- [Usage](#usage) | ||
|
||
## Installation | ||
|
||
To run the repository, ROS1 Noetic is required. To install ROS Noetic, follow the instructions in the [official ROS website](http://wiki.ros.org/noetic/Installation/Ubuntu). | ||
|
||
To install ros dependencies from the workspace, run the following commands: | ||
```bash | ||
cd ws/ | ||
rosdep update | ||
rosdep install --from-paths src --ignore-src -r -y | ||
``` | ||
|
||
Additional dependencies are required for the xArm submodules and the Grasping Pose Detector. For xArm, follow the instructions in the [xArm ROS repository](https://github.com/xArm-Developer/xarm_ros#3-preparations-before-using-this-package), and for the Grasping Pose Detector, follow the instructions in the [Grasping Pose Detector repository](https://github.com/atenpas/gpd#install). For GPD, install eigen and pcl with the following commands: | ||
```bash | ||
sudo apt update && apt install -y libeigen3-dev libpcl-dev | ||
``` | ||
|
||
## Docker Installation | ||
The repository contains a Makefile and scripts for running the workspace in a Docker container. It also allows for use with CUDA (which helps in running the object detection YOLO models). | ||
|
||
### Docker Pre-requisites | ||
- [Docker Engine](https://docs.docker.com/engine/install/ubuntu/#install-using-the-repository) | ||
- [Post-installation steps for Linux](https://docs.docker.com/engine/install/linux-postinstall/) | ||
If using GPU: | ||
- NVIDIA Driver | ||
- [NVIDIA Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/index.html)) | ||
|
||
### Image and Container Creation | ||
To create the Docker image, run the following command: | ||
```bash | ||
make manipulation.build # for CPU | ||
make manipulation.build.cuda # for GPU | ||
``` | ||
|
||
To create the Docker container, run the following command: | ||
```bash | ||
make manipulation.create # for CPU | ||
make manipulation.create.cuda # for GPU | ||
``` | ||
|
||
### Running the Container | ||
To run the Docker container, run the following command: | ||
```bash | ||
make manipulation.up | ||
make manipulation.shell # for GPU | ||
``` | ||
|
||
### Removing the Container | ||
To remove the Docker container, run the following command: | ||
```bash | ||
make manipulation.down | ||
make manipulation.remove | ||
``` | ||
|
||
## Usage | ||
|
||
### Robot Description | ||
The robot description is located in the `ws/src/robot_description` package. It contains the URDF and mesh files for the robot, including the mobile base and arm. After sourcing the workspace, visualize the robot in RViz with the following command: | ||
```bash | ||
roslaunch robot_description display.dashgo.launch | ||
``` | ||
|
||
### MoveIt Configuration | ||
To turn on the xArm6 and allow the robot to make movements with MoveIt planning and other planning and movement options, run the following command: | ||
```bash | ||
roslaunch dashgo_moveit_config dashgo_xarm.launch robot_ip:=<robot_ip> | ||
``` | ||
|
||
### Manipulation Pipelines | ||
The current, most recent and stable manipulation pipelines are located in the `ws/src/pick_and_place` package. The pipelines include the object detection, pointcloud processing and manipulation tasks. To run the manipulation pipeline, run the following command: | ||
```bash | ||
roslaunch pick_and_place cartesian.manipulation.launch | ||
``` | ||
|
||
This command also requires the ZED2 camera, or any other camera that publishes a pointcloud in the `/zed2/zed_node/point_cloud/cloud_registered` topic. The camera can be opened with the following command from the [zed-ros-wrapper](https://github.com/RoBorregos/zed-ros-wrapper) repository in the machine where the camera is connected: | ||
```bash | ||
roslaunch zed_wrapper zed2_robot.launch | ||
``` |