The project is aimed to allow a robot to anticipate human intentions while interaction partners are still performing their action. Specifically, two actions are anticipated from the UR5 robot. When the human decides for an object to pick and reaches to it, the robot arm must pick the other object. This way, a feeling of cooperation between the interaction partner and the robot will be created.
One Ubuntu 16.04 LTS machine and a Windows 10 machine is needed.
- ROS-kinetic: http://wiki.ros.org/kinetic/Installation/Ubuntu
- Catkin workspace: http://wiki.ros.org/ROS/Tutorials/InstallingandConfiguringROSEnvironment
- ROS Industrial Universal Robots: https://github.com/ros-industrial/universal_robot
- Visual Studio 2017: https://docs.microsoft.com/en-us/visualstudio/ide/?view=vs-2017
- Xbox Kinect v2 sensor
- Kinect to PC adaptor
-
Get ROS kinetic on Linux 16.04
-
Follow instructions to create catkin workspace
-
Install ROS Industrial Universal Robots
Open terminal and execute:
sudo apt-get install ros-kinetic-universal-robot
- Download files from my GIT repository
File 'listener.py' is needed into the Ubuntu machine inside catkin enviroment, into /src folder.
-
Download Visual Studio 2017 on Windows Machine
-
Connect Xbox Kinect v2 to the Windows Machine using an adaptor
-
On Ubuntu Machine open terminal and run
hostname -I
- Go to BodyBasics-WPF folder on Windows
Certain changes are needed to use the application.
Open MainWindow.xaml.cs using Visual studio or any other text editor.
Go to line 200, change the IP, to the IP of the Ubuntu Machine obtained in step 7.
The application is ready to be simulated using Gazebo simulator.
On Ubuntu Machine side:
Make sure for every terminal/shell you are directed to the catkin enviroment.
Example:
cd catkin_ws
and then source the setup file for your shell using the command:
source devel/setup.bash
- Open terminal and run Gazebo simulation
roslaunch ur_gazebo ur5.launch
- Open new terminal for setting up the MoveIt! nodes to allow motion planning
roslaunch ur5_moveit_config ur5_moveit_planning_execution.launch sim:=true
- Open new terminal for starting up RViz with a configuration
roslaunch ur5_moveit_config moveit_rviz.launch config:=true
- Open new terminal to run listener node
cd src
python listener.py
On Windows Machine side:
Open the BodyBasics-WPF folder and then open the BodyBasics-WPF.sln application using Visual Studio 2017 and click start.
The photos were obtained in the Advanced Robotics at Queen Mary using the UR5 robot
The video was obtained in the Advanced Robotics at Queen Mary using the UR5 robot