Skip to content

Latest commit

 

History

History
200 lines (152 loc) · 12.4 KB

README.md

File metadata and controls

200 lines (152 loc) · 12.4 KB

Pretrained DLCHOMP Networks for Manipulator Motion Planning

This repository provides pretrained networks for Deep-Learning-Based Covariant Hamiltonian Optimization for Motion Planning (DLCHOMP) of robotic manipulators for MATLAB®. These pretrained networks can output intermediate trajectory guesses for desired start to goal configurations in a given spherical obstacle environment. Open in MATLAB Online

DLCHOMP High Level Visualization

Creator: MathWorks Development

Includes un-trained model: ❌

Includes transfer learning script: ✅
(See this page)

Supported Robots: ✅
(See this page)

Requirements

  • MATLAB® R2024a or later
  • Robotics System Toolbox™
  • Deep Learning Toolbox™

Getting Started

Download or clone this repository to your machine and open it in MATLAB®.

Setup

Add path to the source directory.

addpath("src");

Download Pretrained Network

Use the code below to download the pretrained network for a supported robot. For a list of supported robots, see the table in the Metrics and Evaluation section. To demonstrate this workflow, download the pretrained DLCHOMP network for the KUKA LBR iiwa 7.

robotName = "kukaIiwa7";
data = helper.downloadPretrainedDLCHOMPForRobot(robotName);

Summarize Pretrained Network

Extract the pretrained network. You can summarize and analyze the network as needed to better understand the architecture of the network.

pretrainedNetwork = data.trainedNetwork;
summary(pretrainedNetwork);
analyzeNetwork(pretrainedNetwork);

If you only need the pretrained network and not the pretrained optimizer, skip to the DLCHOMP Details section. Otherwise, continue to obtain a pretrained optimizer.

Obtain Pretrained DLCHOMP Optimizer

Extract the pretrained optimizer.

pretrainedDLCHOMP = data.trainedDLCHOMP;

Predict Trajectory Using Pretrained DLCHOMP Optimizer

Use the pretrained optimizer to predict a trajectory between a start joint configuration and goal joint configuration in an obstacle environment.

% Extract obstacle environment that was used to train the DLCHOMP optimizer.
pretrainedDLCHOMP.SphericalObstacles = data.unseenObstacles;

% Predict a trajectory using the optimizer.
[optimWpts,optimTpts,solinfo] = optimize(pretrainedDLCHOMP,data.unseenStart,data.unseenGoal);

% Visualize results.
show(pretrainedDLCHOMP,optimWpts);

DLCHOMP Output Prediction

Create and Train DLCHOMP Optimizer for New Applications

To generate data and train a DLCHOMP optimizer to suit your application or task, follow the Train Deep-Learning-Based CHOMP Optimizer for Motion Planning example.

Train Custom DLCHOMP Optimizer Using Transfer Learning

Transfer learning enables you to adapt a pretrained DLCHOMP optimizer to your dataset. Follow these examples to create a custom DLCHOMP optimizer and train it for transfer learning to:

DLCHOMP Details

You can enhance the efficiency of optimization-based motion planning tasks by applying deep learning [1]. dlCHOMP is one such MATLAB® feature that utilizes a neural network initial guesser to provide an educated initial guess for a robot's intermediate start to goal trajectory, which is then optimized using the Covariant Hamiltonian Optimization for Motion Planning (CHOMP)[2] algorithm.

DLCHOMP Overview

  • Env: Spherical obstacle environment in which robot motion planning is to be performed. This is provided as a 4-by-N numeric input matrix to dlCHOMP. This input is the fed to the CHOMP and BPS Encoder modules.
  • BPS Encoder: Basis point set encoder for the obstacle environment. The basis point set is a set of fixed points that are used to convert an arbitrary size obstacle environment into a fixed-size encoding vector. This encoding vector is then fed as an input to the Pretrained Network module along with the desired start and goal configurations of the robot.
  • Pretrained Network: Feed-forward neural network that guesses an initial intermediate trajectory for a robot by taking a start configuration, an end configuration, and an obstacle environment encoding vector as inputs. An intermediate trajectory is a trajectory that does not include the start and goal configurations. More details on the architecture of this neural network can be seen in the Neural Network Details section below.
  • CHOMP: Motion planning optimizer that uses the Covariant Hamiltonian Optimization for Motion Planning algorithm [2]. It takes the initial intermediate trajectory guess output of the Pretrained Network and the spherical obstacle environment Env as its inputs to then output an optimized start to goal trajectory.

Neural Network Details

This figure shows the architecture of the DLCHOMP neural network [2].

DLCHOMP Network Architecture

It takes a given motion task (a world obstacle encoding vector WB, a start configuration q1 and an end configuration qNt) to output an initial guess Q. Blocks of tapered Fully Connected Layers (gray) are combined like the DenseNet architecture [3] via skip-connections and concatenations (circular nodes).

Metrics and Evaluation

The test dataset for each pretrained network consists of 1000 data samples, identical to the validation dataset created during the network training phase. To augment the data, these 1000 test samples were flipped, capitalizing on the symmetric nature of the motion planning problem, resulting in a total of 2000 test data samples. The results in these tables are for these 2000 test data sample sets that were created for each robot.

Size and Accuracy Metrics

Header Definition
DLCHOMP Optimizer Name of the supported robot for whom the metrics are being listed in the current row. This name is a short-hand name used to quickly identify each robot. To obtain the full robot name, and hence determine the exact robot model, see the [Pretrained Optimizers](https://www.mathworks.com/help/releases/R2024a/robotics/ref/dlchomp.html#mw_93957c22-f6cc-4e8c-ac45-1ae16cfcb2ef) section of the dlCHOMP page.
Size (MB) Memory footprint of the DLCHOMP optimizer object in megabytes.
% of samples with DLCHOMP Itns < CHOMP Percentage of data samples where the dlCHOMP optimizer took lesser number of iterations than an equivalent manipulatorCHOMP optimizer with similar optimization options.
Mean % of Itns Saved The mean percentage of iterations saved by the dlCHOMP optimizer for the data samples where the dlCHOMP optimizer took lesser iterations than the equivalent manipulatorCHOMP optimizer with similar optimization options.
% of samples with DLCHOMP Inference Time < CHOMP Percentage of data samples where the dlCHOMP optimizer's optimization time was lesser than that of an equivalent manipulatorCHOMP optimizer with similar optimization options.
Mean % of Inference Time Saved The mean percentage of inference time saved by the dlCHOMP optimizer for the data samples where the dlCHOMP optimizer took lesser iterations than that of an equivalent manipulatorCHOMP optimizer with similar optimization options. Inference time of a dlCHOMP optimizer is the sum of the network guess time and subsequent the optimization time. Inference time of a manipulatorCHOMP optimizer is the same as its optimization time since it does not have a neural network component.
Feasibility The percentage of test data samples where the dlCHOMP optimizer gave a collision free optimized trajectory.

The table above defines the headers present in the table below:

DLCHOMP Optimizer Size (MB) % of samples with DLCHOMP Itns < CHOMP Mean % of Itns Saved % of samples with DLCHOMP Inference Time < CHOMP Mean % of Time Saved Feasibility
abbYuMi 25 84.70 78.36 78.20 73.95 74.50
fanucLRMate200ib 25 87.20 80.42 79.10 72.30 78.65
fanucM16ib 25 75.30 82.17 67.40 76.09 73.95
frankaEmikaPanda 25 84.40 83.10 77.60 77.93 77.25
kinovaJacoJ2S7S300 25 99.00 78.63 74.00 75.07 68.55
kinovaGen3 25 77.90 74.56 63.70 67.44 72.15
kukaIiwa7 25 83.80 79.02 74.90 72.38 80.40
meca500r3 25 85.00 79.24 74.90 71.41 65.15
techmanTM5-700 25 78.40 74.49 67.40 66.38 71.20
universalUR5e 25 73.60 76.44 62.20 70.21 71.05

CPU Time Metrics

Header Definition
DLCHOMP Model Without Codegen Name of the supported robot for whom the metrics are being listed in the current row. These metrics were computed in MATLAB without using its code generation feature. This name is a short-hand name used to quickly identify each robot. To obtain the full robot name, and hence determine the exact robot model, [Pretrained Optimizers](https://www.mathworks.com/help/releases/R2024a/robotics/ref/dlchomp.html#mw_93957c22-f6cc-4e8c-ac45-1ae16cfcb2ef).
Mean Network Guess Time (secs) The mean time taken by the dlCHOMP optimizer to obtain its neural network's intermediate guess trajectory in seconds.
Mean Inference Time (secs) The mean of the total time taken by the dlCHOMP optimizer to obtain its neural network's intermediate guess trajectory and then optimize it using CHOMP, in seconds.

The table above defines the headers present in the table below:

DLCHOMP Model without Codegen Mean Network Guess Time (secs) Mean Inference Time (secs)
abbYuMi 0.0072 19.2465
fanucLRMate200ib 0.0100 0.6899
fanucM16ib 0.0069 1.2242
frankaEmikaPanda 0.0075 1.6912
kinovaJacoJ2S7S300 0.0098 4.4300
kinovaGen3 0.0072 3.0774
kukaIiwa7 0.0060 1.5289
meca500r3 0.0057 0.5911
techmanTM5-700 0.0052 1.2719
universalUR5e 0.0075 1.6614

Note: Dual-arm YuMi® (abbYuMi) being the only two-armed robot in this list, takes much longer for optimization as compared to other robots due to the higher probability of self-collisions.

References

[1] J. Tenhumberg, D. Burschka and B. Bäuml, "Speeding Up Optimization-based Motion Planning through Deep Learning," 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 2022, pp. 7182-7189, doi: 10.1109/IROS47612.2022.9981717.

[2] N. Ratliff, M. Zucker, J. A. Bagnell and S. Srinivasa, "CHOMP: Gradient optimization techniques for efficient motion planning," 2009 IEEE International Conference on Robotics and Automation, 2009, pp. 489-494, doi: 10.1109/ROBOT.2009.5152817.

[3] S. J´egou et al., “The one hundred layers tiramisu: Fully convolutional densenets for semantic segmentation,” CoRR, vol. abs/1611.09326, 2016

Copyright 2024 The MathWorks, Inc.