iLSPR: A Learning-based Scene Point-cloud Registration Method for Robotic Spatial Awareness in Industrial Manufacturing
This repository contains supporting materials of the paper "iLSPR: A Learning-based Scene Point-cloud Registration Method for Robotic Spatial Awareness in Industrial Manufacturing", by Yusen Wan and Xu Chen, Department of Mechanical Engineering, University of Washington, 3900 E Stevens Way NE, Seattle, 98195, Washington, USA.
This project includes the following items:
- the dir named "geometric primitives" stores the geometric primitives used in GPDG method.
- the dir "digital model library" stores the CAD models of the digital model library.
- the dir named "ISOPR" store the ISOPR dataset and corresponding documents.
- the dir named "Images" store some example images.
All the CAD models are in ".stl" format. The ISOPR Dataset is a pkl file.
- Introduction
- Industrial Scene Object Point-cloud Registration (ISOPR) dataset
- Geometric Primitives
- Digital Model Library
A critical capability for intelligent manufacturing is the ability for robotic systems to understand the spatial operation environment – the ability for robots to precisely recognize and estimate the spatial positions and orientations of objects in the industrial scenes. Existing scene reconstruction methods are designed for general settings with low precision needs of objects and abundant data. However, manufacturing hinges on high object precision and operates with limited data. Addressing such challenges and limitations, we propose a novel learning-based Scene Point-cloud Registration framework for automatic industrial scene reconstruction (iLSPR). The proposed iLSPR framework leverages point cloud representation and integrates three key innovations: (i) a Multi-Feature Robust Point Matching Network (MF-RPMN) that learn from both raw data and deep features of the objects to accurately align point clouds, (ii) a Geometric-Primitive-based Data Generation (GPDG) method for efficient synthetic data generation, and (iii) a digital model library of industrial target objects. During operation, vision sensors capture point clouds in the scenes, and the iLSPR method registers high-fidelity object models in the scenes using MF-RPMN, pre-trained with GPDG-generated data. We introduce an Industrial Scene Object Pointcloud Registration (ISOPR) dataset in IsaacSim to benchmark performance. Experimental results demonstrate that iLSPR significantly outperforms existing methods in accuracy and robustness. We further validate the approach on a real-world robotic manufacturing system, demonstrating reliable digital reconstruction of industrial scenes.
With the rise of robotic manufacturing, vision-based scene reconstruction methods have become essential, high-precision reconstruction of manufacturing scenes, particularly target workpieces, has emerged as a cornerstone of robotic intelligence. In the manuscript, we proposed a Learning-based Scene Point-cloud Registration Method for automatic industrial scene reconstruction, which leverages point clouds as 3D representation and can estimate the positions and orientations of workpieces in the industrial scenes. For the experimental evaluation, we conducted Industrial Scene Object Point-cloud Registration (ISOPR) dataset. Because, though there are large datasets (ModelNet40 [1], ABC [2], and so on) for benchmarking, they lack simulation of authentic depth camera sampling, which means the experimental formulations in the previous literature are unsuitable.
To simulate the depth camera and generate sensor-realistic point clouds for industrial scene pointcloud registration, we construct the ISOPR Dataset using NVIDIA Isaac Sim simulator, a robotics simulation platform developed on the NVIDIA Omniverse framework.
In IsaacSim, We built a robitc manufacturing scene, which consists of a Universal Robot Arm UR10, a semi-enclosed rectangular platform, a depth camera provided by Isaac Sim fixed on the platform, and the workpiece (object), as shown in Figure 2.. In ISOPR, the resolution of the depth camera is 640×480 pixels, and the horizontal field of view is 30°. The workpieces are selected from the digital model library. Sample workpieces, including various connectors, supporters, and shells, are shown in Figure 3..
Figure 2. shows the process of data collection of ISOPR. The data collection involves the following key steps: (i) place the work-ieces from digital model library randomly within [−0.2m, 0.2m]; (ii) rotate the workpieces randomly within [−45.0◦, 45.0◦]; (iii) capture the point clouds of the whole scene by the depth camera; (iv) segment the point clouds by a predefined bounding box to obtain the point clouds of the workpieces; (v) down-sample the point clouds by voxelization, whose resolution of each sample is selected from {50, 100, 200, 500, 1000, 2000} to make the number of points exceed 1024 while remaining as close as possible, because previous work was developed based on point clouds around 1024. Finally, we obtain 2000 samples for testing. Some samples are shown in Figure 4..
ISOPR can be used to evaluate the point-cloud registration methods. The methods need to align the source point cloud and reference point cloud and estimate the rotation matrixes and translation vectors. Then, compared with the GT value, the performance of methods can be evaluated.
The folder contains all data for ISOPR, structured across 5 ".pkl" files. Each file comprises 400 samples, totally 2000 samples. Every sample is stored as a dictionary with the following keys:
- 'gt': point cloud of Ground truth CAD model in ndarray. (shape: N1×3)
- 'gt_normal': Normal vectors of the ground truth point cloud in ndarray. (shape: N1×3)
- 'det': Detected point cloud in ndarray. (shape: N2×3)
- 'det_normal': Normal vectors of the detected point cloud in ndarray. (shape: N2×3)
- 'R': Ground truth rotation matrix in ndarray. (shape: 3×3)
- 't': Ground truth translation vector in ndarray. (shape: 3×1)
The 'gt' point cloud and the 'det' point cloud can be aligned by:
This dataset can be downloaded in terminal by:
git clone https://github.com/macs-lab/iLSPR-inspection.git
cd iLSPR-inspection/ISOPR
This dataset can be opened by "pickle" package in Python:
import pickle as pkl
with open('ISOPR_0.pkl', 'rb') as p:
ISOPR = pkl.load(p)
Our dataset is released under MIT License (see './ISOPR/License.txt' file for more details)
In mechanical designing and computer-assisted design, geometric primitives are the simplest, irreducible geometric shapes that serve as foundational building blocks for constructing complex 3D models or scenes. Geometric primitives include triangular, quadrangular, pentagonal, hexagonal, heptagonal, and octagonal prisms and pyramids, as well as cylinders, cones, and spheres.
In this manuscript, we generates synthetic part point clouds using a set of geometric primitives, grounded in the engineering design principle that mechanical components are typically constructed through CAD operations—such as Boolean unions, subtractions, and intersections—applied to geometric primitives.
This dir stores the geometric primitives used in our manuscript.
The digital model library stores the ground truth CAD models of each production line’s specific products that serve as the target objects. In our work, we collected 67 CAD models and built the digital model library, which are used to collect ISOPR dataset in IsaacSim and generate authentic point clouds for model training.
[1] Z. Wu, S. Song, A. Khosla, et al, 3D Shapenets: A Deep Representation for Volumetric Shapes, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 1912-1920
[2] S. Koch, A. Matveev, Z. Jiang, et al, ABC: A Big CAD Model Dataset for Geometric Deep Learning. in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 9601-9611




