Advanced computer vision system for PCB component pick-and-place operations with 98.8% success rate, ±0.05mm placement tolerance, and real-time 6-DOF grasp planning for SMD components.
A consumer electronics contract manufacturer in Shenzhen with UK engineering office
7-month engagement, April – October 2023
1 robotics vision specialist, 1 controls engineer
The client manufactures consumer electronics for several major brands and needed to increase throughput on their SMD (surface-mount device) component placement lines. Their existing pick-and-place machines handled standard components well, but a growing proportion of their product mix involved irregularly-shaped connectors, shielding cans, and non-standard packages that defeated the fixed-template vision systems on their legacy machines. The UK engineering office managed the vision system R&D, with deployment and integration handled at the Shenzhen production facility.
Developed a sophisticated robotic vision system for automated PCB component pick-and-place operations. The system picks SMD components from feeder trays and places them on PCBs with ±0.05mm tolerance, combining advanced computer vision, machine learning, and 6-DOF robotic control to handle 50+ component types including irregularly-shaped connectors and shielding cans that defeated the client's legacy template-matching systems.
Universal Robots UR5e, Intel RealSense, Industrial Cameras
ROS2, OpenCV, PyTorch, MoveIt
YOLOv7, Point Cloud Processing, 3D Reconstruction
PID Control, Trajectory Planning, Force Feedback
The client needed to automate the placement of non-standard SMD components — irregularly-shaped connectors, metal shielding cans, and custom packages — that their existing template-based pick-and-place machines could not handle. Key challenges included:
The client's legacy pick-and-place machines used fixed template matching for component detection. This worked reliably for standard rectangular passives and ICs but failed on the growing range of non-standard components in their product mix — shielding cans with variable reflectivity, connectors with protruding pins, and custom mechanical parts. The template library required manual re-tuning for each new component, taking 2–3 days per component type, and even after tuning the detection rate for irregular shapes rarely exceeded 88%. The client had also trialled a commercial AI-based vision add-on, but it was optimised for warehouse logistics (large objects, loose tolerance) and could not achieve the sub-millimetre precision required for PCB assembly.
Deployed a synchronised multi-camera setup with RGB and structured-light depth cameras to capture comprehensive 3D information about components in feeder trays and on PCBs. The system uses structured light rather than stereo vision for depth estimation.
Why structured light over stereo: At the sub-millimetre scale of SMD components, stereo matching struggles with the low-texture surfaces of bare PCBs and metallic component bodies. Structured-light projection provides reliable depth at 0.02mm resolution regardless of surface texture, which was essential for accurate pose estimation of reflective shielding cans.
Developed a custom YOLOv7 model trained on a large-scale dataset of electronic components. 100,000+ training images were collected over 4 weeks using an automated capture rig that photographed components in feeder trays under varied orientations and lighting. Synthetic augmentation generated an additional 300,000 samples covering edge-case orientations and partial occlusions. Manual annotation of edge cases — particularly components with ambiguous orientation markers — required 2 weeks of specialist labelling by engineers familiar with PCB assembly. The model achieves 98.8% detection and classification accuracy across all 50+ component types.
Why YOLOv7 over alternatives: We benchmarked YOLOv7 against EfficientDet and Detectron2 (Faster R-CNN). EfficientDet achieved comparable accuracy but at 3x the inference latency, which would have pushed cycle time above the 2.5-second target. Faster R-CNN was more accurate on the smallest components (0201 passives) but could not run at the required frame rate on the target GPU (NVIDIA RTX A4000). YOLOv7 provided the best accuracy-latency trade-off for our component size range.
Implemented 6-DOF pose estimation combining point cloud registration with learned keypoint detection, enabling the robot to determine each component's exact position and orientation with sub-millimetre accuracy for precise placement on PCB pads.
Why hybrid over pure learned pose: Pure deep-learning pose estimators (e.g., PoseCNN) had difficulty generalising across the wide range of component geometries without per-component fine-tuning. Our hybrid approach uses learned keypoints to initialise an ICP (Iterative Closest Point) refinement step against CAD models, achieving consistent ±0.05mm accuracy across all component types without component-specific training.
Created a grasp planning system that analyses component geometry from the 3D point cloud and selects optimal grasp points based on stability, accessibility, and collision avoidance with adjacent components in feeder trays.
Why analytical over learned grasping: Learned grasp planners (e.g., GraspNet) are effective for novel objects but introduce variability that is unacceptable at ±0.05mm placement tolerance. Our analytical planner uses component CAD models to compute deterministic grasp poses, guaranteeing repeatable pick orientation. For components without CAD models, we fall back to a constrained learned planner that is limited to a pre-validated set of grasp templates.
The vision system processes data through multiple stages:
Advanced control algorithms ensure precise manipulation:
Success Rate (up from 88% with legacy template matching)
Placement Accuracy (was ±0.5mm with legacy system)
Full Pick-Place-Verify Cycle Time
Component Types (was 12 with legacy system)
The 2.3-second cycle time reflects the full pick-place-verify loop for precision PCB assembly, not just the pick or place motion. The breakdown is as follows:
For standard rectangular passives with known orientation, grasp planning is deterministic and the cycle time drops to 1.6 seconds.
The robotic vision system can handle various tasks:
Seamlessly integrated with existing manufacturing infrastructure:
The system is deployed across the client's PCB assembly operations:
Active development and planned near-term improvements: