A Robotic Automatic Assembly System Based On Visio
A Robotic Automatic Assembly System Based On Visio
sciences
Article
A Robotic Automatic Assembly System
Based on Vision
Rui Song 1,2, * , Fengming Li 2 , Tianyu Fu 2 and Jie Zhao 1, *
1 School of Mechanical and Electrical Engineering, Harbin Institute of Technology, Harbin 150006, China
2 Center for Robotics, School of Control Science and Engineering, Shandong University, No. 17923,
Jingshi Road, Jinan 250061, China; lifengming@sucro.org (F.L.); 201934472@mail.sdu.edu.cn (T.F.)
* Correspondence: rsong@sdu.edu.cn (R.S.); jzhao@hit.edu.cn (J.Z.); Tel.: +86-531-88396083 (R.S.)
Received: 12 January 2020; Accepted: 30 January 2020 ; Published: 8 February 2020
Abstract: At present, the production lines of mobile phones are mainly manual and semi-automatic.
Robots are the most important tools used to improve the intelligence level of industry production.
The design of an intelligent robotic assembly system is presented in this paper, which takes the
assembly of screen components and shell components as examples. There are major factors restricting
the application of robots, such as the calibration of diversified objects, the moving workpiece with
incomplete information, and diverse assembly types. A method was proposed to solve these
technological difficulties. The multi-module calibration method is used to transform the pose
relationship between the robot, the conveyor belt, and the two cameras in the robot assembly
system. Aiming at a workpiece moving with incomplete information, the minimum matrix method is
proposed to detect the position. Then dynamic fetching is realized based on pose estimation of the
moving workpiece. At last, the template matching method is used to identify the assembly area of
the workpiece type. The proposed method was verified on a real platform with a LINKHOU LR4-R5
60 robot. Results showed that the assembly success rate was above 98%, and the intelligent assembly
system of the robot can realize the assembly of mobile phone screens and back shells without any
staff.
1. Introduction
Continuous progress of product design concepts has had a great impact on Computer,
Communication and Consumer (3C) industry manufacturing technology. Many enterprises are trying
to shift production lines from mass production to mass customization. A production system should
have the capability to handle smaller batch sizes [1]. As one of the representatives of the 3C industry,
mobile phones are an indispensable communication tool in modern life, which have the characteristics
of compact design and rapid updating. Phones assembly task is complicated and varied, usually
involving many parts and tools (see Figure 1). and there are several challenges [2], which include
the display module, camera, battery, and main board. After investigation, the assembly process of
mobile phone is as shown in the Figure 2. The mobile phone manufacturing process is manual or
semi-automatic, which also leads to high manufacturing costs [3]. The application of robot technology
brings opportunities to production line automation. The automation level of phone production lines
requires the improvement of robotic technology’s application. If robots replace humans, the labor cost
can be greatly saved [4]. However, the assembly parts of mobile phones are various and the robot
operation ability is weak [5,6].
Many robotic companies and researchers have been focused on dual arm robots, such as the
Baxter robot [7], Yumi robot [8], and NEXTAGE robot [9]. They have been applied in some industrial
tasks [10]. However, the dual arm robot with precise hardware is of high cost [11]. Therefore, the
signal-arm robot is popular with some companies. There are many difficulties in the process of mobile
phone assembly, including pose detection, dynamic grabbing, multi-type assembly, and so on. Vision
on the production line is a common set of sensors used to obtain information. The calibration process
of a camera is tedious. In addition, sometimes the information of a workpiece being assembled is
incomplete. For example, there is often a situation beyond the camera’s field of view [12], due to the
cost constraints of the industrial site and the limited manufacturing process of the camera. In addition,
the sampling frequency of the camera under a relay trigger does not match the speed of the conveyor
belt [13], which also easily causes incomplete image information of the workpiece.
Vision and force sensors are the two common ways for a robot perceive the environment [14].
The mechanical constraints may be used to guarantee the initial position of the workpiece. If the model
of the assembled workpiece is changed, the mechanical tools should be different. Machine vision
was used to detect the different types of workpieces, which improved the flexibility of the production
line [15]. In 2013, the personal household robotic systems, such as HERB and PR2, perceived the
surrounding environment and completed some actions, such as cleaning and self-charging with
Appl. Sci. 2020, 10, 1157 3 of 19
the help of visual sensors [16]. Cai et al. achieved complex movements such as grabbing and
placing objects, which combined the multi-degree-of-freedom robot with the active visual positioning
system [17]. The target 3D point cloud is segmented by machine vision, and then the trained artificial
neural network is combined with machine vision to find the most suitable position for grasping the
object [18]. Stephen et al. [19] developed the assembly platform of dynamic binocular visual feedback
system. Image servo data acquisition is used as the basis to judge the next visual measurement
and guide the assembly of the manipulator. A “precision assembly robot” with a real-time vision
system [20] has been developed, and a dynamic target recognition and tracking algorithm based on an
improved Camshaft was proposed.
Therefore, the identification and positioning of the workpiece is very important and a lot of
work has been done [21]. In [22], a multi-workpiece recognition algorithm was proposed to locate the
three-dimensional workpiece of monocular vision in a complex industrial environment, which was
based on adaptive sub-pattern manifold learning. Wan et al. [23] taught robots to perform object
assembly using multi-modal 3D vision. Compared to 3-D vision, the monocular vision is low cost
and easy to implement. For incomplete information workpiece detection, such as exceeding-field
of view, Yang et al. designed a super-field component image measuring instrument by using image
stitching technology [12]. The adaptive image stitching algorithm was proposed in [24] for panoramic
image acquisition of super-field components. In [13], Liu et al. realized the effective recognition of the
exceeding-field workpiece by extracting the feature points on the workpiece and conducting shape
matching.
In addition, the force control or hybrid position/force control method is commonly used for
assembly tasks. In previous studies, there were three classes of control methods: position control,
force control, and hybrid position/force control [10,25]. Position control is used to complete the
assembly task and vision sensors are widely used to get the positions and orientations of assembly
objects [26,27]. The electronic circuit assembly was completed using vision based on OpenCV for an
industry robot [26]. The peg-hole assembly was completed through a vision system estimation [27].
The interaction force was obtained by force sensors to guide the robot to complete the assembly [28,29].
Most of the assembly strategies are based on the idea of modeling. A three-point contact model
is built, and the pose misalignment between the peg and hole is estimated by force and geometric
analysis [30]. A novel modeling method of geometric errors was proposed by Zhang et al. for precision
assembly [31]. A control scheme was used to teach the contact states during operation [10]. A robotic
assembly parameter optimization method [23,32] was proposed to enable industrial robots to assemble
a workpiece.
Compared with the previous work, this paper is aimed at a mobile phone production line process
using a robot. A comparative overview Table 1 is listed, which shows the key differences between the
different previous systems and the proposed system. The automatic system has also been extensively
used in the food industry [33]. There are few references for applying this system to a mobile phone
assembly line.
Table 1. A comparative overview between the previous systems and the proposed system.
In this work, the robot system is proposed to realize the mobile phone assembly process’s
automation. At first, a minimum rectangle fitting algorithm is proposed to solve pose estimation
problem of moving workpieces with incomplete information. The camera-belt-robot multi-hybrid
modular calibration is used to calibrate the assembly system, which could obtain the pose relationship
Appl. Sci. 2020, 10, 1157 4 of 19
among the robot, the conveyor belt, and the two cameras. Then a posture estimation algorithm
on a moving conveyor belt is proposed to track the objects with different speeds. At last, the
effectiveness of the method verified that back shell and screen components were assembled with
a six-degree-of-freedom LINKHOU.
The rest of this paper is laid out as follows. Section 2 introduces the assembly system description
of phone screen components and housing cover. In Section 3, the proposed method is described in
detail. Experiments were performed to validate the proposed method, and the experimental results
are presented and discussed in Section 4. Finally, the results of the current work were summarized and
future directions were discussed in Section 5.
2. System Description
The mobile phone assembly mainly includes three parts: the main board, the screen, and
the back shell. The intelligent assembly system is mainly composed of a manipulator, motion module,
vision module, and assembly module, as shown in Figure 3. The assembly producer, as shown in
Figure 4, should be divided into four steps.
1. Multiple modular hybrid calibration. This mainly includes calibration of the visual system,
calibration of the visual system and belt, and calibration of the belt and robot. The whole system
coordinate system is realized through transformation.
2. Position and posture detection of moving parts. As the motion speed is fast and the shooting lags
behind, there are repeated shootings and incomplete information of the target object in the field
of view of the camera.
3. Dynamic fetching. Due to the different speed of belt movement, the grasping position of the
suction cup is deviated during the grasping process. Posture estimation is important for tracking
belt speeds.
4. Recognize the workpiece to be assembled and the assembly itself. In the area of the workpiece to
be assembled, different types of workpieces should be placed. The correct assembly area should
be identified first, and then the assembly should be performed.
3. Proposed Method
This section describes the method of the calibration, position detection, dynamic capture, and
template matching in detail. Visual guidance is carried out to realize intelligent robot assembly of
parts and components of different types under fast beats.
First the vision system uses the calibration plate, combines the HALCON calibration operator,
takes Zhang’s calibration [34] method to obtain the camera parameters, and then adopts the nine-point
method [35] to obtain the camera’s external parameters. The pose data of the industrial camera
coordinate system {C } are represented by CT P in the conveyor coordinate system { T }, and their
expression is as follows:
Appl. Sci. 2020, 10, 1157 6 of 19
" #
T TC RC
C P= (1)
0 1
RC denotes a rotation vector rotated about the Z axis of the conveyor coordinate system, and TC
represents the translation vector along the X axis and Y axis of the conveyor belt coordinate system.
The calibration of the vision system and the belt is completed through the coordinate values in the
pixel coordinate system converting into coordinate values in the conveyor coordinate system.
B
Then the marker is placed on the conveyor belt. The initial pose T1 P of the marker is recorded, as
B2
are the pose T P after the belt has moved a distance and the number of pulses of the encoder L1 ,L2 .
The relationship between the two positions is as follows:
B2 B2 B1
T P = B1 P × T P (2)
As the process of moving the workpiece on the conveyor belt is only x axis translation, with no
rotation, the matrix expression of BB21 P is defined as follows:
1 0 x B1 B
2
B2
B1 P = 0 1 0 (3)
0 0 1
where x B1 B represents the translation distance of the workpiece along the X axis of the conveyor belt
2
coordinate system. The calculation method is as follows:
x B1 B = M · L (4)
2
where M indicates the change of X axis value of an encoder pulse corresponding to the conveyor
belt coordinate system. L= L2 − L1 represents the sum of pulses turned by the encoder during the
workpiece’s movement.
When the workpiece is detected by the industrial camera, with the advance of the conveyor belt,
its position and pose data in the robot base coordinate system CB2 P of the mechanical arm are expressed
as follows:
B2
C P = B2 P × TB1 P × CT P
B1
1 0 M·L " #" #
RB TB RC TC (5)
= 0 1 0 .
0 1 0 1
0 0 1
According to the rotation translation relationship of the coordinate system and the speed of the
conveyor belt, the calibration of the conveyor and the robot is completed.
As shown in Figure 6, B is the fixed base coordinate system of the robot arm, E is the end coordinate
system of the arm, and C is the camera coordinate system. B HW represents the transformation
relationship of the base coordinate system to the calibration coordinate system, including the rotation
matrix and the translation vector. C HE can be derived from the robot system. Then the robot was
moved to position 1,
−1
PE = B HE1 ∗ E HC ∗ C
HC1 PW (6)
BH = BH ∗ EH ∗ W H
−1
W E1 C C1
−1 (7)
= B HE2 ∗ E HC ∗ W HC2 .
Appl. Sci. 2020, 10, 1157 7 of 19
Make:
A1 ∗ X ∗ B1−1 = A2−1 ∗ A1 ∗ X
A2−1 ∗ A1 ∗ X = X ∗ B2−1 ∗ B1 (8)
AX = XB.
Therefore:
A1 ∗ X ∗ B1−1 = A2−1 ∗ A1 ∗ X
A2−1 ∗ A1 ∗ X = X ∗ B2−1 ∗ B1 (9)
AX = XB.
Among them, A is known, X is to be sought, and B can be obtained by camera calibration (Zhang’s
calibration method) [34].
Set the pixel on the image larger than the threshold to be maximal, and the other pixels to be
minimal. The region where the pixel is maximal in the binarized image is the region of interest (ROI).
In order to overcome the problem of some tiny black holes in the binarized image, a closed operation
process of first dilating and then eroding is required. For the dilate operation:
Appl. Sci. 2020, 10, 1157 8 of 19
The target contour is extracted by Canny edge detection algorithm [36]. The performance of
this algorithm depends largely on the Gaussian filter and the high and low threshold settings. The
Gaussian filter is as follows:
1 x 2 + y2
G ( x, y) = e 2σ2 (13)
2πσ2
Aiming at the incomplete image problems, an algorithm of the minimum rectangle fitting of
feature regions was proposed. Combined with the shape characteristics of the grabbing workpiece,
the minimum circumscribed rectangle is used to fit the ROI of the workpiece appearing in the camera
field of view, and the actual pose of the workpiece is solved by the eigenvalue of the rectangle.
In the pixel coordinate system, the camera field size is X*Y, the actual length of the workpiece is
H pixels, the workpiece center point is Pr , and the position is (U, V ). The length of the rectangular
workpiece fitted in the field of view is h, and the center point Pf coordinate is (u, v). The rotation angle
of the rectangular area is θ. The specific mathematical schematic is shown in Figure 7.
the robot arm coordinate system are obtained [37], along with the incoming belt displacement and
real-time prediction of the pose sequence of the target workpiece. Solving the motion trajectory of the
end effector of the robot based on the Kalman filter algorithm [38,39] achieves dynamic tracking of
a target workpiece position. Based on the pseudo-inverse solution of the Jacobian matrix, the joint
motion velocity of the robot arm is calculated, and a smooth tracking trajectory is generated to achieve
smooth grabbing of the target. The crawling process is divided into the following three steps [40]:
1. Approaching the target object. The end of the arm reaches the ready position, and the amount
of change in the position of the end effector and the target object is calculated. N path points
are planned between the two, and the end is accelerated first, reaching the first path point and
decelerating to the workpiece speed to synchronize with the target object.
2. Tracking target object. The trajectory of the object is the desired trajectory of the end effector of the
arm in the Cartesian coordinate space, which is controlled to keep moving, following the object.
The joint angle and velocity corresponding to each path point can be solved by the Jacobian
pseudo inverse matrix.
3. Grab the target object. When the distance from the workpiece meets the requirements, the object
is sucked by the air pump to complete the grasping process.
2
∑ x0 ,y0 ( T ( x 0 , y0 ) − I ( x + x 0 , y + y0 ))
R( x, y) = q (15)
2 2
∑ x0 ,y0 T 0 ( x 0 , y0 ) · ∑ x0 ,y0 I ( x + x 0 , y + y0 )
In the above, T ( x, y) is the chosen template, I ( x, y) is the target object, R( x, y) is the chosen
template, I ( x, y) is the target object, and R( x, y) is a function of similarity.
5. A recognition result is provided. The performance, based on template matching, depends to a
large extent on the quality of the template database.
sets the robot pre-crawl position, calculates the conveyor moving distance, waits for the workpiece
to arrive, and dynamically grabs the workpiece to be loaded based on the model predictive control
method to realize the pre-fetch function. Through the calibration of the hand end of the robot arm, the
relationship between the position of the arm and the area to be loaded is established. The method of
template matching is used to solve the position of the position to be assembled, and the robot arm is
guided to the assembly site to realize the assembly function.
4. Experiments
workpiece to be loaded, and to guide the robot to reach the target position. The assembly module
mainly includes an end fixture and an assembly workbench. The system uses suction cup grippers
to grab and place mobile phone parts, and the workbench is used to place the parts to be assembled.
The system uses the front and rear covers of the mobile phone as the assembly object. Algorithm
testing relies on industrial computers with the win10 operating system. Through Hikvision machine
vision camera client MVS SDK V2.3.1, C++ language was used to establish communication with the
camera. The images collected by the camera are processed by the Halcon operator using C# language.
Communication was established with the robot through Transmission Control Protocol/Internet
Protocol (TCP/IP), and control the robot to grasp the workpiece. The programs were run in Visual
Studio 2013.
field of view of the industrial camera to trigger the industrial camera to take pictures and calculate the
theoretical calculation pose that the robot should reach. Single-step execution of the robotic program:
the robotic arrives at the pre-grasp point to wait, opens the conveyor belt to move the back cover of
the mobile phone to the optimal gripping position of the robot, and manually controls the end clamp
of the robot to reach the center of the back cover of the mobile phone. The actual pose of the robot at
this time is recorded and compared to the two pose data.
From Table 4, the maximum detection error of the uniaxial coordinate data of each workpiece is
within 0.5 mm, and the attitude detection error is within 0.8 degree, which meets the technological
requirements of the mechanical arm to automatically absorb the back shell of the mobile phone and put
it into the area to be assembled with a vacuum chuck. In Liu’s system [13] for workpieces exceeding
the visual field, it takes about 6 s to recognize and grab the workpiece. In our method, the speed of the
conveyor can be matched to achieve real-time fast grabbing. It costs about 2 s to grasp the workpiece; it
can reach with the conveyor speed of 90 mm/s with incomplete information. In terms of accuracy, the
horizontal error of our grasping position is about 1mm, which is better to Liu’s system(2s). Compared
with special equipment, our robot system is more versatile and can be reused to suit different module
assembly requirements in engineering. The cost of industrial production will be reduced, and it is
more suitable for small batch, order-based manufacturing needs.
Calibration Parameters
Parameters Data Unit
Camera calibration error 0.0467 pixel
RC (0.656, 359.961) degree
TC (−22.521, −2.042) mm
RB (89.325, 179.651) degree
TB (321.251, 251.325) mm
M 10.305 line/mm
1 2 3
Camera detectes pose (488.396, 507.995, −88.45◦ ) (394.760, 467.125, 84.74◦ ) (609.061, 543.858, −76.69◦ )
Theoretical calculation pose (152.458, 245.584, −88.45 ) (171.821, 245.584, 84.74 ) (131.254, 245.584, −76.69◦ )
◦ ◦
Actual test pose (152.953, 245.557, −87.76◦ ) (171.411, 245.544, 83.97◦ ) (131.695, 245.567, −76.06◦ )
X axis error 0.495 mm 0.41 mm 0.441 mm
Angle error 0.69◦ 0.77◦ 0.63◦
Appl. Sci. 2020, 10, 1157 14 of 19
arm through the coordinates of the hand eye calibration, and the assembly of the back cover of the
mobile phone is completed.
5. Conclusions
This article focuses on the demand of intelligent mobile phone manufacturing in the 3C
industry. Certain problems, such as multivariate calibration, dynamic fetching, and identification
of the workpiece to be installed, were solved by the proposed methods. Combined with the
Appl. Sci. 2020, 10, 1157 17 of 19
camera-conveyor-robot hybrid modular calibration, the algorithm of the minimum rectangular fitting
of the feature area is used to solve the incomplete workpiece information on the conveyor belt. The
motion trajectory of the robot end effector achieves dynamic tracking of the target workpiece position
based on the Kalman filter algorithm. The template matching is used to identify the workpiece to be
assembled. The proposed methods were performed on the built platform. The experimental results
show that the maximum detection error is within 0.5mm and 0.8 degrees. The robot could pick up the
workpiece with a success rate of 100% at different conveyor speeds, even though the target component
provided incomplete information. Two different phone types were successfully assembled. The key
technologies developed in this paper will be applied to the key equipment in the automatic robot
production line. This will effectively improve the qualified rate of products, significantly improve
market competitiveness of the target products, and have a broad market space.
Author Contributions: R.S. and F.L. conducted the literature review; R.S. and F.L. conceived and designed the
algorithm; R.S. and T.F. formulated the mathematical model; R.S. and J.Z. developed the mathematical proof; R.S.
and T.F. designed and performed the experiments. All authors have read and agreed to the published version of
the manuscript.
Funding: This research was funded by the National Nature Science Foundation of China, grant number
61673245, the Major Science and Technology Innovation Project of Shandong Province, China, grant number
2019JZZY010430, and Major scientific and technological innovation projects in Shandong Province, China, grant
number 2017CXGC0915.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Kramberger, A.; Gams, A.; Nemec, B.; Schou, C.; Chrysostomou, D.; Madsen, O.; Ude, A. Transfer of
contact skills to new environmental conditions. In Proceedings of the 2016 IEEE-RAS 16th International
Conference on Humanoid Robots (Humanoids), Cancun, Mexico, 15–17 November 2016; pp. 668–675,
doi:10.1109/HUMANOIDS.2016.7803346.
2. Gu, Y.; Sheng, W.; Crick, C.; Ou, Y. Automated assembly skill acquisition and implementation through
human demonstration. Robot. Auton. Syst. 2018, 99, 1–16. doi:10.1016/j.robot.2017.10.002.
3. Guo, R. Research on Circuit Board Precision Assembly System Based on Assembly Robot and Vision.
Ph.D. Thesis, Beijing Institute of Mechanical Industry Automation, Beijing, China, 2014.
4. Sierla, S.; Kyrki, V.; Aarnio, P.; Vyatkin, V. Automatic assembly planning based on digital product
descriptions. Comput. Ind. 2018, 97, 34–46. doi:10.1016/j.compind.2018.01.013.
5. Yin, Z. Research on Assembly Technology of Small Electronic Product Robot. Ph.D. Thesis, Ningxia
University, Ningxia, China, 2016.
6. Ma, D. Design and Simulation of iNtelligent Positioning and Handling Device for Mobile Phone Production
Line. Ph.D. Thesis, North China University of Water Resources and Hydropower, Zhengzhou, China, 2014.
7. Rethink Robotics Inc. Baxter Robot. Available online: http://www.rethinkrobotics.com/baxter/ (accessed
on 1 March 2016).
8. ABB Inc. Yumi Robot. Available online: http://new.abb.com/products/robotics/industrial-robots/yumi
(accessed on 1 March 2016).
9. Kawada Inc. Nextage Robot. Available online: http://nextage.kawada.jp/en/ (accessed on 1 March 2016).
10. Ortegaaranda, D.; Lopezjuarez, I.; Nathsaha, B.; Osoriocomparan, R.; Penacabrera, M.; Lefranc, G. Towards
learning contact states during peg-in-hole assembly with a dual-arm robot. In Proceedings of the Chilean
Conference on Electrical Electronics Engineering Information and Communication Technologies, Pucon,
Chile, 18–20 October 2017; pp. 1–6.
11. Huang, Y.; ; Zhang, X.; Chen, X.; Ota, J. Vision-guided peg-in-hole assembly by Baxter robot. Adv. Mech.
Eng. 2017, 9, 1–9. doi:10.1177/1687814017748078.
12. Yang, J.; Zhou, H. Image acquisition optimization method for super field of view based on image pyramid.
Mach. Made 2012, 50, 38–41.
13. Liu, Z.; Wan, P.; Ling, L.; Chen, L.; Li, X.; ZHOU, W. Recognition and Grabbing System for Workpieces
Exceeding the Visual Field Based on Machine Vision. ROBOT 2018, 40, 40–46, 54.
Appl. Sci. 2020, 10, 1157 18 of 19
14. Liu.; Song.; Xu.; De.; Zhang.; Zhengtao.; Dapeng. High Precision Automatic Assembly Based on Microscopic
Vision and Force Information. IEEE Trans. Autom. Sci. Eng. 2016, 13, 382–393.
15. Chen, Y. Design of “intelligent manufacturing” flexible production line based on industrial robot. Manuf.
Autom. 2017, 55–57. doi:CNKI:SUN:JXGY.0.2017-08-014.
16. Hsiao, E. Addressing Ambiguity in Object Instance Detection. Ph.D. Thesis, Carnegie Mellon University,
Pittsburgh, PA, USA, 2013.
17. Cai*, J.; Ming, X.; Li, J.; Wang, X.; Tan, W.; Yuan, H.; Hu, S. A Multi-Degree of Freedom Manipulator Based
on Active Vision Positioning. Artif. Intell. Robot. Res. 2017, 6, 75–90. doi:10.12677/AIRR.2017.63009.
18. El-Khoury, S.; Sahbani, A. A New Strategy Combining Empirical and Analytical Approaches for Grasping
Unknown 3D Objects. Robot. Auton. Syst. 2010, 58, 497–507. doi:10.1016/j.robot.2010.01.008.
19. Ralis, S.J.; Vikramaditya, B.; Nelson, B.J. Micropositioning of a weakly calibrated microassembly system
using coarse-to-fine visual servoing strategies. IEEE Trans. Electron. Packag. Manuf. 2000, 23, 123–131.
doi:10.1109/6104.846935.
20. Zhong, D.; Yang, Y.; Liu, R.; Han, J. Study and Application of Monocular Vision-Based Assembly Robot. J.
Xi’an Jiaotong Univ. 2018, 52, 81–87.
21. Yang, L.; Chong, M.; Bai, C.; Li, J. A multi-workpieces recognition algorithm based on shape-SVM learning
model. J. Phys. 2018, 1087, 022025. doi:10.1088/1742-6596/1087/2/022025.
22. Li, L.; Wei, X.; Shen, B. 3D workpiece pose estimation based on adaptive sub-pattern manifold learning. J.
Jiangsu Univ. 39, 86–91.
23. Wan, W.; Lu, F.; Wu, Z.; Harada, K. Teaching robots to do object assembly using multi-modal 3D vision.
Neurocomputing 2017, 259, 85–93. doi:10.1016/j.neucom.2017.01.077.
24. Gaochao, T. Reserch and Engineering Application of Batch Parts Automatic Detection System Based on
Maching Vision. Ph.D. Thesis, Donghua University, Shanghai, China, 2014.
25. Wu, B. Industrial Robot High Precision Assembly Based on Force control. Ph.D. Thesis, University of
Chinese Academy of Sciences, Beijing, China, 2017.
26. Korta, J.; Kohut, J.; Uhl, T. OpenCV based vision system for industrial robot-based assembly station:
calibration and testing. Pomiary Autom Kontrola 2014, 60, 35–38.
27. Lin, L.L.; Yang, Y.; Song, Y.T.; Nemec, B.; Ude, A.; Rytz, J.A.; Buch, A.G.; Krüger, N.; Savarimuthu, T.R.
Peg-in-Hole assembly under uncertain pose estimation. In Proceeding of the 11th World Congress
on Intelligent Control and Automation, Shenyang, China, 29 June–4 July 2014; pp. 2842–2847,
doi:10.1109/WCICA.2014.7053179.
28. Xu, Y.; Hu, Y.; Hu, L. Precision Peg-in-Hole Assembly Strategy Using Force-Guided Robot. In Proceeding of
the 2015 3rd International Conference on Machinery, Materials and Information Technology Applications,
Qingdao, China , 28–29 November 2015; pp. 1406–1412, doi:10.2991/icmmita-15.2015.260.
29. Roveda, L. Adaptive Interaction Controller for Compliant Robot Base Applications. IEEE Access 2019,
7, 6553–6561. doi:10.1109/ACCESS.2018.2889849.
30. Tang, T.; Lin, H.; Yu Zhao.; Wenjie Chen.; Tomizuka, M. Autonomous alignment of peg and hole by
force/torque measurement for robotic assembly. In Proceeding of the 2016 IEEE International Conference
on Automation Science and Engineering (CASE), Fort Worth, TX, USA, 21–25 August 2016; pp. 162–167.
doi:10.1109/COASE.2016.7743375.
31. Zhang, Z.; Zhang, Z.; Jin, X.; Zhang, Q. A novel modelling method of geometric errors for precision assembly.
Int. J. Adv. Manuf. Technol. 2018, 94, 1139–1160. doi:10.1007/s00170-017-0936-3.
32. Chen, H.; Xu, J.; Zhang, B.; Fuhlbrigge, T.A. Improved parameter optimization method for complex assembly
process in robotic manufacturing. Ind. Robot-an Int. J. 2017, 44, 21–27.
33. Liu, Z. Research on Fast Sorting Food Packaging System Based on Machine Vision. Ph.D. Thesis, Zhejiang
University of Technology, Hangzhou, Chin, 2015.
34. ZHANG, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22,
1330–1334. doi:10.1109/34.888718.
35. Liu, J. Research on rObot Grasping Simulation Training Technology Based On Deep Learning. Ph.D. Thesis,
Harbin Institute of Technology, China, 2018.
36. Canny, J. A Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. and Mach. Intell. 1986,
PAMI-8, 679–698. doi:10.1109/TPAMI.1986.4767851.
Appl. Sci. 2020, 10, 1157 19 of 19
37. Xie, Y. Research on Robot Dynamic Target Tracking Based on Vision. Ph.D. Thesis, Huaqiao University,
Quanzhou, China, 2017.
38. Gross, R.; Eubanks, T.; Steppe, J.; Freedman, A.P.; Dickey, J.O.; Runge, T.F. A Kalman-filter-based approach
to combining independent Earth-orientation series. J. Geod. 1998, 72, 215–235. doi:10.1007/s001900050162.
39. Kalman, R.E.; Bucy, R.S. New Results in Linear Filtering and Prediction Theory. J. Basic Eng. 1961, 83, 95–108.
doi:10.1115/1.3658902.
40. Zhang, C.; Shang, W.; Cong, S.; Liu, Y. Motion Planning for Robotic Smooth Catching of Moving Object. J.
Mech. Eng. 2018, 54, 193–199.
41. Huang, J.; Li, Q. Research and application of binarization algorithms gray-level image under non-uniform
illumination. Metall. Ind. Autom. 2017, 41, 13.
42. Wijnandts, H. Ecological Energetics of the Long-Eared Owl (Asio-Otus). Ardea 1984, 72, 1–92.
43. Liu, Y. Research on key Technologies for Binocular Positioning System Based on ROI Segmentation. Ph.D.
Thesis, Chongqing University, Chongqing, China, 2018.
44. Shuai, D.; Liu, C.; Wu, X.; Li, H.; Zhang, F. Image Segmentation of Field Rape Based on Template
Matching and K-means Clustering. IOP Conf. Ser. Mater. Sci. Eng. 2018, 466, 012118.
doi:10.1088/1757-899x/466/1/012118.
c 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).