Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

With

2008

method for both calibration and registration of an endoscope

Hybrid method for both calibration and registration of an endoscope with an active optical tracker Gaëtan Martia, Vincent Bettschartb, Jean-Sébastien Billiardc, Charles Baura a VRAI Group, LSRO, IPR, Swiss Federal Institute of Technology (EPFL), Switzerland b Division of General Surgery, University Hospital of Sherbrooke (CHUS), Quebec, Canada c Department of Radiology, University Hospital of Sherbrooke (CHUS), Quebec, Canada Abstract In this paper, we present a hybrid method for calibration of an endoscope and its registration with an active optical tracker. Practically, both operations are done simultaneously by moving an active optical marker in the field of view of the two devices. By segmenting image data, the LEDs composing the marker are extracted and the transformation matrix between the two referentials (homography) is calculated. By reformulating the calibration problem, registration and calibration parameters are extracted from the homography. As camera calibration and registration is an indispensable step for augmented reality or image guided applications, this technique can easily be used in the operating field because it is fast, accurate and reliable. We currently are using this technique with an augmented reality system for laparoscopic procedures. Keywords: calibration, registration, augmented reality, image-guided surgery 1. Introduction In one and a half decade, laparoscopic surgery has gained wide acceptance in both the surgical community and the general population. The obvious advantages of a minimal aggression, including the preservation of the abdominal wall integrity, have given to laparoscopic approaches a widespread use. Operations such as cholecystectomy, appendicetomy, adrenalectomy, colectomy are performed mostly laparoscopically. However some intrinsic limitations of the technique are not been overcame. The perception of the depth is lost, as well as the tactile information. The degree of freedom of the laparoscopic tools is much less than with a hand in open surgical field. Therefore, dissection during laparoscopic surgery is often somewhat hazardous and takes the longest part of the operation. Many operations, like liver or pancreatic resections, cannot yet be habitually performed by laparoscopy. In order to be able to deal safely with such highly vascularized organs, there is a need for a positioning system, allowing the surgeon to know where the relevant structures are, sparing the dissection time. This implies the development of augmented reality tools specifically designed for laparoscopic surgery. Camera calibration is an indispensable step for augmented reality or image guided applications. Among the different calibration techniques, the photometric calibration is current the most used in the medical field [1-6]. It consists of observing a calibration object, like a chessboard or a planar grid, whose geometry in the 3-D space is known with a good precision. This technique uses some snapshot representing the calibration object at different poses and extracts object’s features in the image, to finally correlate them with the 3-D model of the reference object. A calibration algorithm then extracts the homography matrix and the camera intrinsic parameters (focal lengths, optical centre, skew, and optionally radial distortions parameters), as well as extrinsic parameters (rotation and translation of the calibration object relatively to its model coordinates). Registration of a medical camera relatively to an optical tracker is also an important step for image guided applications. On way to perform this operation consist on finding some common cues visible simultaneously by the two devices. In this paper, we present a technique using an active optical marker as the calibration and registration object. By reformulating the calibration problem, it is so possible to express the extrinsic parameters as the registration matrix. Calibration and registration are done simultaneously by moving an active marker both visible by the optical tracker and the endoscope. This technique can easily be deployed in the operating field because it is fast, accurate and reliable. 2. Methods Different optical trackers like easyTrack [7] or Polaris® [8] are using active markers composed of luminescent diodes (LEDs) that are emitting in the near infrared (~ 850 nm). Although, this wavelength is not visible by human eyes, most classic camera sensors are sensitive to the near infrared and can detect without difficulties LEDs’ flashes (figure 1). Active markers, taken at different positions, can thus be used as calibration objects during the procedure. The calibration procedure consists of placing an emitting marker on the endoscope and another one in its field of view. Both markers should be visible by the optical tracking device but only the free one has to be seen by the camera. When moving this marker, the tracker records the 3-D position of its LEDs. Endoscope images are simultaneously acquired and the corresponding 2-D positions of the LEDs are extracted via segmentation. A resampling of the data is usually necessary due to the acquisition rate difference of the two devices. The first calibration step finds the linear transformation from the tracker coordinates (Xi, Yi, Zi) to the endoscope image coordinates (ui, vi). Using a homogeneous 3 x 4 matrix representation for matrix A the following equation can be written:  Xi   ui   a11 a12 a13 a14         Yi   vi  =  a21 a22 a23 a24  ⋅   Z  1  a a32 a33 a34   i  31 { 14 4424443  1  x { A (1) X The matrix A, also called homography, has 11 degrees of freedom (3 rotations, 3 translations, 5 intrinsic parameters) and has an arbitrary scale factor involved, so one of the entries can be set to 1 without loss of generality (i.e. axy = axy / a34). The direct linear transformation, initially developed by Y. Abdel-Aziz [9] is used to solve this system. Assume that the camera is pinhole without radial distortions, the correspondence between the image points and the tracker points is the following projection: a11 X i + a12Yi + a13 Z i + a14  ui = a X + a Y + a Z + a 31 i 32 i 33 i 34 (2)  a21 X i + a22Yi + a23 Z i + a24 vi =  a31 X i + a32Yi + a33 Z i + a34 For N corresponding points (N ≥ 6), the equation 2 can be rewritten: 0 0 0 0 u1 X 1 u1Y1 u1 Z1 u1   a11   − X 1 − Y1 − Z1 − 1     v1Y1 v1 Z1 v1   a12  0 0 0 − X 1 − Y1 − Z1 − 1 v1 X 1  0  ⋅ M  = 0 M     0 0 0 0 u N X N u N YN u N Z N u N   a33   − X N − YN − Z N − 1  0 0 0 0 − X N − YN − Z N − 1 v N X N v N YN v N Z N v N   a34  14444 444444444 444244444444444444443 123 L (3) a Where L is a 2N x 9 matrix and matrix a is the matrix A rewritten as a column vector. The well known solution of this over constrained homogenous system is the right singular vector of L associated with the smallest singular value (or equivalently, the eigenvector of LT L associated with the smallest eigenvalue). This approximate linear solution for L is used as the starting point for a non-linear minimisation of the difference between the measured and the projected points. This optimization will take into account the radial distortions and the camera real model: 2 min A ∑i=1 N  Xi     ui     Yi   vi  − A ⋅   Z 1  i { 1 4 measure 1 42 3 projection (4) Extraction of intrinsic and extrinsic parameters from L is performed using either Faugeras method [10] or QR decomposition [11] depending on the number of points available. For example, using the first method, we have: r r r  − fx   − f x r1t + uc r3t 0 uc   r1t    r r  r r − f y vc  ⋅  r2t t  = I ⋅ E =  − f y r2t + vc r3t A= 0 r r  0   0 1   r3t r3t 144   424 3 42444 3 1 I − f x t x + uc t z   − f y t y + vc t z   tz  (5) E I is the intrinsic matrix, fx and fy are the focal length in the horizontal and vertical directions, (u0, v0) is the optical centre. Note that this technique assumes that the skew (i12) is equal to 0. E is the extrinsic matrix (rotation and translation). Parameters can easily be extracted from the Faugeras decomposition (equation 5). Other decomposition method, like absolute conics, can alternatively be used to extract the intrinsic and extrinsic parameters. Depending on the optics distortions, radial correction may be necessary. The reader is referred to [5] for a more elaborated discussion about this topic. The resulting extrinsic parameters are expressed in the referential of the tracker. The final step consists of formulating this transformation in the referential of the marker fixed on the endoscope. This final transformation is the searched registration. For augmented reality applications, the resulting intrinsic parameters can optionally be used to set the virtual camera and viewport settings. 3. Results We currently are developing an augmented reality system for laparoscopic procedures. Our system allows displaying preoperative 3D models or real-time ultrasound data of the liver in the endoscopic image during minimal invasive interventions. The proposed method is currently used to calibrate and register a 10 mm endoscope with an optical tracker easyTrack 500 [7]. The segmentation of endoscopic images is based on a growing region method (figure 1b). LED’s blobs are sorted depending on their aspect ratio, size, mean intensity and neighbour pixels variance intensity attributes which are statistically pertinent. The LED position used to compute the homography is the barycentre of the blob. To improve detection reliability and speed, the initial homography can be guessed with the first 4 points correspondences by mean of either the described method or a 2D-3D pose estimation technique [12]. This estimate gives a cue of the approximate position of the LEDs in the images and allows improving the search success and speed. In our system, about 20 LEDs positions per second can be processed with this technique and calibration starts with about 150 samples acquired in the full field of view of the camera. Some of these samples contain badly segmented centres and should be filtered before homography calculation. First, the homography is estimated with the whole samples using equation 3. Then, the distance between the measured and the projected points indicates if the sample is an outlier. If so, it is removed and the homography is estimated again. Calibration accuracy is closed to conventional grids methods but is faster and do not require an extra calibration object in the operating field. Results has been compared with the gold-standard “Camera Calibration Toolbox for Matlab” [6] and do not differ for more than one percent. Total time for data acquisition, camera calibration and registration is less than one minute using a modern laptop. This method is about twice faster than the system we previously have developed and which needs a calibration grid. We currently are validating our technique with different kind of endoscopes, as well as, medical microscopes. For example, figure 2 presents a test of augmented reality we have performed during a kidney transplant on living donor. The 3D preoperative model was at bootstrap manually and rigidly registered. The endoscope allows keeping the rigid registration correct while moving. We also plan to quantitatively compare our algorithm with the computer vision gold standard calibration algorithms for early 2004 and to adapt a non-rigid registration algorithm on the augmented reality application. Figure 1: (a) Marker and (b) Segmented marker. Example of illuminated LED (left) and corresponding region-based segmentation. Figure 2: Proposed technique used in an augmented reality application for kidney transplant on living donors. The initial 3D model rigid registration is done manually and is kept while moving the endoscope. 4. Conclusion We present a new method for both calibration and registration of camera images with an active optical tracker. By exploiting the fact that markers’ LEDs are both visible by the endoscope and the optical tracker, the calibration problem can be reformulated to give simultaneously the registration of the camera relatively to the tracker and the intrinsic parameters of the camera. Compared to existing techniques using calibration objects, the proposed approach is ergonomic because medical markers are designed for sterilization and the LEDs can easily be located in the images. This new combined method will also improve time saving during clinical use because technical preparation with this technique is short. 5. References [1] R. Tsai, An Efficient and Accurate Camera Calibration Technique for 3D Machine Vision, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Miami Beach, FL, 1986, pp. 364–374. [2] R. Tsai, A Versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using Off-the-Shelf TV Cameras and Lenses, IEEE Journal of Robotics and Automation, Vol. RA–3, No. 4, August 1987, pp. 323–344. [3] B. Horn, Tsai’s camera calibration method revisited, Copyright 2000, http://www.ai.mit.edu/people/bkph/papers/tsaiexplain.pdf. [4] Z. Zhang, Flexible Camera Calibration by Viewing a Plane from Unknown Orientations. International Conference on Computer Vision (ICCV'99), Corfu, Greece, September 1999, pages 666-673. [5] Z. Zhang, A flexible new technique for camera calibration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(11):1330-1334, 2000. [6] J.-Y. Bouguet, Camera Calibration Toolbox for Matlab, last updated 2004, http://www.vision.caltech.edu/bouguetj/calib_doc/index.html. [7] easyTrack 200 and 500, products of Atracsys sàrl, www.atracsys.com [8] Polaris®, product of Northern Digital - NDI, www.ndi.ca [9] Y. Abdel-Aziz, H. Karara, Direct linear transformation into object space coordinates in close-range photogrammetry, Proc. Symposium on Close-Range Photogrammetry, 1971, Urbana, Illinois, p. 1-18. [10] O. Faugeras, G. Toscani, Camera calibration for 3D computer vision, Proc. International Workshop on Industrial Applications of Machine Vision and Machine Intelligence, Silken, Japan, 1987, p. 240-247. [11] QR Decomposition, http://mathworld.wolfram.com/QRDecomposition.html [12] S. Price, Solving for Perspective Inversion, last updated 1996, http://homepages.inf.ed.ac.uk/rbf/CVonline/LOCAL_COPIES/MARBLE/high/pia/s olving.htm