10
Int J CARS manuscript No. (will be inserted by the editor) Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour Appearance Cues with Preoperative Anatomical Priors Masoud S. Nosrati · Alborz Amir-Khalili · Jean-Marc Peyrat · Julien Abinahed · Osama Al-Alao · Abdulla Al-Ansari · Rafeef Abugharbieh · Ghassan Hamarneh Received: date / Accepted: date Abstract . Purpose Despite great advances in medical image seg- mentation, the accurate and automatic segmentation of endoscopic scenes remains a challenging problem. Two important aspects have to be considered in segment- ing an endoscopic scene: 1) noise and clutter due to light reflection and smoke from cutting tissue, and 2) structure occlusion (e.g. vessels occluded by fat, or en- dophytic tumours occluded by healthy kidney tissue). Methods In this paper, we propose a variational tech- nique to augment a surgeon’s endoscopic view by seg- menting visible as well as occluded structures in the intraoperative endoscopic view. Our method estimates the 3D pose and deformation of anatomical structures segmented from 3D preoperative data in order to align to and segment corresponding structures in 2D intra- operative endoscopic views. Our preoperative to intra- operative alignment is driven by, first, spatio-temporal, signal processing based vessel pulsation cues and, sec- ond, machine learning based analysis of colour and tex- tural visual cues. To our knowledge, this is the first work that utilizes vascular pulsation cues for guiding preoper- M. Nosrati (B), G. Hamarneh Medical Image Analysis Lab, Simon Fraser University, Burn- aby, BC, V5A 1S6, Canada Tel.: +1 778-782-5509 Fax: +1 778-782-3045 E-mail: [email protected] A. Amir-Khalili, R. Abugharbieh BiSICL, University of British Columbia, Vancouver, BC, Canada J.M. Peyrat, J. Abinahed Qatar Robotic Surgery Centre, Qatar Science & Technology Park, Doha, Qatar O. Al-Alao, A. Al-Ansari Urology Department, Hamad General Hospital, Hamad Med- ical Corporation, Doha, Qatar ative to intraoperative registration. In addition, we in- corporate a tissue-specific (i.e. heterogeneous) physically- based deformation model into our framework to cope with the non-rigid deformation of structures that oc- curs during the intervention. Results We validated the utility of our technique on fif- teen challenging clinical cases with 45% improvements in accuracy compared to the state-of-the-art method. Conclusions A new technique for localizing both visi- ble and occluded structures in an endoscopic view was proposed and tested. This method leverages both pre- operative data, as a source of patient-specific prior knowl- edge, as well as vasculature pulsation and endoscopic visual cues in order to accurately segment the highly noisy and cluttered environment of an endoscopic video. Our results on in vivo clinical cases of partial nephrec- tomy illustrate the potential of the proposed framework for augmented reality applications in minimally invasive surgeries. Keywords Robotic surgery · Partial nephrectomy · Image-guided surgery · Segmentation · 3D pose estimation · Endoscopy · Patient-specific model · Occluded vessels · Kidney 1 Introduction Minimally invasive surgeries (MIS) are appreciated for their many advantages over traditional open surgery including decreased risk of infection due to minimal in- cisions, and faster recovery times for patients [15]. The shortcomings of MIS are mainly associated with the loss of direct 3D view of the surgical scene as well as cum- bersome and non-intuitive tool manipulation. With the advent of robotic MIS systems, many of these shortcom- ings have been alleviated. Greater precision, improved

Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

Embed Size (px)

Citation preview

Page 1: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

Int J CARS manuscript No.(will be inserted by the editor)

Endoscopic Scene Labelling and Augmentation using

Intraoperative Pulsatile Motion and Colour Appearance Cues

with Preoperative Anatomical Priors

Masoud S. Nosrati · Alborz Amir-Khalili · Jean-Marc Peyrat · JulienAbinahed · Osama Al-Alao · Abdulla Al-Ansari · Rafeef Abugharbieh ·Ghassan Hamarneh

Received: date / Accepted: date

Abstract .Purpose Despite great advances in medical image seg-mentation, the accurate and automatic segmentation ofendoscopic scenes remains a challenging problem. Twoimportant aspects have to be considered in segment-ing an endoscopic scene: 1) noise and clutter due tolight reflection and smoke from cutting tissue, and 2)structure occlusion (e.g. vessels occluded by fat, or en-dophytic tumours occluded by healthy kidney tissue).Methods In this paper, we propose a variational tech-nique to augment a surgeon’s endoscopic view by seg-menting visible as well as occluded structures in theintraoperative endoscopic view. Our method estimatesthe 3D pose and deformation of anatomical structuressegmented from 3D preoperative data in order to alignto and segment corresponding structures in 2D intra-operative endoscopic views. Our preoperative to intra-operative alignment is driven by, first, spatio-temporal,signal processing based vessel pulsation cues and, sec-ond, machine learning based analysis of colour and tex-tural visual cues. To our knowledge, this is the first workthat utilizes vascular pulsation cues for guiding preoper-

M. Nosrati (B), G. HamarnehMedical Image Analysis Lab, Simon Fraser University, Burn-aby, BC, V5A 1S6, CanadaTel.: +1 778-782-5509 Fax: +1 778-782-3045E-mail: [email protected]

A. Amir-Khalili, R. AbugharbiehBiSICL, University of British Columbia, Vancouver, BC,Canada

J.M. Peyrat, J. AbinahedQatar Robotic Surgery Centre, Qatar Science & TechnologyPark, Doha, Qatar

O. Al-Alao, A. Al-AnsariUrology Department, Hamad General Hospital, Hamad Med-ical Corporation, Doha, Qatar

ative to intraoperative registration. In addition, we in-corporate a tissue-specific (i.e. heterogeneous) physically-based deformation model into our framework to copewith the non-rigid deformation of structures that oc-curs during the intervention.Results We validated the utility of our technique on fif-teen challenging clinical cases with 45% improvementsin accuracy compared to the state-of-the-art method.Conclusions A new technique for localizing both visi-ble and occluded structures in an endoscopic view wasproposed and tested. This method leverages both pre-operative data, as a source of patient-specific prior knowl-edge, as well as vasculature pulsation and endoscopicvisual cues in order to accurately segment the highlynoisy and cluttered environment of an endoscopic video.Our results on in vivo clinical cases of partial nephrec-tomy illustrate the potential of the proposed frameworkfor augmented reality applications in minimally invasivesurgeries.

Keywords Robotic surgery · Partial nephrectomy ·Image-guided surgery · Segmentation · 3D poseestimation · Endoscopy · Patient-specific model ·Occluded vessels · Kidney

1 Introduction

Minimally invasive surgeries (MIS) are appreciated fortheir many advantages over traditional open surgeryincluding decreased risk of infection due to minimal in-cisions, and faster recovery times for patients [15]. Theshortcomings of MIS are mainly associated with the lossof direct 3D view of the surgical scene as well as cum-bersome and non-intuitive tool manipulation. With theadvent of robotic MIS systems, many of these shortcom-ings have been alleviated. Greater precision, improved

Page 2: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

2 Masoud S. Nosrati et al.

dexterity and enhanced 3D immersive visualization forsurgeons are three important advantages of robotic overtraditional MIS.

One of the application areas in which robotic MIS isbeing expanded is partial nephrectomy. Approximately208,000 new cases of kidney cancer are diagnosed in theworld each year [10]. The goal in partial nephrectomy,an e↵ective treatment for localized renal cancers, is toexcise only the cancerous regions and spare as muchhealthy tissue as possible to preserve kidney function.According to Gill et al. [14], a partial nephrectomy pro-cedure is organized into five main stages: 1) Bowel mo-bilization; 2) Hilar dissection and control; 3) Identifica-tion and demarcation of tumour margins; 4) Resectionof tumour; and 5) Reconstruction of the kidney (renor-rhaphy). Among these, hilar dissection and tumour de-marcation stand out as two critical and daunting stagesrequiring significant expertise.

Hilar dissection allows the surgeon access to the re-nal hilum, where the flow of blood into and out of thekidney is controlled by clamping the vessels. Impropervessel control due to overlooked accessory renal vesselscan cause significant bleeding during resection [28]. Theidentification and demarcation stage is also critical as,during this stage, the surgeon localizes the tumour massand identifies the resection margins. Proper demarca-tion speeds up the succeeding stage of tumour resection,improves the preservation of kidney function by sparingas much healthy tissue as possible, and avoids tumourrecurrence by correctly delineating all of the canceroustissue from the healthy ones.

Both of these two stages are complex tasks. In thehilar dissection and vessel clamping stage, some ves-sels might be accidentally missed due to substantialvariability in patients vasculature and the amount offat which surrounds the kidney and hides the vessels[31]. Tumour identification and demarcation is di�cultdue to noise associated with clutter such as bleeding orsmoke from cutting, poor endoscopic image colour/texturecontrast between di↵erent structures, occluding surgi-cal tools, and limited 3D visibility of the structures ofinterest where only surfaces are observable from thecamera feed. If the surgeon’s level of experience is lim-ited, the incidence of missed vessels or incorrect de-marcation may significantly jeopardize the surgical out-come.

To minimize mistakes during the operation, sur-geons typically rely on previously viewed preoperative3D scans, e.g. computed tomography (CT), and men-tally reconstruct locations of various structures duringsurgery. Transferring this mental abstraction from 3Dto 2D data is an error-prone procedure especially ifthe surgeon’s level of experience is limited. To facilitate

this, many e↵orts have been made towards augmentingthe endoscopic views. These methods vary from directlysegmenting the endoscopic scene (e.g. using level sets)to registering the preoperative data onto the intraoper-ative endoscopic scene.

1.1 Related Works

Reducing the di�culty of intraoperative navigation hasbeen attempted by various approaches that rely on multi-modal registration to align the preoperative data ontothe intraoperative view by, for example, fusing a 3Dpreoperative image volume with 2D intraoperative ul-trasound [11], intraoperative MR [16], or 2D X-ray [25,33]. In this work, we focus on augmenting the endo-scopic video as it remains the staple modality in MIS.Some recent works ignored preoperative data and seg-mented the endoscopic view using active contour-basedmethods [12,13] or parameter-sensitive morphologicaloperations and thresholding techniques [21]. Since theseapproaches rely only on colour/intensity information,they often fail due to noise and clutter from bleedingand smoke. In addition, these methods focused on seg-menting only one object in an endoscopic scene.

Other techniques proposed to register 3D preoper-ative data on 2D intraoperative view(s) manually, e.g.by using invasive fiducials and performing tracking [26,27,29,17,18]. While the registration in these methodsis performed manually, the methods proposed in Yimet al. [32] and Merritt et al. [20] are able to automati-cally find the 3D pose. Yet, none of the aforementionedmethods can handle the free-form deformation of tissuesthat usually happens due to respiratory motion and/orsurgical intervention. Furthermore, these techniques areunable to segment critical occluded structures, e.g. ves-sels covered by fat.

To account for non-rigid tissue deformation, Agudoet al. [1] exploited the well-known FEM model anal-ysis and proposed an online framework to model thenon-rigid deformations of shapes in a video sequence.However, their method has been designed for a singlecamera setting. In our previous work [23], we incorpo-rated generative statistical shape models of tissues intheir 3D pose tracking and endoscopic video segmenta-tion framework. Although the statistical model used iscapable of encoding non-rigid deformations, these de-formations are based on a heterogeneous training pop-ulation and are not patient-specific. In addition, ourprevious method required manual 3D-2D registrationfor the first frame of the video and is incapable of lo-calizing hidden vessels.

To localize vessels in the endoscopic view, di↵erentmethods have been proposed. These methods include

Page 3: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

Endoscopic Scene Augmentation using Appearance and Pulsatile Motion 3

Org

an

Data

Method

Cues

Multiple

obj.

Biomechanical

deform

ation

Occ

luded

obj.

Auto

.pose

estimation

Pre-op Intra-op

Figueiredo [12] C 7 EN Active contour CO X 7 7 -Figueiredo [13] C 7 EN Active contour CO 7 7 7 -

Mewes [21] S 7 EN Thresholding GO,TX,CH 7 7 7 -Estepar [11] K/CA CT US LEPART ED X 7 7 7

Gill [16] P MR MR 2D-3D reg. IN 7 7 7 XPickering [25] T/F CT FL 2D-3D reg. IN 7 7 7 X

Zikic [33] H CBR FL 2D-3D reg. IN 7 7 7 XPratt [26] K CT EN 3D-3D (stereo) reg. ML X 7 7 7Yim [32] VF CT EN 3D-2D reg. IN 7 7 7 X

Merritt [20] L CT EN 3D-2D reg. IN 7 7 7 XPuerto [27] K CT EN Feature tracking AL X 7 7 7Tobis [30] K 7 EN Fluorescence imaging C 7 7 7 -Teber [29] K CT EN 3D-2D reg. ML X 7 7 7

Amir-Khalili [2] V 7 EN Motion segmentation PL 7 7 X -Nosrati [23] K CT EN 3D-2D reg. CH,TX X 7 7 7

Our method K CT EN 3D-2D reg. CH,TX,PL X X X XLEPART: Low-pass spectral phase correlation with harmonic selection. op: Operative. obj: Object. reg: Registration. Auto: Automatic.Data={Computed tomography (CT), Ultrasound (US), Magnetic resonance (MR), Endoscopy (EN), Fluoroscopy (FL), Cone beamreconstruction (CBR)}. Organs=

�Kidney (K), Celiac-aorta (CA), Prostate (P), Tibia/Femur (T/F), Head (H), Colon (C), Stomach

(S), Vasculature (V), Vocal folds (VF), Lung (L) . Cues={Edge (ED), Intensity (IN), Colour (CO), Colour histogram (CH), Texture

(TX), Manual landmark (ML), Automatic landmark (AL), Pulsation (PL)}

Table 1: Categorization and comparison between state-of-the-art methods for intraoperative guidance and our proposedmethod.

both hardware solutions, such as near infrared fluores-cence imaging [30], and algorithmic methods based oncolour/intensity information from the endoscope thathighlight vasculature based on perfusion models [8].However, due to sensitivity to the dose administeredand the cost involved, hardware solutions are not widelyused. Moreover, none of these methods [30,8] are ableto identify occluded vessels that are hidden under thicklayers of fat. A detailed survey of such vessel localiza-tion techniques was presented in our recent paper [2].In our recent works [3,2] we proposed a purely image-based method for labelling occluded vasculature by de-tecting minute pulsatile vessel motion. Feedback fromour clinical collaborators revieled that the raw visual-ization of our fuzzy labels alone are not ideal for guid-ance as 1) they are susceptible to noise, 2) they cannotidentify parts of the vasculature that are deeply hiddenunder other tissue, and 3) the visualizations are noteasy to interpret compared to methods based on pre-to intra-operative alignment. This provided the motiva-tion for us to integrate the complimentary advantagesof our previously proposed techniquse in [2] and [23]into a single framework.

Two important aspects have to be considered in seg-menting an endoscopic video: noise/clutter due to lightreflection from uneven tissue surfaces and smoke fromcutting tissue, and structure occlusion either by toolsor other tissues (e.g. vessels occluded by fat, or endo-phytic tumours occluded by healthy kidney tissue). In

this paper, we address these two problems by propos-ing a variational technique to augment the endoscopicsurgical view. Our framework estimates the 3D poseand deformations of multiple structures in the preoper-ative data and uses that to aid the segmentation ofmultiple visible and occluded structures in an endo-scopic scene. Our key contribution in this work is howwe leverage vessel pulsation and endoscopic visual cuesto guide 3D preoperative to endoscopic intraoperativealignment. Our surgical image guidance system com-bines signal processing and machine learning cues. Thefirst performs spatio-temporal colour waveform analysisto detect vessel pulsation, and the second uses randomdecision forests (RF) to learn colour and textural visualpatterns of tissue types in endoscopy. The above con-tributions are used to guide a multi-structure, preoper-ative to intraoperative alignment process that allow forboth rigid and heterogeneous physically-based, patient-specific non-rigid deformations. Table 1 presents a de-tailed summary of the capabilities of state-of-the-artmethods compared to our proposed technique. To demon-strate the potential utility of our method to surgeons,we evaluated our method with a retrospective study offifteen challenging robot-assisted partial nephrectomy(RAPN) clinical cases.

Page 4: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

4 Masoud S. Nosrati et al.

Intraoperative endoscopic video

Preoperative volume (e.g. CT)

Phase-based pulsation analysis using (6)

Visual cues analysis using random forest

3D segmentation

Augmented endoscopy with projections of aligned preoperative data using (2) and (4)

Kidney

TumourVeinArtery

Kidney

Tumour

Vessels

Fig. 1: Our surgical image guidance system leverages vessel pulsation and visual appearance cues to guide the 3D preoperativeto endoscopic intraoperative alignment.

2 Method

2.1 Problem statement and notation

Given 3D preoperative data, let Spre = {S1, · · · , SN}be the set of N segmented structures of interest in thepreoperative spatial domain ⌦pre ⇢ R3, where Si rep-resents the surface of the ith structure. Having M cam-era views of the surgical scene, let Im : ⌦m

2D ⇢ R2 !R3 represent the 3-channel RGB image of mth cameraview. We adopt the 3D geometry defined by first cam-era (m = 1) as the reference domain of the surgicalscene, ⌦srg ⇢ R3. Our objective is to augment the sur-gical intraoperative view by localizing and visualizingthe N structures of interest in the 2D endoscopic im-ages (I1 · · · , IM ). To do so, we transform and deformthe objects of interest in 3D such that their projectionsonto the 2D camera views (e.g. their silhouettes) alignwith the corresponding structures in I1, · · · , IM . Figure1 illustrates the overview of our pipeline.

We implicitly represent the boundary of each struc-ture in Im by a level set function �im : ⌦m

2D ! R suchthat

8><

>:

�im(x) > 0, x is inside the ithstructure in Im

�im(x) = 0, x is on the boundary of the ith structure

�im(x) < 0, x is outside the ithstructure in Im .

(1)

Let P pre,i` = (Xpre,i

` , Y pre,i` , Zpre,i

` ) 2 Si be the coordi-nates of the `th point on Si. The level set �im is calcu-lated as:

�im(x) = SDM

✓@⇣Pm

�T (P pre,i);⇡m

�⌘◆, (2)

where T is a spatial transformation consisting of twoparts: a non-rigid component and a rigid componentthat maps the 3D models (P pre) from ⌦pre to the ref-erence surgical domain ⌦srg. Pm : ⌦srg ! ⌦m

2D isthe projection from the surgical scene frame of refer-ence to ⌦m

2D given the corresponding camera parame-ters ⇡ = {⇡1, · · · ,⇡M}, @ is the spatial 2D derivativeof the projected model (i.e. resulting in the boundaryof the structures), and SDM(.) is the signed distancemap. In less technical terms, we position and deformthe segmented 3D model (using T ), virtually image (orproject) the 3D model using the camera parameters(⇡), detect the boundary of the projected image (using@), and finally represent the boundary using level sets.We define T as:

T (P pre,i) = R

�P

pre,i + iu

i�[3⇥ni]

+ t , (3)

where the columns of i are the ith tissue-specific modesof vibration and u

i are the corresponding weights thatare used to non-rigidly deform the segmented preoper-ative models P pre,i (see Section 2.3 for more details).R and t are the rotation matrix and translation vectorin 3D space, respectively, giving the rigid componentof transformation T . Note that the number of elementsin P

pre,i is 3 ⇥ ni (ni is the number of points in Si),however they are stacked as a long 3ni ⇥ 1 column in(3). The subscript (...)[3⇥ni] is the reshaped version of

(P pre,i + iu

i) from 3ni ⇥ 1 to 3⇥ ni. Also, note that

iu

i deforms the whole ith tissue (or object Si) andnot just a single point on Si. We emphasize that R, t,and u

i in (3) are the same for the 3D models regard-less of which viewpoint of the M cameras is considered.Having N structures appear in M camera images, wedefine and minimize the following energy functional E

Page 5: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

Endoscopic Scene Augmentation using Appearance and Pulsatile Motion 5

to simultaneously align the preoperative models withthe intraoperative images and segment them:

E(T ,U ,�;P pre, I1, · · · IM ,⇡) =

NX

n=1

MX

m=1

Z

⌦m2D

(2⇢nm � 1)(x) H(�nm(x))dx , (4)

where H(.) is the Heaviside function, ⇢nm(x) are theregional terms that provide the cues towards which thelevel set is moved during the optimization and measurethe agreement of the image pixels x with the learntstatistical models (see Section 2.2) of the nth structurein Im, � = {�11, · · · ,�N1 , · · · ,�1M , · · · ,�NM}, and U ={u1, · · · ,uN}. The above energy function is similar tothe Chan-Vese model [7] with the main di↵erence lyingin the optimization; where in the Chan-Vese method,the energy function is minimized with respect to thelevel set function � whereas in our case, we optimize E

with respect to the pose and shape parameters. To findthe optimum pose and deformations of structures, weoptimize E with respect to T and U (with � is updatedaccordingly).

The success of the proposed optimization dependson how accurate we model the 3D non-rigid deforma-tion and how well we identify candidate pixels belongingto di↵erent structures in the endoscopic scenes. Somestructures are hidden (e.g. occluded vessels) while oth-ers have complicated visual appearance. In the follow-ing sections we explain how we choose our data terms(⇢nm in (4)) and how we generate the structures’ defor-mation and the final segmentation.

2.2 Data terms: appearance of structures inendoscopic images

In RAPN, the focus is on three organs: the vascula-ture (v), kidney (k), and tumour (t). In this context,we set the number of objects N = 3 in (4). The re-gional term of vessels, kidney, and tumour are repre-sented with ⇢1 = ⇢v, ⇢2 = ⇢k, and ⇢3 = ⇢t, respectively.

2.2.1 Vascular pulsatile motion:

Computing the regional term ⇢v by appearance aloneis di�cult as blood vessels are typically hidden undera layer of fat. These regions may however be identi-fied by their characteristic pulsatile motion, which isdetectable but invisible to the naked eye. This regionalterm is computed by first extracting pulsatile motionfeatures from local phase information using our methodproposed in [3]. In this section we briefly describe theframework for extracting these features and the corre-sponding data term used to drive our segmentation.

A video frame captured from camera m denotedas Im(x, t) that maps the given pixel x 2 ⌦m

2D attime t to an intensity value can be represented as func-tion of local displacements d(x, t) with respect to thefirst frame of the video Im(x, 0) = f(x) such thatIm(x, t) = f(x + d(x, t)). Our regional term ⇢v is ex-tracted from the local motions specified by d(x, t) andto approximate these motions from the video, we per-form a wavelet decomposition of the video into sub-bands such that

Im(x, t) = f(x+ d(x, t)) ⇡1X

!=�1A!e

i!(x+d(x,t)) , (5)

with each sub-band representing a complex sinusoidS!(x, t) = A!e

i!(x+d(x,t)) at spatial frequency ! andd(x, t) is our approximation to the local motion. Thelocal phase of each sub-band is defined as �!(x, t) =arg(S!) = !(x+ d(x, t)). Since the local !d(x, t) is theonly motion related component of the phase that varieswith time, we can isolate it from the zero-frequencycomponent !x by applying a mean-free temporal band-pass filter

H(x, t) = 2BHsinc(2BHt)� 2BLsinc(2BLt) , (6)

where BL and BH are the low and high frequency cut-o↵, respectively. The response of the temporal bandpassfilter is denoted by B!(x, t) = �! ⇤ H = !dH(x, t),where dH are components of the motion that are inthe passband of the filter. The passband of the filteris tuned to the typical heart rate of a patient so thatwe can isolate components of the local motion that aresynchronous with the heart rate and hence to vascularpulsation.

The bandpassed phases B! are then denoised andcombined across scales and orientations using the meth-ods outlined in [2] to obtain fuzzy labels

L =1

C

X

8!

|Q!|2⇡!

, (7)

where C is a normalizing factor and Q is the denoisedB!. To attenuate these false positives and obtain bettervisual cues of the hidden vessels, we process L in a tem-poral window (i.e. 2 sec. video) by summing L in timeand denoising the results with a spatial median filter.We denote the resulting denoised fuzzy labels by Lavg.Then, we calculate the probability of a pixel belongingto a pulsating vessel (artery and vein) as:

P v(x|Lavg) =1

2+

1

⇡arctan

✓Lavg(x)� 0.3

◆, (8)

where we chose ✏ to be 0.1. The above equation mapsLavg to the probabilistic range of [0, 1]. The value 0.3 in

Page 6: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

6 Masoud S. Nosrati et al.

(a) (b) (c)

Fig. 2: Patch selection and initialization. (a) Patch selection(blue: background/tools, yellow: kidney, green: tumour). (b)Probability map of background, kidney and tumour. (c) Ini-tial pose (before convergence). (d) Recovered pose and shape(after convergence).

(8), chosen empirically, intensifies any value more than0.3 and discards small values in Lavg. The data termcorresponding to vessels is defined as ⇢v = � log(P v(x|Lavg)).

2.2.2 Learning of tissue appearance:

To calculate the data term of a visible structure, weextract a variety of image features (Am) from each im-age m and train a RF classifier to distinguish betweendi↵erent structures. Particularly, we capture the textu-ral patterns and colour variations via local colour his-tograms of the normalized RGB and YCbCr channels.The probability of pixel x belonging to kidney (P k(x))and tumour (P t(x)) are estimated by training a RFconsisting of Nt binary decision trees. To train the RF,we select few 20⇥ 20 patches in Im, i = 1, · · · ,M fromdi↵erent structures. In practice, surgeons may (virtu-ally) select these patches with the help of surgical tools.We emphasize that unlike feature-based methods, e.g.[27], our method does not require any correspondencebetween 3D CT and the 2D intraoperative data. Fig-ure 2(a) shows a sample seeding on an example en-doscopic scene of real clinical data. After training, foreach pixel x, the feature channels, Am(x), are propa-gated through each RF tree resulting in the probabil-ity Pn,j(x|Am(x)), for the jth tree and nth structure.These probabilities are combined into a forest’s jointprobability Pn(x|Im(x)) = 1

Nt

PNt

j=1 pn,j(x|Am(x)) to

determine the probability of x belonging to nth struc-ture. Figure 2(b) illustrates examples of regions proba-bility for the frame shown in Figure 2(a). The regionalterms of kidney and tumour are then calculated as:⇢km = � logP k(x|Im(x)) and ⇢tm = � logP t(x|Im(x)),respectively.

2.3 Patient-specific heterogeneous deformation model

Due to the noisy endoscopic images, the regional termsalone are not able to provide robust cues for guiding the

3D-2D alignment. To obtain a reliable result, we con-strain the space of possible transformations to patient-specific shape models obtained from the preoperativedata (P pre). To account for non-rigid deformation ofstructures, we include the tissue-specific modes of vi-bration ( i) in our framework as described in Section2.1.

The modes of vibration i are obtained by solvingthe generalized eigendecomposition problem: Ki

i =M

i

i⇤

i, where K and M are the sti↵ness and massmatrices, respectively, and ⇤i is a diagonal matrix ofeigenvalues associated with the eigenvectors ( i) wherehigher eigenvalues corresponds to higher frequency. Incontrast to traditional variational methods that updatethe level set function representing the segmentation toadapt it to image data, the use of favours shape defor-mations that are biomechanically plausible (as dictatedby K and M).

As we increase the number of modes, more deforma-tions are allowed, which is useful for accurately mod-elling the deformations. Given all possible modes ofshape variability, we are able to produce/recover allpossible deformations of that shape. However, as weinclude more deformation modes, the computationalcomplexity increases and recovering the proper defor-mation parameters becomes more di�cult. In addition,too many modes of variations may end up with extremeand unrealistic deformations. We empirically found thatchoosing more than 6 modes results in an unjustified in-crease in complexity, as deformations may become un-realistic and no noticeable improvement in accuracy isobserved. According to Pentland et al. [24], since thelowest frequencies correspond to rigid body motionsand global deformations, and we already encoded globaldeformations, translation and rotation, through R andt (eq. (3)), we selected eigenvectors corresponding tolarger eigenvalues to focus on finer deformation details.

We use the average Hounsfield unit (HU) associatedwith each structure in the preoperative CT to approxi-mate the sti↵ness of each tissue. This way the sti↵nessof each structure is proportional to its tissue density.We used a simple linear elastic model to deform the ob-jects of interest to lower the computational complexity,however, exploring more advanced elastic models mightimprove the results but with the cost of complexity. Inthe next section we show how we optimize E in (4).

2.4 Optimization of the multi-structure pose anddeformation

We optimize (4) with respect to T , i.e. with respect toR = {↵,�, �}, t = {tx, ty, tz}, and u = {u1, · · · , uq},where ↵, �, and � are the rotation angles around the

Page 7: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

Endoscopic Scene Augmentation using Appearance and Pulsatile Motion 7

x, y, and z axes, respectively and tx, ty, and tz are thetranslations in x, y, and z directions. The derivative ofE with respect to each of the unknown parameters is:

@E

@⇠`=

X

n2{v,k,t}

MX

m=1

Z

⌦m2D

⇢nm(x)@H(�nm(x))

@⇠`, (9)

where ⇠` = {↵,�, �, tx, ty, tz, u1, · · · , uq} and

@H(�nm(x))

@⇠`=@H(�nm(x))

@�nm

✓@�nm@x

@x

@⇠`+@�nm@y

@y

@⇠`

(10)

= �(�nm)h@�n

m@x

@�nm

@y

i " @x@⇠`@y@⇠`

#.

In (10), �(.) is the Dirac delta function. We use thecentered finite di↵erence to calculate @�

@x and @�@y . Every

2D point x = (x, y) in Im has at least one correspond-ing 3D point P srg = (Xsrg, Y srg, Zsrg) in the surgicaldomain ⌦srg. We calibrated the stereo camera and ob-tained the intrinsic camera parameters. For m = 1, xand P

srg are related by:2

64x

y

1

3

75 =

2

64f1x 0 c1x

0 f1y c1y

0 0 1

3

75

2

64Xsrg

Y srg

Zsrg

3

75 , (11)

where ⇡1 = (f1x, f1y, c1x, c1y) are the first camera pa-rameters. Hence, we have

@x

@⇠`=

f1x

Zsrg2

✓Zsrg @X

srg

@⇠`�Xsrg @Z

srg

@⇠`

◆. (12)

We similarly calculate @y@⇠`

. Also, each 3D point P srg isrelated to P

pre by P

srg = RP

pre + t. Therefore, for⇠` 2 {↵,�, �, tx, ty, tz}, @Xsrg

@⇠`, @Y srg

@⇠`and @Zsrg

@⇠`in (12)

are easily calculated upon the choice of transformationfunction. To update the non-rigid deformation of struc-tures, we need to calculate the derivatives of E withrespect to the shape parameters ui. These shape param-eters ui and the energy functional are related throughthe 2D coordinate x in E (eq.(4)) which are derivedfrom the corresponding 3D point in the surgical do-main P

srg = (Xsrg, Y srg, Zsrg) and the preoperative3D point P pre. According to (3), P pre depends on theweights of the modes of vibration, ui. Therefore, for theshape parameters ⇠` 2 {u1, · · · , uk}, the derivative ofa 3D point P

srg in the surgical domain with respectto ⇠` is @P srg

@⇠`= R · `, where ` is the `th mode of

vibration in . To avoid any irrational shape deforma-tion, we limit the shape parameters to vary not morethan three times the standard deviation (⇤i). For mul-tiple camera views (m > 1), the extrinsic parameters(Rext

m , t

extm ) have to be considered in calculating the

derivatives, i.e. R in the above equations is replaced byR

extm R. The boundary of the segmented structures in

Im are the zero level set of their corresponding level setfunctions �nm that are obtained by (2) after finding theoptimal T .

3 Materials and experiments

For validation, we applied our framework to fifteen dif-ferent clinical cases of robot assisted partial nephrec-tomy. All endoscopic videos were acquired by a da Vincir

Si surgical system (Intuitive Surgical, California, USA)and each frame was resized to 480⇥ 270 pixels for e�-ciency. The default parameters suggested in our previ-ous works [3,2] were used to detect the vascular motioncues. We used Nt = 70 trees to train the RF for learningthe appearance of kidney and tumour. Higher values ofNt did not improve accuracy but increased complexity.

We used the patient-specific 3D segmented kidney,tumor, and vasculature models and to set the tissue-specific modes of vibration, for simplicity, we assumedthat the structures can be modelled as a set of unitmasses mutually interconnected, i.e. M is the identitymatrix and can be removed from the eigendecomposi-tion equation in our case. We set the sti↵ness of each tis-sue to be proportional to its corresponding HU (higherHU means higher density and hence higher sti↵ness).HU is calculated from the preoperative DICOM meta-data as HU = CT pixel value⇥µs +µi, where µs andµi are the rescale slope and rescale intercept values thatare stored in the CT meta-data. We manually initializedT such that the projection of 3D models intersect theorgans. This initialization does not need to be close tothe solution. Fig. 2(c) shows the initial pose, which de-spite being not well placed, results in a reasonable poseas shown in Fig. 2(d). However, we emphasize that anirrational initialization will result in a wrong pose es-timation due to our local optimization framework. Weshould also mention that if most of the surface of theobjects is occluded in the 2D scene, our method cannotfind the correct pose. For our experiments, we askedsurgeons to stop moving the tools for ⇠ 10 seconds soour method can compute the regional term.

We compared the segmentations obtained throughour guidance system with the same ground truth pre-sented in [3,2] (Fig. 3 & 4). Note how the noisy segmen-tations in (c) are improved in (d) by incorporating thepreoperative prior information. We also quantitativelycompared our proposed method with [2] in Table 2.

The average runtime of our unoptimized MATLABcode to process the vessel pulsation in a four-secondclip (120 frames) was 65 seconds. The runtime for poseestimation and segmenting the structures depends on

Page 8: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

8 Masoud S. Nosrati et al.

Method DSC TPR FPR F1-measure AccuracyK V K V K V K V K V

Amir-Khalili et al. [2] - 0.41 - 0.74 - 0.40 - 0.41 - 0.60Our method 0.70 0.61 0.70 0.56 0.07 0.06 0.70 0.61 0.88 0.87

Table 2: Quantitative comparison for kidney (K) and vessel (V) segmentation: our method vs. [2]. DSC: Dice similaritycoe�cient; TPR: True positive ratio; FPR: False positive ratio.

(a) (b) (c) (d)

Fig. 3: First eight cases of the qualitative comparison of ourproposed method with state-of-the-art method [3]. (a) Origi-nal endoscopic image. (b) The ground truth of venous (cyan),arterial (red), kidney (brown) and tumour (green) structuresprovided in [3]. (c) Segmentation results of vessels using [3].(d) Our results. Kidney and tumour are shown in yellow andgreen, respectively.

the initial pose of the organs. The average runtime tofind the pose and segment the structures for an initial-ization similar to Fig. 2(c) is ⇠16 seconds on a standard3.40 GHz CPU.

4 Discussion and conclusions

We proposed a new technique for localizing both visi-ble and occluded structures in an endoscopic view byestimating the 3D pose and deformations of structuresof interest in the 3D surgical space. Our framework

(a) (b) (c) (d)

Fig. 4: Final seven cases of the qualitative comparison ofour proposed method with state-of-the-art method [3]. (a)Original endoscopic image. (b) The ground truth of venous(cyan), arterial (red), kidney (brown) and tumour (green)structures provided in [3]. (c) Segmentation results of vesselsusing [3]. (d) Our results. Kidney and tumour are shown inyellow and green, respectively.

leverages both preoperative data, as a source of patient-specific prior knowledge, as well as vasculature pulsa-tion (by analysing the local phase information) and en-doscopic visual cues (by training a random decision for-est) in order to accurately segment the highly noisy andcluttered environment of an endoscopic video. To han-dle the non-rigid deformation of di↵erent structures, weincorporated a tissue-specific physically-based deforma-tion model. To make the non-rigid deformation of eachstructure closer to reality, we used the HU value of eachstructure in the preoperative CT and assigned a spe-cific sti↵ness to each deformable model. Our results onin vivo clinical cases of partial nephrectomy illustrate

Page 9: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

Endoscopic Scene Augmentation using Appearance and Pulsatile Motion 9

the potential of the proposed framework for augmentedreality applications in MIS.

There are several directions to extend this work. Ourvariational framework is highly prallelizable and we donot foresee any obstacles towards a GPU implementa-tion for real-time pose estimation and endoscopic videosegmentation. In addition, we believe that leveragingstereo views as well as encoding depth information intothe proposed energy functional can improve the perfor-mance.

We attribute some of the observable di↵erences be-tween the ground truth and our results to both the localoptimization framework we used and also to the errorin the alignment of the ground truth. As mentioned inAmir-Khalili et al. [2], due to the fact that the preoper-ative model was rigidly aligned to the endoscopic video,an alignment error of 4�7 mm exist in cases where theorgans have been significantly retracted by the surgicalinstruments or mobilization of other organs. We believethat despite the visible di↵erences between the two, ourcurrent solution is one step closer to an ideal solutioncompared to the ground truth as our current methodallows for non-rigid modes of vibration. Generating aground truth that accounts for the non-rigid deforma-tions due to mobilization and retraction requires volu-metric intraoperative imaging such as cone beam CT orpossibly implanting fiducials. The use of such imagingtechniques is not feasible as it exposes the patient andclinicians to ionizing radiation and implanting fiducialsis intrusive and invasive and hence not recommended.

Also, in our future work, we will explore the use ofan additional shape variation component that is orthog-onal to the restricted shape model, as descried by An-drews and Hamarneh [4], since this allows for exploringlarger shape variability without noticeable increase incomplexity. Although we limited the modes of vibrationto vary not more than three times the correspondingeigenvalue (to avoid any irrational shape deformation),we still might get similar projection from two di↵erent3D deformations. This is due to the fact that we loseinformation during the 3D to 2D transformation. Webelieve that this is another interesting future directionthat worth investigation.

Given that in this proposed method we used a localoptimization technique, leveraging our own group andothers that have worked on convexification techniques[5,4,6,9,19,22] can make the method less sensitive (orinsensitive) to initialization.

Finally, improved estimates of elasticity parameters(e.g. using elastography imaging) will likely more accu-rately constrain the space of non-rigid deformations.

Funding: This publication was made possible byNPRP Grant #4-161-2-056 from the Qatar NationalResearch Fund (a member of the Qatar Foundation).The statements made herein are solely the responsibil-ity of the authors.

Ethical approval:All procedures performed in stud-ies involving human participants were in accordancewith the ethical standards of the institutional and/ornational research committee and with the 1964 Helsinkideclaration and its later amendments or comparableethical standards.

Conflict of Interest: The authors declare thatthey have no conflict of interest.

Informed consent: This articles does not containpatient information.

References

1. Agudo, A., Agapito, L., Calvo, B., Montiel, J.: Good vi-brations: A modal analysis approach for sequential non-rigid structure from motion. In: IEEE Conference onComputer Vision and Pattern Recognition (CVPR), pp.1558–1565 (2014)

2. Amir-Khalili, A., Hamarneh, G., Peyrat, J.M., Abina-hed, J., Al-Alao, O., Al-Ansari, A., Abugharbieh, R.:Automatic segmentation of occluded vasculature via pul-satile motion analysis in endoscopic robot-assisted partialnephrectomy video. Medical image analysis (2015)

3. Amir-Khalili, A., Peyrat, J.M., Abinahed, J., Al-Alao,O., Al-Ansari, A., Hamarneh, G., Abugharbieh, R.: Autolocalization and segmentation of occluded vessels inrobot-assisted partial nephrectomy. In: Medical ImageComputing and Computer-Assisted Intervention (MIC-CAI), pp. 407–414 (2014)

4. Andrews, S., Hamarneh, G.: The generalized log-ratiotransformation: Learning shape and adjacency priors forsimultaneous thigh muscle segmentation. IEEE Transac-tions on Medical Imaging (IEEE TMI) p. 1 (2015)

5. Andrews, S., McIntosh, C., Hamarneh, G.: Convex multi-region probabilistic segmentation with shape prior in theisometric log-ratio transformation space. In: IEEE In-ternational Conference on Computer Vision (ICCV), pp.2096–2103 (2011)

6. Brown, E., Chan, T., Bresson, X.: Convex formulationand exact global solutions for multi-phase piecewise con-stant Mumford-Shah image segmentation. UCLA CAMReport pp. 09–66 (2009)

7. Chan, T.F., Vese, L., et al.: Active contours withoutedges. IEEE transactions on Image processing 10(2),266–277 (2001)

8. Crane, N.J., et al.: Visual enhancement of laparoscopicpartial nephrectomy with 3-charge coupled device cam-era: assessing intraoperative tissue perfusion and vascularanatomy by visible hemoglobin spectral response. TheJournal of Urology 184(4), 1279–1285 (2010)

9. Delong, A., Boykov, Y.: Globally optimal segmentation ofmulti-region objects. In: IEEE International Conferenceon Computer Vision (IEEE ICCV), pp. 285–292 (2009)

Page 10: Endoscopic Scene Labelling and Augmentation using ...hamarneh/ecopy/ijcars2016.pdf · Endoscopic Scene Labelling and Augmentation using Intraoperative Pulsatile Motion and Colour

10 Masoud S. Nosrati et al.

10. Escudier, B., Kataja, V., et al.: Renal cell carcinoma:Esmo clinical practice guidelines for diagnosis, treatmentand follow-up. Ann Oncol 21(Suppl 5), v137–v139 (2010)

11. Estepar, R.S.J., Westin, C.F., Vosburgh, K.G.: Towardsreal time 2D to 3D registration for ultrasound-guidedendoscopic and laparoscopic procedures. InternationalJournal of Computer Assisted Radiology and Surgery4(6), 549–560 (2009)

12. Figueiredo, I.N., Figueiredo, P.N., Stadler, G., Ghattas,O., Araujo, A.: Variational image segmentation for endo-scopic human colonic aberrant crypt foci. IEEE Trans-actions on Medical Imaging 29(4), 998–1011 (2010)

13. Figueiredo, I.N., Moreno, J.C., Prasath, V.S., Figueiredo,P.N.: A segmentation model and application to endo-scopic images. In: Image Analysis and Recognition, pp.164–171 (2012)

14. Gill, I.S., Desai, M.M., Kaouk, J.H., Meraney, A.M.,Murphy, D.P., Sung, G.T., Novick, A.C.: Laparoscopicpartial nephrectomy for renal tumor: duplicating opensurgical techniques. The Journal of Urology 167(2), 469–476 (2002)

15. Gill, I.S., Kavoussi, L.R., Lane, B.R., Blute,M.L., Babineau, D., Colombo Jr, J.R., Frank, I.,Permpongkosol, S., Weight, C.J., Kaouk, J.H., et al.:Comparison of 1,800 laparoscopic and open partialnephrectomies for single renal tumors. The Journal ofUrology 178(1), 41–46 (2007)

16. Gill, S., Abolmaesumi, P., Vikal, S., Mousavi, P.,Fichtinger, G.: Intraoperative prostate tracking withslice-to-volume registration in MR. In: International Con-ference of the Society for Medical Innovation and Tech-nology, pp. 154–158 (2008)

17. Hernes, N., Toril, A., Lindseth, F., Selbekk, T., Wollf,A., Solberg, O.V., Harg, E., Rygh, O.M., Tangen, G.A.,Rasmussen, I., et al.: Computer-assisted 3D ultrasound-guided neurosurgery: technological contributions, in-cluding multimodal registration and advanced display,demonstrating future perspectives. The InternationalJournal of Medical Robotics and Computer AssistedSurgery 2(1), 45–59 (2006)

18. Hummel, J., Figl, M., Bax, M., Bergmann, H., Birkfell-ner, W.: 2D/3D registration of endoscopic ultrasoundto CT volume data. Physics in Medicine and Biology53(16), 4303 (2008)

19. McIntosh, C., Hamarneh, G.: Optimal weights for convexfunctionals in medical image segmentation. In: Advancesin Visual Computing, pp. 1079–1088. Springer (2009)

20. Merritt, S.A., Rai, L., Higgins, W.E.: Real-time CT-videoregistration for continuous endoscopic guidance. In: Med-ical Imaging, pp. 614,313–614,313 (2006)

21. Mewes, P.W., Neumann, D., Licegevic, O., Simon, J.,Juloski, A.L., Angelopoulou, E.: Automatic region-of-interest segmentation and pathology detection in mag-netically guided capsule endoscopy. In: Medical ImageComputing and Computer-Assisted Intervention (MIC-CAI), pp. 141–148 (2011)

22. Nosrati, M.S., Andrews, S., Hamarneh, G.: Bounded la-beling function for global segmentation of multi-part ob-jects with geometric constraints. In: IEEE InternationalConference on Computer Vision (ICCV), pp. 2032–2039(2013)

23. Nosrati, M.S., Peyrat, J.M., Abinahed, J., Al-Alao, O.,Al-Ansari, A., Abugharbieh, R., Hamarneh, G.: E�-cient multi-organ segmentation in multi-view endoscopicvideos using pre-operative priors. In: Medical ImageComputing and Computer-Assisted Intervention (MIC-CAI), pp. 324–331 (2014)

24. Pentland, A., Sclaro↵, S.: Closed-form solutions for phys-ically based shape modeling and recognition. IEEETransactions on Pattern Analysis & Machine Intelligence(IEEE TPAMI) (7), 715–729 (1991)

25. Pickering, M.R., Muhit, A.A., Scarvell, J.M., Smith,P.N.: A new multi-modal similarity measure for fastgradient-based 2D-3D image registration. In: IEEE En-gineering in Medicine and Biology Society (EMBC), pp.5821–5824 (2009)

26. Pratt, P., Mayer, E., Vale, J., Cohen, D., Edwards,E., Darzi, A., Yang, G.Z.: An e↵ective visualisationand registration system for image-guided robotic partialnephrectomy. Journal of Robotic Surgery 6(1), 23–31(2012)

27. Puerto-Souza, G.A., Mariottini, G.L.: Toward long-termand accurate augmented-reality display for minimally-invasive surgery. In: IEEE International Conference onRobotics and Automation (ICRA), pp. 5384–5389 (2013)

28. Singh, I.: Robot-assisted laparoscopic partial nephrec-tomy: Current review of the technique and literature.Journal of minimal access surgery 5(4), 87 (2009)

29. Teber, D., et al.: Augmented reality: a new tool toimprove surgical accuracy during laparoscopic partialnephrectomy? Preliminary in vitro and in vivo results.European Urology 56(2), 332–338 (2009)

30. Tobis, S., et al.: Near infrared fluorescence imaging withrobotic assisted laparoscopic partial nephrectomy: initialclinical experience for renal cortical tumors. The Journalof Urology 186(1), 47–52 (2011)

31. Urban, B.A., et al.: Three-dimensional volume-renderedCT angiography of the renal arteries and veins: Nor-mal anatomy, variants, and clinical applications. Radio-Graphics 21(2), 373–386 (2001)

32. Yim, Y., Wakid, M., Kirmizibayrak, C., Bielamowicz, S.,Hahn, J.: Registration of 3D CT data to 2D endoscopicimage using a gradient mutual information based view-point matching for image-guided medialization laryngo-plasty. Journal of Computing Science and Engineering4(4), 368–387 (2010)

33. Zikic, D., Glocker, B., Kutter, O., Groher, M., Ko-modakis, N., Khamene, A., Paragios, N., Navab, N.:Markov random field optimization for intensity-based2D-3D registration. In: Medical Imaging, pp. 762,334–762,334 (2010)