8
Classification of Tumor Histology via Morphometric Context Hang Chang 1 Alexander Borowsky 2 Paul Spellman 3 Bahram Parvin 1 1 Life Sciences Division, Lawrence Berkeley National Laboratory, Berkeley, California, U.S.A Corresponding authors {hchang, b parvin}@lbl.gov 2 Center for Comparative Medicine, UC Davis, Davis, California, USA [email protected] 3 Center for Spatial Systems Biomedicine, Oregon Health Sciences University, Portland, Oregon, U.S.A [email protected] Abstract Image-based classification of tissue histology, in terms of different components (e.g., normal signature, categories of aberrant signatures), provides a series of indices for tumor composition. Subsequently, aggregation of these indices in each whole slide image (WSI) from a large cohort can pro- vide predictive models of clinical outcome. However, the performance of the existing techniques is hindered as a re- sult of large technical and biological variations that are al- ways present in a large cohort. In this paper, we propose two algorithms for classification of tissue histology based on robust representations of morphometric context, which are built upon nuclear level morphometric features at vari- ous locations and scales within the spatial pyramid match- ing (SPM) framework. These methods have been evaluated on two distinct datasets of different tumor types collected from The Cancer Genome Atlas (TCGA), and the experi- mental results indicate that our methods are (i) extensible to different tumor types; (ii) robust in the presence of wide technical and biological variations; (iii) invariant to differ- ent nuclear segmentation strategies; and (iv) scalable with varying training sample size. In addition, our experiments suggest that enforcing sparsity, during the construction of morphometric context, further improves the performance of the system. 1. Introduction Histology sections provide a wealth of information about the tissue architecture that contains multiple cell types at different states of cell cycles. These sections are often This work was supported by NIH U24 CA1437991 carried out at Lawrence Berkeley National Laboratory under Contract No. DE-AC02- 05CH11231. stained with hematoxylin and eosin (H&E) stains, which label DNA (e.g., nuclei) and protein contents, respectively, in various shades of color. Abberations in the histology ar- chitecture often lead to disease progression. It is desirable to quantify indices associated with these abberations since they can be tested against the clinical outcome, e.g., sur- vival, response to therapy. Even though there are inter- and intra- observer variations [7], a trained pathologist always uses rich content (e.g., various cell types, cellular organiza- tion, cell state and health), in context, to characterize tumor architecture. In this paper, we propose two tissue classification meth- ods based on representations of morphometric context, which are constructed from nuclear morphometric statistics of various locations and scales based on spatial pyramid matching (SPM) [18] and linear spatial pyramid matching (Linear SPM) [28]. Due to the effectiveness of our repre- sentations, our methods achieve excellent performance even with small number of training samples across different seg- mentation strategies and independent datasets of tumors. The performance is further complemented by the fact that one of the methods has a superior result with linear clas- sifiers. These characteristics dramatically improve the (i) effectiveness of our techniques when applied to a large co- hort, and (ii) extensibility to other cell-based assays. Organization of this paper is as follows: Section 2 re- views related works. Sections 3 and 4 describes the de- tails of our proposed mehods and experimental validation. Lastly, section 5 concludes the paper. 2. Related Work For the analysis of the H&E stained sections, several ex- cellent reviews can be found in [13, 8]. Fundamentally, the trend has been based either on nuclear segmentation and corresponding morphometric representaion, or patch-based 2201 2201 2203

Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

Classification of Tumor Histology via Morphometric Context ∗

Hang Chang1� Alexander Borowsky2 Paul Spellman3 Bahram Parvin1�

1Life Sciences Division, Lawrence Berkeley National Laboratory, Berkeley, California, U.S.A� Corresponding authors {hchang, b parvin}@lbl.gov

2Center for Comparative Medicine, UC Davis, Davis, California, [email protected]

3 Center for Spatial Systems Biomedicine, Oregon Health Sciences University, Portland, Oregon, [email protected]

Abstract

Image-based classification of tissue histology, in terms ofdifferent components (e.g., normal signature, categories ofaberrant signatures), provides a series of indices for tumorcomposition. Subsequently, aggregation of these indices ineach whole slide image (WSI) from a large cohort can pro-vide predictive models of clinical outcome. However, theperformance of the existing techniques is hindered as a re-sult of large technical and biological variations that are al-ways present in a large cohort. In this paper, we proposetwo algorithms for classification of tissue histology basedon robust representations of morphometric context, whichare built upon nuclear level morphometric features at vari-ous locations and scales within the spatial pyramid match-ing (SPM) framework. These methods have been evaluatedon two distinct datasets of different tumor types collectedfrom The Cancer Genome Atlas (TCGA), and the experi-mental results indicate that our methods are (i) extensibleto different tumor types; (ii) robust in the presence of widetechnical and biological variations; (iii) invariant to differ-ent nuclear segmentation strategies; and (iv) scalable withvarying training sample size. In addition, our experimentssuggest that enforcing sparsity, during the construction ofmorphometric context, further improves the performance ofthe system.

1. Introduction

Histology sections provide a wealth of information about

the tissue architecture that contains multiple cell types at

different states of cell cycles. These sections are often

∗This work was supported by NIH U24 CA1437991 carried out at

Lawrence Berkeley National Laboratory under Contract No. DE-AC02-

05CH11231.

stained with hematoxylin and eosin (H&E) stains, which

label DNA (e.g., nuclei) and protein contents, respectively,

in various shades of color. Abberations in the histology ar-

chitecture often lead to disease progression. It is desirable

to quantify indices associated with these abberations since

they can be tested against the clinical outcome, e.g., sur-

vival, response to therapy. Even though there are inter- and

intra- observer variations [7], a trained pathologist always

uses rich content (e.g., various cell types, cellular organiza-

tion, cell state and health), in context, to characterize tumor

architecture.

In this paper, we propose two tissue classification meth-

ods based on representations of morphometric context,

which are constructed from nuclear morphometric statistics

of various locations and scales based on spatial pyramid

matching (SPM) [18] and linear spatial pyramid matching

(Linear SPM) [28]. Due to the effectiveness of our repre-

sentations, our methods achieve excellent performance even

with small number of training samples across different seg-

mentation strategies and independent datasets of tumors.

The performance is further complemented by the fact that

one of the methods has a superior result with linear clas-

sifiers. These characteristics dramatically improve the (i)

effectiveness of our techniques when applied to a large co-

hort, and (ii) extensibility to other cell-based assays.

Organization of this paper is as follows: Section 2 re-

views related works. Sections 3 and 4 describes the de-

tails of our proposed mehods and experimental validation.

Lastly, section 5 concludes the paper.

2. Related Work

For the analysis of the H&E stained sections, several ex-

cellent reviews can be found in [13, 8]. Fundamentally, the

trend has been based either on nuclear segmentation and

corresponding morphometric representaion, or patch-based

2013 IEEE Conference on Computer Vision and Pattern Recognition

1063-6919/13 $26.00 © 2013 IEEE

DOI 10.1109/CVPR.2013.286

2201

2013 IEEE Conference on Computer Vision and Pattern Recognition

1063-6919/13 $26.00 © 2013 IEEE

DOI 10.1109/CVPR.2013.286

2201

2013 IEEE Conference on Computer Vision and Pattern Recognition

1063-6919/13 $26.00 © 2013 IEEE

DOI 10.1109/CVPR.2013.286

2203

Page 2: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

Figure 1. Computational workflows: (a) Morphometric nonlinear

kernel SPM; (b) Sparse Morphometric linear SPM. In both ap-

proaches, the nuclear segmentation could be based on any of the

existing methods.

representation of the histology sections that aids in clinical

association. For example, a recent study indicates that de-

tailed segmentation and multivariate representation of nu-

clear features from H&E stained sections can predict DCIS

recurrence [1] in patients with more than one nuclear grade.

The major challenge for tissue classification is the large

amounts of technical and biological variations in the data,

which typically results in techniques that are tumor type

specific. To overcome this problem, recent studies have fo-

cused on either fine tuning human engineered features [2,

16, 17], or applying automatic feature learning [19, 14], for

robust representation.

In the context of image categorization research, the tra-

ditional bag of features (BoF) model has been widely stud-

ied and improved through different variations, e.g., mod-

eling of co-occurrence of descriptors based on generative

methods [4, 3, 20, 24], improving dictionary construction

through discriminative learning [9, 22], modeling the spatial

layout of local descriptors based on spatial pyramid match-

ing kernel (SPM) [18]. It is clear that SPM has become

the major component of the state-of-art systems [10] for its

effectiveness in practice.

Pathologists often use “context” to assess the disease

state. At the same time, SPM partially captures context be-

cause of its hierarchical nature. Motivated by the works of

[18, 28], we encode morphometric signatures, at different

locations and scales, within the SPM framework. The end

results are highly robust and effective systems across multi-

ple tumor types with limited number of training samples.

3. ApproachesThe computational workflows for the proposed methods

are shown in Figure 1, where the nuclear segmentation can

be based on any of the existing methods for delineating nu-

clei from background. For some tissue images and their

corresponding nuclear segmentation, let:

1. M be the total number of segmented nuclei;

2. N be the number of morphometric descriptors ex-

tracted from each segmented nucleus, e.g. nuclear size,

and nuclear intensity;

3. X be the set of morphometric descriptors for all seg-

mented nuclei, where X = [x1, ...,xM ]� ∈ RM×N .

Our proposed methods are described in detail as follows.

3.1. Morphometric nonlinear kernel SPM (MK-SPM)

In this approach, we utilize the nuclear morphometric in-

formation within the SPM framework to construct the mor-

phometric context at various locations and scales for tissue

image representation and classification. It consists of the

following steps:

1. Construct dictionary (D), where D = [d1, ...,dK ]�

are the K nuclear morphometric types to be learned

by the following optimization:

minD,Z

M∑m=1

||xm − zmD||2 (1)

subject to card(zm) = 1, |zm| = 1, zm � 0, ∀mwhere Z = [z1, ..., zM ]� indicates the assignment of

the nuclear morphometric type, card(zm) is a cardi-

nality constraint enforcing only one nonzero element

of zm, zm � 0 is a non-negative constraint on the ele-

ments of zm, and |zm| is the L1-norm of zm. During

training, Equation 1 is optimized with respect to both

Z and D; In the coding phase, for a new set of X, the

learned D is applied, and Equation 1 is optimized with

respect to Z only.

2. Construct spatial histogram as the descriptor for the

morphometric context for SPM [18]. This is done by

repeatedly subdividing an image and computing the

histograms of different nuclear morphometric types

over the resulting subregions. As a result, the spatial

histogram, H , is formed by concatenating the appro-

priately weighted histograms of all nuclear morpho-

metric types at all resolutions. The SPM kernel is then

implemented as a single histogram intersection as fol-

lows,

κ(Hi, Hj) =

Q∑q=1

min(Hi(q), Hj(q)) (2)

where Hi and Hj are the spatial histograms for image

Ii and Ij , respectively, and Q is the length of the spatial

220222022204

Page 3: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

histogram. For more details about SPM, please refer to

[18, 12].

3. Transfer a χ2 support vector machine (SVM) into a

linear SVM based on a homogeneous kernel map [26].

In practice, the intersection kernel and χ2 kernel have

been found to be the most suitable for histogram repre-

sentations [28]. In this step, a homogenous kernel map

is applied to approximate the χ2 kernels, which en-

ables the efficiency by adopting learning methods for

linear kernels, e.g., linear SVM. For more details about

the homogeneous kernel map, please refer to [26].

4. Construct multi-class linear SVM for classification. In

our implementation, the classifier is trained using the

LIBLINEAR [11] package.

3.2. Sparse Morphometric Linear SPM (SMLSPM)

Inspired by the work in [28], which indicates that the

sparse coding of the SIFT features may serve as a better

local descriptor in general image processing tasks, we pro-

pose to utilize the sparse coding of the nuclear morphomet-

ric information within a linear SPM framework to construct

the morphometric context at various locations and scales for

tissue image representation and classification. It consists of

the following steps:

1. Construct dictionary (D), where D = [d1, ...,dK ]�

are the K nuclear morphmetric types to be learned by

the following sparse coding optimization:

minD,Z

M∑m=1

||xm − zmD||2 + λ|zm| (3)

subject to ||dk|| ≤ 1, ∀kwhere ||dk|| is a unit L2-norm constraint for avoid-

ing trivial solutions, and |zm| is the L1-norm enforc-

ing the sparsity of zm. During training, Equation 3 is

optimized with respect to both Z and D; In the coding

phase, the learned D will be applied to a new set of X,

with Equation 3 optimized with respect to Z only.

2. Construct spatial pyramid representation as the de-

scriptor of morphometric context for the linear

SPM [28]. Let Z be the sparse codes calculated

through Equation 3 for a descriptor set X. Based on

pre-learned and fixed dictionary D, the image descrip-

tor is computed based on a pre-determined pooling

function as follows,

f = P(Z) (4)

In our implementation, P is selected to be the maxpooling function on the absolute sparse codes

fj = max{|z1j |, |z2j |, ..., |zMj |} (5)

where fj is the j-th element of f , zij is the matrix el-

ement at i-th row and j-th column of Z, and M is the

number of nuclei in the region. The choice of maxpooling procedure is justified by biophysical evidence

in the visual cortex [25], algorithms in image catego-

rization [28], and our experimental comparison with

other common pooling strategies (see Table 7). Similar

to the construction of SPM, the pooled features from

various locations and scales are then concatenated to

form a spatial pyramid representation of the image, and

a linear SPM kernel is applied as follows,

κ(fi, fj) = f�i fj =2∑

l=0

2l∑s=1

2l∑t=1

〈f li (s, t), f lj(s, t)〉 (6)

where fi and fj are spatial pyramid representations for

image Ii and Ij , respectively, 〈fi, fj〉 = f�i fj , and

f li (s, t) and f lj(s, t) are the max pooling statistics of

the sparse codes in the (s, t)-th segment of image Iiand Ij in the scale level l, respectively.

3. Construct multi-class linear SVM for classification. In

our implementation, the classifier is trained using the

LIBLINEAR [11] package.

4. Experiments And ResultsWe have evaluated five classification methods on two

distinct datasets, curated from (i) Glioblastoma Multiforme

(GBM) and (ii) Kidney Renal Clear Cell Carcinoma (KIRC)

from The Cancer Genome Atlas (TCGA), which are pub-

licly available from the NIH (National Institute of Health)

repository. The five methods are:

1. SMLSPM: the linear SPM that uses linear kernel

on spatial-pyramid pooling of morphometric sparse

codes;

2. MKSPM: the nonlinear kernel SPM that uses spatial-

pyramid histograms of morphometric features and χ2

kernels;

3. ScSPM [28]: the linear SPM that uses linear kernel on

spatial-pyramid pooling of SIFT sparse codes;

4. KSPM [18]: the nonlinear kernel SPM that uses

spatial-pyramid histograms of SIFT features and χ2

kernels;

5. CTSPM: the nonlinear kernel SPM that uses spatial-

pyramid histograms of color and texture features and

χ2 kernels;

In the implementations of SMLSPM and MKSPM, mor-

phometric features were extracted and normalized indepen-

dently with zero mean and unit variance based on three dif-

ferent segmentation strategies:

220322032205

Page 4: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

Approach Precision Recall F-Measure

MRGC 0.75 0.85 0.797SRCD 0.63 0.75 0.685

OTGR NA NA NA

Table 1. Comparison of average segmentation performance among

MRGC [6], SRCD [5], and OTGR. Note: 1) the information above

is quoted from [6]; 2) the OTGR performance is not listed due to

the unavailability of the data used in [6], however, based on our ex-

perience with histological sections, simple thresholding methods

typically generate less favorable results than sophisticated ones.

1. MRGC [6]: A multi-reference graph cut approach for

nuclear segmentation in histology tissue sections;

2. SRCD [5]: A single-reference color decomposition

approach for nuclear segmentation in histology tissue

sections;

3. OTGR: A simple Otsu thresholding [23] approach for

nuclear segmentation in histology tissue sections. In

our implementation, nuclear mask was generated by

applying Otsu thresholding on gray-scale image, and

refined by geometric reasoning [27]. The same re-

finement was also used in the MRGC and SRCD ap-

proaches.

A comparison of the segmentation performance, for the

above methods, is quoted from [6], and listed in Table 1,

and the computed morphometric features are listed in Ta-

ble 2.

In the implementation of ScSPM and KSPM, the dense

SIFT features were extracted on 16 × 16 patches sampled

from each image on a grid with stepsize 8 pixels. In the

implementation of CTSPM, color features were extracted

in the RGB color space; texture features were extracted

via steerable filters [29] with 4 directions and 5 scales

(σ ∈ {1, 2, 3, 4, 5}) on the grayscale image; and the fea-

ture vector was a concatenation of texture and mean color

on 20× 20 patches.

For both SMLSPM and ScSPM, we fixed the sparse con-

straint parameter λ to be 0.15, empirically, to achieve the

best performance. For MKSPM, KSPM and CTSPM, we

used standard K-means clustering for the construction of

dictionaries. Additionally, for all five methods, we fixed the

level of pyramid to be 3, and used linear SVM for classifi-

cation. All experimental processes were repeated 10 times

with randomly selected training and testing images. The fi-

nal results were reported as the mean and standard deviation

of the classification rates.

4.1. GBM Dataset

The GBM dataset contains 3 classes: Tumor, Necrosis,

and Transition to Necrosis, which were curated from whole

MRGC SRCD OTGR

SMLSPM 92.91 ± 0.84 93.56 ± 0.91 91.03 ± 1.15

MKSPM 91.95 ± 0.83 93.33 ± 0.90 90.94 ± 0.87

Table 4. Performance of SMLSPM and MKSPM on the GBM

dataset based on three different segmentation approaches, where

the number of training images per category was fixed to be 160,

and the dictionary sizes for SMLSPM and MKSPM were fixed to

be 1024 and 512, respectively, to achieve optimal performance.

MRGC SRCD OTGR

SMLSPM 98.50 ± 0.42 98.30 ± 0.34 97.66 ± 0.49

MKSPM 97.34 ± 0.48 97.66 ± 0.45 95.90 ± 0.54

Table 6. Performance of SMLSPM and MKSPM on the KIRC

dataset based on three different segmentation approaches, where

the number of training images per category was fixed to be 280,

and the dictionary sizes for both SMLSPM and MKSPM were

fixed to be 512, to achieve the optimal performance.

slide images (WSI) scanned with a 20X objective (0.502

micron/pixel). Examples can be found in Figure 2. The

number of images per category are 628, 428 and 324, re-

spectively. Most images are 1000× 1000 pixels. In this ex-

periment, we trained on 40, 80 and 160 images per category

and tested on the rest, with three different dictionary sizes:

256, 512 and 1024. Detailed comparisons are shown in Ta-

ble 3. For SMLSPM and MKSPM, we also evaluated the

performance based on three different segmentation strate-

gies: MRGC, SRCD and OTGR. Detailed comparisons are

shown in Table 4.

4.2. KIRC Dataset

The KIRC dataset contains 3 classes: Tumor, Normal,

and Stromal, which were curated from whole slide images

(WSI) scanned with a 40X objective (0.252 micron/pixel).

Examples can be found in Figure 3. The number of im-

ages per category are 568, 796 and 784, respectively. Most

images are 1000 × 1000 pixels. In this experiment, we

trained on 70, 140 and 280 images per category and tested

on the rest, with three different dictionary sizes: 256, 512

and 1024. Detailed comparisons are shown in Table 5. For

SMLSPM and MKSPM, we also evaluated the performance

based on three different segmentation strategies: MRGC,

SRCD and OTGR. Detailed comparisons are shown in Ta-

ble 6

The experiments, conducted on two distinct datasets,

demonstrate the following merits of our approach,

1. Extensibility to different tumor types. Tables 3 and 5

indicate that, with the exception of (KIRC; 140 train-

ing; Dictionary size 1024), our methods consistently

outperform ScSPM, KSPM and CTSPM with differ-

ent combinations of experimental factors (e.g., train-

220422042206

Page 5: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

Feature DescriptionNuclear Size #pixels of a segmented nucleus

Nuclear Voronoi Size #pixels of the voronoi region, where the segmented nucleus resides

Aspect Ratio Aspect ratio of the segmented nucleus

Major Axis Length of Major axis of the segmented nucleus

Minor Axis Length of Minor axis of the segmented nucleus

Rotation Angle between major axis and x axis of the segmented nucleus

Bending Energy Mean squared curvature values along nuclear contour

STD Curvature Standard deviation of absolute curvature values along nuclear contour

Abs Max Curvature Maximum absolute curvature values along nuclear contour

Mean Nuclear Intensity Mean intensity in nuclear region measured in gray scale

STD Nuclear Intensity Standard deviation of intensity in nuclear region measured in gray scale

Mean Background Intensity Mean intensity of nuclear background measured in gray scale

STD Background Intensity Standard deviation of intensity of nuclear background measured in gray scale

Mean Nuclear Gradient Mean gradient within nuclear region measured in gray scale

STD Nuclear Gradient Standard deviation of gradient within nuclear region measured in gray scale

Table 2. Morphometric features used in SMLSPM and MKSPM, where the curvature values were computed with σ = 2.0, and the nuclear

background is defined to be outside the nuclear region, but inside the bounding box of nuclear boundary.

Figure 2. GBM Examples. First row: Tumor; Second row: Transition to necrosis; Third row: Necrosis.

Sqrt Abs Max

GBM 92.85 ± 0.94 90.90 ± 1.11 92.91 ± 0.84KIRC 97.60 ± 0.49 97.49 ± 0.38 98.50 ± 0.42

Table 7. Comparison of performance for SMLSPM using differ-

ent pooling strategies on the GBM and KIRC datasets. For GBM,

the number of training images per category was fixed to be 160,

and the dictionary size was fixed to be 1024; for KIRC, the num-

ber of training images per category was fixed to be 280, and the

dictionary size was fixed to be 512.

ing sample size, dictionary size). However, KSPM and

ScSPM appear to be tumor-type dependent, as KSPM

outperforms ScSPM on GBM while ScSPM outper-

forms KSPM on KIRC. Since GBM and KIRC are two

vastly different tumor types with significantly different

signatures, we suggest that the consistency in perfor-

mance assures extensibility to different tumor types.

2. Robustness in the presence of large amounts of tech-

nical and biological variations. With respect to the

GBM dataset, shown in Table 3, the performance of

our methods, based on 40 training samples per cate-

220522052207

Page 6: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

Method DictionarySize=256 DictionarySize=512 DictionarySize=1024

160 training SMLSPM 92.35 ± 0.83 92.57 ± 0.91 92.91 ± 0.84MKSPM 91.85 ± 0.98 91.95 ± 0.83 91.76 ± 0.97

ScSPM [28] 79.58 ± 0.61 81.29 ± 0.86 82.36 ± 1.10

KSPM [18] 85.00 ± 0.79 86.47 ± 0.55 86.81 ± 0.45

CTSPM 78.61 ± 1.33 78.71 ± 1.18 78.69 ± 0.81

80 training SMLSPM 90.82 ± 1.28 90.29 ± 0.68 91.08 ± 0.69MKSPM 89.83 ± 1.15 89.78 ± 1.09 89.44 ± 1.20

ScSPM [28] 77.65 ± 1.43 78.31 ± 1.13 81.00 ± 0.98

KSPM [18] 83.81 ± 1.22 84.32 ± 0.67 84.49 ± 0.34

CTSPM 75.93 ± 1.18 76.06 ± 1.52 76.19 ± 1.33

40 training SMLSPM 88.05 ± 1.38 87.88 ± 1.04 88.54 ± 1.42MKSPM 87.38 ± 1.38 87.06 ± 1.52 86.37 ± 1.73

ScSPM [28] 73.60 ± 1.68 75.58 ± 1.29 76.24 ± 3.05

KSPM [18] 80.54 ± 1.21 80.56 ± 1.24 80.46 ± 0.56

CTSPM 73.10 ± 1.51 72.90 ± 1.09 72.65 ± 1.41

Table 3. Performance of different methods on the GBM dataset, where SMLSPM and MKSPM were evaluated based on the segmentation

method: MRGC [6].

Figure 3. KIRC Examples. First row: Tumor; Second row: Normal; Third row: Stromal.

gory, is better than the performance of ScSPM, KSPM

and CTSPM based on 160 training samples per cat-

egory. With respect to the KIRC dataset, shown in

Table 5, the performance of our methods, based on

70 training samples per category, is comparable to the

performance of ScSPM, KSPM and CTSPM based on

280 training samples per category. Given the fact that

TCGA datasets contain large amount of technical and

biological variations [17, 19], these results clearly indi-

cate the robustness of our morphometric context repre-

sentation, which dramatically improved the reliability

of our approaches.

3. Invariance to different segmentation algorithms. Ta-

bles 4 and 6 indicate that the performance of our ap-

proaches are almost invariant to different segmentation

algorithms, given the fact that the segmentation perfor-

mance itself varies a lot, as shown in Table 1. More im-

portantly, even with the simplest segmentation strategy

OTGR, SMLSPM outperforms the methods in [28, 18]

on both datasets, and MRSPM outperforms the meth-

ods in [28, 18] on the GBM dataset, while generat-

ing comparable results on the KIRC dataset. Given the

fact that, in a lot of studies, both nuclear segmentation

and tissue classification are necessary components, the

220622062208

Page 7: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

Method DictionarySize=256 DictionarySize=512 DictionarySize=1024

280 training SMLSPM 98.15 ± 0.46 98.50 ± 0.42 98.21 ± 0.44

MKSPM 97.37 ± 0.49 97.34 ± 0.48 97.22 ± 0.50

ScSPM [28] 94.52 ± 0.44 96.37 ± 0.45 96.81 ± 0.50

KSPM [18] 93.55 ± 0.31 93.76 ± 0.27 93.90 ± 0.19

CTSPM 87.45 ± 0.59 87.95 ± 0.49 88.53 ± 0.49

140 training SMLSPM 97.40 ± 0.50 97.98 ± 0.35 97.35 ± 0.48

MKSPM 96.56 ± 0.53 96.54 ± 0.50 96.41 ± 0.56

ScSPM [28] 93.46 ± 0.55 95.68 ± 0.36 96.76 ± 0.63

KSPM [18] 92.50 ± 1.12 93.06 ± 0.82 93.26 ± 0.68

CTSPM 86.55 ± 0.99 86.40 ± 0.54 86.49 ± 0.58

70 training SMLSPM 96.20 ± 0.85 96.37 ± 0.85 96.19 ± 0.62

MKSPM 95.62 ± 0.62 95.47 ± 0.71 95.27 ± 0.72

ScSPM [28] 91.93 ± 1.00 93.67 ± 0.72 94.86 ± 0.86

KSPM [18] 90.78 ± 0.98 91.34 ± 1.13 91.59 ± 0.97

CTSPM 84.76 ± 1.32 84.29 ± 1.53 83.71 ± 1.42

Table 5. Performance of different methods on the KIRC dataset, where SMLSPM and MKSPM were evaluated based on the segmentation

mehtod: MRGC [6].

use of pre-computed morphometric features dramati-

cally improve the efficiency by avoiding extra feature

extraction steps. For example, in our implementation,

SIFT costs 1.5 sec/block ( a block is a 1000 × 1000image decomposed from a whole slide tissue section

). For the whole GBM dataset (∼600,000 blocks), by

avoiding SIFT operation, it saves ∼10 days for pro-

cessing.

4. Scalability of training and high speed testing for SML-

SPM. Our experiments show that the morphometric

context representation in SMLSPM works well with

linear SVMs, which dramatically improves the scala-

bility of training and the speed of testing. This is very

important for the analyzing large cohort of whole slide

images.

To study the impact of pooling strategies on the SML-

SPM method, we also provide an experimental comparison

among max pooling and two other common pooling meth-

ods, which are defined as follows,

Sqrt : fj =

√√√√ 1

M

M∑i=1

z2ij

Abs : fj =1

M

M∑i=1

|zij | (7)

where the meaning of the notations are the same as in Equa-

tion 5. As shown in Table 7, the max pooling strategy out-

performs the other two, which is probably due to its robust-

ness to local translations.

The experiments above also indicate a slightly improved

performance of SMLSPM over MKSPM; this is probably

due to the following factors: i) sparse coding has much less

quantization errors than vector quantization; ii) the statis-

tics derived by max pooling are more robust to local trans-

lations compared with the average pooling in the histogram

representation.

5. Conclusion and Future WorkIn this paper, we proposed two spatial pyramid matching

approaches based on morphometric features and morpho-

metric sparse code, respectively, for tissue image classifica-

tion. By modeling the context of the morphometric infor-

mation, these methods outperform traditional ones which

were typically based on pixel- or patch-level features. The

most encouraging results of this paper are that, our methods

are highly i) extensible to different tumor types; ii) robust

in the presence of large amounts of technical and biolog-

ical variations; iii) invariant to different segmentation al-

gorithms; and iv) scalable to extremely large training and

testing sample sizes. Due to i) the effectiveness of our

morphometric context representations; and ii) the impor-

tant role of cellular context for the study of different cell

assays, proposed methods are suggested to be extendable

to image classification tasks for different cell assays. Fu-

ture work will be focused on accelerating the sparse coding

through the sparse auto encoder [15], utilizing supervised

dictionary learning [21] for possible improvement, and fur-

ther validating our methods on other tissue types and other

cell assays.

References[1] D. Axelrod, N. Miller, H. Lickley, J. Qian, W. Christens-

Barry, Y. Yuan, Y. Fu, and J. Chapman. Effect of quantitative

220722072209

Page 8: Classification of Tumor Histology via Morphometric Contextopenaccess.thecvf.com/content_cvpr_2013/papers/Chang_Classification_of... · In this paper, we propose two tissue classification

nuclear features on recurrence of ductal carcinoma in situ

(dcis) of breast. In Cancer Informatics, 4:99–109, 2008. 2

[2] R. Bhagavatula, M. Fickus, W. Kelly, C. Guo, J. Ozolek,

C. Castro, and J. Kovacevic. Automatic identification and

delineation of germ layer components in h&e stained im-

ages of teratomas derived from human and nonhuman pri-

mate embryonic stem cells. In ISBI, pages 1041–1044, 2010.

2

[3] O. Boiman, E. Shechtman, and M. Irani. In defense of

nearest-neighbor based image classification. In Proceedingsof the Conference on Computer Vision and Pattern Recogni-tion, pages 1–8, 2008. 2

[4] A. Bosch, A. Zisserman, and X. Munoz. Scene classifica-

tion using a hybrid generative/discriminative approach. IEEETransactions on Pattern Analysis and Machine Intelligence,

30(4):712–727, Apr. 2008. 2

[5] H. Chang, G. Fontenay, J. Han, G. Cong, F. Baehner, J. Gray,

P. Spellman, and B. Parvin. Morphometric analysis of TCGA

Gliobastoma Multiforme. BMC Bioinformatics, 12(1), 2011.

4

[6] H. Chang, L. Loss, P. Spellman, A. Borowsky, and B. Parvin.

Batch-invariant nuclear segmentation in whole mount histol-

ogy sections. In ISBI, pages 856–859, 2012. 4, 6, 7

[7] L. Dalton, S. Pinder, C. Elston, I. Ellis, D. Page, W. Dupont,

and R. Blamey. Histolgical gradings of breast cancer: link-

age of patient outcome with level of pathologist agreements.

Modern Pathology, 13(7):730–735, 2000. 1

[8] C. Demir and B. Yener. Automated cancer diagnosis based

on histopathological images: A systematic survey, 2009. 1

[9] M. Elad and M. Aharon. Image denoising via sparse and

redundant representations over learned dictionaries. IEEETransactions on Image Processing, 15(12):3736–3745, Dec.

2006. 2

[10] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn,

and A. Zisserman. The PASCAL Visual Object Classes

Challenge 2012 (VOC2012) Results. http://www.pascal-

network.org/challenges/VOC/voc2012/workshop/index.html.

2

[11] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J.

Lin. LIBLINEAR: A library for large linear classification.

Journal of Machine Learning Research, 9:1871–1874, 2008.

3

[12] K. Grauman and T. Darrell. The pyramid match kernel: dis-

criminative classification with sets of image features. In Pro-ceedings of the IEEE International Conference on ComputerVision, volume 2, pages 1458 – 1465, 2005. 3

[13] M. Gurcan, L. Boucheron, A. Can, A. Madabhushi, N. Ra-

jpoot, and Y. Bulent. Histopathological image analysis:

a review. IEEE Transactions on Biomedical Engineering,

2:147–171, 2009. 1

[14] C. Huang, A. Veillard, N. Lomeine, D. Racoceanu, and

L. Roux. Time efficient sparse analysis of histopathologi-

cal whole slide images. Computerized medical imaging andgraphics, 35(7-8):579–591, 2011. 2

[15] K. Kavukcuoglu, M. Ranzato, and Y. LeCun. Fast infer-

ence in sparse coding algorithms with applications to object

recognition. Technical Report CBLL-TR-2008-12-01, Com-

putational and Biological Learning Lab, Courant Institute,

NYU, 2008. 7

[16] J. Kong, L. Cooper, A. Sharma, T. Kurk, D. Brat, and J. Saltz.

Texture based image recognition in microscopy images of

diffuse gliomas with multi-class gentle boosting mechanism.

In ICASSAP, pages 457–460, 2010. 2

[17] S. Kothari, J. Phan, A. Osunkoya, and M. Wang. Biological

interpretation of morphological patterns in histopathological

whole slide images. In ACM Conference on Bioinformatics,Computational Biology and Biomedicine, 2012. 2, 6

[18] S. Lazebnik, C. Schmid, and J. Ponce. Beyond bags of

features: Spatial pyramid matching for recognizing natural

scene categories. In Proceedings of the Conference on Com-puter Vision and Pattern Recognition, pages 2169–2178,

2006. 1, 2, 3, 6, 7

[19] Q. Le, J. Han, J. Gray, P. Spellman, A. Borowsky, and

B. Parvin. Learning invariant features from tumor signature.

In ISBI, pages 302–305, 2012. 2, 6

[20] F.-F. Li and P. Perona. A bayesian hierarchical model for

learning natural scene categories. In Proceedings of the Con-ference on Computer Vision and Pattern Recognition, pages

524–531, Washington, DC, USA, 2005. IEEE Computer So-

ciety. 2

[21] J. Mairal, F. Bach, J. Ponce, G. Sapiro, and A. Zisserman.

Supervised dictionary learning. In NIPS, 2008. 7

[22] F. Moosmann, B. Triggs, and F. Jurie. Randomized cluster-

ing forests for building fast and discriminative visual vocab-

ularies. In NIPS, 2006. 2

[23] N. Otsu. A threshold selection method from gray-level his-

tograms. IEEE Transactions on Systems, Man and Cybernet-ics, 9(1):62–66, 1979. 4

[24] P. Quelhas, F. Monay, J.-M. Odobez, D. Gatica-Perez,

T. Tuytelaars, and L. Van Gool. Modeling scenes with local

descriptors and latent aspects. In Proceedings of the IEEEInternational Conference on Computer Vision, ICCV ’05,

pages 883–890, Washington, DC, USA, 2005. IEEE Com-

puter Society. 2

[25] T. Serre, L. Wolf, and T. Poggio. Object recognition with fea-

tures inspired by visual cortex. In Proceedings of the Confer-ence on Computer Vision and Pattern Recognition, volume 2,

pages 994–1000, 2005. 3

[26] A. Vedaldi and A. Zisserman. Efficient additive kernels via

explicit feature maps. IEEE Transactions on Pattern Analysisand Machine Intelligence, 34(3):480–492, 2012. 3

[27] Q. Wen, H. Chang, and B. Parvin. A Delaunay triangulation

approach for segmenting clumps of nuclei. In ISBI, pages

9–12, 2009. 4

[28] J. Yang, K. Yu, Y. Gong, and T. Huang. Linear spatial pyra-

mid matching using sparse coding for image classification.

In Proceedings of the Conference on Computer Vision andPattern Recognition, pages 1794–1801, 2009. 1, 2, 3, 6, 7

[29] R. A. Young and R. M. Lesperance. The gaussian derivative

model for spatial-temporal vision. I. Cortical Model. SpatialVision, 2001:3–4, 2001. 4

220822082210