Skip to main content
Erschienen in: BMC Cancer 1/2011

Open Access 01.12.2011 | Research article

Multimodal microscopy for automated histologic analysis of prostate cancer

verfasst von: Jin Tae Kwak, Stephen M Hewitt, Saurabh Sinha, Rohit Bhargava

Erschienen in: BMC Cancer | Ausgabe 1/2011

Abstract

Background

Prostate cancer is the single most prevalent cancer in US men whose gold standard of diagnosis is histologic assessment of biopsies. Manual assessment of stained tissue of all biopsies limits speed and accuracy in clinical practice and research of prostate cancer diagnosis. We sought to develop a fully-automated multimodal microscopy method to distinguish cancerous from non-cancerous tissue samples.

Methods

We recorded chemical data from an unstained tissue microarray (TMA) using Fourier transform infrared (FT-IR) spectroscopic imaging. Using pattern recognition, we identified epithelial cells without user input. We fused the cell type information with the corresponding stained images commonly used in clinical practice. Extracted morphological features, optimized by two-stage feature selection method using a minimum-redundancy-maximal-relevance (mRMR) criterion and sequential floating forward selection (SFFS), were applied to classify tissue samples as cancer or non-cancer.

Results

We achieved high accuracy (area under ROC curve (AUC) >0.97) in cross-validations on each of two data sets that were stained under different conditions. When the classifier was trained on one data set and tested on the other data set, an AUC value of ~0.95 was observed. In the absence of IR data, the performance of the same classification system dropped for both data sets and between data sets.

Conclusions

We were able to achieve very effective fusion of the information from two different images that provide very different types of data with different characteristics. The method is entirely transparent to a user and does not involve any adjustment or decision-making based on spectral data. By combining the IR and optical data, we achieved high accurate classification.
Hinweise

Electronic supplementary material

The online version of this article (doi:10.​1186/​1471-2407-11-62) contains supplementary material, which is available to authorized users.

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

JTK contributed to developing algorithms, programming, and data analysis, and drafted the manuscript. SMH made the diagnoses of tissue samples. SS contributed to developing algorithms and data analysis. RB contributed to the data preparation and data analysis. All authors revised the manuscript and approved the final version.

Background

Prostate cancer

Prostate cancer (PCa) is the single most prevalent cancer in US men, accounting for one-third of non-skin cancer diagnoses every year [1]. Screening for the disease is widespread and for almost a million cases a year [24], a biopsy is conducted to detect or rule out cancer [3]. Manually-conducted histologic assessment of tissue upon biopsy forms the definitive diagnosis of PCa [5]. This need places a large demand on pathology services and manual examination limits speed and throughput. Histologic assessment is also critical to scientific progress as it is often the basis for research studies. Alternative methods for histologic recognition can greatly aid in alleviating workloads, assuring quality control and reducing costs [6]. There is no straightforward way, however, to aid pathology in this task and no clinical instrument is available for routine use. Hence, high-throughput, automated and objective tools for prostate pathology - both in clinical practice and in research - are needed.

Optical microscopy and automated PCa detection

Since the tissue does not have appreciable contrast in optical brightfield microscopy (Figure 1A), tissue samples are commonly stained using hematoxylin and eosin (H&E) prior to review by a pathologist. The stain is specific in limited terms - staining protein-rich regions pink and nucleic acid-rich regions of the tissue blue (Figure 1B). A pathologist is trained to recognize, from a stained tissue sample, the morphology and local architecture of glands as well as their structural alterations that indicate disease. The specific cell type that is used to recognize glandular structures is the epithelial sub-type. In prostatic carcinoma, which comprises more than 95% of prostate cancers [5], the cells of interest are epithelial cells [7]. Epithelial cells line 3D ducts in intact tissue and, hence, appear as cells lining empty circular regions (lumens) in images of histologic sections. Patterns of distortions of lumen appearance and spacing, as well as the arrangement of epithelial cells relative to lumens, have been characterized to indicate cancer and characterize its severity (Gleason grade) [8, 9]. The greater the distortion and loss of regular structure, the worse (higher grade) the cancer.
Recognizing structural distortions indicative of disease is a manual pattern recognition process that matches patterns in the tissue sample to standard patterns. Manual examination is powerful in that humans can recognize disease from a wide spectrum of normal and disease states, can overcome confounding artifacts, detect unusual cases and even recognize deficiencies in diagnoses. Manual examination, unfortunately, is time-consuming and leads routinely to variability in grading disease [8]. Computer-aided recognition of disease samples and grade patterns [10], hence, holds the potential for more accurate, reproducible and automated diagnoses [11, 12]. Unfortunately, tissue samples stain variably in populations due to biological diversity, with variations in stain composition, processing conditions and histotechnologists. The net result confounds automated image analysis and human-competitive recognition of cancer has not been automated for routine use. A robust means of automatically detecting epithelium and correlating its spatial patterns to determining cancer presence is highly desirable but yet unsolved.
Several efforts have been made to develop automated systems for the diagnosis and grading of microscopic prostate images. These include methods to identify distinct tissue compositions [13, 14] as well as several methods for automatic grading [1523]. The majority of these methods have extracted texture and/or morphological features to characterize tissue samples. Histologic objects such as nuclei, lumen, or gland have been mainly used to extract morphological features [15, 16, 2022, 24, 25]. Fourier Transform [17], Wavelet Transform [18, 19, 22], and Fractal Analysis [22, 23] have been the techniques commonly used to obtain texture features. In addition to these features, color [22] and graph-based [20] features have also been used. A number of classifiers have been tested on various features and data sets, although the choice of classifiers seems to have been less significant than the feature extraction step [22, 23].
Despite the above-mentioned lines of progress in automated diagnosis, an important concern is that the varying properties of images, due to acquisition settings [19, 25] and staining [26], may affect the classification results substantially. Although the issue of image variation by different acquisition settings has been addressed in [19, 25], to the best of our knowledge, no previous method has been validated across data sets under different staining conditions.
A major roadblock has been the limited information present in the data. For example, different cell types and morphologies are recognized by recognizing colors for empty space (usually close to white), apical portion of epithelial cells (usually pink) and the basal layer of epithelial cells (usually pink-dark blue). Immunohistochemical probes add useful information to diagnostic processes and are effective in understanding specific aspects of the disease, e.g. loss of basement membrane. For routine diagnostic pathology, however, the use of such molecular stains is expensive, time-consuming and does not actually address the need for an operator-free method. Additional molecular data is now available using label-free spectroscopic imaging, also known as chemical imaging [27].

Chemical imaging and automated histologic classification

Prostatic epithelial cells (and other cell types) [28] have recently been automatically recognized using a novel form of chemical imaging based on mid-infrared (IR) spectroscopy. Fourier transform infrared (FT-IR) spectroscopic imaging provides non-perturbing imaging by combining the spatial specificity of optical microscopy with the molecular selectivity of vibrational spectroscopy. Mid-IR spectral frequencies are resonant with the fundamental vibrational mode frequencies in molecules; hence, the IR absorption spectrum at each pixel is a quantitative record of composition [29]. FT-IR imaging has been successfully applied to various biological and biomedical problems such as determining molecular concentrations [30, 31] and structure [32, 33], characterizing cell components [34] and cancer diagnosis [3537]. In particular, the spectral patterns of different cell types being different, computerized pattern recognition can be used to assign each pixel into constituent cell types. The final result of recording data and mathematical analysis is images of tissue that are color coded for cell type. The process is illustrated in Figure 2. The approach has been used by a number of groups and is summarized in recent edited volumes [38, 39]. Since the numerical algorithms are automated, quantification of accuracy and statistical confidence in results is facile [40].
The above approach has been extensively validated in providing histologic recognition using tissue samples from over 1000 patients and tens of millions of pixels using tissue microarrays (TMAs). TMAs consist of multiple tissue samples of a size that assures representative sampling and allow high throughput experimentation in an efficient manner. For this manuscript, we examined two independent data sets from prostate tissue microarrays that were subjected to chemical imaging and histologic classification as outlined above. Images of the data are shown in Figure 2.
While we expected the chemical imaging approach to prove useful in histologic analysis of prostate tissue, its relationship to the existing clinical practice of using H&E stained tissue in PCa diagnosis was not clear a priori. Hence, we sought to examine whether a combination of the two techniques (i.e., optical microscopy following H&E staining, and FT-IR imaging) could provide high accuracy diagnoses that could otherwise not be achieved using H&E images alone.

Overview of this work

We develop a new fully-automated method to classify cancer versus non-cancer prostate tissue samples. The classification algorithm uses morphological features - geometric properties of epithelial cells/nuclei and lumens - that are quantified based on H&E stained images as well as FT-IR images of the tissue samples. By restricting the features used to geometric measures, we sought to mimic the pattern recognition process employed by human experts, and achieve a robust classification procedure that can produce consistently high accuracy across independent data sets. We systematically evaluate the performance of the new method through cross-validation, and examine its robustness across data sets. We also summarize the specific morphological features that prove to be most informative in classification.

Methods

We begin with a description of the computational pipeline. As noted above, a key aspect of our approach is the use of FT-IR imaging data on a serial section that is H&E-stained to enhance the segmentation of nuclei and lumens. The first two components of the pipeline are geared to this functionality, while the next three components exploit the segmented features obtained from image data to classify the tissue sample (Figure 3).

Image Registration

Given two images, the image registration problem can be defined as finding the optimal spatial and intensity transformation [41] of one image to the other. Here, two images are H&E stained (I reference ) and "IR classified" images (I target ) which were acquired from adjacent tissue samples. The IR classified image represents the FT-IR imaging data, processed as indicated in Figure 2, to classify each pixel as a particular cell type. Although the two tissue samples were physically in the same intact tissue and are structurally similar, the two images have different properties (total image and pixel sizes, contrast mechanisms and data values). Hence, features to spatially register the images are not trivial. The H&E image provides detailed morphological information that could ordinarily be used for registration, but the IR image lacks such information. On the other hand, the IR image specifies the exact areas corresponding to each cell type, but the difficulty in precisely extracting such regions from the H&E image hinders us from using cell-type information for registration. The only obvious features are macroscopic tissue sample shape and empty space (lumens) inside the tissue samples. To utilize these two features and to avoid problems due to differences in the two imaging techniques, both images are first converted into binary images. Due to the binarization, the intensity transformation is not necessary. As a spatial transformation, we use an affine transformation ( f ) [41] where a coordinate (x 1 , y 1 ) is transformed to the (x 2 , y 2 ) coordinate after translations (t x , t y ), rotation by θ, and scaling by factor s.
[ x 2 y 2 ] = [ t x t y ] + s [ cos θ sin θ sin θ cos θ ] [ x 1 y 1 ]
Accordingly, we find the optimal parameters of the affine transformation that minimizes the absolute intensity difference between two images (I reference and I target ). In other words, image registration amounts to finding the optimal parameter values ( t x * , t y * , θ * , s * ) = arg min t x , t y , θ , s | I r e f e r e n c e f ( I t a r g e t ; t x , t y , θ , s ) | . The downhill simplex method [42] is applied to solve the above equation. An example of this registration process is shown in Figure 4. (See [Additional file 1: Image Registration] for details.)

Identification of epithelial cells and their morphologic features

While a number of factors are known to be transformed in cancerous tissues, epithelial morphology is utilized as the clinical gold standard. Hence, we focus here on cellular and nuclear morphology of epithelial nuclei and lumens. These structures are different in normal and cancerous tissues, but are not widely used in automated analysis due to a few reasons. First, as described above, simple detection of epithelium from H&E images is difficult. Second, detection of epithelial nuclei may be confounded by a stromal response that is not uniform for all grades and types of cancers. We focused first on addressing these two challenges that hinder automatically parsing morphologic features such as the size and number of epithelial nuclei and lumens, distance from nuclei to lumens, geometry of the nuclei and lumens, and others (Feature Extraction). In order to use these properties, the first step is to detect nuclei and lumens correctly and we sought to develop a robust strategy for the same.

Lumen Detection

In H&E stained images, lumens are recognized to be empty white spaces surrounded by epithelial cells. In normal tissues, lumens are larger in diameter and can have a variety of shapes. In cancerous tissues, lumens are progressively smaller with increasing grade and generally have less distorted elliptical or circular shapes. Our strategy to detect lumens was to find empty areas that are located next to the areas rich in epithelium. White spots inside the tissue sample can be found from the H&E image by using a proper threshold value (>200) for the intensity of Red, Green, and Blue channels, and the pixels corresponding to epithelial cells can be mapped on the H&E image from the IR classified image through image registration. Although restricting the white areas adjacent to epithelial cells, in our observations, many artifactual lumens are still present. Additionally, the size and shape of lumens are examined to eliminate such artifacts. We note that while lumens are ideally completely surrounded by epithelial cells (called complete lumens), some tissue samples have lumens (called incomplete lumens) that violate this criterion because only a part of lumen is present in the tissue sample. To identify these incomplete lumens, we model an entire tissue sample as a circle, and the white spots between the tissue sample and the circle are the candidate incomplete lumens. As did in complete lumen detection, the same threshold value is used to identify white areas. To identify artifacts, we use heuristic criteria based on the size, shape, presence of epithelial cells and background around the areas. In addition, the distances from the center of the tissue to the white spots are examined to identify the artifacts in crescent form which resulted from the small gaps between the tissue sample and the circle fitted to the sample. (See [Additional file 1: Lumen Detection] for details.)

Nucleus Detection - single epithelial cells

Epithelial nucleus detection by automated analysis is more difficult than lumen detection due to variability in staining and experimental conditions under which the entire set of H&E images were acquired. Differences between normal and cancerous tissues, and among different grades of cancerous tissues, also hamper facile detection. To handle such variations and make the contrast of the images consistent, we perform smoothing [43] and adaptive histogram equalization [44] prior to nuclei identification. Nuclei are relatively dark and can be modeled as small elliptical areas in the stained images. This geometrical model is often confounded as multiple nuclei can be so close as to appear like one large, arbitrary-shaped nucleus. Also, small folds or edge staining around lumens can make the darker shaded regions difficult to analyze. Here, we exploit the information provided by the IR classified image to limit ourselves to epithelial cells, and use a thresholding heuristic on a color space-transformed image to identify nuclei with high accuracy. Superimposing the IR classified image on the H&E image, pixels corresponding to epithelium can be identified on the H&E image. These epithelial pixels are dominated by one of two colors: blue or pink, which arise from the nuclear and cytoplasmic component respectively. For nuclei restricted to epithelial cells in this manner, a set of general observations were made that led us to convert the stained image to a new image where each pixel has an intensity value |R + G - B|. (R, G, and B represent the intensity of Red, Green, and Blue channels, respectively.) This transformation, followed by suitable thresholding, was able to successfully characterize the areas where nuclei are present. The threshold values are adaptively determined for Red and Green channels due to the variations in the color intensity. Finally, filling holes and gaps within nuclei by a morphological closing operation [45], the segmentation of each nucleus is accomplished by using a watershed algorithm [45] followed by elimination of false detections. The size, shape, and average intensity are considered to identify and remove artifactual nuclei. Figure 5 details the nucleus detection procedure. (See [Additional file 1: Nucleus Detection] for details.)

Feature Extraction

As mentioned above, the characteristics of nuclei and lumens change in cancerous tissues. In a normal tissue, epithelial cells are located mostly in thin layers around lumens. In cancerous tissue, these cells generally grow to fill lumens, resulting in a decrease in the size of lumens, with the shape of lumens becoming more elliptical or circular. The epithelial association with a lumen becomes inconsistent and epithelial foci may adjoin lumens or may also exist without an apparent lumen. Epithelial cells invading the extra-cellular matrix also result in a deviation from the well-formed lumen structure; this is well-recognized as a hallmark of cancer. Due to filling lumen space and invasion into the extra-cellular space, the number density of epithelial cells increases in tissue. The size of individual epithelial cells and their nuclei also tend to increase as malignancy of a tumor increases. Motivated by such recognized morphological differences between normal and cancerous tissues, we chose to use epithelial nuclei and lumens as the basis of the several quantitative features that our classification system works with. (See examples of such features in Figure 6.) It is notable that these observations are qualitative in actual clinical practice and have not been previously quantified.
Epithelial cell information is available from IR data. However, individual epithelial cells in the tissue are not easily delineated. Therefore, in addition to features directly describing epithelial cells, we also quantify properties of epithelial nuclei, which are available from the segmentation described above. The quantities we measure in defining features are: (1) size of epithelial cells, (2) size of epithelial nuclei, (3) number of nuclei in the tissue sample, (4) distance from a nucleus to the closest lumen, (5) distance from a nucleus to the epithelial cell boundary, (6) number of "isolated" nuclei (nuclei that have no neighboring nucleus within a certain distance), (7) number of nuclei located "far" from lumens, and (8) entropy of spatial distribution of nuclei (Figure 6G). [Additional file 1: Epithelium-related Features] provide specifics of these measures and their calculation.
Features describing glands have been shown to be effective in PCa classification [21, 25]. Here, we try to characterize lumens and mostly focus on the differences in the shape of the lumens. The quantities we measure in defining these features are: (1) size of a lumen, (2) number of lumens, (3) lumen "roundness" [25], defined as L p e r i 2 L a r e a r where L peri is the perimeter of the lumen, L area is the size of the lumen (i.e., number of pixels in the lumen), and r is the radius of a circle of size L area , (4) lumen "distortion" (Figure 6A), computed as S T D ( d L c b ) A V G ( d L c b ) where d L c b is the distance from the center of a lumen to the boundary of the lumen and AVG(·) and STD(·) represent the average and standard deviation, (5) lumen "minimum bounding circle ratio" (Figure 6B), defined as the ratio of the size of a minimum bounding circle of a lumen to the size of the lumen, (6) lumen "convex hull ratio" (Figure 6C), which is the ratio of the size of a convex hull of a lumen to the size of the lumen, (7) symmetric index of lumen boundary (Figure 6E, see [Additional file 1: Lumen-related Features]), (8) symmetric index of lumen area (Figure 6F, see [Additional file 1: Lumen-related Features]), and (9) spatial association of lumens and cytoplasm-rich regions (Figure 6D, see [Additional file 1: Lumen-related Features]). Features (3) - (8) are various ways to summarize lumen shapes, while feature (9) is motivated by the loss of functional polarization of epithelial cells in cancerous tissues.

Global & local tissue features

We have described above the individual measures of epithelium and lumen related quantities that form the basis of the features used by our classification system. Normally, these features have to be summary measures over the entire tissue sample or desired classification area. Hence, we employ average (AVG) or standard deviation (STD), and in some cases the sum total (TOT) of these quantities for further analysis. These features are called "global" features since they are calculated from the entire tissue sample. However, in some cases global features may be misleading, especially where only a part of the tissue sample is indicative of cancer. Therefore, in addition to global features, we define "local" features by sliding a rectangular window of a fixed size (100 × 100 pixels) throughout a tissue sample. For each window, AVG and/or TOT of the epithelium and lumen related quantities are computed. STD or extremal values (MIN or MAX) of the AVG and/or TOT values over all windows become local feature values (Figure 7). In all, 67 features (29 global and 38 local features) are defined capturing various aspects of tissue morphology.

Feature Selection

Feature selection is the step where the classifier examines all available features (67 in our case) with respect to the training data, and selects a subset to use on test data. This selection is generally based on the criterion of high accuracy on training data, but also strives to ensure generalizability beyond the training data. We adopt a two-stage feature selection approach here. In the first stage, we generate a set of candidate features (C candidate ) by using the so-called minimum-redundancy-maximal-relevance (mRMR) criterion [46] (see [Additional file 1: mRMR]). In each iteration, given a feature set chosen thus far, mRMR chooses the single additional feature that is least redundant with the chosen features, while being highly correlated with the class label. C candidate is a set of features that is expected to be close to the optimal feature set for a data set and a classifier under consideration. It is constructed as follows. Given a feature set F = (f1, ..., fM) ordered by mRMR, the area under the ROC curve (AUC) of the set of i top-ranked features is computed for varying values of i. We limit the value of i to be ≤ 30. The feature subset with the best AUC is chosen as the C candidate . In the second stage, feature selection continues with C candidate as the starting point, using the sequential floating forward selection (SFFS) method [47]. This method sequentially adds new features followed by conditional deletion(s) of already selected features. Starting with the C candidate , SFFS searches for a feature x ∉ C candidate that maximizes the AUC among all feature sets C candidate ∪ {x}, and adds it to C candidate . Then, it finds a feature xC candidate that maximizes the AUC among all feature sets C candidate - {x}. If the removal of x improves the highest AUC obtained by C candidate , x is deleted from C candidate . As long as this removal improves upon the highest AUC obtained so far, the removal step is repeated. SFFS repeats the addition and removal steps until AUC reaches 1.0 or the number of additions and deletions exceeds 20, and the feature set with the highest AUC thus far is chosen as the optimal feature set. The classification capability of a feature set, required for feature selection, is measured by AUC, obtained by cross-validation on the training set. SFFS can be directly applied to the original feature set; however, using mRMR may help to reduce the search space and time and to build the optimal classifier by providing a good initial feature set for SFFS.

Classification

We note that there are two levels of classification here. In the first, IR spectral data is used to provide histologic images where each pixel has been classified as a cell type. In the second, the measures from H&E images and IR images are used to classify tissue into disease states. For the first classification task, we used a Bayesian classifier built on 18 spectral features. This previously achieved >0.99 AUC on cell type classification [48, 49]. For the latter task, we used a well established classification algorithm, namely support vector machine (SVM) [50]. As a kernel function, a radial basis function K ( x i , x j ) = exp ( γ x i x j 2 ) with parameter γ = 10, 1, 0.1, 0.01, 0.001 is used. Two cost factors are introduced to deal with an imbalance in training data [51]. The ratio between two cost functions was chosen as
C + C = number of negative training examples number of positive training examples
to make the potential total cost of the false positives and the false negatives the same. (See [Additional file 1: SVM] for details.)

Samples and Data preparation

All of the H&E stained images were acquired on a standard optical microscope at 40x magnification. The size of each pixel is 0.9636 um × 0.9636 um. On the other hand, the pixel size of IR images is 6.25 um × 6.25 um.
Two different tissue microarrays were obtained from two different sources (Tissue microarray research program at the National Institutes of Health and Clinomics Inc.). The first data set ("Data1") consisted of 240 tissue samples from 180 patients, and the second set ("Data2") includes 160 tissue samples from 80 patients. Both sets of tissue samples were sectioned to ~ 7 micron thick sections, with a section being placed on IR transparent BaF2 slides and a serial section on a standard glass slide. The acquisition of data is described elsewhere in [28]. Unfortunately, we were not able to use all of these tissue samples for several reasons. Each data set has two TMAs. One is H&E stained image and the other is IR image. Since these were experimental arrays, some TMA spots were missing in one or both arrays due to processing and plating on the salt plates used for IR analysis. Since our method focuses on epithelial cells, tissue samples which do not have enough epithelial cells (>100) in either of two images (H&E and IR) were not considered in this study. Moreover, some tissue samples in Data2 are spatially displaced and fused with neighboring tissue samples. Eliminating those tissue samples, 66 benign tissue samples and 115 cancer tissue samples are remained for Data1, and 14 benign and 36 cancer tissue samples remained for Data2. An example of H&E images for both data sets is shown in Figure 8.

Results and discussion

The classification system achieves AUC greater than 0.97 on both data sets

We first performed K-fold cross validation on each data set. The data set was divided into K roughly equal-sized partitions, one partition was left out as the "test data", the classifier was trained on the union of the remaining K - 1 partitions (the "training data") and evaluated on the test data. This was repeated K times, with different choices of the left-out partition. (We set K = 10.) In each repetition, cross-validation on the training data was used to select the feature set with the highest AUC as explained in Feature Selection. The correct and incorrect predictions in the test data, across all K repetitions, were summarized into a ROC plot and the AUC was computed, along with specificities when sensitivity equals 90, 95, or 99%. Since the cross-validation exercise makes random choices in partitioning the data set, we examined averages of these performance metrics over 10 repeats of the entire cross-validation pipeline. The average AUC for Data1 and Data2 were 0.982 and 0.974 respectively (Table 1, "feature extraction" = "IR & HE"). At 90%, 95%, and 99% sensitivities, the average specificity achieved on Data1 was 94.76%, 90.91%, and 77.80% respectively, while that on Data2 was 92.53%, 84.19%, and 49.54% respectively. SVM using the kernel parameter γ = 1 is used here. This result is consistent with the classification results using different values of the parameter γ (See [Additional file 1: Supplementary Table S2] for details). We note that other classification methods can be also used. Among various methods, a logistic model tree [52], which combines linear logistic regression with decision tree induction, was used, and achieved slightly lesser performance than SVM (results not shown here).
Table 1
Classification results via cross-validation
Data set
Feature Extraction
AUC
Sensitivity (%)
Specificity (%)
Mf
  
AVG
STD
 
AVG
STD
 
Data1
IR & HE
0.982
0.0030
90
94.76
1.64
13
    
95
90.91
1.62
 
    
99
77.80
5.52
 
 
HE only
0.968
0.0052
90
91.64
2.26
11
    
95
83.90
1.91
 
    
99
53.43
13.65
 
Data2
IR & HE
0.974
0.0145
90
92.53
7.11
7
    
95
84.19
10.84
 
    
99
49.54
22.51
 
 
HE only
0.880
0.0175
90
61.34
10.31
8
    
95
22.21
10.06
 
    
99
11.21
6.01
 
AVG and STD denote average and standard deviation across ten repeats of cross-valdiation. Mf is the median size of the feature set obtained by feature selection from training data. Column "Feature Extraction" indicates if features were obtained using H&E as well as IR data, or with H&E data alone. The parameter γ of a radial basis kernel for SVM is set to 1.
One way to interpret the above values is to examine our automated pipeline as a pre-screening mechanism to identify the samples to be examined by a human pathologist. At a "true positive rate" of 99% (which means that only 1% of the cancer samples will be missed by the screen), the "false positive rate" is 22.2% (i.e., 22.2% of the benign samples will make it through the screen) on average for Data1 (Table 1), thereby reducing the workload of the pathologist by 4.5-fold. While the error rate of manual pathology determinations is generally accepted to be in 1-5% range, inclusion of confounding cancer mimickers raises the rate to as high as 7.5% [53]. Also noteworthy is the observation that the same algorithm performs consistently well on both data sets, that were obtained from different staining conditions. This speaks to the robustness of the classification framework, an attribute that we investigated further in the next exercise.

Classification system is robust to staining conditions

Here, we trained a classifier on Data1 and tested its performance on Data2 (Table 2, "Data set" = "Test") using SVM with γ = 1. We observed an average AUC of 0.956, with average specificity of 88.57%, 81.92%, and 26.86% at sensitivity equaling 90%, 95%, and 99% respectively (Table 2, "feature extraction" = "IR & HE" and "Data set" = "Test"). These values are competitive with the cross-validation results on Data2 (Table 1), where the training and testing were both performed on (disjoint parts of) Data2. It should be noted that in Table 2 "Data set" = "Train" means that the classifier was not only trained but also tested on Data1, and thus the difference between the "Train" and "Test" rows does not refer to a difference in performance on the two data sets. As a classifier is trained on Data2 and tested on Data1, we obtained the average AUC of 0.855 and average specificity of 50.18%, 40.41%, and 12.33% at sensitivity equalling 90%, 95%, and 99% respectively ([Additional file 1: Supplementary Table S4, γ = 1]). The results are worse than both the cross-validation results on Data1 and the validation results on Data2. This may be due to the fact that the number of samples in Data2 is relatively small and much unbalanced. In addition, varying the parameter value γ of SVM, the results, by and large, are the same (See [Additional file 1: Supplementary Table S3 and S4] for details).
Table 2
Validation between data sets
Feature Extraction
Data set
AUC
Sensitivity (%)
Specificity (%)
Mf
  
AVG
STD
 
AVG
STD
 
IR & HE
Train
0.994
0.0006
90
98.30
0.68
13
    
95
96.58
1.10
 
    
99
91.55
2.55
 
 
Test
0.956
0.0089
90
88.57
5.96
 
    
95
81.92
5.28
 
    
99
26.86
15.50
 
HE only
Train
0.986
0.0021
90
97.77
0.97
10
    
95
91.56
2.49
 
    
99
79.29
4.47
 
 
Test
0.918
0.0100
90
65.51
8.37
 
    
95
46.14
7.53
 
    
99
13.29
6.94
 
A classifier is trained on Data1 and tested on Data2. AVG and STD denote the average and standard deviation. Mf is the median size of the optimal feature set. Column "Feature Extraction" indicates if features were obtained using H&E as well as IR data, or with H&E data alone. Column "Data set" indicates if the performance metrics are from training data (Data1) or from test data (Data2). The parameter γ of a radial basis kernel for SVM is set to 1.

Use of IR data improves classification performance

To assess the utility of the IR-based cell-type classification, we repeated the above exercises after extracting features without the guidance of the IR data; i.e., epithelial cells were predicted from the H&E images alone (see [Additional file 1: Epithelium Detection] for details). All of the features defined in Feature Extraction were used, except for "Spatial association of lumens and apical regions", since the distinction between cytoplasm-rich and nuclear-rich region in epithelial cells was unclear in H&E images. The results from this disadvantaged classifier are shown in Tables 1 and 2 ("feature extraction" = "HE only"). For both types of experiments, we obtained lower average AUCs and specificity values. For instance, the AUC of cross-validation in Data2 (Table 1) dropped from 0.974 to 0.880. Similarly, the results of validation between data sets (Table 2) were substantially worse now compared to the IR-guided classification, with the AUC dropping from 0.956 to 0.918. We also observed that the average AUC dropped in the absence of IR data as using different values of parameter γ for SVM (See [Additional file 1] for details). This indicates that the use of IR data, i.e., the improved epithelial identification, helps to attain better classification performance. We also note that other methods, if any, which could achieve high accuracy identification of epithelial cells may have the same impact with the IR data on the classification.
Previously, Tabesh et al. achieved an accuracy of 96.7% via cross validation in cancer/no-cancer classification [22]. Color, morphometric, and texture features were extracted, and all images were acquired under similar conditions. We note that our classification result (Table 1), based solely on morphology, is comparable to their result; however the software developed by Tabesh et al. was not available for evaluation in our data sets. Color and texture features could provide additional information; however, their robustness to different data sets is questionable, and their interpretation is not as obvious as that of morphological features, which are used in clinical practice. Different data sets may have varied properties which may be attributable to staining variations, inconsistent image acquisition settings, and image preparation. The performance of the same method based on texture features has been seen to greatly change from one data set to another [19, 22, 25]. Variations in staining may affect color features. In contrast, morphological features were shown to be robust to varying image acquisition settings [25]. Nonetheless, the quality of morphological features is subject to segmentation of histologic objects. Thus, any method based on morphological features will benefit from the IR cell-type classification.

Examination of discriminative features

We examined the importance of each feature by its rank in the first phase of feature selection, based on its "relevance" to the class label (see [Additional file 1: mRMR]). Since different features (e.g., average or standard deviation, global or local features) based on the same underlying quantity (e.g., "lumen roundness") generally have similar relevance, we examined the average relevance of features in each of 17 feature categories (Figure 9), for each data set. The relevance of features is consistent across cross-validation (see [Additional file 1: Supplementary Figure S1]). The complete list of the individual features and their relevance and mRMR rank (for Data1) is available in Figure 10. For Data1, lumen-related feature categories are most relevant in general, while epithelium-related feature categories are most important for Data2. It is surprising that the top 3 feature categories in Data1 (Figure 9, blue bars) - size of lumen, lumen roundness, and lumen convex hull ratio - have very low relevance in Data2, although we note that this may be in large part due to variations in staining and malignancy of tumors between the two data sets and differences in the size of two data sets. The comparable classification results on Data2 (Table 1, 2), in spite of the maximal relevance differences, may indicate the broadness of our feature set and the accuracy of our feature selection method and facilitate the application of the same classifier on different data sets. Nevertheless, a larger scale study may be necessary to precisely examine the differences between data sets and features. It is, however, noteworthy that examining the features (or feature categories) with highest relevance alone may be slightly misleading, because this examination does not account for redundancy among features.
To further examine the most informative and non-redundant features, we inspected the optimal feature sets selected after both stages of the feature selection component. For both Data1 and Data2, the selection of the features is consistent across all folds of cross validation. (See [Additional file 1: Supplementary Figure S2] for details.) In Figure 11, we show an example of three most frequently selected features for Data1: number of lumens (LSTD), lumen roundness (GAVG), and size of nucleus (GTOT). We note that these include both lumen and epithelium related features. Lumen roundness (GAVG) is the only one ranked high by maximal relevance (Figure 10), yet all three features are consistently chosen by the classifier, since they provide different, complementary information on a tissue: greater circularity of lumens and increase in the number of lumens and the size of nuclei indicate malignancy of a tissue.

Conclusions

In this manuscript, we have presented a means to eliminate epithelium recognition deficiencies in classifying H&E images for presence or absence of cancer. The method is entirely transparent to a user and does not involve any adjustment or decision-making based on spectral data. We were able to achieve very effective fusion of the information from two different modalities, namely optical and IR microscopy, that provide very different types of data with different characteristics. Several features of the tissue were quantified and employed for classification. We found that robust classification could be achieved using a few measures, which are detailed to arise from epithelial/lumen organization and provide a reasonable explanation for the accuracy of the model. The choice of combining the IR and optical data is shown to be necessary for achieving the high accuracy values observed. We anticipate that the combined use of the two microscopies - structural and chemical - will lead to an accurate, robust and automated method for determining cancer within biopsy specimens.

Acknowledgements

The project described was supported by Award Number R01CA138882 from the National Cancer Institute. The project is also supported by National Institutes of Health intramural funding (to S.M.H). The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Cancer Institute or the National Institutes of Health. The project was also supported by a DoD prostate cancer research program young investigator award (to R.B.) and a computational science and engineering (CSE) fellowship (to J.T.K).
Open Access This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License ( https://​creativecommons.​org/​licenses/​by/​2.​0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

JTK contributed to developing algorithms, programming, and data analysis, and drafted the manuscript. SMH made the diagnoses of tissue samples. SS contributed to developing algorithms and data analysis. RB contributed to the data preparation and data analysis. All authors revised the manuscript and approved the final version.
Literatur
1.
Zurück zum Zitat Jemal A, Siegel R, Ward E, Murray T, Xu JQ, Smigal C, Thun MJ: Cancer statistics, 2006. Ca-a Cancer Journal for Clinicians. 2006, 56 (2): 106-130. 10.3322/canjclin.56.2.106.CrossRefPubMed Jemal A, Siegel R, Ward E, Murray T, Xu JQ, Smigal C, Thun MJ: Cancer statistics, 2006. Ca-a Cancer Journal for Clinicians. 2006, 56 (2): 106-130. 10.3322/canjclin.56.2.106.CrossRefPubMed
2.
Zurück zum Zitat Gilbert SM, Cavallo CB, Kahane H, Lowe FC: Evidence suggesting PSA cutpoint of 2.5 ng/mL for prompting prostate biopsy: Review of 36,316 biopsies. Urology. 2005, 65 (3): 549-553. 10.1016/j.urology.2004.10.064.CrossRefPubMed Gilbert SM, Cavallo CB, Kahane H, Lowe FC: Evidence suggesting PSA cutpoint of 2.5 ng/mL for prompting prostate biopsy: Review of 36,316 biopsies. Urology. 2005, 65 (3): 549-553. 10.1016/j.urology.2004.10.064.CrossRefPubMed
3.
Zurück zum Zitat Pinsky PF, Andriole GL, Kramer BS, Hayes RB, Prorok PC, Gohagan JK, P PLCO: Prostate biopsy following a positive screen in the prostate, lung, colorectal and ovarian cancer screening trial. Journal of Urology. 2005, 173 (3): 746-750. 10.1097/01.ju.0000152697.25708.71.CrossRefPubMed Pinsky PF, Andriole GL, Kramer BS, Hayes RB, Prorok PC, Gohagan JK, P PLCO: Prostate biopsy following a positive screen in the prostate, lung, colorectal and ovarian cancer screening trial. Journal of Urology. 2005, 173 (3): 746-750. 10.1097/01.ju.0000152697.25708.71.CrossRefPubMed
4.
Zurück zum Zitat Jacobsen SJ, Katusic SK, Bergstralh EJ, Oesterling JE, Ohrt D, Klee GG, Chute CG, Lieber MM: Incidence of Prostate-Cancer Diagnosis in the Eras before and after Serum Prostate-Specific Antigen Testing. Jama-Journal of the American Medical Association. 1995, 274 (18): 1445-1449. 10.1001/jama.274.18.1445.CrossRef Jacobsen SJ, Katusic SK, Bergstralh EJ, Oesterling JE, Ohrt D, Klee GG, Chute CG, Lieber MM: Incidence of Prostate-Cancer Diagnosis in the Eras before and after Serum Prostate-Specific Antigen Testing. Jama-Journal of the American Medical Association. 1995, 274 (18): 1445-1449. 10.1001/jama.274.18.1445.CrossRef
5.
Zurück zum Zitat Humphrey PA, American Society for Clinical Pathology: Prostate pathology. 2003, Chicago: American Society for Clinical Pathology Humphrey PA, American Society for Clinical Pathology: Prostate pathology. 2003, Chicago: American Society for Clinical Pathology
6.
Zurück zum Zitat Bartels PH, Thompson D, Bartels HG, Montironi R, Scarpelli M, Hamilton PW: Machine vision-based histometry of premalignant and malignant prostatic lesions. Pathol Res Pract. 1995, 191 (9): 935-944.CrossRefPubMed Bartels PH, Thompson D, Bartels HG, Montironi R, Scarpelli M, Hamilton PW: Machine vision-based histometry of premalignant and malignant prostatic lesions. Pathol Res Pract. 1995, 191 (9): 935-944.CrossRefPubMed
7.
Zurück zum Zitat Epstein JI, Netto GJ: Biopsy interpretation of the prostate. 2008, Philadelphia: Wolters Kluwer Health/Lippincott Williams & Wilkins, 4 Epstein JI, Netto GJ: Biopsy interpretation of the prostate. 2008, Philadelphia: Wolters Kluwer Health/Lippincott Williams & Wilkins, 4
8.
Zurück zum Zitat Gleanson DF: Histologic grading and clinical staging of prostate carcinoma. The Prostate. Edited by: Tannenbaum M. 1977, Philadelphia: Lea and Febiger Gleanson DF: Histologic grading and clinical staging of prostate carcinoma. The Prostate. Edited by: Tannenbaum M. 1977, Philadelphia: Lea and Febiger
9.
Zurück zum Zitat Epstein JI, Allsbrook WC, Amin MB, Egevad LL: Update on the Gleason grading system for prostate cancer - Results of an international consensus conference of urologic pathologists. Advances in Anatomic Pathology. 2006, 13 (1): 57-59. 10.1097/01.pap.0000202017.78917.18.CrossRefPubMed Epstein JI, Allsbrook WC, Amin MB, Egevad LL: Update on the Gleason grading system for prostate cancer - Results of an international consensus conference of urologic pathologists. Advances in Anatomic Pathology. 2006, 13 (1): 57-59. 10.1097/01.pap.0000202017.78917.18.CrossRefPubMed
10.
Zurück zum Zitat Gurcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, Yener B: Histopathological Image Analysis: A Review. Biomedical Engineering, IEEE Reviews in. 2009, 2: 147-171. 10.1109/RBME.2009.2034865.CrossRef Gurcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, Yener B: Histopathological Image Analysis: A Review. Biomedical Engineering, IEEE Reviews in. 2009, 2: 147-171. 10.1109/RBME.2009.2034865.CrossRef
11.
Zurück zum Zitat Mulrane L, Rexhepaj E, Penney S, Callanan JJ, Gallagher WM: Automated image analysis in histopathology: a valuable tool in medical diagnostics. Expert Rev Mol Diagn. 2008, 8 (6): 707-725. 10.1586/14737159.8.6.707.CrossRefPubMed Mulrane L, Rexhepaj E, Penney S, Callanan JJ, Gallagher WM: Automated image analysis in histopathology: a valuable tool in medical diagnostics. Expert Rev Mol Diagn. 2008, 8 (6): 707-725. 10.1586/14737159.8.6.707.CrossRefPubMed
12.
Zurück zum Zitat Madabhushi A: Digital pathology image analysis: opportunities and challenges. Imaging in Medicine. 2009, 1 (1): 7-10. 10.2217/iim.09.9.CrossRef Madabhushi A: Digital pathology image analysis: opportunities and challenges. Imaging in Medicine. 2009, 1 (1): 7-10. 10.2217/iim.09.9.CrossRef
13.
Zurück zum Zitat Roula M, Diamond J, Bouridane A, Miller P, Amira A: A multispectral computer vision system for automatic grading of prostatic neoplasia. Biomedical Imaging, 2002 Proceedings 2002 IEEE International Symposium on: 2002. 2002, 193-196. Roula M, Diamond J, Bouridane A, Miller P, Amira A: A multispectral computer vision system for automatic grading of prostatic neoplasia. Biomedical Imaging, 2002 Proceedings 2002 IEEE International Symposium on: 2002. 2002, 193-196.
14.
Zurück zum Zitat Diamond J, Anderson NH, Bartels PH, Montironi R, Hamilton PW: The use of morphological characteristics and texture analysis in the identification of tissue composition in prostatic neoplasia. Human Pathology. 2004, 35 (9): 1121-1131. 10.1016/j.humpath.2004.05.010.CrossRefPubMed Diamond J, Anderson NH, Bartels PH, Montironi R, Hamilton PW: The use of morphological characteristics and texture analysis in the identification of tissue composition in prostatic neoplasia. Human Pathology. 2004, 35 (9): 1121-1131. 10.1016/j.humpath.2004.05.010.CrossRefPubMed
15.
Zurück zum Zitat Stotzka R, Manner R, Bartels PH, Thompson D: A Hybrid Neural and Statistical Classifier System for Histopathologic Grading of Prostatic Lesions. Analytical and Quantitative Cytology and Histology. 1995, 17 (3): 204-218.PubMed Stotzka R, Manner R, Bartels PH, Thompson D: A Hybrid Neural and Statistical Classifier System for Histopathologic Grading of Prostatic Lesions. Analytical and Quantitative Cytology and Histology. 1995, 17 (3): 204-218.PubMed
16.
Zurück zum Zitat Wetzel AW, Crowley R, Kim S, Dawson R, Zheng L, Joo YM, Yagi Y, Gilbertson J, Gadd C, Deerfield DW, et al: Evaluation of prostate tumor grades by content-based image retrieval. 1999; Washington, DC, USA. 1999, SPIE, 244-252. Wetzel AW, Crowley R, Kim S, Dawson R, Zheng L, Joo YM, Yagi Y, Gilbertson J, Gadd C, Deerfield DW, et al: Evaluation of prostate tumor grades by content-based image retrieval. 1999; Washington, DC, USA. 1999, SPIE, 244-252.
17.
Zurück zum Zitat Smith Y, Zajicek G, Werman M, Pizov G, Sherman Y: Similarity measurement method for the classification of architecturally differentiated images. Computers and Biomedical Research. 1999, 32 (1): 1-12. 10.1006/cbmr.1998.1500.CrossRefPubMed Smith Y, Zajicek G, Werman M, Pizov G, Sherman Y: Similarity measurement method for the classification of architecturally differentiated images. Computers and Biomedical Research. 1999, 32 (1): 1-12. 10.1006/cbmr.1998.1500.CrossRefPubMed
18.
Zurück zum Zitat Jafari-Khouzani K, Soltanian-Zadeh H: Multiwavelet grading of pathological images of prostate. Ieee Transactions on Biomedical Engineering. 2003, 50 (6): 697-704. 10.1109/TBME.2003.812194.CrossRefPubMed Jafari-Khouzani K, Soltanian-Zadeh H: Multiwavelet grading of pathological images of prostate. Ieee Transactions on Biomedical Engineering. 2003, 50 (6): 697-704. 10.1109/TBME.2003.812194.CrossRefPubMed
19.
Zurück zum Zitat Farjam R, Slotanian-Zadeh H, Zoroofi RA, Khouzani KJ: Tree-structured grading of pathological images of prostate. Proc SPIE Int Symp Med Imag: 2005; San Diego, CA. 2005, 840-851. Farjam R, Slotanian-Zadeh H, Zoroofi RA, Khouzani KJ: Tree-structured grading of pathological images of prostate. Proc SPIE Int Symp Med Imag: 2005; San Diego, CA. 2005, 840-851.
20.
Zurück zum Zitat Doyle S, Hwang M, Shah K, Madabhushi A, Feldman M, Tomaszeweski J: AUTOMATED GRADING OF PROSTATE CANCER USING ARCHITECTURAL AND TEXTURAL IMAGE FEATURES. Biomedical Imaging: From Nano to Macro, 2007 ISBI 2007 4th IEEE International Symposium on: 2007. 2007, 1284-1287.CrossRef Doyle S, Hwang M, Shah K, Madabhushi A, Feldman M, Tomaszeweski J: AUTOMATED GRADING OF PROSTATE CANCER USING ARCHITECTURAL AND TEXTURAL IMAGE FEATURES. Biomedical Imaging: From Nano to Macro, 2007 ISBI 2007 4th IEEE International Symposium on: 2007. 2007, 1284-1287.CrossRef
21.
Zurück zum Zitat Naik S, Doyle S, Feldman M, Tomaszewski J, Madabhushi A: Gland Segmentation and Computerized {G}leason Grading of Prostate Histology by Integrating Low-, High-level and Domain Specific Information. Proceedings of 2nd Workshop on Microsopic Image Analysis with Applications in Biology, Piscataway, NJ, USA: 2007. 2007 Naik S, Doyle S, Feldman M, Tomaszewski J, Madabhushi A: Gland Segmentation and Computerized {G}leason Grading of Prostate Histology by Integrating Low-, High-level and Domain Specific Information. Proceedings of 2nd Workshop on Microsopic Image Analysis with Applications in Biology, Piscataway, NJ, USA: 2007. 2007
22.
Zurück zum Zitat Tabesh A, Teverovskiy M, Pang HY, Kumar VP, Verbel D, Kotsianti A, Saidi O: Multifeature prostate cancer diagnosis and Gleason grading of histological images. Ieee Transactions on Medical Imaging. 2007, 26 (10): 1366-1378. 10.1109/TMI.2007.898536.CrossRefPubMed Tabesh A, Teverovskiy M, Pang HY, Kumar VP, Verbel D, Kotsianti A, Saidi O: Multifeature prostate cancer diagnosis and Gleason grading of histological images. Ieee Transactions on Medical Imaging. 2007, 26 (10): 1366-1378. 10.1109/TMI.2007.898536.CrossRefPubMed
23.
Zurück zum Zitat Huang PW, Lee CH: Automatic Classification for Pathological Prostate Images Based on Fractal Analysis. Ieee Transactions on Medical Imaging. 2009, 28 (7): 1037-1050. 10.1109/TMI.2009.2012704.CrossRefPubMed Huang PW, Lee CH: Automatic Classification for Pathological Prostate Images Based on Fractal Analysis. Ieee Transactions on Medical Imaging. 2009, 28 (7): 1037-1050. 10.1109/TMI.2009.2012704.CrossRefPubMed
24.
Zurück zum Zitat Arif M, Rajpoot N: Classification of potential nuclei in prostate histology images using shape manifold learning. Machine Vision, 2007 ICMV 2007 International Conference on: 28-29 Dec 2007 2007. 2007, 113-118.CrossRef Arif M, Rajpoot N: Classification of potential nuclei in prostate histology images using shape manifold learning. Machine Vision, 2007 ICMV 2007 International Conference on: 28-29 Dec 2007 2007. 2007, 113-118.CrossRef
25.
Zurück zum Zitat Farjam R, Soltanian-Zadeh H, Jafari-Khouzani K, Zoroofi RA: An image analysis approach for automatic malignancy determination of prostate pathological images. Cytometry Part B: Clinical Cytometry. 2007, 72B (4): 227-240. 10.1002/cyto.b.20162.CrossRef Farjam R, Soltanian-Zadeh H, Jafari-Khouzani K, Zoroofi RA: An image analysis approach for automatic malignancy determination of prostate pathological images. Cytometry Part B: Clinical Cytometry. 2007, 72B (4): 227-240. 10.1002/cyto.b.20162.CrossRef
26.
Zurück zum Zitat Schulte EKW: Standardization of Biological Dyes and Stains - Pitfalls and Possibilities. Histochemistry. 1991, 95 (4): 319-328. 10.1007/BF00266958.CrossRefPubMed Schulte EKW: Standardization of Biological Dyes and Stains - Pitfalls and Possibilities. Histochemistry. 1991, 95 (4): 319-328. 10.1007/BF00266958.CrossRefPubMed
27.
Zurück zum Zitat Levin IW, Bhargava R: Fourier transform infrared vibrational spectroscopic imaging: integrating microscopy and molecular recognition. Annu Rev Phys Chem. 2005, 56: 429-474. 10.1146/annurev.physchem.56.092503.141205.CrossRefPubMed Levin IW, Bhargava R: Fourier transform infrared vibrational spectroscopic imaging: integrating microscopy and molecular recognition. Annu Rev Phys Chem. 2005, 56: 429-474. 10.1146/annurev.physchem.56.092503.141205.CrossRefPubMed
28.
Zurück zum Zitat Fernandez DC, Bhargava R, Hewitt SM, Levin IW: Infrared spectroscopic imaging for histopathologic recognition. Nature Biotechnology. 2005, 23 (4): 469-474. 10.1038/nbt1080.CrossRefPubMed Fernandez DC, Bhargava R, Hewitt SM, Levin IW: Infrared spectroscopic imaging for histopathologic recognition. Nature Biotechnology. 2005, 23 (4): 469-474. 10.1038/nbt1080.CrossRefPubMed
29.
Zurück zum Zitat Ellis DI, Goodacre R: Metabolic fingerprinting in disease diagnosis: biomedical applications of infrared and Raman spectroscopy. Analyst. 2006, 131 (8): 875-885. 10.1039/b602376m.CrossRefPubMed Ellis DI, Goodacre R: Metabolic fingerprinting in disease diagnosis: biomedical applications of infrared and Raman spectroscopy. Analyst. 2006, 131 (8): 875-885. 10.1039/b602376m.CrossRefPubMed
30.
Zurück zum Zitat Budinova G, Salva J, Volka K: Application of molecular spectroscopy in the mid-infrared region to the determination of glucose and cholesterol in whole blood and in blood serum. Appl Spectrosc. 1997, 51 (5): 631-635. 10.1366/0003702971941034.CrossRef Budinova G, Salva J, Volka K: Application of molecular spectroscopy in the mid-infrared region to the determination of glucose and cholesterol in whole blood and in blood serum. Appl Spectrosc. 1997, 51 (5): 631-635. 10.1366/0003702971941034.CrossRef
31.
Zurück zum Zitat Shaw RA, Kotowich S, Mantsch HH, Leroux M: Quantitation of protein, creatinine, and urea in urine by near-infrared spectroscopy. Clin Biochem. 1996, 29 (1): 11-19. 10.1016/0009-9120(95)02011-X.CrossRefPubMed Shaw RA, Kotowich S, Mantsch HH, Leroux M: Quantitation of protein, creatinine, and urea in urine by near-infrared spectroscopy. Clin Biochem. 1996, 29 (1): 11-19. 10.1016/0009-9120(95)02011-X.CrossRefPubMed
32.
Zurück zum Zitat Fabian H, Naumann D: Methods to study protein folding by stopped-flow FT-IR. Methods. 2004, 34 (1): 28-40. 10.1016/j.ymeth.2004.03.004.CrossRefPubMed Fabian H, Naumann D: Methods to study protein folding by stopped-flow FT-IR. Methods. 2004, 34 (1): 28-40. 10.1016/j.ymeth.2004.03.004.CrossRefPubMed
33.
Zurück zum Zitat Petibois C, Deleris G: Evidence that erythrocytes are highly susceptible to exercise oxidative stress: FT-IR spectrometric studies at the molecular level. Cell Biol Int. 2005, 29 (8): 709-716. 10.1016/j.cellbi.2005.04.007.CrossRefPubMed Petibois C, Deleris G: Evidence that erythrocytes are highly susceptible to exercise oxidative stress: FT-IR spectrometric studies at the molecular level. Cell Biol Int. 2005, 29 (8): 709-716. 10.1016/j.cellbi.2005.04.007.CrossRefPubMed
34.
Zurück zum Zitat Helm D, Naumann D: Identification of Some Bacterial-Cell Components by Ft-Ir Spectroscopy. Fems Microbiol Lett. 1995, 126 (1): 75-79.CrossRef Helm D, Naumann D: Identification of Some Bacterial-Cell Components by Ft-Ir Spectroscopy. Fems Microbiol Lett. 1995, 126 (1): 75-79.CrossRef
35.
Zurück zum Zitat Malins DC, Polissar NL, Nishikida K, Holmes EH, Gardner HS, Gunselman SJ: The etiology and prediction of breast cancer. Fourier transform-infrared spectroscopy reveals progressive alterations in breast DNA leading to a cancer-like phenotype in a high proportion of normal women. Cancer. 1995, 75 (2): 503-517. 10.1002/1097-0142(19950115)75:2<503::AID-CNCR2820750213>3.0.CO;2-0.CrossRefPubMed Malins DC, Polissar NL, Nishikida K, Holmes EH, Gardner HS, Gunselman SJ: The etiology and prediction of breast cancer. Fourier transform-infrared spectroscopy reveals progressive alterations in breast DNA leading to a cancer-like phenotype in a high proportion of normal women. Cancer. 1995, 75 (2): 503-517. 10.1002/1097-0142(19950115)75:2<503::AID-CNCR2820750213>3.0.CO;2-0.CrossRefPubMed
36.
Zurück zum Zitat Ly E, Piot O, Wolthuis R, Durlach A, Bernard P, Manfait M: Combination of FTIR spectral imaging and chemometrics for tumour detection from paraffin-embedded biopsies. Analyst. 2008, 133 (2): 197-205. 10.1039/b715924b.CrossRefPubMed Ly E, Piot O, Wolthuis R, Durlach A, Bernard P, Manfait M: Combination of FTIR spectral imaging and chemometrics for tumour detection from paraffin-embedded biopsies. Analyst. 2008, 133 (2): 197-205. 10.1039/b715924b.CrossRefPubMed
37.
Zurück zum Zitat Beleites C, Steiner G, Sowa MG, Baumgartner R, Sobottka S, Schackert G, Salzer R: Classification of human gliomas by infrared imaging spectroscopy and chemometric image processing. Vib Spectrosc. 2005, 38 (1-2): 143-149. 10.1016/j.vibspec.2005.02.020.CrossRef Beleites C, Steiner G, Sowa MG, Baumgartner R, Sobottka S, Schackert G, Salzer R: Classification of human gliomas by infrared imaging spectroscopy and chemometric image processing. Vib Spectrosc. 2005, 38 (1-2): 143-149. 10.1016/j.vibspec.2005.02.020.CrossRef
38.
Zurück zum Zitat Spectrochemical Analysis Using Infrared Multichannel Detectors. Edited by: Rohit Bhargava IWL. 2005, Oxford: Blackwell Publishing, 56-84. Spectrochemical Analysis Using Infrared Multichannel Detectors. Edited by: Rohit Bhargava IWL. 2005, Oxford: Blackwell Publishing, 56-84.
39.
Zurück zum Zitat Diem M, Chalmers JM, Griffiths PR: Vibrational spectroscopy for medical diagnosis. 2008, Chichester, England; Hoboken, NJ: John Wiley & Sons Diem M, Chalmers JM, Griffiths PR: Vibrational spectroscopy for medical diagnosis. 2008, Chichester, England; Hoboken, NJ: John Wiley & Sons
40.
Zurück zum Zitat Bhargava R, Hewitt SM, Levin IW: Unrealistic expectations for IR microspectroscopic imaging - Reply. Nature Biotechnology. 2007, 25 (1): 31-33. 10.1038/nbt0107-31.CrossRef Bhargava R, Hewitt SM, Levin IW: Unrealistic expectations for IR microspectroscopic imaging - Reply. Nature Biotechnology. 2007, 25 (1): 31-33. 10.1038/nbt0107-31.CrossRef
41.
Zurück zum Zitat Brown LG: A Survey of Image Registration Techniques. Computing Surveys. 1992, 24 (4): 325-376. 10.1145/146370.146374.CrossRef Brown LG: A Survey of Image Registration Techniques. Computing Surveys. 1992, 24 (4): 325-376. 10.1145/146370.146374.CrossRef
42.
Zurück zum Zitat Nelder JA, Mead R: A Simplex-Method for Function Minimization. Computer Journal. 1965, 7 (4): 308-313.CrossRef Nelder JA, Mead R: A Simplex-Method for Function Minimization. Computer Journal. 1965, 7 (4): 308-313.CrossRef
43.
Zurück zum Zitat Lee JS: Speckle Suppression and Analysis for Synthetic Aperture Radar Images. Optical Engineering. 1986, 25 (5): 636-643.CrossRef Lee JS: Speckle Suppression and Analysis for Synthetic Aperture Radar Images. Optical Engineering. 1986, 25 (5): 636-643.CrossRef
44.
Zurück zum Zitat Pizer SM, Amburn EP, Austin JD, Cromartie R, Geselowitz A, Greer T, Terhaarromeny B, Zimmerman JB, Zuiderveld K: Adaptive Histogram Equalization and Its Variations. Computer Vision Graphics and Image Processing. 1987, 39 (3): 355-368. 10.1016/S0734-189X(87)80186-X.CrossRef Pizer SM, Amburn EP, Austin JD, Cromartie R, Geselowitz A, Greer T, Terhaarromeny B, Zimmerman JB, Zuiderveld K: Adaptive Histogram Equalization and Its Variations. Computer Vision Graphics and Image Processing. 1987, 39 (3): 355-368. 10.1016/S0734-189X(87)80186-X.CrossRef
45.
Zurück zum Zitat Dougherty ER: An introduction to morphological image processing. 1992, Bellingham, Wash., USA: SPIE Optical Engineering Press Dougherty ER: An introduction to morphological image processing. 1992, Bellingham, Wash., USA: SPIE Optical Engineering Press
46.
Zurück zum Zitat Peng HC, Long FH, Ding C: Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy. Ieee Transactions on Pattern Analysis and Machine Intelligence. 2005, 27 (8): 1226-1238. 10.1109/TPAMI.2005.159.CrossRefPubMed Peng HC, Long FH, Ding C: Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy. Ieee Transactions on Pattern Analysis and Machine Intelligence. 2005, 27 (8): 1226-1238. 10.1109/TPAMI.2005.159.CrossRefPubMed
47.
Zurück zum Zitat Pudil P, Novovicova J, Kittler J: Floating Search Methods in Feature-Selection. Pattern Recognition Letters. 1994, 15 (11): 1119-1125. 10.1016/0167-8655(94)90127-9.CrossRef Pudil P, Novovicova J, Kittler J: Floating Search Methods in Feature-Selection. Pattern Recognition Letters. 1994, 15 (11): 1119-1125. 10.1016/0167-8655(94)90127-9.CrossRef
48.
Zurück zum Zitat Bhargava R: Towards a practical Fourier transform infrared chemical imaging protocol for cancer histopathology. Anal Bioanal Chem. 2007, 389 (4): 1155-1169. 10.1007/s00216-007-1511-9.CrossRefPubMed Bhargava R: Towards a practical Fourier transform infrared chemical imaging protocol for cancer histopathology. Anal Bioanal Chem. 2007, 389 (4): 1155-1169. 10.1007/s00216-007-1511-9.CrossRefPubMed
49.
Zurück zum Zitat Bhargava R, Fernandez DC, Hewitt SM, Levin IW: High throughput assessment of cells and tissues: Bayesian classification of spectral metrics from infrared vibrational spectroscopic imaging data. Biochimica Et Biophysica Acta-Biomembranes. 2006, 1758 (7): 830-845. 10.1016/j.bbamem.2006.05.007.CrossRef Bhargava R, Fernandez DC, Hewitt SM, Levin IW: High throughput assessment of cells and tissues: Bayesian classification of spectral metrics from infrared vibrational spectroscopic imaging data. Biochimica Et Biophysica Acta-Biomembranes. 2006, 1758 (7): 830-845. 10.1016/j.bbamem.2006.05.007.CrossRef
50.
Zurück zum Zitat Vapnik VN: The nature of statistical learning theory. 1995, New York: SpringerCrossRef Vapnik VN: The nature of statistical learning theory. 1995, New York: SpringerCrossRef
51.
Zurück zum Zitat Morik K, Brockhausen P, Joachims T: Combining Statistical Learning with a Knowledge-Based Approach - A Case Study in Intensive Care Monitoring. Proceedings of the Sixteenth International Conference on Machine Learning. 1999, Morgan Kaufmann Publishers Inc, 268-277. Morik K, Brockhausen P, Joachims T: Combining Statistical Learning with a Knowledge-Based Approach - A Case Study in Intensive Care Monitoring. Proceedings of the Sixteenth International Conference on Machine Learning. 1999, Morgan Kaufmann Publishers Inc, 268-277.
52.
Zurück zum Zitat Landwehr N, Hall M, Frank E: Logistic model trees. Lect Notes Artif Int. 2003, 2837: 241-252. Landwehr N, Hall M, Frank E: Logistic model trees. Lect Notes Artif Int. 2003, 2837: 241-252.
53.
Zurück zum Zitat Berney DM, Fisher G, Kattan MW, Oliver RTD, Moller H, Fearn P, Eastham J, Scardino P, Cuzick J, Reuter VE, et al: Pitfalls in the diagnosis of prostatic cancer: retrospective review of 1791 cases with clinical outcome. Histopathology. 2007, 51 (4): 452-457. 10.1111/j.1365-2559.2007.02819.x.CrossRefPubMed Berney DM, Fisher G, Kattan MW, Oliver RTD, Moller H, Fearn P, Eastham J, Scardino P, Cuzick J, Reuter VE, et al: Pitfalls in the diagnosis of prostatic cancer: retrospective review of 1791 cases with clinical outcome. Histopathology. 2007, 51 (4): 452-457. 10.1111/j.1365-2559.2007.02819.x.CrossRefPubMed
Metadaten
Titel
Multimodal microscopy for automated histologic analysis of prostate cancer
verfasst von
Jin Tae Kwak
Stephen M Hewitt
Saurabh Sinha
Rohit Bhargava
Publikationsdatum
01.12.2011
Verlag
BioMed Central
Erschienen in
BMC Cancer / Ausgabe 1/2011
Elektronische ISSN: 1471-2407
DOI
https://doi.org/10.1186/1471-2407-11-62

Weitere Artikel der Ausgabe 1/2011

BMC Cancer 1/2011 Zur Ausgabe

Update Onkologie

Bestellen Sie unseren Fach-Newsletter und bleiben Sie gut informiert.