Texture Characterization of Histopathologic Images Using Ecological Diversity Measures and Discrete Wavelet Transform
Abstract
Breast cancer is a health problem that affects mainly the female population. An early detection increases the chances of effective treatment, improving the prognosis of the disease. In this regard, computational tools have been proposed to assist the specialist in interpreting the breast digital image exam, providing features for detecting and diagnosing tumors and cancerous cells. Nonetheless, detecting tumors with a high sensitivity rate and reducing the false positives rate is still challenging. Texture descriptors have been quite popular in medical image analysis, particularly in histopathologic images (HI), due to the variability of both the texture found in such images and the tissue appearance due to irregularity in the staining process. Such variability may exist depending on differences in staining protocol such as fixation, inconsistency in the staining condition, and reagents, either between laboratories or in the same laboratory. Textural feature extraction for quantifying HI information in a discriminant way is challenging given the distribution of intrinsic properties of such images forms a non-deterministic complex system. This paper proposes a method for characterizing texture across HIs with a considerable success rate. By employing ecological diversity measures and discrete wavelet transform, it is possible to quantify the intrinsic properties of such images with promising accuracy on two HI datasets compared with state-of-the-art methods.
keywords:
Texture Characterization, Texture Classification, Ecological Diversity Measures, Discrete Wavelet Transform1 Introduction
Cancer is a disease caused by an uncontrolled division of cells that can happen anywhere in the body, susceptible to invading neighboring tissues and organs [20]. Cell proliferation does not necessarily imply the presence of malignancy but can simply respond to the body’s specific needs. More than a hundred known types of tumors affect people, but not all tumors are cancerous [46]. Benign tumors grow in an organized, generally slow, expansive manner and have apparent limits. Although they do not invade neighboring tissues, they can compress adjacent organs and tissues and are hardly life-threatening. In contrast, malignant tumors manifest a greater degree of autonomy. They are capable of invading neighboring tissues and causing metastases, which may be resistant to treatment and cause the patient’s death [26].
Breast cancer is the most frequent cancer among women globally, accounting for nearly a quarter of new cases each year [27], and is the most prevalent type of cancer in nearly 140 countries [64]. In addition to resulting in death, related treatment costs are high. It impacts both the public and the private healthcare system, which, in turn, penalizes both the government and the population.
Digital mammography analysis is among the most widely employed and effective methods for the early detection of breast cancer. Mammography is a radiological test that produces images in gray-scale, which are analyzed by radiologists, who need to give more attention and spend more time reliably detecting the cancer information from the images [23]. Likewise, imaging examinations like computerized tomography or ultrasound can diagnose whether there are masses growing in breast tissue, despite the verification of either type of tumor can only be accomplished employing biopsy. Biopsies, in turn, last to provide a result because of the acquisition procedure, which may imply open surgical biopsy or fine-needle aspiration, tissue processing (preparation of slides with a staining process) as well as pathologist visual examination. Innately, pathologist examination is a time-consuming and highly specialized task prone to intra and inter-observer disagreement [7].
Histopathology is the study of how a specific disease affects a set of cells (tissue). Usually, a biopsy study is done using a microscope and dyes. It can also be done during surgery or at an autopsy (death investigation). Histopathologic images (HIs) are a type of medical imaging obtained via microscopy of tissues from biopsies, which allow specialists to observe tissue characteristics on a cell basis. This process consists of tissue processing by chemical fixation or frozen section slides (creating the slide with the staining process). Next, the slides undergo staining process with one or several pigments to envision the tissue by means of a microscope, aiming to reveal cellular components; counterstains are used to provide contrast, and finally, the pathologist analysis. The stain by and large used in histopathology is a combination of hematoxylin and eosin (H&E). The former is employed to stain nuclei (blue), while the latter stains cytoplasm and the extracellular connective tissue matrix (pink). However, the staining process may give rise to a variance in the analysis process because H&E is prone to produce different color intensities being conditioned by the temperature, storage conditions, and brand. Nevertheless, as stated by de Matos et al. [41], HIs continue to be the gold standard for evaluating several types of tumors for cancer diagnosis.
Computer-aided detection (CAD) and computer-assisted diagnostic (CADx) systems are continuously being developed to assist medical image analysis. Clinicians are heavily reliant upon CAD for cancer detection and monitoring. However, given the reliance on CAD and CADx for cancer detection, there is always an extra focus and need for developing systems that improve pathologists’ productivity and ameliorate the reliance on outcomes by adding consistency to the diagnosis process and reducing observer subjectivity. Machine learning (ML) approaches are increasingly being used in CAD and HI analysis to diagnose cancer in various tissues or organs, such as the breast, prostate, skin, brain, bones, liver, and so on. Furthermore, when used in HI analysis, ML approaches reveal potential benefits. As a result, they have seen a lot of use in tasks like feature extraction, classification, and segmentation. The visual properties of macro vision images used in other machine learning applications, such as scene reconstruction, object and face recognition, differ from those of HIs, which contain complex textures and rich geometric structures.
This paper proposes a novel method for texture characterization of HIs with a considerable success rate. We state that it is possible to quantify the intrinsic properties of HIs to the maximum extent by combining biodiversity and taxonomic measures and discrete wavelet transform. Thus, the main contributions of our research are: (i) an information-theoretical measure of ecological diversity indices and measures of biodiversity for texture characterization; (ii) the exploitation of independent wavelet subband coefficients’ non-linear interactions across time; (iii) the mixture of wavelet features and statistical properties of taxonomic indexes representing an unexplored method based on a non-deterministic system analysis; (iv) such a mixture characterizes HIs so that intrinsic properties have provided promising performance for real-world HI datasets such as the CRC and the BreakHis datasets.
This paper is organized as follows: Section 2 presents works that put forth descriptors more popular in medical image analysis. Section 3 puts forth the proposed approach as well as the concepts from which it originated. Section 4 presents the datasets, experiments results, and discussion. Finally, the last section presents the conclusion and perspectives of future work.
2 Related Works
Several feature descriptors have been used to extract a relevant and discriminant information from images. Some are based on shape, texture, fractal, or combination of those mentioned above. Besides natural images, texture descriptors are becoming increasingly popular in medical image analysis, particularly in HIs due to the variability of texture that such images exhibit. Therefore, researchers have been studying a broad range of textural descriptors for the classification of HIs, which are expected to be invariant to translation, scale, rotation, and intensity changes.
Characterizing morphological features from structures observed in HIs, as well as exploring higher-level representations capable of capturing relevant information for medical diagnosis purposes, is one of the most preeminent challenges in extracting features from such images. The above characteristics are related to recognizing tissue alterations (such as cell density or aberrant cell quantity) or cellular changes (e.g. malformed nuclei) caused by mitotic phases. Furthermore, morphological traits are related to how pathologists investigate HIs, looking for specific reasons to categorize them. High-level features, on the other hand, are generalizations of all structures in HIs, not solely cell structures. As a result, most researchers exploit representations or texture descriptors in the frequency domain [41]. This section presents some state-of-the-art works for feature extraction from HIs.
Several authors have utilized the descriptors based on the grey-level co-occurrence matrix (GLCM) to represent texture in HIs. [36] employed GLCM for feature extraction with a segmentation process through unsupervised mean-shift clustering. The latter minimizes color variety to facilitate the segmentation of the image using thresholds. Following that, nuclei are identified, and overlapping is reduced employing contour and area constraints. [10] combined seven feature extraction methods, including GLCM, and constructed a kernel-based representation of the data for each feature type. The kernels are then used within an SVM to detect similarities between data for the implementation of a content retrieval mechanism. [19] proposed a feature extraction approach that uses frequency and spatial textons, implying that images are represented by a limited vocabulary of textures. Texton histograms and GLCM extracted from texton maps are used as classification features. Similarly, the effect of various colormaps on these procedures was assessed. Despite the fact that GLCM requires a gray-level image, the transformation of the H&E color image to gray-level is influenced by the staining color variability, which affects GLCM accordingly. [37] utilized a random forest to determine whether or not GLCM features are susceptible to image variations. The work also emphasized the significance of color normalization.
[57] focused on stroma maturity to assess breast cancer. Image convolution with a bank of derivatives-of-Gaussian filters, basic image features, and LBP with multiple scales for the neighborhood produce the feature vector. [16] proposed geometric- and texture-aware features based on Hu moments and fractal dimensions, respectively. The latter was used to distinguish between mitotic and non-mitotic cells by detecting changes in geometrical and textural nuclei. [15] proposed a patching method on HI slides, intending to create small regions and extract scale-invariant feature transform (SIFT), luminance level, and discrete cosine transform features to generate a bag-of-words.
The local binary pattern (LBP) is one of the most commonly used texture descriptors. LBP was used in the context of multispectral HIs by Peyret et al. [52]. An SVM was used for the evaluation, which aligns all spectra and uses pixels from all other bands. Similarly, it employs a multi-scale kernel size. When compared to the standard LBP and the concatenated spectra LBP, this feature extractor performed better. To deal with multi-scale HIs, Bruno et al. [8] used a curvelet transform. The LBP algorithm was used to extract features from curvelets coefficients, which were then reduced using ANOVA. [54] proposed an algorithm that finds nuclei areas using adaptive and iterative thresholding and extracts texture information using histograms of oriented gradients and LBP.
[25] proposed a two-step feature extraction method comprised of a receptive field for detecting regions of interest and sparse coding in their study. Sparse coding assembles features by combining patches from the same region. Furthermore, the mean and covariance matrix of receptive fields, as well as sparse coding, are utilized as final filters. [47] proposed an approach for distinguishing basal cell carcinoma tumors from squamous cell carcinoma tumors in skin HIs by using Z-transform features derived from a mixture of Fourier transform features. Wan et al. [67] employed a dual-tree complex wavelet transform to represent the HIs in breast cancer detection for mitosis detection. The feature vector is made up of parameters from the symmetric alpha-stable distribution and the generalized Gaussian distribution. Chan and Tuszynski [12] used fractal dimension features to detect breast cancer. These features distinguish between malignant and benign tumors satisfactorily on HIs at 40 magnification. Finally, Ataky and Lameiras Koerich [2] presented a bio-inspired texture (BiT) descriptor based on biodiversity and taxonomic indices. The authors mapped breast cancer HI images to an abstraction model of an ecosystem, from which measures of species diversity, richness, evenness, and taxonomic distinctiveness were extracted. The resulting texture descriptor has been shown to be rotation, translation, and scale invariant. Experiments on HI datasets revealed that the BiT achieved competitive results when compared to deep methods, and outperformed traditional texture descriptors.
In the last years, deep features have become very popular in several image classification tasks, including HIs. Khalid Khan Niazi et al. [33] put forward a CAD system for bladder cancer with a focus on extracting epithelium features with segmentation through an automatic color deconvolution matrix construction. Spanhol et al. [60] employed deep features extracted with a pre-trained AlexNet for the classification of breast tumors into benign and malignant. The method proposed by Vo et al. [66] put forth a feature extraction based on the combination of CNNs and boosting tree classifiers. The latter employs an ensemble of inception CNNs for visual features extraction from multi-scale images. Firstly, data augmentation methods were employed. Subsequently, ensembles of CNNs were trained to extract multi-context information from multi-scale images. The last-mentioned stage extracted both global and local features of breast cancer tumors. George et al. [22] presented an approach for breast cancer diagnosis, which extracts features from nuclei based on CNNs. The methodology comprises different approaches for extracting nucleus features from HIs and selecting the most discriminative spatially sparse nucleus patches. A pre-trained set of CNNs was used to extract features from such patches. Afterward, features belonging to individual images are fused using 3-norm pooling to obtain image-level features.
Various works employed and combined different categories of features for capturing information from geometrical structures and textures from HIs. Leo et al. [37] introduced a method aiming to quantify features’ instability across a few prostate cancer datasets with known variations caused by staining, preparation, and scanning platforms. The author evaluated five groups of features, such as graph-based features, gland shape features, co-occurring gland tensor features, subgraph features, and Haralick texture features. Yu et al. [69] investigated features that characterize lung cancer the best. They extracted the objective quantitative image features such as Haralick texture features of the nuclei, nuclei edge intensity, texture features of the cytoplasm, etc.
The work of Caicedo et al. [9] put forth a low-level to high-level mapping to facilitate image retrieval. Such mapping stage consists of color and gray-level histograms, Tamura texture histogram, LBP, Sobel histogram, and invariant feature histograms. Pang et al. [51] raised a CAD system for lung cancer detection. Such a CAD system utilizes textural features, such as GLCM, Tamura, and LBP, and shape features, namely global features, SIFT, and morphological features. The work of Kruk et al. [35] employed textural, morphometric, and statistical (histogram) features to describe nuclei for clear-cell renal carcinoma grading. To this end, the Genetic algorithm and the Fisher discriminant were utilized to select essential features. Basavanhally et al. [5] introduced a multi-field-of-view classification method to determine low against high-grade ductal carcinoma from breast HIs. This method utilizes a multiple patch size procedure for WSI to analyze which textural, morphological, and graph-based features are the most relevant to each patch size. Tashk et al. [63] put forth a comprehensive framework for breast HI classification that evaluates mitotic pixels in L*a*b color space. A mixture of LBP, statistical features, and morphometrics is extracted from mitotic candidates. The work of Cruz-Roa et al. [15] introduced a patching approach on HI slides to form small regions and extract SIFT, discrete cosine transform, and luminance level features to produce a bag-of-words. Besides, semantic features represent high-level information that can be associated with HIs to aid their classification.
Orlov et al. [50] compared four color spaces (RGB, L*a*b, gray-scale and RGB) with H&E representation and eleven features such as GLCM, Zernike, Tamura, Chebychev, color histograms, Gabor, edge statistics, Chebyshev-Fourier, and others to represent lymph node HIs. De et al. [17] propose a fusion of several feature types for uterine cervical cancer HI classification. The authors employed a feature vector based on Delaunay triangulation, GLCM, and weighted density distribution. The work of Vanderbeck et al. [65] utilized textural, pixel neighboring statistics, and morphological features to represent seven categories of white regions of liver HIs. Kandemir et al. [30] put forth a MIL method to detect Barrett’s cancer from HIs. They utilized cell-level morphometric features, to wit, radius, perimeter, central power sums, area, the roundness of segments, and so forth, within regions and patch-level features such as color histograms, SIFT, LBP, and from segmented images employing the watershed algorithm.
Coatelen et al. [14, 13] proposed a feature selection method of liver HI classification based on morphometric features and graph-based features. Pair of greedy algorithms such as fselector and in-house recursive was employed to select features in a collection of 200 features where an SVM classifier implemented the fitness function. The work of Michail et al. [43] highlighted nuclei using connected-component labeling to classify non-centroblast and centroblast cells. Das et al. [16] presented the so-called geometric- and texture-aware features based on Hu moments and fractal dimensional, respectively. The features set thereof was applied to detect geometrical and textural changes in nuclei to discriminate between non-mitotic and mitotic cells. The method introduced by Kong et al. [34] classifies neuroblastomas utilizing textural and morphological features. The authors consider that pathologists employ morphological features for their analysis, and textural features can be easily extracted. GLCM features and sequential floating forward selection were then utilized.
These are some examples of works developed to detect breast cancer through HIs. However, in spite of the substantial efforts and countless methods that have been proposed in recent years, accurate classification of HIs remains a challenge. Existing approaches developed to cope with such a purpose still arise performance issues in light of noise, different image resolutions, and a lack of a considerable amount of data. Furthermore, methods based on CNNs present a lack of explainability and interpretability, which are both necessary for understanding the behavior of such models, and trusting their decisions. Likewise, pre-trained CNN architectures designed for object classification still require fine-tuning some of their layers on a large amount of data to achieve good performance, including tiny architectures such as T-CNN Inception and T-CNN [42, 3].
3 Proposed Approach
This work proposes an efficient method for texture characterization of HIs combining ecology diversity measures and multi-resolution analysis. Combining biodiversity and taxonomic measures with the discrete wavelet transform (DWT) makes it possible to quantify the intrinsic properties of HIs to the maximum extent. Thus, the fundamental research highlights of the proposed method are: (i) an information-theoretical measure of ecological diversity indices and measures of biodiversity; (ii) the exploitation of non-linear interactions of single and independent wavelet subband coefficients over time; (iii) the mixture of wavelet features and statistical properties of taxonomic indexes representing an unexplored method for non-deterministic pattern’s systems analysis.
This section presents the concepts of information-theoretical measures of ecological diversity, multi-resolution analysis, and DTW, which are the base concepts of our method, before presenting our method for texture characterization of HIs.
3.1 Information-Theoretical Measure of Ecological Diversity
The information-theoretical measures of ecological diversity are used in biology to compare behavioral patterns between species in different areas and within-neighborhood. Similarly, diversity indices based on species richness are of an underlying use when describing an all-inclusive behavior of an ecosystem, constituting a non-deterministic system of patterns. Ataky and Lameiras Koerich [2] proposed the bio-inspired texture descriptor (BiT) based on an information-theoretical measure of taxonomic diversity that includes the species’ richness, abundances, and taxonomic distinctness. They consider an image as an abstract model of an ecosystem where pixels correspond to individuals, gray levels correspond to species, the number of different gray levels corresponds to species richness, and the number of distinct gray levels in a specific region corresponds to species abundance.
Species richness of an image represents the number of gray levels therein contained. The higher the richness index, the more diverse the system is. The richness indices regard but the abundance of each species (gray levels) and among those present in BiT are Margalef’s diversity index () and Menhinick’s diversity index (). Both are the ratio between the total number of gray levels recorded () and the total number of pixels in the image ():
(1) |
(2) |
BiT integrates Shannon-Wiener () and McIntosh’s () indices to account for the spatial variance of diversity. They are defined as the proportion of pixels of gray level in terms of , and the ratio between the number of pixels in the -th gray level and the total number of pixels, and the number of gray level in an image, respectively.
(3) |
(4) |
where denotes the proportion of pixels with the -th gray level, and denotes the number of pixels of the -th gray level (the summation is over all gray levels).
Other diversity measures integrated into BiT are Berger-Parker dominance (), which represents the ratio between the number of pixels in the most abundant gray level () and , Fisher’s alpha diversity metric (), which represents the number of groups of closely related pixels, and Kempton-Taylor index of alpha diversity (), which quantifies the interquartile slope of the cumulative abundance curve.
(5) |
(6) |
(7) |
where is nearly equal to the number of gray levels represented by a single pixel, denotes the number of gray levels with abundance , and are the 25% and 75% quartiles of the cumulative gray level curve, is the number of pixels in the class where falls, and is the number of pixels in the class where falls.
Because such indices may be insensitive to taxonomic differences or similarities, Ataky and Lameiras Koerich [2] have also integrated taxonomic indices, such as diversity and distinctness, which consider the taxonomic relationship between different pixels within an image. Taxonomic diversity () includes aspects of taxonomic relatedness considering the abundance of different gray levels and the taxonomic relationship between them. Its value represents the average taxonomic distance between any two pixels, chosen randomly from an image. Taxonomic distinctiveness (), in turn, denotes a measure of pure taxonomic relatedness. It is the average taxonomic distance between two pixels of different gray levels.
(8) |
(9) |
where and denote the number of pixels that have the -th and -th gray level in the image, respectively, and denotes the ’distinctness weight’ (distance) given to the path length linking pixels and in the hierarchical classification.
Other taxonomic indices integrated into BiT are the sum of phylogenetic distances (), which denotes the sum of phylogenetic distances between pairs of gray levels, the average distance from the nearest neighbor (), which denotes the average distance to the nearest group of one or more gray level of the image to form a unit, the extensive quadratic entropy (), which denotes the sum of the differences between gray levels, the intensive quadratic entropy (), which denotes the number of gray levels and their taxonomic relationships. It establishes a possible link between the diversity indices and the biodiversity measurement indices by expressing the average taxonomic distance between two randomly chosen gray levels. The relationships between the latter influence the entropy, in contrast to other diversity indices. Finally, the last taxonomic index is the total taxonomic distinctness (), which denotes the average phylogenetic distinctiveness added across all gray levels.
(10) |
(11) |
(12) |
(13) |
(14) |
where and denote two distinct gray levels, and is the number of pixels that have such gray levels, and is the distance between the gray levels and ; is the abundance of the referred gray level, and the total number of gray levels.
Such ecological diversity measures may characterize the texture of HIs through its second-order statistical properties, which implies comparing neighboring pixels and defining how a pixel at a specific location relates statistically to pixels at different locations. Furthermore, because they rely on group analysis, they enable a behavioral exploration of the neighborhood of regions displaced from a reference location. In addition, the BiT descriptor takes advantage of the invariance characteristics of ecological patterns such as rotation, scale, and reflection.
The BiT descriptor was developed to estimate diversity and parenthood between gray levels from images. The fundamental idea of combining biodiversity measures and taxonomic indices is to obtain a quantitative estimate of textural variability in space or time that can compare textural entities composed of several components. Applying these measures to HIs mapped as an ecosystem, they explore different periodicities and attempt to characterize their texture. This analysis is constrained to the neighboring of individual pixels, and the within-neighborhood periodicity properties can be used to determine texture differences between different regions.
3.2 Multi-Resolution Analysis
The multi-resolution analysis is a signal processing strategy that employs filter banks to extract relevant information from signals, such as the frequencies and their locations depending on the duration of the signal, and at different resolutions [11]. The motivation behind the multi-resolution analysis of HIs is to define each subband of the wavelet decomposition as a system modeled by ecological ecosystem diversity.
Like natural images, texture across HIs presents non-deterministic patterns that resemble a complex system. The wavelet decomposition plays a fundamental role in defining the spatial dependence structure (correlations between elements) because there is a direct analogy between wavelet subbands and physical systems of particles [38]. Considering that an image is modeled as an ecosystem, we exploit several wavelet subbands to reflect seasonal variation in the context of fluctuation-mediated coexistence. Such a variation in species may relate to the phenology, particularly in the context of season changes (time/frequency). Integrating statistical properties of taxonomic indexes from different variations (subbands) may help broader the view of ecosystems (intrinsic properties of an HI). Thus, with the wavelet decomposition, we exploit non-linear interactions of individual subband coefficients over time for capturing the textural information of HIs based on the principle that most ecosystems work in a cause-effect relationship.
The succinct description of the multi-resolution analysis allows presenting two functions responsible for generating an entire wavelet system: the primary wavelet and the scale function. The scale functions and wavelets , are orthogonal because they respect the following condition:
(15) |
where represents the scale of the function, and corresponds to the translation of concerning the scale function and the primary wavelet, given by and . The scale function and the wavelet are defined in . The translation parameter corresponds to the time information in the transform domain, and the scaling parameter represents the signal compression and expansion process [39].

Fig. 1 illustrates the decomposition of an HI into three levels through the wavelet transform. The useful information obtained from the initial image of the vessel, on a smaller scale, is found in the six adjacent squares. The information contained in these squares is called detail or resolution, which is the information needed to move from one degree of refinement (or ”sharpness”) to another. By adding the information related to the squares, it is possible to recompose the image. This way of decomposing and recomposing images can be implemented quickly and effectively, employing wavelet transforms.
The continuous wavelet transform can be expressed as:
(16) |
where and represent the translation and scale parameters, respectively. However, this transformation requires infinite translations and scaling.
In the discrete wavelet transform (DWT), wavelets are not scaled or translated continuously but at discrete intervals, which is achieved by modifying the continuous wavelet as:
(17) |
(18) |
where and are integers, 1 is a fixed expansion parameter, and is the translation factor that depends on the expansion factor. Generally, 2 is chosen for a frequency sampling called dyadic sampling, and 1 is chosen for temporal sampling, also dyadic [49]. This results in:
(19) |
A wavelet can be seen as a low-pass filter, and a series of scaled wavelets can be seen as a bank of band-pass filters with a factor (filter bank fidelity factor). DWT can be applied to decompose an HI into two other HIs through low-pass filters (scaling signals) and high-pass filters (wavelet signals). Therefore, DWT can be constructed as a perfect reconstruction filter bank with pairwise quadrature mirror filters and [48]. The synthesis filters and , used to reconstruct the original signal from wavelet coefficients, are required for a complete DWT specification. The inverse discrete wavelet transform is the term for this technique.

Fig. 2 shows the decomposition of an HI, and it consists of four subbands: approximation (a), horizontal detail (h), vertical detail (v), and diagonal detail (d). In the next scale, the subband image (a) is used for DWT computation.
3.3 Ecological Modeling of Wavelet Subbands
The proposed method is employed and integrated for texture classification as follows: (1) image channel splitting; (2′) wavelet subband decomposition; (2′′) computation of biodiversity measures and information theory from each channel (R, G, and B) to form a features vector; (3) computation of taxonomic indexes and information theory from each wavelet subband to form a features vector; (4) feature vector concatenation; and (5) classification and performance evaluation. Fig. 3 shows the general overview of the proposed scheme.
Channel Splitting:
we applied the integrative method to make each image channel (R, G, B) a separate input. The motivation is to exploit color information. Therefore, we represent and characterize an input image by a set of local descriptors generated from the interaction of a pixel with its neighborhood from a given channel (R, G, or B) as well as wavelet subbands thereof.
Feature Extraction:
after the channel splitting stage, each image channel undergoes feature extraction. Intrinsic properties and discriminant characteristics within each input channel are extracted as follows. First, for each channel (R, B, and G), we perform a multi-resolution analysis of a texture employing a wavelet transform, generating four subbands: a, h, v, and d. Since we used three levels, the subband a is used for DWT computation in the next scale. After, we compute taxonomic measures, Shannon entropy, and total information, resulting in a 9-dimensional vector for each subband. This sequence can be seen in Fig. 3 (steps 2′ and 3). Taxonomic measures comprise taxonomic diversity, taxonomic distinctness, the sum of phylogenetic distances, average distance from the nearest neighbor, extensive quadratic entropy, intensive quadratic entropy, and total taxonomic distinctness [2]. Because we have split an input image into three channels and chosen a 3-level wavelet decomposition (leading to 10 subbands), step 3 will produce a 270-dimensional feature vector (9310). In parallel, we compute biodiversity measures and again Shannon entropy and total information directly from each original image channel resulting in a 9-dimensional vector for each channel. Biodiversity measures from step 2′′ comprise Margalef’s and Menhinick’s diversity indexes, Berger-Parker dominance, Fisher’s alpha diversity metric, Kempton-Taylor index of alpha diversity, McIntosh’s evenness measure, and Shannon-Wiener diversity index [2]. The reason for not employing DWT in this stage is that low-pass decomposition filters may present negative coefficients in the subband a depending on filter coefficients used for decomposition. Nevertheless, biodiversity measures such as abundance and richness are non-negative. Because we have split an input image into three channels, step 2′′ will produce a 27-dimensional feature vector (93). Finally, feature vectors resulting from steps 2′′ and 3 are concatenated to form the final feature vector (step 4 in Fig. 3). We named it BiTW descriptor because it results from the concatenation of biodiversity measures information theory, and taxonomic indexes extracted from wavelet subbands. The BiTW is a 297-dimensional feature vector, which may leave a whole path to a possible feature selection step. However, this is out of the scope of this paper.
Normalization:
the feature vectors are split into training and test sets before the training step. Then, normalization and scaling occur independently on each feature in the training set, where values are normalized to the range using the min-max normalization. Minimum and maximum are then stored to be used on feature normalization over the testing data. The same procedure is used for the -fold cross-validation (CV), where feature vectors are split into folds and computing the min-max pairs in the merged training folds. The min-max pairs obtained on the trained data are employed to normalize the training and the test folds This procedure is repeated for each new training/test fold during the CV procedure.
Training and Classification:
after feature extraction, the resulting feature vectors are taken through the classification process utilizing seven classifiers: histogram-based algorithm for building gradient boosting ensembles of decision trees (HistoB), light gradient boosting decision trees (LightB), fast, scalable, high-performance gradient boosting decision trees (CatBoost), extra trees (ExtraT), random forest (RF), gradient boosting decision trees (GB), and linear discriminant analysis (LDA). The performance analysis is conducted afterward using accuracy and area under the receiver operating characteristic curve (AUC).

4 Experiments and Results
This section presents the two datasets used to assess the performance of the BiTW descriptor and the experimental protocol to evaluate the properties of the BiTW descriptor and its performance on classification tasks considering different classification and ensemble methods. We compare the BiTW descriptor’s performance with shallow and deep state-of-the-art approaches.
4.1 Image Datasets
Two medical datasets were used in the experiments:
-
1.
CRC [31](Fig. 4): is a dataset of colorectal cancer histopathology images of 5,0005,000 pixels that were patched into 150150 images and labeled according to the structure they contain. Eight types of structures are labeled: stroma (ST), tumor (T), complex stroma (C), immune or lymphoid cell (L), mucosa (M), debris (D), adipose (AD), and empty or Background (E). There is a total of 625 images per structure type, resulting in 5,000 images.
Figure 4: Example of HIs from the CRC dataset: (a) Tumor, (b) Stroma, (c) Complex, (d) Lympho, (e) Debris, (f) Mucosa, (g) Adipose, (h) Empty. -
2.
BreakHis [61] (Fig. 5): comprises 9,109 microscopic images of breast tumor tissue collected from 82 patients using different magnification factors (40, 100, 200, and 400). To date, it contains 2,480 benign and 5,429 malignant samples (700 460 pixels, 3-channel RGB, 8-bit depth in each channel, PNG format). The dataset is imbalanced.
Figure 5: Example of HIs from the BreakHis dataset: (a) Adenosis, (b) Fibroadenoma, (c) Phyllodes, (d) Tabular adenomaa, (e) Ductal carcinoma, (f) Lobular carcinoma, (g) Mucinous carcinoma, (h) Papillary carcinoma, where (a-d) are benign tumors and (e-f) malignant tumors.
4.2 Experimental Results
Table 1 presents the accuracy of both the monolithic classifiers and ensemble methods trained with feature vectors from BiTW on the CRC dataset. GB yielded the best results for both train-test split and -fold CV in the group of all classification algorithms. Likewise, we have computed the area under the ROC curve (AUC), another important metric commonly used in medical images, which accounts for the true-positive rate against the false-positive rate at different threshold levels. The AUC achieved by the BiTW description with GB was 0.99. It is worthy of note that AUCs of 0.7 to 0.8 are considered acceptable, 0.8 to 0.9 are considered excellent, and greater than 0.9 is deemed to be exceptional.
Experimental Protocol (Metric) | Classification Algorithms | ||||||
---|---|---|---|---|---|---|---|
HistoB | LightB | LDA | CatB | ExtraT | RF | GB | |
70/30 (Acc) | 92.42 | 92.20 | 89.27 | 92.20 | 92.42 | 92.42 | 93.28 |
70/30 (AUC) | 0.991 | 0.992 | 0.988 | 0.993 | 0.993 | 0.991 | 0.991 |
10-fold (Acc) | 91.12 | 91.81 | 89.45 | 91.33 | 91.53 | 90.59 | 93.73 |
10-fold (SD) | 0.05 | 0.01 | 0.08 | 0.02 | 0.03 | 0.02 | 0.02 |
10-fold (AUC) | 0.990 | 0.992 | 0.989 | 0.993 | 0.993 | 0.991 | 0.994 |
Acc: Accuracy; SD: Standard Deviation. |
Table 2 compares the results achieved by BiTW + GB with the state-of-the-art for the CRC dataset. The BiTW slightly outperforms the accuracy achieved by almost all other methods. For instance, the difference in accuracy to the second-best method (shallow) is 0.77%, and with the third-best method (CNN) is 1.33%, considering an 8-class classification task and 10-fold CV. For 5-fold CV, BiTW slightly outperformed the second-best method (CNN) with the difference of 1.0%. It is noteworthy to highlight that CNNs generally require massive labeled datasets and, when not possible, may need pre-trained models and/or data augmentation to learn high-quality representations. In the medical field, however, this is not always possible. The shallow approaches using the BiTW descriptor, unlike the latter, did not require any data augmentation on the CRC dataset and has proven to be promising relative to CNNs as well as other shallow methods, despite HIs having other structures than textures.
Reference | Approach | Accuracy (%) | AUC | ||
---|---|---|---|---|---|
10-fold | 5-fold | ||||
Ribeiro et al. [58] | Shallow | 97.60∗ | – | 0.994 | |
Sarkar and Acton [59] | Shallow | 73.60 | – | – | |
Jørgensen et al. [28] | Shallow | – | – | 0.960 | |
Kather et al. [32] | Shallow | 96.90∗ | – | – | |
Kather et al. [32] | Shallow | 87.40 | – | – | |
Naiyar et al. [44] | Shallow | – | – | 0.960 | |
Rathore et al. [56] | Shallow | – | – | 0.970 | |
Kalkan et al. [29] | Shallow | – | – | 0.950 | |
Masood and Rajpoot [40] | Shallow | – | – | 0.900 | |
Ataky and Lameiras Koerich [2] | Shallow | 92.96 | – | – | |
BiTW+GB | Shallow | 93.73 | 93.60 0.2 | 0.994 | |
Wang et al. [68] | CNN | – | 92.60 | – | |
Pham [53] | CNN | – | 84.00 | – | |
Raczkowski et al. [55] | CNN | 92.40 | 92.20 | – | |
∗Used 2-classes classification instead (malignant and benign). |
Table 3 presents the accuracy of both the monolithic classifiers and ensemble methods trained with BiTW feature vectors on the BreakHis dataset. By employing the train-test split, the LightB achieved the best accuracy of 99.26% and 98.62% for 40 and 200 magnifications, respectively. The ExtraT achieved 98.50% and HistoB 98.38% accuracy for 100 and 400 magnifications, respectively. Furthermore, the AUC is nearly 0.98 or above regardless of the classifier or ensemble method. Considering the -fold CV, ExtraT yielded the best accuracy of 98.75% and 98.63% for 40 and 100 magnifications, respectively. The LightB, in turn, achieved an accuracy of 98.72% for 200 magnification. Finally, HistoB yielded an accuracy of 98.38% for 400 magnification. Furthermore, we carried out experiments with -fold cross-validation to ensure that every sample from the original dataset has the chance of appearing in the training and test set, which is a best practice in the presence of limited data. Table 4 presents the average accuracy with the BiTW descriptor on the BreakHis dataset at image level with a 10-fold CV. The results for both types of dataset splitting are very similar for nearly all the classifiers and ensemble methods.
Image Magnification | Classification Algorithm | |||||||
Metric | HistoB | LightB | LDA | CatB | ExtraT | RF | GB | |
40 | Acc | 98.97 | 99.26 | 99.26 | 98.25 | 98.97 | 98.25 | 98.18 |
AUC | 0.992 | 0.995 | 0.993 | 0.981 | 0.985 | 0.988 | 0.987 | |
100 | Acc | 98.43 | 98.30 | 92.41 | 97.19 | 98.50 | 97.19 | 96.64 |
AUC | 0.991 | 0.990 | 0.982 | 0.987 | 0.992 | 0.987 | 0.980 | |
200 | Acc | 98.55 | 98.62 | 90.88 | 98.62 | 98.55 | 98.00 | 97.58 |
AUC | 0.989 | 0.993 | 0.963 | 0.994 | 0.992 | 0.989 | 0.988 | |
400 | Acc | 98.38 | 97.98 | 90.29 | 97.74 | 97.98 | 97.19 | 97.74 |
AUC | 0.993 | 0.990 | 0.983 | 0.987 | 0.989 | 0.982 | 0.986 |
Image Magnification | Classification Algorithm | ||||||
---|---|---|---|---|---|---|---|
HistoB | LightB | LDA | CatB | ExtraT | RF | GB | |
40 | 98.62 | 98.61 | 98.61 | 98.06 | 98.75 | 98.04 | 98.12 |
100 | 98.48 | 98.45 | 91.60 | 97.74 | 98.63 | 97.87 | 97.76 |
200 | 98.54 | 98.72 | 91.55 | 97.98 | 98.65 | 97.98 | 98.53 |
400 | 98.76 | 98.74 | 91.00 | 98.08 | 98.33 | 97.58 | 97.77 |
Table 5 compares the results achieved by the proposed approach with the state-of-the-art for the BreakHis dataset. Though different classifiers outperformed others for different magnifications, for the sake of fairness, we choose a classifier to present for comparison with related work based on the average performance of the four magnifications. Computing the average we obtain 98.58%, 98.54%, 93.21%, 97.95%, 98.50%, 97.65% and 97.53%, for HistoB, LightB, LDA, CatB, ExtraT, RF and GB. Thus, HistoB is chosen for comparison purposes because it presented the highest average accuracy. The BiTW descriptor with HistoB achieved a substantial accuracy of 98.97%, 98.43%, 98.55% and 98.38% for 40, 100, 200, and 400 magnifications, respectively. It is important to notice that the proposed approach outperforms the accuracy of shallow and deep methods, regardless of the magnification. The differences in accuracy between the proposed method and the second- and third-best methods are 1.47% (Shallow) and 1.97% (CNN), 0.93% (CNN) and 1.63% (Shallow), 1.35% (CNN) and 2.75% (Shallow), 1.18% and 3.18% (Shallow) for 40, 100, 200, and 400 magnifications, respectively.
Image Magnification | ||||||
Reference | Method | 40 | 100 | 200 | 400 | |
Alom et al. [1] | CNN | 97.00 | 97.50 | 97.20 | 97.20 | |
Han et al. [24] | CNN | 92.80 | 93.90 | 93.70 | 92.90 | |
Bayramoglu et al. [6] | CNN | 83.00 | 83.10 | 84.60 | 82.10 | |
Spanhol et al. [62] | CNN | 90.00 | 88.40 | 84.60 | 86.10 | |
Gandomkar et al. [21] | CNN | 94.10 | 93.20 | 94.70 | 93.50 | |
Bardou et al. [4] | CNN | 88.23 | 84.64 | 83.31 | 8.98 | |
Nawaz et al. [45] | CNN | – | 95.00 | – | – | |
Spanhol et al. [61]∗ | Shallow | 75.60 | 73.00 | 72.90 | 71.20 | |
Spanhol et al. [61]+ | Shallow | 74.70 | 76.80 | 83.40 | 81.70 | |
Erfankhah et al. [18]∗ | Shallow | 88.30 | 88.30 | 87.10 | 83.40 | |
Ataky and Lameiras Koerich [2]† | Shallow | 97.50 | 96.80 | 95.80 | 95.20 | |
BiTW + HistoB | Shallow | 98.97 | 98.43 | 98.55 | 98.38 | |
∗LBP descriptor; +GLCM descriptor;†BiT descriptor. |
4.3 Discussion
The proposed approach was assessed with two HI datasets, both with eight classes. The experiment protocol employed a train-test split (70/30) and a -fold CV. For either experimental protocol, the results led to the following findings:
-
(1)
Exploiting information-theoretical measures of ecological diversity indices in conjunction with non-linear interactions of single and independent wavelet subband coefficients throughout time yielded promising results.
-
(2)
Although HIs contain other structures than texture, it was possible to characterize texture and achieve a good discriminating capability by employing biodiversity measures and taxonomic indexes together with multi-resolution analysis through DWT.
-
(3)
Such a mixture allowed the characterization of HIs to such an extent that intrinsic properties have provided a promising performance for the real-world datasets classification, reaching 93.73% accuracy and 0.994 AUC for the CRC dataset. Regarding the results on the BreakHis dataset, the accuracy were 99.26%, 98.50%, 98.62%, and 98.76% for 40, 100, 200 and 400 magnification, respectively. The AUC of all the magnifications was above 0.98.
Overall, the proposed approach outperformed state-of-the-art shallow and deep approaches on CRC and BreakHis datasets, regardless of the non-textural information that HIs may contain.
5 Conclusion
The current research leveraged the information-theoretical measure of ecological diversity indices together with a discrete wavelet transform to characterize texture across HIs. We explored the interactions of individual wavelet subband coefficients over time and modeled each as an ecosystem from which measures of biodiversity and statistical properties of taxonomic indexes are extracted to represent HI texture effectively. We stated that by combining measurements of biodiversity from each HI channel and taxonomic indexes extracted from different wavelet subbands, it should be possible to quantify the intrinsic properties of such images to the maximum extent. The mixture of wavelet features and statistical properties of ecology diversity indexes represent a novel method and a promising tool for quantifying intrinsic properties of texture across HIs. Wherefore, the experimental results have shown an increase in terms of texture discrimination over both HI datasets. Moreover, the proposed method outperformed several shallow and deep state-of-the-art methods.
In future work, we intend to further improve the classification accuracy by exploring various color spaces and different families of wavelets to find an optimized wavelet that will bring on discriminative information for characterizing HI categories. Furthermore, because an increasing number of DWT-level decompositions can result in high-dimensional feature vectors, we want to look into appropriate feature selection approaches or dimensionality reduction techniques to find feature subspaces that not only improve texture features but also improve the proposed method’s class discrimination capability.
Acknowledgment
This work was funded by the Regroupement Strategique REPARTI - Fonds de Recherche du Québec - Nature et Technologie (FRQNT) and by the Natural Sciences and Engineering Research Council of Canada (NSERC) under Grant RGPIN 2016-04855.
References
- Alom et al. [2019] Alom, M.Z., Yakopcic, C., Nasrin, M.S., Taha, T.M., Asari, V.K., 2019. Breast cancer classification from histopathological images with inception recurrent residual convolutional neural network. Journal of Digital Imaging 32, 605–617.
- Ataky and Lameiras Koerich [2022] Ataky, S.T.M., Lameiras Koerich, A., 2022. A novel bio-inspired texture descriptor based on biodiversity and taxonomic measures. Pattern Recognition 123, 108382. URL: https://www.sciencedirect.com/science/article/pii/S0031320321005628, doi:https://doi.org/10.1016/j.patcog.2021.108382.
- Ataky et al. [2020] Ataky, S.T.M., de Matos, J., de Souza Britto Jr., A., Oliveira, L.E.S., Koerich, A.L., 2020. Data augmentation for histopathological images based on gaussian-laplacian pyramid blending, in: Intl Joint Conf on Neural Networks, (IJCNN), Glasgow, UK, pp. 1–8. URL: https://arxiv.org/abs/2002.00072, arXiv:2002.00072.
- Bardou et al. [2018] Bardou, D., Zhang, K., Ahmad, S.M., 2018. Classification of breast cancer based on histology images using convolutional neural networks. Ieee Access 6, 24680--24693.
- Basavanhally et al. [2013] Basavanhally, A., Ganesan, S., Feldman, M., Shih, N., Mies, C., Tomaszewski, J., Madabhushi, A., 2013. Multi-Field-of-View Framework for Distinguishing Tumor Grade in ER+ Breast Cancer From Entire Histopathology Slides. IEEE Transactions on Biomedical Engineering 60, 2089--2099. doi:10.1109/TBME.2013.2245129.
- Bayramoglu et al. [2016] Bayramoglu, N., Kannala, J., Heikkilä, J., 2016. Deep learning for magnification independent breast cancer histopathology image classification, in: 23rd Int’l Conf on Pattern Recognition, pp. 2440--2445.
- Bellocq et al. [2011] Bellocq, J.P., Anger, E., Camparo, P., Capron, F., Chenard, M.P., Chetritt, J., Chigot, J.P., Cochand-Priollet, B., Coindre, J.M., Copin, M.C., Fléjou, J.F., Galateau, F., Gaulard, P., Guiu, M., Michiels, J.F., Saint-André, J.P., Scoazec, J.Y., Vacher-Lavenu, M.C., 2011. Sécuriser le diagnostic en anatomie et cytologie pathologiques en 2011. l’erreur diagnostique: entre discours et réalité. Annales de Pathologie 31, S92 -- S94. doi:https://doi.org/10.1016/j.annpat.2011.08.006. carrefour pathologie - 21 au 25 novembre 2011.
- Bruno et al. [2016] Bruno, D.O.T., do Nascimento, M.Z., Ramos, R.P., Batista, V.R., Neves, L.A., Martins, A.S., 2016. LBP operators on curvelet coefficients as an algorithm to describe texture in breast cancer tissues. Expert Systems with Applications 55, 329--340. doi:https://doi.org/10.1016/j.eswa.2016.02.019.
- Caicedo et al. [2008] Caicedo, J.C., Gonzalez, F.A., Romero, E., 2008. A semantic content-based retrieval method for histopathology images, in: Information Retrieval Technology, pp. 51+.
- Caicedo et al. [2011] Caicedo, J.C., González, F.A., Romero, E., 2011. Content-based histopathology image retrieval using a kernel-based semantic annotation framework. Journal of Biomedical Informatics 44, 519--528. doi:https://doi.org/10.1016/j.jbi.2011.01.011.
- Castleman [1996] Castleman, K.R., 1996. Digital image processing. Pearson.
- Chan and Tuszynski [2016] Chan, A., Tuszynski, J.A., 2016. Automatic prediction of tumour malignancy in breast cancer with fractal dimension. Royal Society Open Science 3. doi:10.1098/rsos.160558.
- Coatelen et al. [2015] Coatelen, J., Albouy-Kissi, A., Albouy-Kissi, B., Coton, J.P., Maunier-Sifre, L., Joubert-Zakeyh, J., Dechelotte, P., Abergel, A., 2015. A subset-search and ranking based feature-selection for histology image classification using global and local quantification, in: Intl Conf on Image Processing Theory, Tools and Applications (IPTA), pp. 313--318. doi:10.1109/IPTA.2015.7367154.
- Coatelen et al. [2014] Coatelen, J., Albouy-Kissi, A., Albouy-Kissi, B., Coton, J.P., Sifre, L., Joubert-Zakeyh, J., Dechelotte, P., Abergel, A., 2014. A feature selection based framework for histology image classification using global and local heterogeneity quantification, in: 36th Annual Intl Conf of the IEEE Engineering in Medicine and Biology Society, pp. 1937--1940. doi:10.1109/EMBC.2014.6943991.
- Cruz-Roa et al. [2011] Cruz-Roa, A., Caicedo, J.C., González, F.A., 2011. Visual pattern mining in histology image collections using bag of features. Artificial Intelligence in Medicine 52, 91--106. doi:https://doi.org/10.1016/j.artmed.2011.04.010.
- Das et al. [2017] Das, D.K., Mitra, P., Chakraborty, C., Chatterjee, S., Maiti, A.K., Bose, S., 2017. Computational approach for mitotic cell detection and its application in oral squamous cell carcinoma. Multidimensional Systems and Signal Processing 28, 1031--1050. doi:10.1007/s11045-017-0488-6.
- De et al. [2013] De, S., Stanley, R.J., Lu, C., Long, R., Antani, S., Thoma, G., Zuna, R., 2013. A fusion-based approach for uterine cervical cancer histology image classification. Computerized Medical Imaging and Graphics 37, 475--487. doi:10.1016/j.compmedimag.2013.08.001.
- Erfankhah et al. [2019] Erfankhah, H., Yazdi, M., Babaie, M., Tizhoosh, H.R., 2019. Heterogeneity-aware local binary patterns for retrieval of histopathology images. IEEE Access 7, 18354--18367.
- Fernández-Carrobles et al. [2015] Fernández-Carrobles, M.M., Bueno, G., Déniz, O., Salido, J., García-Rojo, M., Gonzández-López, L., 2015. Frequential versus spatial colour textons for breast {TMA} classification. Computerized Medical Imaging and Graphics 42, 25--37. doi:https://doi.org/10.1016/j.compmedimag.2014.11.009.
- Ferreira [2004] Ferreira, A.F., 2004. Conceitos gerais do câncer e do tratamento quimioterápico. Psicooncologia e interdisciplinaridade: uma experiência na educação à distância. Porto Alegre: Edipucrs , 59--72.
- Gandomkar et al. [2018] Gandomkar, Z., Brennan, P.C., Mello-Thoms, C., 2018. Mudern: Multi-category classification of breast histopathological image using deep residual networks. Artificial intelligence in medicine 88, 14--24.
- George et al. [2019] George, K., Faziludeen, S., Sankaran, P., Paul, J.K., 2019. Deep learned nucleus features for breast cancer histopathological image analysis based on belief theoretical classifier fusion, in: IEEE Region 10 Conf (TENCON), pp. 344--349.
- Giger [2000] Giger, M.L., 2000. Computer-aided diagnosis of breast lesions in medical images. Computing in Science & Engineering 2, 39--45.
- Han et al. [2017] Han, Z., Wei, B., Zheng, Y., Yin, Y., Li, K., Li, S., 2017. Breast cancer multi-classification from histopathological images with structured deep learning model. Scientific Reports 7, 1--10.
- Huang et al. [2011] Huang, C.H., Veillard, A., Roux, L., Loménie, N., Racoceanu, D., 2011. Time-efficient sparse analysis of histopathological whole slide images. Computerized Medical Imaging and Graphics 35, 579--591. doi:https://doi.org/10.1016/j.compmedimag.2010.11.009.
- INCA [2019a] INCA, 2019a. ABC do câncer: abordagens básicas para o controle do câncer. 2 ed., Ministério da Saúde, Rio de Janeiro, Brasil.
- INCA [2019b] INCA, 2019b. Tipo de câncer: próstata. http://www2.inca.gov.br/wps/wcm/connect/tiposdecancer/site/home/prostata. Online; acessado em 13 de Março de 2019.
- Jørgensen et al. [2017] Jørgensen, A.S., Rasmussen, A.M., Andersen, N.K.M., Andersen, S.K., Emborg, J., Røge, R., Østergaard, L.R., 2017. Using cell nuclei features to detect colon cancer tissue in hematoxylin and eosin stained slides. Cytometry Part A 91, 785--793.
- Kalkan et al. [2012] Kalkan, H., Nap, M., Duin, R.P., Loog, M., 2012. Automated classification of local patches in colon histopathology, in: Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), IEEE. pp. 61--64.
- Kandemir et al. [2014] Kandemir, M., Feuchtinger, A., Walch, A., Hamprecht, F.A., 2014. Digital pathology: Multiple instance learning can detect Barrett’s cancer, in: IEEE 11th Intl Symp on Biomedical Imaging, pp. 1348--1351. doi:10.1109/ISBI.2014.6868127.
- Kather et al. [2016a] Kather, J.N., Weis, C.A., Bianconi, F., Melchers, S.M., Schad, L.R., Gaiser, T., Marx, A., Zöllner, F.G., 2016a. Multi-class texture analysis in colorectal cancer histology. Scientific Reports 6, 27988. doi:10.1038/srep27988.
- Kather et al. [2016b] Kather, J.N., Weis, C.A., Bianconi, F., Melchers, S.M., Schad, L.R., Gaiser, T., Marx, A., Zöllner, F.G., 2016b. Multi-class texture analysis in colorectal cancer histology. Scientific Reports 6, 27988.
- Khalid Khan Niazi et al. [2016] Khalid Khan Niazi, M., Parwani, A.V., Gurcan, M.N., 2016. Computer-Assisted bladder cancer grading: -shapes for color space decomposition, in: Progress in Biomedical Optics and Imaging - Proceedings of SPIE. doi:10.1117/12.2216967.
- Kong et al. [2009] Kong, J., Sertel, O., Shimada, H., Boyer, K.L., Saltz, J.H., Gurcan, M.N., 2009. Computer-aided evaluation of neuroblastoma on whole-slide histology images: Classifying grade of neuroblastic differentiation. Pattern Recognition 42, 1080--1092. doi:10.1016/j.patcog.2008.10.035.
- Kruk et al. [2017] Kruk, M., Kurek, J., Osowski, S., Koktysz, R., Swiderski, B., Markiewicz, T., 2017. Ensemble of classifiers and wavelet transformation for improved recognition of Fuhrman grading in clear-cell renal carcinoma. Biocybernetics and Biomedical Engineering 37, 357--364. doi:https://doi.org/10.1016/j.bbe.2017.04.005.
- Kuse et al. [2010] Kuse, M., Sharma, T., Gupta, S., 2010. A classification scheme for lymphocyte segmentation in H&E stained histology images. Lecture Notes in Computer Science 6388 LNCS, 235--243. doi:10.1007/978-3-642-17711-8_24.
- Leo et al. [2016] Leo, P., Lee, G., Shih, N.N.C., Elliott, R., Feldman, M.D., Madabhushi, A., 2016. Evaluating stability of histomorphometric features across scanner and staining variations: prostate cancer diagnosis from whole slide images. Journal of Medical Imaging 3. doi:10.1117/1.JMI.3.4.047502.
- Levada [2014] Levada, A., 2014. Learning from complex systems: On the roles of entropy and fisher information in pairwise isotropic gaussian markov random fields. Entropy 16, 1002--1036. URL: https://www.mdpi.com/1099-4300/16/2/1002, doi:10.3390/e16021002.
- Mallat [1999] Mallat, S., 1999. A wavelet tour of signal processing. Elsevier.
- Masood and Rajpoot [2009] Masood, K., Rajpoot, N., 2009. Texture based classification of hyperspectral colon biopsy samples using clbp, in: 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, pp. 1011--1014. doi:10.1109/ISBI.2009.5193226.
- de Matos et al. [2021] de Matos, J., Ataky, S.T.M., de Souza Britto, A., Soares de Oliveira, L.E., Lameiras Koerich, A., 2021. Machine learning methods for histopathological image analysis: A review. Electronics 10, 562.
- de Matos et al. [2019] de Matos, J., de Souza Britto Jr., A., de Oliveira, L.E.S., Koerich, A.L., 2019. Texture CNN for histopathological image classification, in: 32nd IEEE Intl Symp on Computer-Based Medical Systems (CBMS), Cordoba, Spain, pp. 580--583. doi:10.1109/CBMS.2019.00120.
- Michail et al. [2014] Michail, E., Dimitropoulos, K., Koletsa, T., Kostopoulos, I., Grammalidis, N., 2014. Morphological and textural analysis of centroblasts in low-thickness sliced tissue biopsies of follicular lymphoma, in: Annual Intl Conf of the IEEE Engineering in Medicine and Biology Society, pp. 3374--3377.
- Naiyar et al. [2015] Naiyar, M., Asim, Y., Shahid, A., 2015. Automated colon cancer detection using structural and morphological features, in: 2015 13th International Conference on Frontiers of Information Technology (FIT), IEEE. pp. 240--245.
- Nawaz et al. [2018] Nawaz, W., Ahmed, S., Tahir, A., Khan, H.A., 2018. Classification of breast cancer histology images using alexnet, in: International conference image analysis and recognition, Springer. pp. 869--876.
- NCI [2019] NCI, 2019. Defining Cancer. https://www.cancer.gov/about-cancer/understanding/what-is-cancer. Online; accessed on 12 March 2020.
- Noroozi and Zakerolhosseini [2016] Noroozi, N., Zakerolhosseini, A., 2016. Computer assisted diagnosis of basal cell carcinoma using Z-transform features. Journal of Visual Communication and Image Representation 40, Part A, 128--148. doi:10.1016/j.jvcir.2016.06.014.
- Nsimba and Levada [2019] Nsimba, C.B., Levada, A., 2019. An information-theoretic wavelet-based texture descriptor using gaussian markov random field models. Multimedia Tools and Applications URL: https://doi.org/10.1007/s11042-019-07916-3, doi:10.1007/s11042-019-07916-3.
- Oliveira [2007] Oliveira, H., 2007. Análise de fourier e wavelets: sinais estacionários e não estacionários. Editora Universitária da UFPE .
- Orlov et al. [2010] Orlov, N.V., Chen, W.W., Eckley, D.M., Macura, T.J., Shamir, L., Jaffe, E.S., Goldberg, I.G., 2010. Automatic Classification of Lymphoma Images With Transform-Based Global Features. IEEE Transactions on Information Technology in Biomedicine 14, 1003--1013. doi:10.1109/TITB.2010.2050695.
- Pang et al. [2017] Pang, W., Jiang, H., Li, S., 2017. Sparse Contribution Feature Selection and Classifiers Optimized by Concave-Convex Variation for HCC Image Recognition. BioMed Research Intl 2017. doi:10.1155/2017/9718386.
- Peyret et al. [2018] Peyret, R., Bouridane, A., Khelifi, F., Tahir, M.A., Al-Maadeed, S., 2018. Automatic classification of colorectal and prostatic histologic tumor images using multiscale multispectral local binary pattern texture features and stacked generalization. Neurocomputing 275, 83--93. doi:https://doi.org/10.1016/j.neucom.2017.05.010.
- Pham [2017] Pham, T.D., 2017. Scaling of texture in training autoencoders for classification of histological images of colorectal cancer, in: Int’l Symposium on Neural Networks, Springer. pp. 524--532.
- Phoulady et al. [2016] Phoulady, H.A., Zhou, M., Goldgof, D.B., Hall, L.O., Mouton, P.R., 2016. Automatic quantification and classification of cervical cancer via Adaptive Nucleus Shape Modeling, in: IEEE Intl Conf on Image Processing (ICIP), pp. 2658--2662. doi:10.1109/ICIP.2016.7532841.
- Raczkowski et al. [2019] Raczkowski, Ł., Możejko, M., Zambonelli, J., Szczurek, E., 2019. Ara: accurate, reliable and active histopathological image classification framework with bayesian deep learning. Scientific Reports 9, 1--12.
- Rathore et al. [2013] Rathore, S., Iftikhar, M.A., Hussain, M., Jalil, A., 2013. Classification of colon biopsy images based on novel structural features, in: 2013 IEEE 9th International Conference on Emerging Technologies (ICET), IEEE. pp. 1--6.
- Reis et al. [2017] Reis, S., Gazinska, P., Hipwell, J.H., Mertzanidou, T., Naidoo, K., Williams, N., Pinder, S., Hawkes, D.J., 2017. Automated Classification of Breast Cancer Stroma Maturity from Histological Images. IEEE Transactions on Biomedical Engineering 64, 2344--2352. doi:10.1109/TBME.2017.2665602.
- Ribeiro et al. [2019] Ribeiro, M.G., Neves, L.A., do Nascimento, M.Z., Roberto, G.F., Martins, A.S., Tosta, T.A.A., 2019. Classification of colorectal cancer based on the association of multidimensional and multiresolution features. Expert Systems with Applications 120, 262--278.
- Sarkar and Acton [2017] Sarkar, R., Acton, S.T., 2017. Sdl: Saliency-based dictionary learning framework for image similarity. IEEE Trans on Image Processing 27, 749--763.
- Spanhol et al. [2017] Spanhol, F.A., Oliveira, L.S., Cavalin, P.R., Petitjean, C., Heutte, L., 2017. Deep features for breast cancer histopathological image classification, in: 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 1868--1873.
- Spanhol et al. [2016a] Spanhol, F.A., Oliveira, L.S., Petitjean, C., Heutte, L., 2016a. A Dataset for Breast Cancer Histopathological Image Classification. IEEE Trans on Biomedical Engineering 63, 1455--1462. doi:10.1109/TBME.2015.2496264.
- Spanhol et al. [2016b] Spanhol, F.A., Oliveira, L.S., Petitjean, C., Heutte, L., 2016b. Breast cancer histopathological image classification using convolutional neural networks, in: Int’l Joint Conf on Neural Networks, pp. 2560--2567.
- Tashk et al. [2015] Tashk, A., Helfroush, M.S., Danyali, H., Akbarzadeh-jahromi, M., 2015. Automatic detection of breast cancer mitotic cells based on the combination of textural, statistical and innovative mathematical features. Applied Mathematical Modelling 39, 6165--6182. doi:10.1016/j.apm.2015.01.051.
- Torre et al. [2017] Torre, L.A., Islami, F., Siegel, R.L., Ward, E.M., Jemal, A., 2017. Global cancer in women: Burden and trends. CEBP Focus: Global Cancer in Women 26, 444--457. doi:10.1158/1055-9965.EPI-16-0858.
- Vanderbeck et al. [2014] Vanderbeck, S., Bockhorst, J., Komorowski, R., Kleiner, D.E., Gawrieh, S., 2014. Automatic classification of white regions in liver biopsies by supervised machine learning. Human Pathology 45, 785--792. doi:10.1016/j.humpath.2013.11.011.
- Vo et al. [2019] Vo, D.M., Nguyen, N.Q., Lee, S.W., 2019. Classification of breast cancer histology images using incremental boosting convolution networks. Information Sciences 482, 123--138.
- Wan et al. [2017] Wan, T., Zhang, W., Zhu, M., Chen, J., Achim, A., Qin, Z., 2017. Automated mitosis detection in histopathology based on non-gaussian modeling of complex wavelet coefficients. Neurocomputing 237, 291--303. doi:https://doi.org/10.1016/j.neucom.2017.01.008.
- Wang et al. [2017] Wang, C., Shi, J., Zhang, Q., Ying, S., 2017. Histopathological image classification with bilinear convolutional neural networks, in: 39th Annual Int’l Conf of the IEEE Engineering in Medicine and Biology Society, pp. 4050--4053.
- Yu et al. [2016] Yu, K.H., Zhang, C., Berry, G.J., Altman, R.B., Re, C., Rubin, D.L., Snyder, M., 2016. Predicting non-small cell lung cancer prognosis by fully automated microscopic pathology image features. Nature Communications 7. doi:10.1038/ncomms12474.
Supplementary Material
Supplementary material that may be helpful in the review process should be prepared and provided as a separate electronic file. That file can then be transformed into PDF format and submitted along with the manuscript and graphic files to the appropriate editorial office.