Bio-inspired Computer Visual System using GPU

is rather complex and usually it is not possible to represented it by a single image ... This biological mechanism results in the color space of ..... similar to Karklin's model: first, the basis vectors of various .... Table I. It can be understood easily.
2MB taille 2 téléchargements 254 vues
Bio-inspired Computer Visual System using GPU and Visual Pattern Assessment Language (ViPAL): Application on Breast Cancer Prognosis (Invited Paper) Chao-Hui HUANG∗∗†¶ , Daniel RACOCEANU∗k† , Ludovic ROUX∗‡ , and Thomas C. PUTTI§ ∗ Image

& Pervasive Access Lab (IPAL, UMI2955, CNRS), France Institute, Agency for Science, Technology and Research, Singapore k Centre National de la Recherche Scientifique (CNRS), France † School of Computing, National University of Singapore, Singapore ‡ Universit´e Joseph Fourier, Grenoble, France § Department of Pathology of the National University Hospital, Singapore ¶ EMail: [email protected]

∗∗ Bioinformatics

Abstract—Bio-inspired computer vision is an emerging field. It aims to reproduce the capabilities of biological vision systems, eventually to simulate the visual functions for various purposes. In this paper, we propose a bio-inspired computer visual system using Graphical Processing Unit (GPU), and its application on breast cancer prognosis. The system simulates the major biological mechanisms of human visual system, such as encoding the edges, textures, and shapes of natural scenes. The system extracts visual features from an input image using a mechanism which is similar to human visual system. Then those visual features are converted to artificial neural activity signals which can be classified by a machine learning algorithm. As a result, the elements of the input image, which might be related to particular knowledge concepts, can be identified. In order to operate the bio-inspired computer visual system, in this work, a new computer language, named Visual Pattern Assessment Language (ViPAL), is proposed. The ViPAL provides a shell between the bio-inspired computer visual system and the users. The complexity of the low-level image feature assessment can be managed by the shell. As a result, the workload of the users can be reduced. Based on the bio-inspired computer visual system and the ViPAL, we propose an application for breast cancer prognosis. Currently, there is no computer-aided system to help pathologists making their final decision on breast cancer prognosis. Thus prognosis depends only on the medical human doctors. A welltrained pathologist usually can complete the analysis of a biopsy in 10 to 20 seconds. However, in a big medical institute, sometimes a pathologist need to analyse more than one hundred biopsies per day. So, for the sake of improving the reliability of breast cancer prognosis, we implemented the proposed bio-inspired computer visual system using GPU for detection of breast cancer invasive areas. As a result, object recognition and scene understanding of digitized breast biopsies are performed. Consequently, the system can roughly simulate the decision-making as the procedures of the breast cancer prognosis.

I. I NTRODUCTION Researchers have been interested for years in trying to mimic biological vision systems. The traditional way to perform it on image processing and computer vision usually

follows a top-down approach: an engineer evaluates the requirements, creates an algorithm and imbues it with its intelligence. Bio-inspired computer vision, on the other hand, takes a more bottom-up, decentralised approach. Most of the bioinspired systems are starting from the simulations of simple organisms by computing a set of simple rules. Those simple rules collaborate with each other and produce information for a higher level application. As a result, some forms of complex behaviour arise [1]. In this study, we aim to create a bio-inspired computer visual system which simulates some mechanisms of human visual system in a biological way. We simulate neural activity of the retina and of the primary visual cortex, in order to mimic some visual functions such as encoding the edges, textures, and shapes of natural scenes. The system extracts visual features from an input image by using mechanisms similar to human visual system. Those visual features are converted to artificial neural activity signals which are then classified by a machine learning algorithm. As a result, the contents of the input image, which might be related to particular knowledge concepts, can be identified. By using the bio-inspired computer visual system, it is possible to make connections between image features and visual knowledge concepts. However, a visual knowledge concept (for example: the invasive area on a digitized biospy) is rather complex and usually it is not possible to represented it by a single image feature. Thus, in this paper, we propose a new computer language, named Visual Pattern Assessment Language (ViPAL), in order to provide to the users a highlevel user interface for the application. The application we propose is a computer-aided invasive area detection system for breast cancer prognosis. Breast cancer is the second most common type of cancer and the fifth most common cause of death from cancer. Early signs of breast cancer are often hidden within the breast tissues. Thus, breast cancer prognosis relies highly on the medical human

doctors [2]. During breast cancer prognosis, the pathologists need to identify the tumours and to analyse their histological appearance. Usually they first isolate the most interesting regions, the so-called Regions of Interest (ROI), at lower magnification. Then, they analyse the ROI at higher magnification and make their prognosis. A well-trained pathologist usually can complete the analysis of a biopsy in 10 to 20 seconds. However a doctor in a big medical institute sometimes need to evaluate more than one hundred biopsies per day. Thus, to support their works for breast cancer prognosis becomes an important issue. Although many medical modalities are used to support the prognosis of breast cancer, histopathology is the golden standard to make the final decision for the prognosis and the following treatments. Thus, some researchers proposed their studies based on histopathology [3], [4], [5]. However, to the authors’ knowledge, not many works have been published about computer-aided breast cancer prognosis on histopathology. This is probably because of the difficulty of the problem itself. One of the key issues is that, since the image size of the digitized breast biopsies is quite huge, the implementations in those researches usually requires a lot of computation time [6], often several hours, sometimes even several days. Some researchers even suggested to use specific integrated circuits to implement the algorithms [6]. In this paper, we introduce an invasive area detection algorithm based on bio-inspired visual simulation. A part of the human visual system is simulated in order to approximately mimic the behaviours of the visual system of the pathologists. Thus, the number of parallelizable components can be maximized. Consequently, the Graphic Processing Unit (GPU) can be used. In this study, NVIDIA’s GPU and its Compute Unified Device Architecture (CUDA) were used to accelerate the algorithm. We have applied our method on 15 breast biopsies. All these biopsies have been annotated locally and globally by pathologists of the Department of Pathology of the National University Hospital (NUH), Singapore.

(a) A ganglion cell collects information from a set of neighboring photoreceptors.

(b) The RF of a center-green- (c) The RF of a center-yellowon/surround-red-off ganglion. on/surround-blue-off ganglion. Fig. 1.

The receptive field of a ganglion cell.

II. M ETHODS The proposed model is based on the human visual system. Human visual system includes two major parts: the retina and the primary visual cortex. In the retina, there are two major photoreceptors: the cone- and the rod-cells. The cone-cells can be classified into three types: L-, M-, and S-cells. These kinds of photoreceptors react to various light wavelengths. A set of opponental photoreceptors form a Receptive Field (RF) (see Fig. 1). These photoreceptors form a field which is called ganglion RF since they collect visual information and send neural spikes to a ganglion cell. The ganglion RF can be considered as the fundamental element of the human visual system. Eventually, the ganglion cells produce various stimulations which are sent to the primary visual cortex [7]. This biological mechanism results in the color space of human visual sensory. Hering’s color opponent theory [8] is a popular way to describe the color space [7]. This theory

Fig. 2. The flow of breast cancer invasive area detection. The pathologist selects training patterns via the observation of the breast biopsies, and sends the patterns to the system. The system converts the image into artificial neural activity signals. The first and the second order visual features are then extracted. These visual features are used either to train the system during the training phase, or to detect the invasive areas in the testing phase.

interprets the results of color mixing by proposing the existence of three opponent processes in the eye, the brain, or both. These three processes are the red-green, the yellow-blue, and the black-white sensations. These color stimulations are transferred to the primary visual cortex [7]. In the primary visual cortex, there are two major kinds of cells: the simple- and the complex-cells [7]. Generally speaking, these cells produce two kinds of visual features: the first- and the second-order features [9], [7]. The first-

order feature contains the information of intensities of various color channels, and the second-order feature includes spatial variance of visual signal [10], [11]. This knowledge of human visual system can help us to discover the strategies of decision-making in cognitive process [12] when the pathologists are analysing a breast biopsy for the prognosis of breast cancer. In this study, we simulate some mechanisms of human visual system, generate the first- and second- order features, and classify these features by Support Vector Machine (SVM). After the SVM has been trained, the system is capable to detect the ROIs. In order to accelerate the computing, the GPU technology is used in some major components of the system. Consequently, the result of each input image can be obtained in a reasonable period of time. In the experiments, the training patterns were given by the researchers under the supervision of the pathologist. Those patterns came from a set of golden samples which have been identified by the pathologist. These golden samples contain typical features of invasive area. During the learning phase, the system was sometimes unable to identify some positive cases. After we discussed with the pathologist, some of these cases were considered as the training patterns and the system was retrained in order to improve its performance. The flow of breast cancer invasive areas detection is shown on Fig. 2. A. Color Representation In this study, we aim to discover the relationship between a breast biopsy image displayed on a screen, and the decisionmaking procedure of a pathologist who is analysing the image on the screen. In order to simulate the related visual reactions of human visual system, we have to obtain few parameters. First, the sensitivities of the photoreceptors in visible spectrum. A wellknown cone-cell sensitivity factor in visible spectrum is presented in Fig. 3(a) [13], [14]. Second, the radiance factor of the screen in the visible spectrum is required in order to evaluate the energy which might be received by the human visual system. Fig. 3(b) shows a radiance factor in the visible spectrum of a typical CRT screen. Most of the modern screens provide Gamma correction. An example of Gamma Display Coefficient (GDC) functions is shown in Fig. 3(c). The GDC functions of the screen are also required. These coefficients can be used to evaluate the radiance which is emitting from the screen. Based on what we introduced above, a transform operator can be obtained:     l Γred (r)  m  = s · T ·  Γgreen (g)  , (1) s Γblue (b) where r, g, b are the colors of the pixel, Γred (·), Γgreen (·), and Γblue (·), are GDC functions used to reconstruct the Gamma correction, s is a necessary scale in order to normalize the input values, and T is a 3 × 3 matrix which is the linear combination of the cone-cell sensitivity factor and the radiance

Photoreceptor

z(α)

hR

Bipolar

zR (α)

Ganglion

y(α)

Σ g −b

Horizontal

hH

zH (α)

Fig. 4. The computer fovea model, proposed by Huang et al. [15], including the photoreceptor cells, the horizontal cells, the bipolar cells and the ganglion cells.

factor in visible spectrum. Further details can be found in [13], [14], [15]. In order to represent a color image from a psychological point of view, we propose the use of Ewald Hering’s opponent color theory [8]: ured = log(l), ugreen = log(m), ublue = log(s), uyellow = log( 21 (l + m)).

(2)

Note that logarithm is used to approximate the human visual system. Here (ured , ugreen ) and (ublue , uyellow ) are used to describe the Hering’s opponent color, Red-Green (RG), Blue-Yellow (BY), and Luminance (L): vRG = ured − ugreen , vBY = ublue − uyellow , vL = 23 (ured + ugreen + ublue ) − 1.

(3)

Those color pairs and the luminance information can be used to describe the visual signal in the human visual system [7], [16], [15]. Based on the color information, the human visual system is able to extract features from the image. In our method, these feature extraction algorithms include intensity, colors and texture perception. B. Computer Fovea Model In order to model the retina, various models are suggested for various purposes [17], [18], [19]. Huang et al. proposed a computer fovea model with various applications [15]. The computer fovea model aims to model a simplified version of the full retina system. First, a general assumption of center/surround receptive field (RF) of ganglion can be considered as a reference. Some physiological experiments indicated that the RF of the ganglion exhibits a center/surround characteristic. Furthermore, various publications stated that the RF of ganglions can be modeled as follows [20], [7], [21], [17], [18], [19]: hG (δ(α)) , ▽2 (GσG (α)),

(4)

−3

1

3.5

x 10

1

0.9

0.9 3 0.8

0.7 0.6 0.5 0.4 0.3

2.5

0.7 gamma rate

radiance (w/sr*m2/nm)

response (normalized)

0.8

2 1.5

0.6 0.5 0.4 0.3

1

0.2

0.2 0.5

0.1 0 350

0.1 400

450

500 550 600 wavelength (nm)

650

700

750

0 350

400

450

500 550 600 wavelength (nm)

650

700

750

0 0

50

100

150

200

250

intensity

(a) Cone cell sensitivities in various wavelengths (b) The radiance spectrum of a standard CRT (c) Gamma display coefficient of a standard CRT of light. monitor. monitor. Fig. 3.

The Related coefficients.

where ▽2 (·) denotes a Laplace filter and GσG (·) is a Gaussian filter with standard deviation σG . 1) Photoreceptor: Photoreceptor includes various cone and rod cells. There cells react to different wavelength of visible light. Cone cells can be roughly classified into three types: short wavelength (S/blue cell), middle wavelength (M/green cell), and long wavelength (L/red cell). Generally speaking, a bipolar cell collects the spiking from a set of cone cells and form a diffuse pathway. The use of a Gaussian function to model the diffusion is suggested in various publications [17], [19]. Thus, (5) hR (δ(α)) , GσR (α). where GσR (·) represents a Gaussian filter. As mentioned in [17], σR represents the standard deviation with a range from 1.5 to 12 (cell space). 2) Bipolar Cell: Bipolar cells collect the signal from a number of cone cells and transmit the spiking to ganglion cells. Although there are various types of bipolar cells, in this model the simplest one is chosen. That is, the bipolar cell maps to one cone cell and its opponent channel of cone cell with a bias −b. 3) Horizontal Cell: Horizontal cells have been considered as a set of cells to contribute to the surround response of bipolar cells. The horizontal cells have been shown to be color opponent in response. 1 1 hH (α) , (δ(α) − hG (α) ⊗ h−1 R (α)). b g

(6)

4) Ganglion: In most cases, a ganglion cell collects signal from only one bipolar cell. Thus, in this model, only a bias g is used to represent the function of a ganglion cell. C. First Order Feature 1) Receptive Field: In human visual system, the Receptive Field (RF) of ganglion is considered as the fundamental element [7]. In the first order extraction, the RF is defined as a set of pixels which are sampled on a particular area. Generally speaking, we need to compute all of the pixels in this area. However, in order to reduce the computational costs, we suggest to pick up only some pixels in this area according

to a sampling distribution. The sampling distribution ps (l) is based on Gaussian distribution such that ps (l) , N (µs , Σs ),

(7)

where l = (l1 . . . li . . . lm ), li ∈ R2 is a set of the locations of the data points, µs = [µ1 , µ2 ]T is the mean, and Σs is the covariance matrix:   σ11 σ12 . (8) Σs = σ21 σ22 In the experiments, σ11 = σ22 = 1000 pixels and σ12 = σ21 = 0 (see Fig. 5). 2) First Order Feature Extraction: When a RF is chosen, a set of pixel locations l = (l1 . . . li . . . lm ), li ∈ R2 is generated. We obtain the so-called Hering’s opponent color: Red-Green (RG), Blue-Yellow (BY), and the Luminance (L) information as follows: vi = [vRG (li ), vBY (li ), vL (li )], i = 1 . . . m.

(9)

According to Geusebroek, et al., the opponent color theory can be applied to computer vision and implemented as the Gaussian color model [16], [22], [23], [24]. Thus, all vi can be used to generated a multivariate Gaussian distribution: p1 (v) , N (µ1 , Σ1 ).

(10)

The mean µ1 and the covariance matrix Σ1 can be obtained by Expectation Conditional Maximization (ECM) [25]. In our study, µ1 and Σ1 are defined as the first order features. D. Second Order Feature Second-order features cannot be detected by mechanisms sensitive only to intensity changes. It is ubiquitous in visual scenes, but the neuronal mechanisms mediating perception of such stimuli are not well understood [26], [27]. Most of the researchers agree that the second-order feature includes spatial variance of visual signal [10], [11]. In order to extract the basis, various methods were proposed. Generally speaking, most of them invoked one or several basis analysing algorithms. Those algorithms include Gabor filtering, Principal Component Analysis (PCA), Independent Component Analysis (ICA), Sparse Coding, etc. Each of them has various

TABLE I A N E XAMPLE OF V I PAL. PROGRAM example BEGIN LOAD IMAGE FROM ’biospy.bmp’ INTO input_image; LOAD FEATURE ’invasive.feature’ INTO feature; SELECT FEATURE FROM input_image INTO result; SHOW IMAGE FROM result; END.

where sj,i ∈ si . The basis set can be over-complete (n > k) if the maximum k exists. Basis vectors b1 . . . bn and coefficients s1 . . . sn are obtained using a training set x1 . . . xm . Thus, an optimization problem for b and s corresponds to the minimization of: m X n n m X X X 1 2 φ(sj,i ), b s k + β kx − j j,i i 2σ 2 i=1 j=1 j=1 i=1

subject to Fig. 5. An example of breast biopsy (histopathology) image from our virtual microscopic platform. The resolution of the image is 8.3µm/pixel at 1.2X (12 times) magnification. One point is chosen by a user click and the rest of the 50 points are calculated based on Multivariate Gaussian Distribution. The covariances are Σ11 = Σ22 = 1000 pixels and Σ12 = Σ21 = 0.

advantages and disadvantages. These algorithms are compared in various publications: Olsausen et al. introduced the Sparse Coding for natural images in [28]. In their article, the PCA was compared (see Fig. 1 in [28]). They concluded that the principal components do not arise as a result of the oriented structures in whitened natural images. Karklin compared the costs and performances in [29], and Willmore et al. compared various methods including Sparse Coding, Gabor filtering, PCA, and ICA in [30]. In this work, we suggest a model based on Sparse Coding, similar to Karklin’s model: first, the basis vectors of various channels are extracted by Sparse Coding. Next, the related coefficients of these basis vectors for various images can be obtained. Third, the Gaussian distribution of these coefficients can be generated. Finally, the parameters µ and σ are obtained and are considered as the second order features. The advantages of the proposed model are: first, this model has the same color space as the first order feature extraction. As a result, the time consumption of the algorithm can be reduced. Second, Sparse Coding can be implemented in GPU. As a result, the power of GPU can be brought to this work. Third, the Sparse Coding can be over-complete, hence, the number of basis vectors is not limited. Sparse Coding suggested that a set of natural scenes defined by x1 . . . xm ∈ Rk , can be described in terms of a linear combination of basis vectors b1 . . . bn ∈ Rk and coefficients s1 . . . sm ∈ Rn such that [28] X bj sj,i , i = 1 . . . m. (11) xi ≈ j

kbj k2 ≤ c, j = 1 . . . n,

(12)

where σ 2 is the variance of the noise, φ(·) is a sparsity function, and β is a constant [28], [31]. There are many solutions able to solve this typical optimization problem. In this study, we consider GPUs to implement projective gradient descent [32]. As a result, b and s computation time has consistently been reduced. 1) Receptive Field: The receptive field of the second order feature extraction is more complex than the first order feature extraction. First, following (7), a set of data point locations l1 . . . lm ∈ R2 is generated. Let li be the center of an image patch, from which a set of patches x1 . . . xm can be obtained such that: xi = I(li − d), d = [n1 , n2 ]T , (13) where I represents a channel which is obtained from the input image, −N ≤ n1 < N , and −N ≤ n2 < N , where N is the size of the patches. In the experiments, N = 10. Since all of the patches x1 . . . xm are captured from the same region, they are sharing the same features. Thus, the texture can be encoded by coefficients si,j in (12) with a set of basis vectors bj , which need to be obtained previously. All of si,j of RG, BY, and L channels are obtained, such as: si,j = [sRG,j (li ), sBY,j (li ), sL,j (li )], i = 1 . . . m and j = 1 . . . n.

(14)

All si,j generate a multivariate Gaussian probability distribution: p2 , N (µ2 , Σ2 ). (15) Like the first order feature, the mean µ2 and the covariance matrix Σ2 are obtained by Expectation Conditional Maximization (ECM) [25]. In our study, µ2 and Σ2 are defined as the second order features.

False Negativ e True Positive (Sensibility ) ) False Positiv e

ROC curve r ROC curve 1

PARTEST GRAPH

0.8 0.8 Parameters proportion

True positive rate (Sensitivity )

1

0.6

0.4

0.2

0

0.6

0.4

0.2

0

0.2

0.4

0.6

0.8 )

1

0

0

0.2

0.4 0.6 Subjects proportion

0.8

1

(a) The Receiver Operating Charac- (b) The Clinical Test Perforteristic (ROC) curve. mance. Fig. 8.

Performance of our system. The AUC is 0.9186.

III. E XPERIMENTAL R ESULTS AND VALIDATION

Fig. 6. 400 basis vectors on 20×20 pixels image patches, extracted from digitized breast biopsies (virtual microscope).

TABLE II T HE PERFORMANCE OF GPU ACCELERATION . Sparse Coding SVM EM1

CPU 1406.759 92.265 26.314

GPU 56.297 5.427 2.58

Rate 24.988 17.001 10.199

E. Visual Pattern Assessment Language (ViPAL) The Visual Pattern Assessment Language (ViPAL) is a new computer language which provides the users to access and retrieve the image features by using a script. The design of the ViPAL is based on the well-known database script language: SQL. The ViPAL provides an environment where the users can retrieve images in the same way as they would access to the tables of a database. As a result, this language serves as the interface for the system operation. In order to reduce the learning curve of the language, the design of ViPAL was based on the famous database manipulating script named SQL. An example is shown in Table I. It can be understood easily. In the example, first, an image is loaded into the memory and represented by a variable named input image, and a feature file is loaded into feature. Then, the SELECT command performs the feature selection. Finally the result is shown by the SHOW IMAGE command. The ViPAL provides the functions of file operation, image preview, routine operation, etc. As a result, the complexity of the system is covered by a shell. So the workload on the users can be reduced.

In the testing phase, more than 25 images are obtained from 15 digitized breast biopsy slides. Two of the images are used as the training patterns. The global Receiver Operating Characteristic (ROC) curve and the Clinical Test Performance are shown in Fig. 8. The global Area Under the Curve (AUC) is 0.9186. The optimal cut point is −1.1348. We also computed the difference of ROIs between the results of our ROI detection and the ground truth by: R∩G , (16) R∪G where R is our result, G is the ground truth provided by the experts, and c ∈ (0, 1) is the covering rate. The average of all of the testing results is 0.7077. Some of the results are presented in Table III. Our results also present an interesting phenomenon. In the design of our system, the results are determined by the sign of the hyper plane H(·) which has been produced by SVM classification. However, the test performance suggested that the optimal cut point is −1.1348. This is because the pathologists were asked to give three regions for each image, the invasive area, the normal area, and unknown area. However, some suspect regions which are difficult to determinate, tend to be considered as invasive areas by the pathologists. This results as a higher false-positive rate. In other words, the results reflect the exact behaviors of the pathologists who provided the evaluation patterns. The performance of GPU acceleration is presented in Table II. The GPU is a GeForce 9400M from NVIDIA. The computer is an Apple Macbook with Intel CPU Core 2 Duo with 4G memory. With a pre-trained kernel, the computation time of low resolution image ROI detection for one slide is about 120 seconds. It depends on the size of the slide. c=

IV. C ONCLUSION In this paper, a bio-inspired computer visual system using GPU is presented. It is implemented for breast cancer prognosis. Based on a predefined image database, the invasive

(a) Ground truth provided by pathologists on the input image. Usually the (b) Result of feature extraction and classification methods on a set of equally ground truth contains three regions: invasive, normal, and other areas. distributed testing points. The results are shown as circles. The red-circles indicate positive areas, and the blue-circles are negative areas. Note that the size of the circles is related to the hyper-plane H(·).

(c) Smoothing the results from Fig. 7(b). Fig. 7.

(d) The region of interest is obtained by applying a threshold 0.

An example of construction of region of interest.

cancerous areas can be identified on the digitized breast biopsies. This study came out with an interesting topic, the inter/intra observation source variability. For the same biopsy, different doctors can make different prognosis. Sometimes even the same doctor makes a different prognosis for the same biopsy. Based on this study, we may be able to evaluate the variability between the prognosis. The proposed algorithm can even be applied to different areas (for example: ROI detection on X-ray images). Those will be in our future works. The ViPAL language provides a shell between the bioinspired computer visual system and the users. The complexity of the low-level image feature assessment can be managed by the shell. As a result, the workload of the users can be reduced. ACKNOWLEDGMENT This research is supported by A*STAR SERC 052 101 0103 (NUS R-252-000-319-305). The authors thank Dr. Yan Karklin from New York University for this support on this work. R EFERENCES [1] M. Riesenhuber and T. Poggio, “Neural mechanisms of object recognition,” Current Opinion in Neurobiology, vol. 12, pp. 162–168, 2002. [2] M. E. Lippman, “Breast cancer,” in Harrison’s Principles of Internal Medicine. McGraw-Hill Professional, 2004.

[3] J.-R. Dalle, W. K. Leow, D. Racoceanu, A. E. Tutac, and T. C. Putti, “Automatic breast cancer grading of histopathological images,” in 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2008. EMBS 2008., Vancouver, BC,, 2008, pp. 3052–3055. [4] J. Z. Zhang, S. Petushi, W. C. Regli, F. U. Garcia, and D. E. Breen, “A study of shape distributions for estimating histologic grade,” in 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2008. EMBS 2008., 2008, pp. 1200–1205. [5] S. Maskery, Y. Zhang, R. Jordan, H. Hu, J. Hooke, C. Shriver, and M. Liebman, “Co-occurrence analysis for discovery of novel breast cancer pathology patterns,” IEEE Transactions on Information Technology in Biomedicine, vol. 10, no. 3, pp. 497–503, 2006. [6] M. Tahir and A. Bouridane, “A fpga based coprocessor for cancer classification using nearest neighbour classifier,” in IEEE International Conference on Acoustics, Speech and Signal Processing, vol. 3, 2006. [7] D. Hubel, “Eye, brain, and vision,” Scienceific American, 1988. [8] E. Hering, “Outlines of a theory of the light sense,” Harvard University, Cambridge, Mass, Tech. Rep., 1964. [9] G. Rhodes, “Looking at faces: First-order and second-order features as determinants of facial appearance,” Perception, vol. 17, no. 1, pp. 43–63, 1988. [10] I. A. Shevelev, “Second-order features extraction in the cat visual cortex: Selective and invariant sensitivity of neurons to the shape and orientation of crosses and corners,” Biosystems, vol. 48, no. 1, pp. 195–204, 1998. [11] C. C. Pack, B. R. Conway, R. T. Born, and M. S. Livingstone, “Spatiotemporal structure of nonlinear subunits in macaque visual cortex,” The Journal of Neuroscience, vol. 26, no. 3, pp. 893–907, January 2006. [12] J. A. Bednar, “Understanding neural maps with Topographica,” Brains, Minds, and Media, vol. 3, p. bmm1402, 2008. [Online]. Available: http://www.brains-minds-media.org/archive/1402 [13] “An analytical model for describing the influence of lighting parameters upon visual performance – vol. 1,” Technical Fundations, CIE 19/2.1, Technical Committee 3.1 1, Tech. Rep., 1981.

TABLE III S OME ROI S WHICH HAVE BEEN DETECTED BY OUR SYSTEM . T HE FULL IMAGE SET CAN BE OBTAINED FROM IPAL WEB SITE . Source

Result

ROI

ROC/AUC ROC curve

True positive rate (Sensitivity)

1

0.8

0.6

0.4

0.2

0

0

0.2 0.4 0.6 0.8 False positive rate (1−Specificity)

1

0.87476 ROC curve

True positive rate (Sensitivity)

1

0.8

0.6

0.4

0.2

0

0

0.2 0.4 0.6 0.8 False positive rate (1−Specificity)

1

0.87203 ROC curve

True positive rate (Sensitivity)

1

0.8

0.6

0.4

0.2

0

0

0.2 0.4 0.6 0.8 False positive rate (1−Specificity)

1

0.96631 ROC curve

True positive rate (Sensitivity)

1

0.8

0.6

0.4

0.2

0

0

0.2 0.4 0.6 0.8 False positive rate (1−Specificity)

1

0.89279

[14] X. Zhang, “S-CIELAB: A spatial extension to the CIE l*a*b* DeltaE color difference metric,” Tech. Rep., Jan 10 2010. [Online]. Available: http://white.stanford.edu/ brian/scielab/ [15] C. H. Huang and C. T. Lin, “Bio-inspired computer fovea model based on hexagonal-type cellular neural network,” IEEE Transactions on Circuits and Systems - I: Regular Papers, vol. 54, no. 1, pp. 35–47, 2007. [16] J.-M. Geusebroek, R. van den Boomgaard, A. Smeulders, and H. Geerts, “Color invariance,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 12, pp. 1338–1350, 2001. [17] S. Shah and M. D. Levine, “Visual information processing in primate cone pathways – part i: a model,” IEEE Transactions on Systems, Man, and Cybernetics - Part B: Cybernetics, vol. 26, no. 2, pp. 259–274, 1996. [18] ——, “Visual information processing in primate cone pathways – part ii: experiments,” IEEE Transactions on Systems, Man, and Cybernetics - Part B: Cybernetics, vol. 26, no. 2, pp. 275–289, 1996. [19] J. Thiem and G. Hartmann, “Biology-inspired design of digital Gabor filters upon a hexagonal sampling scheme,” pp. 445–448, 2000. [20] D. Hubel and T. N. Wiesel, “Receptive fields, binocular interaction,and functional architecture in the cat’s visual cortex,” J. Physiol. (Lond.), vol. 160, pp. 106–154, 1962. [21] S. Shah and M. D. Levine, “The primate retina: a biological summary,” Ph.D. dissertation, 1992.

[22] J.-M. Geusebroek, R. v. d. Boomgaard, A. W. M. Smeulders, and T. Gevers, “Color constancy from physical principles,” Pattern Recognition Letters: special issue on colour image processing and analysis, vol. 24, no. 11, pp. 1653–1662, 2003. [23] M. A. Hoang, J.-M. Geusebroek, and A. W. M. Smeulders, “Color texture measurement and segmentation,” vol. 85, no. 2, 2005, pp. 265– 275. [24] H. M¨uller, N. Michoux, D. Bandon, and A. Geissbuhler, “A review of content-based image retrieval systems in medical applications – clinical benefits and future directions,” International Journal of Medical Informatics, vol. 73, no. 1, pp. 1–23, 2004. [25] W. Press, S. Teukolsky, W. Vetterling, and B. Flannery, Numerical Recipes in C, 2nd ed. Cambridge, UK: Cambridge University Press, 1992. [26] J. Larsson, M. S. Landy, and D. J. Heeger, “Orientation-selective adaptation to first- and second-order patterns in human visual cortex,” Journal of Neurophysiology, vol. 95, pp. 862–881, 2006. ¨ [27] H.-P. F. K?nig and Peter, “The role of first- and second-order stimulus features for human overt attention,” Perception & Psychophysics, vol. 69, no. 2, pp. 153–161, 2007. [28] B. A. Olshausen and D. J. Field, “Emergence of simple-cell receptive field properties by learning a sparse code for nature images,” Nature, vol. 381, no. 13, pp. 607–609, 1996. [29] Y. Karklin, “Hierarchical statistical models of computation in the visual

cortex,” 2007. [30] B. Willmore and D. J. Tolhurst, “Characterizing the sparseness of neural codes,” Network: Comput. Neural Syst. 12 (2001) 255V270, vol. 12, pp. 255–270, 2001. [31] H. Lee, A. Battle, R. Raina, and A. Y. Ng, “Efficient sparse coding algorithms,” in Natural Information Processing Systems Conference, 2007, pp. 801–808. [32] A. Madhavan, “Using GPUs to speedup sparse coding algorithms applied to self-taught learning problems,” Stanford University, Tech. Rep., 2008.