A computer diagnostic system and related method are disclosed for automatically classifying tissue types in an original tissue image captured by an imaging device based on texture analysis. In one embodiment, the system receives and divides the tissue image into multiple smaller tissue block images. A combination of local binary pattern (LBP), average LBP (ALBP), and block-based LBP (BLBP) feature extractions are performed on each tissue block. The extractions generate a set of LBP, ALBP, and BLBP features for each block which are used to classify its tissue type. The classification results are visually displayed in a digitally enhanced map of the original tissue image. In one embodiment, a tissue type of interest is displayed in the original tissue image. In another or the same embodiment, the map displays each of the different tissue types present in the original tissue image.
|
17. A computer aided diagnostic method for automatically classifying tissue types in a digital tissue image, the method implemented by a processor executing program instructions and comprising steps of:
receiving an original tissue image captured by an optical imaging device, wherein the imaging device captures the original tissue image by optical coherence microscopy, optical coherence tomography, confocal microscopy, or two photon microscopy;
dividing the original tissue image into a plurality of smaller tissue blocks which collectively represent the captured original tissue image, each tissue block having a texture;
performing on each tissue block a combination of a plurality of texture analysis methods comprising local binary pattern extraction that generates a local binary pattern feature, average local binary pattern extraction that generates an average local binary pattern feature, and block based local binary pattern extraction that compares the average intensity value of pixels in blocks of certain shape in a neighborhood around the center pixel to generate block based local binary pattern features;
classifying a type of tissue found in each tissue block based on its texture; and
displaying a digitally enhanced map of the original tissue image in which at least one tissue type is shown in a visually discernible manner from other portions of the original image.
11. A computer-aided diagnostic system for analyzing tissue image data, comprising:
a non-transitory computer readable medium having software program instructions stored thereon;
a computer processor communicating with the computer readable medium, the processor when configured with and executing the program instructions being operable to:
receive an original tissue image captured by an optical imaging device, wherein the imaging device captures the original tissue image by optical coherence microscopy, optical coherence tomography, confocal microscopy, or two photon microscopy;
divide the original tissue image into a plurality of smaller tissue blocks which collectively represent the captured original tissue image, each tissue block having a texture;
perform on each tissue block a combination of a plurality of texture analysis methods comprising local binary pattern extraction that generates a local binary pattern feature, average local binary pattern extraction that generates an average local binary pattern feature, and block based local binary pattern extraction that compares the average intensity value of pixels in blocks of certain shape in a neighborhood around the center pixel to generate block based local binary pattern features;
classify a type of tissue based on the texture features extracted from each tissue block according to pre-defined types of contextually relevant tissues found in the original tissue image; and
display a digitally enhanced image of the original tissue image which shows at least one type of tissue in a visually discernible manner.
1. A computer-aided diagnostic system for analyzing tissue image data, comprising:
a non-transitory computer readable medium having software program instructions stored thereon;
a computer processor communicating with the computer readable medium, the processor when configured with and executing the program instructions being operable to:
receive an original tissue image captured by an optical imaging device, wherein the imaging device captures the original tissue image by optical coherence microscopy, optical coherence tomography, confocal microscopy, or two photon microscopy;
divide the original tissue image into a plurality of smaller tissue blocks which collectively represent the captured original tissue image, each tissue block having a texture;
analyze each tissue block based on its texture to classify a type of tissue found in each tissue block; and
generate a digitally enhanced map overlaid on the original tissue image displaying a predetermined tissue type of interest in a visually discernible highlighted manner;
wherein the processor is further operable to perform on each tissue block a combination of a plurality of texture analysis methods; and
wherein the texture analysis methods comprise local binary pattern extraction that generates a local binary pattern feature, average local binary pattern extraction that generates an average local binary pattern feature, and block based local binary pattern extraction that compares the average intensity value of pixels in blocks of certain shape in a neighborhood around the center pixel to generate block based local binary pattern features.
2. The system according to
3. The system according to
4. The system according to
5. The system according to
6. The system according to
7. The system according to
8. The system according to
9. The system according to
12. The system according to
13. The system according to
14. The system according to
15. The system according to
16. The system according to
18. The method according to
integrating the local binary pattern feature, average local binary pattern feature, and block based local binary pattern features for each of the tissue blocks; and
generating a multi-class integrated feature for each of the tissue blocks comprising its respective local binary pattern feature and average local binary pattern feature.
19. The method according to
20. The method according to
21. The method according to
22. The method according to
|
The present application is a continuation-in-part of PCT/US2014/064218 filed Nov. 6, 2014, which claims the benefit of U.S. Provisional Patent Application Ser. No. 61/900,828 filed Nov. 6, 2013, the entireties of which are incorporated herein by reference.
This invention was made with government support under National Institute of Health/National Institute of Biomedical Imaging and Bioengineering (NIH/NIBIB) Pathway to Independence Award (R00-EB010071) and under National Institutes of Health/National Library of Medicine (NLM) award (NLM-HHSN276201000693P). The government has certain rights in the invention.
The present invention generally relates to biological tissue imaging, and more particularly to a computer-based system and method for analyzing the texture of tissue samples and classifying the tissue to detect abnormalities which provides a medical diagnostic tool.
Breast cancer is a high incidence cancer in women worldwide. The survival rate of breast cancer improves with screening and early detection. Biopsy is a frequently used medical test in which tissue samples are removed from a human subject and then examined by a pathologist under a microscope to determine the presence or extent of a disease. Traditionally, the tissue is processed to extremely thin slices and stained before being observed under a microscope. Optical coherence tomography (OCT) provides an alternative non-invasive optical imaging modality that can provide 3D, high-resolution images of biopsy sample tissues without staining. Optical coherence microscopy (OCM) combines the advantages of OCT and confocal microscopy using high numerical aperture objectives to provide cellular resolution images. OCT, OCM and other images captured by optical imaging devices are label free, in the sense that they do not use staining of samples to indicate different tissue areas. In contrast, traditional histological images rely on staining of samples with at least one biomarker. Thus, in order to classify different tissue types in OCT, OCM and other optical images, advanced computer algorithms are required to analyze the appearance differences in different tissue areas.
To improve accuracy and efficiency, numerous computer aided diagnostic methods implementing software program algorithms for automated classification of tissue areas in medical images have been proposed. The same level of accuracy offered by such automated techniques and speed cannot reasonably be duplicated by manual analysis of tissue images alone. Experiments show that a spectral texture analysis technique is good at extracting distinctive texture features from OCT images, even if the images have few structure features. Image processing and data mining techniques are applied on a large number of medical images to distinguish different patterns of tissues.
Texture analysis is a commonly used technique for medical image classification. There are mainly two different kinds of texture analysis techniques: structure based methods and statistical methods, including spectrum analysis and feature distributions.
In the structure based methods, typical texture patterns of an image are extracted and clustered into several groups. The frequency of the occurrence of each pattern is used to represent the texture. Structure based methods are good at classification of texture images that are highly repetitive, since the clustering step reserves more information from the texture image if a certain kind of texture pattern appears in the image frequently. But for OCT/OCM images of human tissue, very few texture patterns are exactly the same, so that structure based methods do not perform well in OCT/OCM image based tissue classification.
In statistical methods, required features are extracted and the frequency of the occurrence of each feature in the whole texture image is used to represent the texture. Different texture features are used in statistical methods, including grey-level co-occurrence matrix, center-symmetric auto-correlation, gray level difference, and local binary patterns (LBP). The statistical distribution of different features are calculated and used as feature vectors for texture representation. Since the statistical methods do not require texture images to be highly repetitive, they are suitable to be used for the OCT/OCM image based tissue classification.
LBP is a popular statistical texture analysis technique pursued in recent years. In LBP features, the frequency of different local image intensity feature patterns is calculated and used as feature vectors. LBP features have been applied in many applications, such as texture analysis, face recognition, and description of regions of interest. LBP is rotation invariant and uses a fixed set of patterns to represent the texture. It has high classification accuracy on different texture image datasets. Although LBP is relatively effective in representing and classifying textures in OCT/OCM images, further improvement in tissue classification accuracy however are desired particularly for medical diagnostic purposes.
A computer aided diagnostic system is disclosed that provides tissue classification and abnormality detection techniques based on texture analysis of a tissue specimen or sample imaged using an imaging device. A novel image analysis technique is applied on tissue images to achieve automatic classification and display of tissue types found in the sample. The image analysis technique uses a new variant of local binary pattern (LBP) features, average local binary pattern (ALBP) features and block based local binary pattern (BLBP) features for use in texture analysis of tissue images. The ALBP AND BLBP features are integrated with the LBP features to achieve more accurate image and tissue classification. In various embodiments, the imaging device may be without limitation an optical coherence microscopy (OCM) system, optical coherence tomography (OCT) system, confocal microscopy system, or two photon microscopy system. The present automated diagnostic system may be used for identifying diseases manifested by visually observable tissue abnormalities such as cancer detection, and particularly in one non-limiting example for breast cancer detection. Other applications are possible.
The present tissue classification process generally includes two primary steps: training and testing. In the training process, a series of LBP features which represent image textures are extracted from digital tissue images. Similarly, a series of ALBP features which represent image textures are also extracted from digital tissue images. In LBP texture analysis, the grayscale values of a certain number of neighbor pixels are compared with the grayscale value of a center pixel to generate a binary code pattern. By contrast, ALBP texture analysis compares the grayscale value of a neighbor with the average grayscale value of all neighbors, instead of comparing it to the grayscale value of the center pixel. The results of the LBP and ABLP are combined and integrated in the present system. By using integrated ABLP and LBP features instead of LBP features alone, the inventors have discovered that classification accuracy can be improved.
According to further aspects of the invention, the diagnostic system can generate a tumor probability heat map based on the original captured tissue image in which highly suspect areas are digitally enhanced on the original grayscale tissue image, thereby advantageously providing rapid diagnostic capabilities at a glance. The system can also further generate an enhanced multi-class tissue classification map based on the LBP and ALBP integrated features which identifies all different tissue types found in the sample which is helpful for diagnostic purposes. In one embodiment, the heat map and tissue classification map may be colored and/or patterned.
In one aspect, a computer-aided diagnostic system for analyzing tissue image data includes a non-transitory computer readable medium having software program instructions stored thereon, a computer processor communicating with the computer readable medium, the processor when configured with and executing the program instructions being operable to: receive an original tissue image captured by a label free optical imaging device; divide the original tissue image into a plurality of smaller tissue blocks which collectively represent the captured original tissue image, each tissue block having a texture; analyze each tissue block based on its texture to classify a type of tissue found in each tissue block; and generate a digitally enhanced map overlaid on the original tissue image displaying a predetermined tissue type of interest in a visually discernible highlighted manner.
In another aspect, a computer-aided diagnostic system for analyzing tissue image data includes a non-transitory computer readable medium having software program instructions stored thereon, a computer processor communicating with the computer readable medium, the processor when configured with and executing the program instructions being operable to: receive an original tissue image captured by a label free optical imaging device; divide the original tissue image into a plurality of smaller tissue blocks which collectively represent the captured original tissue image, each tissue block having a texture; extract a local binary pattern feature, an average local binary pattern feature, and block based local binary pattern features for each tissue block; classify a type of tissue based on the texture features extracted from each tissue block according to pre-defined types of contextually relevant tissues found in the original tissue image; and display a digitally enhanced image of the original tissue image which shows at least one type of tissue in a visually discernible manner.
In another aspect, a computer aided diagnostic method for automatically classifying tissue types in a digital tissue image is provided. The method is implemented by a processor executing program instructions and comprising steps of: receiving an original tissue image captured by a label free optical imaging device; dividing the original tissue image into a plurality of smaller tissue blocks which collectively represent the captured original tissue image, each tissue block having a texture; classifying a type of tissue found in each tissue block based on its texture; and displaying a digitally enhanced map of the original tissue image in which at least one tissue type is shown in a visually discernible manner from other portions of the original image.
The features of the exemplary embodiments will be described with reference to the following drawings where like elements are labeled similarly, and in which:
All drawings shown herein are schematic and not to scale. Parts given a reference number in one figure may be considered to be the same parts where they appear in other figures without a reference number for brevity unless specifically labeled with a different part number and/or described herein.
The features and benefits of the invention are illustrated and described herein by reference to non-limiting exemplary embodiments. This description of the embodiments is intended to be read in connection with the accompanying drawings, which are to be considered part of the entire written description. Accordingly, the invention expressly should not be limited to such embodiments illustrating some possible non-limiting combination of features that may exist alone or in other combinations of features; the scope of the invention being defined by the claims appended hereto.
In the description of embodiments disclosed herein, any reference to direction or orientation is merely intended for convenience of description and is not intended in any way to limit the scope of the present invention. Relative terms such as “lower,” “upper,” “horizontal,” “vertical,” “above,” “below,” “up,” “down,” “top” and “bottom” as well as derivative thereof (e.g., “horizontally,” “downwardly,” “upwardly,” etc.) should be construed to refer to the orientation as then described or as shown in the drawing under discussion. These relative terms are for convenience of description only and do not require that the apparatus be constructed or operated in a particular orientation. Terms such as “attached,” “affixed,” “connected,” “coupled,” “interconnected,” and similar refer to a relationship wherein structures may be secured or attached to one another either directly or indirectly through intervening structures, as well as both movable or rigid attachments or relationships, unless expressly described otherwise.
Aspects of the present invention may be implemented in software, hardware, firmware, or combinations thereof. The present invention may be embodied in the form of computer-implemented processes and apparatuses for practicing those processes. The computer programs described herein are not limited to any particular embodiment, and may be implemented in an operating system, application program, foreground or background process, driver, or any combination thereof, executing on a single computer or server processor or multiple computer or server processors. Processors described herein may be one or more of any of a computer/server central processing unit (CPU), multi-core processor including two or more independent CPUs, microprocessor, micro-controller, or other computational device or circuit configured for executing computer program instructions (e.g. control logic or code).
The present invention may also be embodied in the form of computer program code embodied in tangible non-transitory computer readable media, such as without limitation random access memories (RAM), read only memories (ROMs), CD-ROMs, DVDs, hard drives, ZIP™ disks, magnetic floppy drives, memory sticks, magnetic tape, or any other computer-readable storage media which, when the computer program code is loaded into and executed by a computer processor, the processor becomes configured as an apparatus for practicing the invention. When implemented on a general-purpose processor, the computer program code segments configure and customize the processor forming control logic circuits that transform the processor into a special purpose processor for practicing the present invention.
A non-limiting exemplary embodiment of the present invention will now be described for convenience without limitation to investigation of human breast tissue for the presence of a carcinoma. However, the invention is not limited thereto in its use and may find broad application to other types of animal tissue examination (e.g. human or non-human) for any purpose whether the tissue sample images captured by an imaging device are obtained ex vivo or in vivo.
Terms describing actions herein of the present system generally refer to actions performed and executed by specially configured data processing hardware such as a processor of an electronic computer or similar system, unless stated otherwise. Such data processing systems manipulate and transform data represented as physical quantities (e.g., electrical/electronic) within registers and memory into other forms of data represented as physical quantities within registers and memory of the computer system memories or registers when the processor is programmed with and executes appropriately configured control logic or software program instructions.
In one embodiment, the computer system 120 may include at least one programmable computer processor 122 such as a central processing unit (CPU) or other type processor, non-transitory computer readable medium (CRM) 124, input devices 126 (e.g. keyboard, mouse, audio microphone, etc.), visual display (VD) device 128 (e.g. graphic display monitor), and input/output interface 130 operably connecting and establishing communication between the foregoing system devices. The computer readable medium 124 is accessible to processor 122 via input/output interface 130 for exchanging data and program instructions. Various circuits and ancillary or peripheral devices (e.g. power supply, clock drive, etc.) may be provided to form a fully functional system as will be known to those skilled in the art. Auxiliary electronic components may be also provided and operably connected with the computer system 120 (e.g. data storage devices, printers, etc.) via wired or wireless connections.
Computer processor 122 may be connected to a wired or wireless communication infrastructure (e.g., a communications bus, cross-over bar, local area network (LAN), or wide area network (WAN). Processor 122 is programmable with control logic steps or instructions (i.e. computer program code or instructions) to form a special purpose processor for performing processor-based processes which are particularly adapted to analyzing and classifying tissues and displaying results as described herein.
Computer-readable medium 124 may be any media accessible to processor 122 that is operable to store data and computer program instructions used by the processor to implement the present invention. The computer readable medium 124 may include removable or non-removable types of both volatile and nonvolatile media. For example, computer readable medium 124 that may be used include without limitation random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory, or other type memory technology, CD-ROM, digital versatile disk (DVD), Blu-ray disk, internal/external hard disk drives, or data storage devices such as optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage, ZIP™ drive, or other magnetic storage devices, or any other medium which can be used to store data and information that is readable, writeable, and retrievable by computer processor 122. Other types of readable data storage media may be used.
The computer system 120 may also include appropriately configured communication interfaces. Communications interfaces allow software and data to be transferred between computer processor 122 and external devices. Examples of communications interfaces that may be provided include without limitation a modem, a network interface (such as an Ethernet or wireless network card), a communications port, a Personal Computer Memory Card International Association (PCMCIA) slot and card, bar code reader, USB, Blue Tooth, or the like. Software and data transferred via communications interface are in the form of signals which may be electronic, electromagnetic, optical, or any other signal capable of being received by the communications interface. These signals are provided to communications interface via a communications path or channel. The path or channel that carries the signals may be wired or wireless and implemented using wire or cable, fiber optics, a telephone line, a cellular link, a radio frequency (RF) link, or the like.
The computer program software or instructions for analyzing and classifying tissues and transforming an original captured tissue image into a digitally-enhanced displayed tissue image that facilitates a medical diagnosis are stored on computer readable medium 124. Computer processor 122 may be a general purpose computer system that becomes a specific purpose computer system when executing the program instructions describe herein. The program instructions are not limited to any particular programming language or implementation.
The imaging device 110 is a stand-alone imaging system operable to capture an image of a tissue sample which is relayed to computer system 100 for processing and analysis, as further described herein. The captured image may be a digital pixelated image formed by image data comprising a plurality of pixels. The image is communicating and transmitted electronically to the computer system 100 via wired or wireless connections for tissue analysis and classification. In some embodiments, without limitation, the imaging device 110 may be a commercially-available system using optical coherence microscopy (OCM), optical coherence tomography (OCT), confocal microscopy, or two photon microscopy to capture the image. In one preferred embodiment, OCM may be used which combines the benefits of both OCT and confocal microscopy. The imaging device captures the tissue sample with sufficient resolution to permit identification of different tissue types for analysis and classification as normal or diseased tissue.
Feature Extraction
The present invention in one embodiment uses a combination of LBP (local binary patterns) and ALBP (average local binary patterns) techniques to extract features from original OCM tissue images. Features extracted from OCM images are utilized for automatic identification of different tissue types using computer system 120.
LBP are used as features to describe textures in a grayscale image. The LBP algorithm divides the image into blocks and compares the intensity of a center pixel with the intensity values of its neighbors.
where P is the number of neighbor pixels, R is the radius of the neighborhood, gc is the grayscale value of the center pixel, gp is the gray scale value of neighbor pixels. LBPP,R selects P neighbors of the center pixel with the radius R.
So, the LBP feature shown in
Among all the LBP patterns, the most frequent patterns are the uniform patterns which have limited discontinuity in the circular binary representations thus have small U values. Based on the U values, a rotation-invariant feature LBPP,Rrui2 is defined as:
There are a total of P+2 possible values for LBPP,Rrui2. For each pixel in a N×N image, the LBPPRrui2 value is calculated. Excluding those pixels that have distance less than R from image boundary, we get (N−2R)2 LBPP,Rrui2 values. The number of occurrences of each LBPP,Rrui2 value in an image is calculated and represented in a histogram. The histogram is then normalized and saved as a LBP feature vector for the image, as shown in
According to the present invention, in addition to the above LBP features, a new variant of LBP feature is used, namely the Average LBP (ALBP) feature, which compares each neighbor pixel with the average grayscale value of all neighbors:
We calculate the rotation-invariant features ALBPP,Rrui2 by applying the same operations we described above which transform LBPP,R to LBPP,Rrui2:
Then, a variety of ALBP features can be generated for the image.
In comparison, a LBP feature can reveal the relationship between the center pixel and its neighbors, whereas an ALBP feature reveals more specific relationship among the neighbors than the original LBP feature.
LBP and ALBP are complementary. LBP features are powerful at representing the relationship of grayscale intensity values between the center pixel and the neighbors, while ALBP features are more sensitive to the intensity change in the neighborhood area.
In addition to ALBP, to further increase the discriminatory power of LBP texture features, alternative embodiments of the present invention may include yet another new variant of LBP, namely block based LBP (BLBP) features. BLBP compares the average intensity value of pixels in blocks of a certain shape in a neighborhood around the center pixel. Two different shapes of pixel blocks, namely “Spoke” and “Ring,” may be used as described herein, and as shown in
The Spoke-shaped BLBP compares the intensity of the center pixel with the average intensities of neighbor pixels along P different directions, as demonstrated in
where gp,r is the gray value of the pth neighbor pixel on the circle with radius r, gp,s is the average intensity value of all the pixels along the pth neighbor's direction for all radii in [0,R]. We calculate the rotation-invariant features SBLBPriu2 by applying the following transformation:
There are a total of P+2 different binary-number patterns for SBLBPriu2.
The Ring-shaped BLBP compares the intensity of the center pixel with the average intensities of neighbors in ring shaped areas around the center pixel, as demonstrated in
where i is the index of the ith ring between radius i*R and radius (i+1)*R around the center pixel, R is the difference of radius between two adjacent rings, N is the number of rings around the center pixel. The number of different patterns for RBLBP is 2N.
By calculating the frequency of each pattern with every pixel in the image as a center pixel, we can get SBLBP and RBLBP feature vectors which are normalized histograms of pattern occurrence. The overall BLBP feature vector is the concatenation of the SBLBP and RBLBP feature vectors as:
BLBP={SBLBP,RBLBP}
Comparing BLBP with LBP, a BLBP feature encodes richer information about the intensity distribution in blocks of different shape and scale in a neighborhood, whereas a LBP feature reveals more about the intensity differences between the center pixel and surrounding neighbor pixels. The BLBP feature is also more robust to noise than the LBP feature.
Although LBP and ALBP texture analysis methods are used in the foregoing example to extract and generate tissue features from the digital images, the invention is not limited in this regard. Accordingly, in some embodiments other tissue feature extraction techniques (texture analysis methods) may be used in addition to or instead of LBP, ALBP and BLBP alone to further improve tissue classification accuracy.
Integrated Feature and Attribute Selection
Texture features of different scales appear in human breast tissue OCM images. LBP and ALBP feature with different radius parameters capture texture features of different scales. In the present invention, a series of LBP and ALBP features with different radius parameters may be used. A multi-scale feature is constructed by integrating LBP and ALBP features with different radius parameters. These features are then evaluated in the attribute selection phase and the most representative attributes in the integrated feature are selected to be used to construct the classifiers.
LBP and its new ALBP and BLBP variants with different radius parameters can be used as features to capture the characteristics of texture patterns at different scales. A multi-scale feature is constructed by integrating LBP, ALBP and BLBP features obtained with different radius parameters.
In order to increase the speed of the classification algorithm, attribute selection can be performed before classification is carried out based on the extracted features. In attribute selection step, attributes that can represent distinctive characteristics of different types of tissues should be kept, while less relevant attributes are eliminated, in order to lower the dimension of the feature vector. After attribute selection, computation complexity will reduce for both training and testing phases, while classification accuracy remains at a high level.
Attribute selection consists of two parts: a search method which extracts a new subset of attributes from the original attribute space, and an evaluation method which assigns a score to different attribute subsets. Commonly used search methods include greedy forward selection, simulated annealing, genetic algorithm, among others. Commonly used evaluation methods include information gain, chi-squared, among others. In the present invention, a genetic algorithm may be used as the search method and information gain as the evaluation method to achieve efficient and accurate attribute subset selection. For the integrated feature shown in
Image Classification
Different classification methods are used to classify the OCM images based on the selected image feature subsets. The classification results of each classifier are evaluated by comparing to ground truth class labels obtained from histology images. Typical methods that may be used include k-nearest neighbors, naive bayes, C4.5 decision trees, and neural network classifiers.
The k-nearest neighbors is an extensively studied classification method. It has promising performance in a variety of classification applications including image based classification. Since the features in this application are numeric features, Euclidean distance is used as the distance metric when constructing the k-nearest neighbors classification model. Naive bayes is a probabilistic classifier based on Bayes' theorem. C4.5 is an decision tree based classifier. It builds a predictive model which maps the feature value to an output class which the feature belongs. Neural networks is especially useful in applications that are hard to solve by rule-based classification methods, such as image classification. In one embodiment of the present invention, a three-layer neural network may be used as the classification model. Other methods, however, may be used.
The present computer-aided diagnostic system 100 and related processes implemented by computer system 120 used for tissue analysis and disease detection will now be described in further detail.
Referring to
In an alternate or parallel process in step 230, the processor 122 generates a multiclass map showing all of the different tissue types and respective regions present in the original tissue image. In various embodiments, the map may be colored, gray scale, patterned, or combinations thereof with each tissue type found represented in various regions/locations by a visually distinct color, shading, and/or pattern. In step 235, the processor 122 overlays the multiclass tissue map onto the original tissue image to form a composite image that is digitally enhanced. This map is useful to quickly identify all of the different tissue types/structures found in the image sample. The various foregoing steps in process 200 will be further described herein.
Training Process
In order for the computer-aided diagnostic system 100 to accurately identify and classify tissue structures present in the image sample, a machine learning technique is used to program computer system 120, and more specifically to train an artificial neural network classifier of the system. Neural network classifiers are known in the art and described in U.S. Pat. No. 5,262,871, which is incorporated herein by reference in its entirety. Neural network classifiers comprise a software routine which builds a computational model based on image training inputs (e.g. histology images) that allows the computer system to identify different tissue types.
Referring to
In step 315, the digital tissue image is retrieved from the database by processor 122. In step 320, the original digital tissue image 401 is divided into a plurality of smaller tissue blocks or simply tissue blocks 402 (see, e.g.
For each tissue block 402, the processor 122 then runs a parallel set of tissue texture analysis using LBP (local binary pattern), ALBP (average local binary pattern), and BLBP (block-based local binary pattern) to generate features. The LBP feature generation is shown in step 330. For each tissue block 402, LBP extraction is performed using the methodology and algorithms described herein to obtain a LBP feature having a LBP pattern which is represented by a binary number string (see, e.g.
The ALBP feature generation is shown in step 329. For each tissue block 402, ALBP extraction is performed using the novel methodology and algorithms described herein to obtain an ALBP feature having an ALBP pattern which is represented by a binary number string (see, e.g.
For alternative embodiments combining BLBP (block-based local binary pattern) features with ALBP features, the additional steps shown in
In step 350, the computer system 120 compiles and integrates all of the LBP and ALBP features (and SBLBP AND RBLBP features in the alternative embodiment), and generates an integrated multi-scale feature of the original digital tissue image 401 in step 360.
With continuing reference to
In step 365, ground truth tissue classification is performed by processor 122. Registration is performed on the digital tissue images and their corresponding histology images, so that every block 402 in the digital tissue image has its correspondence in a histology image block 404 (see also
In step 370, the results generated in step 365 is transmitted to and stored in the training dataset database 370 for retrieval and use by processor in executing the neural network classifier for classifying the tissue types found in original digital tissue image 401 (see, e.g.
Testing Process
In step 505, the process starts in both the ALBP and BLBP processes of
Subsequent steps 515 through 540 are essentially similar to steps 315 through 360 describe above with respect to testing process 300.
In step 515, the original digital tissue image 401 is divided into a plurality of smaller tissue blocks 402 (see, e.g.
For each tissue block 402, the processor 122 then runs a parallel set of tissue texture analysis using both LBP (local binary pattern) and ALBP (average local binary pattern) to generate features. The LBP feature generation is shown in step 525. For each tissue block 402, LBP extraction is performed using the methodology and algorithms described herein to obtain a LBP feature having a LBP pattern which is represented by a binary number string (see, e.g.
The ALBP feature generation is shown in step 526. For each tissue block 402, ALBP extraction is performed using the novel methodology and algorithms described herein to obtain an ALBP feature having an ALBP pattern which is represented by a binary number string (see, e.g.
For alternative embodiments combining BLBP (block-based local binary pattern) features with ALBP features, the additional steps shown in
In step 535, the computer system 120 compiles and integrates all of the LBP and ALBP features, and generates an integrated multi-scale feature of the original digital tissue image 401 in step 540.
In step 545, the feature or attribute selection phase already described herein is performed in which the most representative attributes in the integrated feature are selected for the neural network classifier to increase the processing speed of the classification algorithm.
In step 375, the pre-trained neural network classifier identifies each of the different tissue types or classes found in the original digital tissue image 401 based on the previous LBP and ALBP (and BLBP in some embodiments) feature extraction process integrated results. The neural network classifier, using the training dataset of images from database 370, operates to classify each of the tissue types in the original digital tissue image according to pre-assigned labels for each type from the training process 300. In this example of investigating breast tissue for cancer detection, the tissue types/classes are tumor, fat, stroma, lobule, and background tissue not belonging to the other tissue classes.
The process continues with the neural network classifier then generating a heat map identifying the presence and extent of a tissue of interest (e.g. a carcinoma) in step 550 and/or generating a multi-class tissue classification map in step 551 showing all tissue types present, as further described herein. The process then ends in step 555 and may be repeated for the next tissue image to be analyzed.
It will be appreciated that in some other processes possible, SBLBP or RBLBP feature generation may be used alone in the training and testing processes of
Heat Map Generation
In step 375, the neural network classifier predicts the probability of the presence of each type or class of tissue for each tissue block image from the original tissue image. In this example of investigating breast tissue for cancer detection, the tissue types/classes may be tumor (carcinomas), fat, stroma, lobule, and background tissue. In decision step 610, the neural network classifier determines, for each tissue block image, if the probability is the largest that the tissue in the image is a tumor. If a “No” response is returned, then the tissue is classified as “normal” in step 615 and that tissue block image will not be displayed in a digitally enhanced manner (e.g. colored) in the heat map overlaid onto the original tissue image 401.
If a “Yes” is returned instead in decision step 610, control passes to decision step 620 in which the neural network classifier then determines if the probability of a tumor is larger than a preprogrammed threshold value in step 620. By setting a threshold Θ to the output value that corresponds to the tissue type of carcinoma for example, the tissue blocks in the digital tissue image which represent tumor regions may be obtained and displayed. If a digital tissue image contains blocks that are classified as tumor tissue, the whole digital tissue image is classified as tumor; otherwise, the whole digital tissue image is classified as benign.
To improve robustness of the algorithm, any isolated block less than 0.4 mm×0.4 mm in physical size in one non-limiting example that is classified as Carcinoma is considered as noise and not taken into account during image classification. Further, since the areas in the OCM image that represent regions of cancerous tissue (i.e. Carcinoma) are determined by setting a threshold θ to the output value corresponding to Carcinoma in the neural network classifier, we can obtain a receiver operating characteristic (ROC) curve by adjusting the threshold value θ, see
Feature Type
Integrated
MBLBP
MALBP
LBP
CLBP
Images
Containing
100
100
94.7
57.9
73.7
Tumor
All Benign
85.2
85.2
85.2
74.1
81.5
Overall
91.3
91.3
89.1
67.4
78.2
Accuracy
With continuing reference to
It should be noted that the foregoing process is performed for each tissue image block 401 to determine whether each block is a carcinoma which is displayed in a highlighted manner on the heat map.
In one embodiment, the tumor regions in the heat map are highlighted in the image displayed by computer system 120 on visual display device 128 as a colorized and/or darker region on top of the grayscale original tissue image 401. For each tissue block determined to be a tumor, a color is overlaid onto that tissue block in the original grayscale tissue image output and displayed by the computer system 120. The color may appear as entirely opaque in the digitally enhanced tissue image in some embodiments, and in other embodiments may be translucent to allow the original tissue structure to show through the color. Groups of colorized tissue blocks will form colorized regions of tumor tissue in the digitally enhanced version of the original tissue image.
It will be appreciated that the tumor region may be displayed in other visually discernible ways by system 120 and is not limited to the non-limiting examples described herein. Advantageously, the heat map allows the health care professional to quickly detect in a visual manner if a suspected tumor region with significant probability of being cancerous has been found by the testing for further medical investigation.
It should be noted that the processor 122 makes a determination for every single tissue block image whether the tissue found is classified as a tumor or normal. The digitally enhanced version of the original tissue image (collectively comprised of the plurality of individual tissue block images) displays each tissue block therefore as either grayscale (normal tissue) or colorized (tumor). If none of the tissue block images are classified by neural network classifier as a tumor, then no colorized heat map is produced in some embodiments and the entirety of the original tissue image appears in grayscale.
Multi-Class Tissue Classification Map
In step 375, the neural network classifier routine run by processor 122 predicts the probabilities for each type of tissue (from the preprogrammed tissue types) in each tissue image block 402 of the original OCM tissue image 401. In this example of investigating breast tissue for cancer detection, the tissue types/classes used may be tumor, fat, stroma, lobule, and background tissue. For each image block 402, the neural network classifier generates a set of tissue type probabilities for the foregoing different types learned from the training process 300. In step 710, the processor 122 compares the tissue type probabilities for each image block to determine which tissue type has the highest probability for each image block (i.e. which tissue type is most likely to be found in each image block). In step 715, the neural network classifier labels each tissue image block with the tissue type corresponding to the highest probability (i.e. tumor, fat, stroma, lobule, and background).
In step 720, the tissue type of each tissue image block 402 of the original digital tissue image is rendered in a different visually discernible manner on display device 128 of computer system 120. For each image block, the system 120 may render each tissue type in various different predetermined visual manners so that different tissue types may be readily distinguished. In one embodiment, regions of different tissue types/classes may be graphically rendered and displayed in different colors and/or patterns. An exemplary multi-class tissue classification is shown in
In step 551, a digitally enhanced version of the entire original tissue image 401 is displayed collectively showing each tissue image block in its corresponding color/pattern.
Experimental Test
In one configuration and test of the present invention, a high speed OCM system was used to image freshly excised human breast tissue based on time domain detection techniques. The setup included a femtosecond Nd:Glass laser was spectrally broadened to generate an output spectrum centered at 1060 nm with a broad spectrum bandwidth of ˜200 nm, providing an axial resolution of less than 4 μm. A 40× water immersion IR-Achroplan objective (e.g. Carl Zeiss Co.) was used to provide OCM images with a transverse resolution of ˜2 μm. Individual OCM image covering an imaging area of 500 μm×400 μm was acquired at a frame rate of 2 frames per second. A high precision three-dimensional translational stage (e.g. Newport Inc.) was used allowing rapid adjustment of focus position inside the tissue and also the imaging area over the tissue surface. The entire surface of the breast tissue specimen was imaged by translating the specimen horizontally.
In this experiment, ex vivo human breast tissue OCM images were used. The images utilized in the experiments are en face OCM images which are contrast adjusted and displayed with an inverse grayscale color map. That means darker areas in OCM images represents increased reflectivity. These OCM images have axial resolution of 4 μm, and transverse resolution of 2 μm, respectively.
The OCM images were preprocessed for the purpose of training and testing. In the training process, each OCM image is divided into blocks of 500×500 pixels. Each block is used as an instance for the training. Each of the tissue blocks is labeled with one of the five different tissue classes: adipose tissue, fibrous stroma, breast lobules, carcinomas, and background. Blocks of different classes show distinctive texture patterns in the OCM images which is used as a clue for tissue classification.
Histology images of the same human breast tissue are used to help in collecting the ground truth data. It bears noting, however, that histology images themselves are not directly used or analyzed for classifying the tissue types found which is instead based on the original tissue image captured by the label free optical imaging device (e.g. OCM images). Registration is performed on the OCM images and their corresponding histology images, so that every block in the OCM image has its correspondence in the histology image. Different types of tissues show different features on the histology images. Based on these features, each histology image is assigned a class label manually by an experienced pathologist. Then each OCM image is assigned the same class label with its corresponding histology image.
For the purpose of testing, OCM images are divided into two different sizes. In the 10 fold cross-validation test, the OCM images are divided into blocks of 500×500 pixels. It is the same size as in the training process. In another experiment, the OCM images are divided into blocks of 400×400 pixels. These blocks are used to classify different areas of the OCM image into different types of tissue. Different sizes are used in this experiment, in order to avoid generating the same tissue blocks in the training process, thus increasing the reliability of the experimental results.
In order to test and verify the performance of the present computer-aided diagnostic system 100 and integrated LBP/ALBP feature generated on classifying human tissue OCM images, classification on OCM image tissue blocks obtained from the image preprocessing step was performed. Tissue blocks were classified into five categories: lobule, stroma, carcinoma, adipose and background. In the experimental test, 4359 OCM image tissue blocks were used for the experiments (198 lobule, 2065 stroma, 1127 carcinoma, 347 adipose, 572 background). Classification was performed by using the neural network classifier.
The classification accuracy (%) using 10 fold cross-validation by using different features is shown in the table below:
Back-
Overall
Features
Lobule
Stroma
Carcinoma
Adipose
ground
Accuracy
LBP2,8
7.1
81.5
73.4
30.5
58.7
68.8
LBP4,8
23.2
82.2
75.7
69.2
77.1
76.1
LBP8,2
53.5
86.0
72.8
46.7
88.3
78.2
LBP8,4
39.9
86.3
77.7
79.5
87.4
81.6
LBP8,8
29.3
86.2
78.0
80.1
92.1
81.7
LBP8,16
25.3
81.6
60.5
69.7
87.6
73.4
LBP8,8,
31.8
86.9
79.8
87.9
97.7
84.0
ALBP8,8
MLBP,
61.1
92.3
85.4
94.2
99.3
90.2
MALBP
BLBP
59.3
88.2
84.6
88.8
99.1
87.4
MBLBP
77.4
96.2
83.2
95.9
100
92.4
Integrated
68.3
94.0
93.5
97.1
100
93.8
LTP8,8
10.6
86.4
81.4
90.5
93.9
82.9
CLBP
57.3
89.6
73.8
81.8
99.0
84.6
The features shown in the above table are:
One can see that, the multi-scale integrated feature (MLBP+MALBP+MBLBP) gives the best result. Since the new ALBP and BLBP features are more robust to image noise and can extract richer, more distinctive information from images based on intensity variance, they demonstrated superior performance in distinguishing image blocks of different types of tissue, when integrated with original LBP. Using multi-scale features improves performance because OCM images of different tissue types exhibit different texture features at multiple scales. The classification performance was also compared using the new features with that of using the original LBP feature alone, the local ternary patterns (LTP) of other researchers (Tan and Triggs, 2007), and complete local binary patterns (CLBP) (Guo et al., 2010). The present results showed that multi-scale integrated features incorporating the newly proposed ALBP and BLBP features outperformed other features in distinguishing different types of breast tissue in OCM images.
While the foregoing description and drawings represent exemplary embodiments of the present invention, it will be understood that various additions, modifications and substitutions may be made therein without departing from the spirit and scope and range of equivalents of the accompanying claims. In particular, it will be clear to those skilled in the art that the present invention may be embodied in other forms, structures, arrangements, proportions, sizes, and with other elements, materials, and components, without departing from the spirit or essential characteristics thereof. In addition, numerous variations in the methods/processes as applicable described herein may be made without departing from the spirit of the invention. One skilled in the art will further appreciate that the invention may be used with many modifications of structure, arrangement, proportions, sizes, materials, and components and otherwise, used in the practice of the invention, which are particularly adapted to specific environments and operative requirements without departing from the principles of the present invention. The presently disclosed embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being defined by the appended claims and equivalents thereof, and not limited to the foregoing description or embodiments. Rather, the appended claims should be construed broadly, to include other variants and embodiments of the invention, which may be made by those skilled in the art without departing from the scope and range of equivalents of the invention.
Zhou, Chao, Huang, Xiaolei, Wan, Sunhua
Patent | Priority | Assignee | Title |
11017207, | Aug 28 2019 | Applied Materials, Inc | System for automatic tumor detection and classification |
11688188, | Aug 28 2019 | Applied Materials, Inc. | System for automatic tumor detection and classification |
11988596, | Oct 23 2020 | Becton, Dickinson and Company | Systems and methods for imaging and image-based analysis of test devices |
11996183, | Sep 17 2020 | SCANWELL HEALTH, INC | Methods of analyzing diagnostic test kits |
D970033, | Oct 23 2020 | Becton, Dickinson and Company | Cartridge imaging background device |
Patent | Priority | Assignee | Title |
7298897, | Feb 11 2004 | NATIONAL AERONAUTICS AND SPACE ADMINISTRATION, U S GOVERNMENT AS REPRESENTED BY THE ADMINISTRATOR OF | Optimal binarization of gray-scaled digital images via fuzzy reasoning |
8047990, | Jan 19 2006 | Collagen density and structural change measurement and mapping in tissue | |
8473035, | Sep 15 2003 | Beth Israel Deaconess Medical Center | Medical imaging systems |
20080033657, | |||
20090324051, | |||
20100098306, | |||
20120130243, | |||
20130286038, | |||
20140044361, | |||
20140247972, | |||
20160203599, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Dec 05 2013 | HUANG, XIAOLEI | Lehigh University | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038270 | /0124 | |
Dec 05 2013 | WAN, SUNHUA | Lehigh University | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038270 | /0124 | |
Dec 05 2013 | ZHOU, CHAO | Lehigh University | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 038270 | /0124 | |
Apr 13 2016 | Lehigh University | (assignment on the face of the patent) | / | |||
Aug 04 2020 | Lehigh University | NATIONAL INSTITUTES OF HEALTH NIH , U S DEPT OF HEALTH AND HUMAN SERVICES DHHS , U S GOVERNMENT | CONFIRMATORY LICENSE SEE DOCUMENT FOR DETAILS | 053452 | /0120 |
Date | Maintenance Fee Events |
Jan 28 2022 | MICR: Entity status set to Micro. |
Feb 28 2022 | REM: Maintenance Fee Reminder Mailed. |
Jul 05 2022 | M3551: Payment of Maintenance Fee, 4th Year, Micro Entity. |
Jul 05 2022 | M3554: Surcharge for Late Payment, Micro Entity. |
Date | Maintenance Schedule |
Jul 10 2021 | 4 years fee payment window open |
Jan 10 2022 | 6 months grace period start (w surcharge) |
Jul 10 2022 | patent expiry (for year 4) |
Jul 10 2024 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jul 10 2025 | 8 years fee payment window open |
Jan 10 2026 | 6 months grace period start (w surcharge) |
Jul 10 2026 | patent expiry (for year 8) |
Jul 10 2028 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jul 10 2029 | 12 years fee payment window open |
Jan 10 2030 | 6 months grace period start (w surcharge) |
Jul 10 2030 | patent expiry (for year 12) |
Jul 10 2032 | 2 years to revive unintentionally abandoned end. (for year 12) |