CN114037671B - Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN - Google Patents
Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN Download PDFInfo
- Publication number
- CN114037671B CN114037671B CN202111282117.XA CN202111282117A CN114037671B CN 114037671 B CN114037671 B CN 114037671B CN 202111282117 A CN202111282117 A CN 202111282117A CN 114037671 B CN114037671 B CN 114037671B
- Authority
- CN
- China
- Prior art keywords
- spectrum
- network
- image
- feature
- layer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 210000000265 leukocyte Anatomy 0.000 title claims abstract description 44
- 238000001514 detection method Methods 0.000 title claims abstract description 37
- 230000006872 improvement Effects 0.000 title claims abstract description 11
- 238000001228 spectrum Methods 0.000 claims abstract description 78
- 238000000605 extraction Methods 0.000 claims abstract description 48
- 210000004369 blood Anatomy 0.000 claims abstract description 8
- 239000008280 blood Substances 0.000 claims abstract description 8
- 238000013075 data extraction Methods 0.000 claims abstract description 8
- 238000004519 manufacturing process Methods 0.000 claims abstract description 3
- 210000004027 cell Anatomy 0.000 claims description 41
- 230000003595 spectral effect Effects 0.000 claims description 23
- 230000004927 fusion Effects 0.000 claims description 17
- 238000012937 correction Methods 0.000 claims description 15
- 238000002310 reflectometry Methods 0.000 claims description 14
- 238000000034 method Methods 0.000 claims description 12
- 238000011176 pooling Methods 0.000 claims description 12
- 238000012549 training Methods 0.000 claims description 7
- 238000010276 construction Methods 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 4
- 238000012216 screening Methods 0.000 claims description 4
- 238000010606 normalization Methods 0.000 claims description 3
- 238000012795 verification Methods 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 2
- 210000005056 cell body Anatomy 0.000 claims description 2
- 230000008859 change Effects 0.000 claims description 2
- 238000013527 convolutional neural network Methods 0.000 abstract description 2
- 210000000440 neutrophil Anatomy 0.000 description 13
- 238000004422 calculation algorithm Methods 0.000 description 8
- 210000003651 basophil Anatomy 0.000 description 7
- 210000004698 lymphocyte Anatomy 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 238000012986 modification Methods 0.000 description 5
- 230000004048 modification Effects 0.000 description 5
- 230000001575 pathological effect Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 210000003979 eosinophil Anatomy 0.000 description 4
- 210000001616 monocyte Anatomy 0.000 description 4
- 208000024893 Acute lymphoblastic leukemia Diseases 0.000 description 3
- 208000014697 Acute lymphocytic leukaemia Diseases 0.000 description 3
- 208000006664 Precursor Cell Lymphoblastic Leukemia-Lymphoma Diseases 0.000 description 3
- 230000008901 benefit Effects 0.000 description 3
- 210000001185 bone marrow Anatomy 0.000 description 3
- 238000000701 chemical imaging Methods 0.000 description 3
- 230000006835 compression Effects 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 230000007613 environmental effect Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 210000000601 blood cell Anatomy 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 210000003743 erythrocyte Anatomy 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 238000003909 pattern recognition Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 206010000830 Acute leukaemia Diseases 0.000 description 1
- 208000035143 Bacterial infection Diseases 0.000 description 1
- 208000028018 Lymphocytic leukaemia Diseases 0.000 description 1
- 206010028980 Neoplasm Diseases 0.000 description 1
- 208000036142 Viral infection Diseases 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 208000022362 bacterial infectious disease Diseases 0.000 description 1
- 210000003969 blast cell Anatomy 0.000 description 1
- 210000001772 blood platelet Anatomy 0.000 description 1
- 201000011510 cancer Diseases 0.000 description 1
- 210000003855 cell nucleus Anatomy 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 244000078673 foodborn pathogen Species 0.000 description 1
- 238000009432 framing Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 230000001926 lymphatic effect Effects 0.000 description 1
- 208000003747 lymphoid leukemia Diseases 0.000 description 1
- 210000003563 lymphoid tissue Anatomy 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 208000025113 myeloid leukemia Diseases 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 210000004940 nucleus Anatomy 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 210000004976 peripheral blood cell Anatomy 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 230000003612 virological effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/17—Systems in which incident light is modified in accordance with the properties of the material investigated
- G01N21/25—Colour; Spectral properties, i.e. comparison of effect of material on the light at two or more different wavelengths or wavelength bands
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10032—Satellite or aerial image; Remote sensing
- G06T2207/10036—Multispectral image; Hyperspectral image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10056—Microscopic image
- G06T2207/10061—Microscopic image from scanning electron microscope
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30024—Cell structures in vitro; Tissue sections in vitro
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30204—Marker
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Probability & Statistics with Applications (AREA)
- Chemical & Material Sciences (AREA)
- Analytical Chemistry (AREA)
- Biochemistry (AREA)
- Immunology (AREA)
- Pathology (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Investigating Or Analysing Biological Materials (AREA)
- Image Processing (AREA)
Abstract
The invention belongs to the technical field of medical treatment, and provides a white blood cell detection method based on an improved FASTER RCNN microscopic hyperspectral image, which uses a hyperspectral microscope to obtain a white blood cell detection image of a blood smear and carries out identification and classification on white blood cells based on an improved FASTER RCNN. And (3) obtaining pseudo-color images and spectrum data of the blood smear by using a hyperspectral microscope, marking different types of white blood cells, and manufacturing a data set. The traditional FASTER RCNN network is improved, namely the VGG16 in the FASTER RCNN original network is replaced by Resnet and is used as a new pseudo-color image feature extraction network; and establishing a spectrum data extraction module aiming at blood smear hyperspectral data, extracting spectrum features by utilizing a one-dimensional convolutional neural network, and fusing white blood cell image features and spectrum features by utilizing a FASTER RCNN network on the basis of the improvement, so as to finally realize white blood cell identification and classification. Compared with the traditional FASTER RCNN network, the invention has obvious improvement on the identification precision and classification accuracy of the white blood cells.
Description
Technical Field
The invention belongs to the technical field of medical treatment, and particularly relates to a microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN.
Background
White blood cells are an important component of blood, produced by bone marrow and lymphoid tissue, and function to combat viral and bacterial infections. Can be classified into neutrophils, lymphocytes, monocytes, eosinophils, and basophils. Traditional manual microscopic detection methods are overly complex, and manual differential counting of leukocytes is more prone to error, particularly when detecting large numbers of samples. In recent years, corresponding computer vision algorithms and systems have been widely used in the field of automatic classification and detection of blood cells.
Computer vision inspection can be divided into two major categories at present, the first category is to simply stack traditional algorithm modules according to the order of preprocessing, feature extraction and classification. Rawat et al propose a new technique for distinguishing acute lymphoblastic leukemia (Acute Lymphoblastic Leukemia) from healthy lymphocytes, using an SVM classifier to classify fusion features including texture and shape with an accuracy of 89.8%. Mohapatra et al adopts a two-stage color segmentation strategy based on fuzzy clustering to segment white blood cells in an image, and proposes two new shape features of Hausdorff dimension and outline feature, and uses SVM to classify lymphatic nuclei, so that the accuracy reaches 95%. Alferez et al utilized color component clustering and watershed transformation to segment peripheral blood cell images, and extracted 113 features altogether for different types of lymphocyte recognition. The accuracy of the training set is 98.07%, and the accuracy of the verification set is 85.33%.
However, most of the conventional pattern recognition algorithms are used for sequentially completing the representation and classification modules, the super parameters of the conventional pattern recognition algorithms are usually given by experience before training, and the generalization capability and stability of the model cannot be ensured when the sample space is small or the inter-class gap is not obvious.
The second category is to analyze pathological pictures based on a deep learning algorithm, and the computer can extract multidimensional features more comprehensively and automatically without manually designating the extraction of a certain feature, so that after a large amount of training, the targets in the images can be identified and classified more accurately, efficiently and stably, and strong strength is shown in a plurality of fields, particularly in the medical field with a large amount of image analysis and processing, and the method has the advantage incomparable with the traditional machine learning algorithm. Boldu et AL in the study of acute leukemia detection, a dual-module deep learning architecture (AL Net) was used to identify and classify abnormal promyelocytic and blast cells, respectively, and the accuracy of identification of myeloid leukemia and lymphoid leukemia reached 93.7% and 100%, respectively. SHAKARAMI et al used a fast and efficient YOLOv (FED) detector with EFFICIENTNET convolutional neural network as the backbone, enabling blood cell identification based on BCCD datasets. Finally, the classification accuracy of platelets, erythrocytes and leukocytes was 90.25%, 80.41% and 98.92%, respectively.
In the published patent, it is referred to the detection of cells using FASTER RCNN networks as follows:
Patent CN109598224a proposes a method for detecting white blood cells in bone marrow section based on region recommended convolution, which uses Resnet and FPN as a characteristic extraction network of FASTER RCNN, and bone marrow section image as input, and the accuracy can reach more than 90%. Patent CN110059672a proposes a method for performing class-increasing learning on a microscope cell image detection model by using incremental learning, calculates the distance between an intermediate feature layer obtained by predicting new class cells by using a FASTER RCNN model trained by an old class cell image and an intermediate feature layer of an incremental model, and obtains a new class-increasing prediction model. Patent CN110580699a proposes a pathological image cell nucleus detection method based on an improved FASTER RCNN algorithm, uses ZF as a FASTER RCNN feature extraction network, processes sample pictures by data enhancement transformation, difficult sample mining and small target detection optimization, achieves a monitoring speed of 1 s/sheet, and has good robustness. Patent CN111598849a proposes a pathological image cell counting method, device and medium based on target detection, which refers to two different FASTER RCNN models trained by bockbone to perform cell counting detection on pathological images, and fuses the detection results to accurately, rapidly and efficiently realize pathological cell counting.
However, these methods are only based on gray scale or RGB images with cellular space characteristics, and are susceptible to various environmental factors such as optical conditions of a microscope, thickness of a slide, etc. The hyperspectral imaging technology is a combination of the traditional imaging technology and the spectrum technology, so that not only the spatial characteristics of a monitored target but also the spectrum characteristics of the monitored target can be extracted. In the medical field, it has been used for the identification and diagnosis of acute lymphoblastic leukemia, food borne pathogens and cancer. Although hyperspectral imaging provides rich spectral information on the basis of spatial features, hundreds of narrow continuous bands make the data dimension redundant, requiring manual extraction of spectral values of a region of interest (ROI), which is extremely cumbersome in practical monitoring. Therefore, there is a strong need for an automatic, fast and efficient method of white blood cell detection based on hyperspectral imaging.
According to the invention, the original VGG16 characteristic extraction network FASTER RCNN is replaced by Resnet, except for stacking on a simple convolution layer, the cross arrangement of two basic blocks of Conv Block and Identity Block in the network can avoid losing important characteristics due to mismatch between the number of convolution layers and data, and the generalization of the network is far superior to that of VGG16. Besides the replacement of the feature extraction network, the invention is different from the original FASTER RCNN network frame in that the biggest innovation point of the patents and researches is that a complete set of spectrum feature extraction network and feature fusion network are constructed and embedded into FASTER RCNN, thereby realizing the novel automatic white blood cell identification and classification technology based on the combined detection of space and spectrum features.
Disclosure of Invention
In order to solve the technical problems, the invention provides a microscopic hyperspectral leukocyte detection method based on an improved FASTER RCNN algorithm, which adds a detection network branch of spectrum data in a traditional FASTER RCNN network architecture, realizes the joint detection of the spatial characteristics and the spectral characteristics of blood images, and effectively improves the classification precision.
The technical scheme of the invention is as follows:
the microscopic hyperspectral leukocyte detection method based on the improvement FASTER RCNN comprises the following steps:
1. Establishing a leukocyte dataset: collecting a pseudo-color image (an image synthesized by 667.4,557.2 and 440.5 wave bands) and a spectrum image of a blood smear by using a microscopic hyperspectral system, correcting the color of the pseudo-color image, marking five types of white blood cells by frame selection and category, correcting the black and white of the spectrum image, and manufacturing a training and verification data set;
2. And (3) constructing a pseudo-color image feature extraction network: resnet18 which can extract more effective features is adopted as a feature extraction network of FASTER RCNN;
3. Constructing an RPN proposal network, which is responsible for automatically generating an initial target priori frame, carrying out preliminary classification on the priori frame, and then finally screening according to the priori frame to obtain a proposal frame, and intercepting corresponding features at a public feature layer;
4. And constructing a suggested frame coordinate return branch, constructing a suggested frame coordinate return network branch by adopting a full connection layer (Dense), and decoding the suggested frame coordinate return network branch to obtain coordinate values of the target feature on the original image.
5. And (3) constructing a spectrum data extraction module: using FASTER RCNN to detect four coordinate values obtained by a network terminal positioning frame coordinate branch as input of a spectrum extraction module, intercepting full-band data of a corresponding region in a spectrum image according to the coordinate, and then calculating an average spectrum value of the full-band data to obtain an average spectrum curve as input of a spectrum feature extraction network;
6. Constructing a spectrum feature extraction network, additionally constructing a one-dimensional spectrum data feature extraction network, and outputting one-dimensional spectrum features;
7. Construction of a feature fusion network: the characteristics extracted from the tail end of the spectrum characteristic extraction network are fused with a FASTER RCNN tail end of a characteristic layer based on cell pseudo-color image classification;
8. Construction of cell sorting layer: calculating the obtained one-dimensional fusion characteristics by using the full-connection layer and the softmax classification layer to obtain a final cell classification result;
Further, in step (1), the hyperspectral camera cannot obtain a conventional RGB image due to the limitation of the band range, so that the RGB image, i.e. the pseudo-color image, is synthesized by extracting 667.4,557.2 and 440.5 three bands representing R, G, and B channels, respectively. The Gamma correction is used for correcting the color of the pseudo-color image, so that the influence of various environmental factors, such as the optical condition of a microscope, the thickness of a slide and the like, on the brightness of images in different batches is reduced. Black and white correction of spectral images specifically refers to: collecting a pair of blank slide spectral images as white reference images for correction, namely, defining the spectral reflectivity of a blank slide as the highest value; and covering a lens cover of the spectrum camera, closing a light source, and collecting a spectrum image in a pure black environment as a corrected black reference image, namely, defining the spectrum reflectivity in the dark environment as the minimum value. And then carrying out normalization processing on the data of all the spectral images based on the black-white correction data to obtain corrected spectral images. Each white blood cell within the image field of view is framed and based on the corresponding class label, the final dataset includes for each image: a pseudo color image; a spectral image; a property file (containing the coordinates of the positioning frame and the classification label of the white blood cells contained in the image).
Further, in the step (2), the feature extraction network Resnet of the pseudo-color image is specifically built by using a small convolution kernel of 3*3, two basic blocks of Conv Block and Identity Block in the network are arranged in a crossing manner, the Conv Block is used for compressing the image, the Identity Block is not used for changing the size of the image, and the key point is that the two feature extraction modules combine the extracted features with the original features, thereby not only ensuring the extraction of important features, but also ensuring that the original features are not missed due to deepening of the layer number. The spectrum feature extraction network is built in the form of Conv1D+ MaxPooling D layer based on 1*3 convolution kernels.
Further, in step (3), the common feature obtained by the pseudo-color image after passing through the feature extraction network has two application paths, one of which is to sweep it through a 3*3-size convolution to obtain a feature of a deeper layer, and then to divide it into two network branches, one of which is used to generate 9 prior frames for each point. And secondly, generating a deviation value of each priori frame and the true annotation frame. These two sets of features are then passed sequentially into propasal and ROI Pooling layers for preliminary classification of the prior frames and screening of the final suggested frames, respectively. The second purpose of the public feature layer is to input the public feature layer into the ROI Pooling layer, intercept corresponding feature blocks according to the screened suggestion boxes, unify the feature fast sizes and obtain final target features.
Further, in step (4) and step (5), after the pseudo color image is input to FASTER RCNN network and features are extracted by Resnet, the pseudo color image passes through RPN network, the network determines those areas in the image are cell bodies, and feature blocks with different sizes are obtained, and after ROI Pooling layers, the feature blocks proposed by the network are converted into uniform sizes. Then FASTER RCNN is divided into two branches, one branch is a return network for predicting the coordinates of the positioning frame and is used for learning and returning the four coordinates of the predicted cell positioning frame and the deviation value of the coordinates of the labeling positioning frame in the attribute file, and the other branch is a classification network for five classifications of cells. The spectrum data is stored in a picture file of a raw type, the file type is that reflectivity values of all pixels in the spectrum picture corresponding to all wave bands are arranged end to form a one-dimensional list, the one-dimensional list is rearranged into a three-dimensional data block according to a format of length multiplied by height multiplied by wave band number of an image, then the corresponding pixel points and the reflectivity values of the corresponding pixel points in all wave bands are cut out by utilizing four coordinate values of a cell prediction coordinate frame obtained based on a pseudo-color image, and the reflectivity average value of all the pixel points is obtained under each wave band to obtain an average spectrum curve which is used as input of a spectrum characteristic network.
Further, in the step (6), the average spectrum curve of each cell obtained by the spectrum data extraction module is one-dimensional data, so that a spectrum-oriented feature extraction network is built by using Conv 1D. The network is built by a Conv1D+ Max Pooling structure, 350-dimensional spectrum data are compressed, after a feature layer is obtained, the Flatten operation is carried out, a one-dimensional full feature layer is obtained, a full-connection Dense layer is connected, and then the layer is connected with a pseudo-color image feature layer in a FASTER RCNN tail cell classification branch.
Further, in the step (7), before the classification of FASTER RCNN is performed by using the softmax layer, a plurality of feature blocks obtained by the RPN layer are subjected to a flat operation to be changed into a one-dimensional feature layer, the last layer of the spectrum feature extraction network is also a one-dimensional feature layer, and the two feature layers are spliced according to the weight of 1:1, so that a fusion feature layer comprising space and spectrum features is obtained.
Further, in the step (8), after a one-dimensional fusion feature layer is obtained, the features are transferred into a full-connection layer and a classification layer of softmax, so that classification based on the fusion features is realized.
Compared with the prior art, the invention provides a microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN, and particularly creates innovation in the aspect of feature Fusion detection of multi-sensor data, and provides a multi-class data feature Fusion detection method, namely a multi-feature Fusion network Fusion-Net.
Drawings
FIG. 1 is a schematic block diagram of the present invention;
FIG. 2A is a microscopic hyperspectral image of class 1 (neutrophils) and class 5 (monocytes) cells;
FIG. 2B is a microscopic hyperspectral image of class 1 (neutrophils) and class 3 (basophils) cells;
FIG. 2C is a microscopic hyperspectral image of class 1 (neutrophils), class 3 (basophils) and class 4 (lymphocytes) cells;
FIG. 2D is a microscopic hyperspectral image of class 1 (neutrophils) and class 2 (eosinophils) cells;
FIG. 3Resnet is a schematic diagram of a feature extraction network; (a) is Conv Block; (b) is Identity Block; (c) is Resnet < 18 >;
FIG. 4 is a spectral feature extraction network;
FIG. 5 is a schematic diagram of a network structure of FASTER RCNN after modification;
FIG. 6A is a graph showing the results of conventional FASTER RCNN-class 1 (neutrophils) and 5-class 5 (monocytes) cell detection classification;
FIG. 6B is a graph showing the results of conventional FASTER RCNN-based classification of class 1 (neutrophils) and class 3 (basophils) cells;
FIG. 6C is a graph showing the results of conventional FASTER RCNN-based classification of class 1 (neutrophils), class 3 (basophils) and class 4 (lymphocytes) cells;
FIG. 6D is a graph showing the results of conventional FASTER RCNN-based classification of class 1 (neutrophils) and class 2 (eosinophils) cells;
FIG. 7A is a graph showing the results of the detection classification of FASTER RCNN cells of class 1 (neutrophils) and class 5 (monocytes) after modification of the present invention;
FIG. 7B is a graph showing the results of the detection classification of FASTER RCNN versus class 1 (neutrophils) and class 3 (basophils) cells after modification of the present invention;
FIG. 7C is a graph showing the results of the detection classification of FASTER RCNN cells of class 1 (neutrophils), class 3 (basophils) and class 4 (lymphocytes) after modification of the invention;
FIG. 7D is a graph showing the results of the detection classification of FASTER RCNN versus class 1 (neutrophils) and class 2 (eosinophils) cells after modification of the present invention.
Detailed Description
The following describes the embodiments of the present invention further with reference to the drawings and technical schemes.
As shown in fig. 1, a microscopic hyperspectral leukocyte detection method based on the improvement FASTER RCNN comprises the following specific implementation steps:
s1, establishing a leukocyte data set:
the hyperspectral camera is arranged on a three-eye student microscope, and the wave band used by the hyperspectral camera is a near infrared short wave light region between 382nm and 1020 nm. The method comprises the steps of finding a visual field containing white blood cells through an eyepiece, enabling a detection target in the visual field to be clearly visible through focusing and other operations, and then collecting images and spectrum information in the visual field by using a hyperspectral camera.
And carrying out color correction on the acquired pseudo-color image. And Gamma correction is adopted, so that the influence of various environmental factors, such as the optical condition of a microscope, the thickness of a slide and the like, on the brightness of images of different batches is reduced. The gamma correction is a method of editing a gamma curve of an image to perform nonlinear tone editing of the image, and increases the ratio of dark color portions to light color portions in an image signal to thereby improve the image contrast effect.
And performing box selection and category labeling on the five types of white blood cells. And using marking software to carry out frame selection on each cell on each image, giving corresponding classification labels, and generating an attribute file containing four coordinates of the positioning frame and the corresponding labels. The principle of the frame selection is that background elements of non-leucocytes appear in the frame as little as possible, and when two leucocytes are overlapped, the content of non-own cell elements in the frame is reduced as much as possible.
Black and white correction is performed on the spectral image. To reduce the sample position differences and the effect of the light source on the image, the original hyperspectral image is corrected to the reflectance mode. Collecting a pair of blank slide spectral images as white reference images for correction, namely, defining the spectral reflectivity of a blank plectrum to be the highest value (the reflectivity is close to 100%); the lens cover of the spectrum camera is covered, the light source is turned off, and a spectrum image under a pure black environment is collected and used as a corrected black reference image, namely, the spectrum reflectivity under the dark environment is regulated to be the minimum value (the reflectivity is close to 0%). And then carrying out normalization processing on the data of all the spectral images based on the black-white correction data to obtain corrected spectral images. The correction process is as follows:
R=(I-B)/(W-B)×100%
Wherein I represents an original image; w represents a full white calibration image; b represents a full black calibration image; r represents the corrected relative image.
For each image, the final dataset includes: a pseudo color image; a spectral image; a property file (containing the coordinates of the positioning frame and the classification label of the white blood cells contained in the image). The image of five types of white blood cells collected is shown in fig. 2.
S2, pseudo color image feature extraction network:
Resnet18 was employed as the feature extraction network for FASTER RCNN. As shown in fig. 3, resnet has two basic blocks, namely Conv Block and Identity Block, wherein the dimensions of Conv Block input and output are different, so that the Conv Block cannot be serially connected, and the effect of the Conv Block is to change the dimension of a network; the Identity Block input dimension and the output dimension are the same and can be connected in series to deepen the network. The pseudo-color image feature extraction network of the improvement FASTERRCNN only comprises four image size compressions, and the content after the fifth compression is used after ROI Pooling layers. The specific structure is shown in the Resnet block diagram in fig. 3, and after the 600×600×3 picture at the time of input is sent to Resnet, a feature block with a size of 38×38×256 is obtained.
S3.rpn proposed network:
As shown in the network structure schematic diagram of FASTER RCNN after improvement in fig. 5, the public feature obtained by the pseudo-color image after passing through the feature extraction network has two application ways, one of which is to sweep it through one 3*3-sized convolution to obtain a feature of a deeper layer, and then to divide it into two network branches, one of which is to sweep it through 9-channel convolution with a size of 1*1, that is, the number of prior frames generated by each pixel point in FASTER RCNN is 9. The second is to sweep using 36-channel convolution of size 1*1, i.e., each a priori frame contains 4 coordinate variables, so 9*4 =36 channels are produced per pixel. Both sets of features are input to proposal of FASTER RCNN, where the objects in each prior box are initially classified, i.e., whether the object is included or not is determined, and then the object to be detected is found in the prior box determined to be an object. At this time, the feature sizes of the extracted targets are different, and then the sizes of the prior frames are unified at ROI Pooling layers to obtain the final suggestion frame. Simultaneously, the common features extracted from the previous feature network are also input into the ROI Pooling layers, the features of 14 x 14 are cut out from the common features according to the suggested frame coordinates after the unified size, and then the features of each suggested frame are subjected to the fifth compression of the original Resnet to obtain the features of 7*7. So far, the whole RPN network is ended, and a plurality of characteristic blocks which are judged to be detection targets by the RPN network are obtained.
S4-S5, suggesting a frame coordinate return branch and a spectrum data extraction module:
After obtaining a number 7*7 of feature blocks through the RPN network, the flag operation is performed to become one-dimensional data, and then there are two branches, namely a return branch of the coordinates of the suggestion frame, which is constructed by a full connection layer, as shown in fig. 5, and is used for adjusting the corresponding suggestion frame, and 20 (the number of classification×4) neurons are included, that is, each cell includes 4 coordinate parameters no matter which class is divided into. Finally, after decoding the layer of data on the original image, coordinate values of four vertexes can be obtained, one of the purposes is to draw a recognition frame on the original image, and the other is to use four coordinate values obtained by suggesting frame coordinate branches as input of a spectrum extraction module. The spectrum data extraction module is written by using OpenCV and mainly comprises three parts: reading and arranging spectrum data in a raw file; intercepting a corresponding data block according to the coordinate value of the suggestion frame; and carrying out average calculation on the intercepted spectrum data on each wave band to obtain average spectrum data of the target in each suggestion frame, wherein the average spectrum data is used as input of a spectrum characteristic extraction network.
S6, spectral feature extraction network:
the average spectrum curve of each cell obtained through the spectrum data extraction module is one-dimensional data, so that a spectrum-oriented characteristic extraction network is built by using Conv 1D. A specific network structure is shown in fig. 4. The network is built by a Conv1D+ Max Pooling structure, 350-dimensional spectrum data are compressed, after a characteristic layer of 1 x 173 x 64 is obtained, a flat operation is carried out, a full characteristic layer of 1 x 5568 is obtained, a full connection Dense layer is connected, and then the layer is connected with a pseudo-color image characteristic layer in a FASTER RCNN tail cell classification branch.
S7, feature fusion network:
As shown in fig. 5, before classifying by softmax in the classification branch of FASTER RCNN, the features obtained by the RPN layer are subjected to the flatten operation to become a one-dimensional feature layer, the last layer of the spectral feature extraction network is also a one-dimensional feature layer, and the two feature layers are spliced according to a weight of 1:1, so that a fused feature layer comprising spatial and spectral features is obtained.
S8, cell classification network:
after a one-dimensional fusion feature layer is obtained, the features are transferred into a full-connection layer and a classification layer of softmax, so that classification based on the fusion features is realized.
The above is all the steps of the flow of the invention, and during training, the data set is randomly divided into a training set and a testing set according to the ratio of 7:3. Meanwhile, a conventional FASTER RCNN network which is characterized by VGG16 extraction and is trained only based on pseudo-color images is used as a control group, the detection result is shown in fig. 6, detection confusion can occur on two white blood cells which are closer to each other, two white blood cells are used as one cell for framing, and due to interference of a large number of red blood cells on a blood picture, part of white blood cells cannot be identified, besides the identification accuracy, the conventional FASTER RCNN network classifies all the identified white blood cells into one type, and the fact that the conventional FASTER RCNN network is poor in identification and classification accuracy is reflected in the condition that a data set is smaller. Fig. 7 shows the improved FASTER RCNN network detection result based on the combined detection of spatial and spectral features, and the same test image, each white blood cell is accurately identified and positioned, which shows the advantage of Resnet in feature extraction, and each white blood cell is accurately classified into 5 classes, which is because the white blood cells have obvious advantages in the inter-class distinction of white blood cells under the conditions that the difference between the white blood cells on the image is small and the data set is not huge enough.
Claims (1)
1. The microscopic hyperspectral leukocyte detection method based on the improvement FASTER RCNN is characterized by comprising the following steps:
(1) Establishing a leukocyte dataset: collecting a pseudo-color image and a spectrum image of a blood smear by using a microscopic hyperspectral system, carrying out color correction on the pseudo-color image, carrying out frame selection and category marking on five types of white blood cells, carrying out black-and-white correction on the spectrum image, and manufacturing a training data set and a verification data set;
The pseudo-color image is synthesized by images corresponding to three wavebands 667.4, 557.2 and 440.5, the three wavebands correspond to R, G and B channels of the image respectively, and Gamma is used for correcting the color of the pseudo-color image;
The black-and-white correction of the spectrum image specifically refers to: collecting a pair of blank slide spectral images as white reference images for correction, namely, defining the spectral reflectivity of a blank slide as the highest value; covering a lens cover of a spectrum camera, closing a light source, and collecting a spectrum image in a pure black environment as a corrected black reference image, namely, defining the spectrum reflectivity in the dark environment as the minimum value; then, carrying out normalization processing on the data of all the spectrum images based on the white reference image and the black reference image to obtain corrected spectrum images; each white blood cell in the image view is framed and selected based on the corresponding classification label; for each image, the final dataset comprises a pseudo-color image, a spectral image, a property file containing the location frame coordinates and classification tags of the white blood cells contained in the image;
(2) Constructing a pseudo color image feature extraction network: resnet18 is adopted as a feature extraction network of FASTER RCNN;
Resnet 18 the construction is carried out by using a 3*3 small convolution kernel, a characteristic extraction module in the network is a Conv Block and an Identity Block which are arranged in a crossing way, the Conv Block is used for compressing an image, the Identity Block cannot change the size of the image, and the two characteristic extraction modules combine the extracted characteristics with original characteristics to obtain a common characteristic layer;
The characteristic extraction network of the spectrum image is built in a form of Conv1D+ MaxPooling D layer based on 1*3 convolution kernels;
(3) Constructing an RPN proposal network, which is responsible for automatically generating an initial target priori frame, carrying out preliminary classification on the target priori frame, and then finally screening according to the target priori frame to obtain a proposal frame, and intercepting corresponding features at a public feature layer;
The public feature layer obtained after the pseudo-color image passes through the image feature extraction network has two applications, and the first purpose of the public feature layer is to divide the public feature layer into two network branches after carrying out 3*3-dimension convolution once, wherein one network branch is used for generating 9 priori frames for each point; secondly, generating a deviation value of each priori frame and a true annotation frame; then the two groups of features are sequentially transmitted into propasal layers and ROI Pooling layers to respectively carry out preliminary classification of the prior frame and screening of the final suggestion frame; the second purpose of the public feature layer is to input the public feature layer into the ROI Pooling layer, intercept corresponding feature blocks according to the selected suggestion frame, and unify the feature fast size to obtain the final target feature;
(4) Building suggested box coordinates return network branches: constructing a suggestion frame coordinate return network branch by adopting a full-connection layer, and decoding the suggestion frame coordinate return network branch to obtain a coordinate value of a target feature block on an original image;
(5) Constructing a spectrum data extraction module: intercepting corresponding spectrum data according to coordinate values obtained by returning the suggested frame coordinates to the network branches; average calculation is carried out on the intercepted spectrum data on each wave band to obtain average spectrum data of the target in each suggestion frame, and the average spectrum data is used as input of a spectrum characteristic extraction network;
after the pseudo-color image is input into FASTER RCNN network and features are extracted through Resnet, the RPN network is used for judging which areas in the image are cell bodies and obtaining feature blocks with different sizes, and after the image passes through ROI Pooling layers, the feature blocks proposed by the network are converted into uniform sizes; then FASTER RCNN is divided into two branches, one branch is a return network for predicting the coordinates of the positioning frame and is used for learning and returning the four coordinates of the predicted cell positioning frame and the deviation value of the coordinates of the marking positioning frame in the attribute file; the other branch is a classification network of five classifications of cells; the spectrum data is stored in a picture file of a raw type, the file type is that reflectivity values of all pixels in the spectrum picture corresponding to all wave bands are arranged end to form a one-dimensional list, the data is rearranged into a three-dimensional data block according to a format of length multiplied by height multiplied by wave band number of an image, then the corresponding pixel points and the reflectivity values of the corresponding pixel points in all wave bands are cut out by utilizing four coordinate values of a cell prediction coordinate frame obtained based on a pseudo-color image, and the reflectivity average value of all the pixel points is obtained under each wave band to obtain an average spectrum curve which is used as input of a spectrum characteristic network;
(6) Constructing a spectrum feature extraction network: additionally constructing a one-dimensional spectrum data feature extraction network, and outputting one-dimensional spectrum features;
The average spectrum curve of each cell obtained by the spectrum data extraction module is one-dimensional data, so that a spectrum-oriented characteristic extraction network is built by using Conv 1D; the network is built by a Conv1D+ Max Pooling structure, 350-dimensional spectrum data are compressed, after a feature layer is obtained, the method is subjected to the Flatten operation, a one-dimensional full feature layer is obtained, a full-connection Dense layer is connected, and then the full-connection Dense layer is connected with a pseudo-color image feature layer in a FASTER RCNN tail cell classification branch;
(7) Construction of a feature fusion network: the characteristics extracted from the tail end of the spectrum characteristic extraction network are fused with a characteristic layer of FASTER RCNN tail end based on cell pseudo-color image classification, and then the whole classification network is connected, so that the purpose of classification based on fusion characteristics of cell image space and spectrum is realized;
before classifying by using a softmax layer in a FASTER RCNN classification branch, performing a flatten operation on a feature block obtained by an RPN layer to obtain a one-dimensional feature layer, wherein the last layer of a spectrum feature extraction network is a one-dimensional feature layer, splicing the two feature layers according to a weight of 1:1 to obtain a fusion feature layer containing space and spectrum features, and then connecting the softmax classification layer to realize classification based on fusion features.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111282117.XA CN114037671B (en) | 2021-11-01 | 2021-11-01 | Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111282117.XA CN114037671B (en) | 2021-11-01 | 2021-11-01 | Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114037671A CN114037671A (en) | 2022-02-11 |
CN114037671B true CN114037671B (en) | 2024-10-29 |
Family
ID=80142425
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111282117.XA Active CN114037671B (en) | 2021-11-01 | 2021-11-01 | Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114037671B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115791640B (en) * | 2023-02-06 | 2023-06-02 | 杭州华得森生物技术有限公司 | Tumor cell detection equipment and method based on spectroscopic spectrum |
CN116310870A (en) * | 2023-03-16 | 2023-06-23 | 江苏电子信息职业学院 | Object detection classification method, system and storage medium based on hyperspectral image |
CN116563848B (en) * | 2023-07-12 | 2023-11-10 | 北京大学 | Abnormal cell identification method, device, equipment and storage medium |
CN118658011A (en) * | 2024-08-19 | 2024-09-17 | 长春理工大学 | Hyperspectral image classification method and device and computer-readable terminal |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112580670A (en) * | 2020-12-31 | 2021-03-30 | 中国人民解放军国防科技大学 | Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10282589B2 (en) * | 2017-08-29 | 2019-05-07 | Konica Minolta Laboratory U.S.A., Inc. | Method and system for detection and classification of cells using convolutional neural networks |
CN109598224A (en) * | 2018-11-27 | 2019-04-09 | 微医云(杭州)控股有限公司 | Recommend white blood cell detection method in the Sections of Bone Marrow of convolutional neural networks based on region |
CN113343975A (en) * | 2021-04-22 | 2021-09-03 | 山东师范大学 | Deep learning-based white blood cell classification system and method |
CN113239755B (en) * | 2021-04-28 | 2022-06-21 | 湖南大学 | Medical hyperspectral image classification method based on space-spectrum fusion deep learning |
-
2021
- 2021-11-01 CN CN202111282117.XA patent/CN114037671B/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112580670A (en) * | 2020-12-31 | 2021-03-30 | 中国人民解放军国防科技大学 | Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning |
Also Published As
Publication number | Publication date |
---|---|
CN114037671A (en) | 2022-02-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114037671B (en) | Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN | |
Anandhakrishnan et al. | Deep Convolutional Neural Networks for image based tomato leaf disease detection | |
CN103034838B (en) | A kind of special vehicle instrument type identification based on characteristics of image and scaling method | |
US10121245B2 (en) | Identification of inflammation in tissue images | |
CN107315990B (en) | Pedestrian detection algorithm based on XCS-LBP characteristics | |
Shirazi et al. | Extreme learning machine based microscopic red blood cells classification | |
Huang et al. | GCS-YOLOV4-Tiny: A lightweight group convolution network for multi-stage fruit detection | |
CN112464983A (en) | Small sample learning method for apple tree leaf disease image classification | |
CN114170511A (en) | Pavement crack disease identification method based on Cascade RCNN | |
CN111062346A (en) | Automatic leukocyte positioning detection and classification recognition system and method | |
CN112580748A (en) | Method for counting cancer cells of Ki67 stained image | |
CN114299324A (en) | Pathological image classification method and system based on multi-scale domain confrontation network | |
Yildiz et al. | Hybrid image improving and CNN (HIICNN) stacking ensemble method for traffic sign recognition | |
CN110910497B (en) | Method and system for realizing augmented reality map | |
CN116468935A (en) | Multi-core convolutional network-based stepwise classification and identification method for traffic signs | |
CN115775226A (en) | Transformer-based medical image classification method | |
CN115471838A (en) | Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction | |
CN114972202A (en) | Ki67 pathological cell rapid detection and counting method based on lightweight neural network | |
CN114037868A (en) | Image recognition model generation method and device | |
Karthik et al. | GrapeLeafNet: A Dual-Track Feature Fusion Network with Inception-ResNet and Shuffle-Transformer for Accurate Grape Leaf Disease Identification | |
US12051254B2 (en) | Typing biological cells | |
CN116704241A (en) | Full-channel 3D convolutional neural network hyperspectral remote sensing image classification method | |
CN112949634B (en) | Railway contact net nest detection method | |
Priya et al. | Detection of leukemia in blood microscopic images using fuzzy logic | |
CN115661042A (en) | Hierarchical classification defect detection method based on attention mechanism guidance |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant |