CN114037671B - Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN - Google Patents

Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN Download PDF

Info

Publication number
CN114037671B
CN114037671B CN202111282117.XA CN202111282117A CN114037671B CN 114037671 B CN114037671 B CN 114037671B CN 202111282117 A CN202111282117 A CN 202111282117A CN 114037671 B CN114037671 B CN 114037671B
Authority
CN
China
Prior art keywords
spectrum
network
image
feature
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111282117.XA
Other languages
Chinese (zh)
Other versions
CN114037671A (en
Inventor
李良军
曾凡一
黄杰
苏颖
石爽
蔡虹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Second Hospital of Dalian Medical University
Original Assignee
Second Hospital of Dalian Medical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Second Hospital of Dalian Medical University filed Critical Second Hospital of Dalian Medical University
Priority to CN202111282117.XA priority Critical patent/CN114037671B/en
Publication of CN114037671A publication Critical patent/CN114037671A/en
Application granted granted Critical
Publication of CN114037671B publication Critical patent/CN114037671B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/17Systems in which incident light is modified in accordance with the properties of the material investigated
    • G01N21/25Colour; Spectral properties, i.e. comparison of effect of material on the light at two or more different wavelengths or wavelength bands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10036Multispectral image; Hyperspectral image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10056Microscopic image
    • G06T2207/10061Microscopic image from scanning electron microscope
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30024Cell structures in vitro; Tissue sections in vitro
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Chemical & Material Sciences (AREA)
  • Analytical Chemistry (AREA)
  • Biochemistry (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Investigating Or Analysing Biological Materials (AREA)
  • Image Processing (AREA)

Abstract

The invention belongs to the technical field of medical treatment, and provides a white blood cell detection method based on an improved FASTER RCNN microscopic hyperspectral image, which uses a hyperspectral microscope to obtain a white blood cell detection image of a blood smear and carries out identification and classification on white blood cells based on an improved FASTER RCNN. And (3) obtaining pseudo-color images and spectrum data of the blood smear by using a hyperspectral microscope, marking different types of white blood cells, and manufacturing a data set. The traditional FASTER RCNN network is improved, namely the VGG16 in the FASTER RCNN original network is replaced by Resnet and is used as a new pseudo-color image feature extraction network; and establishing a spectrum data extraction module aiming at blood smear hyperspectral data, extracting spectrum features by utilizing a one-dimensional convolutional neural network, and fusing white blood cell image features and spectrum features by utilizing a FASTER RCNN network on the basis of the improvement, so as to finally realize white blood cell identification and classification. Compared with the traditional FASTER RCNN network, the invention has obvious improvement on the identification precision and classification accuracy of the white blood cells.

Description

Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN
Technical Field
The invention belongs to the technical field of medical treatment, and particularly relates to a microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN.
Background
White blood cells are an important component of blood, produced by bone marrow and lymphoid tissue, and function to combat viral and bacterial infections. Can be classified into neutrophils, lymphocytes, monocytes, eosinophils, and basophils. Traditional manual microscopic detection methods are overly complex, and manual differential counting of leukocytes is more prone to error, particularly when detecting large numbers of samples. In recent years, corresponding computer vision algorithms and systems have been widely used in the field of automatic classification and detection of blood cells.
Computer vision inspection can be divided into two major categories at present, the first category is to simply stack traditional algorithm modules according to the order of preprocessing, feature extraction and classification. Rawat et al propose a new technique for distinguishing acute lymphoblastic leukemia (Acute Lymphoblastic Leukemia) from healthy lymphocytes, using an SVM classifier to classify fusion features including texture and shape with an accuracy of 89.8%. Mohapatra et al adopts a two-stage color segmentation strategy based on fuzzy clustering to segment white blood cells in an image, and proposes two new shape features of Hausdorff dimension and outline feature, and uses SVM to classify lymphatic nuclei, so that the accuracy reaches 95%. Alferez et al utilized color component clustering and watershed transformation to segment peripheral blood cell images, and extracted 113 features altogether for different types of lymphocyte recognition. The accuracy of the training set is 98.07%, and the accuracy of the verification set is 85.33%.
However, most of the conventional pattern recognition algorithms are used for sequentially completing the representation and classification modules, the super parameters of the conventional pattern recognition algorithms are usually given by experience before training, and the generalization capability and stability of the model cannot be ensured when the sample space is small or the inter-class gap is not obvious.
The second category is to analyze pathological pictures based on a deep learning algorithm, and the computer can extract multidimensional features more comprehensively and automatically without manually designating the extraction of a certain feature, so that after a large amount of training, the targets in the images can be identified and classified more accurately, efficiently and stably, and strong strength is shown in a plurality of fields, particularly in the medical field with a large amount of image analysis and processing, and the method has the advantage incomparable with the traditional machine learning algorithm. Boldu et AL in the study of acute leukemia detection, a dual-module deep learning architecture (AL Net) was used to identify and classify abnormal promyelocytic and blast cells, respectively, and the accuracy of identification of myeloid leukemia and lymphoid leukemia reached 93.7% and 100%, respectively. SHAKARAMI et al used a fast and efficient YOLOv (FED) detector with EFFICIENTNET convolutional neural network as the backbone, enabling blood cell identification based on BCCD datasets. Finally, the classification accuracy of platelets, erythrocytes and leukocytes was 90.25%, 80.41% and 98.92%, respectively.
In the published patent, it is referred to the detection of cells using FASTER RCNN networks as follows:
Patent CN109598224a proposes a method for detecting white blood cells in bone marrow section based on region recommended convolution, which uses Resnet and FPN as a characteristic extraction network of FASTER RCNN, and bone marrow section image as input, and the accuracy can reach more than 90%. Patent CN110059672a proposes a method for performing class-increasing learning on a microscope cell image detection model by using incremental learning, calculates the distance between an intermediate feature layer obtained by predicting new class cells by using a FASTER RCNN model trained by an old class cell image and an intermediate feature layer of an incremental model, and obtains a new class-increasing prediction model. Patent CN110580699a proposes a pathological image cell nucleus detection method based on an improved FASTER RCNN algorithm, uses ZF as a FASTER RCNN feature extraction network, processes sample pictures by data enhancement transformation, difficult sample mining and small target detection optimization, achieves a monitoring speed of 1 s/sheet, and has good robustness. Patent CN111598849a proposes a pathological image cell counting method, device and medium based on target detection, which refers to two different FASTER RCNN models trained by bockbone to perform cell counting detection on pathological images, and fuses the detection results to accurately, rapidly and efficiently realize pathological cell counting.
However, these methods are only based on gray scale or RGB images with cellular space characteristics, and are susceptible to various environmental factors such as optical conditions of a microscope, thickness of a slide, etc. The hyperspectral imaging technology is a combination of the traditional imaging technology and the spectrum technology, so that not only the spatial characteristics of a monitored target but also the spectrum characteristics of the monitored target can be extracted. In the medical field, it has been used for the identification and diagnosis of acute lymphoblastic leukemia, food borne pathogens and cancer. Although hyperspectral imaging provides rich spectral information on the basis of spatial features, hundreds of narrow continuous bands make the data dimension redundant, requiring manual extraction of spectral values of a region of interest (ROI), which is extremely cumbersome in practical monitoring. Therefore, there is a strong need for an automatic, fast and efficient method of white blood cell detection based on hyperspectral imaging.
According to the invention, the original VGG16 characteristic extraction network FASTER RCNN is replaced by Resnet, except for stacking on a simple convolution layer, the cross arrangement of two basic blocks of Conv Block and Identity Block in the network can avoid losing important characteristics due to mismatch between the number of convolution layers and data, and the generalization of the network is far superior to that of VGG16. Besides the replacement of the feature extraction network, the invention is different from the original FASTER RCNN network frame in that the biggest innovation point of the patents and researches is that a complete set of spectrum feature extraction network and feature fusion network are constructed and embedded into FASTER RCNN, thereby realizing the novel automatic white blood cell identification and classification technology based on the combined detection of space and spectrum features.
Disclosure of Invention
In order to solve the technical problems, the invention provides a microscopic hyperspectral leukocyte detection method based on an improved FASTER RCNN algorithm, which adds a detection network branch of spectrum data in a traditional FASTER RCNN network architecture, realizes the joint detection of the spatial characteristics and the spectral characteristics of blood images, and effectively improves the classification precision.
The technical scheme of the invention is as follows:
the microscopic hyperspectral leukocyte detection method based on the improvement FASTER RCNN comprises the following steps:
1. Establishing a leukocyte dataset: collecting a pseudo-color image (an image synthesized by 667.4,557.2 and 440.5 wave bands) and a spectrum image of a blood smear by using a microscopic hyperspectral system, correcting the color of the pseudo-color image, marking five types of white blood cells by frame selection and category, correcting the black and white of the spectrum image, and manufacturing a training and verification data set;
2. And (3) constructing a pseudo-color image feature extraction network: resnet18 which can extract more effective features is adopted as a feature extraction network of FASTER RCNN;
3. Constructing an RPN proposal network, which is responsible for automatically generating an initial target priori frame, carrying out preliminary classification on the priori frame, and then finally screening according to the priori frame to obtain a proposal frame, and intercepting corresponding features at a public feature layer;
4. And constructing a suggested frame coordinate return branch, constructing a suggested frame coordinate return network branch by adopting a full connection layer (Dense), and decoding the suggested frame coordinate return network branch to obtain coordinate values of the target feature on the original image.
5. And (3) constructing a spectrum data extraction module: using FASTER RCNN to detect four coordinate values obtained by a network terminal positioning frame coordinate branch as input of a spectrum extraction module, intercepting full-band data of a corresponding region in a spectrum image according to the coordinate, and then calculating an average spectrum value of the full-band data to obtain an average spectrum curve as input of a spectrum feature extraction network;
6. Constructing a spectrum feature extraction network, additionally constructing a one-dimensional spectrum data feature extraction network, and outputting one-dimensional spectrum features;
7. Construction of a feature fusion network: the characteristics extracted from the tail end of the spectrum characteristic extraction network are fused with a FASTER RCNN tail end of a characteristic layer based on cell pseudo-color image classification;
8. Construction of cell sorting layer: calculating the obtained one-dimensional fusion characteristics by using the full-connection layer and the softmax classification layer to obtain a final cell classification result;
Further, in step (1), the hyperspectral camera cannot obtain a conventional RGB image due to the limitation of the band range, so that the RGB image, i.e. the pseudo-color image, is synthesized by extracting 667.4,557.2 and 440.5 three bands representing R, G, and B channels, respectively. The Gamma correction is used for correcting the color of the pseudo-color image, so that the influence of various environmental factors, such as the optical condition of a microscope, the thickness of a slide and the like, on the brightness of images in different batches is reduced. Black and white correction of spectral images specifically refers to: collecting a pair of blank slide spectral images as white reference images for correction, namely, defining the spectral reflectivity of a blank slide as the highest value; and covering a lens cover of the spectrum camera, closing a light source, and collecting a spectrum image in a pure black environment as a corrected black reference image, namely, defining the spectrum reflectivity in the dark environment as the minimum value. And then carrying out normalization processing on the data of all the spectral images based on the black-white correction data to obtain corrected spectral images. Each white blood cell within the image field of view is framed and based on the corresponding class label, the final dataset includes for each image: a pseudo color image; a spectral image; a property file (containing the coordinates of the positioning frame and the classification label of the white blood cells contained in the image).
Further, in the step (2), the feature extraction network Resnet of the pseudo-color image is specifically built by using a small convolution kernel of 3*3, two basic blocks of Conv Block and Identity Block in the network are arranged in a crossing manner, the Conv Block is used for compressing the image, the Identity Block is not used for changing the size of the image, and the key point is that the two feature extraction modules combine the extracted features with the original features, thereby not only ensuring the extraction of important features, but also ensuring that the original features are not missed due to deepening of the layer number. The spectrum feature extraction network is built in the form of Conv1D+ MaxPooling D layer based on 1*3 convolution kernels.
Further, in step (3), the common feature obtained by the pseudo-color image after passing through the feature extraction network has two application paths, one of which is to sweep it through a 3*3-size convolution to obtain a feature of a deeper layer, and then to divide it into two network branches, one of which is used to generate 9 prior frames for each point. And secondly, generating a deviation value of each priori frame and the true annotation frame. These two sets of features are then passed sequentially into propasal and ROI Pooling layers for preliminary classification of the prior frames and screening of the final suggested frames, respectively. The second purpose of the public feature layer is to input the public feature layer into the ROI Pooling layer, intercept corresponding feature blocks according to the screened suggestion boxes, unify the feature fast sizes and obtain final target features.
Further, in step (4) and step (5), after the pseudo color image is input to FASTER RCNN network and features are extracted by Resnet, the pseudo color image passes through RPN network, the network determines those areas in the image are cell bodies, and feature blocks with different sizes are obtained, and after ROI Pooling layers, the feature blocks proposed by the network are converted into uniform sizes. Then FASTER RCNN is divided into two branches, one branch is a return network for predicting the coordinates of the positioning frame and is used for learning and returning the four coordinates of the predicted cell positioning frame and the deviation value of the coordinates of the labeling positioning frame in the attribute file, and the other branch is a classification network for five classifications of cells. The spectrum data is stored in a picture file of a raw type, the file type is that reflectivity values of all pixels in the spectrum picture corresponding to all wave bands are arranged end to form a one-dimensional list, the one-dimensional list is rearranged into a three-dimensional data block according to a format of length multiplied by height multiplied by wave band number of an image, then the corresponding pixel points and the reflectivity values of the corresponding pixel points in all wave bands are cut out by utilizing four coordinate values of a cell prediction coordinate frame obtained based on a pseudo-color image, and the reflectivity average value of all the pixel points is obtained under each wave band to obtain an average spectrum curve which is used as input of a spectrum characteristic network.
Further, in the step (6), the average spectrum curve of each cell obtained by the spectrum data extraction module is one-dimensional data, so that a spectrum-oriented feature extraction network is built by using Conv 1D. The network is built by a Conv1D+ Max Pooling structure, 350-dimensional spectrum data are compressed, after a feature layer is obtained, the Flatten operation is carried out, a one-dimensional full feature layer is obtained, a full-connection Dense layer is connected, and then the layer is connected with a pseudo-color image feature layer in a FASTER RCNN tail cell classification branch.
Further, in the step (7), before the classification of FASTER RCNN is performed by using the softmax layer, a plurality of feature blocks obtained by the RPN layer are subjected to a flat operation to be changed into a one-dimensional feature layer, the last layer of the spectrum feature extraction network is also a one-dimensional feature layer, and the two feature layers are spliced according to the weight of 1:1, so that a fusion feature layer comprising space and spectrum features is obtained.
Further, in the step (8), after a one-dimensional fusion feature layer is obtained, the features are transferred into a full-connection layer and a classification layer of softmax, so that classification based on the fusion features is realized.
Compared with the prior art, the invention provides a microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN, and particularly creates innovation in the aspect of feature Fusion detection of multi-sensor data, and provides a multi-class data feature Fusion detection method, namely a multi-feature Fusion network Fusion-Net.
Drawings
FIG. 1 is a schematic block diagram of the present invention;
FIG. 2A is a microscopic hyperspectral image of class 1 (neutrophils) and class 5 (monocytes) cells;
FIG. 2B is a microscopic hyperspectral image of class 1 (neutrophils) and class 3 (basophils) cells;
FIG. 2C is a microscopic hyperspectral image of class 1 (neutrophils), class 3 (basophils) and class 4 (lymphocytes) cells;
FIG. 2D is a microscopic hyperspectral image of class 1 (neutrophils) and class 2 (eosinophils) cells;
FIG. 3Resnet is a schematic diagram of a feature extraction network; (a) is Conv Block; (b) is Identity Block; (c) is Resnet < 18 >;
FIG. 4 is a spectral feature extraction network;
FIG. 5 is a schematic diagram of a network structure of FASTER RCNN after modification;
FIG. 6A is a graph showing the results of conventional FASTER RCNN-class 1 (neutrophils) and 5-class 5 (monocytes) cell detection classification;
FIG. 6B is a graph showing the results of conventional FASTER RCNN-based classification of class 1 (neutrophils) and class 3 (basophils) cells;
FIG. 6C is a graph showing the results of conventional FASTER RCNN-based classification of class 1 (neutrophils), class 3 (basophils) and class 4 (lymphocytes) cells;
FIG. 6D is a graph showing the results of conventional FASTER RCNN-based classification of class 1 (neutrophils) and class 2 (eosinophils) cells;
FIG. 7A is a graph showing the results of the detection classification of FASTER RCNN cells of class 1 (neutrophils) and class 5 (monocytes) after modification of the present invention;
FIG. 7B is a graph showing the results of the detection classification of FASTER RCNN versus class 1 (neutrophils) and class 3 (basophils) cells after modification of the present invention;
FIG. 7C is a graph showing the results of the detection classification of FASTER RCNN cells of class 1 (neutrophils), class 3 (basophils) and class 4 (lymphocytes) after modification of the invention;
FIG. 7D is a graph showing the results of the detection classification of FASTER RCNN versus class 1 (neutrophils) and class 2 (eosinophils) cells after modification of the present invention.
Detailed Description
The following describes the embodiments of the present invention further with reference to the drawings and technical schemes.
As shown in fig. 1, a microscopic hyperspectral leukocyte detection method based on the improvement FASTER RCNN comprises the following specific implementation steps:
s1, establishing a leukocyte data set:
the hyperspectral camera is arranged on a three-eye student microscope, and the wave band used by the hyperspectral camera is a near infrared short wave light region between 382nm and 1020 nm. The method comprises the steps of finding a visual field containing white blood cells through an eyepiece, enabling a detection target in the visual field to be clearly visible through focusing and other operations, and then collecting images and spectrum information in the visual field by using a hyperspectral camera.
And carrying out color correction on the acquired pseudo-color image. And Gamma correction is adopted, so that the influence of various environmental factors, such as the optical condition of a microscope, the thickness of a slide and the like, on the brightness of images of different batches is reduced. The gamma correction is a method of editing a gamma curve of an image to perform nonlinear tone editing of the image, and increases the ratio of dark color portions to light color portions in an image signal to thereby improve the image contrast effect.
And performing box selection and category labeling on the five types of white blood cells. And using marking software to carry out frame selection on each cell on each image, giving corresponding classification labels, and generating an attribute file containing four coordinates of the positioning frame and the corresponding labels. The principle of the frame selection is that background elements of non-leucocytes appear in the frame as little as possible, and when two leucocytes are overlapped, the content of non-own cell elements in the frame is reduced as much as possible.
Black and white correction is performed on the spectral image. To reduce the sample position differences and the effect of the light source on the image, the original hyperspectral image is corrected to the reflectance mode. Collecting a pair of blank slide spectral images as white reference images for correction, namely, defining the spectral reflectivity of a blank plectrum to be the highest value (the reflectivity is close to 100%); the lens cover of the spectrum camera is covered, the light source is turned off, and a spectrum image under a pure black environment is collected and used as a corrected black reference image, namely, the spectrum reflectivity under the dark environment is regulated to be the minimum value (the reflectivity is close to 0%). And then carrying out normalization processing on the data of all the spectral images based on the black-white correction data to obtain corrected spectral images. The correction process is as follows:
R=(I-B)/(W-B)×100%
Wherein I represents an original image; w represents a full white calibration image; b represents a full black calibration image; r represents the corrected relative image.
For each image, the final dataset includes: a pseudo color image; a spectral image; a property file (containing the coordinates of the positioning frame and the classification label of the white blood cells contained in the image). The image of five types of white blood cells collected is shown in fig. 2.
S2, pseudo color image feature extraction network:
Resnet18 was employed as the feature extraction network for FASTER RCNN. As shown in fig. 3, resnet has two basic blocks, namely Conv Block and Identity Block, wherein the dimensions of Conv Block input and output are different, so that the Conv Block cannot be serially connected, and the effect of the Conv Block is to change the dimension of a network; the Identity Block input dimension and the output dimension are the same and can be connected in series to deepen the network. The pseudo-color image feature extraction network of the improvement FASTERRCNN only comprises four image size compressions, and the content after the fifth compression is used after ROI Pooling layers. The specific structure is shown in the Resnet block diagram in fig. 3, and after the 600×600×3 picture at the time of input is sent to Resnet, a feature block with a size of 38×38×256 is obtained.
S3.rpn proposed network:
As shown in the network structure schematic diagram of FASTER RCNN after improvement in fig. 5, the public feature obtained by the pseudo-color image after passing through the feature extraction network has two application ways, one of which is to sweep it through one 3*3-sized convolution to obtain a feature of a deeper layer, and then to divide it into two network branches, one of which is to sweep it through 9-channel convolution with a size of 1*1, that is, the number of prior frames generated by each pixel point in FASTER RCNN is 9. The second is to sweep using 36-channel convolution of size 1*1, i.e., each a priori frame contains 4 coordinate variables, so 9*4 =36 channels are produced per pixel. Both sets of features are input to proposal of FASTER RCNN, where the objects in each prior box are initially classified, i.e., whether the object is included or not is determined, and then the object to be detected is found in the prior box determined to be an object. At this time, the feature sizes of the extracted targets are different, and then the sizes of the prior frames are unified at ROI Pooling layers to obtain the final suggestion frame. Simultaneously, the common features extracted from the previous feature network are also input into the ROI Pooling layers, the features of 14 x 14 are cut out from the common features according to the suggested frame coordinates after the unified size, and then the features of each suggested frame are subjected to the fifth compression of the original Resnet to obtain the features of 7*7. So far, the whole RPN network is ended, and a plurality of characteristic blocks which are judged to be detection targets by the RPN network are obtained.
S4-S5, suggesting a frame coordinate return branch and a spectrum data extraction module:
After obtaining a number 7*7 of feature blocks through the RPN network, the flag operation is performed to become one-dimensional data, and then there are two branches, namely a return branch of the coordinates of the suggestion frame, which is constructed by a full connection layer, as shown in fig. 5, and is used for adjusting the corresponding suggestion frame, and 20 (the number of classification×4) neurons are included, that is, each cell includes 4 coordinate parameters no matter which class is divided into. Finally, after decoding the layer of data on the original image, coordinate values of four vertexes can be obtained, one of the purposes is to draw a recognition frame on the original image, and the other is to use four coordinate values obtained by suggesting frame coordinate branches as input of a spectrum extraction module. The spectrum data extraction module is written by using OpenCV and mainly comprises three parts: reading and arranging spectrum data in a raw file; intercepting a corresponding data block according to the coordinate value of the suggestion frame; and carrying out average calculation on the intercepted spectrum data on each wave band to obtain average spectrum data of the target in each suggestion frame, wherein the average spectrum data is used as input of a spectrum characteristic extraction network.
S6, spectral feature extraction network:
the average spectrum curve of each cell obtained through the spectrum data extraction module is one-dimensional data, so that a spectrum-oriented characteristic extraction network is built by using Conv 1D. A specific network structure is shown in fig. 4. The network is built by a Conv1D+ Max Pooling structure, 350-dimensional spectrum data are compressed, after a characteristic layer of 1 x 173 x 64 is obtained, a flat operation is carried out, a full characteristic layer of 1 x 5568 is obtained, a full connection Dense layer is connected, and then the layer is connected with a pseudo-color image characteristic layer in a FASTER RCNN tail cell classification branch.
S7, feature fusion network:
As shown in fig. 5, before classifying by softmax in the classification branch of FASTER RCNN, the features obtained by the RPN layer are subjected to the flatten operation to become a one-dimensional feature layer, the last layer of the spectral feature extraction network is also a one-dimensional feature layer, and the two feature layers are spliced according to a weight of 1:1, so that a fused feature layer comprising spatial and spectral features is obtained.
S8, cell classification network:
after a one-dimensional fusion feature layer is obtained, the features are transferred into a full-connection layer and a classification layer of softmax, so that classification based on the fusion features is realized.
The above is all the steps of the flow of the invention, and during training, the data set is randomly divided into a training set and a testing set according to the ratio of 7:3. Meanwhile, a conventional FASTER RCNN network which is characterized by VGG16 extraction and is trained only based on pseudo-color images is used as a control group, the detection result is shown in fig. 6, detection confusion can occur on two white blood cells which are closer to each other, two white blood cells are used as one cell for framing, and due to interference of a large number of red blood cells on a blood picture, part of white blood cells cannot be identified, besides the identification accuracy, the conventional FASTER RCNN network classifies all the identified white blood cells into one type, and the fact that the conventional FASTER RCNN network is poor in identification and classification accuracy is reflected in the condition that a data set is smaller. Fig. 7 shows the improved FASTER RCNN network detection result based on the combined detection of spatial and spectral features, and the same test image, each white blood cell is accurately identified and positioned, which shows the advantage of Resnet in feature extraction, and each white blood cell is accurately classified into 5 classes, which is because the white blood cells have obvious advantages in the inter-class distinction of white blood cells under the conditions that the difference between the white blood cells on the image is small and the data set is not huge enough.

Claims (1)

1. The microscopic hyperspectral leukocyte detection method based on the improvement FASTER RCNN is characterized by comprising the following steps:
(1) Establishing a leukocyte dataset: collecting a pseudo-color image and a spectrum image of a blood smear by using a microscopic hyperspectral system, carrying out color correction on the pseudo-color image, carrying out frame selection and category marking on five types of white blood cells, carrying out black-and-white correction on the spectrum image, and manufacturing a training data set and a verification data set;
The pseudo-color image is synthesized by images corresponding to three wavebands 667.4, 557.2 and 440.5, the three wavebands correspond to R, G and B channels of the image respectively, and Gamma is used for correcting the color of the pseudo-color image;
The black-and-white correction of the spectrum image specifically refers to: collecting a pair of blank slide spectral images as white reference images for correction, namely, defining the spectral reflectivity of a blank slide as the highest value; covering a lens cover of a spectrum camera, closing a light source, and collecting a spectrum image in a pure black environment as a corrected black reference image, namely, defining the spectrum reflectivity in the dark environment as the minimum value; then, carrying out normalization processing on the data of all the spectrum images based on the white reference image and the black reference image to obtain corrected spectrum images; each white blood cell in the image view is framed and selected based on the corresponding classification label; for each image, the final dataset comprises a pseudo-color image, a spectral image, a property file containing the location frame coordinates and classification tags of the white blood cells contained in the image;
(2) Constructing a pseudo color image feature extraction network: resnet18 is adopted as a feature extraction network of FASTER RCNN;
Resnet 18 the construction is carried out by using a 3*3 small convolution kernel, a characteristic extraction module in the network is a Conv Block and an Identity Block which are arranged in a crossing way, the Conv Block is used for compressing an image, the Identity Block cannot change the size of the image, and the two characteristic extraction modules combine the extracted characteristics with original characteristics to obtain a common characteristic layer;
The characteristic extraction network of the spectrum image is built in a form of Conv1D+ MaxPooling D layer based on 1*3 convolution kernels;
(3) Constructing an RPN proposal network, which is responsible for automatically generating an initial target priori frame, carrying out preliminary classification on the target priori frame, and then finally screening according to the target priori frame to obtain a proposal frame, and intercepting corresponding features at a public feature layer;
The public feature layer obtained after the pseudo-color image passes through the image feature extraction network has two applications, and the first purpose of the public feature layer is to divide the public feature layer into two network branches after carrying out 3*3-dimension convolution once, wherein one network branch is used for generating 9 priori frames for each point; secondly, generating a deviation value of each priori frame and a true annotation frame; then the two groups of features are sequentially transmitted into propasal layers and ROI Pooling layers to respectively carry out preliminary classification of the prior frame and screening of the final suggestion frame; the second purpose of the public feature layer is to input the public feature layer into the ROI Pooling layer, intercept corresponding feature blocks according to the selected suggestion frame, and unify the feature fast size to obtain the final target feature;
(4) Building suggested box coordinates return network branches: constructing a suggestion frame coordinate return network branch by adopting a full-connection layer, and decoding the suggestion frame coordinate return network branch to obtain a coordinate value of a target feature block on an original image;
(5) Constructing a spectrum data extraction module: intercepting corresponding spectrum data according to coordinate values obtained by returning the suggested frame coordinates to the network branches; average calculation is carried out on the intercepted spectrum data on each wave band to obtain average spectrum data of the target in each suggestion frame, and the average spectrum data is used as input of a spectrum characteristic extraction network;
after the pseudo-color image is input into FASTER RCNN network and features are extracted through Resnet, the RPN network is used for judging which areas in the image are cell bodies and obtaining feature blocks with different sizes, and after the image passes through ROI Pooling layers, the feature blocks proposed by the network are converted into uniform sizes; then FASTER RCNN is divided into two branches, one branch is a return network for predicting the coordinates of the positioning frame and is used for learning and returning the four coordinates of the predicted cell positioning frame and the deviation value of the coordinates of the marking positioning frame in the attribute file; the other branch is a classification network of five classifications of cells; the spectrum data is stored in a picture file of a raw type, the file type is that reflectivity values of all pixels in the spectrum picture corresponding to all wave bands are arranged end to form a one-dimensional list, the data is rearranged into a three-dimensional data block according to a format of length multiplied by height multiplied by wave band number of an image, then the corresponding pixel points and the reflectivity values of the corresponding pixel points in all wave bands are cut out by utilizing four coordinate values of a cell prediction coordinate frame obtained based on a pseudo-color image, and the reflectivity average value of all the pixel points is obtained under each wave band to obtain an average spectrum curve which is used as input of a spectrum characteristic network;
(6) Constructing a spectrum feature extraction network: additionally constructing a one-dimensional spectrum data feature extraction network, and outputting one-dimensional spectrum features;
The average spectrum curve of each cell obtained by the spectrum data extraction module is one-dimensional data, so that a spectrum-oriented characteristic extraction network is built by using Conv 1D; the network is built by a Conv1D+ Max Pooling structure, 350-dimensional spectrum data are compressed, after a feature layer is obtained, the method is subjected to the Flatten operation, a one-dimensional full feature layer is obtained, a full-connection Dense layer is connected, and then the full-connection Dense layer is connected with a pseudo-color image feature layer in a FASTER RCNN tail cell classification branch;
(7) Construction of a feature fusion network: the characteristics extracted from the tail end of the spectrum characteristic extraction network are fused with a characteristic layer of FASTER RCNN tail end based on cell pseudo-color image classification, and then the whole classification network is connected, so that the purpose of classification based on fusion characteristics of cell image space and spectrum is realized;
before classifying by using a softmax layer in a FASTER RCNN classification branch, performing a flatten operation on a feature block obtained by an RPN layer to obtain a one-dimensional feature layer, wherein the last layer of a spectrum feature extraction network is a one-dimensional feature layer, splicing the two feature layers according to a weight of 1:1 to obtain a fusion feature layer containing space and spectrum features, and then connecting the softmax classification layer to realize classification based on fusion features.
CN202111282117.XA 2021-11-01 2021-11-01 Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN Active CN114037671B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111282117.XA CN114037671B (en) 2021-11-01 2021-11-01 Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111282117.XA CN114037671B (en) 2021-11-01 2021-11-01 Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN

Publications (2)

Publication Number Publication Date
CN114037671A CN114037671A (en) 2022-02-11
CN114037671B true CN114037671B (en) 2024-10-29

Family

ID=80142425

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111282117.XA Active CN114037671B (en) 2021-11-01 2021-11-01 Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN

Country Status (1)

Country Link
CN (1) CN114037671B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115791640B (en) * 2023-02-06 2023-06-02 杭州华得森生物技术有限公司 Tumor cell detection equipment and method based on spectroscopic spectrum
CN116310870A (en) * 2023-03-16 2023-06-23 江苏电子信息职业学院 Object detection classification method, system and storage medium based on hyperspectral image
CN116563848B (en) * 2023-07-12 2023-11-10 北京大学 Abnormal cell identification method, device, equipment and storage medium
CN118658011A (en) * 2024-08-19 2024-09-17 长春理工大学 Hyperspectral image classification method and device and computer-readable terminal

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112580670A (en) * 2020-12-31 2021-03-30 中国人民解放军国防科技大学 Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10282589B2 (en) * 2017-08-29 2019-05-07 Konica Minolta Laboratory U.S.A., Inc. Method and system for detection and classification of cells using convolutional neural networks
CN109598224A (en) * 2018-11-27 2019-04-09 微医云(杭州)控股有限公司 Recommend white blood cell detection method in the Sections of Bone Marrow of convolutional neural networks based on region
CN113343975A (en) * 2021-04-22 2021-09-03 山东师范大学 Deep learning-based white blood cell classification system and method
CN113239755B (en) * 2021-04-28 2022-06-21 湖南大学 Medical hyperspectral image classification method based on space-spectrum fusion deep learning

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112580670A (en) * 2020-12-31 2021-03-30 中国人民解放军国防科技大学 Hyperspectral-spatial-spectral combined feature extraction method based on transfer learning

Also Published As

Publication number Publication date
CN114037671A (en) 2022-02-11

Similar Documents

Publication Publication Date Title
CN114037671B (en) Microscopic hyperspectral leukocyte detection method based on improvement FASTER RCNN
Anandhakrishnan et al. Deep Convolutional Neural Networks for image based tomato leaf disease detection
CN103034838B (en) A kind of special vehicle instrument type identification based on characteristics of image and scaling method
US10121245B2 (en) Identification of inflammation in tissue images
CN107315990B (en) Pedestrian detection algorithm based on XCS-LBP characteristics
Shirazi et al. Extreme learning machine based microscopic red blood cells classification
Huang et al. GCS-YOLOV4-Tiny: A lightweight group convolution network for multi-stage fruit detection
CN112464983A (en) Small sample learning method for apple tree leaf disease image classification
CN114170511A (en) Pavement crack disease identification method based on Cascade RCNN
CN111062346A (en) Automatic leukocyte positioning detection and classification recognition system and method
CN112580748A (en) Method for counting cancer cells of Ki67 stained image
CN114299324A (en) Pathological image classification method and system based on multi-scale domain confrontation network
Yildiz et al. Hybrid image improving and CNN (HIICNN) stacking ensemble method for traffic sign recognition
CN110910497B (en) Method and system for realizing augmented reality map
CN116468935A (en) Multi-core convolutional network-based stepwise classification and identification method for traffic signs
CN115775226A (en) Transformer-based medical image classification method
CN115471838A (en) Cervical squamous lesion cell detection method based on depth self-adaptive feature extraction
CN114972202A (en) Ki67 pathological cell rapid detection and counting method based on lightweight neural network
CN114037868A (en) Image recognition model generation method and device
Karthik et al. GrapeLeafNet: A Dual-Track Feature Fusion Network with Inception-ResNet and Shuffle-Transformer for Accurate Grape Leaf Disease Identification
US12051254B2 (en) Typing biological cells
CN116704241A (en) Full-channel 3D convolutional neural network hyperspectral remote sensing image classification method
CN112949634B (en) Railway contact net nest detection method
Priya et al. Detection of leukemia in blood microscopic images using fuzzy logic
CN115661042A (en) Hierarchical classification defect detection method based on attention mechanism guidance

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant