patents.google.com

CN113887503A - A five-class classification method of white blood cells based on an improved attention convolutional neural network - Google Patents

  • ️Tue Jan 04 2022
A five-class classification method of white blood cells based on an improved attention convolutional neural network Download PDF

Info

Publication number
CN113887503A
CN113887503A CN202111231648.6A CN202111231648A CN113887503A CN 113887503 A CN113887503 A CN 113887503A CN 202111231648 A CN202111231648 A CN 202111231648A CN 113887503 A CN113887503 A CN 113887503A Authority
CN
China
Prior art keywords
attention
output
network
white blood
neural network
Prior art date
2021-10-22
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111231648.6A
Other languages
Chinese (zh)
Other versions
CN113887503B (en
Inventor
王慧慧
邵卫东
张旭
曾凡一
康家铭
张春旭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian Polytechnic University
Original Assignee
Dalian Polytechnic University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
2021-10-22
Filing date
2021-10-22
Publication date
2022-01-04
2021-10-22 Application filed by Dalian Polytechnic University filed Critical Dalian Polytechnic University
2021-10-22 Priority to CN202111231648.6A priority Critical patent/CN113887503B/en
2022-01-04 Publication of CN113887503A publication Critical patent/CN113887503A/en
2022-06-14 Application granted granted Critical
2022-06-14 Publication of CN113887503B publication Critical patent/CN113887503B/en
Status Active legal-status Critical Current
2041-10-22 Anticipated expiration legal-status Critical

Links

  • 210000000265 leukocyte Anatomy 0.000 title claims abstract description 63
  • 238000000034 method Methods 0.000 title claims abstract description 33
  • 238000013527 convolutional neural network Methods 0.000 title claims abstract description 14
  • 238000012549 training Methods 0.000 claims description 26
  • 230000007246 mechanism Effects 0.000 claims description 16
  • 230000008569 process Effects 0.000 claims description 16
  • 210000004369 blood Anatomy 0.000 claims description 9
  • 239000008280 blood Substances 0.000 claims description 9
  • 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 8
  • 238000012360 testing method Methods 0.000 claims description 8
  • 230000006870 function Effects 0.000 claims description 7
  • 210000003651 basophil Anatomy 0.000 claims description 3
  • 210000004027 cell Anatomy 0.000 claims description 3
  • 210000003979 eosinophil Anatomy 0.000 claims description 3
  • 210000004698 lymphocyte Anatomy 0.000 claims description 3
  • 210000000440 neutrophil Anatomy 0.000 claims description 3
  • 238000011176 pooling Methods 0.000 claims description 3
  • ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 2
  • 230000004913 activation Effects 0.000 claims description 2
  • 238000004364 calculation method Methods 0.000 claims description 2
  • 239000011159 matrix material Substances 0.000 claims description 2
  • 230000009467 reduction Effects 0.000 claims description 2
  • 238000007781 pre-processing Methods 0.000 claims 1
  • 238000013528 artificial neural network Methods 0.000 abstract description 7
  • 238000012935 Averaging Methods 0.000 abstract description 2
  • 238000013135 deep learning Methods 0.000 abstract description 2
  • 238000010586 diagram Methods 0.000 abstract description 2
  • 230000006872 improvement Effects 0.000 abstract description 2
  • 210000000601 blood cell Anatomy 0.000 abstract 1
  • 230000000977 initiatory effect Effects 0.000 description 4
  • 238000009534 blood test Methods 0.000 description 2
  • 201000010099 disease Diseases 0.000 description 2
  • 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 2
  • 230000004048 modification Effects 0.000 description 2
  • 238000012986 modification Methods 0.000 description 2
  • 210000001616 monocyte Anatomy 0.000 description 2
  • 238000003062 neural network model Methods 0.000 description 2
  • 230000002159 abnormal effect Effects 0.000 description 1
  • 230000004075 alteration Effects 0.000 description 1
  • 238000003491 array Methods 0.000 description 1
  • 238000013473 artificial intelligence Methods 0.000 description 1
  • 230000001174 ascending effect Effects 0.000 description 1
  • 230000008901 benefit Effects 0.000 description 1
  • 230000008859 change Effects 0.000 description 1
  • 238000003759 clinical diagnosis Methods 0.000 description 1
  • 238000011281 clinical therapy Methods 0.000 description 1
  • 230000003247 decreasing effect Effects 0.000 description 1
  • 238000001514 detection method Methods 0.000 description 1
  • 238000003745 diagnosis Methods 0.000 description 1
  • 230000008034 disappearance Effects 0.000 description 1
  • 239000000284 extract Substances 0.000 description 1
  • 230000003862 health status Effects 0.000 description 1
  • 210000000987 immune system Anatomy 0.000 description 1
  • 238000007689 inspection Methods 0.000 description 1
  • 238000002372 labelling Methods 0.000 description 1
  • 238000001000 micrograph Methods 0.000 description 1
  • 238000003058 natural language processing Methods 0.000 description 1
  • 244000052769 pathogen Species 0.000 description 1
  • 231100000915 pathological change Toxicity 0.000 description 1
  • 230000036285 pathological change Effects 0.000 description 1
  • 238000002360 preparation method Methods 0.000 description 1
  • 238000012545 processing Methods 0.000 description 1
  • 230000009466 transformation Effects 0.000 description 1

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Investigating Or Analysing Biological Materials (AREA)
  • Image Processing (AREA)

Abstract

The invention belongs to the field of medical microscopic image classification, and provides an improved attention convolution neural network leukocyte five-classification-based method for recognizing blood cell images by using deep learning. The method takes ResNeXt-50 as a backbone network, a residual error module uses packet convolution to reduce the number of model parameters, an independent attention module structure is added at the end of each stage of the network in parallel, aiming at a white blood cell characteristic diagram output by a convolutional neural network at different stages, a white blood cell key region is extracted by using the attention part of the attention module, a prediction type and a confidence score are output by the output part of the attention module, the output of a final network model is obtained by weighted averaging of the prediction type and the confidence score, and five white blood cell classification is realized based on improvement under the original ResNeXt-50 network framework. The invention utilizes the parallel attention modules to output class prediction and confidence score at different stages of the network, thereby improving the accuracy of leukocyte classification.

Description

Improved attention convolution neural network-based five-classification method for white blood cells

Technical Field

The invention belongs to the field of medical microscopic image classification, and relates to a leukocyte classification method of a convolutional neural network embedded with an attention module in parallel.

Background

The statements in this section merely provide background information related to the present disclosure and may not necessarily constitute prior art.

Leukocytes are part of the immune system and are responsible for destroying and removing old or abnormal cells and cell debris, as well as attacking pathogens and foreign bodies. Leukocytes commonly found in the blood are mature neutrophils, lymphocytes and monocytes, with a lower number of eosinophils and basophils. An increase or decrease in the number of leukocytes may represent a sign of the onset of certain diseases, and the morphology and proportion of the various leukocytes may reflect a person's health status. The precise classification of leukocytes thus plays a crucial role in clinical diagnosis and therapy. The artificial microscopic examination is the 'gold standard' of clinical leukocyte detection, can accurately classify leukocytes and can observe pathological changes of the leukocytes. However, this method requires the blood sample to be processed into a blood smear, which is then examined under a microscope by a professional examiner, and the examination is complicated, takes a long time, and has a high repeatability, which may lead to fatigue of the examiner and misjudgment or missed judgment of the leukocyte category, thereby affecting diagnosis and treatment of diseases.

In recent years, convolutional neural networks have been widely used in the field of medical image classification because they exhibit good performance in processing data presented by multiple arrays. However, many current deep learning methods cannot fully utilize key features and important information when performing convolution operations, which greatly limits the performance of neural networks. Currently, attention mechanisms have been widely applied in artificial intelligence related fields such as natural language processing and computer vision. The use of the attention mechanism allows the model to learn attention-being able to ignore irrelevant information and focus on critical information, thereby improving the performance of the computer vision task. Some researchers have used convolutional neural networks with attention added mechanisms for the classification of leukocytes, mainly including:

the patent "automatic recognition method of multiple types of white blood cells based on deep convolutional neural network" (CN 110059568A) proposes a deep convolutional neural network with attention added to classify white blood cells. The deep convolutional neural network is formed by cascading 9 initiation modules, auxiliary classifiers are added to the 4 th initiation module and the 7 th initiation module respectively, and the attention of an SE-Net channel is added when each initiation module is cascaded. And pre-training the model in an ILSVRC data set, and finely adjusting the network on a training set to realize the classification of the white blood cells.

The disadvantages are as follows: the auxiliary classifier added in the method utilizes two intermediate feature layers to predict the classes only to prevent the problems of gradient disappearance and the like and does not provide more help for the final decision of the network; in addition, the concept network has too many over-parameters and poor generalization capability, and does not necessarily have a better result.

The patent "method and system for realizing automatic classification of leukocytes based on mixed residual attention residual network" (CN 113343799a) proposes a residual network of a mixed attention mechanism to automatically classify leukocytes, the residual network of the mixed attention mechanism forms a mixed attention residual network by stacking and using residual attention modules, and the feature graph obtained after convolution is used for extracting key features through the mixed attention module to realize classification of leukocytes.

The disadvantages are as follows: the network is formed by stacking a plurality of residual attention modules, the influence of noise on low-level features after traversing a plurality of residual attention connections is ignored, and the low-level features (such as textures) are important in fine-grained identification and can help to distinguish two similar classes.

Disclosure of Invention

Aiming at the problems of noise influence caused by stacking and using a residual error attention module and incapability of completely utilizing different hierarchical features in the existing method, the invention provides an improved attention convolution neural network-based five-classification method for white blood cells, wherein ResNext-50 is used as a backbone network, the number of model parameters can be greatly reduced by grouping convolution in the network, an independent attention mechanism module is embedded in the backbone network in parallel to extract key information from a low-level feature map and a high-level feature map, prediction categories and confidence scores of the different hierarchical feature maps are obtained through the attention module respectively, and finally, model output categories are obtained through confidence weighting and averaging for all prediction categories, so that the purpose of improving the accuracy of white blood cell classification is achieved.

In order to achieve the purpose, the invention adopts the technical scheme that:

a leukocyte classification method based on an improved attention convolution neural network comprises the following steps:

step (1): collecting a leukocyte image, cutting the complete blood microscopic image into an individual image, and labeling the leukocyte by a blood inspection expert;

step (2): carrying out image enhancement operation on the leukocyte micrographs acquired in the step (1), and carrying out pretreatment;

and (3): dividing the white blood cell image data set processed in the step (2) into a training set and a testing set randomly according to a certain proportion;

and (4): constructing an improved attention convolution neural network model, and performing patrol on the model by using the training set divided in the step (3), wherein the process is a forward propagation process;

and (5): after one-time forward propagation, calculating an error between a predicted value and a true value by using a cross entropy loss function, continuously reducing the loss error by using an Adam algorithm, and updating parameters of each layer of the network model, wherein the process is a one-time backward propagation process;

and (6): repeatedly performing the forward propagation in the step (4) and the backward propagation in the step (5), continuously updating the network layer parameters, converging the network model when the number of training rounds reaches the set maximum number of training rounds, finishing the training, and storing the network model with the highest accuracy of the training set as the optimal network model;

and (7): and (4) carrying out white blood cell classification by using the optimal network model stored in the step (6), inputting a white blood cell image into the trained model, and outputting the category of the white blood cells.

The specific implementation process of the step (1) comprises the following steps: firstly, a blood smear provided by a hospital is subjected to white blood cell microscopic image collection by using a microscope, a white blood cell image with the size of 256 x 256 is cut by taking a complete white blood cell as a center on the whole microscopic image, and the type of the cut white blood cell image is marked by a blood specialty.

The specific implementation process of the step (2) comprises the following steps: and performing data enhancement operation on the white blood cells, including up-down left-right cutting, random rotation, image contrast enhancement and mirror image turning.

And (3) randomly dividing the image data set processed in the step (2) into a training set and a testing set according to the proportion of 7: 3.

The network model takes a residual error network ResNeXt-50 as a framework, reduces the model parameter number by using grouping convolution, uses a light attention mechanism to be embedded into each stage of the residual error network ResNeXt-50 in parallel, generates a spatial attention heat map for feature maps of different layers of the network model, and outputs category prediction and confidence score based on local information. The final output prediction is weighted by all class predictions and normalized confidence scores.

The attention mechanism we have adopted can be added after each convolutional layer and does not change the structure of the network as a whole. The attention module mainly comprises two sub-modules: attention head H, it extracts the region where the feature map is most relevant to the class decision; and outputting a head O, generating a category prediction through global pooling and full connection, and outputting a confidence gate score for each attention head. Each attention mechanism will get a class prediction and a confidence score, and finally, all the class predictions are weighted and averaged by the confidence scores to get the final prediction class.

Compared with the prior art, the invention has the following advantages:

1. the improved attention convolution neural network simultaneously adopts an attention mechanism to extract the most interesting parts of the low-level features and the high-level features, and utilizes the features to output class prediction and confidence scores to help the classification decision of the final model.

2. The invention adopts the residual error network model, and uses the grouping convolution in the residual error module, so that the network learns different characteristics, network parameters are reduced, and the model precision is improved.

3. The invention adopts a light-weight attention mechanism, does not bring complex parameters to the model, has flexible attention mechanism modules, can be expanded at different depths and widths, and shows good performance.

Drawings

FIG. 1 is a flow chart of the method of the present invention.

FIG. 2 shows the residual block of ResNeXt-50(34 × 4d) used in the present invention.

FIG. 3 is a schematic view of an attention module for use in the present invention.

FIG. 4 is a schematic diagram of an attention convolution neural network in accordance with the present invention.

Fig. 5 is a microscopic image of five types of leukocytes.

Detailed Description

The following further describes a specific embodiment of the present invention with reference to the drawings and technical solutions.

As shown in fig. 1, the general steps of the present invention are as follows:

step 1: collection and preparation of data sets, blood smears made by blood test experts were subjected to the collection of leucocyte microscopic images under the same conditions using a biomicroscope (magnification 1000) equipped with an industrial camera. The whole image was cropped to have a size of 256 × 256 white blood cell image centered on the whole single white blood cell, and the blood test specialist labeled the white blood cell image with a size of 256 × 256 to accurately classify neutrophils, eosinophils, basophils, monocytes, and lymphocytes.

Step 2: and (2) performing data enhancement operation on the leukocyte images collected and labeled in the step (1), specifically, cutting the leukocyte images, respectively cutting out the leukocyte images with the size of 224 × 224 from the four corners and the center of the images, performing mirror image turning, and rotating by 30 degrees and 60 degrees to enhance the image contrast, so as to obtain the leukocyte images at different positions and under different environments, prevent overfitting of the model and increase the generalization capability of the model.

And step 3: randomly dividing the leukocyte data set enhanced in the step 2 into a training set and a test set according to the proportion of 7:3, wherein the training set is used for the parameter training process of the convolution network model, and the test set is used for checking the efficiency of the whole leukocyte five-classification recognition algorithm and updating the parameter weight;

and 4, step 4: an improved attention convolution neural network model is constructed, the network model designed by the invention takes ResNeXt-50 as a backbone network, and an attention mechanism is embedded in parallel after each stage to generate class prediction and confidence score to assist the decision of a final model by utilizing the most useful parts of different level features.

1) The ResNeXt-50 network deduces and deduces the excellent ideas in the indications, ResNet and VGG to obtain a powerful network structure with simple structure. ResNeXt-50 consists of a common convolution structure, some residual blocks, and a full link layer. As shown in fig. 2, the left half of each residual module is a convolution operation composed of two convolution kernels of 1 × 1 and 3 × 3, the right half is a quick connection operation, and the results of the two parts are output through an activation function by an addition operation. Specifically, for the convolution operation of the left half, firstly, 1 × 1 convolution kernel is used to realize the overall dimension of ascending and descending, then the idea of grouping convolution is adopted to divide the channel into 32 branches, the 4-channel feature maps of each 32 branches are respectively operated through 3 × 3 convolution kernel, and the obtained transformation results (feature maps) are aggregated. Similar to Resnet, there are 4 layer layers for the entire ResNeXt-50, layer1 contains 3 residual blocks, layer2 contains 4 residual blocks, layer3 contains 6 residual blocks, and layer4 contains 3 residual blocks.

2) Attention modules are embedded at the end of different stages of ResNeXt-50, as shown in FIG. 3, and each attention module comprises two major parts, namely an attention head and an output head. And (3) performing convolution operation on the feature map Z obtained after convolution by using a convolution kernel of 1 × 1, and outputting an attention heat map M by using a spatial softmax, wherein the product of the attention heat map M and the input feature map Z channel obtains the output H of the attention head through a broadcasting mechanism, wherein M is a 2-dimensional plane, and the spatial softmax is used for the most relevant region in the model learning image. The output H of the attention head of each attention module consists of a spatial dimension reduction layer (namely, a global pooling layer), and then a category prediction o is generated through a full-connection layer, and each attention module makes the category prediction o according to local information of the attention module. However, in some cases, the local features are not sufficient to output a good prediction. In order to alleviate the problem, each attention module and the skeleton network output are led to predict a confidence score c through an inner product of the confidence score and a weight matrix, then the confidence score is normalized through a softmax function to obtain a weight g, the final output of the network is a weighted sum of the class prediction and the confidence score of each output, and a calculation formula is that

output=gnet·outputnet+∑∑gl k·ol k

Where output is the final output of the entire network model, gnetNormalized confidence score, output, for the output of the backbone networknetFor class prediction of skeletal networks, gl kNormalized confidence score, o, for the output of each attention modulel kA class prediction for each attention module.

And (4) training the network model designed in the step (4) by using the leukocyte training set data obtained in the step (3), wherein the process is a forward propagation process.

After one-time forward propagation, the error between the predicted value and the true value is calculated by using a cross entropy loss function, the loss error is continuously reduced by using a random gradient descent algorithm, the parameter of each layer of the network model is updated, a learning rate fixed step size decreasing strategy is adopted, the step size is 7 epochs, the gamma coefficient is 0.1, and the process is a one-time backward propagation process.

And (5) repeatedly carrying out forward propagation in the step 5 and backward propagation in the step 6, continuously updating network layer parameters, converging the network model when the number of training rounds reaches the set maximum number of training rounds, finishing training, and storing the network model with the highest accuracy of the training set as the optimal network model.

And (4) performing five-classification prediction on the leukocyte test set by using the optimal model stored in the step (7), wherein the test accuracy is shown in the following table.

Figure BDA0003316132030000051

It should be understood that parts of the specification not set forth in detail are well within the prior art. Various modifications and alterations to this invention will become apparent to those skilled in the art. Any modification, equivalent replacement, or improvement made without departing from the spirit and principle of the present invention shall fall within the protection scope of the present invention.

Claims (3)

1.一种基于改进的注意力卷积神经网络白细胞五分类方法,其特征在于,包括以下步骤:1. a five-classification method based on improved attention convolutional neural network leukocyte, is characterized in that, comprises the following steps: 步骤(1):采集白细胞图像,并把完整的血液显微图像裁剪为单独的图像,并对白细胞进行类别标注,分为嗜中性粒细胞、嗜酸性粒细胞、嗜碱性粒细胞、单核细胞和淋巴细胞;Step (1): collect leukocyte images, crop the complete blood microscopic image into separate images, and label the leukocytes into neutrophils, eosinophils, basophils, single nuclear cells and lymphocytes; 步骤(2):对于步骤(1)采集到的白细胞显微图像进行图像增强操作,并进行预处理;Step (2): perform image enhancement on the leukocyte microscopic image collected in step (1), and perform preprocessing; 步骤(3):将步骤(2)处理后的白细胞显微图像数据集随机地按照比例划分为训练集和测试集;Step (3): randomly divide the white blood cell microscopic image data set processed in step (2) into a training set and a test set according to the proportion; 步骤(4):构建改进的注意力卷积神经网络模型,使用步骤(3)划分的训练集对改进的注意力卷积神经网络模型进行训练,此过程为一次前向传播过程;Step (4): constructing an improved attention convolutional neural network model, using the training set divided in step (3) to train the improved attentional convolutional neural network model, and this process is a forward propagation process; 改进的注意力卷积神经网络模型以ResNeXt-50为骨干网络,在每个阶段后面并行嵌入一种注意力机制来利用不同层次特征最有用的部分生成类别预测和置信度得分辅助最终模型的决策;The improved attention convolutional neural network model takes ResNeXt-50 as the backbone network, and embeds an attention mechanism in parallel after each stage to utilize the most useful parts of different levels of features to generate class predictions and confidence scores to assist the decision of the final model ; 1)ResNeXt-50由一个普通卷积结构、一些残差块和一个全连接层组成;每个残差块的左半部分由两个1*1的卷积核和3*3的卷积核组成卷积操作,右半部分则是一个快速连接操作,两部分的结果经过加操作经激活函数得到输出;1) ResNeXt-50 consists of a common convolution structure, some residual blocks and a fully connected layer; the left half of each residual block consists of two 1*1 convolution kernels and 3*3 convolution kernels The convolution operation is formed, and the right half is a fast connection operation, and the results of the two parts are output through the activation function through the addition operation; 2)在ResNeXt-50不同阶段末嵌入注意力模块,每个注意力模块包括注意头和输出头两大部分;注意头部分对卷积过后得到的特征图Z使用1*1的卷积核进行卷积操作和空间softmax输出注意力热图M,注意力热图M与输入特征图Z通道的乘积通过广播机制得到注意头的输出H,其中M是一个2维平面,空间softmax用于模型学习图像中最相关的区域;每个注意力模块注意头的输出H由一个空间降维层即全局池化层组成,后面通过一个全连接层产生类别预测o,每个注意力模块根据其局部信息做出类别预测o;然而,在某些情况下,局部特征并不足以输出一个好的预测;为了缓解这个问题,让每个注意力模块以及骨架网络输出,通过与权矩阵的内积来预测置信度分数c,然后通过softmax函数对置信度分数进行归一化,得到权值g,网络的最终输出是各输出的类别预测和置信度分数的加权和,计算公式为2) Embed attention modules at the end of different stages of ResNeXt-50. Each attention module includes two parts: attention head and output head; the attention head part uses a 1*1 convolution kernel for the feature map Z obtained after convolution. The convolution operation and spatial softmax output attention heatmap M, and the product of the attention heatmap M and the input feature map Z channel gets the output H of the attention head through the broadcasting mechanism, where M is a 2-dimensional plane, and the spatial softmax is used for model learning The most relevant area in the image; the output H of the attention head of each attention module is composed of a spatial dimension reduction layer, that is, a global pooling layer, followed by a fully connected layer to generate category prediction o, and each attention module is based on its local information. Make a class prediction o; however, in some cases, local features are not enough to output a good prediction; to alleviate this problem, let each attention module as well as the skeleton network output, predict by inner product with the weight matrix The confidence score c is then normalized by the softmax function to obtain the weight g. The final output of the network is the weighted sum of the category prediction and confidence score of each output. The calculation formula is output=gnet·outputnet+∑∑gl ol k output=g net · output net +∑∑g l k · o l k 其中,output为整个网络模型的最终输出,gnet为骨干网络的输出的归一化后的置信度分数,outputnet为骨架网络的的类别预测,gl k为每个注意力模块的输出的归一化后的置信度分数,ol k为每个注意力模块的类别预测;Among them, output is the final output of the entire network model, g net is the normalized confidence score of the output of the backbone network, output net is the category prediction of the backbone network, and g l k is the output of each attention module. The normalized confidence score, ol k is the category prediction of each attention module; 使用步骤3得到的白细胞训练集数据对步骤4设计的网络模型进行训练,此过程为一次前向传播;Use the white blood cell training set data obtained in step 3 to train the network model designed in step 4, and this process is a forward propagation; 步骤(5):经过一次向前传播后,使用交叉熵损失函数计算预测值和真实值之间的误差,通过使用Adam算法来不断地减小损失误差,并更新网络模型每层的参数,此过程为一次后向传播过程;Step (5): After one forward propagation, use the cross entropy loss function to calculate the error between the predicted value and the true value, and use the Adam algorithm to continuously reduce the loss error and update the parameters of each layer of the network model. The process is a backward propagation process; 步骤(6):反复经过步骤(4)的前向传播和步骤(5)的后向传播,不断更新网络层参数,当训练轮数达到设置的最大训练轮数时,网络模型收敛,训练结束,保存训练集准确率最高的网络模型为最优网络模型;Step (6): Repeat the forward propagation of step (4) and the backward propagation of step (5), and continuously update the network layer parameters. When the number of training rounds reaches the set maximum number of training rounds, the network model converges and the training ends. , save the network model with the highest accuracy in the training set as the optimal network model; 步骤(7):使用步骤(6)保存的最优网络模型进行白细胞分类,输入一张白细胞图像到训练好的模型中,输出白细胞的类别。Step (7): Use the optimal network model saved in step (6) to classify white blood cells, input a white blood cell image into the trained model, and output the type of white blood cells. 2.根据权利要求1所述的注意力卷积神经网络白细胞五分类方法,其特征在于,步骤(2)中针对白细胞进行数据增强操作,包括上下左右裁剪、随机旋转、增强图像对比度、镜像翻转。2. The five-classification method for white blood cells of attention convolutional neural network according to claim 1, is characterized in that, in step (2), data enhancement operation is carried out for white blood cells, including up and down, left and right cropping, random rotation, enhanced image contrast, mirror flipping . 3.根据权利要求1或2所述的注意力卷积神经网络白细胞五分类方法,其特征在于,步骤(3)中将步骤(2)处理过后的图像数据集随机地按照7:3的比例划分为训练集和测试集。3. The attention convolutional neural network five classification method for white blood cells according to claim 1 or 2, wherein in step (3), the image data set processed in step (2) is randomly according to the ratio of 7:3 Divided into training set and test set.

CN202111231648.6A 2021-10-22 2021-10-22 Improved attention convolution neural network-based five-classification method for white blood cells Active CN113887503B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111231648.6A CN113887503B (en) 2021-10-22 2021-10-22 Improved attention convolution neural network-based five-classification method for white blood cells

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111231648.6A CN113887503B (en) 2021-10-22 2021-10-22 Improved attention convolution neural network-based five-classification method for white blood cells

Publications (2)

Publication Number Publication Date
CN113887503A true CN113887503A (en) 2022-01-04
CN113887503B CN113887503B (en) 2022-06-14

Family

ID=79004139

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111231648.6A Active CN113887503B (en) 2021-10-22 2021-10-22 Improved attention convolution neural network-based five-classification method for white blood cells

Country Status (1)

Country Link
CN (1) CN113887503B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114638878A (en) * 2022-03-18 2022-06-17 北京安德医智科技有限公司 Two-dimensional echocardiogram pipe diameter detection method and device based on deep learning
CN114820545A (en) * 2022-05-09 2022-07-29 天津大学 A method for primary screening and detection of onychomycosis based on improved residual network
CN116503661A (en) * 2023-05-10 2023-07-28 桂林电子科技大学 ResNeXt structure based on attention mechanism and image classification algorithm using this structure
CN117422645A (en) * 2023-11-14 2024-01-19 中国科学院长春光学精密机械与物理研究所 Confidence aggregation-based radar point cloud shape completion method

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104751462A (en) * 2015-03-29 2015-07-01 嘉善加斯戴克医疗器械有限公司 White cell segmentation method based on multi-feature nonlinear combination
CN106897682A (en) * 2017-02-15 2017-06-27 电子科技大学 Leucocyte automatic identifying method in a kind of leukorrhea based on convolutional neural networks
CN109034045A (en) * 2018-07-20 2018-12-18 中南大学 A kind of leucocyte automatic identifying method based on convolutional neural networks
CN110059568A (en) * 2019-03-21 2019-07-26 中南大学 Multiclass leucocyte automatic identifying method based on deep layer convolutional neural networks
CN110059656A (en) * 2019-04-25 2019-07-26 山东师范大学 The leucocyte classification method and system for generating neural network are fought based on convolution
US20200340908A1 (en) * 2017-12-22 2020-10-29 Imec Vzw Fast and Robust Fourier Domain-Based Cell Differentiation
US20210139960A1 (en) * 2017-06-23 2021-05-13 FUNDAClÓ INSTITUT DE CIÈNCIES FOTÒNIQUES Method for quantifying protein copy-number
CN113343975A (en) * 2021-04-22 2021-09-03 山东师范大学 Deep learning-based white blood cell classification system and method

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104751462A (en) * 2015-03-29 2015-07-01 嘉善加斯戴克医疗器械有限公司 White cell segmentation method based on multi-feature nonlinear combination
CN106897682A (en) * 2017-02-15 2017-06-27 电子科技大学 Leucocyte automatic identifying method in a kind of leukorrhea based on convolutional neural networks
US20210139960A1 (en) * 2017-06-23 2021-05-13 FUNDAClÓ INSTITUT DE CIÈNCIES FOTÒNIQUES Method for quantifying protein copy-number
US20200340908A1 (en) * 2017-12-22 2020-10-29 Imec Vzw Fast and Robust Fourier Domain-Based Cell Differentiation
CN109034045A (en) * 2018-07-20 2018-12-18 中南大学 A kind of leucocyte automatic identifying method based on convolutional neural networks
CN110059568A (en) * 2019-03-21 2019-07-26 中南大学 Multiclass leucocyte automatic identifying method based on deep layer convolutional neural networks
CN110059656A (en) * 2019-04-25 2019-07-26 山东师范大学 The leucocyte classification method and system for generating neural network are fought based on convolution
CN113343975A (en) * 2021-04-22 2021-09-03 山东师范大学 Deep learning-based white blood cell classification system and method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
DARIA BORTOLOTTI等: "Impact of HLA-G analysis in prevention, diagnosis and treatment of pathological conditions", 《WORLD JOURNAL OF METHODOLOGY》 *
陈畅等: "基于卷积神经网络的外周血白细胞分类", 《中国生物医学工程学报》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114638878A (en) * 2022-03-18 2022-06-17 北京安德医智科技有限公司 Two-dimensional echocardiogram pipe diameter detection method and device based on deep learning
CN114638878B (en) * 2022-03-18 2022-11-11 北京安德医智科技有限公司 Two-dimensional echocardiogram pipe diameter detection method and device based on deep learning
CN114820545A (en) * 2022-05-09 2022-07-29 天津大学 A method for primary screening and detection of onychomycosis based on improved residual network
CN116503661A (en) * 2023-05-10 2023-07-28 桂林电子科技大学 ResNeXt structure based on attention mechanism and image classification algorithm using this structure
CN117422645A (en) * 2023-11-14 2024-01-19 中国科学院长春光学精密机械与物理研究所 Confidence aggregation-based radar point cloud shape completion method

Also Published As

Publication number Publication date
CN113887503B (en) 2022-06-14

Similar Documents

Publication Publication Date Title
CN113887503A (en) 2022-01-04 A five-class classification method of white blood cells based on an improved attention convolutional neural network
Dong et al. 2017 Classification of cataract fundus image based on deep learning
CN106778902B (en) 2020-01-21 Dairy cow individual identification method based on deep convolutional neural network
CN111951246B (en) 2022-03-04 Multidirectional X-ray chest radiography pneumonia diagnosis method based on deep learning
CN106408001B (en) 2019-10-11 A fast detection method for regions of interest based on deep kernel hashing
Bani-Hani et al. 2018 Classification of leucocytes using convolutional neural network optimized through genetic algorithm
CN108830282A (en) 2018-11-16 A kind of the breast lump information extraction and classification method of breast X-ray image
Shivaprasad et al. 2023 Deep learning-based plant leaf disease detection
CN113012163A (en) 2021-06-22 Retina blood vessel segmentation method, equipment and storage medium based on multi-scale attention network
CN110059568A (en) 2019-07-26 Multiclass leucocyte automatic identifying method based on deep layer convolutional neural networks
Wahid et al. 2019 Deep convolutional neural network for microscopic bacteria image classification
Magpantay et al. 2022 A transfer learning-based deep CNN approach for classification and diagnosis of acute lymphocytic leukemia cells
CN118135278A (en) 2024-06-04 Breast cancer pathology full-section image classification method based on double-layer multi-instance learning
CN116778250A (en) 2023-09-19 Coronary artery lesion classification method based on transfer learning and CBAM
Guefrachi et al. 2024 Diabetic retinopathy detection using deep learning multistage training method
Alam et al. 2024 Benchmarking deep learning frameworks for automated diagnosis of ocular toxoplasmosis: A comprehensive approach to classification and segmentation
Tobias et al. 2020 Android application for chest X-ray health classification from a CNN deep learning TensorFlow model
CN118430790A (en) 2024-08-02 Mammary tumor BI-RADS grading method based on multi-modal-diagram neural network
Li et al. 2022 Fine-grained classification of grape leaves via a pyramid residual convolution neural network
Zou et al. 2020 Deep learning and its application in diabetic retinopathy screening
Sevinç et al. 2022 An effective medical image classification: transfer learning enhanced by auto encoder and classified with SVM
Dey et al. 2023 A Butterfly Malar Rash Detection Model for Early Systemic Lupus Erythematosus Diagnosis
Kalaivani et al. 2023 Deep Learning with Multi-Class Classification for Detection of Covid-19 and Pneumonia
Vashisht et al. 2023 Pneumonia classification model using Deep Learning algorithm
Taha et al. 2021 Automatic identification of malaria-infected cells using deep convolutional neural network

Legal Events

Date Code Title Description
2022-01-04 PB01 Publication
2022-01-04 PB01 Publication
2022-01-21 SE01 Entry into force of request for substantive examination
2022-01-21 SE01 Entry into force of request for substantive examination
2022-06-14 GR01 Patent grant
2022-06-14 GR01 Patent grant