CN112561865B - Method, system and storage medium for training detection model of constant molar position - Google Patents
Method, system and storage medium for training detection model of constant molar position Download PDFInfo
- Publication number
- CN112561865B CN112561865B CN202011406995.3A CN202011406995A CN112561865B CN 112561865 B CN112561865 B CN 112561865B CN 202011406995 A CN202011406995 A CN 202011406995A CN 112561865 B CN112561865 B CN 112561865B
- Authority
- CN
- China
- Prior art keywords
- image
- information
- confidence coefficient
- features
- feature
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 88
- 238000000034 method Methods 0.000 title claims abstract description 51
- 238000012549 training Methods 0.000 title claims abstract description 29
- 230000004913 activation Effects 0.000 claims abstract description 39
- 238000000605 extraction Methods 0.000 claims abstract description 36
- 230000003213 activating effect Effects 0.000 claims abstract description 18
- 230000008569 process Effects 0.000 claims abstract description 16
- 230000008197 oral dentition Effects 0.000 claims abstract description 8
- 238000011176 pooling Methods 0.000 claims description 32
- 210000000214 mouth Anatomy 0.000 claims description 16
- 210000004513 dentition Anatomy 0.000 claims description 15
- 230000007246 mechanism Effects 0.000 claims description 15
- 230000036346 tooth eruption Effects 0.000 claims description 15
- 230000006870 function Effects 0.000 claims description 11
- 238000004364 calculation method Methods 0.000 claims description 9
- 230000004927 fusion Effects 0.000 claims description 8
- 238000012935 Averaging Methods 0.000 claims description 3
- 238000007499 fusion processing Methods 0.000 claims description 3
- 239000011159 matrix material Substances 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 3
- 208000002925 dental caries Diseases 0.000 description 7
- 241000894006 Bacteria Species 0.000 description 5
- 238000012216 screening Methods 0.000 description 5
- 238000013528 artificial neural network Methods 0.000 description 4
- 238000007789 sealing Methods 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 3
- 210000004763 bicuspid Anatomy 0.000 description 3
- 230000002378 acidificating effect Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 230000006378 damage Effects 0.000 description 2
- 201000010099 disease Diseases 0.000 description 2
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 2
- 235000013305 food Nutrition 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 208000017667 Chronic Disease Diseases 0.000 description 1
- 239000002253 acid Substances 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 150000001720 carbohydrates Chemical class 0.000 description 1
- 235000014633 carbohydrates Nutrition 0.000 description 1
- 230000001013 cariogenic effect Effects 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 230000006806 disease prevention Effects 0.000 description 1
- 230000003670 easy-to-clean Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000013100 final test Methods 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000002207 metabolite Substances 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000000149 penetrating effect Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000001681 protective effect Effects 0.000 description 1
- 239000003566 sealing material Substances 0.000 description 1
- 230000028327 secretion Effects 0.000 description 1
- 235000021058 soft food Nutrition 0.000 description 1
- 235000021147 sweet food Nutrition 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30036—Dental; Teeth
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a method, a system and a storage medium for training a detection model of a constant molar position, wherein the method comprises the following steps: acquiring a plurality of oral dentition images as first images; extracting first features of the first image to obtain a first feature image; selectively activating the first characteristic image in a channel dimension and a space dimension respectively; performing second feature extraction on the first feature image after the activation; a detection frame for predicting the constant molar position according to the image features extracted by the second features; fusing the information of the detection frames to generate a first confidence coefficient; taking the maximum first confidence coefficient as a prediction output result; and reversely updating parameters of the detection model according to the prediction output result. The detection model provided by the invention has pertinence in the application process and can be more suitable for different application situations, so that the accuracy of the detection result of the detection model in the application process is improved. The invention can be widely applied to the technical field of model training.
Description
Technical Field
The invention relates to the technical field of model training, in particular to a method, a system and a storage medium for training a detection model of a constant molar position.
Background
Caries is a chronic disease in which bacteria (cariogenic bacteria) produce acid from carbohydrates in food, causing progressive destruction of hard tissues of teeth, and is a major common disease of the oral cavity, and one of the most common diseases in humans. Children of school age 6-12 years old often have a preference for sweet food, soft food and sticky food and are easy to stick to teeth, however, oral hygiene habits of children at this stage are often poor or a cleaning method is not well mastered, teeth can be cleaned effectively, and the children are high-incidence people suffering from caries. In addition, during the period of replacing teeth, the newly erupted permanent teeth, especially the newly erupted permanent molar fossa are often deeper, bacteria are easy to accumulate and are not easy to clean, acidic secretion produced by the bacteria can cause the damage of hard tissues of teeth, and dental caries is easy to produce. Thus, timely screening and early intervention in the oral condition of children is the root cause of the disease prevention and treatment of caries. Pit and groove sealing is the best method recommended by world health organization for preventing constant dental caries, has been widely popularized in many countries and regions, and widely popularized in China for school-age children. The pit and groove sealing refers to a method that pit and groove sealing materials are coated on pit and groove point gaps of occlusal surfaces and buccal surfaces of dental crowns without damaging dental tissues, and are solidified and hardened after flowing into and penetrating into the pit and groove to form a protective barrier which covers the deep pit and groove, so that caries causing bacteria and acidic metabolites can be prevented from corroding dental bodies, and the pit and groove caries preventing method is achieved.
The first step in the occlusion of the sulcus is to screen out permanent molars that are in compliance with the indication. In the past, the pit and fissure sealing project is screened by the manpower of an oral doctor, and a great deal of manpower, material resources and financial resources are required to be spent. In order to save manpower, material resources and financial resources of pit and trench sealing projects, an online tooth detection and early warning method is provided. However, due to the lack of pertinence and adaptation of the current online detection model, the accuracy of the recognition result of the constant molars is not high in the application process.
Disclosure of Invention
In order to solve one of the above technical problems, the present invention aims to: a method, system and storage medium for training a detection model of a constant molar position are provided, which can improve the accuracy of a detection result of the detection model in an application process.
In a first aspect, embodiments of the present invention provide:
a method for training a detection model of a constant molar position comprises the following steps:
acquiring a plurality of oral dentition images as first images;
extracting first features of the first image to obtain a first feature image;
selectively activating the first characteristic image in a channel dimension and a space dimension respectively;
performing second feature extraction on the first feature image after the activation;
a detection frame for predicting the constant molar position according to the image features extracted by the second features;
fusing the information of the detection frames to generate a first confidence coefficient;
taking the maximum first confidence coefficient as a prediction output result;
and reversely updating parameters of the detection model according to the prediction output result.
Further, the acquiring a plurality of oral dentition images as the first image includes:
acquiring a plurality of oral dentition images;
the sizes of the oral cavity dentition images are processed into preset sizes;
interlacing sampling is carried out on the oral cavity dentition image after the size processing to obtain a plurality of sub-images;
and taking the spliced image of the plurality of sub-images as a first image.
Further, the performing the first feature extraction on the first image to obtain a first feature image includes:
performing downsampling feature extraction on the first image to generate feature images with different sizes;
the feature images of different sizes are taken as first feature images.
Further, the selectively activating the first feature image in the channel dimension includes:
carrying out average pooling and maximum pooling on the first characteristic image;
and selectively activating the first characteristic images after the average pooling and the maximum pooling in the channel dimension by adopting an attention mechanism. Further, the selectively activating the first feature image in the spatial dimension specifically includes:
the first image processed by the convolution kernel attention mechanism is selectively activated in the spatial dimension.
Further, the second feature extraction is an abstract feature extraction.
Further, the fusing the information of the detection frame to generate a first confidence coefficient includes:
fusing the position information of the detection frame, the second confidence coefficient and the image characteristic information;
and generating a first confidence coefficient according to the fusion result.
In a second aspect, embodiments of the present invention provide:
a model training system for detecting a constant molar position, comprising:
the acquisition module is used for acquiring a plurality of oral cavity dentition images as first images;
the first feature extraction module is used for carrying out first feature extraction on the first image to obtain a first feature image;
the activation module is used for selectively activating the first characteristic image in the channel dimension and the space dimension respectively;
the second feature extraction module is used for carrying out second feature extraction on the first feature image after the selection activation;
the prediction module is used for predicting a detection frame of the constant molar position according to the image features after the second features are extracted;
the fusion module is used for fusing the information of the detection frame to generate a first confidence coefficient;
the confidence coefficient selection module is used for taking the maximum first confidence coefficient as a prediction output result;
and the parameter updating module is used for reversely updating the parameters of the detection model according to the prediction output result.
In a third aspect, embodiments of the present invention provide:
a model training system for detecting a constant molar position, comprising:
at least one memory for storing a program;
at least one processor for loading the program to perform the method of model training for detection of the position of the permanent molars.
In a fourth aspect, embodiments of the present invention provide:
a storage medium having stored therein a processor executable program which when executed by a processor is for performing the method of model training for detection of a constant molar position.
The embodiment of the invention has the beneficial effects that: according to the embodiment of the invention, the first characteristic extraction is firstly carried out on a plurality of oral cavity dentition images, then the first characteristic images are respectively and selectively activated in the channel dimension and the space dimension, the second characteristic extraction is carried out on the first characteristic images after the selective activation, then the detection frame of the constant molar position is predicted according to the image characteristics after the second characteristic extraction, the information of the detection frame is fused, the first confidence coefficient is generated, finally the maximum first confidence coefficient is used as a prediction output result, and the parameters of the detection model are reversely updated according to the prediction output result, so that the detection model after the training of the embodiment has pertinence and can be more suitable for different application situations in the application process, and the accuracy of the detection result of the detection model in the application process is improved.
Drawings
FIG. 1 is a flow chart of a method for training a model for detecting the position of a permanent molar in accordance with an embodiment of the present invention.
Detailed Description
The invention will now be described in further detail with reference to the drawings and to specific examples. The step numbers in the following embodiments are set for convenience of illustration only, and the order between the steps is not limited in any way, and the execution order of the steps in the embodiments may be adaptively adjusted according to the understanding of those skilled in the art.
In the following description, reference is made to "some embodiments" which describe a subset of all possible embodiments, but it is to be understood that "some embodiments" can be the same subset or different subsets of all possible embodiments and can be combined with one another without conflict.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs. The terminology used herein is for the purpose of describing embodiments of the present application only and is not intended to be limiting of the present application.
Referring to fig. 1, the embodiment of the invention provides a method for training a detection model of a constant molar position, which can be applied to a server or a background processor of various platforms, so that a user only needs to perform a designated step operation on a man-machine interface in an application process, and the constant molar position detection can be completed.
The embodiment comprises the following steps:
s1, acquiring a plurality of oral cavity dentition images as first images; in this step, the oral cavity dentition image may be an image of the person to be detected acquired in the actual detection process, where the acquisition process may be that the person to be acquired is photographed by a mobile phone or other devices with a photographing function.
When the plurality of oral cavity dentition images are applied to the training process of the detection model, in some embodiments, data enhancement modes such as data overturning and rotation can be adopted to carry out data enhancement on the oral cavity dentition images so as to provide more training sample data and improve training precision.
In some embodiments, the step S1 may be implemented by:
acquiring a plurality of oral dentition images; the plurality of oral cavity dentition images are images shot by different mobile intelligent equipment, and the sizes of the plurality of oral cavity dentition images are processed into the preset sizes because the shot images are inconsistent in size and the input size of the neural network is of a fixed size. In this process, the image size may be set to 512×512, i.e. the input image data size is c×w×h, where c=3 represents the number of input image channels is 3, and each channel sequentially represents the color values on the red, green, and blue components, respectively; w=512 is the width of the image; h=512 is the height of the image. Then, carrying out interlaced sampling on the oral cavity dentition image subjected to the size processing to obtain a plurality of sub-images; for example, for an image of 512 x 512 size, starting from each point in a 2 x 2 region in the upper left corner of the image, the original image is gap sampled at intervals of length 2, generating 4 sub-images of 256 x 256. And finally, taking the image formed by splicing the plurality of sub-images as a first image input by the detection model to fully utilize the information on the image.
S2, performing first feature extraction on the first image to obtain a first feature image.
In this step, a first feature extraction is performed on the first image by the feature extraction module. The feature extraction modules are formed by overlapping 8 layers of network modules, 2 layers of the 8 layers of networks form a downsampling feature extraction module group as a whole, the downsampling feature extraction module group is divided into four groups, and each group of operation generates a feature map with corresponding size. After feature extraction is performed by the downsampling feature extraction module group, 4 feature images with the sizes of 512 x 512, 256 x 256, 128 x 128 and 64 x 64 can be obtained, and finally feature images with different sizes are used as first feature images.
S3, selectively activating the first characteristic image in the channel dimension and the space dimension respectively;
in some embodiments, the first feature image is selectively activated in the channel dimension by:
carrying out average pooling and maximum pooling on the first characteristic image;
and selectively activating the first characteristic images after the average pooling and the maximum pooling in the channel dimension by adopting an attention mechanism.
In some embodiments, the first feature image is selectively activated in a spatial dimension, which may be in the following manner:
the first image processed by the convolution kernel attention mechanism is selectively activated in the spatial dimension.
In the above embodiment, the attention mechanism and the convolution kernel form an attention mechanism module, specifically, the selective activation of the channel dimension is performed first, the feature map is processed by the maximum pooling and the average pooling respectively, then the feature maps obtained by the two operations are selectively activated to corresponding feature channels respectively, wherein the selection of the activation channel is a parameter that can be iterated through a neural network to perform self-learning, and finally the two features obtained after the activation of the channel dimension are overlapped and fused together to be used as the feature after the activation of the channel dimension. And secondly, selectively activating the space dimension, wherein the part adopts a strategy of simultaneously utilizing the characteristic map after average pooling and maximum pooling as input, and expands the receptive field by utilizing a convolution kernel of 7*7, and maps the convolved output to the importance degree of each position in the characteristic map relative to the prediction target through a sigmoid activation function.
Specifically, in the selective activation of the channel dimension, the calculation is performed using equation 1:
wherein F is S For features that are selectively activated through the channel dimension,representing the image features after averaging pooling, < >>Representing the image characteristics after maximum pooling, omega 0 And omega 1 For two different matrix parameters, which are used to represent the weight sizes for different channel numbers, σ represents the activation function sigmoid.
In the step, two identical parameters are applied to the feature map processed in different modes, so that the parameters have stronger robustness, and the feature map is mapped by continuously applying the two parameters, so that more neurons can be included to fit the feature as complex as possible on the basis of ensuring the feature map to be unchanged in size. Finally, mapping from real domain R to [0,1] is accomplished with an activation function represented by σ to achieve normalization of the output.
In the selective activation of the spatial dimension, the calculation is performed using equation 2:
wherein f 7*7 A convolution operation is represented as 7*7,representing image features averaged pooling on the basis of features obtained by selective activation of channel dimensions, a>Representing image features obtained by maximum pooling based on features obtained by selective activation of channel dimensions, σ representing the activation function sigmoid.
In this step, the receptive field is enlarged by applying a larger convolution kernel, so that the model is helped to infer the importance degree of the current position according to the content of the current pixel point and the field thereof.
S4, extracting second features on the first feature images after the selection and activation; in this step, the second feature extraction is an abstract feature extraction.
S5, predicting a detection frame of the constant molar position according to the image features extracted by the second features; in this step, it specifically predicts the position and size of the detection frame on the first image.
And S6, fusing the information of the detection frames to generate a first confidence coefficient.
In some embodiments, step S6, which may be specifically implemented by:
fusing the position information of the detection frame, the second confidence coefficient and the image characteristic information;
and generating a first confidence coefficient according to the fusion result.
In this embodiment, the second confidence coefficient is the confidence coefficient of the image itself, and the first confidence coefficient is the confidence coefficient after the same image fusion process as the second confidence coefficient.
Since the existing model prediction detection frame may have a certain deviation, for example, two teeth of the same type appear on the same side, which is a phenomenon that is impossible in practical situations. Thus, in some embodiments, a detection frame screening mechanism is employed that incorporates multimodal information fusion. Specifically, the existing detection frame screening method only judges according to the confidence coefficient generated by the regression model and the overlapping area between different detection frames, and ignores the image characteristic information of the candidate detection frames. In this embodiment, the position information and the image feature of the candidate frame are regarded as information of two different modes, and the detection frame can be effectively screened by fusing the information of the two different modes.
Specifically, when information of different modalities is fused, formula 3 and formula 4 are adopted:
wherein f A For the image features after the confidence information is fused, f I To fuse the confidence information after image features,and->The original image characteristics and the confidence information are respectively, and the rest variables are parameters of the neural network self-learning.
On the basis of fusing information of different modes, calculating by combining the position information of the image to obtain the confidence coefficient of the detection frame fusing the multi-mode information, wherein the specific calculation process is shown in a formula 5-a formula 8:
f S =σ(W S f total )+b S equation 8
Wherein,representing features fused with confidence information and image feature information, < >>For the coded detection frame position information, f total,k Then indicate at +.>Further fusing the characteristics of the position information of the detection frame, f S And representing the new confidence coefficient finally obtained after the multi-mode information is fused, wherein the rest letters or letter combinations represent parameters required by the neural network.
S7, taking the maximum first confidence coefficient as a prediction output result; in this step, the prediction output result with the highest confidence is obtained for each region.
And S8, reversely updating parameters of the detection model according to the prediction output result, so that the detection model has pertinence in the application process and can be more suitable for different application situations, and the accuracy of the detection result of the detection model in the application process is improved.
The embodiment is specifically applied and comprises the following steps:
color images acquired by the mobile intelligent device are acquired, and angles, light rays and covered area conditions of the images are different. In this embodiment, the size of the image is unified to 512×512.
The image dataset was divided into training and testing sets at a ratio of 4:1, i.e., 3316 pictures in the dataset were used to train the model, and the remaining 829 pictures were used to verify the performance of the model. Final test results as shown in table 1:
TABLE 1
Method | AP | AP 50 | AP 75 | AR | AR 50 | SR 75 | AP molar1 | AP molar2 | Time(ms) |
Baseline(D) | 46.2 | 92.5 | 40.2 | 48.4 | 99.7 | 70.3 | 92.7 | 94.3 | 6.6 |
Baseline(O) | 44.5 | 89.7 | 37.2 | 46.7 | 98.1 | 64.3 | - | 90.1 | 6.1 |
Baseline(A+D) | 46.7 | 93.2 | 40.8 | 48.6 | 99.7 | 70.8 | 93.5 | 95.4 | 8.9 |
Baseline(AN+D) | 47.9 | 94.5 | 41.6 | 48.5 | 98.9 | 72.5 | 94.2 | 97.3 | 10.4 |
Baseline(AN+A+D) | 49.1 | 95.6 | 42.3 | 49.1 | 99.2 | 72.6 | 96.1 | 98.5 | 12.4 |
In table 1, AP and AR are taken as evaluation indexes, that is, the IOU values of the detection frame and the target frame are regarded as correct when they are greater than the threshold value. Wherein Baseline (D) represents the results when the first premolars and the second premolars are detected simultaneously, baseline (O) represents the results when only the second premolars are detected, A represents the add-on mechanism, AN represents the add-on A-NMS detection frame screening mechanism.
In summary, the above embodiment limits the range of target detection by adding the attention mechanism, which greatly increases the adaptability and robustness of the method. Meanwhile, a detection frame screening mechanism is set, the image features and the position features of the target are synthesized to further select the detection frame, and candidate frames with the image features matched with the position features are reserved, so that the accuracy of the detection result is improved on the same data set.
The embodiment of the invention provides a constant molar position detection model training system corresponding to the method of fig. 1, which comprises the following steps:
the acquisition module is used for acquiring a plurality of oral cavity dentition images as first images;
the first feature extraction module is used for carrying out first feature extraction on the first image to obtain a first feature image;
the activation module is used for selectively activating the first characteristic image in the channel dimension and the space dimension respectively;
the second feature extraction module is used for carrying out second feature extraction on the first feature image after the selection activation;
the prediction module is used for predicting a detection frame of the constant molar position according to the image features after the second features are extracted;
the fusion module is used for fusing the information of the detection frame to generate a first confidence coefficient;
the confidence coefficient selection module is used for taking the maximum first confidence coefficient as a prediction output result;
and the parameter updating module is used for reversely updating the parameters of the detection model according to the prediction output result.
The content of the method embodiment of the invention is suitable for the system embodiment, the specific function of the system embodiment is the same as that of the method embodiment, and the achieved beneficial effects are the same as those of the method.
The embodiment of the invention provides a detection model training system for a constant molar position, which comprises the following steps:
at least one memory for storing a program;
at least one processor for loading the program to perform the method of model training for detection of the position of the permanent molars.
The content of the method embodiment of the invention is suitable for the system embodiment, the specific function of the system embodiment is the same as that of the method embodiment, and the achieved beneficial effects are the same as those of the method.
Embodiments of the present invention provide a storage medium having stored therein a processor-executable program which, when executed by a processor, is adapted to perform the method of model training for detection of a constant molar position.
Embodiments of the present invention also disclose a computer program product or computer program comprising computer instructions stored in a computer readable storage medium. The computer instructions may be read from a computer-readable storage medium by a processor of a computer device, and executed by the processor, to cause the computer device to perform the method shown in fig. 1.
While the preferred embodiment of the present invention has been described in detail, the present invention is not limited to the embodiments described above, and various equivalent modifications and substitutions can be made by those skilled in the art without departing from the spirit of the present invention, and these equivalent modifications and substitutions are intended to be included in the scope of the present invention as defined in the appended claims.
Claims (7)
1. The method for training the detection model of the constant molar position is characterized by comprising the following steps of:
acquiring a plurality of oral dentition images as first images;
extracting first features of the first image to obtain a first feature image;
selectively activating the first characteristic image in a channel dimension and a space dimension respectively;
performing second feature extraction on the first feature image after the activation;
a detection frame for predicting the constant molar position according to the image features extracted by the second features;
fusing the information of the detection frames to generate a first confidence coefficient;
taking the maximum first confidence coefficient as a prediction output result;
reversely updating parameters of the detection model according to the prediction output result;
the selectively activating the first feature image in the channel dimension includes:
carrying out average pooling and maximum pooling on the first characteristic image;
selectively activating the first characteristic images after the average pooling and the maximum pooling in the channel dimension by adopting an attention mechanism;
in the selective activation of channel dimensions, the calculation is performed using equation 1:
equation 1
Wherein,for features after selective activation via channel dimension, < ->Representing the image features after the averaging pooling,representing the image features after maximum pooling, < >>And->For two different matrix parameters representing the weight sizes for different channel numbers,/->Representing an activation function sigmoid;
the selective activation of the first feature image in the spatial dimension is specifically:
selectively activating the first image processed by adopting a convolution check attention mechanism in a space dimension;
in the selective activation of the spatial dimension, the calculation is performed using equation 2:
equation 2
Wherein,representing 7*7 convolution operation,/->Representing image features averaged pooling on the basis of features obtained by selective activation of channel dimensions, a>Representing image features resulting from maximum pooling based on features obtained by selective activation of channel dimensions, and>representing an activation function sigmoid;
the fusing the information of the detection frame to generate a first confidence coefficient comprises the following steps:
fusing the position information of the detection frame, the second confidence coefficient and the image characteristic information;
generating a first confidence coefficient according to the fusion result, wherein the second confidence coefficient is the confidence coefficient of the image, and the first confidence coefficient is the confidence coefficient of the same image after fusion processing with the second confidence coefficient;
when information of the detection frames is fused, a formula 3 and a formula 4 are adopted when information of different modes is fused:
equation 3
Equation 4
Wherein,for the image feature after fusing the second confidence information,/I>To fuse the confidence information after image features,and->Original image characteristics and confidence information respectively;
on the basis of fusing information of different modes, calculating by combining the position information of the image to obtain the confidence coefficient of the detection frame fusing the multi-mode information, wherein the specific calculation process is shown in a formula 5-a formula 8:
equation 5
Equation 6
Equation 7
Equation 8
Wherein,representing features fused with the second confidence information and the image feature information,/for the feature>For the position information of the encoded detection frame, < >>Then indicate at +.>Further fusing the features of the position information of the detection frame,/on the basis of (a)>And representing the new confidence coefficient finally obtained after the multi-mode information is fused as a first confidence coefficient.
2. The method of claim 1, wherein the acquiring a plurality of oral dentition images as the first image comprises:
acquiring a plurality of oral dentition images;
the sizes of the oral cavity dentition images are processed into preset sizes;
interlacing sampling is carried out on the oral cavity dentition image after the size processing to obtain a plurality of sub-images;
and taking the spliced image of the plurality of sub-images as a first image.
3. The method for training a model for detecting a constant molar position according to claim 1, wherein the performing the first feature extraction on the first image to obtain a first feature image comprises:
performing downsampling feature extraction on the first image to generate feature images with different sizes;
the feature images of different sizes are taken as first feature images.
4. The method of claim 1, wherein the second feature extraction is an abstract feature extraction.
5. A model training system for detecting a constant molar position, comprising:
the acquisition module is used for acquiring a plurality of oral cavity dentition images as first images;
the first feature extraction module is used for carrying out first feature extraction on the first image to obtain a first feature image;
the activation module is used for selectively activating the first characteristic image in the channel dimension and the space dimension respectively;
the second feature extraction module is used for carrying out second feature extraction on the first feature image after the selection activation;
the prediction module is used for predicting a detection frame of the constant molar position according to the image features after the second features are extracted;
the fusion module is used for fusing the information of the detection frame to generate a first confidence coefficient;
the confidence coefficient selection module is used for taking the maximum first confidence coefficient as a prediction output result;
the parameter updating module is used for reversely updating the parameters of the detection model according to the prediction output result;
the selectively activating the first feature image in the channel dimension includes:
carrying out average pooling and maximum pooling on the first characteristic image;
selectively activating the first characteristic images after the average pooling and the maximum pooling in the channel dimension by adopting an attention mechanism;
in the selective activation of channel dimensions, the calculation is performed using equation 1:
equation 1
Wherein,for features after selective activation via channel dimension, < ->Representing the image features after the averaging pooling,representing the image features after maximum pooling, < >>And->For two different matrix parameters representing the weight sizes for different channel numbers,/->Representing an activation function sigmoid;
the selective activation of the first feature image in the spatial dimension is specifically:
selectively activating the first image processed by adopting a convolution check attention mechanism in a space dimension;
in the selective activation of the spatial dimension, the calculation is performed using equation 2:
equation 2
Wherein,representing 7*7 convolution operation,/->Representing image features averaged pooling on the basis of features obtained by selective activation of channel dimensions, a>Representing image features resulting from maximum pooling based on features obtained by selective activation of channel dimensions, and>representing an activation function sigmoid;
the fusing the information of the detection frame to generate a first confidence coefficient comprises the following steps:
fusing the position information of the detection frame, the second confidence coefficient and the image characteristic information;
generating a first confidence coefficient according to the fusion result, wherein the second confidence coefficient is the confidence coefficient of the image, and the first confidence coefficient is the confidence coefficient of the same image after fusion processing with the second confidence coefficient;
when information of the detection frames is fused, a formula 3 and a formula 4 are adopted when information of different modes is fused:
equation 3
Equation 4
Wherein,for the image feature after fusing the second confidence information,/I>To fuse the confidence information after image features,and->Original image characteristics and confidence information respectively;
on the basis of fusing information of different modes, calculating by combining the position information of the image to obtain the confidence coefficient of the detection frame fusing the multi-mode information, wherein the specific calculation process is shown in a formula 5-a formula 8:
equation 5
Equation 6
Equation 7
Equation 8
Wherein,representing features fused with the second confidence information and the image feature information,/for the feature>For the position information of the encoded detection frame, < >>Then indicate at +.>Further fusing the features of the position information of the detection frame,/on the basis of (a)>And representing the new confidence coefficient finally obtained after the multi-mode information is fused as a first confidence coefficient.
6. A model training system for detecting a constant molar position, comprising:
at least one memory for storing a program;
at least one processor for loading the program to perform the method of model training for detection of permanent molar positions according to any one of claims 1-4.
7. A storage medium having stored therein a processor executable program which when executed by a processor is for performing the constant molar position detection model training method according to any one of claims 1-4.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011406995.3A CN112561865B (en) | 2020-12-04 | 2020-12-04 | Method, system and storage medium for training detection model of constant molar position |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011406995.3A CN112561865B (en) | 2020-12-04 | 2020-12-04 | Method, system and storage medium for training detection model of constant molar position |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112561865A CN112561865A (en) | 2021-03-26 |
CN112561865B true CN112561865B (en) | 2024-03-12 |
Family
ID=75048486
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011406995.3A Active CN112561865B (en) | 2020-12-04 | 2020-12-04 | Method, system and storage medium for training detection model of constant molar position |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112561865B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113343853B (en) * | 2021-06-08 | 2024-06-14 | 深圳格瑞健康科技有限公司 | Intelligent screening method and device for dental caries of children |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110246571A (en) * | 2019-07-30 | 2019-09-17 | 深圳市倍康美医疗电子商务有限公司 | Tooth data processing method |
CN110349224A (en) * | 2019-06-14 | 2019-10-18 | 众安信息技术服务有限公司 | A kind of color of teeth value judgment method and system based on deep learning |
CN110619350A (en) * | 2019-08-12 | 2019-12-27 | 北京达佳互联信息技术有限公司 | Image detection method, device and storage medium |
CN110930421A (en) * | 2019-11-22 | 2020-03-27 | 电子科技大学 | Segmentation method for CBCT (Cone Beam computed tomography) tooth image |
CN111882002A (en) * | 2020-08-06 | 2020-11-03 | 桂林电子科技大学 | MSF-AM-based low-illumination target detection method |
CN111986217A (en) * | 2020-09-03 | 2020-11-24 | 北京大学口腔医学院 | Image processing method, device and equipment |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7717708B2 (en) * | 2001-04-13 | 2010-05-18 | Orametrix, Inc. | Method and system for integrated orthodontic treatment planning using unified workstation |
EP2593036B1 (en) * | 2010-07-12 | 2017-12-20 | Centre De Recherche Medico Dentaire Am Inc. | Dental analysis method and system |
WO2018022752A1 (en) * | 2016-07-27 | 2018-02-01 | James R. Glidewell Dental Ceramics, Inc. | Dental cad automation using deep learning |
US10751548B2 (en) * | 2017-07-28 | 2020-08-25 | Elekta, Inc. | Automated image segmentation using DCNN such as for radiation therapy |
WO2019204520A1 (en) * | 2018-04-17 | 2019-10-24 | VideaHealth, Inc. | Dental image feature detection |
-
2020
- 2020-12-04 CN CN202011406995.3A patent/CN112561865B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110349224A (en) * | 2019-06-14 | 2019-10-18 | 众安信息技术服务有限公司 | A kind of color of teeth value judgment method and system based on deep learning |
CN110246571A (en) * | 2019-07-30 | 2019-09-17 | 深圳市倍康美医疗电子商务有限公司 | Tooth data processing method |
CN110619350A (en) * | 2019-08-12 | 2019-12-27 | 北京达佳互联信息技术有限公司 | Image detection method, device and storage medium |
CN110930421A (en) * | 2019-11-22 | 2020-03-27 | 电子科技大学 | Segmentation method for CBCT (Cone Beam computed tomography) tooth image |
CN111882002A (en) * | 2020-08-06 | 2020-11-03 | 桂林电子科技大学 | MSF-AM-based low-illumination target detection method |
CN111986217A (en) * | 2020-09-03 | 2020-11-24 | 北京大学口腔医学院 | Image processing method, device and equipment |
Non-Patent Citations (2)
Title |
---|
基于SIFT算法的图像特征点提取和匹配研究;王心醉;董宁宁;李欢利;;南京医科大学学报(自然科学版)(第02期);全文 * |
基于特征融合和中心预测的目标检测算法;段志伟;;现代计算机(第09期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN112561865A (en) | 2021-03-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12086964B2 (en) | Selective image modification based on sharpness metric and image domain | |
US11730569B2 (en) | Generation of synthetic post treatment images of teeth | |
US20240005567A1 (en) | Generation of images with tooth color determined using depth information | |
US20220218449A1 (en) | Dental cad automation using deep learning | |
CN111563887B (en) | Intelligent analysis method and device for oral cavity image | |
CN110473243B (en) | Tooth segmentation method and device based on depth contour perception and computer equipment | |
CN107644225A (en) | Pulmonary lesionses recognition methods, device and realization device | |
US11887209B2 (en) | Method for generating objects using an hourglass predictor | |
US10219703B2 (en) | Method and system for intra-oral imagine using HDR imaging and highlight removal | |
CN113379697B (en) | Color image caries identification method based on deep learning | |
KR20200110878A (en) | A system and method for early diagnosing dental caries based on the deep learning | |
CN112561865B (en) | Method, system and storage medium for training detection model of constant molar position | |
CN114586069A (en) | Method for generating dental images | |
FR3111066A1 (en) | Method and device for three-dimensional reconstruction of a face with toothed part from a single image | |
CN113343853B (en) | Intelligent screening method and device for dental caries of children | |
WO2024208188A1 (en) | Underwater image enhancement method based on brightness-mask-guided multi-attention mechanism | |
EP3821393B1 (en) | Method for transforming a view of a 3d model of a dental arcade into a photorealistic view | |
CN116630230A (en) | Focal region segmentation system based on ophthalmic ultra-wide angle image | |
CN116309432A (en) | Lightweight caries detection system and method based on YOLOv5 and IGCC 3 fusion | |
KR20210011050A (en) | Method for predicting dental caries area and device for predicting dental caries area using the same | |
CN116993702A (en) | Method and related device for realizing single frame data gum separation by deep learning | |
TWI843109B (en) | Method for identifying medical image, computer device and computer readable storage medium | |
US20220198613A1 (en) | Automatic Reduction of the Appearance of Mach Band Effect on Radiographs | |
CN118537316A (en) | Real-time ultrasonic image interpretation system and method based on machine learning model | |
CN116681780A (en) | Tooth colorimetry method, apparatus, medium and program based on variable component self-encoder |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
CB02 | Change of applicant information |
Address after: 805, Building B, Second Unified Building, Houhai Neighborhood Committee, No.1 Nanyou Huaming Road, Nanshan District, Guangdong Province, 518000 Applicant after: Shenzhen Gree Health Technology Co.,Ltd. Address before: 805, block B, No.2 Tongjian building, Houhai neighborhood committee, No.1 Huayou Huaming Road, Nanshan District, Shenzhen, Guangdong 518000 Applicant before: Shenzhen Gree Health Management Co.,Ltd. |
|
CB02 | Change of applicant information | ||
GR01 | Patent grant | ||
GR01 | Patent grant |