US20170262598A1 - Content-based medical image rendering based on machine learning - Google Patents
Content-based medical image rendering based on machine learning Download PDFInfo
- Publication number
- US20170262598A1 US20170262598A1 US15/191,043 US201615191043A US2017262598A1 US 20170262598 A1 US20170262598 A1 US 20170262598A1 US 201615191043 A US201615191043 A US 201615191043A US 2017262598 A1 US2017262598 A1 US 2017262598A1
- Authority
- US
- United States
- Prior art keywords
- rendering
- machine
- image
- training
- parameters
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000009877 rendering Methods 0.000 title claims abstract description 196
- 238000010801 machine learning Methods 0.000 title claims description 30
- 238000012549 training Methods 0.000 claims abstract description 81
- 238000013473 artificial intelligence Methods 0.000 claims abstract description 29
- 238000013135 deep learning Methods 0.000 claims abstract description 16
- 238000000034 method Methods 0.000 claims description 44
- 230000015654 memory Effects 0.000 claims description 36
- 238000013461 design Methods 0.000 claims description 27
- 230000006870 function Effects 0.000 claims description 18
- 230000002787 reinforcement Effects 0.000 claims description 17
- 238000012546 transfer Methods 0.000 claims description 17
- 238000012545 processing Methods 0.000 claims description 14
- 230000002452 interceptive effect Effects 0.000 claims description 9
- 239000000463 material Substances 0.000 claims description 9
- 238000013528 artificial neural network Methods 0.000 claims description 5
- 230000003190 augmentative effect Effects 0.000 claims description 5
- 230000003094 perturbing effect Effects 0.000 claims description 3
- 238000003384 imaging method Methods 0.000 abstract description 9
- 230000009471 action Effects 0.000 description 15
- 238000012800 visualization Methods 0.000 description 14
- 239000000523 sample Substances 0.000 description 12
- 210000001519 tissue Anatomy 0.000 description 8
- 230000000007 visual effect Effects 0.000 description 8
- 230000005540 biological transmission Effects 0.000 description 7
- 239000003795 chemical substances by application Substances 0.000 description 6
- 238000003745 diagnosis Methods 0.000 description 5
- 230000003993 interaction Effects 0.000 description 5
- 238000013459 approach Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 238000005286 illumination Methods 0.000 description 4
- 239000011159 matrix material Substances 0.000 description 4
- 238000005259 measurement Methods 0.000 description 4
- 230000007170 pathology Effects 0.000 description 4
- 238000013442 quality metrics Methods 0.000 description 4
- 208000000913 Kidney Calculi Diseases 0.000 description 3
- 206010029148 Nephrolithiasis Diseases 0.000 description 3
- 238000002059 diagnostic imaging Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 230000003902 lesion Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000002604 ultrasonography Methods 0.000 description 3
- 238000007794 visualization technique Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000001934 delay Effects 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- 238000002156 mixing Methods 0.000 description 2
- 238000002600 positron emission tomography Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000004088 simulation Methods 0.000 description 2
- 238000002603 single-photon emission computed tomography Methods 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 206010006187 Breast cancer Diseases 0.000 description 1
- 208000026310 Breast neoplasm Diseases 0.000 description 1
- 238000010521 absorption reaction Methods 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 238000000149 argon plasma sintering Methods 0.000 description 1
- 230000000740 bleeding effect Effects 0.000 description 1
- 230000018678 bone mineralization Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 230000000747 cardiac effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 238000013329 compounding Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000002591 computed tomography Methods 0.000 description 1
- 238000005094 computer simulation Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 230000001605 fetal effect Effects 0.000 description 1
- 238000002594 fluoroscopy Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 210000003734 kidney Anatomy 0.000 description 1
- 238000002357 laparoscopic surgery Methods 0.000 description 1
- 230000031700 light absorption Effects 0.000 description 1
- 210000004185 liver Anatomy 0.000 description 1
- 238000010339 medical test Methods 0.000 description 1
- 230000003278 mimic effect Effects 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000002310 reflectometry Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
- 230000003746 surface roughness Effects 0.000 description 1
- 230000008685 targeting Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Images
Classifications
-
- G06F19/345—
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24133—Distances to prototypes
- G06F18/24137—Distances to cluster centroïds
- G06F18/2414—Smoothing the distance, e.g. radial basis function networks [RBFN]
-
- G06F19/321—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/004—Artificial life, i.e. computing arrangements simulating life
- G06N3/006—Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/005—General purpose rendering architectures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/06—Ray-tracing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/08—Volume rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
- G06T15/506—Illumination models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/20—ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computing arrangements based on specific mathematical models
- G06N7/01—Probabilistic graphical models, e.g. probabilistic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/41—Medical
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
Definitions
- the present embodiments relate to medical imaging of three-dimensional (3D) scans.
- Data representing a volume is rendered for visualization. Due to the many different scan settings and patient variability, renderings for different patients or at different times appear different.
- Variability is particularly troublesome for physically-based volume rendering, which relies on the physical simulation of light propagation (e.g., unbiased path tracing).
- Physically-based visualization techniques produce global illumination effects in computer-generated graphics that mimic the real-world interaction of light with various 3D objects or tissues. This results in physically plausible images that are often easier for the human brain to interpret when compared to the more analytical images from traditional rendering. For example, subtle ambient occlusions, shadows, and color bleeding provide important depth cues for the spatial understanding of 3D relationships between structures in a single 2D image, whereas simpler visualization techniques may require additional interaction with the viewing parameters (e.g., moving the virtual camera around the 3D data) to obtain the same spatial information from the image parallax.
- the preferred embodiments described below include methods, systems, instructions, and computer readable media for machine training an artificial intelligence and use of that artificial intelligence to provide rendering settings.
- settings of multiple rendering parameters may be provided for consistent imaging or visualization even in physically-based rendering.
- a method for content-based rendering based on machine learning in a rendering system.
- a medical dataset representing a three-dimensional region of a patient is loaded from memory.
- a machine applies the medical dataset to a machine-learnt non-linear model.
- the machine-learned model is trained with deep learning to extract features from the medical dataset and trained to output values for two or more physically-based rendering parameters based on input of the medical dataset.
- At least one of the two or more physically-based rendering parameters are controls for consistent data handling, lighting design, viewing design, material propriety, or internal renderer property.
- a physically-based renderer renders a photorealistic image of the three-dimensional region of the patient using the output values resulting from the applying. The photorealistic image is transmitted.
- a method for machine training for content-based rendering in a machine training system.
- First volume data of a volume of a patient, a first image of the volume, and first values of path tracing rendering parameters are input to training of an artificial intelligence.
- a machine trains the artificial intelligence to output second values of the path tracing rendering parameters for a second volume data where the second values provide a second rendered image of the second volume modeled on the first image.
- the trained artificial intelligence is stored.
- a system for content-based rendering based on machine learning.
- a medical scanner is configured to scan a patient.
- a machine is configured to output settings for rendering parameters by application of data from the scan to a machine-learnt model.
- the rendering parameters include windowing, transfer function, and lighting, and the settings are learned to provide a first image from the data similar to one or more second images for a same diagnostic context.
- a graphics processing unit is configured to render the first image using the settings.
- FIG. 1 is a flow chart diagram of one embodiment of a method for machine training for content-based rendering
- FIG. 2 is a flow chart diagram of an embodiment of a method for content-based rendering based on machine learning in a rendering system
- FIG. 3 is a block diagram of one embodiment of a system for content-based rendering based on machine learning.
- Content-based photorealistic rendering of medical images is based on machine learning.
- the machine learnt non-linear model determines rendering settings based on the content of the 3D medical scan. Due to the variability between scan parameters, data contrast, noise, and/or the large number of rendering parameters, producing rendered images that allow for consistent visual interpretation has been a long standing challenge, which has prevented wider adoption 3D images for definitive decision support.
- deep learning generates the settings for the rendering parameters for the photorealistic rendering components, establishing a consistent visualization design.
- This design may be applied to new 3D image data in order to provide reliable visual interpretation within the diagnostic workflow.
- Image features, and optionally, available non-image data are mapped to sets of rendering parameters that produce optimal images, where the relationship between the two is highly non-linear in the general case.
- Consistent visual interpretation for 3D images may potentially simplify diagnostic workflow and may promote the clinical use of 3D data for decision support.
- Consistent photorealistic rendered images may be provided in new platforms and workflows, such as fully automated reporting and cloud-based rendering services.
- Consistent visualization design as a machine learning problem may be applied 1) to optimize rendering preset designs tailored to a large variety of data types in a unified and scalable way, 2) to adapt rendering parameters to produce target consistent results based on the provided input, 3) to adapt rendering parameters to specific individual preferences, 4) to transfer expert trained results to non-expert users, 5) to increase the conspicuity of visual features in the rendered images (e.g., pathologies in medical images) and 6) to provide automatic rendering presets for specific workflows.
- a scalable framework is provided for fully automated image generation in the clinical workflow.
- FIG. 1 shows one embodiment of a method for machine training for content-based rendering in a machine training system.
- Optimal rendering parameters may be obtained via training an artificial intelligence system.
- the input of the artificial intelligence system may include, but is not limited to, the 3D scan data to be rendered, images rendered from the 3D scan data, information on the patient, and/or information on the diagnostic purpose.
- the machine training learns to provide rendering settings for photorealistic rendering based on input data for a specific patient.
- the output of the trained artificial intelligence system is the optimal set of settings for the rendering parameters.
- the method of FIG. 1 is implemented by a processor and a graphics processing unit in a server, computer, or other machine.
- the system of FIG. 3 implements the method of FIG. 1 , with or without the medical imager 56 and/or display 54 .
- data is input by loading from a memory
- a processor performs the training based on the data
- the memory stores the learnt artificial intelligence.
- the medical scanner and/or memory input the data to a server, which performs the training and outputs to a memory.
- results may be fed back as input in a repetition of the acts.
- act 22 is not provided, such as where the trained artificial intelligence is used rather than stored for later use.
- training data is input.
- the training data is gathered from one or more memories, sensors, and/or scanners.
- the gathered data may be stored and/or transmitted to a buffer, memory, cache, processor, or other device for training. This gathered training data is input to the machine learning algorithm.
- scan data is provided.
- the scan data represents a patient.
- a two-dimensional (2D), 3D, 2D+time sequence, 3D+time sequence, and/or other image or scan data may be used.
- Patient non-image data may be input to the training.
- Higher dimensional data includes but is not limited to patient information such as a) natural language medical documents or information extracted therefrom of previous diagnosis and/or clinical procedures; b) previous medical test results; c) previously processed medical images and measurements; d) computational modeling results; e) parametric image information derived from medical image analysis algorithms (e.g. cardiac strain map or elasticity); and/or f) other information (e.g., scanner settings).
- the goal or ground truth information is input.
- the artificial intelligence system is trained with labeled data (e.g., pairs of input and corresponding desired output (e.g., rendering parameters, optimal rendered images, and/or photographs/movies of the real object)). Based on the provided data, the artificial intelligence system is trained to produce the desired output from any input.
- the ground truth is provided by desired settings of the rendering parameters for each scan dataset. The values of the path tracing rendering parameters for desired results of each sample are input to the training. Alternatively or additionally, the ground truth is provided as a desired rendered image for each scan dataset. A rendered image for each scan dataset is provided. In other embodiments, an image with desired visualization is provided for multiple datasets. Using a renderer, the machine training may learn the settings that produce similar visualization.
- the input samples may provide both negative and positive training samples. For example, scan datasets and settings providing desired and undesired images are input. The training learns to distinguish the desired settings from both good and bad settings.
- additional samples are created by perturbing the path tracing rendering parameters of an input sample.
- the perturbing creates a collection of sets of the path tracing rendering parameters for each scan data set.
- the artificial intelligence system is applied on the input scan dataset to generate rendering parameters, which are then perturbed to generate a pool of rendering parameters.
- Rendering is then performed using the sets of rendering parameters in the pool to produce a pool of corresponding images. Human users or other operators then visually examine the produced images and select the satisfying images.
- the rendering parameters for the selected images are then used as desired output and/or a sub-set as positive ground-truth.
- the rendering parameters for the unselected sub-set of images are not used or are used as negative ground-truth.
- a web-based or network-based service In another approach to gather samples for input, a web-based or network-based service is provided.
- a website or network portal allows a user to upload the input data (i.e., image and additional information).
- the server or other processor then runs the artificial intelligence system and/or perturbs possible rendering settings to renderer a pool of images.
- the service then provides the generated images to the user and asks the user to select the satisfying images.
- the rendering parameters corresponding to the user selected images are then stored as the desired output for the input (i.e., ground truth).
- the input may be repeated.
- the training of the artificial intelligence system is performed periodically or multiple different times. After additional samples are collected (e.g., through the web-based service), the training is repeated using the additional samples with or without previous samples used for earlier iterations of the training.
- the artificial intelligence or intelligences are machine trained.
- the collected samples including ground truth, are used to train.
- a machine such as a processor, computer, server, or other device learns from the samples.
- complex statistical relationships between large numbers e.g., tens, hundreds, thousands, or more
- input variables e.g., tens, hundreds, thousands, or more
- output variables e.g., tens, hundreds, thousands, or more
- the training is for imaging in any context.
- One model may be learned for any number of imaging situations.
- different models are trained for different situations.
- the different situations may include different scan modalities (e.g., different model for computed tomography, magnetic resonance, ultrasound, positron emission tomography, and single photon emission computed tomography).
- the different situations may include different types of tissue of interest (e.g., liver versus kidney), different diagnostic purpose or workflow (e.g., cancerous lesion versus bone calcification), and/or different users (e.g., different operators may have different preferences for visualization).
- Regression, classification, and/or reinforcement learning are used.
- Regression training learns a range or continuous output by minimization of a metric.
- Classification learns disparate outputs.
- Reinforcement learning learns through a sequence of actions with feedback.
- Neural network, Bayes network, or support vector machine training may be used.
- Hierarchal or other approaches may be used.
- Supervised or semi-supervised machine learning may be used.
- features are extracted from the input data.
- Haar wavelet, steerable, or other features may be extracted from the scan dataset or images. Measures or other information may be extracted from non-image data.
- the input data itself e.g., scalar values of the scan dataset.
- the machine learns features as part of the training.
- deep learning e.g., deep structured learning, hierarchical learning, or deep machine learning
- a deep neural network processes the input via multiple layers of feature extraction to produce features used to derive outputs of settings for rendering.
- the deep learning provides the features used by other machine training to learn to output the rendering settings.
- Other deep learnt, sparse auto-encoding models may be trained and applied.
- the machine training is unsupervised in learning the features to use and how to classify given an input sample (i.e., feature vector).
- Deep learning uses deep learning to improve the input features required for the machine learning. This is especially useful with render settings, which may not have an explicit natural relationship with the input data. Deep learning figures out the data features that end up providing a good classifier or model. Deep learning is known to be effective in extracting high level abstractions (features) from low level input data (e.g., raw images) that is suitable for the trained task, so may be more convenient than and superior to hand-crafted features for identifying features for render settings. In addition, learning-based automatically rendering setting selection is a less studied problem, with no hand-craft feature specifically designed for this problem. Standard hand-craft features, like HAAR feature or SIFT feature, designed for other general computer vision tasks may not be suitable for rendering setting.
- HAAR feature or SIFT feature designed for other general computer vision tasks may not be suitable for rendering setting.
- the machine trains to output rendering settings to provide consistent imaging.
- the training learns to provide rendering settings that result in a rendered image with similar or same visualization despite differences in the content of the data.
- the path tracing or other physically-based rendering settings so that the rendered image for a given input scan dataset looks and feels similar to an ideal are learned.
- the machine learns to output renderer settings to model the resulting rendered image for a given scan dataset on a desired image.
- the learned sets of parameters establish a visualization design that may be reused when new data is introduced to the system in order to provide consistent rendered images. In this way, the machine may learn renderer settings to provide rendering to a standard. Rather than render with the same preset values for a given context, the values are based on a given input dataset.
- consistency refers to consistent data handling, consistent viewing, and/or consistent visual styles, which may be utilized both in automated and interactive clinical workflows.
- the ground truth provides consistency. Where the ground truth uses similar or consistent images and/or settings that result in similar or consistent images despite variance in the scan datasets, the training learns to output settings providing the consistency.
- the notion of optimal rendering parameters and optimal output images relates to using objective quality measurements, subjective quality measurements, clinical significance ratings, and/or other indicators of consistency.
- the consistency is learned using a measured or calculated metric.
- the training uses a metric of similarity. Any metric of similarity may be used.
- Example computed metrics may be visual perception metrics based on visual systems, comparative differences, perceptual hash, mutual information, correlation, computed feature signals, or simple error metrics.
- the metrics may be automatically computed or provided by the users.
- the metric is used as an indicator of strength of importance of a given training sample.
- one or more metrics are used in reinforcement learning.
- deep reinforcement learning uses a similarity of the rendered image from a sample to a reference image. The similarity is used as a reinforcement.
- the output image rendered from a given set of rendering settings for a sample of scan data and the associated quality metrics are fed-back into the learning system for additional refinement of the learned parameters (e.g., in deep supervised and/or reinforcement learning).
- a metric measuring the similarity of one rendered image to one or a group of rendered images with the desired quality is calculated, and used as the “reward” to train the artificial intelligent agent using the deep reinforcement learning technique.
- quality metrics than similarity may be used.
- the choice of metric to use may depend on the types of features relevant to a specific clinical data domain. Depending on the diagnosis of interest or medical workflow, different pathologies may be of interest. Different rendering settings increase the conspicuity of different pathologies.
- the quality metric in the reinforcement learning is used to learn to provide the optimal rendering parameters for increasing the conspicuity of the pathology of interest.
- the artificial intelligent agent is a deep neural network that takes in the output of the reinforcement agent as its input, and outputs one quality or similarity metric (referred to as “action-value”) for each possible action (e.g., set of rendering settings), indicating the effect of taking these actions (i.e., higher values are associated with actions that can improve the rendering quality).
- action-value quality or similarity metric
- the agent also analyzes the action-values, and selects one action to perform.
- the set of rendering parameters providing the highest implementable quality rendering are selected as the output.
- a probability distribution of different values of the metric is used for the selection.
- the selected action is the action with highest action-value or randomly selected from the actions with a probability distribution assigned based on the rank of action-values. In a typical scenario, actions with higher action-values are assigned higher probability, but any assignment may be used.
- the training is modeled after steps taken by an expert or other user to achieve the desired visualization for the training data.
- the user's sequence of adjustments to provide the desired rendering from the scan dataset is monitored (e.g., select transfer function T 1 , then window function W 1 , then select a different transfer function T 2 , then select a material reflectance MR 1 , . . . ).
- the learning uses metrics for each stage or step to learn the settings used for each.
- the actions of the expert are extrapolated to different training datasets or included for each training dataset.
- the training learns to provide final rendering settings based on the monitored sequence.
- a photograph or video of the imaged tissue, imaged object, or other real-world references is used.
- the physically-based rendering provides photorealistic images.
- the training learns to provide rendering settings modeled after real images.
- the desired output is labeled using actual optical images so that the rendering parameters result in a matching look and/or feel.
- Other modes of images may be used for the ground truth or reinforcement metric, so that the training learns to output rendering settings modeled after the different modality (e.g., render CT data to match the look of rendered MR data).
- one model is trained to output settings for multiple rendering parameters.
- a hierarchy or other grouping of models are trained to output the settings for multiple rendering parameters.
- the machine is trained to output settings of rendering parameters. By outputting settings for two or more rendering parameters for a given scan dataset, less manual setting of parameters is needed. In path tracing or other physically-based rendering, the amount of processing and time required to provide the photorealistic imaging may cause delays. Compounding such delays with user adjustment, it may be overly time consuming to provide a desired rendering manually.
- the machine trained model contributes to less delay in achieving the desired rendered image.
- the model may be trained to output any number and/or type of rendering parameters.
- one or more data consistency parameters are to be output.
- Data consistency parameters include windowing, scaling, level compression, data normalization, or others.
- one or more transfer function design parameters are to be output. Transfer function design parameters include classification look-up tables, multi-dimensional transfer functions, tissue-specific transfer functions, or other transfer functions.
- one or more lighting design parameters are to be output. Lighting design parameters include type of virtual lights, position of the virtual light sources, orientation of the virtual light sources, image-based lighting sources, or others.
- one or more viewing design parameters are to be output. Viewing design parameters include type of camera, position of the camera, orientation of the camera, intrinsic parameters for viewing, or others.
- one or more use-case specific parameters are to be output. Use-case specific parameters are settings specific to a given use, such as a particular camera position for a given type of medical report or use of two cameras for stereoscopic viewing.
- the parameters may be specific to physically-based rendering, such as internal render properties.
- the setting for any parameter for path tracing, unbiased path tracing, Monte Carlo rendering, global illumination, or other simulations of light propagation may be learned.
- a global albedo, or surface reflectance coefficient may be learned for the volume data, which describes the relative importance of light scattering versus light absorption.
- Settings for any parameters for material property controlling the interaction with light may be learned.
- the transparency, reflectivity, surface roughness, and/or other properties may be modeled in the physically-based rendering, so included as rendering parameters.
- Albedo may also be learned as a per-material property for the different tissue classes in the volume data. Since the effect of various settings of material type or light interaction may not be instinctive to a user, training the model with machine learning to provide the initial settings may assist the user.
- All or a sub-set of the rendering parameters are to be output by the machine-learnt model.
- the model is trained to output any number of data consistency, transfer function, lighting, and viewing parameters.
- the model is trained to provide settings for all of the rendering parameters.
- the model is trained to provide settings for some but not all of the rendering parameters. For example, all of the rendering parameters are output except for the view design parameters, which are under the control of the user. The viewing camera parameters are under the control of the user, so the machine learning is for settings of other parameters. The user controls the view design parameters independently.
- the model is trained to output settings for multiple parameters regardless of the user-controlled parameters.
- the user-selected parameters are an input to the classification so that the output settings account for the user-selected settings.
- the model learns to output rendering parameter settings based, in part, on settings for one or more parameters under the control of the user.
- the artificial intelligence is trained to output the values for the rendering parameters based on user selected viewing camera parameters. Subsets of the rendering parameters are used as additional inputs in the training phase of the machine learning algorithm. In that case, the machine learning algorithm output is an optimized subset of the rendering parameters.
- the user input settings may be optimized or not while the other settings are optimized by the model.
- the input to the training is a higher dimension vector including the input scan dataset and reference viewing parameters.
- the artificial intelligence is trained to account for different ambient conditions in the real-world viewing environment of the user.
- the settings for “ambient-varying” rendering of a volume are learned.
- the training data includes different ambient light conditions and measures of the ambient light using light probes.
- the model is trained to account for ambient luminosity of the room. The luminosity is sent to the trained model, which adapts the rendering parameters automatically to provide the same content-based realistic rendering.
- the training may be repeated.
- the user may adjust some of the output settings. This adjustment, the setting, and/or image resulting from the adjustment are added to the training data.
- the machine learning model is refined by repeating the learning with the addition feedback as samples from which to learn.
- the manually adjusted settings of the parameters define personalized visualization designs. Other sources of further training data may be used.
- the repetition of the training may be performed separately for different users, physicians, medical groups, hospitals, or other grouping. Personalized training is used so that consistent images for a relevant group are provided. In other embodiments, the repetition is provided for all users.
- the trained artificial intelligence is stored.
- the result of the training is a matrix.
- the matrix represents the learned knowledge through machine training.
- the matrix includes an input vector and outputs for the settings of the rendering parameters.
- Other machine-learnt model representations may be used, such as a hierarchy of matrices or other non-linear models.
- the trained model is stored in a memory. Any memory may be used.
- the memory used for the training data may be used.
- the memory may be in other devices.
- the trained model is stored in a memory of a server.
- the server uses the trained model to output rendering parameters to clients or to a server-based renderer which outputs rendered images to the clients.
- multiple copies of the trained model are provided to different physicians, medical scanners, and/or workstations for use by different physicians.
- FIG. 2 shows one embodiment of a method for content-based rendering based on machine learning in a rendering system.
- the machine-learnt model or artificial intelligence agent is applied.
- a rendering of a scan dataset is to be performed.
- the artificial intelligence agent is applied to the scan dataset with or without other input information, outputting values for the rendering parameters.
- the values result in consistency in photorealistic rendering, possibly avoiding time consuming manual adjustment of various rendering parameters to achieve the desired visual.
- the method is implemented by a computer, server, or other processor with a graphics processing unit or other renderer and a display.
- medical data is loaded from memory.
- a light sensor measures the ambient light or captures a light probe image, and a computer applies the machine-learnt model, outputting values for rendering.
- a renderer uses the values to render from the medical data, resulting in transmission of the rendered image to a display.
- Different devices may be used, such as a controller or medical scanner processor performing the application and/or rendering.
- acts may be provided.
- the ambient light is not measured in act 32 .
- the rendering of act 42 and transmission of the photorealistic image of act 44 are not provided, but instead the values of the rendering parameters are stored or transmitted.
- acts for manual adjustment and/or initial setting of one or more rendering parameters are provided, such as for an interactive rendering workflow.
- none, one or two of acts 36 - 40 are performed without others of that set of acts.
- the acts are performed in the order shown (top to bottom) or other orders.
- the ambient light may be measured in act 32 prior to loading the medical data in act 30 and/or in parallel with performing act 34 .
- medical data is loaded from a memory, sensors, and/or other source.
- a medical scanner may provide the data, such as a medical dataset representing a 3D region of the patient. Mining of a patient's computerized medical record may be used. Medical information from a picture archiving and communications server may be obtained. Any source may be used.
- the medical data is for a given patient. While the training may use samples from many patients to learn features and/or learn to classify input medical data to provide values for rendering, the learnt model is applied to the medical data for a patient to output values for rendering parameters for that patient. The same or different machine-learnt models are applied to data for any number of patients.
- Scan data representing a 3D volume is loaded as a medical dataset.
- the scan data may be from multiple two-dimensional scans or may be formatted from a 3D scan.
- the medical data may include non-image or non-scan data, such as patient test results, physician notes, lab results, diagnosis information, billing codes, clinical data, patient history, or other information.
- the medical data includes the information learned to be used by the machine-trained model. Ground truth information and/or similarity or quality metric measures may be not provided. Where the machine learning learned to use feedback of predicted ground truth, then that feedback may be created and used as an input.
- the content to be used to determine the settings of the rendering parameters is provided to the machine-learnt model. For example, just the 3D scan data is provided as a medical dataset for a given patient. As another example, the 3D scan data and diagnostic information are provided as the medical dataset for the patient. The diagnostic information may be used to select the machine-learnt model to use and/or be part of the input feature vector.
- Some machine-learnt models may be robust or still operate with some of the input feature vector missing. Alternatively, all of the information needed for the input feature vector is loaded.
- a sensor measures ambient light. Any luminosity sensor may be used, such as a charge coupled device. The sensor provides a measure of the luminosity. This measure may be used as an input in the input feature vector and/or to select a particular machine-learnt model to use. In alternative embodiments, the rendering is not based on ambient lighting conditions, so no measurement of ambient light is provided.
- a light probe image is obtained as the measure of ambient light with a camera as the sensor.
- the image is obtained using any of a variety of ways and then used for high dynamic range image-based illumination during rendering. Examples ways for obtaining light probe images include capturing high-dynamic range photographs of a mirrored ball, stitching multiple photographs, or using self-contained panoramic cameras.
- the medical dataset is applied to the machine-learnt model.
- the loaded medical dataset is processed by a machine, such as a computer, processor, or server.
- the machine uses the internal representation of the machine-learnt model.
- the input feature vector is created and/or is the medical dataset.
- the machine inputs the feature vector, resulting in the machine-learnt model outputting the rendering settings based on the learned knowledge.
- Acts 36 - 40 represent different acts in applying the machine-learnt model of presets for one or more rendering parameters. Additional, different, or fewer acts may be provided.
- the machine-learned model was previously learned using any approach for machine learning, so may be a regression, classification, or reinforcement-learnt model.
- the machine-learned model is trained with deep learning (e.g., deep neural network) to extract features from the medical dataset.
- deep learning e.g., deep neural network
- features are extracted from the medical dataset.
- One or more other features may be extracted from the medical dataset using manually or user configured features.
- These features may then be used by another machine-learnt model and/or another hierarchal stage to output the values for the rendering. For example, a further reinforcement-learnt model is applied.
- the extracted features are input to the machine-learnt model for application.
- the application of the machine-learnt model is more likely to output values in act 38 resulting in the photorealistic image corresponding to a standard image despite differences in the medical dataset.
- the photorealistic image is created using path tracing, other physically-based rendering, or reference photographs.
- the model uses the input feature vector to predict the values for rendering that result in the rendering of the scan data as an image with desired characteristics.
- the standard image may be a group of images having similar characteristics, such as resolution and/or contrast for a particular type of tissue or lesion.
- the rendering values output were learnt in order to provide the same or similar resolution and/or contrast for that particular type of tissue.
- Different machine-learnt models may be used for different diagnostic and/or workflow situations, such as outputting values for breast cancer imaging or for fetal imaging.
- the ambient light measured in act 32 is applied to the machine-learnt model with or as part of the medical dataset.
- the machine-learnt model is trained to output the values based in part on the ambient light.
- the machine-learnt model outputs values for the rendering parameters that account for ambient light.
- Photorealistic images may be output as if lit by the ambient light in a room or to which the patient is subjected. Where the photorealistic images are used for augmented reality, use of ambient light and consistency in imaging accounting for ambient light may result in better blending or greater visual appeal.
- the machine-learnt model is trained to output any number of values for a respective number of rendering parameters in act 40 .
- values for two or more physically-based rendering parameters are output based on the input medical dataset.
- Three or more, all, or a sub-set of any number of rendering parameters may have their values output by application to the machine-trained model or group of models.
- Example rendering parameters include controls for consistent data handling, transfer function, lighting design, viewing design, material propriety, or internal renderer parameters.
- One or more of the output values are for non-transfer function parameters.
- the machine-learnt model or set of models output values for all of the rendering parameters except for viewing design.
- One or more parameters for viewing design are left to the user, such as for user navigation or setting of the camera position, orientation, and/or scale.
- the user inputs the values for all or some of the viewing design parameters, and the application of the medical dataset to the machine-learnt model provides the values for all of the other rendering parameters. Additional, different, or fewer types of rendering parameters may use initial input by the user.
- the user may adjust model output values. Since physically-based rendering may be more time consuming than traditional rendering, providing a comprehensive set of values for the rendering parameters more likely to result in the desired image initially may avoid time-consuming adjustments. The user may, however, make any number of adjustments. The number and/or amount of change may be less due to the assistance by the machine-learnt model.
- a physically-based renderer renders the photorealistic image of the 3D region of the patient using the values output from the application.
- the output values of the machine-learnt model are used as settings by the physically-based renderer. For example, a sub-set or all of the rendering parameters for rendering with unbiased path tracing have values established by the output of the machine-learnt model.
- the physically-based renderer is a graphics processing unit, array of graphics processing units, workstation, server, computer, or other machine capable of the required computations. Any physically-based rendering algorithm capable of computing the light transport may be used, such as path tracing, unbiased path tracing, Monte-Carlo path tracing, or Metropolis light transport.
- the physically-based rendering simulates the physics of light propagation to create an image instead of accumulating volume samples with local illumination along viewing rays from the camera, as is done for traditional volume rendering.
- path tracing implemented for physically-based rendering, the paths of light or photons, including due to scattering and absorption, are modeled and used for rendering.
- the physically-based rendering result may be built up over time as the rendering may rely on probabilistic scattering and tracing millions of light paths.
- the rendering results in a photorealistic image A sequence of images may be provided as the image is built or rendered. Alternatively, for a given set of values of rendering parameters, a single image is output. In either case, the photorealistic image is rendered using presets from, at least in part, the machine-learnt model. A sequence of images due to alteration of values for one or more rendering parameters may be output. The user or renderer may change the values, such as rotation of a camera or for modeling fly-through.
- the renderer, host computer of the renderer, or other device transmits the photorealistic image or images.
- the transmission is over a bus, through a network interface, over a network, to a memory, or to a display.
- the image is loaded into a buffer and output from the buffer to a display for viewing by a physician to aid diagnosis or pre-operative planning.
- a computer-generated representation of the medical data suitable for viewing by human users e.g., 2D images, videos, stereoscopic image pairs, volumes, holograms, or lightfields
- the output may be for storage and/or for further machine analysis.
- the transmission may be for any one or more of various purposes.
- the transmission of the photorealistic image is part of a diagnostic report.
- a given diagnostic report may be specific to context, such as a report for kidney stones. Any kidney stones should be visible in an image.
- the photorealistic image shows kidney stones based on the values of the rendering parameters output by the model trained to create such images.
- the optimal context-specific images are provided for an automated diagnostic report in the clinical workflow.
- the transmission is to a display as an initial image for interactive viewing.
- the machine-learnt model is used to initialize interactive viewing applications in the clinical workflow. By providing values for rendering parameters so that the images are consistent with an established visualization design, the amount of user interaction may be reduced.
- the transmission is to an augmented reality device.
- a physician may wear augmented reality glasses.
- the photorealistic image with or without added annotations or graphics, is projected onto the glasses as an overlay on the actual view seen by the wearer.
- the photorealistic image is combined with a camera image and/or is used as a replacement for the camera image.
- the rendered image may be overlaid onto a live stream of real-time medical images (ultrasound, x-ray, optical laparoscopy, etc.).
- FIG. 3 shows a block diagram of one embodiment of a system for content-based rendering based on machine learning.
- the system is for training with machine learning and/or application of a machine-learnt model.
- a machine-learnt model is trained to output settings for rendering parameters given a new volume for a patient.
- the system implements the method of FIGS. 1 and/or 2 .
- Other methods or acts may be implemented, such as providing a user input (e.g., mouse, trackball, touch pad, and/or keyboard) and user interface for interactive rendering.
- a user input e.g., mouse, trackball, touch pad, and/or keyboard
- user interface for interactive rendering.
- the system includes a machine 50 , a memory 52 , a display 54 , a medical imager 56 , and a renderer 58 . Additional, different, or fewer components may be provided. For example, the medical imager 56 and/or memory 52 are not provided. In another example, a network or network connection is provided, such as for networking with a medical imaging network or data archival system. A user interface may be provided for interacting with the machine 50 , renderer 58 , or other components.
- the machine 50 , memory 52 , renderer 58 , and/or display 54 are part of the medical imager 56 .
- the machine 50 , memory 52 , renderer 58 , and/or display 54 are part of a server, workstation, or computer separate from the medical imager 56 .
- the machine 50 , memory 52 , renderer 58 , and/or display 54 are a personal computer, such as desktop or laptop, a workstation, a server, a network, or combinations thereof.
- the machine 50 and memory 52 are part of a separate computer from the renderer 58 .
- the medical imager 56 is a medical diagnostic imaging system. Ultrasound, CT, x-ray, fluoroscopy, positron emission tomography (PET), single photon emission computed tomography (SPECT), and/or MR systems may be used.
- the medical imager 56 may include a transmitter and includes a detector for scanning or receiving data representative of the interior of the patient.
- the medical imager 56 acquires scan data representing the patient.
- the scan data may represent a volume of the patient. For example, scan data is acquired and used for diagnosis or surgical planning, such as identifying a lesion or treatment location.
- the medical imager 56 is not provided, but a previously acquired dataset for a patient is stored in the memory 52 .
- many medical images 56 are provided in the memory 52 as the training data, which is gathered and stored in the memory 52 .
- the machine 50 is a computer, workstation, server, processor, or other device configured to apply machine learning and/or to apply a machine-learnt model.
- the machine 50 is configured by software, hardware, and/or firmware.
- the machine 50 is configured by one or more machine learning algorithms.
- the machine 50 is configured, in part, by a learnt matrix or matrices associating input data to output values.
- the machine 50 applies a machine-learnt model, such as one learnt with deep learning.
- the machine-learnt model as implemented by the machine 50 , generates input features and/or outputs settings for rendering parameters based on application of medical data from a scan of a patient.
- the rendering parameters include windowing, transfer function, lighting, and/or material properties.
- the settings are learnt to provide an image similar in look and/or feel to one or more images from a same diagnostic context. Consistent images, such as to highlight or show anatomy of interest, are created.
- the diagnostic context provides the standard or information of interest.
- the memory 52 is a graphics processing memory, a video random access memory, a random access memory, system memory, cache memory, hard drive, optical media, magnetic media, flash drive, buffer, database, combinations thereof, or other now known or later developed memory device for storing training data, rendering parameters values, medical datasets, and/or other information.
- the memory 52 is part of the medical imager 56 , part of a computer associated with the machine 50 , part of a computer associated with the renderer 58 , a database, part of another system, a picture archival memory, or a standalone device.
- the memory 52 stores scan data representing one or more patients. For example, data from the medical imager 56 is stored. The data is in a scan format or reconstructed to a volume or three-dimensional grid format. The scan data is stored for training and/or stored for rendering an image of a patient.
- the memory 52 or other memory is alternatively or additionally a non-transitory computer readable storage medium storing data representing instructions executable by the programmed machine 50 for learning or applying the machine-learnt model.
- the instructions for implementing the processes, methods, and/or techniques discussed herein are provided on non-transitory computer-readable storage media or memories, such as a cache, buffer, RAM, removable media, hard drive, or other computer readable storage media.
- Non-transitory computer readable storage media include various types of volatile and nonvolatile storage media.
- the functions, acts or tasks illustrated in the figures or described herein are executed in response to one or more sets of instructions stored in or on computer readable storage media.
- processing strategies may include multiprocessing, multitasking, parallel processing, and the like.
- the instructions are stored on a removable media device for reading by local or remote systems.
- the instructions are stored in a remote location for transfer through a computer network or over telephone lines.
- the instructions are stored within a given computer, CPU, GPU, or system.
- the renderer 58 is a graphics processing unit, graphics card, graphic chip, multi-core processor, or other processor for running a software package that implements the photorealistic image generation functionality.
- the renderer 58 is configured by an application programming interface to render an image from the 3D scan data representing a patient. Using physically-based rendering, a photorealistic image is rendered. Using the settings output by the machine-learnt model, path tracing is applied to render the image. For feedback in training, the renderer 58 may render training images. Similarity is calculated by the renderer 58 and/or the machine 50 .
- the display 54 is a monitor, LCD, projector, plasma display, CRT, printer, or other now known or later developed device for displaying the photorealistic image or images.
- the display 54 receives images from the machine 50 , memory 52 , renderer 58 , or medical imager 56 .
- the images of the tissue captured by the medical imager 56 are displayed.
- Other information may be displayed as well, such as generated graphics, text, or quantities as a virtual overlay.
- Additional images may be displayed.
- scan data represents a sequence of scans over time
- a corresponding sequence of images may be generated.
- new images may be generated as settings for one or more rendering parameters are changed by a user.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Graphics (AREA)
- General Engineering & Computer Science (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- Databases & Information Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- Mathematical Physics (AREA)
- Multimedia (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Biophysics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Pathology (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Medical Treatment And Welfare Office Work (AREA)
Abstract
Description
- The present patent document claims the benefit of the filing date under 35 U.S.C. §119(e) of Provisional U.S. Patent Application Ser. No. 62/306,407, filed Mar. 10, 2016, which is hereby incorporated by reference.
- The present embodiments relate to medical imaging of three-dimensional (3D) scans. Data representing a volume is rendered for visualization. Due to the many different scan settings and patient variability, renderings for different patients or at different times appear different.
- In an effort to reduce variability, most existing medical rendering provides a set of static visualization presets for specific workflows or diagnostic contexts. Even with static presets, the rendering may require manual adjustment and may not provide consistent rendered results. Semi-automatic machine learning approaches have been used to provide transfer function design in traditional volume rendering (e.g., projection or alpha blending), but may still not sufficiently reduce variability.
- Variability is particularly troublesome for physically-based volume rendering, which relies on the physical simulation of light propagation (e.g., unbiased path tracing). Physically-based visualization techniques produce global illumination effects in computer-generated graphics that mimic the real-world interaction of light with various 3D objects or tissues. This results in physically plausible images that are often easier for the human brain to interpret when compared to the more analytical images from traditional rendering. For example, subtle ambient occlusions, shadows, and color bleeding provide important depth cues for the spatial understanding of 3D relationships between structures in a single 2D image, whereas simpler visualization techniques may require additional interaction with the viewing parameters (e.g., moving the virtual camera around the 3D data) to obtain the same spatial information from the image parallax.
- These physically-based visualization techniques are more sensitive to changes in the classification or the interpretation of the underlying medical data. As a result, small changes to the rendering parameters may have a more pronounced impact on the perception and interpretation of 3D structures in the final image as compared to the traditional volume rendering techniques based on compositing of classified voxels along viewing rays. Obtaining very high quality reproducible images with diagnostic significance is then partially at the user's discretion. While existing techniques for providing visualization presets can help, the workflow is not fully automated and the resulting rendered images may not have consistent quantitative properties (color, hue, reflectance, etc.) across datasets.
- By way of introduction, the preferred embodiments described below include methods, systems, instructions, and computer readable media for machine training an artificial intelligence and use of that artificial intelligence to provide rendering settings. By using deep learning and/or other machine training, settings of multiple rendering parameters may be provided for consistent imaging or visualization even in physically-based rendering.
- In a first aspect, a method is provided for content-based rendering based on machine learning in a rendering system. A medical dataset representing a three-dimensional region of a patient is loaded from memory. A machine applies the medical dataset to a machine-learnt non-linear model. The machine-learned model is trained with deep learning to extract features from the medical dataset and trained to output values for two or more physically-based rendering parameters based on input of the medical dataset. At least one of the two or more physically-based rendering parameters are controls for consistent data handling, lighting design, viewing design, material propriety, or internal renderer property. A physically-based renderer renders a photorealistic image of the three-dimensional region of the patient using the output values resulting from the applying. The photorealistic image is transmitted.
- In a second aspect, a method is provided for machine training for content-based rendering in a machine training system. First volume data of a volume of a patient, a first image of the volume, and first values of path tracing rendering parameters are input to training of an artificial intelligence. A machine trains the artificial intelligence to output second values of the path tracing rendering parameters for a second volume data where the second values provide a second rendered image of the second volume modeled on the first image. The trained artificial intelligence is stored.
- In a third aspect, a system is provided for content-based rendering based on machine learning. A medical scanner is configured to scan a patient. A machine is configured to output settings for rendering parameters by application of data from the scan to a machine-learnt model. The rendering parameters include windowing, transfer function, and lighting, and the settings are learned to provide a first image from the data similar to one or more second images for a same diagnostic context. A graphics processing unit is configured to render the first image using the settings.
- The present invention is defined by the following claims, and nothing in this section should be taken as a limitation on those claims. Further aspects and advantages of the invention are discussed below in conjunction with the preferred embodiments and may be later claimed independently or in combination.
- The components and the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the invention. Moreover, in the figures, like reference numerals designate corresponding parts throughout the different views.
-
FIG. 1 is a flow chart diagram of one embodiment of a method for machine training for content-based rendering; -
FIG. 2 is a flow chart diagram of an embodiment of a method for content-based rendering based on machine learning in a rendering system; and -
FIG. 3 is a block diagram of one embodiment of a system for content-based rendering based on machine learning. - Content-based photorealistic rendering of medical images is based on machine learning. To produce consistent rendered images from 3D medical scans, particularly when using physically-based rendering techniques, the machine learnt non-linear model determines rendering settings based on the content of the 3D medical scan. Due to the variability between scan parameters, data contrast, noise, and/or the large number of rendering parameters, producing rendered images that allow for consistent visual interpretation has been a long standing challenge, which has prevented wider adoption 3D images for definitive decision support.
- In one embodiment, deep learning generates the settings for the rendering parameters for the photorealistic rendering components, establishing a consistent visualization design. This design may be applied to new 3D image data in order to provide reliable visual interpretation within the diagnostic workflow. Image features, and optionally, available non-image data are mapped to sets of rendering parameters that produce optimal images, where the relationship between the two is highly non-linear in the general case. By targeting the entire or multiple parts of the rendering pipeline, a scalable solution leverages machine learning and supports consistent photorealistic rendering.
- Consistent visual interpretation for 3D images may potentially simplify diagnostic workflow and may promote the clinical use of 3D data for decision support. Consistent photorealistic rendered images may be provided in new platforms and workflows, such as fully automated reporting and cloud-based rendering services. Consistent visualization design as a machine learning problem may be applied 1) to optimize rendering preset designs tailored to a large variety of data types in a unified and scalable way, 2) to adapt rendering parameters to produce target consistent results based on the provided input, 3) to adapt rendering parameters to specific individual preferences, 4) to transfer expert trained results to non-expert users, 5) to increase the conspicuity of visual features in the rendered images (e.g., pathologies in medical images) and 6) to provide automatic rendering presets for specific workflows. A scalable framework is provided for fully automated image generation in the clinical workflow.
-
FIG. 1 shows one embodiment of a method for machine training for content-based rendering in a machine training system. Optimal rendering parameters may be obtained via training an artificial intelligence system. The input of the artificial intelligence system may include, but is not limited to, the 3D scan data to be rendered, images rendered from the 3D scan data, information on the patient, and/or information on the diagnostic purpose. Using many examples, the machine training learns to provide rendering settings for photorealistic rendering based on input data for a specific patient. The output of the trained artificial intelligence system is the optimal set of settings for the rendering parameters. - The method of
FIG. 1 is implemented by a processor and a graphics processing unit in a server, computer, or other machine. In one embodiment, the system ofFIG. 3 implements the method ofFIG. 1 , with or without themedical imager 56 and/or display 54. For example, data is input by loading from a memory, a processor performs the training based on the data, and the memory stores the learnt artificial intelligence. As another example, the medical scanner and/or memory input the data to a server, which performs the training and outputs to a memory. - The acts are performed in the order shown (top to bottom) or another order. For some forms of learning, results may be fed back as input in a repetition of the acts.
- Additional, different, or fewer acts may be provided. For example, act 22 is not provided, such as where the trained artificial intelligence is used rather than stored for later use.
- In
act 12, training data is input. The training data is gathered from one or more memories, sensors, and/or scanners. The gathered data may be stored and/or transmitted to a buffer, memory, cache, processor, or other device for training. This gathered training data is input to the machine learning algorithm. - Many samples of the same type of data are input. To learn statistical or other relationships that may be sufficiently complex that a human cannot perceive the relationships in a same level, tens, hundreds, or thousands of samples are provided.
- Any data may be used. Since the artificial intelligence is to be trained to provide settings for rendering parameters, scan data is provided. For each sample, the scan data represents a patient. A two-dimensional (2D), 3D, 2D+time sequence, 3D+time sequence, and/or other image or scan data may be used.
- Other information than scan data may be provided for each sample. Patient non-image data may be input to the training. Higher dimensional data includes but is not limited to patient information such as a) natural language medical documents or information extracted therefrom of previous diagnosis and/or clinical procedures; b) previous medical test results; c) previously processed medical images and measurements; d) computational modeling results; e) parametric image information derived from medical image analysis algorithms (e.g. cardiac strain map or elasticity); and/or f) other information (e.g., scanner settings).
- For training, the goal or ground truth information is input. The artificial intelligence system is trained with labeled data (e.g., pairs of input and corresponding desired output (e.g., rendering parameters, optimal rendered images, and/or photographs/movies of the real object)). Based on the provided data, the artificial intelligence system is trained to produce the desired output from any input. The ground truth is provided by desired settings of the rendering parameters for each scan dataset. The values of the path tracing rendering parameters for desired results of each sample are input to the training. Alternatively or additionally, the ground truth is provided as a desired rendered image for each scan dataset. A rendered image for each scan dataset is provided. In other embodiments, an image with desired visualization is provided for multiple datasets. Using a renderer, the machine training may learn the settings that produce similar visualization.
- The input samples may provide both negative and positive training samples. For example, scan datasets and settings providing desired and undesired images are input. The training learns to distinguish the desired settings from both good and bad settings.
- In one embodiment, additional samples are created by perturbing the path tracing rendering parameters of an input sample. The perturbing creates a collection of sets of the path tracing rendering parameters for each scan data set. The artificial intelligence system is applied on the input scan dataset to generate rendering parameters, which are then perturbed to generate a pool of rendering parameters. Rendering is then performed using the sets of rendering parameters in the pool to produce a pool of corresponding images. Human users or other operators then visually examine the produced images and select the satisfying images. The rendering parameters for the selected images are then used as desired output and/or a sub-set as positive ground-truth. The rendering parameters for the unselected sub-set of images are not used or are used as negative ground-truth.
- In another approach to gather samples for input, a web-based or network-based service is provided. A website or network portal allows a user to upload the input data (i.e., image and additional information). The server or other processor then runs the artificial intelligence system and/or perturbs possible rendering settings to renderer a pool of images. The service then provides the generated images to the user and asks the user to select the satisfying images. The rendering parameters corresponding to the user selected images are then stored as the desired output for the input (i.e., ground truth). By providing such a service, additional samples for training may be gathered over time.
- The input may be repeated. The training of the artificial intelligence system is performed periodically or multiple different times. After additional samples are collected (e.g., through the web-based service), the training is repeated using the additional samples with or without previous samples used for earlier iterations of the training.
- In
act 14, the artificial intelligence or intelligences are machine trained. The collected samples, including ground truth, are used to train. A machine, such as a processor, computer, server, or other device learns from the samples. Using machine-learning, complex statistical relationships between large numbers (e.g., tens, hundreds, thousands, or more) of input variables to any number of output variables are extracted from the large number of samples. - The training is for imaging in any context. One model may be learned for any number of imaging situations. Alternatively, different models are trained for different situations. The different situations may include different scan modalities (e.g., different model for computed tomography, magnetic resonance, ultrasound, positron emission tomography, and single photon emission computed tomography). The different situations may include different types of tissue of interest (e.g., liver versus kidney), different diagnostic purpose or workflow (e.g., cancerous lesion versus bone calcification), and/or different users (e.g., different operators may have different preferences for visualization).
- Any now known or later developed machine learning may be used. Regression, classification, and/or reinforcement learning are used. Regression training learns a range or continuous output by minimization of a metric. Classification learns disparate outputs. Reinforcement learning learns through a sequence of actions with feedback. Neural network, Bayes network, or support vector machine training may be used. Hierarchal or other approaches may be used. Supervised or semi-supervised machine learning may be used.
- To train, features are extracted from the input data. Haar wavelet, steerable, or other features may be extracted from the scan dataset or images. Measures or other information may be extracted from non-image data. Alternatively, the input data itself (e.g., scalar values of the scan dataset) is used.
- In an alternative represented in
act 16, the machine learns features as part of the training. For example, deep learning (e.g., deep structured learning, hierarchical learning, or deep machine learning) models high-level abstractions in data by using multiple processing layers with structures composed of multiple non-linear transformations, where the input data features are not engineered explicitly. A deep neural network processes the input via multiple layers of feature extraction to produce features used to derive outputs of settings for rendering. The deep learning provides the features used by other machine training to learn to output the rendering settings. Other deep learnt, sparse auto-encoding models may be trained and applied. The machine training is unsupervised in learning the features to use and how to classify given an input sample (i.e., feature vector). - Using deep learning, the input features required for the machine learning are not engineered by the user explicitly, as in shallow learning algorithms. This is especially useful with render settings, which may not have an explicit natural relationship with the input data. Deep learning figures out the data features that end up providing a good classifier or model. Deep learning is known to be effective in extracting high level abstractions (features) from low level input data (e.g., raw images) that is suitable for the trained task, so may be more convenient than and superior to hand-crafted features for identifying features for render settings. In addition, learning-based automatically rendering setting selection is a less studied problem, with no hand-craft feature specifically designed for this problem. Standard hand-craft features, like HAAR feature or SIFT feature, designed for other general computer vision tasks may not be suitable for rendering setting.
- In
act 18, the machine trains to output rendering settings to provide consistent imaging. To reduce variability, the training learns to provide rendering settings that result in a rendered image with similar or same visualization despite differences in the content of the data. The path tracing or other physically-based rendering settings so that the rendered image for a given input scan dataset looks and feels similar to an ideal are learned. The machine learns to output renderer settings to model the resulting rendered image for a given scan dataset on a desired image. The learned sets of parameters establish a visualization design that may be reused when new data is introduced to the system in order to provide consistent rendered images. In this way, the machine may learn renderer settings to provide rendering to a standard. Rather than render with the same preset values for a given context, the values are based on a given input dataset. In various embodiments of the system, consistency refers to consistent data handling, consistent viewing, and/or consistent visual styles, which may be utilized both in automated and interactive clinical workflows. - The ground truth provides consistency. Where the ground truth uses similar or consistent images and/or settings that result in similar or consistent images despite variance in the scan datasets, the training learns to output settings providing the consistency. The notion of optimal rendering parameters and optimal output images relates to using objective quality measurements, subjective quality measurements, clinical significance ratings, and/or other indicators of consistency.
- In one embodiment, the consistency is learned using a measured or calculated metric. The training uses a metric of similarity. Any metric of similarity may be used. Example computed metrics may be visual perception metrics based on visual systems, comparative differences, perceptual hash, mutual information, correlation, computed feature signals, or simple error metrics. The metrics may be automatically computed or provided by the users. The metric is used as an indicator of strength of importance of a given training sample.
- Alternatively or additionally, one or more metrics are used in reinforcement learning. For example, deep reinforcement learning uses a similarity of the rendered image from a sample to a reference image. The similarity is used as a reinforcement. The output image rendered from a given set of rendering settings for a sample of scan data and the associated quality metrics (i.e., similarity) are fed-back into the learning system for additional refinement of the learned parameters (e.g., in deep supervised and/or reinforcement learning). A metric measuring the similarity of one rendered image to one or a group of rendered images with the desired quality is calculated, and used as the “reward” to train the artificial intelligent agent using the deep reinforcement learning technique.
- Other quality metrics than similarity may be used. The choice of metric to use may depend on the types of features relevant to a specific clinical data domain. Depending on the diagnosis of interest or medical workflow, different pathologies may be of interest. Different rendering settings increase the conspicuity of different pathologies. The quality metric in the reinforcement learning is used to learn to provide the optimal rendering parameters for increasing the conspicuity of the pathology of interest.
- In one embodiment, the artificial intelligent agent is a deep neural network that takes in the output of the reinforcement agent as its input, and outputs one quality or similarity metric (referred to as “action-value”) for each possible action (e.g., set of rendering settings), indicating the effect of taking these actions (i.e., higher values are associated with actions that can improve the rendering quality). The agent also analyzes the action-values, and selects one action to perform. The set of rendering parameters providing the highest implementable quality rendering are selected as the output.
- In another embodiment, a probability distribution of different values of the metric (i.e., different similarities) is used for the selection. The selected action is the action with highest action-value or randomly selected from the actions with a probability distribution assigned based on the rank of action-values. In a typical scenario, actions with higher action-values are assigned higher probability, but any assignment may be used.
- In other or additional reinforcement learning, the training is modeled after steps taken by an expert or other user to achieve the desired visualization for the training data. The user's sequence of adjustments to provide the desired rendering from the scan dataset is monitored (e.g., select transfer function T1, then window function W1, then select a different transfer function T2, then select a material reflectance MR1, . . . ). The learning uses metrics for each stage or step to learn the settings used for each. The actions of the expert are extrapolated to different training datasets or included for each training dataset. The training learns to provide final rendering settings based on the monitored sequence.
- Rather than using a rendered image or multiple rendered images as the ground truth and/or for calculating similarity, a photograph or video of the imaged tissue, imaged object, or other real-world references is used. The physically-based rendering provides photorealistic images. By comparing or training to real photos or video, the training learns to provide rendering settings modeled after real images. The desired output is labeled using actual optical images so that the rendering parameters result in a matching look and/or feel. Other modes of images may be used for the ground truth or reinforcement metric, so that the training learns to output rendering settings modeled after the different modality (e.g., render CT data to match the look of rendered MR data).
- In
act 20, one model is trained to output settings for multiple rendering parameters. In other embodiments, a hierarchy or other grouping of models are trained to output the settings for multiple rendering parameters. Using the training data and ground truth, the machine is trained to output settings of rendering parameters. By outputting settings for two or more rendering parameters for a given scan dataset, less manual setting of parameters is needed. In path tracing or other physically-based rendering, the amount of processing and time required to provide the photorealistic imaging may cause delays. Compounding such delays with user adjustment, it may be overly time consuming to provide a desired rendering manually. By using machine learning to provide the initial settings for multiple parameters, the machine trained model contributes to less delay in achieving the desired rendered image. - The model may be trained to output any number and/or type of rendering parameters. For example, one or more data consistency parameters are to be output. Data consistency parameters include windowing, scaling, level compression, data normalization, or others. As another example, one or more transfer function design parameters are to be output. Transfer function design parameters include classification look-up tables, multi-dimensional transfer functions, tissue-specific transfer functions, or other transfer functions. In another example, one or more lighting design parameters are to be output. Lighting design parameters include type of virtual lights, position of the virtual light sources, orientation of the virtual light sources, image-based lighting sources, or others. In yet another example, one or more viewing design parameters are to be output. Viewing design parameters include type of camera, position of the camera, orientation of the camera, intrinsic parameters for viewing, or others. In other examples, one or more use-case specific parameters are to be output. Use-case specific parameters are settings specific to a given use, such as a particular camera position for a given type of medical report or use of two cameras for stereoscopic viewing.
- The parameters may be specific to physically-based rendering, such as internal render properties. The setting for any parameter for path tracing, unbiased path tracing, Monte Carlo rendering, global illumination, or other simulations of light propagation may be learned. For example, a global albedo, or surface reflectance coefficient, may be learned for the volume data, which describes the relative importance of light scattering versus light absorption. Settings for any parameters for material property controlling the interaction with light may be learned. For example, the transparency, reflectivity, surface roughness, and/or other properties may be modeled in the physically-based rendering, so included as rendering parameters. Albedo may also be learned as a per-material property for the different tissue classes in the volume data. Since the effect of various settings of material type or light interaction may not be instinctive to a user, training the model with machine learning to provide the initial settings may assist the user.
- All or a sub-set of the rendering parameters are to be output by the machine-learnt model. For example, the model is trained to output any number of data consistency, transfer function, lighting, and viewing parameters. For fully automatic, the model is trained to provide settings for all of the rendering parameters. For interactive or other semi-automatic operation, the model is trained to provide settings for some but not all of the rendering parameters. For example, all of the rendering parameters are output except for the view design parameters, which are under the control of the user. The viewing camera parameters are under the control of the user, so the machine learning is for settings of other parameters. The user controls the view design parameters independently.
- The model is trained to output settings for multiple parameters regardless of the user-controlled parameters. Alternatively, the user-selected parameters are an input to the classification so that the output settings account for the user-selected settings. The model learns to output rendering parameter settings based, in part, on settings for one or more parameters under the control of the user. The artificial intelligence is trained to output the values for the rendering parameters based on user selected viewing camera parameters. Subsets of the rendering parameters are used as additional inputs in the training phase of the machine learning algorithm. In that case, the machine learning algorithm output is an optimized subset of the rendering parameters. The user input settings may be optimized or not while the other settings are optimized by the model. In an interactive viewing application, the input to the training is a higher dimension vector including the input scan dataset and reference viewing parameters.
- In another embodiment, the artificial intelligence is trained to account for different ambient conditions in the real-world viewing environment of the user. The settings for “ambient-varying” rendering of a volume are learned. The training data includes different ambient light conditions and measures of the ambient light using light probes. The model is trained to account for ambient luminosity of the room. The luminosity is sent to the trained model, which adapts the rendering parameters automatically to provide the same content-based realistic rendering.
- The training may be repeated. As the machine-trained model is used, the user may adjust some of the output settings. This adjustment, the setting, and/or image resulting from the adjustment are added to the training data. The machine learning model is refined by repeating the learning with the addition feedback as samples from which to learn. The manually adjusted settings of the parameters define personalized visualization designs. Other sources of further training data may be used.
- The repetition of the training may be performed separately for different users, physicians, medical groups, hospitals, or other grouping. Personalized training is used so that consistent images for a relevant group are provided. In other embodiments, the repetition is provided for all users.
- In
act 22, the trained artificial intelligence is stored. The result of the training is a matrix. The matrix represents the learned knowledge through machine training. The matrix includes an input vector and outputs for the settings of the rendering parameters. Other machine-learnt model representations may be used, such as a hierarchy of matrices or other non-linear models. - The trained model is stored in a memory. Any memory may be used. The memory used for the training data may be used. For application, the memory may be in other devices. For example, the trained model is stored in a memory of a server. The server uses the trained model to output rendering parameters to clients or to a server-based renderer which outputs rendered images to the clients. As another example, multiple copies of the trained model are provided to different physicians, medical scanners, and/or workstations for use by different physicians.
-
FIG. 2 shows one embodiment of a method for content-based rendering based on machine learning in a rendering system. The machine-learnt model or artificial intelligence agent is applied. For a given patient, a rendering of a scan dataset is to be performed. The artificial intelligence agent is applied to the scan dataset with or without other input information, outputting values for the rendering parameters. The values result in consistency in photorealistic rendering, possibly avoiding time consuming manual adjustment of various rendering parameters to achieve the desired visual. - The method is implemented by a computer, server, or other processor with a graphics processing unit or other renderer and a display. For example, medical data is loaded from memory. A light sensor measures the ambient light or captures a light probe image, and a computer applies the machine-learnt model, outputting values for rendering. A renderer uses the values to render from the medical data, resulting in transmission of the rendered image to a display. Different devices may be used, such as a controller or medical scanner processor performing the application and/or rendering.
- Additional, different, or fewer acts may be provided. For example, the ambient light is not measured in
act 32. As another example, the rendering ofact 42 and transmission of the photorealistic image ofact 44 are not provided, but instead the values of the rendering parameters are stored or transmitted. In yet another example, acts for manual adjustment and/or initial setting of one or more rendering parameters are provided, such as for an interactive rendering workflow. In other examples, none, one or two of acts 36-40 are performed without others of that set of acts. - The acts are performed in the order shown (top to bottom) or other orders. For example, the ambient light may be measured in
act 32 prior to loading the medical data inact 30 and/or in parallel with performingact 34. - In
act 30, medical data is loaded from a memory, sensors, and/or other source. A medical scanner may provide the data, such as a medical dataset representing a 3D region of the patient. Mining of a patient's computerized medical record may be used. Medical information from a picture archiving and communications server may be obtained. Any source may be used. - The medical data is for a given patient. While the training may use samples from many patients to learn features and/or learn to classify input medical data to provide values for rendering, the learnt model is applied to the medical data for a patient to output values for rendering parameters for that patient. The same or different machine-learnt models are applied to data for any number of patients.
- Any type of medical data is obtained. Scan data representing a 3D volume is loaded as a medical dataset. The scan data may be from multiple two-dimensional scans or may be formatted from a 3D scan. The medical data may include non-image or non-scan data, such as patient test results, physician notes, lab results, diagnosis information, billing codes, clinical data, patient history, or other information.
- The medical data includes the information learned to be used by the machine-trained model. Ground truth information and/or similarity or quality metric measures may be not provided. Where the machine learning learned to use feedback of predicted ground truth, then that feedback may be created and used as an input. The content to be used to determine the settings of the rendering parameters is provided to the machine-learnt model. For example, just the 3D scan data is provided as a medical dataset for a given patient. As another example, the 3D scan data and diagnostic information are provided as the medical dataset for the patient. The diagnostic information may be used to select the machine-learnt model to use and/or be part of the input feature vector.
- Some machine-learnt models may be robust or still operate with some of the input feature vector missing. Alternatively, all of the information needed for the input feature vector is loaded.
- In
act 32, a sensor measures ambient light. Any luminosity sensor may be used, such as a charge coupled device. The sensor provides a measure of the luminosity. This measure may be used as an input in the input feature vector and/or to select a particular machine-learnt model to use. In alternative embodiments, the rendering is not based on ambient lighting conditions, so no measurement of ambient light is provided. - In an alternative embodiment for
act 32, a light probe image is obtained as the measure of ambient light with a camera as the sensor. The image is obtained using any of a variety of ways and then used for high dynamic range image-based illumination during rendering. Examples ways for obtaining light probe images include capturing high-dynamic range photographs of a mirrored ball, stitching multiple photographs, or using self-contained panoramic cameras. - In
act 34, the medical dataset is applied to the machine-learnt model. The loaded medical dataset is processed by a machine, such as a computer, processor, or server. The machine uses the internal representation of the machine-learnt model. The input feature vector is created and/or is the medical dataset. The machine inputs the feature vector, resulting in the machine-learnt model outputting the rendering settings based on the learned knowledge. - Acts 36-40 represent different acts in applying the machine-learnt model of presets for one or more rendering parameters. Additional, different, or fewer acts may be provided.
- The machine-learned model was previously learned using any approach for machine learning, so may be a regression, classification, or reinforcement-learnt model. In one embodiment, the machine-learned model is trained with deep learning (e.g., deep neural network) to extract features from the medical dataset. By applying the deep-learnt model in
act 36, features are extracted from the medical dataset. One or more other features may be extracted from the medical dataset using manually or user configured features. These features may then be used by another machine-learnt model and/or another hierarchal stage to output the values for the rendering. For example, a further reinforcement-learnt model is applied. The extracted features are input to the machine-learnt model for application. - Where the training was for consistency, the application of the machine-learnt model is more likely to output values in
act 38 resulting in the photorealistic image corresponding to a standard image despite differences in the medical dataset. The photorealistic image is created using path tracing, other physically-based rendering, or reference photographs. The model uses the input feature vector to predict the values for rendering that result in the rendering of the scan data as an image with desired characteristics. The standard image may be a group of images having similar characteristics, such as resolution and/or contrast for a particular type of tissue or lesion. The rendering values output were learnt in order to provide the same or similar resolution and/or contrast for that particular type of tissue. Different machine-learnt models may be used for different diagnostic and/or workflow situations, such as outputting values for breast cancer imaging or for fetal imaging. - In another embodiment, the ambient light measured in
act 32 is applied to the machine-learnt model with or as part of the medical dataset. The machine-learnt model is trained to output the values based in part on the ambient light. By applying the ambient light measure, the machine-learnt model outputs values for the rendering parameters that account for ambient light. Photorealistic images may be output as if lit by the ambient light in a room or to which the patient is subjected. Where the photorealistic images are used for augmented reality, use of ambient light and consistency in imaging accounting for ambient light may result in better blending or greater visual appeal. - The machine-learnt model is trained to output any number of values for a respective number of rendering parameters in
act 40. For example, values for two or more physically-based rendering parameters are output based on the input medical dataset. Three or more, all, or a sub-set of any number of rendering parameters may have their values output by application to the machine-trained model or group of models. Example rendering parameters include controls for consistent data handling, transfer function, lighting design, viewing design, material propriety, or internal renderer parameters. One or more of the output values are for non-transfer function parameters. - In an example embodiment, the machine-learnt model or set of models output values for all of the rendering parameters except for viewing design. One or more parameters for viewing design are left to the user, such as for user navigation or setting of the camera position, orientation, and/or scale. The user inputs the values for all or some of the viewing design parameters, and the application of the medical dataset to the machine-learnt model provides the values for all of the other rendering parameters. Additional, different, or fewer types of rendering parameters may use initial input by the user.
- The user may adjust model output values. Since physically-based rendering may be more time consuming than traditional rendering, providing a comprehensive set of values for the rendering parameters more likely to result in the desired image initially may avoid time-consuming adjustments. The user may, however, make any number of adjustments. The number and/or amount of change may be less due to the assistance by the machine-learnt model.
- In
act 42, a physically-based renderer renders the photorealistic image of the 3D region of the patient using the values output from the application. The output values of the machine-learnt model are used as settings by the physically-based renderer. For example, a sub-set or all of the rendering parameters for rendering with unbiased path tracing have values established by the output of the machine-learnt model. - The physically-based renderer is a graphics processing unit, array of graphics processing units, workstation, server, computer, or other machine capable of the required computations. Any physically-based rendering algorithm capable of computing the light transport may be used, such as path tracing, unbiased path tracing, Monte-Carlo path tracing, or Metropolis light transport. The physically-based rendering simulates the physics of light propagation to create an image instead of accumulating volume samples with local illumination along viewing rays from the camera, as is done for traditional volume rendering. In path tracing implemented for physically-based rendering, the paths of light or photons, including due to scattering and absorption, are modeled and used for rendering. The physically-based rendering result may be built up over time as the rendering may rely on probabilistic scattering and tracing millions of light paths.
- The rendering results in a photorealistic image. A sequence of images may be provided as the image is built or rendered. Alternatively, for a given set of values of rendering parameters, a single image is output. In either case, the photorealistic image is rendered using presets from, at least in part, the machine-learnt model. A sequence of images due to alteration of values for one or more rendering parameters may be output. The user or renderer may change the values, such as rotation of a camera or for modeling fly-through.
- In
act 44, the renderer, host computer of the renderer, or other device transmits the photorealistic image or images. The transmission is over a bus, through a network interface, over a network, to a memory, or to a display. For example, the image is loaded into a buffer and output from the buffer to a display for viewing by a physician to aid diagnosis or pre-operative planning. A computer-generated representation of the medical data suitable for viewing by human users (e.g., 2D images, videos, stereoscopic image pairs, volumes, holograms, or lightfields) is output. Alternatively, the output may be for storage and/or for further machine analysis. - The transmission may be for any one or more of various purposes. In one embodiment, the transmission of the photorealistic image is part of a diagnostic report. A given diagnostic report may be specific to context, such as a report for kidney stones. Any kidney stones should be visible in an image. The photorealistic image shows kidney stones based on the values of the rendering parameters output by the model trained to create such images. The optimal context-specific images are provided for an automated diagnostic report in the clinical workflow.
- In another embodiment, the transmission is to a display as an initial image for interactive viewing. The machine-learnt model is used to initialize interactive viewing applications in the clinical workflow. By providing values for rendering parameters so that the images are consistent with an established visualization design, the amount of user interaction may be reduced.
- In yet another embodiment, the transmission is to an augmented reality device. For example, a physician may wear augmented reality glasses. The photorealistic image, with or without added annotations or graphics, is projected onto the glasses as an overlay on the actual view seen by the wearer. In other example, the photorealistic image is combined with a camera image and/or is used as a replacement for the camera image. In an augmented reality setup, the rendered image may be overlaid onto a live stream of real-time medical images (ultrasound, x-ray, optical laparoscopy, etc.).
-
FIG. 3 shows a block diagram of one embodiment of a system for content-based rendering based on machine learning. The system is for training with machine learning and/or application of a machine-learnt model. Using deep learning, learning to provide consistent images despite variability of input scan data, and/or learning a plurality of physically-based rendering parameters, a machine-learnt model is trained to output settings for rendering parameters given a new volume for a patient. - The system implements the method of
FIGS. 1 and/or 2 . Other methods or acts may be implemented, such as providing a user input (e.g., mouse, trackball, touch pad, and/or keyboard) and user interface for interactive rendering. - The system includes a
machine 50, amemory 52, adisplay 54, amedical imager 56, and arenderer 58. Additional, different, or fewer components may be provided. For example, themedical imager 56 and/ormemory 52 are not provided. In another example, a network or network connection is provided, such as for networking with a medical imaging network or data archival system. A user interface may be provided for interacting with themachine 50,renderer 58, or other components. - The
machine 50,memory 52,renderer 58, and/ordisplay 54 are part of themedical imager 56. Alternatively, themachine 50,memory 52,renderer 58, and/ordisplay 54 are part of a server, workstation, or computer separate from themedical imager 56. Themachine 50,memory 52,renderer 58, and/ordisplay 54 are a personal computer, such as desktop or laptop, a workstation, a server, a network, or combinations thereof. In yet other embodiments, themachine 50 andmemory 52 are part of a separate computer from therenderer 58. - The
medical imager 56 is a medical diagnostic imaging system. Ultrasound, CT, x-ray, fluoroscopy, positron emission tomography (PET), single photon emission computed tomography (SPECT), and/or MR systems may be used. Themedical imager 56 may include a transmitter and includes a detector for scanning or receiving data representative of the interior of the patient. Themedical imager 56 acquires scan data representing the patient. The scan data may represent a volume of the patient. For example, scan data is acquired and used for diagnosis or surgical planning, such as identifying a lesion or treatment location. - In alternative embodiments, the
medical imager 56 is not provided, but a previously acquired dataset for a patient is stored in thememory 52. In yet other alternatives, manymedical images 56 are provided in thememory 52 as the training data, which is gathered and stored in thememory 52. - The
machine 50 is a computer, workstation, server, processor, or other device configured to apply machine learning and/or to apply a machine-learnt model. Themachine 50 is configured by software, hardware, and/or firmware. For learning, themachine 50 is configured by one or more machine learning algorithms. For applying a learnt model, themachine 50 is configured, in part, by a learnt matrix or matrices associating input data to output values. - The
machine 50 applies a machine-learnt model, such as one learnt with deep learning. The machine-learnt model, as implemented by themachine 50, generates input features and/or outputs settings for rendering parameters based on application of medical data from a scan of a patient. The rendering parameters include windowing, transfer function, lighting, and/or material properties. The settings are learnt to provide an image similar in look and/or feel to one or more images from a same diagnostic context. Consistent images, such as to highlight or show anatomy of interest, are created. The diagnostic context provides the standard or information of interest. - The
memory 52 is a graphics processing memory, a video random access memory, a random access memory, system memory, cache memory, hard drive, optical media, magnetic media, flash drive, buffer, database, combinations thereof, or other now known or later developed memory device for storing training data, rendering parameters values, medical datasets, and/or other information. Thememory 52 is part of themedical imager 56, part of a computer associated with themachine 50, part of a computer associated with therenderer 58, a database, part of another system, a picture archival memory, or a standalone device. - The
memory 52 stores scan data representing one or more patients. For example, data from themedical imager 56 is stored. The data is in a scan format or reconstructed to a volume or three-dimensional grid format. The scan data is stored for training and/or stored for rendering an image of a patient. - The
memory 52 or other memory is alternatively or additionally a non-transitory computer readable storage medium storing data representing instructions executable by the programmedmachine 50 for learning or applying the machine-learnt model. The instructions for implementing the processes, methods, and/or techniques discussed herein are provided on non-transitory computer-readable storage media or memories, such as a cache, buffer, RAM, removable media, hard drive, or other computer readable storage media. Non-transitory computer readable storage media include various types of volatile and nonvolatile storage media. The functions, acts or tasks illustrated in the figures or described herein are executed in response to one or more sets of instructions stored in or on computer readable storage media. The functions, acts or tasks are independent of the particular type of instructions set, storage media, processor or processing strategy and may be performed by software, hardware, integrated circuits, firmware, micro code and the like, operating alone, or in combination. Likewise, processing strategies may include multiprocessing, multitasking, parallel processing, and the like. - In one embodiment, the instructions are stored on a removable media device for reading by local or remote systems. In other embodiments, the instructions are stored in a remote location for transfer through a computer network or over telephone lines. In yet other embodiments, the instructions are stored within a given computer, CPU, GPU, or system.
- The
renderer 58 is a graphics processing unit, graphics card, graphic chip, multi-core processor, or other processor for running a software package that implements the photorealistic image generation functionality. Therenderer 58 is configured by an application programming interface to render an image from the 3D scan data representing a patient. Using physically-based rendering, a photorealistic image is rendered. Using the settings output by the machine-learnt model, path tracing is applied to render the image. For feedback in training, therenderer 58 may render training images. Similarity is calculated by therenderer 58 and/or themachine 50. - The
display 54 is a monitor, LCD, projector, plasma display, CRT, printer, or other now known or later developed device for displaying the photorealistic image or images. Thedisplay 54 receives images from themachine 50,memory 52,renderer 58, ormedical imager 56. The images of the tissue captured by themedical imager 56 are displayed. Other information may be displayed as well, such as generated graphics, text, or quantities as a virtual overlay. - Additional images may be displayed. Where scan data represents a sequence of scans over time, a corresponding sequence of images may be generated. For interactive rendering, new images may be generated as settings for one or more rendering parameters are changed by a user.
- While the invention has been described above by reference to various embodiments, it should be understood that many changes and modifications can be made without departing from the scope of the invention. It is therefore intended that the foregoing detailed description be regarded as illustrative rather than limiting, and that it be understood that it is the following claims, including all equivalents, that are intended to define the spirit and scope of this invention.
Claims (24)
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/191,043 US9760690B1 (en) | 2016-03-10 | 2016-06-23 | Content-based medical image rendering based on machine learning |
CN201780016345.8A CN108701370B (en) | 2016-03-10 | 2017-03-08 | Content-based medical imaging rendering based on machine learning |
EP17711584.7A EP3427230A1 (en) | 2016-03-10 | 2017-03-08 | Content-based medical imaging rendering based on machine learning |
PCT/EP2017/055384 WO2017153445A1 (en) | 2016-03-10 | 2017-03-08 | Content-based medical imaging rendering based on machine learning |
US15/643,973 US10339695B2 (en) | 2016-03-10 | 2017-07-07 | Content-based medical image rendering based on machine learning |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201662306407P | 2016-03-10 | 2016-03-10 | |
US15/191,043 US9760690B1 (en) | 2016-03-10 | 2016-06-23 | Content-based medical image rendering based on machine learning |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/643,973 Continuation US10339695B2 (en) | 2016-03-10 | 2017-07-07 | Content-based medical image rendering based on machine learning |
Publications (2)
Publication Number | Publication Date |
---|---|
US9760690B1 US9760690B1 (en) | 2017-09-12 |
US20170262598A1 true US20170262598A1 (en) | 2017-09-14 |
Family
ID=59752844
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/191,043 Active US9760690B1 (en) | 2016-03-10 | 2016-06-23 | Content-based medical image rendering based on machine learning |
US15/643,973 Active 2036-08-08 US10339695B2 (en) | 2016-03-10 | 2017-07-07 | Content-based medical image rendering based on machine learning |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/643,973 Active 2036-08-08 US10339695B2 (en) | 2016-03-10 | 2017-07-07 | Content-based medical image rendering based on machine learning |
Country Status (4)
Country | Link |
---|---|
US (2) | US9760690B1 (en) |
EP (1) | EP3427230A1 (en) |
CN (1) | CN108701370B (en) |
WO (1) | WO2017153445A1 (en) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3553695A1 (en) * | 2018-04-13 | 2019-10-16 | Alpine Electronics, Inc. | Method and system for displaying image data from at least one night vision camera of a vehicle |
US10592743B2 (en) * | 2017-08-24 | 2020-03-17 | International Business Machines Corporation | Machine learning to predict cognitive image composition |
US10643331B2 (en) | 2017-05-03 | 2020-05-05 | Siemens Healthcare Gmbh | Multi-scale deep reinforcement machine learning for N-dimensional segmentation in medical imaging |
WO2020139775A1 (en) * | 2018-12-27 | 2020-07-02 | Exo Imaging, Inc. | Methods to maintain image quality in ultrasound imaging at reduced cost, size, and power |
US10835209B2 (en) | 2016-12-04 | 2020-11-17 | Exo Imaging Inc. | Configurable ultrasonic imager |
US20210093301A1 (en) * | 2018-04-09 | 2021-04-01 | Koninklijke Philips N.V. | Ultrasound system with artificial neural network for retrieval of imaging parameter settings for recurring patient |
US11132775B2 (en) * | 2018-11-16 | 2021-09-28 | Samsung Electronics Co., Ltd. | Image processing apparatus and method of operating the same |
US11199623B2 (en) | 2020-03-05 | 2021-12-14 | Exo Imaging, Inc. | Ultrasonic imaging device with programmable anatomy and flow imaging |
US11364013B2 (en) * | 2017-01-05 | 2022-06-21 | Koninklijke Philips N.V. | Ultrasound imaging system with a neural network for image formation and tissue characterization |
US11488297B2 (en) * | 2018-12-18 | 2022-11-01 | Canon Medical Systems Corporation | Medical information processing apparatus and medical information processing system |
US11574739B2 (en) * | 2018-01-29 | 2023-02-07 | Function, Inc. | Systems and methods for formulating personalized skincare products |
US11971477B2 (en) | 2018-09-25 | 2024-04-30 | Exo Imaging, Inc. | Imaging devices with selectively alterable characteristics |
US12092735B2 (en) | 2019-10-16 | 2024-09-17 | Koninklijke Philips N.V. | Method and apparatus for deep learning-based ultrasound beamforming |
Families Citing this family (63)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9069725B2 (en) | 2011-08-19 | 2015-06-30 | Hartford Steam Boiler Inspection & Insurance Company | Dynamic outlier bias reduction system and method |
JP6795488B2 (en) | 2014-04-11 | 2020-12-02 | ハートフォード スチーム ボイラー インスペクション アンド インシュランス カンパニー | Improved future reliability forecasts based on modeling system operation and performance data |
US9569736B1 (en) * | 2015-09-16 | 2017-02-14 | Siemens Healthcare Gmbh | Intelligent medical image landmark detection |
US10991070B2 (en) | 2015-12-18 | 2021-04-27 | OrthoGrid Systems, Inc | Method of providing surgical guidance |
US20180025664A1 (en) * | 2016-07-25 | 2018-01-25 | Anna Clarke | Computerized methods and systems for motor skill training |
US10049301B2 (en) * | 2016-08-01 | 2018-08-14 | Siemens Healthcare Gmbh | Medical scanner teaches itself to optimize clinical protocols and image acquisition |
CN110392903B (en) | 2016-08-19 | 2024-08-30 | 莫维迪厄斯有限公司 | Dynamic culling of matrix operations |
US10466714B2 (en) * | 2016-09-01 | 2019-11-05 | Ford Global Technologies, Llc | Depth map estimation with stereo images |
US10453165B1 (en) * | 2017-02-27 | 2019-10-22 | Amazon Technologies, Inc. | Computer vision machine learning model execution service |
US10977397B2 (en) | 2017-03-10 | 2021-04-13 | Altair Engineering, Inc. | Optimization of prototype and machine design within a 3D fluid modeling environment |
US10607393B2 (en) | 2017-03-10 | 2020-03-31 | Siemens Healthcare Gmbh | Consistent 3D rendering in medical imaging |
US10867085B2 (en) | 2017-03-10 | 2020-12-15 | General Electric Company | Systems and methods for overlaying and integrating computer aided design (CAD) drawings with fluid models |
US10409950B2 (en) | 2017-03-10 | 2019-09-10 | General Electric Company | Systems and methods for utilizing a 3D CAD point-cloud to automatically create a fluid model |
US11004568B2 (en) * | 2017-03-10 | 2021-05-11 | Altair Engineering, Inc. | Systems and methods for multi-dimensional fluid modeling of an organism or organ |
US10803211B2 (en) | 2017-03-10 | 2020-10-13 | General Electric Company | Multiple fluid model tool for interdisciplinary fluid modeling |
US10475165B2 (en) * | 2017-04-06 | 2019-11-12 | Disney Enterprises, Inc. | Kernel-predicting convolutional neural networks for denoising |
US10643368B2 (en) | 2017-06-27 | 2020-05-05 | The Boeing Company | Generative image synthesis for training deep learning machines |
US10699410B2 (en) * | 2017-08-17 | 2020-06-30 | Siemes Healthcare GmbH | Automatic change detection in medical images |
US11264135B2 (en) | 2017-11-10 | 2022-03-01 | Siemens Medical Solutions Usa, Inc. | Machine-aided workflow in ultrasound imaging |
US10593099B2 (en) | 2017-11-14 | 2020-03-17 | Siemens Healthcare Gmbh | Transfer function determination in medical imaging |
US10643320B2 (en) | 2017-11-15 | 2020-05-05 | Toyota Research Institute, Inc. | Adversarial learning of photorealistic post-processing of simulation with privileged information |
CN107910061A (en) * | 2017-12-01 | 2018-04-13 | 中南大学 | A kind of medical data processing method and system |
US10722210B2 (en) * | 2017-12-14 | 2020-07-28 | Siemens Healthcare Gmbh | Method for memorable image generation for anonymized three-dimensional medical image workflows |
US11379715B2 (en) * | 2017-12-15 | 2022-07-05 | Meta Platforms, Inc. | Deep learning based distribution of content items describing events to users of an online system |
US10521908B2 (en) | 2017-12-20 | 2019-12-31 | International Business Machines Corporation | User interface for displaying simulated anatomical photographs |
US10438351B2 (en) | 2017-12-20 | 2019-10-08 | International Business Machines Corporation | Generating simulated photographic anatomical slices |
US10614570B2 (en) | 2017-12-20 | 2020-04-07 | International Business Machines Corporation | Medical image exam navigation using simulated anatomical photographs |
US10665007B2 (en) * | 2018-01-15 | 2020-05-26 | Siemens Healthcare Gmbh | Hybrid interactive mode for rendering medical images with ray tracing |
CN108495110B (en) * | 2018-01-19 | 2020-03-17 | 天津大学 | Virtual viewpoint image generation method based on generation type countermeasure network |
US11443201B2 (en) * | 2018-05-21 | 2022-09-13 | Siemens Healthcare Gmbh | Artificial intelligence-based self-learning in medical imaging |
CN109147003A (en) * | 2018-08-01 | 2019-01-04 | 北京东方畅享科技有限公司 | Method, equipment and the storage medium painted to line manuscript base picture |
CN109389667B (en) * | 2018-08-23 | 2023-07-18 | 北京大学 | High-efficiency global illumination drawing method based on deep learning |
JP7466928B2 (en) | 2018-09-12 | 2024-04-15 | オルソグリッド システムズ ホールディング,エルエルシー | Artificial intelligence intraoperative surgical guidance systems and methods of use |
US11540794B2 (en) | 2018-09-12 | 2023-01-03 | Orthogrid Systesm Holdings, LLC | Artificial intelligence intra-operative surgical guidance system and method of use |
US11636292B2 (en) | 2018-09-28 | 2023-04-25 | Hartford Steam Boiler Inspection And Insurance Company | Dynamic outlier bias reduction system and method |
CN109273084B (en) * | 2018-11-06 | 2021-06-22 | 中山大学附属第一医院 | Method and system based on multi-mode ultrasound omics feature modeling |
CN109671140B (en) * | 2018-12-26 | 2024-02-02 | 上海赞奇文化科技有限公司 | Cloud rendering service processing method adopting micro-service |
US10973472B2 (en) * | 2019-03-05 | 2021-04-13 | Siemens Healthcare Gmbh | Artificial intelligence-based material decomposition in medical imaging |
CN110051434A (en) * | 2019-04-25 | 2019-07-26 | 厦门强本科技有限公司 | AR operation piloting method and terminal in conjunction with endoscope |
US11132826B2 (en) * | 2019-05-16 | 2021-09-28 | Caterpillar Inc. | Artificial image generation for training an object detection system |
CN110277166B (en) * | 2019-06-28 | 2020-05-15 | 屈兴玲 | Auxiliary diagnosis system and method for palace laparoscope |
KR20210012730A (en) | 2019-07-26 | 2021-02-03 | 삼성전자주식회사 | Learning method of artificial intelligence model and electronic apparatus |
US11328177B2 (en) | 2019-09-18 | 2022-05-10 | Hartford Steam Boiler Inspection And Insurance Company | Computer-based systems, computing components and computing objects configured to implement dynamic outlier bias reduction in machine learning models |
JP7399269B2 (en) | 2019-09-18 | 2023-12-15 | ハートフォード スチーム ボイラー インスペクション アンド インシュアランス カンパニー | Computer-based systems, computer components and computer objects configured to implement dynamic outlier bias reduction in machine learning models |
US11615348B2 (en) | 2019-09-18 | 2023-03-28 | Hartford Steam Boiler Inspection And Insurance Company | Computer-based systems, computing components and computing objects configured to implement dynamic outlier bias reduction in machine learning models |
ES2815623A1 (en) * | 2019-09-27 | 2021-03-30 | Delgado Oscar Ruesga | PROCEDURE FOR PERFORMING A DIAGNOSIS and DIGITAL MEDICAL FORECAST BASED ON DIGITAL ANATOMY (Machine-translation by Google Translate, not legally binding) |
US11398072B1 (en) * | 2019-12-16 | 2022-07-26 | Siemens Healthcare Gmbh | Method of obtaining a set of values for a respective set of parameters for use in a physically based path tracing process and a method of rendering using a physically based path tracing process |
CN111353967B (en) * | 2020-03-06 | 2021-08-24 | 浙江杜比医疗科技有限公司 | Image acquisition method and device, electronic equipment and readable storage medium |
US20230281959A1 (en) * | 2020-03-25 | 2023-09-07 | Yissum Research Development Company Of The Hebrew University Of Jerusalem Ltd. | Deep learning-based anomaly detection in images |
CN111310859A (en) * | 2020-03-26 | 2020-06-19 | 上海景和国际展览有限公司 | Rapid artificial intelligence data training system used in multimedia display |
CN111739617B (en) * | 2020-06-15 | 2023-07-11 | 中国医学科学院医学信息研究所 | Medical image artificial intelligence quality control labeling method, device, equipment and storage medium |
CN112308888B (en) * | 2020-10-15 | 2024-07-12 | 复旦大学附属中山医院 | Full-mode medical image sequence grouping method based on deep learning physical sign structure |
US11635878B2 (en) * | 2021-01-31 | 2023-04-25 | Walmart Apollo, Llc | Systems and methods for altering a graphical user interface |
US20240156439A1 (en) * | 2021-04-01 | 2024-05-16 | Koninklijke Philips N.V. | Systems and methods for self-sensing mobile medical devices to auto-correct an output |
US11848095B2 (en) | 2021-04-29 | 2023-12-19 | Lymphatech, Inc. | Identifying body part or body area anatomical landmarks from digital imagery for the fitting of compression garments for a person in need thereof |
US11527320B2 (en) | 2021-04-29 | 2022-12-13 | Lymphatech, Inc. | Methods and systems for generating accurately fitting compression garments having glove or glove-like features for a person in need thereof |
US11211162B1 (en) | 2021-04-29 | 2021-12-28 | Lymphatech, Inc. | Methods and systems for identifying body part or body area anatomical landmarks from digital imagery for the fitting of compression garments for a person in need thereof |
US11720228B2 (en) | 2021-08-05 | 2023-08-08 | Evernorth Strategic Development, Inc. | Systems and methods for transforming an interactive graphical user interface according to machine learning models |
CN114332594B (en) * | 2021-08-27 | 2024-09-06 | 吉林大学 | Method for classifying unbalanced data of haptic material based on DDQN |
CN113706583B (en) * | 2021-09-01 | 2024-03-22 | 上海联影医疗科技股份有限公司 | Image processing method, device, computer equipment and storage medium |
CN114120062B (en) * | 2021-11-26 | 2023-07-28 | 北京百度网讯科技有限公司 | Sample generation method and device |
EP4254350A1 (en) * | 2022-04-01 | 2023-10-04 | Siemens Healthcare GmbH | Determination of illumination parameters in medical image rendering |
CN117274473B (en) * | 2023-11-21 | 2024-02-02 | 北京渲光科技有限公司 | Multiple scattering real-time rendering method and device and electronic equipment |
Family Cites Families (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6463438B1 (en) * | 1994-06-03 | 2002-10-08 | Urocor, Inc. | Neural network for cell image analysis for identification of abnormal cells |
US6690371B1 (en) * | 2000-05-03 | 2004-02-10 | Ge Medical Systems Global Technology, Llc | Relevant image data extraction from a medical image data volume |
US6826568B2 (en) * | 2001-12-20 | 2004-11-30 | Microsoft Corporation | Methods and system for model matching |
US6999549B2 (en) * | 2002-11-27 | 2006-02-14 | Ge Medical Systems Global Technology, Llc | Method and apparatus for quantifying tissue fat content |
US8711144B2 (en) | 2006-08-01 | 2014-04-29 | Siemens Medical Solutions Usa, Inc. | Perception-based artifact quantification for volume rendering |
US7840076B2 (en) * | 2006-11-22 | 2010-11-23 | Intel Corporation | Methods and apparatus for retrieving images from a large collection of images |
US8520906B1 (en) * | 2007-09-24 | 2013-08-27 | Videomining Corporation | Method and system for age estimation based on relative ages of pairwise facial images of people |
US10108620B2 (en) * | 2010-04-29 | 2018-10-23 | Google Llc | Associating still images and videos |
CN103220974A (en) * | 2010-07-21 | 2013-07-24 | 阿敏·E·莫尔勒 | Image reporting method |
US20120078062A1 (en) * | 2010-09-24 | 2012-03-29 | International Business Machines Corporation | Decision-support application and system for medical differential-diagnosis and treatment using a question-answering system |
US9043316B1 (en) * | 2011-03-28 | 2015-05-26 | Google Inc. | Visual content retrieval |
US8509525B1 (en) * | 2011-04-06 | 2013-08-13 | Google Inc. | Clustering of forms from large-scale scanned-document collection |
US9020233B2 (en) | 2011-09-19 | 2015-04-28 | Siemens Aktiengesellschaft | Method and system for up-vector detection for ribs in computed tomography volumes |
US8831358B1 (en) * | 2011-11-21 | 2014-09-09 | Google Inc. | Evaluating image similarity |
US9235875B2 (en) | 2012-11-01 | 2016-01-12 | Google Inc. | Image enhancement using learned non-photorealistic effects |
WO2014152339A1 (en) * | 2013-03-14 | 2014-09-25 | Robert Bosch Gmbh | Time and environment aware graphical displays for driver information and driver assistance systems |
US9317777B2 (en) * | 2013-10-04 | 2016-04-19 | Monotype Imaging Inc. | Analyzing font similarity for presentation |
US9679258B2 (en) * | 2013-10-08 | 2017-06-13 | Google Inc. | Methods and apparatus for reinforcement learning |
US9008391B1 (en) * | 2013-10-22 | 2015-04-14 | Eyenuk, Inc. | Systems and methods for processing retinal images for screening of diseases or abnormalities |
WO2015099810A1 (en) * | 2013-12-29 | 2015-07-02 | Hewlett-Packard Development Company, L.P. | Learning graph |
US9324022B2 (en) * | 2014-03-04 | 2016-04-26 | Signal/Sense, Inc. | Classifying data with deep learning neural records incrementally refined through expert input |
US9412046B2 (en) * | 2014-10-10 | 2016-08-09 | Facebook, Inc. | Training image adjustment preferences |
US9454713B2 (en) * | 2014-12-30 | 2016-09-27 | Ebay Inc. | Similar item detection |
US9842390B2 (en) * | 2015-02-06 | 2017-12-12 | International Business Machines Corporation | Automatic ground truth generation for medical image collections |
US10905400B2 (en) * | 2015-02-23 | 2021-02-02 | Canon Medical Systems Corporation | Apparatus and method for optimization of ultrasound images |
US20160321523A1 (en) * | 2015-04-30 | 2016-11-03 | The Regents Of The University Of California | Using machine learning to filter monte carlo noise from images |
US20170161607A1 (en) * | 2015-12-04 | 2017-06-08 | Pilot Ai Labs, Inc. | System and method for improved gesture recognition using neural networks |
US10068385B2 (en) * | 2015-12-15 | 2018-09-04 | Intel Corporation | Generation of synthetic 3-dimensional object images for recognition systems |
-
2016
- 2016-06-23 US US15/191,043 patent/US9760690B1/en active Active
-
2017
- 2017-03-08 WO PCT/EP2017/055384 patent/WO2017153445A1/en active Application Filing
- 2017-03-08 EP EP17711584.7A patent/EP3427230A1/en not_active Ceased
- 2017-03-08 CN CN201780016345.8A patent/CN108701370B/en active Active
- 2017-07-07 US US15/643,973 patent/US10339695B2/en active Active
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11759175B2 (en) | 2016-12-04 | 2023-09-19 | Exo Imaging, Inc. | Configurable ultrasonic imager |
US10835209B2 (en) | 2016-12-04 | 2020-11-17 | Exo Imaging Inc. | Configurable ultrasonic imager |
US11712222B2 (en) | 2016-12-04 | 2023-08-01 | Exo Imaging, Inc. | Configurable ultrasonic imager |
US11058396B2 (en) | 2016-12-04 | 2021-07-13 | Exo Imaging Inc. | Low voltage, low power MEMS transducer with direct interconnect capability |
US11364013B2 (en) * | 2017-01-05 | 2022-06-21 | Koninklijke Philips N.V. | Ultrasound imaging system with a neural network for image formation and tissue characterization |
US10643331B2 (en) | 2017-05-03 | 2020-05-05 | Siemens Healthcare Gmbh | Multi-scale deep reinforcement machine learning for N-dimensional segmentation in medical imaging |
US10592743B2 (en) * | 2017-08-24 | 2020-03-17 | International Business Machines Corporation | Machine learning to predict cognitive image composition |
US11468664B2 (en) * | 2017-08-24 | 2022-10-11 | International Business Machines Corporation | Machine learning to predict cognitive image composition |
US11574739B2 (en) * | 2018-01-29 | 2023-02-07 | Function, Inc. | Systems and methods for formulating personalized skincare products |
US20210093301A1 (en) * | 2018-04-09 | 2021-04-01 | Koninklijke Philips N.V. | Ultrasound system with artificial neural network for retrieval of imaging parameter settings for recurring patient |
US11948345B2 (en) * | 2018-04-09 | 2024-04-02 | Koninklijke Philips N.V. | Ultrasound system with artificial neural network for retrieval of imaging parameter settings for recurring patient |
EP3553695A1 (en) * | 2018-04-13 | 2019-10-16 | Alpine Electronics, Inc. | Method and system for displaying image data from at least one night vision camera of a vehicle |
US11971477B2 (en) | 2018-09-25 | 2024-04-30 | Exo Imaging, Inc. | Imaging devices with selectively alterable characteristics |
US11132775B2 (en) * | 2018-11-16 | 2021-09-28 | Samsung Electronics Co., Ltd. | Image processing apparatus and method of operating the same |
US11488297B2 (en) * | 2018-12-18 | 2022-11-01 | Canon Medical Systems Corporation | Medical information processing apparatus and medical information processing system |
WO2020139775A1 (en) * | 2018-12-27 | 2020-07-02 | Exo Imaging, Inc. | Methods to maintain image quality in ultrasound imaging at reduced cost, size, and power |
US12061257B2 (en) | 2018-12-27 | 2024-08-13 | Exo Imaging, Inc. | Methods to maintain image quality in ultrasound imaging at reduced cost, size, and power |
US12092735B2 (en) | 2019-10-16 | 2024-09-17 | Koninklijke Philips N.V. | Method and apparatus for deep learning-based ultrasound beamforming |
US11199623B2 (en) | 2020-03-05 | 2021-12-14 | Exo Imaging, Inc. | Ultrasonic imaging device with programmable anatomy and flow imaging |
US12019155B2 (en) | 2020-03-05 | 2024-06-25 | Exo Imaging, Inc. | Ultrasonic imaging device with programmable anatomy and flow imaging |
Also Published As
Publication number | Publication date |
---|---|
US20170308656A1 (en) | 2017-10-26 |
EP3427230A1 (en) | 2019-01-16 |
WO2017153445A1 (en) | 2017-09-14 |
CN108701370A (en) | 2018-10-23 |
US10339695B2 (en) | 2019-07-02 |
US9760690B1 (en) | 2017-09-12 |
CN108701370B (en) | 2020-01-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10339695B2 (en) | Content-based medical image rendering based on machine learning | |
US10957098B2 (en) | Consistent 3D rendering in medical imaging | |
US10643401B2 (en) | Two-dimensional cinematic medical imaging in color based on deep learning | |
US10643331B2 (en) | Multi-scale deep reinforcement machine learning for N-dimensional segmentation in medical imaging | |
US10893262B2 (en) | Lightfield rendering based on depths from physically-based volume rendering | |
US10997475B2 (en) | COPD classification with machine-trained abnormality detection | |
US12062429B2 (en) | Salient visual explanations of feature assessments by machine learning models | |
US11727086B2 (en) | Multimodality image processing techniques for training image data generation and usage thereof for developing mono-modality image inferencing models | |
US12033247B2 (en) | Three-dimensional shape reconstruction from a topogram in medical imaging | |
US20190183451A1 (en) | Method for memorable image generation for anonymized three-dimensional medical image workflows | |
CN113256553A (en) | System and method for consistently presenting medical images using deep neural networks | |
US20220076829A1 (en) | Method and apparatus for analyzing medical image data in a latent space representation | |
CN114387380A (en) | Method for generating a computer-based visualization of 3D medical image data | |
US20240037731A1 (en) | Image processing method and apparatus | |
CN112541882A (en) | Implicit surface coloring in medical volume rendering |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PETKOV, KALOIAN;MIAO, SHUN;YU, DAPHNE;AND OTHERS;SIGNING DATES FROM 20160603 TO 20160621;REEL/FRAME:039120/0696 |
|
AS | Assignment |
Owner name: SIEMENS HEALTHCARE GMBH, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS MEDICAL SOLUTIONS USA, INC.;REEL/FRAME:040511/0509 Effective date: 20161114 Owner name: SIEMENS HEALTHCARE GMBH, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ENGEL, KLAUS;REEL/FRAME:040511/0542 Effective date: 20161115 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: SIEMENS HEALTHINEERS AG, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS HEALTHCARE GMBH;REEL/FRAME:066267/0346 Effective date: 20231219 |