US20170262598A1 - Content-based medical image rendering based on machine learning - Google Patents

Content-based medical image rendering based on machine learning Download PDF

Info

Publication number
US20170262598A1
US20170262598A1 US15/191,043 US201615191043A US2017262598A1 US 20170262598 A1 US20170262598 A1 US 20170262598A1 US 201615191043 A US201615191043 A US 201615191043A US 2017262598 A1 US2017262598 A1 US 2017262598A1
Authority
US
United States
Prior art keywords
rendering
machine
image
training
parameters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/191,043
Other versions
US9760690B1 (en
Inventor
Kaloian Petkov
Shun Miao
Daphne Yu
Bogdan Georgescu
Klaus Engel
Tommaso Mansi
Dorin Comaniciu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Healthineers AG
Original Assignee
Siemens Healthcare GmbH
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US15/191,043 priority Critical patent/US9760690B1/en
Application filed by Siemens Healthcare GmbH filed Critical Siemens Healthcare GmbH
Assigned to SIEMENS MEDICAL SOLUTIONS USA, INC. reassignment SIEMENS MEDICAL SOLUTIONS USA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MIAO, SHUN, GEORGESCU, BOGDAN, MANSI, TOMMASO, COMANICIU, DORIN, PETKOV, KALOIAN, YU, DAPHNE
Assigned to SIEMENS HEALTHCARE GMBH reassignment SIEMENS HEALTHCARE GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ENGEL, KLAUS
Assigned to SIEMENS HEALTHCARE GMBH reassignment SIEMENS HEALTHCARE GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS MEDICAL SOLUTIONS USA, INC.
Priority to CN201780016345.8A priority patent/CN108701370B/en
Priority to EP17711584.7A priority patent/EP3427230A1/en
Priority to PCT/EP2017/055384 priority patent/WO2017153445A1/en
Priority to US15/643,973 priority patent/US10339695B2/en
Publication of US9760690B1 publication Critical patent/US9760690B1/en
Application granted granted Critical
Publication of US20170262598A1 publication Critical patent/US20170262598A1/en
Assigned to Siemens Healthineers Ag reassignment Siemens Healthineers Ag ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS HEALTHCARE GMBH
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • G06F19/345
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • G06F18/24137Distances to cluster centroïds
    • G06F18/2414Smoothing the distance, e.g. radial basis function networks [RBFN]
    • G06F19/321
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/06Ray-tracing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/08Volume rendering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/50Lighting effects
    • G06T15/506Illumination models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/20ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/41Medical
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/03Recognition of patterns in medical or anatomical images

Definitions

  • the present embodiments relate to medical imaging of three-dimensional (3D) scans.
  • Data representing a volume is rendered for visualization. Due to the many different scan settings and patient variability, renderings for different patients or at different times appear different.
  • Variability is particularly troublesome for physically-based volume rendering, which relies on the physical simulation of light propagation (e.g., unbiased path tracing).
  • Physically-based visualization techniques produce global illumination effects in computer-generated graphics that mimic the real-world interaction of light with various 3D objects or tissues. This results in physically plausible images that are often easier for the human brain to interpret when compared to the more analytical images from traditional rendering. For example, subtle ambient occlusions, shadows, and color bleeding provide important depth cues for the spatial understanding of 3D relationships between structures in a single 2D image, whereas simpler visualization techniques may require additional interaction with the viewing parameters (e.g., moving the virtual camera around the 3D data) to obtain the same spatial information from the image parallax.
  • the preferred embodiments described below include methods, systems, instructions, and computer readable media for machine training an artificial intelligence and use of that artificial intelligence to provide rendering settings.
  • settings of multiple rendering parameters may be provided for consistent imaging or visualization even in physically-based rendering.
  • a method for content-based rendering based on machine learning in a rendering system.
  • a medical dataset representing a three-dimensional region of a patient is loaded from memory.
  • a machine applies the medical dataset to a machine-learnt non-linear model.
  • the machine-learned model is trained with deep learning to extract features from the medical dataset and trained to output values for two or more physically-based rendering parameters based on input of the medical dataset.
  • At least one of the two or more physically-based rendering parameters are controls for consistent data handling, lighting design, viewing design, material propriety, or internal renderer property.
  • a physically-based renderer renders a photorealistic image of the three-dimensional region of the patient using the output values resulting from the applying. The photorealistic image is transmitted.
  • a method for machine training for content-based rendering in a machine training system.
  • First volume data of a volume of a patient, a first image of the volume, and first values of path tracing rendering parameters are input to training of an artificial intelligence.
  • a machine trains the artificial intelligence to output second values of the path tracing rendering parameters for a second volume data where the second values provide a second rendered image of the second volume modeled on the first image.
  • the trained artificial intelligence is stored.
  • a system for content-based rendering based on machine learning.
  • a medical scanner is configured to scan a patient.
  • a machine is configured to output settings for rendering parameters by application of data from the scan to a machine-learnt model.
  • the rendering parameters include windowing, transfer function, and lighting, and the settings are learned to provide a first image from the data similar to one or more second images for a same diagnostic context.
  • a graphics processing unit is configured to render the first image using the settings.
  • FIG. 1 is a flow chart diagram of one embodiment of a method for machine training for content-based rendering
  • FIG. 2 is a flow chart diagram of an embodiment of a method for content-based rendering based on machine learning in a rendering system
  • FIG. 3 is a block diagram of one embodiment of a system for content-based rendering based on machine learning.
  • Content-based photorealistic rendering of medical images is based on machine learning.
  • the machine learnt non-linear model determines rendering settings based on the content of the 3D medical scan. Due to the variability between scan parameters, data contrast, noise, and/or the large number of rendering parameters, producing rendered images that allow for consistent visual interpretation has been a long standing challenge, which has prevented wider adoption 3D images for definitive decision support.
  • deep learning generates the settings for the rendering parameters for the photorealistic rendering components, establishing a consistent visualization design.
  • This design may be applied to new 3D image data in order to provide reliable visual interpretation within the diagnostic workflow.
  • Image features, and optionally, available non-image data are mapped to sets of rendering parameters that produce optimal images, where the relationship between the two is highly non-linear in the general case.
  • Consistent visual interpretation for 3D images may potentially simplify diagnostic workflow and may promote the clinical use of 3D data for decision support.
  • Consistent photorealistic rendered images may be provided in new platforms and workflows, such as fully automated reporting and cloud-based rendering services.
  • Consistent visualization design as a machine learning problem may be applied 1) to optimize rendering preset designs tailored to a large variety of data types in a unified and scalable way, 2) to adapt rendering parameters to produce target consistent results based on the provided input, 3) to adapt rendering parameters to specific individual preferences, 4) to transfer expert trained results to non-expert users, 5) to increase the conspicuity of visual features in the rendered images (e.g., pathologies in medical images) and 6) to provide automatic rendering presets for specific workflows.
  • a scalable framework is provided for fully automated image generation in the clinical workflow.
  • FIG. 1 shows one embodiment of a method for machine training for content-based rendering in a machine training system.
  • Optimal rendering parameters may be obtained via training an artificial intelligence system.
  • the input of the artificial intelligence system may include, but is not limited to, the 3D scan data to be rendered, images rendered from the 3D scan data, information on the patient, and/or information on the diagnostic purpose.
  • the machine training learns to provide rendering settings for photorealistic rendering based on input data for a specific patient.
  • the output of the trained artificial intelligence system is the optimal set of settings for the rendering parameters.
  • the method of FIG. 1 is implemented by a processor and a graphics processing unit in a server, computer, or other machine.
  • the system of FIG. 3 implements the method of FIG. 1 , with or without the medical imager 56 and/or display 54 .
  • data is input by loading from a memory
  • a processor performs the training based on the data
  • the memory stores the learnt artificial intelligence.
  • the medical scanner and/or memory input the data to a server, which performs the training and outputs to a memory.
  • results may be fed back as input in a repetition of the acts.
  • act 22 is not provided, such as where the trained artificial intelligence is used rather than stored for later use.
  • training data is input.
  • the training data is gathered from one or more memories, sensors, and/or scanners.
  • the gathered data may be stored and/or transmitted to a buffer, memory, cache, processor, or other device for training. This gathered training data is input to the machine learning algorithm.
  • scan data is provided.
  • the scan data represents a patient.
  • a two-dimensional (2D), 3D, 2D+time sequence, 3D+time sequence, and/or other image or scan data may be used.
  • Patient non-image data may be input to the training.
  • Higher dimensional data includes but is not limited to patient information such as a) natural language medical documents or information extracted therefrom of previous diagnosis and/or clinical procedures; b) previous medical test results; c) previously processed medical images and measurements; d) computational modeling results; e) parametric image information derived from medical image analysis algorithms (e.g. cardiac strain map or elasticity); and/or f) other information (e.g., scanner settings).
  • the goal or ground truth information is input.
  • the artificial intelligence system is trained with labeled data (e.g., pairs of input and corresponding desired output (e.g., rendering parameters, optimal rendered images, and/or photographs/movies of the real object)). Based on the provided data, the artificial intelligence system is trained to produce the desired output from any input.
  • the ground truth is provided by desired settings of the rendering parameters for each scan dataset. The values of the path tracing rendering parameters for desired results of each sample are input to the training. Alternatively or additionally, the ground truth is provided as a desired rendered image for each scan dataset. A rendered image for each scan dataset is provided. In other embodiments, an image with desired visualization is provided for multiple datasets. Using a renderer, the machine training may learn the settings that produce similar visualization.
  • the input samples may provide both negative and positive training samples. For example, scan datasets and settings providing desired and undesired images are input. The training learns to distinguish the desired settings from both good and bad settings.
  • additional samples are created by perturbing the path tracing rendering parameters of an input sample.
  • the perturbing creates a collection of sets of the path tracing rendering parameters for each scan data set.
  • the artificial intelligence system is applied on the input scan dataset to generate rendering parameters, which are then perturbed to generate a pool of rendering parameters.
  • Rendering is then performed using the sets of rendering parameters in the pool to produce a pool of corresponding images. Human users or other operators then visually examine the produced images and select the satisfying images.
  • the rendering parameters for the selected images are then used as desired output and/or a sub-set as positive ground-truth.
  • the rendering parameters for the unselected sub-set of images are not used or are used as negative ground-truth.
  • a web-based or network-based service In another approach to gather samples for input, a web-based or network-based service is provided.
  • a website or network portal allows a user to upload the input data (i.e., image and additional information).
  • the server or other processor then runs the artificial intelligence system and/or perturbs possible rendering settings to renderer a pool of images.
  • the service then provides the generated images to the user and asks the user to select the satisfying images.
  • the rendering parameters corresponding to the user selected images are then stored as the desired output for the input (i.e., ground truth).
  • the input may be repeated.
  • the training of the artificial intelligence system is performed periodically or multiple different times. After additional samples are collected (e.g., through the web-based service), the training is repeated using the additional samples with or without previous samples used for earlier iterations of the training.
  • the artificial intelligence or intelligences are machine trained.
  • the collected samples including ground truth, are used to train.
  • a machine such as a processor, computer, server, or other device learns from the samples.
  • complex statistical relationships between large numbers e.g., tens, hundreds, thousands, or more
  • input variables e.g., tens, hundreds, thousands, or more
  • output variables e.g., tens, hundreds, thousands, or more
  • the training is for imaging in any context.
  • One model may be learned for any number of imaging situations.
  • different models are trained for different situations.
  • the different situations may include different scan modalities (e.g., different model for computed tomography, magnetic resonance, ultrasound, positron emission tomography, and single photon emission computed tomography).
  • the different situations may include different types of tissue of interest (e.g., liver versus kidney), different diagnostic purpose or workflow (e.g., cancerous lesion versus bone calcification), and/or different users (e.g., different operators may have different preferences for visualization).
  • Regression, classification, and/or reinforcement learning are used.
  • Regression training learns a range or continuous output by minimization of a metric.
  • Classification learns disparate outputs.
  • Reinforcement learning learns through a sequence of actions with feedback.
  • Neural network, Bayes network, or support vector machine training may be used.
  • Hierarchal or other approaches may be used.
  • Supervised or semi-supervised machine learning may be used.
  • features are extracted from the input data.
  • Haar wavelet, steerable, or other features may be extracted from the scan dataset or images. Measures or other information may be extracted from non-image data.
  • the input data itself e.g., scalar values of the scan dataset.
  • the machine learns features as part of the training.
  • deep learning e.g., deep structured learning, hierarchical learning, or deep machine learning
  • a deep neural network processes the input via multiple layers of feature extraction to produce features used to derive outputs of settings for rendering.
  • the deep learning provides the features used by other machine training to learn to output the rendering settings.
  • Other deep learnt, sparse auto-encoding models may be trained and applied.
  • the machine training is unsupervised in learning the features to use and how to classify given an input sample (i.e., feature vector).
  • Deep learning uses deep learning to improve the input features required for the machine learning. This is especially useful with render settings, which may not have an explicit natural relationship with the input data. Deep learning figures out the data features that end up providing a good classifier or model. Deep learning is known to be effective in extracting high level abstractions (features) from low level input data (e.g., raw images) that is suitable for the trained task, so may be more convenient than and superior to hand-crafted features for identifying features for render settings. In addition, learning-based automatically rendering setting selection is a less studied problem, with no hand-craft feature specifically designed for this problem. Standard hand-craft features, like HAAR feature or SIFT feature, designed for other general computer vision tasks may not be suitable for rendering setting.
  • HAAR feature or SIFT feature designed for other general computer vision tasks may not be suitable for rendering setting.
  • the machine trains to output rendering settings to provide consistent imaging.
  • the training learns to provide rendering settings that result in a rendered image with similar or same visualization despite differences in the content of the data.
  • the path tracing or other physically-based rendering settings so that the rendered image for a given input scan dataset looks and feels similar to an ideal are learned.
  • the machine learns to output renderer settings to model the resulting rendered image for a given scan dataset on a desired image.
  • the learned sets of parameters establish a visualization design that may be reused when new data is introduced to the system in order to provide consistent rendered images. In this way, the machine may learn renderer settings to provide rendering to a standard. Rather than render with the same preset values for a given context, the values are based on a given input dataset.
  • consistency refers to consistent data handling, consistent viewing, and/or consistent visual styles, which may be utilized both in automated and interactive clinical workflows.
  • the ground truth provides consistency. Where the ground truth uses similar or consistent images and/or settings that result in similar or consistent images despite variance in the scan datasets, the training learns to output settings providing the consistency.
  • the notion of optimal rendering parameters and optimal output images relates to using objective quality measurements, subjective quality measurements, clinical significance ratings, and/or other indicators of consistency.
  • the consistency is learned using a measured or calculated metric.
  • the training uses a metric of similarity. Any metric of similarity may be used.
  • Example computed metrics may be visual perception metrics based on visual systems, comparative differences, perceptual hash, mutual information, correlation, computed feature signals, or simple error metrics.
  • the metrics may be automatically computed or provided by the users.
  • the metric is used as an indicator of strength of importance of a given training sample.
  • one or more metrics are used in reinforcement learning.
  • deep reinforcement learning uses a similarity of the rendered image from a sample to a reference image. The similarity is used as a reinforcement.
  • the output image rendered from a given set of rendering settings for a sample of scan data and the associated quality metrics are fed-back into the learning system for additional refinement of the learned parameters (e.g., in deep supervised and/or reinforcement learning).
  • a metric measuring the similarity of one rendered image to one or a group of rendered images with the desired quality is calculated, and used as the “reward” to train the artificial intelligent agent using the deep reinforcement learning technique.
  • quality metrics than similarity may be used.
  • the choice of metric to use may depend on the types of features relevant to a specific clinical data domain. Depending on the diagnosis of interest or medical workflow, different pathologies may be of interest. Different rendering settings increase the conspicuity of different pathologies.
  • the quality metric in the reinforcement learning is used to learn to provide the optimal rendering parameters for increasing the conspicuity of the pathology of interest.
  • the artificial intelligent agent is a deep neural network that takes in the output of the reinforcement agent as its input, and outputs one quality or similarity metric (referred to as “action-value”) for each possible action (e.g., set of rendering settings), indicating the effect of taking these actions (i.e., higher values are associated with actions that can improve the rendering quality).
  • action-value quality or similarity metric
  • the agent also analyzes the action-values, and selects one action to perform.
  • the set of rendering parameters providing the highest implementable quality rendering are selected as the output.
  • a probability distribution of different values of the metric is used for the selection.
  • the selected action is the action with highest action-value or randomly selected from the actions with a probability distribution assigned based on the rank of action-values. In a typical scenario, actions with higher action-values are assigned higher probability, but any assignment may be used.
  • the training is modeled after steps taken by an expert or other user to achieve the desired visualization for the training data.
  • the user's sequence of adjustments to provide the desired rendering from the scan dataset is monitored (e.g., select transfer function T 1 , then window function W 1 , then select a different transfer function T 2 , then select a material reflectance MR 1 , . . . ).
  • the learning uses metrics for each stage or step to learn the settings used for each.
  • the actions of the expert are extrapolated to different training datasets or included for each training dataset.
  • the training learns to provide final rendering settings based on the monitored sequence.
  • a photograph or video of the imaged tissue, imaged object, or other real-world references is used.
  • the physically-based rendering provides photorealistic images.
  • the training learns to provide rendering settings modeled after real images.
  • the desired output is labeled using actual optical images so that the rendering parameters result in a matching look and/or feel.
  • Other modes of images may be used for the ground truth or reinforcement metric, so that the training learns to output rendering settings modeled after the different modality (e.g., render CT data to match the look of rendered MR data).
  • one model is trained to output settings for multiple rendering parameters.
  • a hierarchy or other grouping of models are trained to output the settings for multiple rendering parameters.
  • the machine is trained to output settings of rendering parameters. By outputting settings for two or more rendering parameters for a given scan dataset, less manual setting of parameters is needed. In path tracing or other physically-based rendering, the amount of processing and time required to provide the photorealistic imaging may cause delays. Compounding such delays with user adjustment, it may be overly time consuming to provide a desired rendering manually.
  • the machine trained model contributes to less delay in achieving the desired rendered image.
  • the model may be trained to output any number and/or type of rendering parameters.
  • one or more data consistency parameters are to be output.
  • Data consistency parameters include windowing, scaling, level compression, data normalization, or others.
  • one or more transfer function design parameters are to be output. Transfer function design parameters include classification look-up tables, multi-dimensional transfer functions, tissue-specific transfer functions, or other transfer functions.
  • one or more lighting design parameters are to be output. Lighting design parameters include type of virtual lights, position of the virtual light sources, orientation of the virtual light sources, image-based lighting sources, or others.
  • one or more viewing design parameters are to be output. Viewing design parameters include type of camera, position of the camera, orientation of the camera, intrinsic parameters for viewing, or others.
  • one or more use-case specific parameters are to be output. Use-case specific parameters are settings specific to a given use, such as a particular camera position for a given type of medical report or use of two cameras for stereoscopic viewing.
  • the parameters may be specific to physically-based rendering, such as internal render properties.
  • the setting for any parameter for path tracing, unbiased path tracing, Monte Carlo rendering, global illumination, or other simulations of light propagation may be learned.
  • a global albedo, or surface reflectance coefficient may be learned for the volume data, which describes the relative importance of light scattering versus light absorption.
  • Settings for any parameters for material property controlling the interaction with light may be learned.
  • the transparency, reflectivity, surface roughness, and/or other properties may be modeled in the physically-based rendering, so included as rendering parameters.
  • Albedo may also be learned as a per-material property for the different tissue classes in the volume data. Since the effect of various settings of material type or light interaction may not be instinctive to a user, training the model with machine learning to provide the initial settings may assist the user.
  • All or a sub-set of the rendering parameters are to be output by the machine-learnt model.
  • the model is trained to output any number of data consistency, transfer function, lighting, and viewing parameters.
  • the model is trained to provide settings for all of the rendering parameters.
  • the model is trained to provide settings for some but not all of the rendering parameters. For example, all of the rendering parameters are output except for the view design parameters, which are under the control of the user. The viewing camera parameters are under the control of the user, so the machine learning is for settings of other parameters. The user controls the view design parameters independently.
  • the model is trained to output settings for multiple parameters regardless of the user-controlled parameters.
  • the user-selected parameters are an input to the classification so that the output settings account for the user-selected settings.
  • the model learns to output rendering parameter settings based, in part, on settings for one or more parameters under the control of the user.
  • the artificial intelligence is trained to output the values for the rendering parameters based on user selected viewing camera parameters. Subsets of the rendering parameters are used as additional inputs in the training phase of the machine learning algorithm. In that case, the machine learning algorithm output is an optimized subset of the rendering parameters.
  • the user input settings may be optimized or not while the other settings are optimized by the model.
  • the input to the training is a higher dimension vector including the input scan dataset and reference viewing parameters.
  • the artificial intelligence is trained to account for different ambient conditions in the real-world viewing environment of the user.
  • the settings for “ambient-varying” rendering of a volume are learned.
  • the training data includes different ambient light conditions and measures of the ambient light using light probes.
  • the model is trained to account for ambient luminosity of the room. The luminosity is sent to the trained model, which adapts the rendering parameters automatically to provide the same content-based realistic rendering.
  • the training may be repeated.
  • the user may adjust some of the output settings. This adjustment, the setting, and/or image resulting from the adjustment are added to the training data.
  • the machine learning model is refined by repeating the learning with the addition feedback as samples from which to learn.
  • the manually adjusted settings of the parameters define personalized visualization designs. Other sources of further training data may be used.
  • the repetition of the training may be performed separately for different users, physicians, medical groups, hospitals, or other grouping. Personalized training is used so that consistent images for a relevant group are provided. In other embodiments, the repetition is provided for all users.
  • the trained artificial intelligence is stored.
  • the result of the training is a matrix.
  • the matrix represents the learned knowledge through machine training.
  • the matrix includes an input vector and outputs for the settings of the rendering parameters.
  • Other machine-learnt model representations may be used, such as a hierarchy of matrices or other non-linear models.
  • the trained model is stored in a memory. Any memory may be used.
  • the memory used for the training data may be used.
  • the memory may be in other devices.
  • the trained model is stored in a memory of a server.
  • the server uses the trained model to output rendering parameters to clients or to a server-based renderer which outputs rendered images to the clients.
  • multiple copies of the trained model are provided to different physicians, medical scanners, and/or workstations for use by different physicians.
  • FIG. 2 shows one embodiment of a method for content-based rendering based on machine learning in a rendering system.
  • the machine-learnt model or artificial intelligence agent is applied.
  • a rendering of a scan dataset is to be performed.
  • the artificial intelligence agent is applied to the scan dataset with or without other input information, outputting values for the rendering parameters.
  • the values result in consistency in photorealistic rendering, possibly avoiding time consuming manual adjustment of various rendering parameters to achieve the desired visual.
  • the method is implemented by a computer, server, or other processor with a graphics processing unit or other renderer and a display.
  • medical data is loaded from memory.
  • a light sensor measures the ambient light or captures a light probe image, and a computer applies the machine-learnt model, outputting values for rendering.
  • a renderer uses the values to render from the medical data, resulting in transmission of the rendered image to a display.
  • Different devices may be used, such as a controller or medical scanner processor performing the application and/or rendering.
  • acts may be provided.
  • the ambient light is not measured in act 32 .
  • the rendering of act 42 and transmission of the photorealistic image of act 44 are not provided, but instead the values of the rendering parameters are stored or transmitted.
  • acts for manual adjustment and/or initial setting of one or more rendering parameters are provided, such as for an interactive rendering workflow.
  • none, one or two of acts 36 - 40 are performed without others of that set of acts.
  • the acts are performed in the order shown (top to bottom) or other orders.
  • the ambient light may be measured in act 32 prior to loading the medical data in act 30 and/or in parallel with performing act 34 .
  • medical data is loaded from a memory, sensors, and/or other source.
  • a medical scanner may provide the data, such as a medical dataset representing a 3D region of the patient. Mining of a patient's computerized medical record may be used. Medical information from a picture archiving and communications server may be obtained. Any source may be used.
  • the medical data is for a given patient. While the training may use samples from many patients to learn features and/or learn to classify input medical data to provide values for rendering, the learnt model is applied to the medical data for a patient to output values for rendering parameters for that patient. The same or different machine-learnt models are applied to data for any number of patients.
  • Scan data representing a 3D volume is loaded as a medical dataset.
  • the scan data may be from multiple two-dimensional scans or may be formatted from a 3D scan.
  • the medical data may include non-image or non-scan data, such as patient test results, physician notes, lab results, diagnosis information, billing codes, clinical data, patient history, or other information.
  • the medical data includes the information learned to be used by the machine-trained model. Ground truth information and/or similarity or quality metric measures may be not provided. Where the machine learning learned to use feedback of predicted ground truth, then that feedback may be created and used as an input.
  • the content to be used to determine the settings of the rendering parameters is provided to the machine-learnt model. For example, just the 3D scan data is provided as a medical dataset for a given patient. As another example, the 3D scan data and diagnostic information are provided as the medical dataset for the patient. The diagnostic information may be used to select the machine-learnt model to use and/or be part of the input feature vector.
  • Some machine-learnt models may be robust or still operate with some of the input feature vector missing. Alternatively, all of the information needed for the input feature vector is loaded.
  • a sensor measures ambient light. Any luminosity sensor may be used, such as a charge coupled device. The sensor provides a measure of the luminosity. This measure may be used as an input in the input feature vector and/or to select a particular machine-learnt model to use. In alternative embodiments, the rendering is not based on ambient lighting conditions, so no measurement of ambient light is provided.
  • a light probe image is obtained as the measure of ambient light with a camera as the sensor.
  • the image is obtained using any of a variety of ways and then used for high dynamic range image-based illumination during rendering. Examples ways for obtaining light probe images include capturing high-dynamic range photographs of a mirrored ball, stitching multiple photographs, or using self-contained panoramic cameras.
  • the medical dataset is applied to the machine-learnt model.
  • the loaded medical dataset is processed by a machine, such as a computer, processor, or server.
  • the machine uses the internal representation of the machine-learnt model.
  • the input feature vector is created and/or is the medical dataset.
  • the machine inputs the feature vector, resulting in the machine-learnt model outputting the rendering settings based on the learned knowledge.
  • Acts 36 - 40 represent different acts in applying the machine-learnt model of presets for one or more rendering parameters. Additional, different, or fewer acts may be provided.
  • the machine-learned model was previously learned using any approach for machine learning, so may be a regression, classification, or reinforcement-learnt model.
  • the machine-learned model is trained with deep learning (e.g., deep neural network) to extract features from the medical dataset.
  • deep learning e.g., deep neural network
  • features are extracted from the medical dataset.
  • One or more other features may be extracted from the medical dataset using manually or user configured features.
  • These features may then be used by another machine-learnt model and/or another hierarchal stage to output the values for the rendering. For example, a further reinforcement-learnt model is applied.
  • the extracted features are input to the machine-learnt model for application.
  • the application of the machine-learnt model is more likely to output values in act 38 resulting in the photorealistic image corresponding to a standard image despite differences in the medical dataset.
  • the photorealistic image is created using path tracing, other physically-based rendering, or reference photographs.
  • the model uses the input feature vector to predict the values for rendering that result in the rendering of the scan data as an image with desired characteristics.
  • the standard image may be a group of images having similar characteristics, such as resolution and/or contrast for a particular type of tissue or lesion.
  • the rendering values output were learnt in order to provide the same or similar resolution and/or contrast for that particular type of tissue.
  • Different machine-learnt models may be used for different diagnostic and/or workflow situations, such as outputting values for breast cancer imaging or for fetal imaging.
  • the ambient light measured in act 32 is applied to the machine-learnt model with or as part of the medical dataset.
  • the machine-learnt model is trained to output the values based in part on the ambient light.
  • the machine-learnt model outputs values for the rendering parameters that account for ambient light.
  • Photorealistic images may be output as if lit by the ambient light in a room or to which the patient is subjected. Where the photorealistic images are used for augmented reality, use of ambient light and consistency in imaging accounting for ambient light may result in better blending or greater visual appeal.
  • the machine-learnt model is trained to output any number of values for a respective number of rendering parameters in act 40 .
  • values for two or more physically-based rendering parameters are output based on the input medical dataset.
  • Three or more, all, or a sub-set of any number of rendering parameters may have their values output by application to the machine-trained model or group of models.
  • Example rendering parameters include controls for consistent data handling, transfer function, lighting design, viewing design, material propriety, or internal renderer parameters.
  • One or more of the output values are for non-transfer function parameters.
  • the machine-learnt model or set of models output values for all of the rendering parameters except for viewing design.
  • One or more parameters for viewing design are left to the user, such as for user navigation or setting of the camera position, orientation, and/or scale.
  • the user inputs the values for all or some of the viewing design parameters, and the application of the medical dataset to the machine-learnt model provides the values for all of the other rendering parameters. Additional, different, or fewer types of rendering parameters may use initial input by the user.
  • the user may adjust model output values. Since physically-based rendering may be more time consuming than traditional rendering, providing a comprehensive set of values for the rendering parameters more likely to result in the desired image initially may avoid time-consuming adjustments. The user may, however, make any number of adjustments. The number and/or amount of change may be less due to the assistance by the machine-learnt model.
  • a physically-based renderer renders the photorealistic image of the 3D region of the patient using the values output from the application.
  • the output values of the machine-learnt model are used as settings by the physically-based renderer. For example, a sub-set or all of the rendering parameters for rendering with unbiased path tracing have values established by the output of the machine-learnt model.
  • the physically-based renderer is a graphics processing unit, array of graphics processing units, workstation, server, computer, or other machine capable of the required computations. Any physically-based rendering algorithm capable of computing the light transport may be used, such as path tracing, unbiased path tracing, Monte-Carlo path tracing, or Metropolis light transport.
  • the physically-based rendering simulates the physics of light propagation to create an image instead of accumulating volume samples with local illumination along viewing rays from the camera, as is done for traditional volume rendering.
  • path tracing implemented for physically-based rendering, the paths of light or photons, including due to scattering and absorption, are modeled and used for rendering.
  • the physically-based rendering result may be built up over time as the rendering may rely on probabilistic scattering and tracing millions of light paths.
  • the rendering results in a photorealistic image A sequence of images may be provided as the image is built or rendered. Alternatively, for a given set of values of rendering parameters, a single image is output. In either case, the photorealistic image is rendered using presets from, at least in part, the machine-learnt model. A sequence of images due to alteration of values for one or more rendering parameters may be output. The user or renderer may change the values, such as rotation of a camera or for modeling fly-through.
  • the renderer, host computer of the renderer, or other device transmits the photorealistic image or images.
  • the transmission is over a bus, through a network interface, over a network, to a memory, or to a display.
  • the image is loaded into a buffer and output from the buffer to a display for viewing by a physician to aid diagnosis or pre-operative planning.
  • a computer-generated representation of the medical data suitable for viewing by human users e.g., 2D images, videos, stereoscopic image pairs, volumes, holograms, or lightfields
  • the output may be for storage and/or for further machine analysis.
  • the transmission may be for any one or more of various purposes.
  • the transmission of the photorealistic image is part of a diagnostic report.
  • a given diagnostic report may be specific to context, such as a report for kidney stones. Any kidney stones should be visible in an image.
  • the photorealistic image shows kidney stones based on the values of the rendering parameters output by the model trained to create such images.
  • the optimal context-specific images are provided for an automated diagnostic report in the clinical workflow.
  • the transmission is to a display as an initial image for interactive viewing.
  • the machine-learnt model is used to initialize interactive viewing applications in the clinical workflow. By providing values for rendering parameters so that the images are consistent with an established visualization design, the amount of user interaction may be reduced.
  • the transmission is to an augmented reality device.
  • a physician may wear augmented reality glasses.
  • the photorealistic image with or without added annotations or graphics, is projected onto the glasses as an overlay on the actual view seen by the wearer.
  • the photorealistic image is combined with a camera image and/or is used as a replacement for the camera image.
  • the rendered image may be overlaid onto a live stream of real-time medical images (ultrasound, x-ray, optical laparoscopy, etc.).
  • FIG. 3 shows a block diagram of one embodiment of a system for content-based rendering based on machine learning.
  • the system is for training with machine learning and/or application of a machine-learnt model.
  • a machine-learnt model is trained to output settings for rendering parameters given a new volume for a patient.
  • the system implements the method of FIGS. 1 and/or 2 .
  • Other methods or acts may be implemented, such as providing a user input (e.g., mouse, trackball, touch pad, and/or keyboard) and user interface for interactive rendering.
  • a user input e.g., mouse, trackball, touch pad, and/or keyboard
  • user interface for interactive rendering.
  • the system includes a machine 50 , a memory 52 , a display 54 , a medical imager 56 , and a renderer 58 . Additional, different, or fewer components may be provided. For example, the medical imager 56 and/or memory 52 are not provided. In another example, a network or network connection is provided, such as for networking with a medical imaging network or data archival system. A user interface may be provided for interacting with the machine 50 , renderer 58 , or other components.
  • the machine 50 , memory 52 , renderer 58 , and/or display 54 are part of the medical imager 56 .
  • the machine 50 , memory 52 , renderer 58 , and/or display 54 are part of a server, workstation, or computer separate from the medical imager 56 .
  • the machine 50 , memory 52 , renderer 58 , and/or display 54 are a personal computer, such as desktop or laptop, a workstation, a server, a network, or combinations thereof.
  • the machine 50 and memory 52 are part of a separate computer from the renderer 58 .
  • the medical imager 56 is a medical diagnostic imaging system. Ultrasound, CT, x-ray, fluoroscopy, positron emission tomography (PET), single photon emission computed tomography (SPECT), and/or MR systems may be used.
  • the medical imager 56 may include a transmitter and includes a detector for scanning or receiving data representative of the interior of the patient.
  • the medical imager 56 acquires scan data representing the patient.
  • the scan data may represent a volume of the patient. For example, scan data is acquired and used for diagnosis or surgical planning, such as identifying a lesion or treatment location.
  • the medical imager 56 is not provided, but a previously acquired dataset for a patient is stored in the memory 52 .
  • many medical images 56 are provided in the memory 52 as the training data, which is gathered and stored in the memory 52 .
  • the machine 50 is a computer, workstation, server, processor, or other device configured to apply machine learning and/or to apply a machine-learnt model.
  • the machine 50 is configured by software, hardware, and/or firmware.
  • the machine 50 is configured by one or more machine learning algorithms.
  • the machine 50 is configured, in part, by a learnt matrix or matrices associating input data to output values.
  • the machine 50 applies a machine-learnt model, such as one learnt with deep learning.
  • the machine-learnt model as implemented by the machine 50 , generates input features and/or outputs settings for rendering parameters based on application of medical data from a scan of a patient.
  • the rendering parameters include windowing, transfer function, lighting, and/or material properties.
  • the settings are learnt to provide an image similar in look and/or feel to one or more images from a same diagnostic context. Consistent images, such as to highlight or show anatomy of interest, are created.
  • the diagnostic context provides the standard or information of interest.
  • the memory 52 is a graphics processing memory, a video random access memory, a random access memory, system memory, cache memory, hard drive, optical media, magnetic media, flash drive, buffer, database, combinations thereof, or other now known or later developed memory device for storing training data, rendering parameters values, medical datasets, and/or other information.
  • the memory 52 is part of the medical imager 56 , part of a computer associated with the machine 50 , part of a computer associated with the renderer 58 , a database, part of another system, a picture archival memory, or a standalone device.
  • the memory 52 stores scan data representing one or more patients. For example, data from the medical imager 56 is stored. The data is in a scan format or reconstructed to a volume or three-dimensional grid format. The scan data is stored for training and/or stored for rendering an image of a patient.
  • the memory 52 or other memory is alternatively or additionally a non-transitory computer readable storage medium storing data representing instructions executable by the programmed machine 50 for learning or applying the machine-learnt model.
  • the instructions for implementing the processes, methods, and/or techniques discussed herein are provided on non-transitory computer-readable storage media or memories, such as a cache, buffer, RAM, removable media, hard drive, or other computer readable storage media.
  • Non-transitory computer readable storage media include various types of volatile and nonvolatile storage media.
  • the functions, acts or tasks illustrated in the figures or described herein are executed in response to one or more sets of instructions stored in or on computer readable storage media.
  • processing strategies may include multiprocessing, multitasking, parallel processing, and the like.
  • the instructions are stored on a removable media device for reading by local or remote systems.
  • the instructions are stored in a remote location for transfer through a computer network or over telephone lines.
  • the instructions are stored within a given computer, CPU, GPU, or system.
  • the renderer 58 is a graphics processing unit, graphics card, graphic chip, multi-core processor, or other processor for running a software package that implements the photorealistic image generation functionality.
  • the renderer 58 is configured by an application programming interface to render an image from the 3D scan data representing a patient. Using physically-based rendering, a photorealistic image is rendered. Using the settings output by the machine-learnt model, path tracing is applied to render the image. For feedback in training, the renderer 58 may render training images. Similarity is calculated by the renderer 58 and/or the machine 50 .
  • the display 54 is a monitor, LCD, projector, plasma display, CRT, printer, or other now known or later developed device for displaying the photorealistic image or images.
  • the display 54 receives images from the machine 50 , memory 52 , renderer 58 , or medical imager 56 .
  • the images of the tissue captured by the medical imager 56 are displayed.
  • Other information may be displayed as well, such as generated graphics, text, or quantities as a virtual overlay.
  • Additional images may be displayed.
  • scan data represents a sequence of scans over time
  • a corresponding sequence of images may be generated.
  • new images may be generated as settings for one or more rendering parameters are changed by a user.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Graphics (AREA)
  • General Engineering & Computer Science (AREA)
  • Public Health (AREA)
  • Biomedical Technology (AREA)
  • Databases & Information Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Primary Health Care (AREA)
  • Epidemiology (AREA)
  • Mathematical Physics (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Pathology (AREA)
  • Image Analysis (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Medical Treatment And Welfare Office Work (AREA)

Abstract

An artificial intelligence agent is machine trained and used to provide physically-based rendering settings. By using deep learning and/or other machine training, settings of multiple rendering parameters may be provided for consistent imaging even in physically-based rendering.

Description

    RELATED APPLICATIONS
  • The present patent document claims the benefit of the filing date under 35 U.S.C. §119(e) of Provisional U.S. Patent Application Ser. No. 62/306,407, filed Mar. 10, 2016, which is hereby incorporated by reference.
  • BACKGROUND
  • The present embodiments relate to medical imaging of three-dimensional (3D) scans. Data representing a volume is rendered for visualization. Due to the many different scan settings and patient variability, renderings for different patients or at different times appear different.
  • In an effort to reduce variability, most existing medical rendering provides a set of static visualization presets for specific workflows or diagnostic contexts. Even with static presets, the rendering may require manual adjustment and may not provide consistent rendered results. Semi-automatic machine learning approaches have been used to provide transfer function design in traditional volume rendering (e.g., projection or alpha blending), but may still not sufficiently reduce variability.
  • Variability is particularly troublesome for physically-based volume rendering, which relies on the physical simulation of light propagation (e.g., unbiased path tracing). Physically-based visualization techniques produce global illumination effects in computer-generated graphics that mimic the real-world interaction of light with various 3D objects or tissues. This results in physically plausible images that are often easier for the human brain to interpret when compared to the more analytical images from traditional rendering. For example, subtle ambient occlusions, shadows, and color bleeding provide important depth cues for the spatial understanding of 3D relationships between structures in a single 2D image, whereas simpler visualization techniques may require additional interaction with the viewing parameters (e.g., moving the virtual camera around the 3D data) to obtain the same spatial information from the image parallax.
  • These physically-based visualization techniques are more sensitive to changes in the classification or the interpretation of the underlying medical data. As a result, small changes to the rendering parameters may have a more pronounced impact on the perception and interpretation of 3D structures in the final image as compared to the traditional volume rendering techniques based on compositing of classified voxels along viewing rays. Obtaining very high quality reproducible images with diagnostic significance is then partially at the user's discretion. While existing techniques for providing visualization presets can help, the workflow is not fully automated and the resulting rendered images may not have consistent quantitative properties (color, hue, reflectance, etc.) across datasets.
  • BRIEF SUMMARY
  • By way of introduction, the preferred embodiments described below include methods, systems, instructions, and computer readable media for machine training an artificial intelligence and use of that artificial intelligence to provide rendering settings. By using deep learning and/or other machine training, settings of multiple rendering parameters may be provided for consistent imaging or visualization even in physically-based rendering.
  • In a first aspect, a method is provided for content-based rendering based on machine learning in a rendering system. A medical dataset representing a three-dimensional region of a patient is loaded from memory. A machine applies the medical dataset to a machine-learnt non-linear model. The machine-learned model is trained with deep learning to extract features from the medical dataset and trained to output values for two or more physically-based rendering parameters based on input of the medical dataset. At least one of the two or more physically-based rendering parameters are controls for consistent data handling, lighting design, viewing design, material propriety, or internal renderer property. A physically-based renderer renders a photorealistic image of the three-dimensional region of the patient using the output values resulting from the applying. The photorealistic image is transmitted.
  • In a second aspect, a method is provided for machine training for content-based rendering in a machine training system. First volume data of a volume of a patient, a first image of the volume, and first values of path tracing rendering parameters are input to training of an artificial intelligence. A machine trains the artificial intelligence to output second values of the path tracing rendering parameters for a second volume data where the second values provide a second rendered image of the second volume modeled on the first image. The trained artificial intelligence is stored.
  • In a third aspect, a system is provided for content-based rendering based on machine learning. A medical scanner is configured to scan a patient. A machine is configured to output settings for rendering parameters by application of data from the scan to a machine-learnt model. The rendering parameters include windowing, transfer function, and lighting, and the settings are learned to provide a first image from the data similar to one or more second images for a same diagnostic context. A graphics processing unit is configured to render the first image using the settings.
  • The present invention is defined by the following claims, and nothing in this section should be taken as a limitation on those claims. Further aspects and advantages of the invention are discussed below in conjunction with the preferred embodiments and may be later claimed independently or in combination.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The components and the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the invention. Moreover, in the figures, like reference numerals designate corresponding parts throughout the different views.
  • FIG. 1 is a flow chart diagram of one embodiment of a method for machine training for content-based rendering;
  • FIG. 2 is a flow chart diagram of an embodiment of a method for content-based rendering based on machine learning in a rendering system; and
  • FIG. 3 is a block diagram of one embodiment of a system for content-based rendering based on machine learning.
  • DETAILED DESCRIPTION OF THE DRAWINGS AND PRESENTLY PREFERRED EMBODIMENTS
  • Content-based photorealistic rendering of medical images is based on machine learning. To produce consistent rendered images from 3D medical scans, particularly when using physically-based rendering techniques, the machine learnt non-linear model determines rendering settings based on the content of the 3D medical scan. Due to the variability between scan parameters, data contrast, noise, and/or the large number of rendering parameters, producing rendered images that allow for consistent visual interpretation has been a long standing challenge, which has prevented wider adoption 3D images for definitive decision support.
  • In one embodiment, deep learning generates the settings for the rendering parameters for the photorealistic rendering components, establishing a consistent visualization design. This design may be applied to new 3D image data in order to provide reliable visual interpretation within the diagnostic workflow. Image features, and optionally, available non-image data are mapped to sets of rendering parameters that produce optimal images, where the relationship between the two is highly non-linear in the general case. By targeting the entire or multiple parts of the rendering pipeline, a scalable solution leverages machine learning and supports consistent photorealistic rendering.
  • Consistent visual interpretation for 3D images may potentially simplify diagnostic workflow and may promote the clinical use of 3D data for decision support. Consistent photorealistic rendered images may be provided in new platforms and workflows, such as fully automated reporting and cloud-based rendering services. Consistent visualization design as a machine learning problem may be applied 1) to optimize rendering preset designs tailored to a large variety of data types in a unified and scalable way, 2) to adapt rendering parameters to produce target consistent results based on the provided input, 3) to adapt rendering parameters to specific individual preferences, 4) to transfer expert trained results to non-expert users, 5) to increase the conspicuity of visual features in the rendered images (e.g., pathologies in medical images) and 6) to provide automatic rendering presets for specific workflows. A scalable framework is provided for fully automated image generation in the clinical workflow.
  • FIG. 1 shows one embodiment of a method for machine training for content-based rendering in a machine training system. Optimal rendering parameters may be obtained via training an artificial intelligence system. The input of the artificial intelligence system may include, but is not limited to, the 3D scan data to be rendered, images rendered from the 3D scan data, information on the patient, and/or information on the diagnostic purpose. Using many examples, the machine training learns to provide rendering settings for photorealistic rendering based on input data for a specific patient. The output of the trained artificial intelligence system is the optimal set of settings for the rendering parameters.
  • The method of FIG. 1 is implemented by a processor and a graphics processing unit in a server, computer, or other machine. In one embodiment, the system of FIG. 3 implements the method of FIG. 1, with or without the medical imager 56 and/or display 54. For example, data is input by loading from a memory, a processor performs the training based on the data, and the memory stores the learnt artificial intelligence. As another example, the medical scanner and/or memory input the data to a server, which performs the training and outputs to a memory.
  • The acts are performed in the order shown (top to bottom) or another order. For some forms of learning, results may be fed back as input in a repetition of the acts.
  • Additional, different, or fewer acts may be provided. For example, act 22 is not provided, such as where the trained artificial intelligence is used rather than stored for later use.
  • In act 12, training data is input. The training data is gathered from one or more memories, sensors, and/or scanners. The gathered data may be stored and/or transmitted to a buffer, memory, cache, processor, or other device for training. This gathered training data is input to the machine learning algorithm.
  • Many samples of the same type of data are input. To learn statistical or other relationships that may be sufficiently complex that a human cannot perceive the relationships in a same level, tens, hundreds, or thousands of samples are provided.
  • Any data may be used. Since the artificial intelligence is to be trained to provide settings for rendering parameters, scan data is provided. For each sample, the scan data represents a patient. A two-dimensional (2D), 3D, 2D+time sequence, 3D+time sequence, and/or other image or scan data may be used.
  • Other information than scan data may be provided for each sample. Patient non-image data may be input to the training. Higher dimensional data includes but is not limited to patient information such as a) natural language medical documents or information extracted therefrom of previous diagnosis and/or clinical procedures; b) previous medical test results; c) previously processed medical images and measurements; d) computational modeling results; e) parametric image information derived from medical image analysis algorithms (e.g. cardiac strain map or elasticity); and/or f) other information (e.g., scanner settings).
  • For training, the goal or ground truth information is input. The artificial intelligence system is trained with labeled data (e.g., pairs of input and corresponding desired output (e.g., rendering parameters, optimal rendered images, and/or photographs/movies of the real object)). Based on the provided data, the artificial intelligence system is trained to produce the desired output from any input. The ground truth is provided by desired settings of the rendering parameters for each scan dataset. The values of the path tracing rendering parameters for desired results of each sample are input to the training. Alternatively or additionally, the ground truth is provided as a desired rendered image for each scan dataset. A rendered image for each scan dataset is provided. In other embodiments, an image with desired visualization is provided for multiple datasets. Using a renderer, the machine training may learn the settings that produce similar visualization.
  • The input samples may provide both negative and positive training samples. For example, scan datasets and settings providing desired and undesired images are input. The training learns to distinguish the desired settings from both good and bad settings.
  • In one embodiment, additional samples are created by perturbing the path tracing rendering parameters of an input sample. The perturbing creates a collection of sets of the path tracing rendering parameters for each scan data set. The artificial intelligence system is applied on the input scan dataset to generate rendering parameters, which are then perturbed to generate a pool of rendering parameters. Rendering is then performed using the sets of rendering parameters in the pool to produce a pool of corresponding images. Human users or other operators then visually examine the produced images and select the satisfying images. The rendering parameters for the selected images are then used as desired output and/or a sub-set as positive ground-truth. The rendering parameters for the unselected sub-set of images are not used or are used as negative ground-truth.
  • In another approach to gather samples for input, a web-based or network-based service is provided. A website or network portal allows a user to upload the input data (i.e., image and additional information). The server or other processor then runs the artificial intelligence system and/or perturbs possible rendering settings to renderer a pool of images. The service then provides the generated images to the user and asks the user to select the satisfying images. The rendering parameters corresponding to the user selected images are then stored as the desired output for the input (i.e., ground truth). By providing such a service, additional samples for training may be gathered over time.
  • The input may be repeated. The training of the artificial intelligence system is performed periodically or multiple different times. After additional samples are collected (e.g., through the web-based service), the training is repeated using the additional samples with or without previous samples used for earlier iterations of the training.
  • In act 14, the artificial intelligence or intelligences are machine trained. The collected samples, including ground truth, are used to train. A machine, such as a processor, computer, server, or other device learns from the samples. Using machine-learning, complex statistical relationships between large numbers (e.g., tens, hundreds, thousands, or more) of input variables to any number of output variables are extracted from the large number of samples.
  • The training is for imaging in any context. One model may be learned for any number of imaging situations. Alternatively, different models are trained for different situations. The different situations may include different scan modalities (e.g., different model for computed tomography, magnetic resonance, ultrasound, positron emission tomography, and single photon emission computed tomography). The different situations may include different types of tissue of interest (e.g., liver versus kidney), different diagnostic purpose or workflow (e.g., cancerous lesion versus bone calcification), and/or different users (e.g., different operators may have different preferences for visualization).
  • Any now known or later developed machine learning may be used. Regression, classification, and/or reinforcement learning are used. Regression training learns a range or continuous output by minimization of a metric. Classification learns disparate outputs. Reinforcement learning learns through a sequence of actions with feedback. Neural network, Bayes network, or support vector machine training may be used. Hierarchal or other approaches may be used. Supervised or semi-supervised machine learning may be used.
  • To train, features are extracted from the input data. Haar wavelet, steerable, or other features may be extracted from the scan dataset or images. Measures or other information may be extracted from non-image data. Alternatively, the input data itself (e.g., scalar values of the scan dataset) is used.
  • In an alternative represented in act 16, the machine learns features as part of the training. For example, deep learning (e.g., deep structured learning, hierarchical learning, or deep machine learning) models high-level abstractions in data by using multiple processing layers with structures composed of multiple non-linear transformations, where the input data features are not engineered explicitly. A deep neural network processes the input via multiple layers of feature extraction to produce features used to derive outputs of settings for rendering. The deep learning provides the features used by other machine training to learn to output the rendering settings. Other deep learnt, sparse auto-encoding models may be trained and applied. The machine training is unsupervised in learning the features to use and how to classify given an input sample (i.e., feature vector).
  • Using deep learning, the input features required for the machine learning are not engineered by the user explicitly, as in shallow learning algorithms. This is especially useful with render settings, which may not have an explicit natural relationship with the input data. Deep learning figures out the data features that end up providing a good classifier or model. Deep learning is known to be effective in extracting high level abstractions (features) from low level input data (e.g., raw images) that is suitable for the trained task, so may be more convenient than and superior to hand-crafted features for identifying features for render settings. In addition, learning-based automatically rendering setting selection is a less studied problem, with no hand-craft feature specifically designed for this problem. Standard hand-craft features, like HAAR feature or SIFT feature, designed for other general computer vision tasks may not be suitable for rendering setting.
  • In act 18, the machine trains to output rendering settings to provide consistent imaging. To reduce variability, the training learns to provide rendering settings that result in a rendered image with similar or same visualization despite differences in the content of the data. The path tracing or other physically-based rendering settings so that the rendered image for a given input scan dataset looks and feels similar to an ideal are learned. The machine learns to output renderer settings to model the resulting rendered image for a given scan dataset on a desired image. The learned sets of parameters establish a visualization design that may be reused when new data is introduced to the system in order to provide consistent rendered images. In this way, the machine may learn renderer settings to provide rendering to a standard. Rather than render with the same preset values for a given context, the values are based on a given input dataset. In various embodiments of the system, consistency refers to consistent data handling, consistent viewing, and/or consistent visual styles, which may be utilized both in automated and interactive clinical workflows.
  • The ground truth provides consistency. Where the ground truth uses similar or consistent images and/or settings that result in similar or consistent images despite variance in the scan datasets, the training learns to output settings providing the consistency. The notion of optimal rendering parameters and optimal output images relates to using objective quality measurements, subjective quality measurements, clinical significance ratings, and/or other indicators of consistency.
  • In one embodiment, the consistency is learned using a measured or calculated metric. The training uses a metric of similarity. Any metric of similarity may be used. Example computed metrics may be visual perception metrics based on visual systems, comparative differences, perceptual hash, mutual information, correlation, computed feature signals, or simple error metrics. The metrics may be automatically computed or provided by the users. The metric is used as an indicator of strength of importance of a given training sample.
  • Alternatively or additionally, one or more metrics are used in reinforcement learning. For example, deep reinforcement learning uses a similarity of the rendered image from a sample to a reference image. The similarity is used as a reinforcement. The output image rendered from a given set of rendering settings for a sample of scan data and the associated quality metrics (i.e., similarity) are fed-back into the learning system for additional refinement of the learned parameters (e.g., in deep supervised and/or reinforcement learning). A metric measuring the similarity of one rendered image to one or a group of rendered images with the desired quality is calculated, and used as the “reward” to train the artificial intelligent agent using the deep reinforcement learning technique.
  • Other quality metrics than similarity may be used. The choice of metric to use may depend on the types of features relevant to a specific clinical data domain. Depending on the diagnosis of interest or medical workflow, different pathologies may be of interest. Different rendering settings increase the conspicuity of different pathologies. The quality metric in the reinforcement learning is used to learn to provide the optimal rendering parameters for increasing the conspicuity of the pathology of interest.
  • In one embodiment, the artificial intelligent agent is a deep neural network that takes in the output of the reinforcement agent as its input, and outputs one quality or similarity metric (referred to as “action-value”) for each possible action (e.g., set of rendering settings), indicating the effect of taking these actions (i.e., higher values are associated with actions that can improve the rendering quality). The agent also analyzes the action-values, and selects one action to perform. The set of rendering parameters providing the highest implementable quality rendering are selected as the output.
  • In another embodiment, a probability distribution of different values of the metric (i.e., different similarities) is used for the selection. The selected action is the action with highest action-value or randomly selected from the actions with a probability distribution assigned based on the rank of action-values. In a typical scenario, actions with higher action-values are assigned higher probability, but any assignment may be used.
  • In other or additional reinforcement learning, the training is modeled after steps taken by an expert or other user to achieve the desired visualization for the training data. The user's sequence of adjustments to provide the desired rendering from the scan dataset is monitored (e.g., select transfer function T1, then window function W1, then select a different transfer function T2, then select a material reflectance MR1, . . . ). The learning uses metrics for each stage or step to learn the settings used for each. The actions of the expert are extrapolated to different training datasets or included for each training dataset. The training learns to provide final rendering settings based on the monitored sequence.
  • Rather than using a rendered image or multiple rendered images as the ground truth and/or for calculating similarity, a photograph or video of the imaged tissue, imaged object, or other real-world references is used. The physically-based rendering provides photorealistic images. By comparing or training to real photos or video, the training learns to provide rendering settings modeled after real images. The desired output is labeled using actual optical images so that the rendering parameters result in a matching look and/or feel. Other modes of images may be used for the ground truth or reinforcement metric, so that the training learns to output rendering settings modeled after the different modality (e.g., render CT data to match the look of rendered MR data).
  • In act 20, one model is trained to output settings for multiple rendering parameters. In other embodiments, a hierarchy or other grouping of models are trained to output the settings for multiple rendering parameters. Using the training data and ground truth, the machine is trained to output settings of rendering parameters. By outputting settings for two or more rendering parameters for a given scan dataset, less manual setting of parameters is needed. In path tracing or other physically-based rendering, the amount of processing and time required to provide the photorealistic imaging may cause delays. Compounding such delays with user adjustment, it may be overly time consuming to provide a desired rendering manually. By using machine learning to provide the initial settings for multiple parameters, the machine trained model contributes to less delay in achieving the desired rendered image.
  • The model may be trained to output any number and/or type of rendering parameters. For example, one or more data consistency parameters are to be output. Data consistency parameters include windowing, scaling, level compression, data normalization, or others. As another example, one or more transfer function design parameters are to be output. Transfer function design parameters include classification look-up tables, multi-dimensional transfer functions, tissue-specific transfer functions, or other transfer functions. In another example, one or more lighting design parameters are to be output. Lighting design parameters include type of virtual lights, position of the virtual light sources, orientation of the virtual light sources, image-based lighting sources, or others. In yet another example, one or more viewing design parameters are to be output. Viewing design parameters include type of camera, position of the camera, orientation of the camera, intrinsic parameters for viewing, or others. In other examples, one or more use-case specific parameters are to be output. Use-case specific parameters are settings specific to a given use, such as a particular camera position for a given type of medical report or use of two cameras for stereoscopic viewing.
  • The parameters may be specific to physically-based rendering, such as internal render properties. The setting for any parameter for path tracing, unbiased path tracing, Monte Carlo rendering, global illumination, or other simulations of light propagation may be learned. For example, a global albedo, or surface reflectance coefficient, may be learned for the volume data, which describes the relative importance of light scattering versus light absorption. Settings for any parameters for material property controlling the interaction with light may be learned. For example, the transparency, reflectivity, surface roughness, and/or other properties may be modeled in the physically-based rendering, so included as rendering parameters. Albedo may also be learned as a per-material property for the different tissue classes in the volume data. Since the effect of various settings of material type or light interaction may not be instinctive to a user, training the model with machine learning to provide the initial settings may assist the user.
  • All or a sub-set of the rendering parameters are to be output by the machine-learnt model. For example, the model is trained to output any number of data consistency, transfer function, lighting, and viewing parameters. For fully automatic, the model is trained to provide settings for all of the rendering parameters. For interactive or other semi-automatic operation, the model is trained to provide settings for some but not all of the rendering parameters. For example, all of the rendering parameters are output except for the view design parameters, which are under the control of the user. The viewing camera parameters are under the control of the user, so the machine learning is for settings of other parameters. The user controls the view design parameters independently.
  • The model is trained to output settings for multiple parameters regardless of the user-controlled parameters. Alternatively, the user-selected parameters are an input to the classification so that the output settings account for the user-selected settings. The model learns to output rendering parameter settings based, in part, on settings for one or more parameters under the control of the user. The artificial intelligence is trained to output the values for the rendering parameters based on user selected viewing camera parameters. Subsets of the rendering parameters are used as additional inputs in the training phase of the machine learning algorithm. In that case, the machine learning algorithm output is an optimized subset of the rendering parameters. The user input settings may be optimized or not while the other settings are optimized by the model. In an interactive viewing application, the input to the training is a higher dimension vector including the input scan dataset and reference viewing parameters.
  • In another embodiment, the artificial intelligence is trained to account for different ambient conditions in the real-world viewing environment of the user. The settings for “ambient-varying” rendering of a volume are learned. The training data includes different ambient light conditions and measures of the ambient light using light probes. The model is trained to account for ambient luminosity of the room. The luminosity is sent to the trained model, which adapts the rendering parameters automatically to provide the same content-based realistic rendering.
  • The training may be repeated. As the machine-trained model is used, the user may adjust some of the output settings. This adjustment, the setting, and/or image resulting from the adjustment are added to the training data. The machine learning model is refined by repeating the learning with the addition feedback as samples from which to learn. The manually adjusted settings of the parameters define personalized visualization designs. Other sources of further training data may be used.
  • The repetition of the training may be performed separately for different users, physicians, medical groups, hospitals, or other grouping. Personalized training is used so that consistent images for a relevant group are provided. In other embodiments, the repetition is provided for all users.
  • In act 22, the trained artificial intelligence is stored. The result of the training is a matrix. The matrix represents the learned knowledge through machine training. The matrix includes an input vector and outputs for the settings of the rendering parameters. Other machine-learnt model representations may be used, such as a hierarchy of matrices or other non-linear models.
  • The trained model is stored in a memory. Any memory may be used. The memory used for the training data may be used. For application, the memory may be in other devices. For example, the trained model is stored in a memory of a server. The server uses the trained model to output rendering parameters to clients or to a server-based renderer which outputs rendered images to the clients. As another example, multiple copies of the trained model are provided to different physicians, medical scanners, and/or workstations for use by different physicians.
  • FIG. 2 shows one embodiment of a method for content-based rendering based on machine learning in a rendering system. The machine-learnt model or artificial intelligence agent is applied. For a given patient, a rendering of a scan dataset is to be performed. The artificial intelligence agent is applied to the scan dataset with or without other input information, outputting values for the rendering parameters. The values result in consistency in photorealistic rendering, possibly avoiding time consuming manual adjustment of various rendering parameters to achieve the desired visual.
  • The method is implemented by a computer, server, or other processor with a graphics processing unit or other renderer and a display. For example, medical data is loaded from memory. A light sensor measures the ambient light or captures a light probe image, and a computer applies the machine-learnt model, outputting values for rendering. A renderer uses the values to render from the medical data, resulting in transmission of the rendered image to a display. Different devices may be used, such as a controller or medical scanner processor performing the application and/or rendering.
  • Additional, different, or fewer acts may be provided. For example, the ambient light is not measured in act 32. As another example, the rendering of act 42 and transmission of the photorealistic image of act 44 are not provided, but instead the values of the rendering parameters are stored or transmitted. In yet another example, acts for manual adjustment and/or initial setting of one or more rendering parameters are provided, such as for an interactive rendering workflow. In other examples, none, one or two of acts 36-40 are performed without others of that set of acts.
  • The acts are performed in the order shown (top to bottom) or other orders. For example, the ambient light may be measured in act 32 prior to loading the medical data in act 30 and/or in parallel with performing act 34.
  • In act 30, medical data is loaded from a memory, sensors, and/or other source. A medical scanner may provide the data, such as a medical dataset representing a 3D region of the patient. Mining of a patient's computerized medical record may be used. Medical information from a picture archiving and communications server may be obtained. Any source may be used.
  • The medical data is for a given patient. While the training may use samples from many patients to learn features and/or learn to classify input medical data to provide values for rendering, the learnt model is applied to the medical data for a patient to output values for rendering parameters for that patient. The same or different machine-learnt models are applied to data for any number of patients.
  • Any type of medical data is obtained. Scan data representing a 3D volume is loaded as a medical dataset. The scan data may be from multiple two-dimensional scans or may be formatted from a 3D scan. The medical data may include non-image or non-scan data, such as patient test results, physician notes, lab results, diagnosis information, billing codes, clinical data, patient history, or other information.
  • The medical data includes the information learned to be used by the machine-trained model. Ground truth information and/or similarity or quality metric measures may be not provided. Where the machine learning learned to use feedback of predicted ground truth, then that feedback may be created and used as an input. The content to be used to determine the settings of the rendering parameters is provided to the machine-learnt model. For example, just the 3D scan data is provided as a medical dataset for a given patient. As another example, the 3D scan data and diagnostic information are provided as the medical dataset for the patient. The diagnostic information may be used to select the machine-learnt model to use and/or be part of the input feature vector.
  • Some machine-learnt models may be robust or still operate with some of the input feature vector missing. Alternatively, all of the information needed for the input feature vector is loaded.
  • In act 32, a sensor measures ambient light. Any luminosity sensor may be used, such as a charge coupled device. The sensor provides a measure of the luminosity. This measure may be used as an input in the input feature vector and/or to select a particular machine-learnt model to use. In alternative embodiments, the rendering is not based on ambient lighting conditions, so no measurement of ambient light is provided.
  • In an alternative embodiment for act 32, a light probe image is obtained as the measure of ambient light with a camera as the sensor. The image is obtained using any of a variety of ways and then used for high dynamic range image-based illumination during rendering. Examples ways for obtaining light probe images include capturing high-dynamic range photographs of a mirrored ball, stitching multiple photographs, or using self-contained panoramic cameras.
  • In act 34, the medical dataset is applied to the machine-learnt model. The loaded medical dataset is processed by a machine, such as a computer, processor, or server. The machine uses the internal representation of the machine-learnt model. The input feature vector is created and/or is the medical dataset. The machine inputs the feature vector, resulting in the machine-learnt model outputting the rendering settings based on the learned knowledge.
  • Acts 36-40 represent different acts in applying the machine-learnt model of presets for one or more rendering parameters. Additional, different, or fewer acts may be provided.
  • The machine-learned model was previously learned using any approach for machine learning, so may be a regression, classification, or reinforcement-learnt model. In one embodiment, the machine-learned model is trained with deep learning (e.g., deep neural network) to extract features from the medical dataset. By applying the deep-learnt model in act 36, features are extracted from the medical dataset. One or more other features may be extracted from the medical dataset using manually or user configured features. These features may then be used by another machine-learnt model and/or another hierarchal stage to output the values for the rendering. For example, a further reinforcement-learnt model is applied. The extracted features are input to the machine-learnt model for application.
  • Where the training was for consistency, the application of the machine-learnt model is more likely to output values in act 38 resulting in the photorealistic image corresponding to a standard image despite differences in the medical dataset. The photorealistic image is created using path tracing, other physically-based rendering, or reference photographs. The model uses the input feature vector to predict the values for rendering that result in the rendering of the scan data as an image with desired characteristics. The standard image may be a group of images having similar characteristics, such as resolution and/or contrast for a particular type of tissue or lesion. The rendering values output were learnt in order to provide the same or similar resolution and/or contrast for that particular type of tissue. Different machine-learnt models may be used for different diagnostic and/or workflow situations, such as outputting values for breast cancer imaging or for fetal imaging.
  • In another embodiment, the ambient light measured in act 32 is applied to the machine-learnt model with or as part of the medical dataset. The machine-learnt model is trained to output the values based in part on the ambient light. By applying the ambient light measure, the machine-learnt model outputs values for the rendering parameters that account for ambient light. Photorealistic images may be output as if lit by the ambient light in a room or to which the patient is subjected. Where the photorealistic images are used for augmented reality, use of ambient light and consistency in imaging accounting for ambient light may result in better blending or greater visual appeal.
  • The machine-learnt model is trained to output any number of values for a respective number of rendering parameters in act 40. For example, values for two or more physically-based rendering parameters are output based on the input medical dataset. Three or more, all, or a sub-set of any number of rendering parameters may have their values output by application to the machine-trained model or group of models. Example rendering parameters include controls for consistent data handling, transfer function, lighting design, viewing design, material propriety, or internal renderer parameters. One or more of the output values are for non-transfer function parameters.
  • In an example embodiment, the machine-learnt model or set of models output values for all of the rendering parameters except for viewing design. One or more parameters for viewing design are left to the user, such as for user navigation or setting of the camera position, orientation, and/or scale. The user inputs the values for all or some of the viewing design parameters, and the application of the medical dataset to the machine-learnt model provides the values for all of the other rendering parameters. Additional, different, or fewer types of rendering parameters may use initial input by the user.
  • The user may adjust model output values. Since physically-based rendering may be more time consuming than traditional rendering, providing a comprehensive set of values for the rendering parameters more likely to result in the desired image initially may avoid time-consuming adjustments. The user may, however, make any number of adjustments. The number and/or amount of change may be less due to the assistance by the machine-learnt model.
  • In act 42, a physically-based renderer renders the photorealistic image of the 3D region of the patient using the values output from the application. The output values of the machine-learnt model are used as settings by the physically-based renderer. For example, a sub-set or all of the rendering parameters for rendering with unbiased path tracing have values established by the output of the machine-learnt model.
  • The physically-based renderer is a graphics processing unit, array of graphics processing units, workstation, server, computer, or other machine capable of the required computations. Any physically-based rendering algorithm capable of computing the light transport may be used, such as path tracing, unbiased path tracing, Monte-Carlo path tracing, or Metropolis light transport. The physically-based rendering simulates the physics of light propagation to create an image instead of accumulating volume samples with local illumination along viewing rays from the camera, as is done for traditional volume rendering. In path tracing implemented for physically-based rendering, the paths of light or photons, including due to scattering and absorption, are modeled and used for rendering. The physically-based rendering result may be built up over time as the rendering may rely on probabilistic scattering and tracing millions of light paths.
  • The rendering results in a photorealistic image. A sequence of images may be provided as the image is built or rendered. Alternatively, for a given set of values of rendering parameters, a single image is output. In either case, the photorealistic image is rendered using presets from, at least in part, the machine-learnt model. A sequence of images due to alteration of values for one or more rendering parameters may be output. The user or renderer may change the values, such as rotation of a camera or for modeling fly-through.
  • In act 44, the renderer, host computer of the renderer, or other device transmits the photorealistic image or images. The transmission is over a bus, through a network interface, over a network, to a memory, or to a display. For example, the image is loaded into a buffer and output from the buffer to a display for viewing by a physician to aid diagnosis or pre-operative planning. A computer-generated representation of the medical data suitable for viewing by human users (e.g., 2D images, videos, stereoscopic image pairs, volumes, holograms, or lightfields) is output. Alternatively, the output may be for storage and/or for further machine analysis.
  • The transmission may be for any one or more of various purposes. In one embodiment, the transmission of the photorealistic image is part of a diagnostic report. A given diagnostic report may be specific to context, such as a report for kidney stones. Any kidney stones should be visible in an image. The photorealistic image shows kidney stones based on the values of the rendering parameters output by the model trained to create such images. The optimal context-specific images are provided for an automated diagnostic report in the clinical workflow.
  • In another embodiment, the transmission is to a display as an initial image for interactive viewing. The machine-learnt model is used to initialize interactive viewing applications in the clinical workflow. By providing values for rendering parameters so that the images are consistent with an established visualization design, the amount of user interaction may be reduced.
  • In yet another embodiment, the transmission is to an augmented reality device. For example, a physician may wear augmented reality glasses. The photorealistic image, with or without added annotations or graphics, is projected onto the glasses as an overlay on the actual view seen by the wearer. In other example, the photorealistic image is combined with a camera image and/or is used as a replacement for the camera image. In an augmented reality setup, the rendered image may be overlaid onto a live stream of real-time medical images (ultrasound, x-ray, optical laparoscopy, etc.).
  • FIG. 3 shows a block diagram of one embodiment of a system for content-based rendering based on machine learning. The system is for training with machine learning and/or application of a machine-learnt model. Using deep learning, learning to provide consistent images despite variability of input scan data, and/or learning a plurality of physically-based rendering parameters, a machine-learnt model is trained to output settings for rendering parameters given a new volume for a patient.
  • The system implements the method of FIGS. 1 and/or 2. Other methods or acts may be implemented, such as providing a user input (e.g., mouse, trackball, touch pad, and/or keyboard) and user interface for interactive rendering.
  • The system includes a machine 50, a memory 52, a display 54, a medical imager 56, and a renderer 58. Additional, different, or fewer components may be provided. For example, the medical imager 56 and/or memory 52 are not provided. In another example, a network or network connection is provided, such as for networking with a medical imaging network or data archival system. A user interface may be provided for interacting with the machine 50, renderer 58, or other components.
  • The machine 50, memory 52, renderer 58, and/or display 54 are part of the medical imager 56. Alternatively, the machine 50, memory 52, renderer 58, and/or display 54 are part of a server, workstation, or computer separate from the medical imager 56. The machine 50, memory 52, renderer 58, and/or display 54 are a personal computer, such as desktop or laptop, a workstation, a server, a network, or combinations thereof. In yet other embodiments, the machine 50 and memory 52 are part of a separate computer from the renderer 58.
  • The medical imager 56 is a medical diagnostic imaging system. Ultrasound, CT, x-ray, fluoroscopy, positron emission tomography (PET), single photon emission computed tomography (SPECT), and/or MR systems may be used. The medical imager 56 may include a transmitter and includes a detector for scanning or receiving data representative of the interior of the patient. The medical imager 56 acquires scan data representing the patient. The scan data may represent a volume of the patient. For example, scan data is acquired and used for diagnosis or surgical planning, such as identifying a lesion or treatment location.
  • In alternative embodiments, the medical imager 56 is not provided, but a previously acquired dataset for a patient is stored in the memory 52. In yet other alternatives, many medical images 56 are provided in the memory 52 as the training data, which is gathered and stored in the memory 52.
  • The machine 50 is a computer, workstation, server, processor, or other device configured to apply machine learning and/or to apply a machine-learnt model. The machine 50 is configured by software, hardware, and/or firmware. For learning, the machine 50 is configured by one or more machine learning algorithms. For applying a learnt model, the machine 50 is configured, in part, by a learnt matrix or matrices associating input data to output values.
  • The machine 50 applies a machine-learnt model, such as one learnt with deep learning. The machine-learnt model, as implemented by the machine 50, generates input features and/or outputs settings for rendering parameters based on application of medical data from a scan of a patient. The rendering parameters include windowing, transfer function, lighting, and/or material properties. The settings are learnt to provide an image similar in look and/or feel to one or more images from a same diagnostic context. Consistent images, such as to highlight or show anatomy of interest, are created. The diagnostic context provides the standard or information of interest.
  • The memory 52 is a graphics processing memory, a video random access memory, a random access memory, system memory, cache memory, hard drive, optical media, magnetic media, flash drive, buffer, database, combinations thereof, or other now known or later developed memory device for storing training data, rendering parameters values, medical datasets, and/or other information. The memory 52 is part of the medical imager 56, part of a computer associated with the machine 50, part of a computer associated with the renderer 58, a database, part of another system, a picture archival memory, or a standalone device.
  • The memory 52 stores scan data representing one or more patients. For example, data from the medical imager 56 is stored. The data is in a scan format or reconstructed to a volume or three-dimensional grid format. The scan data is stored for training and/or stored for rendering an image of a patient.
  • The memory 52 or other memory is alternatively or additionally a non-transitory computer readable storage medium storing data representing instructions executable by the programmed machine 50 for learning or applying the machine-learnt model. The instructions for implementing the processes, methods, and/or techniques discussed herein are provided on non-transitory computer-readable storage media or memories, such as a cache, buffer, RAM, removable media, hard drive, or other computer readable storage media. Non-transitory computer readable storage media include various types of volatile and nonvolatile storage media. The functions, acts or tasks illustrated in the figures or described herein are executed in response to one or more sets of instructions stored in or on computer readable storage media. The functions, acts or tasks are independent of the particular type of instructions set, storage media, processor or processing strategy and may be performed by software, hardware, integrated circuits, firmware, micro code and the like, operating alone, or in combination. Likewise, processing strategies may include multiprocessing, multitasking, parallel processing, and the like.
  • In one embodiment, the instructions are stored on a removable media device for reading by local or remote systems. In other embodiments, the instructions are stored in a remote location for transfer through a computer network or over telephone lines. In yet other embodiments, the instructions are stored within a given computer, CPU, GPU, or system.
  • The renderer 58 is a graphics processing unit, graphics card, graphic chip, multi-core processor, or other processor for running a software package that implements the photorealistic image generation functionality. The renderer 58 is configured by an application programming interface to render an image from the 3D scan data representing a patient. Using physically-based rendering, a photorealistic image is rendered. Using the settings output by the machine-learnt model, path tracing is applied to render the image. For feedback in training, the renderer 58 may render training images. Similarity is calculated by the renderer 58 and/or the machine 50.
  • The display 54 is a monitor, LCD, projector, plasma display, CRT, printer, or other now known or later developed device for displaying the photorealistic image or images. The display 54 receives images from the machine 50, memory 52, renderer 58, or medical imager 56. The images of the tissue captured by the medical imager 56 are displayed. Other information may be displayed as well, such as generated graphics, text, or quantities as a virtual overlay.
  • Additional images may be displayed. Where scan data represents a sequence of scans over time, a corresponding sequence of images may be generated. For interactive rendering, new images may be generated as settings for one or more rendering parameters are changed by a user.
  • While the invention has been described above by reference to various embodiments, it should be understood that many changes and modifications can be made without departing from the scope of the invention. It is therefore intended that the foregoing detailed description be regarded as illustrative rather than limiting, and that it be understood that it is the following claims, including all equivalents, that are intended to define the spirit and scope of this invention.

Claims (24)

1. A method for content-based rendering based on machine learning in a rendering system, the method comprising:
loading, from memory, a medical dataset representing a three-dimensional region of a patient;
applying, by a machine, the medical dataset to a machine-learnt model, the machine-learned model trained with deep learning to extract features from the medical dataset and trained to output values for two or more physically-based rendering parameters based on input of the medical dataset, the two or more physically-based rendering parameters being settings of a physically-based renderer, the settings used by the physically-based renderer to control rendering from three dimensions to two-dimensions, at least one of the two or more rendering parameters being controls for data consistency handling, lighting design, viewing design, material propriety, or internal renderer property;
rendering, by the physically-based renderer, a photorealistic image of the three-dimensional region of the patient using the output values resulting from the applying as the settings to control the rendering from the medical dataset, the rendering of the medical dataset of the three-dimensional region being to the photorealistic image in the two-dimensions; and
transmitting the photorealistic image.
2. The method of claim 1 further comprising loading patient information other than the medical dataset representing the three-dimensional region of the patient.
3. The method of claim 1 wherein applying comprises applying to output the values for the two or more physically-based rendering parameters as all of the physically-based rendering parameters other than viewing design based on applying user input of the viewing design and the medical dataset.
4. The method of claim 1 wherein applying comprises applying with the machine-learnt model trained to output the values resulting in the photorealistic image corresponding to a standard image despite differences in the medical dataset.
5. The method of claim 1 wherein applying comprises applying with the machine-learnt model comprising a regression, classification, or reinforcement learnt model.
6. The method of claim 1 wherein applying comprises applying with the deep learning as a deep neural network.
7. The method of claim 1 wherein applying comprises applying with the machine-learnt model as a deep reinforcement learnt model.
8. The method of claim 1 wherein rendering comprises rendering with unbiased path tracing.
9. The method of claim 1 wherein transmitting comprises transmitting as part of a diagnostic report, as an initial image of an interactive viewing, or as an overlay in augmented reality.
10. The method of claim 1 further comprising:
measuring ambient light with a light sensor;
wherein applying comprises applying the ambient light and the medical dataset, the machine-learnt model trained to output the values based in part on the ambient light.
11. A method for machine training for content-based rendering in a machine training system, the method comprising:
inputting first volume data of a volume of a patient, a first image of the volume, and first values of path tracing rendering parameters to training of an artificial intelligence, the path tracing rendering parameters being settings to control rendering from the volume to a two-dimensional image;
machine training, with a machine, the artificial intelligence to output second values of the path tracing rendering parameters for a second volume data where the second values control the rendering from the volume to provide a second rendered image of the second volume modeled on the first image; and
storing the trained artificial intelligence.
12. The method of claim 11 wherein inputting comprises inputting the patient non-image information to the training.
13. The method of claim 11 wherein training comprises training the artificial intelligence to output the second values as two or more of data consistency, transfer function, lighting, and viewing parameters.
14. The method of claim 11 wherein training comprises training the artificial intelligence to output the second values based on user selected viewing camera parameters, the path tracing rendering parameters for which second values are to be output being other than the viewing camera parameters.
15. The method of claim 11 wherein training so the second rendered image is modeled after the first rendered image comprises training with a metric of similarity.
16. The method of claim 11 wherein the first image comprises a reference photograph or video of the patient.
17. The method of claim 11 wherein inputting comprises perturbing the path tracing rendering parameters, creating a collection of sets of the path tracing rendering parameters, and wherein training comprises training based on selection of a sub-set of the sets.
18. The method of claim 17 wherein selection of the sub-set comprises selection by a user based on images rendered using the sets.
19. The method of claim 11 wherein machine training comprises deep learning with regression, classification, or reinforcement learning.
20. The method of claim 19 wherein machine training comprises deep reinforcement learning with a similarity of the second rendered image to the first image as a reinforcement.
21. The method of claim 20 wherein deep reinforcement learning comprises selecting with a probability distribution of different similarities including the similarity.
22. A system for content-based rendering based on machine learning, the system comprising:
a medical scanner configured to scan a patient;
a machine configured to output settings for rendering parameters by application of data from the scan to a machine-learnt model, the rendering parameters being controls for performing rendering from a volume to a two-dimensional image, the rendering parameters including windowing, transfer function, and lighting, and the settings learned to provide a first image from the data similar to one or more second images for a same diagnostic context; and
a graphics processing unit configured to render the first image from the data using the settings output by the application of the data to the machine-learnt model, the first image being a two-dimensional representation.
23. The system of claim 22 wherein the machine-learnt model is machine learnt with deep learning.
24. The system of claim 22 wherein the rendering parameters further comprise material properties, viewing properties, lighting properties and internal renderer properties, and wherein the graphics processing unit is configured to render with path tracing using the settings.
US15/191,043 2016-03-10 2016-06-23 Content-based medical image rendering based on machine learning Active US9760690B1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US15/191,043 US9760690B1 (en) 2016-03-10 2016-06-23 Content-based medical image rendering based on machine learning
CN201780016345.8A CN108701370B (en) 2016-03-10 2017-03-08 Content-based medical imaging rendering based on machine learning
EP17711584.7A EP3427230A1 (en) 2016-03-10 2017-03-08 Content-based medical imaging rendering based on machine learning
PCT/EP2017/055384 WO2017153445A1 (en) 2016-03-10 2017-03-08 Content-based medical imaging rendering based on machine learning
US15/643,973 US10339695B2 (en) 2016-03-10 2017-07-07 Content-based medical image rendering based on machine learning

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201662306407P 2016-03-10 2016-03-10
US15/191,043 US9760690B1 (en) 2016-03-10 2016-06-23 Content-based medical image rendering based on machine learning

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/643,973 Continuation US10339695B2 (en) 2016-03-10 2017-07-07 Content-based medical image rendering based on machine learning

Publications (2)

Publication Number Publication Date
US9760690B1 US9760690B1 (en) 2017-09-12
US20170262598A1 true US20170262598A1 (en) 2017-09-14

Family

ID=59752844

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/191,043 Active US9760690B1 (en) 2016-03-10 2016-06-23 Content-based medical image rendering based on machine learning
US15/643,973 Active 2036-08-08 US10339695B2 (en) 2016-03-10 2017-07-07 Content-based medical image rendering based on machine learning

Family Applications After (1)

Application Number Title Priority Date Filing Date
US15/643,973 Active 2036-08-08 US10339695B2 (en) 2016-03-10 2017-07-07 Content-based medical image rendering based on machine learning

Country Status (4)

Country Link
US (2) US9760690B1 (en)
EP (1) EP3427230A1 (en)
CN (1) CN108701370B (en)
WO (1) WO2017153445A1 (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3553695A1 (en) * 2018-04-13 2019-10-16 Alpine Electronics, Inc. Method and system for displaying image data from at least one night vision camera of a vehicle
US10592743B2 (en) * 2017-08-24 2020-03-17 International Business Machines Corporation Machine learning to predict cognitive image composition
US10643331B2 (en) 2017-05-03 2020-05-05 Siemens Healthcare Gmbh Multi-scale deep reinforcement machine learning for N-dimensional segmentation in medical imaging
WO2020139775A1 (en) * 2018-12-27 2020-07-02 Exo Imaging, Inc. Methods to maintain image quality in ultrasound imaging at reduced cost, size, and power
US10835209B2 (en) 2016-12-04 2020-11-17 Exo Imaging Inc. Configurable ultrasonic imager
US20210093301A1 (en) * 2018-04-09 2021-04-01 Koninklijke Philips N.V. Ultrasound system with artificial neural network for retrieval of imaging parameter settings for recurring patient
US11132775B2 (en) * 2018-11-16 2021-09-28 Samsung Electronics Co., Ltd. Image processing apparatus and method of operating the same
US11199623B2 (en) 2020-03-05 2021-12-14 Exo Imaging, Inc. Ultrasonic imaging device with programmable anatomy and flow imaging
US11364013B2 (en) * 2017-01-05 2022-06-21 Koninklijke Philips N.V. Ultrasound imaging system with a neural network for image formation and tissue characterization
US11488297B2 (en) * 2018-12-18 2022-11-01 Canon Medical Systems Corporation Medical information processing apparatus and medical information processing system
US11574739B2 (en) * 2018-01-29 2023-02-07 Function, Inc. Systems and methods for formulating personalized skincare products
US11971477B2 (en) 2018-09-25 2024-04-30 Exo Imaging, Inc. Imaging devices with selectively alterable characteristics
US12092735B2 (en) 2019-10-16 2024-09-17 Koninklijke Philips N.V. Method and apparatus for deep learning-based ultrasound beamforming

Families Citing this family (63)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9069725B2 (en) 2011-08-19 2015-06-30 Hartford Steam Boiler Inspection & Insurance Company Dynamic outlier bias reduction system and method
JP6795488B2 (en) 2014-04-11 2020-12-02 ハートフォード スチーム ボイラー インスペクション アンド インシュランス カンパニー Improved future reliability forecasts based on modeling system operation and performance data
US9569736B1 (en) * 2015-09-16 2017-02-14 Siemens Healthcare Gmbh Intelligent medical image landmark detection
US10991070B2 (en) 2015-12-18 2021-04-27 OrthoGrid Systems, Inc Method of providing surgical guidance
US20180025664A1 (en) * 2016-07-25 2018-01-25 Anna Clarke Computerized methods and systems for motor skill training
US10049301B2 (en) * 2016-08-01 2018-08-14 Siemens Healthcare Gmbh Medical scanner teaches itself to optimize clinical protocols and image acquisition
CN110392903B (en) 2016-08-19 2024-08-30 莫维迪厄斯有限公司 Dynamic culling of matrix operations
US10466714B2 (en) * 2016-09-01 2019-11-05 Ford Global Technologies, Llc Depth map estimation with stereo images
US10453165B1 (en) * 2017-02-27 2019-10-22 Amazon Technologies, Inc. Computer vision machine learning model execution service
US10977397B2 (en) 2017-03-10 2021-04-13 Altair Engineering, Inc. Optimization of prototype and machine design within a 3D fluid modeling environment
US10607393B2 (en) 2017-03-10 2020-03-31 Siemens Healthcare Gmbh Consistent 3D rendering in medical imaging
US10867085B2 (en) 2017-03-10 2020-12-15 General Electric Company Systems and methods for overlaying and integrating computer aided design (CAD) drawings with fluid models
US10409950B2 (en) 2017-03-10 2019-09-10 General Electric Company Systems and methods for utilizing a 3D CAD point-cloud to automatically create a fluid model
US11004568B2 (en) * 2017-03-10 2021-05-11 Altair Engineering, Inc. Systems and methods for multi-dimensional fluid modeling of an organism or organ
US10803211B2 (en) 2017-03-10 2020-10-13 General Electric Company Multiple fluid model tool for interdisciplinary fluid modeling
US10475165B2 (en) * 2017-04-06 2019-11-12 Disney Enterprises, Inc. Kernel-predicting convolutional neural networks for denoising
US10643368B2 (en) 2017-06-27 2020-05-05 The Boeing Company Generative image synthesis for training deep learning machines
US10699410B2 (en) * 2017-08-17 2020-06-30 Siemes Healthcare GmbH Automatic change detection in medical images
US11264135B2 (en) 2017-11-10 2022-03-01 Siemens Medical Solutions Usa, Inc. Machine-aided workflow in ultrasound imaging
US10593099B2 (en) 2017-11-14 2020-03-17 Siemens Healthcare Gmbh Transfer function determination in medical imaging
US10643320B2 (en) 2017-11-15 2020-05-05 Toyota Research Institute, Inc. Adversarial learning of photorealistic post-processing of simulation with privileged information
CN107910061A (en) * 2017-12-01 2018-04-13 中南大学 A kind of medical data processing method and system
US10722210B2 (en) * 2017-12-14 2020-07-28 Siemens Healthcare Gmbh Method for memorable image generation for anonymized three-dimensional medical image workflows
US11379715B2 (en) * 2017-12-15 2022-07-05 Meta Platforms, Inc. Deep learning based distribution of content items describing events to users of an online system
US10521908B2 (en) 2017-12-20 2019-12-31 International Business Machines Corporation User interface for displaying simulated anatomical photographs
US10438351B2 (en) 2017-12-20 2019-10-08 International Business Machines Corporation Generating simulated photographic anatomical slices
US10614570B2 (en) 2017-12-20 2020-04-07 International Business Machines Corporation Medical image exam navigation using simulated anatomical photographs
US10665007B2 (en) * 2018-01-15 2020-05-26 Siemens Healthcare Gmbh Hybrid interactive mode for rendering medical images with ray tracing
CN108495110B (en) * 2018-01-19 2020-03-17 天津大学 Virtual viewpoint image generation method based on generation type countermeasure network
US11443201B2 (en) * 2018-05-21 2022-09-13 Siemens Healthcare Gmbh Artificial intelligence-based self-learning in medical imaging
CN109147003A (en) * 2018-08-01 2019-01-04 北京东方畅享科技有限公司 Method, equipment and the storage medium painted to line manuscript base picture
CN109389667B (en) * 2018-08-23 2023-07-18 北京大学 High-efficiency global illumination drawing method based on deep learning
JP7466928B2 (en) 2018-09-12 2024-04-15 オルソグリッド システムズ ホールディング,エルエルシー Artificial intelligence intraoperative surgical guidance systems and methods of use
US11540794B2 (en) 2018-09-12 2023-01-03 Orthogrid Systesm Holdings, LLC Artificial intelligence intra-operative surgical guidance system and method of use
US11636292B2 (en) 2018-09-28 2023-04-25 Hartford Steam Boiler Inspection And Insurance Company Dynamic outlier bias reduction system and method
CN109273084B (en) * 2018-11-06 2021-06-22 中山大学附属第一医院 Method and system based on multi-mode ultrasound omics feature modeling
CN109671140B (en) * 2018-12-26 2024-02-02 上海赞奇文化科技有限公司 Cloud rendering service processing method adopting micro-service
US10973472B2 (en) * 2019-03-05 2021-04-13 Siemens Healthcare Gmbh Artificial intelligence-based material decomposition in medical imaging
CN110051434A (en) * 2019-04-25 2019-07-26 厦门强本科技有限公司 AR operation piloting method and terminal in conjunction with endoscope
US11132826B2 (en) * 2019-05-16 2021-09-28 Caterpillar Inc. Artificial image generation for training an object detection system
CN110277166B (en) * 2019-06-28 2020-05-15 屈兴玲 Auxiliary diagnosis system and method for palace laparoscope
KR20210012730A (en) 2019-07-26 2021-02-03 삼성전자주식회사 Learning method of artificial intelligence model and electronic apparatus
US11328177B2 (en) 2019-09-18 2022-05-10 Hartford Steam Boiler Inspection And Insurance Company Computer-based systems, computing components and computing objects configured to implement dynamic outlier bias reduction in machine learning models
JP7399269B2 (en) 2019-09-18 2023-12-15 ハートフォード スチーム ボイラー インスペクション アンド インシュアランス カンパニー Computer-based systems, computer components and computer objects configured to implement dynamic outlier bias reduction in machine learning models
US11615348B2 (en) 2019-09-18 2023-03-28 Hartford Steam Boiler Inspection And Insurance Company Computer-based systems, computing components and computing objects configured to implement dynamic outlier bias reduction in machine learning models
ES2815623A1 (en) * 2019-09-27 2021-03-30 Delgado Oscar Ruesga PROCEDURE FOR PERFORMING A DIAGNOSIS and DIGITAL MEDICAL FORECAST BASED ON DIGITAL ANATOMY (Machine-translation by Google Translate, not legally binding)
US11398072B1 (en) * 2019-12-16 2022-07-26 Siemens Healthcare Gmbh Method of obtaining a set of values for a respective set of parameters for use in a physically based path tracing process and a method of rendering using a physically based path tracing process
CN111353967B (en) * 2020-03-06 2021-08-24 浙江杜比医疗科技有限公司 Image acquisition method and device, electronic equipment and readable storage medium
US20230281959A1 (en) * 2020-03-25 2023-09-07 Yissum Research Development Company Of The Hebrew University Of Jerusalem Ltd. Deep learning-based anomaly detection in images
CN111310859A (en) * 2020-03-26 2020-06-19 上海景和国际展览有限公司 Rapid artificial intelligence data training system used in multimedia display
CN111739617B (en) * 2020-06-15 2023-07-11 中国医学科学院医学信息研究所 Medical image artificial intelligence quality control labeling method, device, equipment and storage medium
CN112308888B (en) * 2020-10-15 2024-07-12 复旦大学附属中山医院 Full-mode medical image sequence grouping method based on deep learning physical sign structure
US11635878B2 (en) * 2021-01-31 2023-04-25 Walmart Apollo, Llc Systems and methods for altering a graphical user interface
US20240156439A1 (en) * 2021-04-01 2024-05-16 Koninklijke Philips N.V. Systems and methods for self-sensing mobile medical devices to auto-correct an output
US11848095B2 (en) 2021-04-29 2023-12-19 Lymphatech, Inc. Identifying body part or body area anatomical landmarks from digital imagery for the fitting of compression garments for a person in need thereof
US11527320B2 (en) 2021-04-29 2022-12-13 Lymphatech, Inc. Methods and systems for generating accurately fitting compression garments having glove or glove-like features for a person in need thereof
US11211162B1 (en) 2021-04-29 2021-12-28 Lymphatech, Inc. Methods and systems for identifying body part or body area anatomical landmarks from digital imagery for the fitting of compression garments for a person in need thereof
US11720228B2 (en) 2021-08-05 2023-08-08 Evernorth Strategic Development, Inc. Systems and methods for transforming an interactive graphical user interface according to machine learning models
CN114332594B (en) * 2021-08-27 2024-09-06 吉林大学 Method for classifying unbalanced data of haptic material based on DDQN
CN113706583B (en) * 2021-09-01 2024-03-22 上海联影医疗科技股份有限公司 Image processing method, device, computer equipment and storage medium
CN114120062B (en) * 2021-11-26 2023-07-28 北京百度网讯科技有限公司 Sample generation method and device
EP4254350A1 (en) * 2022-04-01 2023-10-04 Siemens Healthcare GmbH Determination of illumination parameters in medical image rendering
CN117274473B (en) * 2023-11-21 2024-02-02 北京渲光科技有限公司 Multiple scattering real-time rendering method and device and electronic equipment

Family Cites Families (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6463438B1 (en) * 1994-06-03 2002-10-08 Urocor, Inc. Neural network for cell image analysis for identification of abnormal cells
US6690371B1 (en) * 2000-05-03 2004-02-10 Ge Medical Systems Global Technology, Llc Relevant image data extraction from a medical image data volume
US6826568B2 (en) * 2001-12-20 2004-11-30 Microsoft Corporation Methods and system for model matching
US6999549B2 (en) * 2002-11-27 2006-02-14 Ge Medical Systems Global Technology, Llc Method and apparatus for quantifying tissue fat content
US8711144B2 (en) 2006-08-01 2014-04-29 Siemens Medical Solutions Usa, Inc. Perception-based artifact quantification for volume rendering
US7840076B2 (en) * 2006-11-22 2010-11-23 Intel Corporation Methods and apparatus for retrieving images from a large collection of images
US8520906B1 (en) * 2007-09-24 2013-08-27 Videomining Corporation Method and system for age estimation based on relative ages of pairwise facial images of people
US10108620B2 (en) * 2010-04-29 2018-10-23 Google Llc Associating still images and videos
CN103220974A (en) * 2010-07-21 2013-07-24 阿敏·E·莫尔勒 Image reporting method
US20120078062A1 (en) * 2010-09-24 2012-03-29 International Business Machines Corporation Decision-support application and system for medical differential-diagnosis and treatment using a question-answering system
US9043316B1 (en) * 2011-03-28 2015-05-26 Google Inc. Visual content retrieval
US8509525B1 (en) * 2011-04-06 2013-08-13 Google Inc. Clustering of forms from large-scale scanned-document collection
US9020233B2 (en) 2011-09-19 2015-04-28 Siemens Aktiengesellschaft Method and system for up-vector detection for ribs in computed tomography volumes
US8831358B1 (en) * 2011-11-21 2014-09-09 Google Inc. Evaluating image similarity
US9235875B2 (en) 2012-11-01 2016-01-12 Google Inc. Image enhancement using learned non-photorealistic effects
WO2014152339A1 (en) * 2013-03-14 2014-09-25 Robert Bosch Gmbh Time and environment aware graphical displays for driver information and driver assistance systems
US9317777B2 (en) * 2013-10-04 2016-04-19 Monotype Imaging Inc. Analyzing font similarity for presentation
US9679258B2 (en) * 2013-10-08 2017-06-13 Google Inc. Methods and apparatus for reinforcement learning
US9008391B1 (en) * 2013-10-22 2015-04-14 Eyenuk, Inc. Systems and methods for processing retinal images for screening of diseases or abnormalities
WO2015099810A1 (en) * 2013-12-29 2015-07-02 Hewlett-Packard Development Company, L.P. Learning graph
US9324022B2 (en) * 2014-03-04 2016-04-26 Signal/Sense, Inc. Classifying data with deep learning neural records incrementally refined through expert input
US9412046B2 (en) * 2014-10-10 2016-08-09 Facebook, Inc. Training image adjustment preferences
US9454713B2 (en) * 2014-12-30 2016-09-27 Ebay Inc. Similar item detection
US9842390B2 (en) * 2015-02-06 2017-12-12 International Business Machines Corporation Automatic ground truth generation for medical image collections
US10905400B2 (en) * 2015-02-23 2021-02-02 Canon Medical Systems Corporation Apparatus and method for optimization of ultrasound images
US20160321523A1 (en) * 2015-04-30 2016-11-03 The Regents Of The University Of California Using machine learning to filter monte carlo noise from images
US20170161607A1 (en) * 2015-12-04 2017-06-08 Pilot Ai Labs, Inc. System and method for improved gesture recognition using neural networks
US10068385B2 (en) * 2015-12-15 2018-09-04 Intel Corporation Generation of synthetic 3-dimensional object images for recognition systems

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11759175B2 (en) 2016-12-04 2023-09-19 Exo Imaging, Inc. Configurable ultrasonic imager
US10835209B2 (en) 2016-12-04 2020-11-17 Exo Imaging Inc. Configurable ultrasonic imager
US11712222B2 (en) 2016-12-04 2023-08-01 Exo Imaging, Inc. Configurable ultrasonic imager
US11058396B2 (en) 2016-12-04 2021-07-13 Exo Imaging Inc. Low voltage, low power MEMS transducer with direct interconnect capability
US11364013B2 (en) * 2017-01-05 2022-06-21 Koninklijke Philips N.V. Ultrasound imaging system with a neural network for image formation and tissue characterization
US10643331B2 (en) 2017-05-03 2020-05-05 Siemens Healthcare Gmbh Multi-scale deep reinforcement machine learning for N-dimensional segmentation in medical imaging
US10592743B2 (en) * 2017-08-24 2020-03-17 International Business Machines Corporation Machine learning to predict cognitive image composition
US11468664B2 (en) * 2017-08-24 2022-10-11 International Business Machines Corporation Machine learning to predict cognitive image composition
US11574739B2 (en) * 2018-01-29 2023-02-07 Function, Inc. Systems and methods for formulating personalized skincare products
US20210093301A1 (en) * 2018-04-09 2021-04-01 Koninklijke Philips N.V. Ultrasound system with artificial neural network for retrieval of imaging parameter settings for recurring patient
US11948345B2 (en) * 2018-04-09 2024-04-02 Koninklijke Philips N.V. Ultrasound system with artificial neural network for retrieval of imaging parameter settings for recurring patient
EP3553695A1 (en) * 2018-04-13 2019-10-16 Alpine Electronics, Inc. Method and system for displaying image data from at least one night vision camera of a vehicle
US11971477B2 (en) 2018-09-25 2024-04-30 Exo Imaging, Inc. Imaging devices with selectively alterable characteristics
US11132775B2 (en) * 2018-11-16 2021-09-28 Samsung Electronics Co., Ltd. Image processing apparatus and method of operating the same
US11488297B2 (en) * 2018-12-18 2022-11-01 Canon Medical Systems Corporation Medical information processing apparatus and medical information processing system
WO2020139775A1 (en) * 2018-12-27 2020-07-02 Exo Imaging, Inc. Methods to maintain image quality in ultrasound imaging at reduced cost, size, and power
US12061257B2 (en) 2018-12-27 2024-08-13 Exo Imaging, Inc. Methods to maintain image quality in ultrasound imaging at reduced cost, size, and power
US12092735B2 (en) 2019-10-16 2024-09-17 Koninklijke Philips N.V. Method and apparatus for deep learning-based ultrasound beamforming
US11199623B2 (en) 2020-03-05 2021-12-14 Exo Imaging, Inc. Ultrasonic imaging device with programmable anatomy and flow imaging
US12019155B2 (en) 2020-03-05 2024-06-25 Exo Imaging, Inc. Ultrasonic imaging device with programmable anatomy and flow imaging

Also Published As

Publication number Publication date
US20170308656A1 (en) 2017-10-26
EP3427230A1 (en) 2019-01-16
WO2017153445A1 (en) 2017-09-14
CN108701370A (en) 2018-10-23
US10339695B2 (en) 2019-07-02
US9760690B1 (en) 2017-09-12
CN108701370B (en) 2020-01-21

Similar Documents

Publication Publication Date Title
US10339695B2 (en) Content-based medical image rendering based on machine learning
US10957098B2 (en) Consistent 3D rendering in medical imaging
US10643401B2 (en) Two-dimensional cinematic medical imaging in color based on deep learning
US10643331B2 (en) Multi-scale deep reinforcement machine learning for N-dimensional segmentation in medical imaging
US10893262B2 (en) Lightfield rendering based on depths from physically-based volume rendering
US10997475B2 (en) COPD classification with machine-trained abnormality detection
US12062429B2 (en) Salient visual explanations of feature assessments by machine learning models
US11727086B2 (en) Multimodality image processing techniques for training image data generation and usage thereof for developing mono-modality image inferencing models
US12033247B2 (en) Three-dimensional shape reconstruction from a topogram in medical imaging
US20190183451A1 (en) Method for memorable image generation for anonymized three-dimensional medical image workflows
CN113256553A (en) System and method for consistently presenting medical images using deep neural networks
US20220076829A1 (en) Method and apparatus for analyzing medical image data in a latent space representation
CN114387380A (en) Method for generating a computer-based visualization of 3D medical image data
US20240037731A1 (en) Image processing method and apparatus
CN112541882A (en) Implicit surface coloring in medical volume rendering

Legal Events

Date Code Title Description
AS Assignment

Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PETKOV, KALOIAN;MIAO, SHUN;YU, DAPHNE;AND OTHERS;SIGNING DATES FROM 20160603 TO 20160621;REEL/FRAME:039120/0696

AS Assignment

Owner name: SIEMENS HEALTHCARE GMBH, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS MEDICAL SOLUTIONS USA, INC.;REEL/FRAME:040511/0509

Effective date: 20161114

Owner name: SIEMENS HEALTHCARE GMBH, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ENGEL, KLAUS;REEL/FRAME:040511/0542

Effective date: 20161115

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

AS Assignment

Owner name: SIEMENS HEALTHINEERS AG, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS HEALTHCARE GMBH;REEL/FRAME:066267/0346

Effective date: 20231219