US20140330421A1 - Single action three-dimensional model printing methods - Google Patents
Single action three-dimensional model printing methods Download PDFInfo
- Publication number
- US20140330421A1 US20140330421A1 US14/332,578 US201414332578A US2014330421A1 US 20140330421 A1 US20140330421 A1 US 20140330421A1 US 201414332578 A US201414332578 A US 201414332578A US 2014330421 A1 US2014330421 A1 US 2014330421A1
- Authority
- US
- United States
- Prior art keywords
- printing
- image data
- image
- voxel
- voxels
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000007639 printing Methods 0.000 title claims abstract description 108
- 238000000034 method Methods 0.000 title claims abstract description 96
- 230000009471 action Effects 0.000 title description 3
- 238000010146 3D printing Methods 0.000 claims description 23
- 238000003384 imaging method Methods 0.000 claims description 19
- 210000000056 organ Anatomy 0.000 claims description 12
- 238000002595 magnetic resonance imaging Methods 0.000 claims description 3
- 238000000386 microscopy Methods 0.000 claims description 2
- 238000002600 positron emission tomography Methods 0.000 claims description 2
- 238000002591 computed tomography Methods 0.000 claims 1
- 238000002604 ultrasonography Methods 0.000 claims 1
- 230000008569 process Effects 0.000 abstract description 30
- 238000006243 chemical reaction Methods 0.000 abstract description 8
- 238000012545 processing Methods 0.000 description 27
- 210000000988 bone and bone Anatomy 0.000 description 25
- 238000003709 image segmentation Methods 0.000 description 18
- 239000000463 material Substances 0.000 description 13
- 210000001519 tissue Anatomy 0.000 description 9
- 238000009877 rendering Methods 0.000 description 7
- 239000010410 layer Substances 0.000 description 6
- 230000011218 segmentation Effects 0.000 description 6
- 238000003325 tomography Methods 0.000 description 6
- 210000003484 anatomy Anatomy 0.000 description 5
- 238000011960 computer-aided design Methods 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 5
- 239000007787 solid Substances 0.000 description 5
- 230000003993 interaction Effects 0.000 description 4
- 210000004556 brain Anatomy 0.000 description 3
- 230000008021 deposition Effects 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 239000007943 implant Substances 0.000 description 3
- 230000033001 locomotion Effects 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 238000000638 solvent extraction Methods 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000010009 beating Methods 0.000 description 2
- 239000004566 building material Substances 0.000 description 2
- 238000003711 image thresholding Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000012014 optical coherence tomography Methods 0.000 description 2
- 210000004872 soft tissue Anatomy 0.000 description 2
- 238000003860 storage Methods 0.000 description 2
- 208000027502 Ankle fracture Diseases 0.000 description 1
- 208000010392 Bone Fractures Diseases 0.000 description 1
- 206010017076 Fracture Diseases 0.000 description 1
- 238000012356 Product development Methods 0.000 description 1
- 239000000654 additive Substances 0.000 description 1
- 230000000996 additive effect Effects 0.000 description 1
- 239000011230 binding agent Substances 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 210000004369 blood Anatomy 0.000 description 1
- 239000008280 blood Substances 0.000 description 1
- 210000004204 blood vessel Anatomy 0.000 description 1
- 210000000746 body region Anatomy 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000005520 cutting process Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000000881 depressing effect Effects 0.000 description 1
- 230000000994 depressogenic effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000001125 extrusion Methods 0.000 description 1
- 210000002683 foot Anatomy 0.000 description 1
- 238000005755 formation reaction Methods 0.000 description 1
- 239000012634 fragment Substances 0.000 description 1
- 238000010100 freeform fabrication Methods 0.000 description 1
- 210000004884 grey matter Anatomy 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000010030 laminating Methods 0.000 description 1
- 210000005228 liver tissue Anatomy 0.000 description 1
- 238000003754 machining Methods 0.000 description 1
- 229940127554 medical product Drugs 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 229910052751 metal Inorganic materials 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000012768 molten material Substances 0.000 description 1
- 210000003205 muscle Anatomy 0.000 description 1
- 230000000399 orthopedic effect Effects 0.000 description 1
- 230000000149 penetrating effect Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 239000000843 powder Substances 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 230000008439 repair process Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000011524 similarity measure Methods 0.000 description 1
- 238000004513 sizing Methods 0.000 description 1
- 210000003625 skull Anatomy 0.000 description 1
- 239000002344 surface layer Substances 0.000 description 1
- 238000001356 surgical procedure Methods 0.000 description 1
- 229920001169 thermoplastic Polymers 0.000 description 1
- 239000004416 thermosoftening plastic Substances 0.000 description 1
- 210000004885 white matter Anatomy 0.000 description 1
Images
Classifications
-
- B29C67/0088—
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B29—WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
- B29C—SHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
- B29C64/00—Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
- B29C64/30—Auxiliary operations or equipment
- B29C64/386—Data acquisition or data processing for additive manufacturing
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B33—ADDITIVE MANUFACTURING TECHNOLOGY
- B33Y—ADDITIVE MANUFACTURING, i.e. MANUFACTURING OF THREE-DIMENSIONAL [3-D] OBJECTS BY ADDITIVE DEPOSITION, ADDITIVE AGGLOMERATION OR ADDITIVE LAYERING, e.g. BY 3-D PRINTING, STEREOLITHOGRAPHY OR SELECTIVE LASER SINTERING
- B33Y50/00—Data acquisition or data processing for additive manufacturing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/20—Finite element generation, e.g. wire-frame surface description, tesselation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B33—ADDITIVE MANUFACTURING TECHNOLOGY
- B33Y—ADDITIVE MANUFACTURING, i.e. MANUFACTURING OF THREE-DIMENSIONAL [3-D] OBJECTS BY ADDITIVE DEPOSITION, ADDITIVE AGGLOMERATION OR ADDITIVE LAYERING, e.g. BY 3-D PRINTING, STEREOLITHOGRAPHY OR SELECTIVE LASER SINTERING
- B33Y80/00—Products made by additive manufacturing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/04—Indexing scheme for image data processing or generation, in general involving 3D image data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/08—Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30008—Bone
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/41—Medical
Definitions
- the present invention is in the technical field of three-dimensional (“3D”) printing and rapid prototyping.
- the present invention is in the technical field of 3D printing and rapid prototyping using three or n-dimensional image data sets, such as CT (computerized tomography) or MRI (magnetic resonance imaging) images.
- CT computerized tomography
- MRI magnetic resonance imaging
- Three-dimensional (“3D”) printing of physical models is useful in a wide variety of settings. Some potential uses include production of anatomical bodies like bones for research and clinical applications, medical product development, machine design, and equipment design, to name just a few.
- 3D printing or rapid prototyping refers to a collection of technologies for producing physical parts directly from digital descriptions. Digital descriptions include output of any software that produces a 3D digital model.
- One example of such software is Computer-Aided Design (CAD) software.
- CAD Computer-Aided Design
- Creating a 3D digital model from a 3D image data set requires specialized imaging or CAD software. Rapid prototyping machines have been commercially available since the early 1990's, the most popular versions of which build a desired structure by adding building material layer-by-layer based on a digital three-dimensional model of the structure.
- FIG. 1 illustrates the current method of creating a physical model from an input image data set.
- the input image data set comes in the form of 3D voxel data or serial, sequenced two-dimensional (“2D”) images.
- a voxel (volumetric pixel or, more correctly, Volumetric Picture Element) is a volume element on a regular grid in a three dimensional space, having one or more numerical values as attributes such as intensity or color. This is analogous to a pixel (Picture Element), which has one or more numerical values as attributes on a regular grid in a 2D image data set.
- a 3D image data set may be organized as a series of 2D images and a voxel in a two-dimensional image plane may be referred to as a pixel.
- FIG. 1 when a user 11 needs to create a physical model 35 from an image data set 10 , the user 11 looks up the image on his/her computer 15 and transfers the image data 10 to an image processing operator 21 .
- the image processing operator 21 loads the image data 10 set on his computer 20 where special image modeling software is available.
- the image processing operator 21 reads the instructions sent by the user 11 to understand what type of model is required. If the image processing operator 21 still has questions or needs additional information, he will communicate with the user 11 to get the information.
- the image processing operator 21 then starts the process to create a 3D digital model 22 from the image data sets 10 on his computer 20 using specialized modeling software.
- the creation of the 3D digital model 22 requires a trained operator 21 , specialized imaging software, and a significant amount of user interaction.
- the image processing operator 21 needs to communicate frequently with the user 11 who has ordered the physical model to understand the requirements and applications of the model. The image processing operator is also required to spend a significant amount of time to perform image segmentation and to trace manually certain image areas.
- the 3D digital model 22 is then saved to a file format supported by a 3D printer or rapid prototyping machine 30 , for example, the STL (stereolithography) file format.
- the digital model file is then sent to the 3D printer or rapid prototyping machine 30 to generate a 3D physical model 35 .
- the three-dimensional (“3D”) printer 30 is likely located at a different location and operated by a 3D printing operator 31 .
- the 3D printing operator 31 sends it to the imaging processing operator 21 who then sends the finished physical model 35 back to the original user 11 .
- the present 3D printing techniques are complex and cost ineffective.
- the physical models may take too long to create to be useful, for example, to an emergency-care doctor.
- the image processing operator 21 After loading the image data, the image processing operator 21 then creates a digital 3D model, for example, a 3D polygonal surface model by using software-based modeling tools.
- a digital 3D model for example, a 3D polygonal surface model by using software-based modeling tools.
- 3D-DOCTOR can be used to produce 3D digital models of anatomical structures, as described in Yecheng Wu, From CT Image to 3 D Model , Advanced Imaging, August 2001, 20-23.
- the image processing operator 21 sends the digital model to a 3D printing service provider.
- the 3D printing operator 30 at the 3D printing service provider loads the digital model data on his computer, controls the 3D printer to produce a physical model, and then delivers the finished physical model to the doctor who ordered the model.
- the above-described process is user intensitive and requires operators to possess advanced software training, knowledge of the intended applications, and a good understanding of the difference between image data formats and the various data formats supported by 3D printer and rapid prototyping machines.
- Image segmentation refers to the delineation and labeling of specific image regions in an image data set that defines distinct structures.
- Image segmentation may include steps such as differentiating a particular structure from adjacent material having different composition and identifying distinct objects having the same or similar composition.
- CT Computerized Tomography
- MR Magnetic Resonance
- bony structures need to be delineated from other structures (soft tissues, blood vessels, etc.) in the images.
- each bone must typically be separated from adjacent bones when modeling anatomical structures such as cervical spine or foot.
- a useful feature is the capability of building a prototype of a patient-specific anatomical region quickly. For example, if a patient comes in with a broken ankle, the surgeon may use a physical model of the bone fragments of the patient to aid surgical planning, if the physical model can be generated rapidly. For orthopedic surgeons, the ability to visualize and manipulate a physical model of a bone or joint in need of repair prior to surgery can aid in the selection and design of surgical implants for fracture fixation or joint replacement. Rapid prototyping of patient specific models increases efficiency and reduces costs by cutting operating room time. Rapid prototyping of patient specific models offers tremendous promise for improved pre-operative planning and preparation.
- Rapid prototyping or 3D printing refers to a collection of technologies for producing physical parts directly from digital descriptions, which frequently are the output from Computer-Aided Design (CAD) software. Rapid prototyping machines have been commercially available since the early 1990's, and the most popular versions involve adding material to build the desired structure layer-by-layer based on a digital three dimensional model of the structure.
- CAD Computer-Aided Design
- a physical model may be fabricated using a rapid prototyping system using stereolithography, fused deposition modeling, or three dimensional printing.
- stereolithography a laser is used to selectively cure successive surface layers in a vat of photopolymer.
- fused deposition modeling a thermal extrusion head is used to print molten material (typically a thermoplastic) that fuses onto the preceding layer.
- molten material typically a thermoplastic
- a typical three-dimensional printer uses a printer head to selectively deposit binder onto the top layer of a powder bed.
- the present application discloses systems and methods for single-action printing of 3D physical models from a three or n-dimensional image data set.
- the methods may be applied to image data set obtained from any of a wide variety of imaging modalities, including Computerized Tomography (“CT”), Magnetic Resonance (“MR”), positron emission tomography (“PET”), optical coherence tomography (“OCT”), ultrasonic imaging, X-ray imaging, sonar, radar including ground penetrating radar, acoustic imaging, microscopy imaging, simulated image data and the like, or combinations of one or more imaging modalities.
- CT Computerized Tomography
- MR Magnetic Resonance
- PET positron emission tomography
- OCT optical coherence tomography
- ultrasonic imaging X-ray imaging
- sonar including ground penetrating radar
- acoustic imaging microscopy imaging
- simulated image data and the like or combinations of one or more imaging modalities.
- the systems and methods are applicable to a wide range
- the systems and methods described herein generally contemplate combining printing template methods with a 3D printer or rapid prototyping machine.
- the printing template methods usually include predefined data processing steps comprising identifying voxels in an image data set, generating a geometric representation, and sending the geometric representation to the 3D printer to produce a 3D physical model.
- a 3D printer or rapid prototyping machine refers to a collection of devices capable of producing three-dimensional physical parts directly from digital models using stereolithography, fused deposition modeling, three dimensional printing, sheet laminating or other technologies.
- FIG. 1 illustrates a current method of printing a 3D model from an image data set.
- FIG. 2 illustrates a proposed method for printing a 3D model from an image data set.
- FIG. 3 illustrates a flowchart of an exemplary single-action 3D model printing method.
- FIG. 4 illustrates a flowchart of image data conversion steps included in a printing template.
- FIG. 5 illustrates a list of 3D printing templates accompanied and described with text.
- FIG. 6 illustrates a list of 3D printing templates accompanied and described with text and graphics.
- FIG. 7 illustrates 3D points as an exemplary geometric representation generated by a printing template from image data.
- FIG. 8 illustrates a 3D contour as an exemplary geometric representation generated by a printing template from image data.
- FIG. 9 illustrates a 3D triangle-based surface model as an exemplary geometric representation generated by a printing template from image data.
- FIG. 10 provides a table of sample CT numbers for various human tissues.
- FIG. 11 illustrates an exemplary printing template of printing a bone structure from a CT image data set.
- FIG. 13 illustrates an exemplary printing template of printing a physical model using predefined value ranges from an image data set.
- FIG. 14 illustrates an exemplary process of printing a physical model of selected organs or parts from an image data set.
- FIG. 15 illustrates an exemplary image with seed voxels marked before image segmentation.
- FIG. 16 illustrates a segmentation result of a first round region growing.
- FIG. 17 illustrates a segmentation result of a second round region growing.
- FIG. 18 illustrates a segmentation result of a last round region growing.
- FIG. 20 illustrates an example of user adjustable physical model printing method.
- FIG. 2 illustrates an exemplary system using a single-action 3D printing method to print a 3D physical model from an input image data set 10 .
- the 3D model can be, for example, a patient-specific anatomical model.
- the image data set 10 such as CT data, MR data etc, is loaded on a computer 15 .
- the image data set 10 is typically a voxel-based image data set depicting a 3D region with each voxel of the image data set 10 encoding at least one image attribute, such as image intensity, color or the like.
- a user 11 at the computer 15 selects one printing template 18 from a list of printing templates ( 300 in FIG. 5 ) for printing a 3D physical model.
- FIG. 3 is a flowchart of a single-action 3D printing method.
- an image data set 10 is first received in step 210 .
- a selected printing template 18 is executed to identify the voxel categories and generate a geometric representation for printing a 3D model.
- the generated geometric representation is sent to a 3D printer 30 and in step 235 , a 3D physical model 35 is produced.
- step 218 represents a single user action involved in the printing process of a 3D model.
- selecting a printing template includes a selection action by using a pointing device to position on a specific printing template from a list of predefined printing templates and select the printing template for execution.
- the single-action may be a clicking of a mouse button when a cursor is positioned over a predefined area of a displayed list of printing templates or a depressing of a key on a key pad to select a specific printing template.
- a printing template as defined herein is a software program for identifying voxels in an image data set, generating a geometric representation of a 3D physical model in a data format supported by a 3D printer, and sending the geometric representation to a 3D printer to create a 3D physical model.
- 3D printers require a geometric representation of an object in order to fabricate the geometric shapes required in making a 3D physical model.
- the geometric representation of an object may include one or a combination of the following forms: a list of 3D points 501 - 506 for the entire body of the object with locational and material information defined at each 3D point ( FIG. 7 ), a group of 3D contours 552 - 561 to define the shape of the object on each image plane ( FIG. 8 ), or surface models 580 ( FIG. 9 ) consisting of triangles or polygons or surface patches delineating the body of the object.
- FIG. 4 illustrates a flowchart of the image data conversion step in a printing template.
- An input image data set 10 is received by a printing template 15 .
- the printing template starts its predefined voxel identifying function 310 to identify voxel categories in the image data. After identifying the voxel categories, the printing template generates one of three geometric representations (3D points 315 , 3D contours 320 or surface models 325 ) supported by a 3D printer and sends the generated geometric representation to the 3D printer 30 to produce a 3D physical model 35 .
- 3D points 315 , 3D contours 320 or surface models 325 supported by a 3D printer
- the image data conversion process generates a geometric representation and any additional data needed for a 3D printer to print out the physical model.
- the voxel identifying process is generally done using image processing techniques such as image segmentation and classification.
- image segmentation and classification One purpose of image segmentation and classification is to identify the voxel categories at each voxel location for the entire image data set.
- Commonly used image classification techniques include trained classifiers (such as artificial neural networks), image clustering using voxel similarity measures, etc.
- Commonly used image segmentation techniques include image thresholding, histogram thresholding, region growing, region splitting, watershed method, graph partitioning, clustering, artificial neural network, and other methods.
- the geometric representation generated from the input image data set for 3D printing may be a list of 3D points 501 - 506 in the body of an object with locational and material information ( FIG. 7 ) specified at each point, or a set of 3D contours 551 - 561 to define the shape of an object in the image planes ( FIG. 8 ), or surface models of an object 580 ( FIG. 9 ), or a combination of them.
- the data generated for 3D printing is not limited to geometrical representations such as points, contours or surfaces as described.
- the data may also be organized as a list of printing instructions, such as “move to a location”, “deposit a specified amount of building material”, “move to a new location”, etc., that can be used to complete the physical model printing process.
- the list of printing templates may be displayed as either text 300 ( FIG. 5 ) or graphics 301 ( FIG. 6 ) on the computer 15 .
- the text may use a description such as “bone structure” or “brain.”
- the graphic display may use pre-drawn graphic icons to indicate “bone”, “skull”, or “brain.”
- the graphic display of a printing template may also use a 3D graphic rendering of the geometric representation generated from the image data by the printing template.
- the number of printing templates is not limited. Additional printing templates may be added for specific physical model printing needs. New printing templates may be created with different processing steps and parameters.
- a printing template may also be implemented as a part of the input image; in which case user interaction is not required.
- This embodiment is implemented as a printing template for printing a physical model of a bone structure from a CT image data set.
- the intensity value at each voxel may be converted to a value in Hounsfield units (HU).
- the Hounsfield unit system measures the attenuation coefficient of tissues in computerized tomography.
- Hounsfield units are also termed CT numbers.
- FIG. 10 provides a table of sample CT numbers for various human tissues. The table lists some of the voxel values of different tissues or materials in Hounsfield units for a typical CT scanner. The values may differ on a different CT image scanner due to specific settings on that particular imaging device and custom calibrations of image data.
- the formula to calculate the CT number in Hounsfield units from the voxel intensity is normally provided as part of the image data. For example, the formula used by many CT scanner vendors is:
- HU is the voxel value in Hounsfield units
- Voxel Intensity is the attribute value of each voxel provided in an image data set
- bone tissues may be identified using a range of CT numbers (>1000).
- the value of every voxel in the image data set can be checked to identify bone tissues. For example, if a voxel has a value above 1000 HU, it is marked as bone tissue. Often an upper limit is used to prevent other hard materials such as metal implants from being marked as bone tissue.
- a similar technique may be applied to other tissues, such as soft tissue (fat, muscle, etc), blood, liver tissue, and white and grey matter in the brain.
- FIG. 11 is a flowchart of an exemplary printing template for printing bone structures from a CT image set.
- the printing template identifies voxels that are part of the bone structure in the CT image 600 , generates a geometric representation (in the format of 3D points 605 , Contours 606 , 606 , or Surfaces 607 ) in a 3D printer supported data format and sends the geometric representation to the 3D printer 608 to generate a physical model 610 .
- the printing template includes the following processing steps:
- Step 601 Go through the entire image data set 600 to check the HU value of each voxel (Step 601 ).
- Step 602 For each voxel with a HU value larger than 1000 HU but less than an upper bound, mark the voxel with value 1 to indicate the voxel as representing bone tissue (Step 602 ). Otherwise, mark it with value 0 to indicate non-bone tissue (Step 603 ).
- Step “a” and “b” until all voxels are checked, in which case a geometric representation is generated (Step 604 ).
- the value 1000 HU is used here as an example. Different values or ranges may be used for different images.
- a geometric representation comprising a list of 3D points for all voxels marked with value 1 may be generated (Step 605 ) and sent to the printer (Step 608 ) to generate a physical model 610 .
- FIG. 7 shows an example of the 3D points generated from an image data set.
- every 3D point has an identification value which is either 0 or 1.
- 0 indicates non-bone tissue and the voxels with value 0 are represented here by a white color.
- 1 indicates bone tissue and the voxels with value 1 are represented here by a dark color.
- every voxel may have one or multiple identification values which may be any value, not limited to 0 or 1.
- the list of 3D points are represented as:
- each point has a data format of (X, Y, Z, Value).
- X, Y, Z are the three-dimensional coordinates of a voxel and Value is the attribute with a value of, in this case, 1 for all the voxels identified as bone tissue and 0 otherwise. Other values may be used for identification purposes. Additional values may be also used to indicate attributes such as color or material.
- a geometric representation comprising the contours are generated by tracing along the outer edge of all voxels marked with the value of 1 (Step 606 ).
- the contour tracing method is straight forward, and is normally done by walking along the edge voxels in a fixed order within each 2D image plane.
- FIG. 8 shows an example of tracing a contour in a 2D image plane.
- the tracing process starts at one edge voxel 551 and the contour starts with no point data.
- Voxel 551 is added to the contour as the starting point.
- the next voxel on the edge to be traced is voxel 552 .
- Voxel 552 is added to the contour.
- the contour may be represented as:
- each point has a data format of (X, Y, Z, Value).
- X, Y, Z are the three-dimensional coordinates of a voxel and Value is the attribute with a value of, in this case, 1 for all the voxels identified as bone tissue and 0 otherwise.
- Other values may be used for the attribute and additional attributes such as color or material may be included as well.
- the particular tracing technique is described as an example. Other tracing methods and variations may be used to generate similar results.
- a geometric representation in the format of a surface model is generated using the “Marching Cubes” (U.S. Pat. No.
- Step 607 4,710,876, 4,751,643, 4,868,748) method or other surface modeling methods.
- the generated geometric representation is sent to the 3D printer to produce a physical model 610 (Step 608 ).
- Most commercially available 3D printers and rapid prototyping machines support the “STL” format, which stores surface geometry data as a set of raw unstructured triangles.
- the surface model 607 is sent to the the three-dimensional (“3D”) printer in the “STL” format.
- “Marching cubes” is a computer graphics algorithm for extracting a polygonal mesh of an isosurface from three-dimensional voxels. The algorithm proceeds through the voxels marked with 1, taking eight neighbor locations at a time (thus forming an imaginary cube) and then determining the polygon(s) needed to represent the part of the isosurface that passes through this cube. The individual polygons are then fused into the desired surface.
- the “Marching Cubes” algorithm generates triangle-based surface models. Additional post processing steps such as surface smoothing and surface decimation may be applied to improve the surface quality but are not required.
- FIG. 9 shows an example of a three dimensional triangle-based surface model 580 .
- the triangle-based surface model has 8 vertexes: P0, P1, P2, P3, P4, P5, P6, P7 and 12 surface triangles with T1, T2, T3, T4, T5, T6 displayed at the front of the model and T7, T8, T9, T10, T11, T12 displayed at the back of the model.
- Each vertex is a 3D point: (X, Y, Z).
- Each triangle has 3 vertexes, for example (P0, P2, P1).
- This surface model may be represented as:
- the surface model representation is similar to the commonly used “STL” format and may be sent to the the 3D printer in the “STL” format for printing a physical model.
- Other representations and variations, such as surface patches or polygon-based surfaces, may also be used.
- the above example describes one embodiment of the single-action 3D image printing methods.
- the steps in the printing template may be combined or varied.
- the voxel checking and marking Steps “a” and “b” can be combined into Step “e” that checks the voxel values and generates the surface triangles without marking the voxels.
- This embodiment is implemented as a printing template for printing a physical model of a solid body from an image data set.
- the voxels in an empty or no-tissue region in an image typically have a known value range.
- CT Computerized Tomography
- MR Magnetic Resonance
- a voxel representing air has a value range around ⁇ 1000 HU as shown in the CT values table ( FIG. 10 ).
- a voxel in a CT image has a value between ⁇ 1000 HU and ⁇ 200 HU (the value below the lowest tissue value in Hounsfield unit)
- the voxel may be identified as air. Otherwise the voxel may be identified as body tissue.
- the same method may be applied to other imaging modalities to identify empty regions that are defined with known voxel value ranges.
- FIG. 12 illustrates an exemplary printing template for printing a solid body structure from an image data set.
- the printing template identifies voxels (Step 621 ) in empty regions and body regions in the image, generates a geometric representation (Steps 625 , 626 , or 627 ), and sends the geometric representation to the 3D printer to create a physical model 629 (Step 628 ).
- the printing template includes the following processing steps:
- Step 621 Go through the entire image data set 620 to check the value of each voxel in Hounsfield units (Step 621 ).
- Step 623 For each voxel, if its value is within the value range of no-tissue (empty region, for example, air), mark the voxel with value 0 to indicate it is empty (Step 623 ). Otherwise, mark it with value 1 to indicate it has tissue (Step 622 ). Repeat Step “a” and “b” until all voxels are checked, in which case, a geometric representation is generated (Step 624 ).
- Step 625 If the output printing device supports input data in the format of 3D points, we then generate a list of 3D points for all voxels marked with value 1 (Step 625 ) and send the list to the printer to generate a physical model 629 (Step 628 ). If other information such as material or color is supported, we can extract such information from the input image data and send it together with the geometric representation in the format of 3D points. See FIG. 7 for an example of the 3D points generated from an image data set. d) If the 3D printer supports input data in the format of contours, then a geometric representation may be generated by tracing the contours along the outer edge of all voxels marked with value 1 (Step 626 ). See FIG.
- Step 8 for an example of tracing a contour in a 2D image plane.
- a geometric representation in the format of a surface model is generated (Step 627 ) using the “Marching Cubes” (U.S. Pat. No. 4,710,876, 4,751,643, 4,868,748) method or other surface modeling methods.
- the geometric representation is then sent to the 3D printer to produce a physical model 629 (Step 628 ).
- the above example describes one embodiment of the single-action 3D image printing methods.
- the steps in the printing template may be combined or varied, for example, the voxel checking and marking Steps “a” and “b” can be combined into Step “e” that checks the voxel values and generates the surface triangles without marking the voxels.
- This embodiment is implemented as a printing template for printing a physical model from an image data set using predefined voxel value ranges.
- FIG. 13 illustrates an exemplary printing template for printing a physical model from an image data set 630 using one or more predefined voxel value ranges.
- a predefined value range may be in voxel intensity, color, texture, location, region, or any derived value from them.
- a typical range has a low value and a high value to define the bounds of the range.
- a list of ranges may be used to define multiple value ranges that are not adjacent to each other.
- the method identifies voxels using the predefined voxel value range (Step 631 ), generates a geometric representation (Steps 635 , 636 , or 637 ), and sends the data to the 3D printer to generate a physical model 639 (Step 638 ). It includes the following steps:
- Step 631 a) Go through the entire image data set to check the value of each voxel against the value ranges defined in the printing template (Step 631 ).
- Step 632 For each voxel with a value within the bound of one of the defined ranges, mark the voxel with value 1 to indicate it is within the specified range (Step 632 ). Otherwise, mark with value 0 to indicate it is outside (Step 633 ). Repeat Step “a” and “b” until all voxels are checked and identified. A geometric representation is generated in Step 634 .
- Step 635 If the 3D printer supports input data in the format of three-dimensional points, we generate a geometric representation comprising a list of 3D points for all voxels marked with value 1 (Step 635 ) and send the geometric representation to the 3D printer (Step 638 ) to create a physical model 639 . If other information such as material or color are supported in the input image data 630 , we may include the information and send it together with the geometric representation to the printer.
- FIG. 7 shows an example of the 3D points generated from an image data set.
- a geometric representation can be generated by tracing the contours along the outer edge of all voxels marked with value 1 (Step 636 ). See FIG.
- a geometric representation may be generated using the “Marching Cubes” (U.S. Pat. No. 4,710,876, 4,751,643, 4,868,748) method or other surface modeling methods (Step 637 ).
- the geometric representation is then sent to the 3D printer (Step 638 ) to produce a physical model 639 .
- Most commercially available 3D printers and rapid prototyping machines support the “STL” format, which stores surface geometry data as a set of raw unstructured triangles.
- the surface model 637 is sent to the 3D printer in the “STL” format.
- the above example describes one embodiment of the single-action 3D image printing methods.
- the steps in the printing template may be combined or varied.
- the voxel checking and marking Steps “a” and “b” can be combined into Step “e” that checks the voxel values and generates the surface triangles without marking the voxels.
- This embodiment is implemented as a printing template for printing a physical model of selected organs or parts from an image data set.
- FIG. 14 illustrates an example of printing a physical model 649 of selected organs or parts from an image set 640 .
- the image regions of the selected organs need to be identified using image segmentation techniques (Step 644 ).
- a typical segmentation technique used for identifying specific image regions starts with either a set of automatically generated (Step 642 ) or user selected seed locations or regions (Step 641 ), grows each region by merging neighboring voxels that are within a certain similarity criterion, and repeats the process until no more neighboring voxels are available for merging.
- the criterion could be a difference of voxel intensity, gray level, texture, or color between the voxels already identified and the ones being checked.
- the identified regions are then used to generated a geometric representation for the 3D printer.
- Other image segmentation methods such as region growing, active contours, graph partitioning, watershed, and clustering, may be used in the image region identifying step of this embodiment.
- the method identifies voxels using a region growing technique (Step 644 ), generates a geometric representation of the identified voxels in a format supported by a 3D printer 648 (Steps 645 , 646 , or 647 ) and sends the geometric representation to the 3D printer to generate a physical model of selected organs or parts 649 (Step 658 ).
- a user 11 selects some voxels or regions on the input image as seed voxels or regions.
- FIG. 15 illustrates an example where two selected seed voxels 701 at location (4, 4, Z) and (5, 4, Z) are marked with dark color. The location is represented by the X, Y, Z coordinates of a voxel.
- FIG. 15 shows the original image as a grayscale image. The grayscale value of this image is from 0 to 255. Seed voxels are checked (Step 641 ) to see whether they are available and whether they are within the bound of the image set.
- Step 642 is carried out to generate seed voxels automatically. For example, a predefined value range may be used to select seed voxels within a value range.
- the image segmentation process through region growing starts at the seed voxel locations.
- the process grows each region by merging neighboring voxels that are within a certain similarity criterion and repeats the process until no more neighboring voxels are available for merging.
- a simple criterion for measuring voxel similarity that is, for a neighboring voxel to qualify as similar to a reference voxel
- the grayscale value difference between the neighboring voxel and the reference voxel must be less than 20. In other words, if the grayscale value difference is less than 20, the neighboring voxel is added to the region and the region grows by one voxel.
- voxel 702 (4, 3, Z) in FIG. 15 has a grayscale value of 103.
- the difference between this value and the grayscale value (102) of the voxel located at (4, 4, Z) 701 is 1.
- the difference is within the similarity criterion, so the neighboring voxel 702 is added to the region.
- FIG. 16 illustrates the first round of region growing for the image example shown in FIG. 15 . All neighboring voxels with grayscale difference less than 20 are marked with a dark color.
- FIG. 17 illustrates the result of the second round region growing.
- 3 voxels 703 (marked with vertical hatch lines) at locations (4, 6, Z), (5, 6, Z), (6, 6, Z) are added to the region.
- FIG. 18 illustrates the result of the last round region growing.
- 1 voxel 704 (marked with diagonal hatch lines) at locations (5, 7, Z) is added to the region. After this round, no more voxels meet the similarity criteria. The region growing process stops. All voxels added to the region are marked with value 1 and the rest of the voxels are marked with value 0 as shown in FIG. 19 .
- a geometric representation is generated in the format of either 3D points, or contours, or surfaces (Step 645 , 646 , or 647 ).
- the geometric representation is then sent to a 3D printer to create a physical model 649 (Step 648 ).
- the above example describes one embodiment of the single-action 3D image printing methods in which the image segmentation technique uses a region growing method. Other image segmentation methods may be used to generate similar results.
- This embodiment is implemented as a printing template for printing a physical model from an image data set using user adjustable image processing parameters and steps.
- FIG. 20 illustrates an exemplary printing template that adopts a user adjustable physical model printing method.
- This printing template provides a user with a list of selectable processing options and adjustable parameters for image segmentation and data conversion.
- the user 11 makes the initial selection of parameters and segmentation methods (Step 651 ).
- the segmentation methods may include the ones described in the above examples, such as region growing, image thresholding, graph partitioning, and others.
- the parameters may include value ranges that are adjustable, user defined seed regions, and others.
- the image segmentation process (Step 654 ) segments the input image using the selected methods and parameters.
- the geometric representation generated from the image segmentation process is used to create a 3D rendering (Step 660 ) to show how a final physical model may look on a computer.
- the computer 15 FIG.
- the 2 can be used to display a three-dimensional rendering 660 of the model generated by the image segmentation process.
- the three-dimensional rendering may be implemented using volume rendering of the segmented image or surface rendering from the surface model 657 . If the rendering meets the user's requirements, the segmented image is converted to a format supported by the 3D printer 658 for printing (Steps 655 , 656 , or 657 ). Otherwise, the user can make additional adjustments. The user can decide to print the 3D model on a 3D printer using the current settings or continue the adjustment (Step 651 ) until the user is satisfied with the settings for printing.
- this embodiment further includes the processing steps of converting an image data set 650 to a geometric representation in a data format supported by the 3D printer and sending the geometric representation to the 3D printer to generate a physical model of selected organs or parts 659 (Step 658 ).
- n-dimensional data where n may be 2, 3, 4 or any number larger than 4.
- the invention may be applied to n-dimensional data in which one of the dimensions is time and there are two or three spatial dimensions. For example, we can use the above described methods to produce multi-dimensional physical models that evolve over time.
- time-sequence 3D image data of a chest containing a beating heart may be used to generate a series of chest models to show the shape and motion of the beating heart at different time points.
- a voice command may be spoken by the user.
- a key may be depressed by the user.
- a button on a 3D printing device may be pressed by the user.
- Selection using any pointing device may be effected by a user to start the execution of a printing template.
- a single-action may be preceded by multiple physical movements of the user (e.g., moving a mouse so that a mouse pointer is over a button)
- the single-action generally refers to a single event received by a system that commands the system to print a physical model from an image data set or a derived representation of the image.
- various techniques for identifying voxel categories and generating a geometric representation can be used to print a physical model from an image data set.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Materials Engineering (AREA)
- Chemical & Material Sciences (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Manufacturing & Machinery (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Mechanical Engineering (AREA)
- Optics & Photonics (AREA)
- Architecture (AREA)
- Evolutionary Computation (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
Abstract
Methods and techniques of using 3D printers to create physical models from image data are discussed. Geometric representations of different physical models are described and complex data conversion processes that convert input image data into geometric representations compatible with third party 3D printers are disclosed. Printing templates are used to encapsulate complex geometric representations and complicated data conversion processes from users for fast and simple 3D physical model printing applications.
Description
- The present application is a divisional application of U.S. patent application Ser. No. 14/036,620 filed on Sep. 25, 2013 (now U.S. Pat. No. 8,579,620 issued Nov. 12, 2013), which is a divisional application of U.S. patent application Ser. No. 13/039,048 filed on Mar. 2, 2011, both of which are hereby incorporated by reference in its entirety.
- The present invention is in the technical field of three-dimensional (“3D”) printing and rapid prototyping. In particular, the present invention is in the technical field of 3D printing and rapid prototyping using three or n-dimensional image data sets, such as CT (computerized tomography) or MRI (magnetic resonance imaging) images.
- Three-dimensional (“3D”) printing of physical models is useful in a wide variety of settings. Some potential uses include production of anatomical bodies like bones for research and clinical applications, medical product development, machine design, and equipment design, to name just a few. 3D printing or rapid prototyping refers to a collection of technologies for producing physical parts directly from digital descriptions. Digital descriptions include output of any software that produces a 3D digital model. One example of such software is Computer-Aided Design (CAD) software. Creating a 3D digital model from a 3D image data set requires specialized imaging or CAD software. Rapid prototyping machines have been commercially available since the early 1990's, the most popular versions of which build a desired structure by adding building material layer-by-layer based on a digital three-dimensional model of the structure.
- However, because of the amount of user interaction time involved and the complexity of data conversion process between image data formats and data formats supported by 3D printers or rapid prototyping machines, applications of the present technology of producing 3D physical models from three or n-dimensional images are rather limited.
-
FIG. 1 illustrates the current method of creating a physical model from an input image data set. The input image data set comes in the form of 3D voxel data or serial, sequenced two-dimensional (“2D”) images. A voxel (volumetric pixel or, more correctly, Volumetric Picture Element) is a volume element on a regular grid in a three dimensional space, having one or more numerical values as attributes such as intensity or color. This is analogous to a pixel (Picture Element), which has one or more numerical values as attributes on a regular grid in a 2D image data set. A 3D image data set may be organized as a series of 2D images and a voxel in a two-dimensional image plane may be referred to as a pixel. - In
FIG. 1 , when auser 11 needs to create aphysical model 35 from an image data set 10, theuser 11 looks up the image on his/hercomputer 15 and transfers theimage data 10 to animage processing operator 21. Theimage processing operator 21 loads theimage data 10 set on hiscomputer 20 where special image modeling software is available. Theimage processing operator 21 reads the instructions sent by theuser 11 to understand what type of model is required. If theimage processing operator 21 still has questions or needs additional information, he will communicate with theuser 11 to get the information. Theimage processing operator 21 then starts the process to create a 3Ddigital model 22 from theimage data sets 10 on hiscomputer 20 using specialized modeling software. The creation of the 3Ddigital model 22 requires a trainedoperator 21, specialized imaging software, and a significant amount of user interaction. Theimage processing operator 21 needs to communicate frequently with theuser 11 who has ordered the physical model to understand the requirements and applications of the model. The image processing operator is also required to spend a significant amount of time to perform image segmentation and to trace manually certain image areas. After the 3Ddigital model 22 has been created, it is then saved to a file format supported by a 3D printer orrapid prototyping machine 30, for example, the STL (stereolithography) file format. The digital model file is then sent to the 3D printer orrapid prototyping machine 30 to generate a 3Dphysical model 35. The three-dimensional (“3D”)printer 30 is likely located at a different location and operated by a3D printing operator 31. When thephysical model 35 is printed or fabricated, the3D printing operator 31 sends it to theimaging processing operator 21 who then sends the finishedphysical model 35 back to theoriginal user 11. The present 3D printing techniques are complex and cost ineffective. The physical models may take too long to create to be useful, for example, to an emergency-care doctor. - As a particular example of the need for an efficient 3D physical model printing process, we consider 3D printing applications in the medical field. In a typical application of 3D printing techniques in the medical field, medical images are first ordered and acquired on a hospital computer by a doctor. The doctor then sends the images to a trained image processing operator to create a digital model. The image processing operator communicates with the doctor to understand the requirements for the model. The image processing operator loads the image data set into a 3D image processing software to identify features such as bones, tissues, etc. by using image segmentation software tools. Because image processing of medical data is complex and time-consuming, it remains a challenging task even to a professional image processing operator.
- After loading the image data, the
image processing operator 21 then creates a digital 3D model, for example, a 3D polygonal surface model by using software-based modeling tools. As an example, one commercially available software solution, “3D-DOCTOR”, can be used to produce 3D digital models of anatomical structures, as described in Yecheng Wu, From CT Image to 3D Model, Advanced Imaging, August 2001, 20-23. After creating the digital 3D model, theimage processing operator 21 sends the digital model to a 3D printing service provider. The3D printing operator 30 at the 3D printing service provider loads the digital model data on his computer, controls the 3D printer to produce a physical model, and then delivers the finished physical model to the doctor who ordered the model. The above-described process is user intensitive and requires operators to possess advanced software training, knowledge of the intended applications, and a good understanding of the difference between image data formats and the various data formats supported by 3D printer and rapid prototyping machines. - In the above described process, one procedure employed in image processing is image segmentation. Image segmentation refers to the delineation and labeling of specific image regions in an image data set that defines distinct structures. Image segmentation may include steps such as differentiating a particular structure from adjacent material having different composition and identifying distinct objects having the same or similar composition. For example, when constructing bone models from Computerized Tomography (“CT”) and/or Magnetic Resonance (“MR”) images, bony structures need to be delineated from other structures (soft tissues, blood vessels, etc.) in the images. Also, each bone must typically be separated from adjacent bones when modeling anatomical structures such as cervical spine or foot.
- In 3D printing applications in the medical field, a useful feature is the capability of building a prototype of a patient-specific anatomical region quickly. For example, if a patient comes in with a broken ankle, the surgeon may use a physical model of the bone fragments of the patient to aid surgical planning, if the physical model can be generated rapidly. For orthopedic surgeons, the ability to visualize and manipulate a physical model of a bone or joint in need of repair prior to surgery can aid in the selection and design of surgical implants for fracture fixation or joint replacement. Rapid prototyping of patient specific models increases efficiency and reduces costs by cutting operating room time. Rapid prototyping of patient specific models offers tremendous promise for improved pre-operative planning and preparation. While the technique of sizing surgical implants using newer imaging modalities such as Computerized Tomography (“CT”) and/or Magnetic Resonance (“MR”) imaging is an improvement over standard X-ray films, the ability to work with an accurate physical model of the region of interest would produce further benefits, such as providing tactile 3D feedback of the relevant patient anatomy. Rapid prototyping or 3D printing refers to a collection of technologies for producing physical parts directly from digital descriptions, which frequently are the output from Computer-Aided Design (CAD) software. Rapid prototyping machines have been commercially available since the early 1990's, and the most popular versions involve adding material to build the desired structure layer-by-layer based on a digital three dimensional model of the structure. For example, a physical model may be fabricated using a rapid prototyping system using stereolithography, fused deposition modeling, or three dimensional printing. In stereolithography, a laser is used to selectively cure successive surface layers in a vat of photopolymer. In fused deposition modeling, a thermal extrusion head is used to print molten material (typically a thermoplastic) that fuses onto the preceding layer. A typical three-dimensional printer uses a printer head to selectively deposit binder onto the top layer of a powder bed.
- This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to identify key features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
- The present application discloses systems and methods for single-action printing of 3D physical models from a three or n-dimensional image data set. The methods may be applied to image data set obtained from any of a wide variety of imaging modalities, including Computerized Tomography (“CT”), Magnetic Resonance (“MR”), positron emission tomography (“PET”), optical coherence tomography (“OCT”), ultrasonic imaging, X-ray imaging, sonar, radar including ground penetrating radar, acoustic imaging, microscopy imaging, simulated image data and the like, or combinations of one or more imaging modalities. The systems and methods are applicable to a wide range of applications from creating physical models of anatomical structures such as bones and organs to creating physical models of mechanical components, archaeological sites, and natural geological formations.
- The systems and methods described herein generally contemplate combining printing template methods with a 3D printer or rapid prototyping machine. The printing template methods usually include predefined data processing steps comprising identifying voxels in an image data set, generating a geometric representation, and sending the geometric representation to the 3D printer to produce a 3D physical model. A 3D printer or rapid prototyping machine refers to a collection of devices capable of producing three-dimensional physical parts directly from digital models using stereolithography, fused deposition modeling, three dimensional printing, sheet laminating or other technologies.
-
FIG. 1 illustrates a current method of printing a 3D model from an image data set. -
FIG. 2 illustrates a proposed method for printing a 3D model from an image data set. -
FIG. 3 illustrates a flowchart of an exemplary single-action 3D model printing method. -
FIG. 4 illustrates a flowchart of image data conversion steps included in a printing template. -
FIG. 5 illustrates a list of 3D printing templates accompanied and described with text. -
FIG. 6 illustrates a list of 3D printing templates accompanied and described with text and graphics. -
FIG. 7 illustrates 3D points as an exemplary geometric representation generated by a printing template from image data. -
FIG. 8 illustrates a 3D contour as an exemplary geometric representation generated by a printing template from image data. -
FIG. 9 illustrates a 3D triangle-based surface model as an exemplary geometric representation generated by a printing template from image data. -
FIG. 10 provides a table of sample CT numbers for various human tissues. -
FIG. 11 illustrates an exemplary printing template of printing a bone structure from a CT image data set. -
FIG. 12 illustrates an exemplary printing template of printing a solid body structure from an image data set. -
FIG. 13 illustrates an exemplary printing template of printing a physical model using predefined value ranges from an image data set. -
FIG. 14 illustrates an exemplary process of printing a physical model of selected organs or parts from an image data set. -
FIG. 15 illustrates an exemplary image with seed voxels marked before image segmentation. -
FIG. 16 illustrates a segmentation result of a first round region growing. -
FIG. 17 illustrates a segmentation result of a second round region growing. -
FIG. 18 illustrates a segmentation result of a last round region growing. -
FIG. 19 illustrates a segmented image using a region growing technique. -
FIG. 20 illustrates an example of user adjustable physical model printing method. - Certain specific details are set forth in the following description and figures to provide a thorough understanding of various embodiments of the invention. Certain well-known details often associated with computing and software technology are not set forth in the following disclosure, however, to avoid unnecessarily obscuring the various embodiments of the invention. Further, those of ordinary skill in the relevant art will understand that they can practice the invention without one or more of the details described below. Finally, while various methods are described with reference to steps and sequences in the following disclosure, the description as such is for providing an implementation of embodiments of the invention, and the steps and sequences of steps should not be taken as required to practice this invention.
-
FIG. 2 illustrates an exemplary system using a single-action 3D printing method to print a 3D physical model from an inputimage data set 10. The 3D model can be, for example, a patient-specific anatomical model. First, theimage data set 10 such as CT data, MR data etc, is loaded on acomputer 15. Theimage data set 10 is typically a voxel-based image data set depicting a 3D region with each voxel of theimage data set 10 encoding at least one image attribute, such as image intensity, color or the like. Auser 11 at thecomputer 15 selects oneprinting template 18 from a list of printing templates (300 inFIG. 5 ) for printing a 3D physical model. Thecomputer 15 applies the selectedprinting template 18 to identify voxels in the image data, generate a geometric representation in a data format supported by a 3D printer, and send the data to a connected ornetworked 3D printer 30 for producing aphysical model 35. For example, the3D printer 30 may comprise a rapid prototyping device as discussed above. The3D printer 30 may be connected to the computer directly through a local computer port, local area network, or the Internet. - When a
3D printer 30 is not directly connected to the computer where a printing template is used, the data generated from the printing template may be saved to a storage media (for example, a CD or DVD) or storage device (for example, a external hard drive). The saved data can then be ported to the3D printer 30 to generate thephysical model 35. -
FIG. 3 is a flowchart of a single-action 3D printing method. InFIG. 3 , animage data set 10 is first received instep 210. Instep 218, a selectedprinting template 18 is executed to identify the voxel categories and generate a geometric representation for printing a 3D model. Instep 230, the generated geometric representation is sent to a3D printer 30 and instep 235, a 3Dphysical model 35 is produced. - In
FIG. 3 ,step 218 represents a single user action involved in the printing process of a 3D model. Instep 218, selecting a printing template includes a selection action by using a pointing device to position on a specific printing template from a list of predefined printing templates and select the printing template for execution. The single-action may be a clicking of a mouse button when a cursor is positioned over a predefined area of a displayed list of printing templates or a depressing of a key on a key pad to select a specific printing template. - A printing template as defined herein is a software program for identifying voxels in an image data set, generating a geometric representation of a 3D physical model in a data format supported by a 3D printer, and sending the geometric representation to a 3D printer to create a 3D physical model.
- In general, 3D printers require a geometric representation of an object in order to fabricate the geometric shapes required in making a 3D physical model. The geometric representation of an object may include one or a combination of the following forms: a list of 3D points 501-506 for the entire body of the object with locational and material information defined at each 3D point (
FIG. 7 ), a group of 3D contours 552-561 to define the shape of the object on each image plane (FIG. 8 ), or surface models 580 (FIG. 9 ) consisting of triangles or polygons or surface patches delineating the body of the object. - In the present application, a 3D
physical model 35 may have one or more pieces and one or multiple colors, and may be made of one or multiple materials. The conversion process from input image date set to a geometric representation understood by a 3D printer may be either dependent on or independent of imaging modality or any other image information. A printing template may be implemented as a software program on a computer, a computer processing board, or the controller board of a 3D printer. It may be implemented as but not limited to: a program script file with processing instructions and parameters, a binary executable program with processing instructions and parameters, a dynamically linked library (DLL), an application plug-in, or a printer device driver. A printing template may be implemented as a stand-alone solution or a component of a system used forprinting 3D physical model from image data sets. A printing template program may be loaded locally on a user's computer or reside on a remote server connected through computer network. -
FIG. 4 illustrates a flowchart of the image data conversion step in a printing template. An input image data set 10 is received by aprinting template 15. The printing template starts its predefinedvoxel identifying function 310 to identify voxel categories in the image data. After identifying the voxel categories, the printing template generates one of three geometric representations (3D points 315,3D contours 320 or surface models 325) supported by a 3D printer and sends the generated geometric representation to the3D printer 30 to produce a 3Dphysical model 35. - The image data conversion process generates a geometric representation and any additional data needed for a 3D printer to print out the physical model. The voxel identifying process is generally done using image processing techniques such as image segmentation and classification. One purpose of image segmentation and classification is to identify the voxel categories at each voxel location for the entire image data set. Commonly used image classification techniques include trained classifiers (such as artificial neural networks), image clustering using voxel similarity measures, etc. Commonly used image segmentation techniques include image thresholding, histogram thresholding, region growing, region splitting, watershed method, graph partitioning, clustering, artificial neural network, and other methods.
- The geometric representation generated from the input image data set for 3D printing may be a list of 3D points 501-506 in the body of an object with locational and material information (
FIG. 7 ) specified at each point, or a set of 3D contours 551-561 to define the shape of an object in the image planes (FIG. 8 ), or surface models of an object 580 (FIG. 9 ), or a combination of them. The data generated for 3D printing is not limited to geometrical representations such as points, contours or surfaces as described. The data may also be organized as a list of printing instructions, such as “move to a location”, “deposit a specified amount of building material”, “move to a new location”, etc., that can be used to complete the physical model printing process. - The list of printing templates may be displayed as either text 300 (
FIG. 5 ) or graphics 301 (FIG. 6 ) on thecomputer 15. For example, the text may use a description such as “bone structure” or “brain.” The graphic display may use pre-drawn graphic icons to indicate “bone”, “skull”, or “brain.” The graphic display of a printing template may also use a 3D graphic rendering of the geometric representation generated from the image data by the printing template. The number of printing templates is not limited. Additional printing templates may be added for specific physical model printing needs. New printing templates may be created with different processing steps and parameters. A printing template may also be implemented as a part of the input image; in which case user interaction is not required. For example, when the the input image is received with a specific printing template attached, the printing process starts automatically by executing the attached printing template. The execution of the attached printing template may include steps of first generating a geometric representation from the image and then producing a physical model without any user interaction. - As described above, rapid prototyping systems build a physical model by adding consecutive layers, as opposed to subtractive rapid prototyping or conventional machining that uses a tool to remove material from blank stock. However, generation of a physical model may just as well use other processes and equipment. For example, rapid prototyping processes may be adapted to produce functional objects (“parts”) rather than just geometric models. In such case, rapid prototyping may be referred to by the alternative names such as additive fabrication, layered manufacturing, and solid free form fabrication.
- Many commercial rapid prototyping machines currently employ standard input formats comprising of a polygonal representation of the boundary of the object. For example, a CAD model or other three-dimensional (“3D”) digital model is converted to a list of triangles defining the surface of the object. The machine slices through the collection of triangles to generate a geometric representation that comprises the boundary of each layer to be printed or deposited. In the following sections, different embodiments of 3D printing templates are discussed.
- (1) Printing Bone Structure from CT Image Date Set
- This embodiment is implemented as a printing template for printing a physical model of a bone structure from a CT image data set.
- In a CT image, the intensity value at each voxel may be converted to a value in Hounsfield units (HU). The Hounsfield unit system measures the attenuation coefficient of tissues in computerized tomography. Hounsfield units are also termed CT numbers.
FIG. 10 provides a table of sample CT numbers for various human tissues. The table lists some of the voxel values of different tissues or materials in Hounsfield units for a typical CT scanner. The values may differ on a different CT image scanner due to specific settings on that particular imaging device and custom calibrations of image data. The formula to calculate the CT number in Hounsfield units from the voxel intensity is normally provided as part of the image data. For example, the formula used by many CT scanner vendors is: -
HU=Voxel Intensity*Scale+Intercept; - where HU is the voxel value in Hounsfield units, Voxel Intensity is the attribute value of each voxel provided in an image data set, and Scale and Intercept are parameters provided with the formula. For example, for many CT images, Scale=1 and Intercept=−1000. Other values for Scale and Intercept may also be used.
- As indicated in the table of
FIG. 10 , bone tissues may be identified using a range of CT numbers (>1000). The value of every voxel in the image data set can be checked to identify bone tissues. For example, if a voxel has a value above 1000 HU, it is marked as bone tissue. Often an upper limit is used to prevent other hard materials such as metal implants from being marked as bone tissue. A similar technique may be applied to other tissues, such as soft tissue (fat, muscle, etc), blood, liver tissue, and white and grey matter in the brain. -
FIG. 11 is a flowchart of an exemplary printing template for printing bone structures from a CT image set. The printing template identifies voxels that are part of the bone structure in theCT image 600, generates a geometric representation (in the format of3D points 605,Contours 3D printer 608 to generate aphysical model 610. The printing template includes the following processing steps: - a) Go through the entire
image data set 600 to check the HU value of each voxel (Step 601).
b) For each voxel with a HU value larger than 1000 HU but less than an upper bound, mark the voxel withvalue 1 to indicate the voxel as representing bone tissue (Step 602). Otherwise, mark it withvalue 0 to indicate non-bone tissue (Step 603). Repeat Step “a” and “b” until all voxels are checked, in which case a geometric representation is generated (Step 604). Thevalue 1000 HU is used here as an example. Different values or ranges may be used for different images.
c) If the 3D printer (608) supports input data in the format of 3D points, a geometric representation comprising a list of 3D points for all voxels marked withvalue 1 may be generated (Step 605) and sent to the printer (Step 608) to generate aphysical model 610. If other information such as material or color is supported, we may include the other information in the geometric representation.FIG. 7 shows an example of the 3D points generated from an image data set. In this example, every 3D point has an identification value which is either 0 or 1. In this case, 0 indicates non-bone tissue and the voxels withvalue 0 are represented here by a white color. 1 indicates bone tissue and the voxels withvalue 1 are represented here by a dark color. In other embodiments, every voxel may have one or multiple identification values which may be any value, not limited to 0 or 1.
InFIG. 7 , the list of 3D points are represented as: - where each point has a data format of (X, Y, Z, Value). X, Y, Z are the three-dimensional coordinates of a voxel and Value is the attribute with a value of, in this case, 1 for all the voxels identified as bone tissue and 0 otherwise. Other values may be used for identification purposes. Additional values may be also used to indicate attributes such as color or material.
d) If the 3D printer supports input data in the format of 3D contours, a geometric representation comprising the contours are generated by tracing along the outer edge of all voxels marked with the value of 1 (Step 606). The contour tracing method is straight forward, and is normally done by walking along the edge voxels in a fixed order within each 2D image plane. For example, we may start the walk on an edge voxel and follow the next edge voxel in a clockwise fashion until the starting position is encountered. The walking process is then repeated for all image planes.FIG. 8 shows an example of tracing a contour in a 2D image plane. In this example, the tracing process starts at oneedge voxel 551 and the contour starts with no point data.Voxel 551 is added to the contour as the starting point. In a clockwise order, the next voxel on the edge to be traced isvoxel 552.Voxel 552 is added to the contour. Repeat the process to addvoxels - where each point has a data format of (X, Y, Z, Value). X, Y, Z are the three-dimensional coordinates of a voxel and Value is the attribute with a value of, in this case, 1 for all the voxels identified as bone tissue and 0 otherwise. Other values may be used for the attribute and additional attributes such as color or material may be included as well. In this example, the particular tracing technique is described as an example. Other tracing methods and variations may be used to generate similar results.
e) If the 3D printer supports input data in the format of a surface model, then a geometric representation in the format of a surface model is generated using the “Marching Cubes” (U.S. Pat. No. 4,710,876, 4,751,643, 4,868,748) method or other surface modeling methods (Step 607). The generated geometric representation is sent to the 3D printer to produce a physical model 610 (Step 608). Most commercially available 3D printers and rapid prototyping machines support the “STL” format, which stores surface geometry data as a set of raw unstructured triangles. For this example, thesurface model 607 is sent to the the three-dimensional (“3D”) printer in the “STL” format. - “Marching cubes” is a computer graphics algorithm for extracting a polygonal mesh of an isosurface from three-dimensional voxels. The algorithm proceeds through the voxels marked with 1, taking eight neighbor locations at a time (thus forming an imaginary cube) and then determining the polygon(s) needed to represent the part of the isosurface that passes through this cube. The individual polygons are then fused into the desired surface. The “Marching Cubes” algorithm generates triangle-based surface models. Additional post processing steps such as surface smoothing and surface decimation may be applied to improve the surface quality but are not required.
-
FIG. 9 shows an example of a three dimensional triangle-basedsurface model 580. In this example, the triangle-based surface model has 8 vertexes: P0, P1, P2, P3, P4, P5, P6, P7 and 12 surface triangles with T1, T2, T3, T4, T5, T6 displayed at the front of the model and T7, T8, T9, T10, T11, T12 displayed at the back of the model. Each vertex is a 3D point: (X, Y, Z). Each triangle has 3 vertexes, for example (P0, P2, P1). This surface model may be represented as: - In this example, the surface model representation is similar to the commonly used “STL” format and may be sent to the the 3D printer in the “STL” format for printing a physical model. Other representations and variations, such as surface patches or polygon-based surfaces, may also be used.
- The above example describes one embodiment of the single-
action 3D image printing methods. The steps in the printing template may be combined or varied. For example, the voxel checking and marking Steps “a” and “b” can be combined into Step “e” that checks the voxel values and generates the surface triangles without marking the voxels. - (2) Printing Solid Body Structure from an Image Data Set
- This embodiment is implemented as a printing template for printing a physical model of a solid body from an image data set.
- For a known imaging modality, such as Computerized Tomography (“CT”) or Magnetic Resonance (“MR”) imaging, the voxels in an empty or no-tissue region in an image typically have a known value range. For example, air would be considered a no-tissue region. A voxel representing air has a value range around −1000 HU as shown in the CT values table (
FIG. 10 ). In other words, we can check the value of each voxel in the image data set to identify whether the voxel represents an empty region or not. For example, if a voxel in a CT image has a value between −1000 HU and −200 HU (the value below the lowest tissue value in Hounsfield unit), the voxel may be identified as air. Otherwise the voxel may be identified as body tissue. The same method may be applied to other imaging modalities to identify empty regions that are defined with known voxel value ranges. -
FIG. 12 illustrates an exemplary printing template for printing a solid body structure from an image data set. The printing template identifies voxels (Step 621) in empty regions and body regions in the image, generates a geometric representation (Steps - a) Go through the entire
image data set 620 to check the value of each voxel in Hounsfield units (Step 621).
b) For each voxel, if its value is within the value range of no-tissue (empty region, for example, air), mark the voxel withvalue 0 to indicate it is empty (Step 623). Otherwise, mark it withvalue 1 to indicate it has tissue (Step 622). Repeat Step “a” and “b” until all voxels are checked, in which case, a geometric representation is generated (Step 624).
c) If the output printing device supports input data in the format of 3D points, we then generate a list of 3D points for all voxels marked with value 1 (Step 625) and send the list to the printer to generate a physical model 629 (Step 628). If other information such as material or color is supported, we can extract such information from the input image data and send it together with the geometric representation in the format of 3D points. SeeFIG. 7 for an example of the 3D points generated from an image data set.
d) If the 3D printer supports input data in the format of contours, then a geometric representation may be generated by tracing the contours along the outer edge of all voxels marked with value 1 (Step 626). SeeFIG. 8 for an example of tracing a contour in a 2D image plane.
e) If the 3D printer supports input data in the format of a surface model, then a geometric representation in the format of a surface model is generated (Step 627) using the “Marching Cubes” (U.S. Pat. No. 4,710,876, 4,751,643, 4,868,748) method or other surface modeling methods. The geometric representation is then sent to the 3D printer to produce a physical model 629 (Step 628). - The above example describes one embodiment of the single-
action 3D image printing methods. The steps in the printing template may be combined or varied, for example, the voxel checking and marking Steps “a” and “b” can be combined into Step “e” that checks the voxel values and generates the surface triangles without marking the voxels. - This embodiment is implemented as a printing template for printing a physical model from an image data set using predefined voxel value ranges.
-
FIG. 13 illustrates an exemplary printing template for printing a physical model from animage data set 630 using one or more predefined voxel value ranges. A predefined value range may be in voxel intensity, color, texture, location, region, or any derived value from them. A typical range has a low value and a high value to define the bounds of the range. A list of ranges may be used to define multiple value ranges that are not adjacent to each other. - In this embodiment, the method identifies voxels using the predefined voxel value range (Step 631), generates a geometric representation (
Steps 635, 636, or 637), and sends the data to the 3D printer to generate a physical model 639 (Step 638). It includes the following steps: - a) Go through the entire image data set to check the value of each voxel against the value ranges defined in the printing template (Step 631).
b) For each voxel with a value within the bound of one of the defined ranges, mark the voxel withvalue 1 to indicate it is within the specified range (Step 632). Otherwise, mark withvalue 0 to indicate it is outside (Step 633). Repeat Step “a” and “b” until all voxels are checked and identified. A geometric representation is generated inStep 634.
c) If the 3D printer supports input data in the format of three-dimensional points, we generate a geometric representation comprising a list of 3D points for all voxels marked with value 1 (Step 635) and send the geometric representation to the 3D printer (Step 638) to create aphysical model 639. If other information such as material or color are supported in theinput image data 630, we may include the information and send it together with the geometric representation to the printer.FIG. 7 shows an example of the 3D points generated from an image data set.
d) If the 3D printer supports input data in the format of contours, a geometric representation can be generated by tracing the contours along the outer edge of all voxels marked with value 1 (Step 636). SeeFIG. 8 for an example of tracing a contour in a 2D image plane.
e) If the 3D printer supports input data in the format of a surface model, a geometric representation may be generated using the “Marching Cubes” (U.S. Pat. No. 4,710,876, 4,751,643, 4,868,748) method or other surface modeling methods (Step 637). The geometric representation is then sent to the 3D printer (Step 638) to produce aphysical model 639. Most commercially available 3D printers and rapid prototyping machines support the “STL” format, which stores surface geometry data as a set of raw unstructured triangles. In a particular example, thesurface model 637 is sent to the 3D printer in the “STL” format. - The above example describes one embodiment of the single-
action 3D image printing methods. The steps in the printing template may be combined or varied. For example, the voxel checking and marking Steps “a” and “b” can be combined into Step “e” that checks the voxel values and generates the surface triangles without marking the voxels. - This embodiment is implemented as a printing template for printing a physical model of selected organs or parts from an image data set.
-
FIG. 14 illustrates an example of printing aphysical model 649 of selected organs or parts from animage set 640. To generated a physical model of selected organs or parts, the image regions of the selected organs need to be identified using image segmentation techniques (Step 644). A typical segmentation technique used for identifying specific image regions starts with either a set of automatically generated (Step 642) or user selected seed locations or regions (Step 641), grows each region by merging neighboring voxels that are within a certain similarity criterion, and repeats the process until no more neighboring voxels are available for merging. For example, the criterion could be a difference of voxel intensity, gray level, texture, or color between the voxels already identified and the ones being checked. After the identification process is complete, the identified regions are then used to generated a geometric representation for the 3D printer. Other image segmentation methods, such as region growing, active contours, graph partitioning, watershed, and clustering, may be used in the image region identifying step of this embodiment. - In this example, the method identifies voxels using a region growing technique (Step 644), generates a geometric representation of the identified voxels in a format supported by a 3D printer 648 (
Steps - In this embodiment, a user 11 (referenced in
FIG. 2 ) selects some voxels or regions on the input image as seed voxels or regions.FIG. 15 illustrates an example where two selectedseed voxels 701 at location (4, 4, Z) and (5, 4, Z) are marked with dark color. The location is represented by the X, Y, Z coordinates of a voxel.FIG. 15 shows the original image as a grayscale image. The grayscale value of this image is from 0 to 255. Seed voxels are checked (Step 641) to see whether they are available and whether they are within the bound of the image set. If seed voxels are available, continue to Step “b” to start the image segmentation process (Step 644) through region growing. If seed voxels are not available,Step 642 is carried out to generate seed voxels automatically. For example, a predefined value range may be used to select seed voxels within a value range. - b) The image segmentation process through region growing starts at the seed voxel locations. The process grows each region by merging neighboring voxels that are within a certain similarity criterion and repeats the process until no more neighboring voxels are available for merging. In this example, we choose a simple criterion for measuring voxel similarity, that is, for a neighboring voxel to qualify as similar to a reference voxel, the grayscale value difference between the neighboring voxel and the reference voxel must be less than 20. In other words, if the grayscale value difference is less than 20, the neighboring voxel is added to the region and the region grows by one voxel. If the grayscale value difference is equal to or greater than 20, the voxel is not added to the region. For example, voxel 702 (4, 3, Z) in
FIG. 15 has a grayscale value of 103. The difference between this value and the grayscale value (102) of the voxel located at (4, 4, Z) 701 is 1. The difference is within the similarity criterion, so the neighboringvoxel 702 is added to the region.FIG. 16 illustrates the first round of region growing for the image example shown inFIG. 15 . All neighboring voxels with grayscale difference less than 20 are marked with a dark color. 8 voxels 702 (marked with horizontal hatch lines) at locations (4, 3, Z), (5, 3, Z), (3, 4, Z), (6, 4, Z), (3, 5, Z), (4, 5, Z), (5, 5, Z), (6, 5, Z) are added to the region. Here the image plane is assumed to be parallel to the XY plane therefore all voxels have the same Z value. -
FIG. 17 illustrates the result of the second round region growing. 3 voxels 703 (marked with vertical hatch lines) at locations (4, 6, Z), (5, 6, Z), (6, 6, Z) are added to the region. -
FIG. 18 illustrates the result of the last round region growing. 1 voxel 704 (marked with diagonal hatch lines) at locations (5, 7, Z) is added to the region. After this round, no more voxels meet the similarity criteria. The region growing process stops. All voxels added to the region are marked withvalue 1 and the rest of the voxels are marked withvalue 0 as shown inFIG. 19 . - In
FIG. 14 , after the image region of the selected organ has been identified (Step 644), a geometric representation is generated in the format of either 3D points, or contours, or surfaces (Step - The above example describes one embodiment of the single-
action 3D image printing methods in which the image segmentation technique uses a region growing method. Other image segmentation methods may be used to generate similar results. - This embodiment is implemented as a printing template for printing a physical model from an image data set using user adjustable image processing parameters and steps.
-
FIG. 20 illustrates an exemplary printing template that adopts a user adjustable physical model printing method. This printing template provides a user with a list of selectable processing options and adjustable parameters for image segmentation and data conversion. Theuser 11 makes the initial selection of parameters and segmentation methods (Step 651). The segmentation methods may include the ones described in the above examples, such as region growing, image thresholding, graph partitioning, and others. The parameters may include value ranges that are adjustable, user defined seed regions, and others. The image segmentation process (Step 654) segments the input image using the selected methods and parameters. The geometric representation generated from the image segmentation process is used to create a 3D rendering (Step 660) to show how a final physical model may look on a computer. The computer 15 (FIG. 2 ) can be used to display a three-dimensional rendering 660 of the model generated by the image segmentation process. The three-dimensional rendering may be implemented using volume rendering of the segmented image or surface rendering from thesurface model 657. If the rendering meets the user's requirements, the segmented image is converted to a format supported by the3D printer 658 for printing (Steps - Similar to other embodiments as described above, this embodiment further includes the processing steps of converting an
image data set 650 to a geometric representation in a data format supported by the 3D printer and sending the geometric representation to the 3D printer to generate a physical model of selected organs or parts 659 (Step 658). - (6) Extension to n-Dimension
- Although the present method has been described with reference to 3D image data sets, it will be immediately apparent to persons of skill in the art that the methods described above are readily applicable to any number of dimensions. It is contemplated that the methods may be applied to n-dimensional data, where n may be 2, 3, 4 or any number larger than 4. In particular, it is contemplated that the invention may be applied to n-dimensional data in which one of the dimensions is time and there are two or three spatial dimensions. For example, we can use the above described methods to produce multi-dimensional physical models that evolve over time.
- It should be appreciated that the present method greatly reduces the time required for printing physical models from an n-dimensional data set, including a 3D data set. Therefore practical applications capable of producing a series of physical models from time-sequence image data sets to show changes of shape or motions can be implemented. For example, time-
sequence 3D image data of a chest containing a beating heart may be used to generate a series of chest models to show the shape and motion of the beating heart at different time points. - Although the present invention has been described in terms of various embodiments, the invention is not limited to these embodiments. Modification within the spirit of the invention will be apparent to those skilled in the art. For example, various different single-actions can be used to effect the printing of a physical model from an image data set. For example, a voice command may be spoken by the user. A key may be depressed by the user. A button on a 3D printing device may be pressed by the user. Selection using any pointing device may be effected by a user to start the execution of a printing template. Although a single-action may be preceded by multiple physical movements of the user (e.g., moving a mouse so that a mouse pointer is over a button), the single-action generally refers to a single event received by a system that commands the system to print a physical model from an image data set or a derived representation of the image. Finally, various techniques for identifying voxel categories and generating a geometric representation can be used to print a physical model from an image data set.
Claims (4)
1. A printing template method for single-action 3D printing of a selected organ part comprising:
receiving an image data set as input;
identifying voxels in the selected organ part by segmenting the image data set using either user selected or automatically generated seed points;
generating a geometric representation data from the identified voxels; and
sending the generated geometric representation to a 3D printer for printing a physical model.
2. The method of claim 1 , wherein the image data set has a plurality of voxels, each voxel having an intensity value; wherein the image data set comprises images of the three dimensional region obtained using one or more imaging modalities selected from magnetic resonance imaging, computed tomography, ultrasound, X-ray imaging, positron emission tomography, microscopy imaging, simulated image.
3. The method of claim 1 , wherein the identifying of voxels comprises segmenting the image data using either user selected or automatically generated seed points.
4. The method of claim 1 , wherein the geometric representation may be a list of 3D points, a list of 3D contours, a list of 3D polygonal surface models, or a combination of them.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/332,578 US20140330421A1 (en) | 2011-03-02 | 2014-07-16 | Single action three-dimensional model printing methods |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/039,048 US8579620B2 (en) | 2011-03-02 | 2011-03-02 | Single-action three-dimensional model printing methods |
US14/036,620 US8817332B2 (en) | 2011-03-02 | 2013-09-25 | Single-action three-dimensional model printing methods |
US14/332,578 US20140330421A1 (en) | 2011-03-02 | 2014-07-16 | Single action three-dimensional model printing methods |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/036,620 Division US8817332B2 (en) | 2011-03-02 | 2013-09-25 | Single-action three-dimensional model printing methods |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140330421A1 true US20140330421A1 (en) | 2014-11-06 |
Family
ID=46753327
Family Applications (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/039,048 Active 2031-05-30 US8579620B2 (en) | 2011-03-02 | 2011-03-02 | Single-action three-dimensional model printing methods |
US14/036,620 Active US8817332B2 (en) | 2011-03-02 | 2013-09-25 | Single-action three-dimensional model printing methods |
US14/332,578 Abandoned US20140330421A1 (en) | 2011-03-02 | 2014-07-16 | Single action three-dimensional model printing methods |
US14/332,616 Abandoned US20140330418A1 (en) | 2011-03-02 | 2014-07-16 | Single action three-dimensional model printing methods |
US14/332,544 Abandoned US20160210779A9 (en) | 2011-03-02 | 2014-07-16 | Single Action Three-Dimensional Model Printing Methods |
US14/332,555 Abandoned US20160207257A9 (en) | 2011-03-02 | 2014-07-16 | Single Action Three-Dimensional Model Printing Methods |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/039,048 Active 2031-05-30 US8579620B2 (en) | 2011-03-02 | 2011-03-02 | Single-action three-dimensional model printing methods |
US14/036,620 Active US8817332B2 (en) | 2011-03-02 | 2013-09-25 | Single-action three-dimensional model printing methods |
Family Applications After (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/332,616 Abandoned US20140330418A1 (en) | 2011-03-02 | 2014-07-16 | Single action three-dimensional model printing methods |
US14/332,544 Abandoned US20160210779A9 (en) | 2011-03-02 | 2014-07-16 | Single Action Three-Dimensional Model Printing Methods |
US14/332,555 Abandoned US20160207257A9 (en) | 2011-03-02 | 2014-07-16 | Single Action Three-Dimensional Model Printing Methods |
Country Status (1)
Country | Link |
---|---|
US (6) | US8579620B2 (en) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140257518A1 (en) * | 2013-03-08 | 2014-09-11 | The Trustees Of Princeton University | Multi-functional hybrid devices/structures using 3d printing |
CN105469442A (en) * | 2015-12-30 | 2016-04-06 | 武汉金玺银杏工业设计有限责任公司 | Printing method and apparatus |
KR20160068452A (en) * | 2014-12-05 | 2016-06-15 | 주식회사 케이티 | Method for recommending 3d printing according to slicing direction in cloud environment, server and computing device |
WO2016153047A1 (en) * | 2015-03-26 | 2016-09-29 | 学校法人慶應義塾 | Three-dimensional object-manufacturing apparatus, three-dimensional object-manufacturing method, and program |
WO2016171717A1 (en) * | 2015-04-24 | 2016-10-27 | Hewlett-Packard Development Company, L.P. | Three-dimensional object representation |
CN106683550A (en) * | 2017-01-10 | 2017-05-17 | 首都医科大学附属北京友谊医院 | Ear structure model manufacturing method and device thereof |
WO2017084369A1 (en) * | 2015-11-18 | 2017-05-26 | 深圳市艾科赛龙科技股份有限公司 | Osteocyte culturing method |
CN108595893A (en) * | 2018-05-16 | 2018-09-28 | 电子科技大学 | A kind of three-dimensional mechanical Modal Analysis analogy method based on three layers of pretreatment |
US20190039322A1 (en) * | 2017-01-19 | 2019-02-07 | Hewlett-Packard Development Company, L.P. | Build unit control |
US10782217B2 (en) | 2016-07-12 | 2020-09-22 | Deka Products Limited Partnership | System and method for applying force to a device |
US11254901B2 (en) | 2016-07-12 | 2022-02-22 | Deka Products Limited Partnership | System and method for printing tissue |
US11299705B2 (en) | 2016-11-07 | 2022-04-12 | Deka Products Limited Partnership | System and method for creating tissue |
US11530380B2 (en) | 2017-07-12 | 2022-12-20 | Deka Products Limited Partnership | System and method for transferring tissue |
Families Citing this family (115)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10795457B2 (en) | 2006-12-28 | 2020-10-06 | D3D Technologies, Inc. | Interactive 3D cursor |
US11275242B1 (en) | 2006-12-28 | 2022-03-15 | Tipping Point Medical Images, Llc | Method and apparatus for performing stereoscopic rotation of a volume on a head display unit |
US11228753B1 (en) | 2006-12-28 | 2022-01-18 | Robert Edwin Douglas | Method and apparatus for performing stereoscopic zooming on a head display unit |
US9349183B1 (en) * | 2006-12-28 | 2016-05-24 | David Byron Douglas | Method and apparatus for three dimensional viewing of images |
US11315307B1 (en) | 2006-12-28 | 2022-04-26 | Tipping Point Medical Images, Llc | Method and apparatus for performing rotating viewpoints using a head display unit |
US8625889B2 (en) * | 2010-12-30 | 2014-01-07 | Samsung Electronics Co., Ltd. | System for food recognition method using portable devices having digital cameras |
US8579620B2 (en) * | 2011-03-02 | 2013-11-12 | Andy Wu | Single-action three-dimensional model printing methods |
US9183764B2 (en) * | 2011-03-31 | 2015-11-10 | National University Corporation Kobe University | Method for manufacturing three-dimensional molded model and support tool for medical treatment, medical training, research, and education |
US9802364B2 (en) * | 2011-10-18 | 2017-10-31 | 3D Systems, Inc. | Systems and methods for construction of an instruction set for three-dimensional printing of a user-customizableimage of a three-dimensional structure |
US9999509B2 (en) | 2011-11-17 | 2018-06-19 | Stratasys Ltd. | System and method for fabricating a body part model using multi-material additive manufacturing |
JP5814853B2 (en) * | 2012-04-18 | 2015-11-17 | 富士フイルム株式会社 | Stereo model data generation apparatus and method, and program |
US9481134B2 (en) * | 2012-06-08 | 2016-11-01 | Makerbot Industries, Llc | Build platform leveling with tactile feedback |
US20140046170A1 (en) * | 2012-08-07 | 2014-02-13 | Chia-Wei Sun | Brain volumetric measuring method and system using the same |
US8963135B2 (en) | 2012-11-30 | 2015-02-24 | Intel Corporation | Integrated circuits and systems and methods for producing the same |
GB201300171D0 (en) * | 2013-01-07 | 2013-02-20 | Bae Systems Plc | Object production and assessment |
AU2014204284B2 (en) | 2013-01-07 | 2017-06-29 | Bae Systems Plc | Object production using an additive manufacturing process and quality assessment of the object |
WO2014121129A1 (en) * | 2013-01-31 | 2014-08-07 | Sandboxr, Llc | Method and system for 3-d printing product customization |
US20140261430A1 (en) | 2013-03-15 | 2014-09-18 | Lucy Carol Davis | Facial Mask Apparatus and Method of Making |
US9364995B2 (en) | 2013-03-15 | 2016-06-14 | Matterrise, Inc. | Three-dimensional printing and scanning system and method |
US20140324204A1 (en) * | 2013-04-18 | 2014-10-30 | Massachusetts Institute Of Technology | Methods and apparati for implementing programmable pipeline for three-dimensional printing including multi-material applications |
US20160287339A1 (en) * | 2013-04-30 | 2016-10-06 | Universiti Malaya | Method for manufacturing a three-dimensional anatomical structure |
GB2515266B (en) * | 2013-05-09 | 2018-02-28 | Disney Entpr Inc | Manufacturing Process for 3D Printed Objects |
US9104298B1 (en) | 2013-05-10 | 2015-08-11 | Trade Only Limited | Systems, methods, and devices for integrated product and electronic image fulfillment |
US9821517B2 (en) * | 2013-06-26 | 2017-11-21 | Microsoft Technology Licensing, Llc | 3D manufacturing platform |
CN103341989B (en) * | 2013-07-08 | 2015-07-29 | 上海大学 | The comprehensive Regenerated Bone stent forming System and method for be shaped is printed based on 3D |
US9760674B2 (en) | 2013-07-26 | 2017-09-12 | Aetrex Worldwide, Inc. | Systems and methods for generating orthotic device models from user-based data capture |
GB201314421D0 (en) * | 2013-08-12 | 2013-09-25 | Materialise Nv | Data Processing |
CN104375106B (en) * | 2013-08-14 | 2017-05-31 | 西门子(深圳)磁共振有限公司 | The method for shimming and shimming system of a kind of MR imaging apparatus |
US9636871B2 (en) * | 2013-08-21 | 2017-05-02 | Microsoft Technology Licensing, Llc | Optimizing 3D printing using segmentation or aggregation |
DE102013110699A1 (en) * | 2013-09-27 | 2015-04-02 | Helmut Hans Boris Pende | Procedures for replicating bones and planning a surgical procedure |
KR102133843B1 (en) * | 2013-10-31 | 2020-07-14 | 엘지전자 주식회사 | Apparatus and Method for Head Mounted Display indicating process of 3D printing |
US10022914B2 (en) * | 2013-11-12 | 2018-07-17 | Adobe Systems Incorporated | Method and apparatus for automatically adding utility holes to printable 3-dimensional models |
KR20160088902A (en) * | 2013-11-20 | 2016-07-26 | 포비아, 인크. | Volume rendering polygons for 3-d printing |
US9873229B2 (en) * | 2013-11-21 | 2018-01-23 | Hankookin, Inc. | Three-dimensional object development |
CN103707507B (en) * | 2013-12-13 | 2016-08-17 | 吉林大学 | The 3D of Polyether-ether-ketobiomimetic biomimetic artificial bone prints manufacture method |
CN103679808A (en) * | 2013-12-24 | 2014-03-26 | 通号通信信息集团有限公司 | Method and system for rebuilding three-dimensional head model by two-dimensional nuclear magnetic images |
US20150190970A1 (en) * | 2014-01-03 | 2015-07-09 | Michael Itagaki | Texturing of 3d medical images |
DE102014200915A1 (en) | 2014-01-20 | 2015-07-23 | Siemens Aktiengesellschaft | A method of providing a spatial anatomical model of a body part of a patient |
TWI514252B (en) * | 2014-02-11 | 2015-12-21 | Xyzprinting Inc | Method and device for composing three-dimension model |
US9669585B2 (en) | 2014-02-11 | 2017-06-06 | Adobe Systems Incorporated | Method and apparatus for embedding a 2-dimensional image in a 3-dimensional model |
US10239229B2 (en) * | 2014-02-18 | 2019-03-26 | Halliburton Energy Services, Inc. | System and method for generating formation cores with realistic geological composition and geometry |
US9946816B2 (en) | 2014-03-18 | 2018-04-17 | Palo Alto Research Center Incorporated | System for visualizing a three dimensional (3D) model as printed from a 3D printer |
US10061870B2 (en) | 2014-03-18 | 2018-08-28 | Palo Alto Research Center Incorporated | Automated metrology and model correction for three dimensional (3D) printability |
US9747394B2 (en) * | 2014-03-18 | 2017-08-29 | Palo Alto Research Center Incorporated | Automated design and manufacturing feedback for three dimensional (3D) printability |
US9079440B1 (en) * | 2014-03-31 | 2015-07-14 | Xerox Corporation | System for detecting inoperative inkjets in printheads ejecting clear ink using a light transmitting substrate |
CN112588116A (en) * | 2014-05-13 | 2021-04-02 | 美国安进公司 | Process control system and method for filter and filtering process |
WO2015195127A1 (en) * | 2014-06-19 | 2015-12-23 | Halliburton Energy Services, Inc. | Forming facsimile formation core samples using three-dimensional printing |
CN104044273B (en) * | 2014-06-26 | 2017-05-03 | 珠海天威飞马打印耗材有限公司 | Calibration method and device of three-dimensional printer |
US10252466B2 (en) | 2014-07-28 | 2019-04-09 | Massachusetts Institute Of Technology | Systems and methods of machine vision assisted additive fabrication |
US9862149B2 (en) | 2014-08-29 | 2018-01-09 | Microsoft Technology Licensing, Llc | Print bureau interface for three-dimensional printing |
US20160070822A1 (en) * | 2014-09-09 | 2016-03-10 | Primesmith Oy | Method, Apparatus and Computer Program Code for Design and Visualization of a Physical Object |
TWI526990B (en) * | 2014-09-15 | 2016-03-21 | 三緯國際立體列印科技股份有限公司 | Image processing method for transforming 2d image into 3d model |
US9895836B2 (en) * | 2014-09-16 | 2018-02-20 | The Boeing Company | Fused filament fabricated part using multi-directional layering |
US9626608B2 (en) | 2014-12-01 | 2017-04-18 | Savannah River Nuclear Solutions, Llc. | Additive manufactured serialization |
WO2016101071A1 (en) * | 2014-12-22 | 2016-06-30 | Dental Wings Inc. | Pre-forms and methods for using same in the manufacture of dental prostheses |
TWI628208B (en) * | 2015-01-12 | 2018-07-01 | 國立台灣科技大學 | Method of stereolithography fabrication and photo-curing photosensitive resin |
US10688727B2 (en) * | 2015-01-30 | 2020-06-23 | Hewlett-Packard Development Company, L.P. | Control data based on sub-regions with non-variable object properties |
CA3018919A1 (en) * | 2015-03-25 | 2016-09-29 | Zaxis Labs | Mixed reality simulation system and method for medical procedure planning |
US20160284122A1 (en) * | 2015-03-26 | 2016-09-29 | Intel Corporation | 3d model recognition apparatus and method |
JP6514013B2 (en) | 2015-04-13 | 2019-05-15 | ローランドディー.ジー.株式会社 | Slice model generation device and 3D modeling system |
EP3230809B1 (en) * | 2015-04-24 | 2023-03-15 | Hewlett-Packard Development Company, L.P. | Method for generating three dimensional object models for an additive manufacturing process |
EP3292487B1 (en) * | 2015-04-24 | 2022-10-05 | Hewlett-Packard Development Company, L.P. | Three dimensional object data |
WO2016175831A1 (en) * | 2015-04-30 | 2016-11-03 | Hewlett-Packard Development Company, L.P. | Object data representations for additive manufacturing |
GB2555268A (en) * | 2015-06-12 | 2018-04-25 | Chand Mathur Ashok | Method and apparatus of very much faster 3D printer |
FR3038254B1 (en) * | 2015-06-30 | 2021-02-12 | Inria Inst Nat Rech Informatique & Automatique | DATA PROCESSING DEVICE FOR ADDITIVE MANUFACTURING |
GB201512304D0 (en) * | 2015-07-13 | 2015-08-19 | Whispering Gibbon Ltd | Preparing a polygon mesh for printing |
GB2542552A (en) * | 2015-08-11 | 2017-03-29 | Naotech Solutions S L | Methods, systems and apparatus for construction of three-dimensional objects |
US9824491B2 (en) | 2015-09-09 | 2017-11-21 | Siemens Healthcare Gmbh | Data driven framework for optimizing artificial organ printing and scaffold selection for regenerative medicine |
US10492761B2 (en) * | 2015-10-14 | 2019-12-03 | General Electric Company | Utilizing depth from ultrasound volume rendering for 3D printing |
US10157500B2 (en) | 2015-10-14 | 2018-12-18 | General Electric Company | Utilizing depth from ultrasound volume rendering for 3D printing |
US10733911B2 (en) | 2015-10-14 | 2020-08-04 | Humanetics Innovative Solutions, Inc. | Three-dimensional ribs and method of three-dimensional printing of ribs for crash test dummy |
US10636184B2 (en) | 2015-10-14 | 2020-04-28 | Fovia, Inc. | Methods and systems for interactive 3D segmentation |
US11222153B2 (en) | 2015-10-30 | 2022-01-11 | Hewlett-Packard Development Company, L.P. | Three-dimensional object generation parameter descriptions |
KR102000486B1 (en) * | 2016-03-03 | 2019-07-17 | 한국전자통신연구원 | Apparatus and Method for Generating 3D Printing Model using Multiple Texture |
KR101660371B1 (en) * | 2016-03-24 | 2016-09-27 | 삼성메디슨 주식회사 | Apparatus and method for processing medical images and computer-readable recording medium |
US10232649B2 (en) * | 2016-04-05 | 2019-03-19 | Océ Holding B.V. | Printing system for printing an object having a surface of varying height |
US10259164B2 (en) | 2016-06-22 | 2019-04-16 | Massachusetts Institute Of Technology | Methods and apparatus for 3D printing of point cloud data |
EP3273430A1 (en) * | 2016-07-19 | 2018-01-24 | Humanetics Innovative Solutions, Inc. | Three-dimensional components and method of three-dimensional printing of components for crash test dummy |
US10157408B2 (en) | 2016-07-29 | 2018-12-18 | Customer Focus Software Limited | Method, systems, and devices for integrated product and electronic image fulfillment from database |
EP3500231A1 (en) * | 2016-08-19 | 2019-06-26 | The Procter and Gamble Company | Polymeric materials and articles manufactured there from |
WO2018048545A1 (en) * | 2016-09-07 | 2018-03-15 | The Procter & Gamble Company | Polymeric materials and articles manufactured there from |
US10872177B2 (en) * | 2016-09-21 | 2020-12-22 | New York University | Embedded security elements for digital models used in additive manufacturing |
US11138790B2 (en) | 2016-10-14 | 2021-10-05 | Axial Medical Printing Limited | Method for generating a 3D physical model of a patient specific anatomic feature from 2D medical images |
GB201617507D0 (en) * | 2016-10-14 | 2016-11-30 | Axial3D Limited | Axial3D UK |
CN106558097B (en) * | 2016-10-15 | 2019-06-14 | 合肥市勘察院有限责任公司 | A kind of underground environment perspective three dimensional method for establishing model |
US11804305B2 (en) * | 2016-10-26 | 2023-10-31 | Mosaic Ventures, Llc | Contralateral image orthopedic implant |
US10520920B2 (en) * | 2016-10-27 | 2019-12-31 | Océ Holding B.V. | Printing system for printing an object having a surface of varying height |
US10777324B2 (en) | 2016-11-03 | 2020-09-15 | The University Of Akron | Engineered 3D lung airway tree |
US10275909B2 (en) * | 2016-11-23 | 2019-04-30 | 3Dsystems, Inc. | Systems and methods for an integrated system for visualizing, simulating, modifying and 3D printing 3D objects |
EP3554798B1 (en) | 2016-12-16 | 2020-12-02 | Massachusetts Institute of Technology | Adaptive material deposition for additive manufacturing |
CN106780743B (en) * | 2016-12-19 | 2018-06-29 | 武汉大学 | It is a kind of remotely to match spoon method and system |
US10838401B2 (en) * | 2017-01-25 | 2020-11-17 | Hewlett-Packard Development Company, L.P. | Producing instructions that control three-dimensional printing from voxels |
US10647061B2 (en) | 2017-05-12 | 2020-05-12 | Lawrence Livermore National Security, Llc | System and method for computed axial lithography (CAL) for 3D additive manufacturing |
CN107330244B (en) * | 2017-06-01 | 2020-10-30 | 福建师范大学 | Arch morphology analysis method and computer-readable storage medium |
US10703549B2 (en) * | 2017-06-30 | 2020-07-07 | The Procter And Gamble Company | Water soluble containers and methods of making them |
CN109242949B (en) * | 2017-07-11 | 2023-04-07 | 周武增 | Intelligent 3D printing system and method |
EP3655238A4 (en) * | 2017-07-19 | 2021-01-20 | Hewlett-Packard Development Company, L.P. | Three-dimensional surface normal determination |
US10248971B2 (en) | 2017-09-07 | 2019-04-02 | Customer Focus Software Limited | Methods, systems, and devices for dynamically generating a personalized advertisement on a website for manufacturing customizable products |
CN107945877B (en) * | 2017-10-19 | 2021-06-25 | 福建师范大学 | Arch bar system structure construction method based on arch living body bone inertia main shaft |
EP3513730A1 (en) * | 2018-01-18 | 2019-07-24 | Koninklijke Philips N.V. | System and method for image decomposition of a projection image |
US11723875B2 (en) * | 2018-04-10 | 2023-08-15 | The Procter & Gamble Company | Polymeric materials and articles manufactured there from |
CN108986119B (en) * | 2018-07-25 | 2020-07-28 | 京东方科技集团股份有限公司 | Image segmentation method and device, computer equipment and readable storage medium |
CN109001237B (en) * | 2018-09-25 | 2019-05-17 | 长安大学 | A kind of rock mass internal structure production method based on CT scan and 3D printing |
CN109614695A (en) * | 2018-12-10 | 2019-04-12 | 可脉检测(南京)有限公司 | A kind of method and its application for analyzing fracture micromorphology by 3D printing technique |
GB201900437D0 (en) | 2019-01-11 | 2019-02-27 | Axial Medical Printing Ltd | Axial3d big book 2 |
CN109872386B (en) * | 2019-01-18 | 2020-10-30 | 北京理工大学 | STL model-based product physical property modeling method for accurate voxelization |
CN113840880A (en) * | 2019-05-22 | 2021-12-24 | 汉高股份有限及两合公司 | Method for cleaning and post-processing 3D printed photocured silicone |
JP7400259B2 (en) * | 2019-08-14 | 2023-12-19 | 富士フイルムビジネスイノベーション株式会社 | 3D shape data generation device, 3D printing device, and 3D shape data generation program |
US11094116B2 (en) | 2019-11-18 | 2021-08-17 | GE Precision Healthcare LLC | System and method for automatic generation of a three-dimensional polygonal model with color mapping from a volume rendering |
US20230112951A1 (en) * | 2020-03-09 | 2023-04-13 | Dignity Health | Systems and methods for a neurosurgical simulation model for surgical training |
GB202101908D0 (en) | 2021-02-11 | 2021-03-31 | Axial Medical Printing Ltd | Axial3D pathology |
EP4098219A1 (en) * | 2021-06-04 | 2022-12-07 | Inteneural Networks Inc. | Machine learning based systems and methods for creating personalized endovascular stents and stent grafts |
IT202100016277A1 (en) * | 2021-06-22 | 2022-12-22 | Univ Degli Studi Milano | METHOD FOR THE MANUFACTURING OF ANATOMIC MODELS ABLE TO SIMULATE ORGANS OR PARTS OF ORGANS OF A PATIENT |
KR20230081466A (en) * | 2021-11-30 | 2023-06-07 | 주식회사 애니캐스팅 소프트웨어 | System and Method of predicting microstructure characteristics of casting parts |
CN116503429B (en) * | 2023-06-28 | 2023-09-08 | 深圳市华海天贸科技有限公司 | Model image segmentation method for biological material 3D printing |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080261165A1 (en) * | 2006-11-28 | 2008-10-23 | Bob Steingart | Systems for haptic design of dental restorations |
US20110222081A1 (en) * | 2010-03-15 | 2011-09-15 | Chen Yi | Printing Three-Dimensional Objects Using Hybrid Format Data |
US20120041446A1 (en) * | 2006-02-06 | 2012-02-16 | Conformis, Inc. | Patient Selectable Joint Arthroplasty Devices and Surgical Tools Incorporating Anatomical Relief |
US8412588B1 (en) * | 2010-09-24 | 2013-04-02 | Amazon Technologies, Inc. | Systems and methods for fabricating products on demand |
US20140052285A1 (en) * | 2011-03-03 | 2014-02-20 | Cornell University | Method for specifying and fabricating an object, associated apparatus, and applications |
Family Cites Families (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5184307A (en) * | 1988-04-18 | 1993-02-02 | 3D Systems, Inc. | Method and apparatus for production of high resolution three-dimensional objects by stereolithography |
US5636334A (en) * | 1994-01-28 | 1997-06-03 | Casio Computer Co., Ltd. | Three-dimensional image creation devices |
US6208347B1 (en) * | 1997-06-23 | 2001-03-27 | Real-Time Geometry Corporation | System and method for computer modeling of 3D objects and 2D images by mesh constructions that incorporate non-spatial data such as color or texture |
US6117061A (en) * | 1997-07-21 | 2000-09-12 | Avery Dennison Corporation | Method for custom printing and forming three-dimensional structures |
JP2001166810A (en) * | 1999-02-19 | 2001-06-22 | Sanyo Electric Co Ltd | Device and method for providing solid model |
JP2001165638A (en) * | 1999-02-19 | 2001-06-22 | Sanyo Electric Co Ltd | Device and method for providing three-dimensional model |
JP4979868B2 (en) * | 1999-12-29 | 2012-07-18 | オルムコ コーポレイション | Custom orthodontic appliance forming method and apparatus |
US7300619B2 (en) * | 2000-03-13 | 2007-11-27 | Objet Geometries Ltd. | Compositions and methods for use in three dimensional model printing |
US6772026B2 (en) * | 2000-04-05 | 2004-08-03 | Therics, Inc. | System and method for rapidly customizing design, manufacture and/or selection of biomedical devices |
US7050876B1 (en) * | 2000-10-06 | 2006-05-23 | Phonak Ltd. | Manufacturing methods and systems for rapid production of hearing-aid shells |
US6654656B2 (en) * | 2001-03-06 | 2003-11-25 | The Research Foundation Of State University Of New York | Rapid informational prototypes, including rapid colored prototypes |
US7034820B2 (en) * | 2001-12-03 | 2006-04-25 | Canon Kabushiki Kaisha | Method, apparatus and program for processing a three-dimensional image |
JP3744442B2 (en) * | 2002-03-04 | 2006-02-08 | ソニー株式会社 | Authentication system, authentication method, authentication medium manufacturing apparatus, and authentication terminal apparatus |
US7149596B2 (en) * | 2004-01-13 | 2006-12-12 | Sensable Technologies, Inc. | Apparatus and methods for modifying a model of an object to enforce compliance with a manufacturing constraint |
US7625198B2 (en) * | 2004-08-11 | 2009-12-01 | Cornell Research Foundation, Inc. | Modular fabrication systems and methods |
US7654826B2 (en) * | 2004-10-04 | 2010-02-02 | Solid Terrain Modeling | Three-dimensional cartographic user interface system |
JP2006202083A (en) * | 2005-01-21 | 2006-08-03 | Seiko Epson Corp | Image data creation apparatus and printer |
WO2007030132A2 (en) * | 2005-02-14 | 2007-03-15 | Mayo Foundation For Medical Education And Research | Electronic stool subtraction in ct colonography |
JP4613636B2 (en) * | 2005-02-21 | 2011-01-19 | セイコーエプソン株式会社 | Print data output device and print data output method |
JP3848354B1 (en) * | 2005-08-31 | 2006-11-22 | 博史 大塚 | X-ray image processing system and method |
GB2455926B (en) * | 2006-01-30 | 2010-09-01 | Axellis Ltd | Method of preparing a medical restraint |
US20090149977A1 (en) * | 2007-11-06 | 2009-06-11 | Schendel Stephen A | Methods, systems, and computer program products for shaping medical implants directly from virtual reality models |
WO2009094576A1 (en) * | 2008-01-23 | 2009-07-30 | Sensable Technologies, Inc. | Haptically enabled dental modeling system |
US8391642B1 (en) * | 2008-05-12 | 2013-03-05 | Hewlett-Packard Development Company, L.P. | Method and system for creating a custom image |
US8249732B2 (en) * | 2008-06-26 | 2012-08-21 | Siemens Product Lifecycle Management Software Inc. | System and method for developing automated templates for knowledge capture |
US7983792B2 (en) * | 2008-10-22 | 2011-07-19 | Xerox Corporation | Method and system for the production of variable-dimensional printed substrates |
US9493024B2 (en) * | 2008-12-16 | 2016-11-15 | Xerox Corporation | System and method to derive structure from image |
CN101833426A (en) * | 2009-03-13 | 2010-09-15 | 鸿富锦精密工业(深圳)有限公司 | Printing system and printing method thereof |
US8970900B2 (en) * | 2009-10-27 | 2015-03-03 | Coveroo, Inc. | Automated layout and design for recording text and images in any of disparate three-dimensional objects |
US8565909B2 (en) * | 2010-02-24 | 2013-10-22 | Disney Enterprises, Inc. | Fabrication of materials with desired characteristics from base materials having determined characteristics |
US8527244B2 (en) * | 2010-08-04 | 2013-09-03 | Medizinische Hochschule Hannover (MMH) | Generating model data representing a biological body section |
US8414280B2 (en) * | 2010-08-18 | 2013-04-09 | Makerbot Industries, Llc | Networked three-dimensional printing |
JP2012160039A (en) * | 2011-02-01 | 2012-08-23 | Fujifilm Corp | Image processor, stereoscopic image printing system, image processing method and program |
US8579620B2 (en) * | 2011-03-02 | 2013-11-12 | Andy Wu | Single-action three-dimensional model printing methods |
-
2011
- 2011-03-02 US US13/039,048 patent/US8579620B2/en active Active
-
2013
- 2013-09-25 US US14/036,620 patent/US8817332B2/en active Active
-
2014
- 2014-07-16 US US14/332,578 patent/US20140330421A1/en not_active Abandoned
- 2014-07-16 US US14/332,616 patent/US20140330418A1/en not_active Abandoned
- 2014-07-16 US US14/332,544 patent/US20160210779A9/en not_active Abandoned
- 2014-07-16 US US14/332,555 patent/US20160207257A9/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120041446A1 (en) * | 2006-02-06 | 2012-02-16 | Conformis, Inc. | Patient Selectable Joint Arthroplasty Devices and Surgical Tools Incorporating Anatomical Relief |
US20080261165A1 (en) * | 2006-11-28 | 2008-10-23 | Bob Steingart | Systems for haptic design of dental restorations |
US20110222081A1 (en) * | 2010-03-15 | 2011-09-15 | Chen Yi | Printing Three-Dimensional Objects Using Hybrid Format Data |
US8412588B1 (en) * | 2010-09-24 | 2013-04-02 | Amazon Technologies, Inc. | Systems and methods for fabricating products on demand |
US20140052285A1 (en) * | 2011-03-03 | 2014-02-20 | Cornell University | Method for specifying and fabricating an object, associated apparatus, and applications |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9517128B2 (en) * | 2013-03-08 | 2016-12-13 | The Trustees Of Princeton University | Multi-functional hybrid devices/structures using 3D printing |
US20140257518A1 (en) * | 2013-03-08 | 2014-09-11 | The Trustees Of Princeton University | Multi-functional hybrid devices/structures using 3d printing |
KR102332927B1 (en) * | 2014-12-05 | 2021-11-30 | 주식회사 케이티 | Method for recommending 3d printing according to slicing direction in cloud environment, server and computing device |
KR20160068452A (en) * | 2014-12-05 | 2016-06-15 | 주식회사 케이티 | Method for recommending 3d printing according to slicing direction in cloud environment, server and computing device |
WO2016153047A1 (en) * | 2015-03-26 | 2016-09-29 | 学校法人慶應義塾 | Three-dimensional object-manufacturing apparatus, three-dimensional object-manufacturing method, and program |
JP2016182745A (en) * | 2015-03-26 | 2016-10-20 | 学校法人慶應義塾 | Three-dimensional object production device, method and program |
US10507616B2 (en) | 2015-03-26 | 2019-12-17 | Keio University | Three-dimensional object-manufacturing apparatus, three-dimensional object-manufacturing method, and program |
WO2016171717A1 (en) * | 2015-04-24 | 2016-10-27 | Hewlett-Packard Development Company, L.P. | Three-dimensional object representation |
CN107209958A (en) * | 2015-04-24 | 2017-09-26 | 惠普发展公司,有限责任合伙企业 | Three dimensional object is represented |
US20180099496A1 (en) * | 2015-04-24 | 2018-04-12 | Hewlett-Packard Development Company, L.P. | Three-dimensional object representation |
US10569527B2 (en) * | 2015-04-24 | 2020-02-25 | Hewlett-Packard Development Company, L.P. | Three-dimensional object presentation |
WO2017084369A1 (en) * | 2015-11-18 | 2017-05-26 | 深圳市艾科赛龙科技股份有限公司 | Osteocyte culturing method |
CN105469442A (en) * | 2015-12-30 | 2016-04-06 | 武汉金玺银杏工业设计有限责任公司 | Printing method and apparatus |
US11254901B2 (en) | 2016-07-12 | 2022-02-22 | Deka Products Limited Partnership | System and method for printing tissue |
US10782217B2 (en) | 2016-07-12 | 2020-09-22 | Deka Products Limited Partnership | System and method for applying force to a device |
US11543336B2 (en) | 2016-07-12 | 2023-01-03 | Deka Products Limited Partnership | System and method for applying force to a device |
US11299705B2 (en) | 2016-11-07 | 2022-04-12 | Deka Products Limited Partnership | System and method for creating tissue |
US11939566B2 (en) | 2016-11-07 | 2024-03-26 | Deka Products Limited Partnership | System and method for creating tissue |
US12024701B2 (en) | 2016-11-07 | 2024-07-02 | Deka Products Limited Partnership | System and method for creating tissue |
CN106683550A (en) * | 2017-01-10 | 2017-05-17 | 首都医科大学附属北京友谊医院 | Ear structure model manufacturing method and device thereof |
US20190039322A1 (en) * | 2017-01-19 | 2019-02-07 | Hewlett-Packard Development Company, L.P. | Build unit control |
US10926474B2 (en) * | 2017-01-19 | 2021-02-23 | Hewlett-Packard Development Company, L.P. | Build unit control |
US11530380B2 (en) | 2017-07-12 | 2022-12-20 | Deka Products Limited Partnership | System and method for transferring tissue |
US11939564B2 (en) | 2017-07-12 | 2024-03-26 | Deka Products Limited Partnership | System and method for transferring tissue |
CN108595893A (en) * | 2018-05-16 | 2018-09-28 | 电子科技大学 | A kind of three-dimensional mechanical Modal Analysis analogy method based on three layers of pretreatment |
Also Published As
Publication number | Publication date |
---|---|
US8817332B2 (en) | 2014-08-26 |
US20140025190A1 (en) | 2014-01-23 |
US20140330418A1 (en) | 2014-11-06 |
US8579620B2 (en) | 2013-11-12 |
US20120224755A1 (en) | 2012-09-06 |
US20150088291A1 (en) | 2015-03-26 |
US20150084956A1 (en) | 2015-03-26 |
US20160207257A9 (en) | 2016-07-21 |
US20160210779A9 (en) | 2016-07-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8817332B2 (en) | Single-action three-dimensional model printing methods | |
US20080030497A1 (en) | Three dimensional modeling of objects | |
CN105608728B (en) | Semantic medical image is to 3D printing anatomical structure | |
Sun et al. | Recent development on computer aided tissue engineering—a review | |
Zachow et al. | 3D reconstruction of individual anatomy from medical image data: Segmentation and geometry processing | |
CN110191806A (en) | Towards image segmentation 3 D-printing method and system | |
WO2014178705A1 (en) | A method for manufacturing a three-dimensional anatomical structure | |
Xin et al. | Image fusion in craniofacial virtual reality modeling based on CT and 3dMD photogrammetry | |
He et al. | A method in the design and fabrication of exact-fit customized implant based on sectional medical images and rapid prototyping technology | |
Shah et al. | Setting up 3D printing services for orthopaedic applications: a step-by-step guide and an overview of 3Dbiosphere | |
Zhao et al. | A frame of 3D printing data generation method extracted from CT data | |
da Silva et al. | Scaffold informatics and biomimetic design: three-dimensional medical reconstruction | |
Lerebours et al. | A voxel-based method for designing a numerical biomechanical model patient-specific with an anatomical functional approach adapted to additive manufacturing | |
Grif et al. | Planning technology for neurosurgical procedures by using a software platform to create an optima configuration of customized titanium implants | |
Shivanna et al. | Diarthrodial joint contact models: finite element model development of the human hip | |
US20230059287A1 (en) | Design and Manufacturing Process to Create Modular Surgical Simulators for Training | |
Ward et al. | Creating a human phantom for the virtual human program | |
Shi et al. | Fabrication of anatomically correct head phantoms Based on processing of CT images with MATLAB and rapid prototyping | |
Uma Maheshwaraa et al. | Three-dimensional reconstruction and rapid prototyping of femur bone using multiple digital X-rays | |
Thakur et al. | Optimization of Direct Slicing Process Using DICOMS for Additive Manufacturing | |
Dotremont | From medical images to 3D model: processing and segmentation | |
Pugalendhi et al. | 3D printed patient-specific bone models for anatomy education from medical imaging | |
Singh et al. | Fabrication of Patient Specific Distal Femur with Additive Manufacturing | |
Kontodina | Digital Fabrication of Patient Specific 3D-Printed Medical Models | |
Prat et al. | Segmentation and 3D Printing of Anatomical Models from CT Angiograms |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |